Cross-modal data augmentation

Discriminative Cross-Modal Data Augmentation for Medical Imaging Applications

This is a research project that aim to mitigate the data deficiency issue via cross-modality data augmentation in the medical imaging domain. Here’s the Paper Arxiv link: Discriminative Cross-Modal Data Augmentation for Medical Imaging Applications.

Abstract

While deep learning methods have shown great success in medical image analysis, they require a number of medical images to train. Due to data privacy concerns and unavailability of medical annotators, it is oftentimes very difficult to obtain a lot of labeled medical images for model training. In this paper, we study cross-modality data augmentation to mitigate the data deficiency issue in the medical imaging domain. We propose a discriminative unpaired image-to-image translation model which translates images in source modality into images in target modality where the translation task is conducted jointly with the downstream prediction task and the translation is guided by the prediction. Experiments on two applications demonstrate the effectiveness of our method.

Methodology

The architecture of our model. The source modality and target modality is CT and MRI respectively. For clarity, we omit the adversarial loss in translating CTs to MRIs.

Results

Comparison between our method and baselines.
On the left, some examples of translating MRIs to CTs by CycleGAN and our method. As shown in the regions marked with ovals, our method can better preserve fine-grained details in the translated images than CycleGAN. On the right, examples of translating MRI and PET images into CTs.

Citation

@article{yang2020discriminative,
  title={Discriminative Cross-Modal Data Augmentation for Medical Imaging Applications},
  author={Yang, Yue and Xie, Pengtao},
  journal={arXiv preprint arXiv:2010.03468},
  year={2020}
}