CC BY-NC-ND 4.0 · Yearb Med Inform 2020; 29(01): 129-138
DOI: 10.1055/s-0040-1702009
Section 4: Sensor, Signal and Imaging Informatics
Survey
Georg Thieme Verlag KG Stuttgart

Advancing Medical Imaging Informatics by Deep Learning-Based Domain Adaptation

Anirudh Choudhary*
1   Department of Computational Science and Engineering, Georgia Institute of Technology, GA, USA
,
Li Tong*
2   Department of Biomedical Engineering, Georgia Institute of Technology and Emory University, GA, USA
,
Yuanda Zhu
3   School of Electrical and Computer Engineering, Georgia Institute of Technology, GA, USA
,
May D. Wang
1   Department of Computational Science and Engineering, Georgia Institute of Technology, GA, USA
2   Department of Biomedical Engineering, Georgia Institute of Technology and Emory University, GA, USA
3   School of Electrical and Computer Engineering, Georgia Institute of Technology, GA, USA
› Author Affiliations
Further Information

Correspondence to

May D. Wang
Whitaker Building Room 4106
313 Ferst Drive, Atlanta, GA, USA, 30332
Phone: +1 404 385 2954   
Fax: +1 404 894 4243   

Publication History

Publication Date:
21 August 2020 (online)

 

Summary

Introduction: There has been a rapid development of deep learning (DL) models for medical imaging. However, DL requires a large labeled dataset for training the models. Getting large-scale labeled data remains a challenge, and multi-center datasets suffer from heterogeneity due to patient diversity and varying imaging protocols. Domain adaptation (DA) has been developed to transfer the knowledge from a labeled data domain to a related but unlabeled domain in either image space or feature space. DA is a type of transfer learning (TL) that can improve the performance of models when applied to multiple different datasets.

Objective: In this survey, we review the state-of-the-art DL-based DA methods for medical imaging. We aim to summarize recent advances, highlighting the motivation, challenges, and opportunities, and to discuss promising directions for future work in DA for medical imaging.

Methods: We surveyed peer-reviewed publications from leading biomedical journals and conferences between 2017-2020, that reported the use of DA in medical imaging applications, grouping them by methodology, image modality, and learning scenarios.

Results: We mainly focused on pathology and radiology as application areas. Among various DA approaches, we discussed domain transformation (DT) and latent feature-space transformation (LFST). We highlighted the role of unsupervised DA in image segmentation and described opportunities for future development.

Conclusion: DA has emerged as a promising solution to deal with the lack of annotated training data. Using adversarial techniques, unsupervised DA has achieved good performance, especially for segmentation tasks. Opportunities include domain transferability, multi-modal DA, and applications that benefit from synthetic data.


#

1 Introduction

Medical imaging informatics utilizes digital imaging processing and machine learning (ML) to improve the efficiency, accuracy, and reliability of imaging-based diagnosis [1]. During the past few years, medical imaging informatics has made remarkable progress due to the increasing availability of data and the rapid development of deep learning (DL) techniques [2]. However, fundamental challenges hinder the effective deployment of deep learning models to clinical settings. Annotated medical datasets are limited due to the tedious labeling process [2] and are not easily shared due to privacy concerns [3] [4]. While multicenter datasets can increase the amount of annotated data, these datasets suffer from heterogeneity due to varying hospital procedures and diverse patient populations [5] [6]. Due to a distribution shift (also known as domain-shift) between the available training dataset and the dataset encountered in clinical practice, pre-trained models trained by one dataset may fail for another dataset.

1.1 What is Transfer Learning and Domain Adaptation

Transfer learning (TL) [7] is a technique that applies knowledge learned from one domain and one task to another related domain and/or another task, when there is insufficient labeled data for traditional supervised learning. For medical imaging, a domain usually refers to images or features, while the task refers to segmentation, classification, etc. Mathematically, X and Y being random variables, where X is d-dimensional feature space with marginal probability distribution p(X) and Y is a labeled vector with conditional probability distribution p(Y|X), we use D ={X, p(X)} to represent domain and T={Y, p(Y|X)} to represent task, where p is learnt using a function (e.g. neural network). If both source (DS) and target domains (DT) are similar, i.e., DS ~ DT, then DS and DT can use the same ML model for similar tasks (TS~TT). However, if DS≠DT or TS≠TT, the ML model trained on the source domain might have decreased performance on the target domain (DT). TL can be categorized into three types based on the relationships between domains and/or tasks:

  1. Inductive TL requires some labeled data. While the two domains may or may not differ (DS~DT or DS≠DT), the target and source tasks are different (TS≠TT), for e.g. 3D organ reconstruction across multiple anatomies;

  2. Transductive TL requires labeled source data and unlabeled target data with related domains (DS~DT) and same tasks (TS=TT), while the marginal probability distributions differ (p(XS)≠p(XT)), for e.g., lung tumor detection across X-Ray and computed tomography images;

  3. Unsupervised TL does not require labeled data in any domain and has different tasks (TS≠TT), for e.g., classifying cancer for different anatomies using unlabeled histology images.

Domain Adaptation (DA) is a transductive TL approach that aims to transfer knowledge across domains by learning domain-invariant transformations, which align the domain distributions (see [Figure 1-b]). DA assumes that the source data is labeled, while the target domain can be (a) fully labeled data (i.e., supervised setting); (b) a small set of labeled data (i.e., semi-supervised setting); or (c) completely unlabeled data (i.e., unsupervised setting).

Zoom Image
Fig. 1 a) Transfer learning (TL) and its different types; b) Overview of domain adaptation; c) Organization of this survey paper.

#

1.2 Using Domain Adaptation to Improve Model Training in Medical Imaging

In biomedical imaging, due to the existence of multi-modality imaging (e.g., magnetic resonance imaging (MRI) and positron emission tomography (PET)), DA has advantages over conventional semi-supervised learning or unsupervised learning. Cross-modal DA transfers labels between distinct, but somewhat related, image modalities (e.g., MRI and computed tomography (CT)). Single-modality DA adapts different image distributions within the same modality [8] ([Figure 2]).

Zoom Image
Fig. 2 a) Summary of domain adaptation methodologies employed in medical imaging; b) Different scenarios encountered in cross-modality [16] [28] and single-modality [25] [29] domain adaptation.

1.2.1 Challenge of Limited Training Data

Developing accurate DL models requires large scale training data covering a wide range of input variations. However, in biomedical imaging, due to concerns over patient privacy [3] [4] and lack of manual annotation of images by clinical experts [9], few well-labeled datasets are available for training. This situation is worse for rare diseases, where a low number of positive cases lead to significantly unbalanced datasets [10].

DA can mitigate the lack of well-annotated data by augmenting target domain data, either by generating synthetic labeled images from source images or aligning source and target image features and training a task network on them [11] [12] [112]. For example, MRI achieves higher resolution for soft tissue imaging compared to CT [13]. As such, MRI is preferred for neuroimaging, and brain MRI annotations are easily accessible. On the other hand, CT imaging is fast and less expensive and may be preferred in trauma situations [14]. Thus, through DA, annotated MRI scans from historical subjects can be combined with CT to reduce the number of image acquisitions needed. As another example, Hematoxylin and Eosin (H&E) stained images are widely available, while immunohistochemistry (IHC) images, which clearly highlight nuclei via specific biomarkers [15], are not. DA methods can translate multi-stained H&E-stained images to the IHC domain, making nuclei detection easier [16].


#

1.2.2 Challenge of Dataset Variations

To train robust DL models, many studies rely on images aggregated from multiple institutes such as NCI/NIH The Cancer Genome Atlas (TCGA) [17] and Stanford’s large chest radiograph dataset, CheXpert [18]. The data in these repositories are heterogeneous due to varying hospital processes (image acquisition platforms or data preparation protocols), different demographics of patient populations (ethnicity, gender, age), or different pathological conditions [5]. Specifically, pathology images have stain variations [19] while MRIs are susceptible to varying magnetic fields and contrast agents [20]. Such intra- or inter-dataset variations cause the training and test dataset to have different distributions, resulting in a domain-shift which impacts model generalization [21] [22]. Diversifying the training data by creating larger datasets is a possible solution, but recent medical imaging studies [21] [23] [24] have shown that it does not guarantee improved generalization. DA methods try to minimize the dataset variation, while retaining the distinguishing aspects for task classifier, and have been shown to generalize well in image segmentation tasks for multiple modalities [25] [26] [27].

The organization of the survey paper is illustrated in [Figure 1-c]. Section 2 introduces our survey methodology in identifying and selecting relevant medical imaging studies. Section 3 presents various DL-based frameworks in the DA literature and the current best practices for medical imaging. Section 4 summarizes the current DA challenges and future opportunities.


#
#
#

2 Materials and Methods

In this survey, we examined publications between 2017-2020. We considered the proceedings of leading peer-reviewed journals and conferences, including IEEE Transactions on Medical Imaging, the International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI), Medical Image Analysis (Elsevier), IEEE International Symposium on Biomedical Imaging (ISBI), Conference on Computer Vision and Pattern Recognition (CVPR), Association for the Advancement of Artificial Intelligence (AAAI), and the International Conference on Medical Imaging with Deep Learning (MIDL). Additionally, we identified a few relevant works from arXiV and PubMed, which were not found in review proceedings. Our search keywords included ‘Domain Adaptation’, ‘Transfer Learning’,’ Cross Modality’, ‘Multimodal’ and ‘Medical Image Adaptation’, and ‘Medical Images’. We found that radiology and pathology were the most common application areas (characteristics of our results are illustrated in [Figure 3-a] and [Figure 3-c]). Cross-modality segmentation is observed more extensively in radiology compared to other areas. MRI, CT, and PET maintain better relative morphological-consistency of organs and provide complementary information for disease detection [30]. On the other hand, histopathology with smaller objects, such as nuclei, are prone to artifacts during cross-modal translation [31].

Zoom Image
Fig. 3 Categorization of medical imaging DA publications as per a) imaging modality; b) anatomy; c) learning scenarios

#

3 Deep Learning-based Domain Adaptation

DL-based DA is achieved using various representation learning strategies such as aligning the domain distributions, learning a mapping between domains, separating normalization statistics, and ensemble-based approaches [32] [33] [34]. As shown in [Figure 2-a], there are two families of DA approaches for medical imaging: (a) Domain Transformation (DT-DA) translates images from one domain to the other domain, so that the obtained models can be directly applied to all images, and (b) Latent Feature Space Transformation (LFST-DA) aligns images from both domains in a common hidden feature space to train the task model on top of the hidden features. These two approaches can work together to improve adaptation performance by preserving finer semantic details [35] [36]. We have summarized the application of these DA methods in medical imaging in [Table 1].

Table 1

Summary of DA studies in medical imaging categorized by DA methodology, task, modality, anatomy, and learning scenarios (S: Segmentation; C: Classification; 3DR: 3D Reconstruction).

3.1 Domain Transformation in Domain Adaptation

DT-DA translates images from one domain to another domain (i.e., image-to-image translation [37]). Such translation is typically done using generative models (e.g., generative adversarial networks (GANs)) [38], which achieve pixel-level mapping by learning the translation at a semantic level.The translation direction is usually decided by the relative ease of translation and modeling in a modality [39]. For example, Dou et al., [36] observed lower performance in adapting CT to MRI for cardiac images, since cardiac MRI is more challenging to segment. The task networks are trained independently or jointly, with the image-translation network, using the labeled source images [35]. DT-DA performs alignment in the image space instead of the latent feature space, leading to better interpretability through visual inspection of synthesized images [40], enforcing semantic consistency, and preserving low-level appearance aspects using shape-consistency [41] and structural-similarity constraints [42].

3.1.1 Unidirectional Translation

Unidirectional translation maps images from the source domain to the target domain or vice versa using GANs (e.g., vanilla GAN and conditional GAN (cGAN)) [43]. Compared with vanilla GAN, cGAN conditions the training of the generator and discriminator on extra information such as the class label. Yoo et al., [44] proposed pixel-level domain transfer using cGAN with a domain discriminator. Liu and Tuzel [45] utilized GANs coupled with shared weights to generate paired synthetic source and target images sharing high-level abstraction. Bousmalis et al., [40] leveraged cGAN with the content-similarity loss to generate realistic target images and jointly trained the GAN discriminator with the task network.

Unidirectional translation has been applied to remove dataset variations. For example, Bentaieb et al., [29] designed a stain normalization approach, using a task conditional GAN to translate H&E images to a reference stain. Madani et al., [46] proposed a semi-supervised approach for cardiac abnormality classification using using GAN discriminator for abnormality classification in minimally labeled X-ray images, and showed that the adversarial loss could reduce domain overfitting. Mahmood et al., [39] translated real endoscopy images to graphically-rendered synthetic colon images with ground-truth annotations, for depth-estimation during surgical navigation. Unidirectional translation has also been applied to cross-modality scenario. For instance, Zhao et al., [47] proposed a modified U-Net to translate paired brain CT to MRI.


#

3.1.2 Bidirectional Translation

Bidirectional image translation (also known as reconstruction-based DT) leverages two GANs, constraining the mapping space by enforcing semantic-consistency between the original and reconstructed images. CycleGAN, by Zhu et al. [48], is one of the most popular architectures for bidirectional translation. CycleGAN utilizes cycle-consistency to constrain the translation mapping and improve the quality of generated images. CycleGAN has been expanded to handle larger domain shifts with semantic-consistency loss functions (CyCADA [35]), multi-domain translation (StarGAN [49]), and translation between two domains with multi-modal conditional distributions (MUNIT [50]). In supervised learning, bidirectional translation expands the training data to make the segmentation task model robust. The translation and segmentation network can be trained either independently (two stages) or jointly. Zhang et al., [51] presented a one-stage framework with an additional shape-consistency loss in CycleGAN to achieve better segmentation masks and lower failures. Chartsias et al., [11] used a two-stage framework to segment MRI images using CT images. Cai et al., [52] combined segmentation loss on generated images as an additional shape constraint for 3D translation and leveraged MRI for pancreas segmentation in CT images. In the unsupervised setting, image translation is used to create labeled data for the target domain. Huo et al., [12] proposed a joint optimization approach for the synthesis and segmentation of CT images using labeled MRI. Their framework achieved comparable performance in comparison to the fully labeled case.

There are a few observations about GANs: (a) CycleGAN does not guarantee consistent translation of minor anatomical structures and boundaries [53], and thus needs additional constraints like gradient [53] and shape consistency [51]. For instance, Jiang et al., [54] incorporated tumor-shape and feature-based losses to preserve tumors while translating CT data to MRI data; (b) Attention networks can account for varying transferability of different image regions [55]. For instance, Liu et al., [56] proposed a novel attention-based U-Net [57] as a GAN generator to translate hard-to-generate textured regions from MRI to CT. For alternate scenarios such as 3D-2D, paired images, or semi-supervised DA-DT, Zhang et al., [51] segmented X-ray images by using synthetic X-ray images created from accessible 3D CT annotations. Nguyen et al., [58] used semi-supervised DA with paired CT images to constrain CycleGAN to generate more realistic images. Pan et al., [30] leveraged MRI to generate missing PET images for patients for Alzheimer’s disease diagnosis. Chen et al., [59] proposed state-of-the-art unsupervised segmentation method using bidirectional DA-DT between MRI and CT, combining CycleGAN with shared feature encoder layers between domains. Their method resembled CyCADA [41] and showed the efficacy of combining DT with feature-based alignment; (c) DA-DT can be used for single-modality medical imaging. Chen et al., [28] leveraged a CycleGAN with semantic-aware adversarial loss to perform lung segmentation across different chest X-ray datasets.


#
#

3.2 Latent Feature Space Transformation in Domain Adaptation

Unlike the image-to-image translation in DT-DA, the LFST-DA transforms the source domain and target domain images to a shared latent feature space to learn a domain-invariant feature representation. The goal is to minimize domain-specific information while preserving the task-related information. The LFST-DA can be trained in an unsupervised fashion to obtain a domain-invariant representation, or in a concurrent manner where the representation network and the task network (e.g., image classification network) are trained simultaneously to improve the performance. LFST-DA is used in three basic implementations: divergence minimization [60] [61] [62] [63] [64], adversarial training [65] [66] [67] [68], and cross-domain reconstruction [69] [70]. Compared to DT-DA, LSFT-DA is more computationally efficient because it focuses on translating relevant information only instead of the complete image [34]. Also, feature-based domain alignment outperforms DT-DA by preserving task-critical features [35].

3.2.1 Divergence Minimization

A simple approach to learn domain-invariant features and remove distribution-shift is to minimize some divergence criterion between source and target data distributions. Common choices include maximum mean discrepancy (MMD) [60], correlation alignment (CORAL) [61] [63], contrastive domain discrepancy (CDD) [64], and Wasserstein distance [62]. MMD, CORAL, and Wasserstein distances are class-agnostic divergence metrics and do not discriminate class labels when aligning samples. CDD-based DA aligns samples based on their labels, by minimizing the intra-class discrepancy and maximizing the inter-class discrepancy. MMD and CORAL are two of the most utilized divergence metrics that match the first-order moment (mean) and the second-order moment (covariance) of distributions. However, the represented hidden features can be complicated in the real world and may not be fully characterized by mean and covariance. Wasserstein distance aligns feature distributions between domains via optimal transport theory. Compared to the adversarial-based approaches, divergence-based DA has not been as widely explored in medical imaging. For cross-modality DA, Zhu et al., [71] utilized maximum mean discrepancy to map MR and PET images to a common space to mitigate missing data. Several works have used same-modality DA to mitigate dataset variations in X-ray [72], retinal fundus [73], and electron microscopy images [74].


#

3.2.2 Adversarial Training

Instead of minimizing a divergence metric, adversarial methods train a discriminator, typically a separate network, in an adversarial fashion against the feature encoder network. The goal of the feature network is to learn a latent representation such that the discriminator is unable to identify the input sample domain from the representation. For medical imaging, feature-based adversarial domain adaptation has been widely utilized for various applications. For example, in cross-modality adaptation, Zhang et al., [75] applied a domain discriminator to adapt models trained for pathology images to microscopy images. LSFT-DA is also used for single-modality adaptation to overcome dataset variations in pathology images, MR images, and ultrasound images. For example, Lafarge et al., [76] have utilized a domain discriminator to mitigate the color variations of histopathology images for mitosis detection in breast cancer. Kamnitsas et al., [77] have applied a domain discriminator to MR images from different scanners and imaging protocols to improve the brain lesion segmentation performance.


#

3.2.3 Reconstruction-based Adaptation

The reconstruction-based adaptation maximizes the inter-domain similarity by encoding images from each domain to reconstruct images in the other domain. The reconstruction network (decoder) performs feature alignment by recreating the feature extractor’s input while the feature extractor (encoder) transforms input image into latent representation. Ghifary et al., [70] proposed DRCN for object detection, using only target domain data reconstruction while Bousmalis et al., [69] proposed a domain separation network that extracts image representations in two subspaces: the private domain features and the shared-domain features, the latter being used to reconstruct input image. For medical imaging, reconstruction-based methods are less developed and are usually combined with adversarial learning. For same-modality adaptation, Oliveira et al., [80] have combined image-to-image translation with a feature-based discriminator to mitigate the variations in X-ray images and improve segmentation performance. For cross-modality adaptation, Ouyang et al., [81] combined variational autoencoder (VAE) with adversarial training to adapt MR to CT scans.


#
#
#

4 Challenges and Opportunities

4.1 Domain Selection and Direction of Domain Adaptation

Selecting related domains for effective knowledge transfer is an open-research area in ML. In medical imaging, domains are often selected based on the type of imaging technique (e.g., radiology), anatomy, availability of labeled data, and whether the modalities are complementary for the underlying task [30]. Regarding whether DA could be performed symmetrically across domains, the potential information loss in a particular direction is critical for assessing task performance. For example, for unsupervised DA from CT to MRI, reverse DA may sometimes be needed to preserve tumors [54]. For supervised DA between multiple H&E stained images, Tellez et al., [82] showed that mitosis-detection and cancer tissue classification in a particular color space leads to higher accuracy. Typically, to assess domain relationship and DA direction, it is necessary to use (a) large-scale empirical studies such as [6] [58] exploring bi-directional DA across multiple datasets, (b) a representation-shift metric [24] to roughly quantify the risk of applying learned-representations from a particular domain to a new domain, or (c) multi-source DA [83], which automatically explores latent source domains in multi-source datasets and quantifies the membership of each target sample. However, such experimentation requires extensive benchmarking studies that are lacking in medical imaging.


#

4.2 Transferability of Individual Samples

Most DA studies for medical imaging assume that all samples are equally transferable across two domains. Thus, they focus on globally aligning domain distributions. However, the ability to transfer (or align) varies across clinical samples because of: (a) intra-domain variations (e.g., in multi-modal DA between MRI and CT, each modality can have contrast variations) [75]; (b) noisy annotations due to human subjectivity; (c) target label space being a subset of source label space [84]; and (d) varying transferability among different image regions [55] (e.g., tumors are difficult to translate and could be missed during CT to MRI image-translation [54]). Some samples in the source domain may be less useful and can lead to negative transferring [84], which adversely impacts DA. Selecting relevant samples or reducing the impact of outlier samples using transferability frameworks is a potential solution. Some strategies include weighting samples based on classifier discrepancy [85], down-weighting outlier classes using the classification probability for target data [86], leveraging open-set based optimization [87], and leveraging an attention mechanism [55] to focus on hard-to-transfer samples or using a noise co-adaption layer [88]. Recent medical imaging studies have explored sample selection and transferability assessment using reverse classification accuracy [89], attention-based U-Net [56], and transferable semantic representations [84].


#

4.3 Limitations of Domain Adaptation in Medical Imaging

For medical imaging, most DL-based DA uses adversarial methods, primarily GAN for unsupervised DA. Adversarial methods are prone to errors because the discriminator can be confused, and there is no guarantee that the domain distributions are sufficiently similar [90]. Moreover, the generator in GAN is prone to “hallucinating” content to convince the discriminator that data belongs to the target distribution [91]. As such, CycleGAN could be trained to synthesize tumors in images of healthy patients. Beyond applying consistency constraints during image translation, artifacts which are not directly visible in synthesized images, are also important for consideration. For example, CycleGANs incorporate high-frequency information in the intermediate representation used by the second generator to translate the image back to the source domain [92]. This high frequency information can interfere with downstream tasks.

DT-DA approaches require translating the entire image, increasing the complexity of the models for large-sized medical images like whole slide images. Few studies [28] [36] have compared adversarial DA methods for MRI-CT translation. However, a comprehensive comparison of various feature-based DA approaches is lacking. Future studies could explore combining DT-DA and LFST-DA approaches [59]. Moreover, current frameworks typically focus on source-target domain pair, while many tasks, such as stain normalization in histopathology images, can be multi-domain [93].


#

4.4 Leveraging Synthetic Data

DA for medical imaging can be applied in relatively under-explored applications such as single-view 3D reconstruction [94] or temporal disease analysis [95]. This could benefit image-guided surgery, in which training data is very scarce and difficult to obtain [96]. One way is to leverage synthetic data with ground truth information, adapting it to the real data. This approach has been successfully applied in natural images [97]. Reverse domain adaptation (i.e., translating real data to synthetic data) is also a promising solution. Mahmood et al., [39] generated synthetic endoscopy data with known depth information by using an anatomical colon model and a virtual endoscope. This simulated data was used for 3D reconstruction of real endoscopic images. Pan et al., [30] translated MR data to generate synthetic PET images to infer missing patient scans for temporal analysis of Alzheimer’s disease [30]. Another area that could benefit from synthetic data is skin lesion detection [98].


#
#

5 Conclusions and Future Directions

Deep learning has been widely applied to medical imaging data analysis. However the lack of well-annotated images and the heterogeneity of multi-center medical imaging datasets are two key challenges for DL performance. DA has emerged as an effective approach for minimizing domain-shift and leveraging labeled data from distinct but related domains. FST-DA and DT-DA are two popular approaches to minimize the distribution divergence in multiple medical imaging studies exploring same-modality or cross-modality scenarios. They have proven to achieve good performance, particularly in unsupervised DA settings and organ segmentation tasks. Current approaches are primarily adversarial with domains being selected based on certain heuristics and underlying tasks. Extensive benchmarking studies are needed to quantify the domain relationship for different imaging modalities and to compare adversarial and non-adversarial approaches. Varying sample transferability and multi-modal domains for medical imaging are two other major issues. One strategy is to explore down-weighting or attention-based networks. Also alternative multi-modal frameworks such as MUNIT [30] can be explored. Finally, for certain application areas in medical imaging such as 3D reconstruction and temporal disease analysis where DA is relatively unexplored, synthetic data can be used.


#
#

Acknowledgments

The work was supported in part by grants from the National Science Foundation EAGER Award NSF1651360, Children’s Healthcare of Atlanta and Georgia Tech Partnership Grant, Giglio Breast Cancer Research Fund, Petit Institute Faculty Fellow Award, and Carol Ann and David D. Flanagan Faculty Fellow Research Fund for Professor May D Wang. This work was also supported in part by the scholarship from China Scholarship Council (CSC) under the Grant CSC NO. 201406010343. The content of this article is solely the responsibility of the authors and does not necessarily represent the official views of the NIH.

* Equal Contributing First Authors.


  • References

  • 1 Mendelson DS, Rubin DL. Imaging informatics: essential tools for the delivery of imaging services. Acad Radiol 2013; 20 (10) 1195-212
  • 2 Litjens G, Kooi T, Bejnordi BE, Setio AAA, Ciompi F, Ghafoorian M. et al. A survey on deep learning in medical image analysis. Med Image Anal 2017; 42: 60-88
  • 3 Adler-Milstein J, Jha AK. Sharing clinical data electronically: a critical challenge for fixing the health care system. JAMA 2012; 307 (16) 1695-6
  • 4 Sharma P, Shamout FE, Clifton DA. Preserving Patient Privacy while Training a Predictive Model of In-hospital Mortality. arXiv preprint arXiv:1912.00354; 2019 Dec 1
  • 5 Zhang Y, Wei Y, Zhao P, Niu S, Wu Q, Tan M, Huang J. Collaborative unsupervised domain adaptation for medical image diagnosis. arXiv preprint arXiv:1911.07293; 2019 Nov 17
  • 6 Yao L, Prosky J, Covington B, Lyman K. A strong baseline for domain adaptation and generalization in medical imaging. arXiv preprint arXiv:1904.01638; 2019 Apr 2
  • 7 Pan SJ, Yan Q. A survey on transfer learning. IEEE Trans Knowl Data Eng 2009; 22 (10) 1345-59
  • 8 Tajbakhsh N, Jeyaseelan L, Li Q, Chiang JN, Wu Z, Ding X. Embracing imperfect datasets: A review of deep learning solutions for medical image segmentation. Med Image Anal 2020; 101693
  • 9 Dai C, Mo Y, Angelini E, Guo Y, Bai W. Transfer Learning from Partial Annotations for Whole Brain Segmentation. In: Domain Adaptation and Representation Transfer and Medical Image Learning with Less Labels and Imperfect Data; 2019 Oct 13. p. 199-206
  • 10 Razzak MI, Naz S, Zaib A. Deep learning for medical image processing: Overview, challenges and the future. In: Classification in BioApps; 2018. p. 323-50
  • 11 Chartsias A, Joyce T, Dharmakumar R, Tsaftaris SA. Adversarial image synthesis for unpaired multi-modal cardiac data. In: International workshop on simulation and synthesis in medical imaging; 2017 Sep 10. p. 3-13
  • 12 Huo Y, Xu Z, Bao S, Assad A, Abramson RG, Landman BA. Adversarial synthesis learning enables segmentation without target modality ground truth. In: IEEE 15th International Symposium on Biomedical Imaging, ISBI 2018; 2018. p. 1217-20
  • 13 Chen C, Dou Q, Chen H, Qin J, Heng PA. Synergistic image and feature adaptation: Towards cross-modality domain adaptation for medical image segmentation. In: Proceedings of the AAAI Conference on Artificial Intelligence 2019; 33: 865-72
  • 14 Lee B, Newberg A. Neuroimaging in Traumatic Brain Imaging. NeuroRx 2005; Apr 2 (02) 372-83
  • 15 Gurcan MN, Boucheron LE, Can A, Madabhushi A, Rajpoot NM, Yener B. Histopathological image analysis: A review. IEEE Rev Biomed Eng 2009 Oct 2009; Oct 30 (02) 147-71
  • 16 Brieu N, Meier A, Kapil A, Schoenmeyer R, Gavriel CG, Caie PD. et al. Domain Adaptation-based Augmentation for Weakly Supervised Nuclei Detection. arXiv preprint arXiv:1907.04681; 2019 Jul 10
  • 17 Tomczak K, Czerwińska P, Wiznerowicz M. The Cancer Genome Atlas (TCGA): an immeasurable source of knowledge. Contemp Oncol 2015; 19 (1A): A68-77
  • 18 Irvin J, Rajpurkar P, Ko M, Yu Y, Ciurea-Ilcus S, Chute C. et al. CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison. arXiv preprint arXiv: 1901.07031; 2019 Jan 21
  • 19 Ciompi F, Geessink O, Bejnordi BE, De Souza GS, Baidoshvili A, Litjens G. et al. The importance of stain normalization in colorectal tissue classification with convolutional networks. In: IEEE 14th International Symposium on Biomedical Imaging, ISBI 2017; 2017. p. 160-3
  • 20 Kushibar K, Valverde S, González-Villà S, Bernal J, Cabezas M, Oliver A. et al. Supervised domain adaptation for automatic sub-cortical brain structure segmentation with minimal user interaction. Sci Rep 2019; 9 (01) 6742
  • 21 Pooch EHP, Ballester PL, Barros RC. Can we trust deep learning models diagnosis? The impact of domain shift in chest radiograph classification. arXiv preprint arXiv: 1909.01940; 2019 Sep 3
  • 22 Lampert T, Merveille O, Schmitz J, Forestier G, Feuerhake F, Wemmert C. Strategies for training stain invariant CNNs. In: IEEE 16th International Symposium on Biomedical Imaging, ISBI 2019; p. 905-9
  • 23 AlBadawy EA, Saha A, Mazurowski MA. Deep learning for segmentation of brain tumors: Impact of cross-institutional training and testing. Med Phys 2018; Mar 45 (03) 1150-8
  • 24 Stacke K, Eilertsen G, Unger J, Lundström C. A Closer Look at Domain Shift for Deep Learning in Histopathology. arXiv preprint arXiv: 1909.11575; 2019 Sep 25
  • 25 Yang X, Dou H, Li R, Wang X, Bian C, Li S. et al. Generalizing Deep Models for Ultrasound Image Segmentation. In: Medical Image Computing and Computer Assisted Intervention, MICCAI 2018; p. 497-505
  • 26 Degel MA, Navab N, Albarqouni S. Domain and Geometry Agnostic CNNs for Left Atrium Segmentation in 3D Ultrasound. ArXiv180500357 Cs. 2018; 11073: 630-7
  • 27 Li H, Loehr T, Wiestler B, Zhang J, Menze B. e-UDA: Efficient Unsupervised Domain Adaptation for Cross-Site Medical Image Segmentation. arXiv preprint arXiv: 2001.09313; 2020 Jan 25
  • 28 Chen C, Dou Q, Chen H, Heng PA. Semantic-aware generative adversarial nets for unsupervised domain adaptation in chest X-ray segmentation. In: International Workshop on Machine Learning in Medical Imaging 2018. p. 143-51
  • 29 Bentaieb A, Hamarneh G. Adversarial stain transfer for histopathology image analysis. IEEE Trans Med Imaging 2018; Mar 37 (03) 792-802
  • 30 Pan Y, Liu M, Lian C, Zhou T, Xia Y, Shen D. Synthesizing missing PET from MRI with cycle-consistent generative adversarial networks for Alzheimer’s disease diagnosis. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2018. p. 455-63
  • 31 Ehteshami Bejnordi B, Litjens G, Timofeeva N, Otte-Holler I, Homeyer A, Karssemeijer N. et al. Stain specific standardization of whole-slide histopathological images. IEEE Trans Med Imaging 2016; Feb; 35 (02) 404-15
  • 32 Csurka G. Domain adaptation for visual applications: A comprehensive survey. arXiv preprint arXiv:1702.05374; 2017 Feb 17
  • 33 Wilson G, Cook DJ. A Survey of Unsupervised Deep Domain Adaptation. arXiv preprint arXiv:1812.02849; 2018 Dec 6
  • 34 Wang M, Deng W. Deep visual domain adaptation: A survey. Neurocomputing 2018; 312: 135-53
  • 35 Hoffman J, Tzeng E, Park T, Zhu JY, Isola P, Saenko K. et al. CyCADA: Cycle-Consistent Adversarial Domain Adaptation. In: International Conference on Machine Learning 2018. p. 1989-98
  • 36 Dou Q, Ouyang C, Chen C, Chen H, Glocker B, Zhuang X. et al. Pnp-adanet: Plug-and-play adversarial domain adaptation network with a benchmark at cross-modality cardiac segmentation. arXiv preprint arXiv:1812.07907; 2018 Dec 19
  • 37 Isola P, Zhu JY, Zhou T, Efros AA. Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition 2017. p. 1125-34
  • 38 Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A. et al. Generative adversarial nets. In: Advances in neural information processing systems 2014. p. 2672-80
  • 39 Mahmood F, Chen R, Durr NJ. Unsupervised reverse domain adaptation for synthetic medical images via adversarial training. IEEE Trans Med Imaging 2018; Dec; 37 (12) 2572-81
  • 40 Bousmalis K, Silberman N, Dohan D, Erhan D, Krishnan D. Unsupervised pixel-level domain adaptation with generative adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition 2017. p. 3722-31
  • 41 Zhang Z, Yang L, Zheng Y. Translating and segmenting multimodal medical volumes with cycle-and shape-consistency generative adversarial network. In: Proceedings of the IEEE conference on computer vision and pattern Recognition 2018. p. 9242-9251
  • 42 de Bel T, Hermsen M, Kers J, Van Der Laak J, Litjens G. Stain-transforming cycle-consistent generative adversarial networks for improved segmentation of renal histopathology. In: Proceedings of the 2nd International Conference on Medical Imaging with Deep Learning; Proceedings of Machine Learning Research 2019 May 24. p. 151-63
  • 43 Mirza M, Osindero S. Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784; 2014 Nov 6
  • 44 Yoo D, Kim N, Park S, Paek AS, Kweon IS. Pixel-level domain transfer. In: European Conference on Computer Vision 2016. p. 517-32
  • 45 Liu MY, Tuzel O. Coupled generative adversarial networks. In Advances in neural information processing systems 2016. p. 469-77
  • 46 Madani A, Moradi M, Karargyris A, Syeda-Mahmood T. Semi-supervised learning with generative adversarial networks for chest X-ray classification with ability of data domain adaptation. In: IEEE 15th International Symposium on Biomedical Imaging, ISBI 2018. p. 1038-42
  • 47 Zhao C, Carass A, Lee J, He Y, Prince JL. Whole brain segmentation and labeling from CT using synthetic MR images. In: International Workshop on Machine Learning in Medical Imaging 2017. p. 291-8
  • 48 Zhu JY, Park T, Isola P, Efros AA. Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE international conference on computer vision 2017. p. 2223-32
  • 49 Choi Y, Choi M, Kim M, Ha JW, Kim S, Choo J. StarGAN: Unified generative adversarial networks for multi-domain image-to-image translation. In: Proceedings of the IEEE conference on computer vision and pattern recognition 2018. p. 8789-97
  • 50 Huang X, Liu MY, Belongie S, Kautz J. Multimodal unsupervised image-to-image translation. In: Proceedings of the European Conference on Computer Vision, ECCV 2018. p. 172-89
  • 51 Zhang Y, Miao S, Mansi T, Liao R. Task driven generative modeling for unsupervised domain adaptation: Application to X-ray image segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2018. p. 599-607
  • 52 Cai J, Zhang Z, Cui L, Zheng Y, Yang L. Towards cross-modal organ translation and segmentation: a cycle-and shape-consistent generative adversarial network. Med Image Anal 2019; 52: 174-84
  • 53 Hiasa Y, Otake Y, Takao M, Matsuoka T, Takashima K, Carass A. et al. Cross-modality image synthesis from unpaired data using CycleGAN. In: International workshop on simulation and synthesis in medical imaging 2018. p. 31-41
  • 54 Jiang J, Hu YC, Tyagi N, Zhang P, Rimner A, Mageras GS. et al. Tumor-aware, adversarial domain adaptation from ct to mri for lung cancer segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2018; pp. 777-785
  • 55 Wang X, Li L, Ye W, Long M, Wang J. Transferable attention for domain adaptation. In: Proceedings of the AAAI Conference on Artificial Intelligence 2019; 33: 5345-52
  • 56 Liu X, Wei X, Yu A, Pan Z. Unpaired Data based Cross-domain Synthesis and Segmentation Using Attention Neural Network. In: Asian Conference on Machine Learning 2019. p. 987-1000
  • 57 Oktay O, Schlemper J, Folgoc LL, Lee M, Heinrich M, Misawa K. et al. Attention u-net: Learning where to look for the pancreas. arXiv preprint arXiv:1804.03999; 2018 Apr 11
  • 58 Nguyen H, Luo S, Ramos F. Semi-supervised Learning Approach to Generate Neuroimaging Modalities with Adversarial Training. In: Pacific-Asia Conference on Knowledge Discovery and Data Mining 2020. p. 409-21
  • 59 Chen C, Dou Q, Chen H, Qin J, Heng PA. Unsupervised bidirectional cross-modality adaptation via deeply synergistic image and feature alignment for medical image segmentation. IEEE Trans Med Imaging 2020; 39 (07) 2494-505
  • 60 Rozantsev A, Salzmann M, Fua P. Beyond sharing weights for deep domain adaptation. IEEE transactions Trans Pattern Anal Mach Intell 2019; 41 (04) 801-14
  • 61 Sun B, Saenko K. Deep coral: Correlation alignment for deep domain adaptation. In: European conference on computer vision 2016. p. 443-50
  • 62 Bhushan Damodaran B, Kellenberger B, Flamary R, Tuia D, Courty N. Deepjdot: Deep joint distribution optimal transport for unsupervised domain adaptation. In: Proceedings of the European Conference on Computer Vision (ECCV) 2018. p. 447-63
  • 63 Sun B, Feng J, Saenko K. Return of frustratingly easy domain adaptation. In: Thirtieth AAAI Conference on Artificial Intelligence 2016 Mar 2
  • 64 Kang G, Jiang L, Yang Y, Hauptmann AG. Contrastive adaptation network for unsupervised domain adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2019. p. 4893-902
  • 65 Tzeng E, Hoffman J, Saenko K, Darrell T. Adversarial discriminative domain adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2017. p. 7167-76
  • 66 Ganin Y, Ustinova E, Ajakan H, Germain P, Larochelle H, Laviolette F. et al. Domain-adversarial training of neural networks. The Journal of Machine Learning Research 2016; 17 (01) 2096-30
  • 67 Tzeng E, Hoffman J, Zhang N, Saenko K, Darrell T. Deep domain confusion: Maximizing for domain invariance. arXiv preprint arXiv:1412.3474. 2014 Dec 10
  • 68 Tsai YH, Hung WC, Schulter S, Sohn K, Yang MH, Chandraker M. Learning to adapt structured output space for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2018. p. 7472-81
  • 69 Bousmalis K, Trigeorgis G, Silberman N, Krishnan D, Erhan D. Domain separation networks. In: Advances in neural information processing systems 2016. p. 343-51
  • 70 Ghifary M, Kleijn WB, Zhang M, Balduzzi D, Li W. Deep reconstruction-classification networks for unsupervised domain adaptation. In: European Conference on Computer Vision 2016. p. 597-613
  • 71 Zhu X, Thung KH, Adeli E, Zhang Y, Shen D. Maximum mean discrepancy based multiple kernel learning for incomplete multimodality neuroimaging data. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2017. p. 72-80
  • 72 Venkataramani R, Ravishankar H, Anamandra S. Towards Continuous Domain Adaptation For Medical Imaging. In: IEEE 16th International Symposium on Biomedical Imaging, ISBI 2019. p. 443-6
  • 73 Zhuang J, Chen Z, Zhang J, Zhang D, Cai Z. Domain adaptation for retinal vessel segmentation using asymmetrical maximum classifier discrepancy. In: Proceedings of the ACM Turing Celebration Conference-China 2019. p. 1-6
  • 74 Bermúdez-Chacón R, Altingövde O, Becker C, Salzmann M, Fua P. Visual Correspondences for Unsupervised Domain Adaptation on Electron Microscopy Images. IEEE Trans Med Imaging 2020; 39 (04) 1256-67
  • 75 Zhang Y, Chen H, Wei Y, Zhao P, Cao J, Fan X. et al. From whole slide imaging to microscopy: Deep microscopy adaptation network for histopathology cancer image classification. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2019. p. 360-8
  • 76 Lafarge MW, Pluim JP, Eppenhof KA, Moeskops P, Veta M. Domain-adversarial neural networks to address the appearance variability of histopathology images. In: Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support 2017. p. 83-91
  • 77 Kamnitsas K, Baumgartner C, Ledig C, Newcombe V, Simpson J, Kane A. et al. Unsupervised domain adaptation in brain lesion segmentation with adversarial networks. In: International conference on information processing in medical imaging 2017. p. 597-609
  • 78 Orbes-Arteaga M, Varsavsky T, Sudre CH, Eaton-Rosen Z, Haddow LJ, Sørensen L. , et al. Multi-domain adaptation in brain MRI through paired consistency and adversarial learning. In: Domain Adaptation and Representation Transfer and Medical Image Learning with Less Labels and Imperfect Data 2019. p. 54-62
  • 79 Shanis Z, Gerber S, Gao M, Enquobahrie A. Intramodality Domain Adaptation Using Self Ensembling and Adversarial Training. In: Domain Adaptation and Representation Transfer and Medical Image Learning with Less Labels and Imperfect Data 2019. p. 28-36
  • 80 Oliveira H, Ferreira E, Dos Santos JA. Truly Generalizable Radiograph Segmentation with Conditional Domain Adaptation. IEEE Access 2020 May 1
  • 81 Ouyang C, Kamnitsas K, Biffi C, Duan J, Rueckert D. Data efficient unsupervised domain adaptation for Cross-Modality image segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2019. p. 669-77
  • 82 Tellez D, Litjens G, Bándi P, Bulten W, Bokhorst JM, Ciompi F, van der Laak J. Quantifying the effects of data augmentation and stain color normalization in convolutional neural networks for computational pathology. Med Image Anal 2019 Dec 2019; Dec 1 58: 101544
  • 83 Peng X, Bai Q, Xia X, Huang Z, Saenko K, Wang B. Moment matching for multi-source domain adaptation. In: Proceedings of the IEEE International Conference on Computer Vision 2019. p. 1406-15
  • 84 Dong J, Cong Y, Sun G, Zhong B, Xu X. What Can Be Transferred: Unsupervised Domain Adaptation for Endoscopic Lesions Segmentation. arXiv preprint arXiv:2004.11500; 2020 Apr 24
  • 85 Saito K, Watanabe K, Ushiku Y, Harada T. Maximum classifier discrepancy for unsupervised domain adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2018. p. 3723-32
  • 86 Cao Z, Ma L, Long M, Wang J. Partial adversarial domain adaptation. In: Proceedings of the European Conference on Computer Vision, ECCV 2018. p. 135-50
  • 87 Panareda Busto P, Gall J. Open set domain adaptation. In: Proceedings of the IEEE International Conference on Computer Vision 2017. p. 754-63
  • 88 Goldberger J, Ben-Reuven E. Training deep neural-networks using a noise adaptation layer. In: International Conference on Learning Representations 2017
  • 89 Valindria VV, Lavdas I, Bai W, Kamnitsas K, Aboagye EO, Rockall AG. et al. Domain adaptation for MRI organ segmentation using reverse classification accuracy. arXiv preprint arXiv:1806.00363; 2018 Jun 1
  • 90 Arora S, Ge R, Liang Y, Ma T, Zhang Y. Generalization and equilibrium in generative adversarial nets (gans). In: Proceedings of the 34th International Conference on Machine Learning Volume 70 2017 Aug 6. p. 224-32
  • 91 Martin A, Lon B. Towards principled methods for training generative adversarial networks. In: NIPS 2016 Workshop on Adversarial Training
  • 92 Chu C, Zhmoginov A, Sandler M. Cyclegan, a master of steganography. arXiv preprint arXiv:1712.02950; 2017 Dec 8
  • 93 Shrivastava A, Adorno W, Ehsan L, Ali SA, Moore SR, Amadi BC. et al. Self-Attentive Adversarial Stain Normalization. arXiv preprint arXiv:1909.01963; 2019 Sep 4
  • 94 Wang Y, Zhong Z, Hua J. DeepOrganNet: On-the-Fly Reconstruction and Visualization of 3D/4D Lung Models from Single-View Projections by Deep Deformation Network. IEEE Trans Vis Comput Graph 2019; 26 (01) 960-70
  • 95 Jac Jr CR, Knopman DS, Jagust WJ, Petersen RC, Weiner MW, Aisen PS. et al. Update on hypothetical model of Alzheimer’s disease biomarkers. Lancet Neurology 2013; 12 (02) 207
  • 96 Mahmood F, Durr NJ. Deep learning and conditional random fields-based depth estimation and topographical reconstruction from conventional endoscopy. Med Image Anal 2018 Aug 2018; Aug 1 48: 230-43
  • 97 Shrivastava A, Pfister T, Tuzel O, Susskind J, Wang W, Webb R. Learning from simulated and unsupervised images through adversarial training. In: Proceedings of the IEEE conference on computer vision and pattern recognition 2017. p. 2107-16
  • 98 Ghorbani A, Natarajan V, Coz D, Liu Y. DermGAN: Synthetic Generation of Clinical Skin Images with Pathology. arXiv preprint arXiv:1911.0871
  • 99 Gholami A, Subramanian S, Shenoy V, Himthani N, Yue X, Zhao S. et al. A novel domain adaptation framework for medical image segmentation. In: International MICCAI Brainlesion Workshop 2018. p. 289-98
  • 100 Tang Y, Tang Y, Sandfort V, Xiao J, Summers RM. TUNA-Net: Task-oriented UNsupervised Adversarial Network for Disease Recognition in Cross-Domain Chest X-rays. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2019. p. 431-40
  • 101 Kapil A, Wiestler T, Lanzmich S, Silva A, Steele K, Rebelatto M. et al. DASGAN--Joint Domain Adaptation and Segmentation for the Analysis of Epithelial Regions in Histopathology PD-L1 Images. arXiv preprint arXiv:1906.11118; 2019 Jun 26
  • 102 Yang J, Dvornek NC, Zhang F, Chapiro J, Lin M, Duncan JS. Unsupervised domain adaptation via disentangled representations: Application to cross-modality liver segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2019. p. 255-63
  • 103 Dou Q, Ouyang C, Chen C, Chen H, Glocker B, Zhuang X. et al. Pnp-adanet: Plug-and-play adversarial domain adaptation network with a benchmark at cross-modality cardiac segmentation. arXiv preprint arXiv:1812.07907; 2018 Dec 19
  • 104 Dong N, Kampffmeyer M, Liang X, Wang Z, Dai W, Xing E. Unsupervised domain adaptation for automatic estimation of cardiothoracic ration. In: International Conference on Medical Image Computing and Computer-assisted intervention 2018. p. 544-52
  • 105 Javanmardi M, Tasdizen T. Domain adaptation for biomedical image segmentation using adversarial training. In: IEEE 15th International Symposium on Biomedical Imaging 2018. p. 554-8
  • 106 Novosad P, Fonov V, Collins DL. Unsupervised domain adaptation for the automated segmentation of neuroanatomy in MRI: a deep learning approach. bioRxiv preprint bioRxiv:845537 2019 Jan 1
  • 107 Ren J, Hacihaliloglu I, Singer EA, Foran DJ, Qi X. Adversarial domain adaptation for classification of prostate histopathology whole-slide images. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2018. p. 201-9
  • 108 Wang S, Yu L, Li K, Yang X, Fu CW, Heng PA. Boundary and Entropy-driven Adversarial Learning for Fundus Image Segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2019. p. 102-10
  • 109 Yan W, Wang Y, Xia M, Tao Q. Edge-Guided Output Adaptor: Highly Efficient Adaptation Module for Cross-Vendor Medical Image Segmentation. IEEE Signal Processing Letters 2019; 26 (11) 1593-7
  • 110 Hou X, Liu J, Xu B, Liu B, Chen X, Ilyas M. et al. Dual Adaptive Pyramid Network for Cross-Stain Histopathology Image Segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2019. p. 101-9
  • 111 Zhao H, Li H, Maurer-Stroh S, Gui Y, Deng Q, Cheng L. Supervised segmentation of un-annotated retinal fundus images by synthesis. IEEE Trans Med Imaging 2018; 38 (01) 46-56
  • 112 Tong L, Wu H, Wang MD. CAESNet: Convolutional AutoEncoder based Semi-supervised Network for improving multiclass classification of endomicroscopic images. J Am Med Inform Assoc 2019; Nov; 26 (11) 1286-96

Correspondence to

May D. Wang
Whitaker Building Room 4106
313 Ferst Drive, Atlanta, GA, USA, 30332
Phone: +1 404 385 2954   
Fax: +1 404 894 4243   

  • References

  • 1 Mendelson DS, Rubin DL. Imaging informatics: essential tools for the delivery of imaging services. Acad Radiol 2013; 20 (10) 1195-212
  • 2 Litjens G, Kooi T, Bejnordi BE, Setio AAA, Ciompi F, Ghafoorian M. et al. A survey on deep learning in medical image analysis. Med Image Anal 2017; 42: 60-88
  • 3 Adler-Milstein J, Jha AK. Sharing clinical data electronically: a critical challenge for fixing the health care system. JAMA 2012; 307 (16) 1695-6
  • 4 Sharma P, Shamout FE, Clifton DA. Preserving Patient Privacy while Training a Predictive Model of In-hospital Mortality. arXiv preprint arXiv:1912.00354; 2019 Dec 1
  • 5 Zhang Y, Wei Y, Zhao P, Niu S, Wu Q, Tan M, Huang J. Collaborative unsupervised domain adaptation for medical image diagnosis. arXiv preprint arXiv:1911.07293; 2019 Nov 17
  • 6 Yao L, Prosky J, Covington B, Lyman K. A strong baseline for domain adaptation and generalization in medical imaging. arXiv preprint arXiv:1904.01638; 2019 Apr 2
  • 7 Pan SJ, Yan Q. A survey on transfer learning. IEEE Trans Knowl Data Eng 2009; 22 (10) 1345-59
  • 8 Tajbakhsh N, Jeyaseelan L, Li Q, Chiang JN, Wu Z, Ding X. Embracing imperfect datasets: A review of deep learning solutions for medical image segmentation. Med Image Anal 2020; 101693
  • 9 Dai C, Mo Y, Angelini E, Guo Y, Bai W. Transfer Learning from Partial Annotations for Whole Brain Segmentation. In: Domain Adaptation and Representation Transfer and Medical Image Learning with Less Labels and Imperfect Data; 2019 Oct 13. p. 199-206
  • 10 Razzak MI, Naz S, Zaib A. Deep learning for medical image processing: Overview, challenges and the future. In: Classification in BioApps; 2018. p. 323-50
  • 11 Chartsias A, Joyce T, Dharmakumar R, Tsaftaris SA. Adversarial image synthesis for unpaired multi-modal cardiac data. In: International workshop on simulation and synthesis in medical imaging; 2017 Sep 10. p. 3-13
  • 12 Huo Y, Xu Z, Bao S, Assad A, Abramson RG, Landman BA. Adversarial synthesis learning enables segmentation without target modality ground truth. In: IEEE 15th International Symposium on Biomedical Imaging, ISBI 2018; 2018. p. 1217-20
  • 13 Chen C, Dou Q, Chen H, Qin J, Heng PA. Synergistic image and feature adaptation: Towards cross-modality domain adaptation for medical image segmentation. In: Proceedings of the AAAI Conference on Artificial Intelligence 2019; 33: 865-72
  • 14 Lee B, Newberg A. Neuroimaging in Traumatic Brain Imaging. NeuroRx 2005; Apr 2 (02) 372-83
  • 15 Gurcan MN, Boucheron LE, Can A, Madabhushi A, Rajpoot NM, Yener B. Histopathological image analysis: A review. IEEE Rev Biomed Eng 2009 Oct 2009; Oct 30 (02) 147-71
  • 16 Brieu N, Meier A, Kapil A, Schoenmeyer R, Gavriel CG, Caie PD. et al. Domain Adaptation-based Augmentation for Weakly Supervised Nuclei Detection. arXiv preprint arXiv:1907.04681; 2019 Jul 10
  • 17 Tomczak K, Czerwińska P, Wiznerowicz M. The Cancer Genome Atlas (TCGA): an immeasurable source of knowledge. Contemp Oncol 2015; 19 (1A): A68-77
  • 18 Irvin J, Rajpurkar P, Ko M, Yu Y, Ciurea-Ilcus S, Chute C. et al. CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison. arXiv preprint arXiv: 1901.07031; 2019 Jan 21
  • 19 Ciompi F, Geessink O, Bejnordi BE, De Souza GS, Baidoshvili A, Litjens G. et al. The importance of stain normalization in colorectal tissue classification with convolutional networks. In: IEEE 14th International Symposium on Biomedical Imaging, ISBI 2017; 2017. p. 160-3
  • 20 Kushibar K, Valverde S, González-Villà S, Bernal J, Cabezas M, Oliver A. et al. Supervised domain adaptation for automatic sub-cortical brain structure segmentation with minimal user interaction. Sci Rep 2019; 9 (01) 6742
  • 21 Pooch EHP, Ballester PL, Barros RC. Can we trust deep learning models diagnosis? The impact of domain shift in chest radiograph classification. arXiv preprint arXiv: 1909.01940; 2019 Sep 3
  • 22 Lampert T, Merveille O, Schmitz J, Forestier G, Feuerhake F, Wemmert C. Strategies for training stain invariant CNNs. In: IEEE 16th International Symposium on Biomedical Imaging, ISBI 2019; p. 905-9
  • 23 AlBadawy EA, Saha A, Mazurowski MA. Deep learning for segmentation of brain tumors: Impact of cross-institutional training and testing. Med Phys 2018; Mar 45 (03) 1150-8
  • 24 Stacke K, Eilertsen G, Unger J, Lundström C. A Closer Look at Domain Shift for Deep Learning in Histopathology. arXiv preprint arXiv: 1909.11575; 2019 Sep 25
  • 25 Yang X, Dou H, Li R, Wang X, Bian C, Li S. et al. Generalizing Deep Models for Ultrasound Image Segmentation. In: Medical Image Computing and Computer Assisted Intervention, MICCAI 2018; p. 497-505
  • 26 Degel MA, Navab N, Albarqouni S. Domain and Geometry Agnostic CNNs for Left Atrium Segmentation in 3D Ultrasound. ArXiv180500357 Cs. 2018; 11073: 630-7
  • 27 Li H, Loehr T, Wiestler B, Zhang J, Menze B. e-UDA: Efficient Unsupervised Domain Adaptation for Cross-Site Medical Image Segmentation. arXiv preprint arXiv: 2001.09313; 2020 Jan 25
  • 28 Chen C, Dou Q, Chen H, Heng PA. Semantic-aware generative adversarial nets for unsupervised domain adaptation in chest X-ray segmentation. In: International Workshop on Machine Learning in Medical Imaging 2018. p. 143-51
  • 29 Bentaieb A, Hamarneh G. Adversarial stain transfer for histopathology image analysis. IEEE Trans Med Imaging 2018; Mar 37 (03) 792-802
  • 30 Pan Y, Liu M, Lian C, Zhou T, Xia Y, Shen D. Synthesizing missing PET from MRI with cycle-consistent generative adversarial networks for Alzheimer’s disease diagnosis. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2018. p. 455-63
  • 31 Ehteshami Bejnordi B, Litjens G, Timofeeva N, Otte-Holler I, Homeyer A, Karssemeijer N. et al. Stain specific standardization of whole-slide histopathological images. IEEE Trans Med Imaging 2016; Feb; 35 (02) 404-15
  • 32 Csurka G. Domain adaptation for visual applications: A comprehensive survey. arXiv preprint arXiv:1702.05374; 2017 Feb 17
  • 33 Wilson G, Cook DJ. A Survey of Unsupervised Deep Domain Adaptation. arXiv preprint arXiv:1812.02849; 2018 Dec 6
  • 34 Wang M, Deng W. Deep visual domain adaptation: A survey. Neurocomputing 2018; 312: 135-53
  • 35 Hoffman J, Tzeng E, Park T, Zhu JY, Isola P, Saenko K. et al. CyCADA: Cycle-Consistent Adversarial Domain Adaptation. In: International Conference on Machine Learning 2018. p. 1989-98
  • 36 Dou Q, Ouyang C, Chen C, Chen H, Glocker B, Zhuang X. et al. Pnp-adanet: Plug-and-play adversarial domain adaptation network with a benchmark at cross-modality cardiac segmentation. arXiv preprint arXiv:1812.07907; 2018 Dec 19
  • 37 Isola P, Zhu JY, Zhou T, Efros AA. Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition 2017. p. 1125-34
  • 38 Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A. et al. Generative adversarial nets. In: Advances in neural information processing systems 2014. p. 2672-80
  • 39 Mahmood F, Chen R, Durr NJ. Unsupervised reverse domain adaptation for synthetic medical images via adversarial training. IEEE Trans Med Imaging 2018; Dec; 37 (12) 2572-81
  • 40 Bousmalis K, Silberman N, Dohan D, Erhan D, Krishnan D. Unsupervised pixel-level domain adaptation with generative adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition 2017. p. 3722-31
  • 41 Zhang Z, Yang L, Zheng Y. Translating and segmenting multimodal medical volumes with cycle-and shape-consistency generative adversarial network. In: Proceedings of the IEEE conference on computer vision and pattern Recognition 2018. p. 9242-9251
  • 42 de Bel T, Hermsen M, Kers J, Van Der Laak J, Litjens G. Stain-transforming cycle-consistent generative adversarial networks for improved segmentation of renal histopathology. In: Proceedings of the 2nd International Conference on Medical Imaging with Deep Learning; Proceedings of Machine Learning Research 2019 May 24. p. 151-63
  • 43 Mirza M, Osindero S. Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784; 2014 Nov 6
  • 44 Yoo D, Kim N, Park S, Paek AS, Kweon IS. Pixel-level domain transfer. In: European Conference on Computer Vision 2016. p. 517-32
  • 45 Liu MY, Tuzel O. Coupled generative adversarial networks. In Advances in neural information processing systems 2016. p. 469-77
  • 46 Madani A, Moradi M, Karargyris A, Syeda-Mahmood T. Semi-supervised learning with generative adversarial networks for chest X-ray classification with ability of data domain adaptation. In: IEEE 15th International Symposium on Biomedical Imaging, ISBI 2018. p. 1038-42
  • 47 Zhao C, Carass A, Lee J, He Y, Prince JL. Whole brain segmentation and labeling from CT using synthetic MR images. In: International Workshop on Machine Learning in Medical Imaging 2017. p. 291-8
  • 48 Zhu JY, Park T, Isola P, Efros AA. Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE international conference on computer vision 2017. p. 2223-32
  • 49 Choi Y, Choi M, Kim M, Ha JW, Kim S, Choo J. StarGAN: Unified generative adversarial networks for multi-domain image-to-image translation. In: Proceedings of the IEEE conference on computer vision and pattern recognition 2018. p. 8789-97
  • 50 Huang X, Liu MY, Belongie S, Kautz J. Multimodal unsupervised image-to-image translation. In: Proceedings of the European Conference on Computer Vision, ECCV 2018. p. 172-89
  • 51 Zhang Y, Miao S, Mansi T, Liao R. Task driven generative modeling for unsupervised domain adaptation: Application to X-ray image segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2018. p. 599-607
  • 52 Cai J, Zhang Z, Cui L, Zheng Y, Yang L. Towards cross-modal organ translation and segmentation: a cycle-and shape-consistent generative adversarial network. Med Image Anal 2019; 52: 174-84
  • 53 Hiasa Y, Otake Y, Takao M, Matsuoka T, Takashima K, Carass A. et al. Cross-modality image synthesis from unpaired data using CycleGAN. In: International workshop on simulation and synthesis in medical imaging 2018. p. 31-41
  • 54 Jiang J, Hu YC, Tyagi N, Zhang P, Rimner A, Mageras GS. et al. Tumor-aware, adversarial domain adaptation from ct to mri for lung cancer segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2018; pp. 777-785
  • 55 Wang X, Li L, Ye W, Long M, Wang J. Transferable attention for domain adaptation. In: Proceedings of the AAAI Conference on Artificial Intelligence 2019; 33: 5345-52
  • 56 Liu X, Wei X, Yu A, Pan Z. Unpaired Data based Cross-domain Synthesis and Segmentation Using Attention Neural Network. In: Asian Conference on Machine Learning 2019. p. 987-1000
  • 57 Oktay O, Schlemper J, Folgoc LL, Lee M, Heinrich M, Misawa K. et al. Attention u-net: Learning where to look for the pancreas. arXiv preprint arXiv:1804.03999; 2018 Apr 11
  • 58 Nguyen H, Luo S, Ramos F. Semi-supervised Learning Approach to Generate Neuroimaging Modalities with Adversarial Training. In: Pacific-Asia Conference on Knowledge Discovery and Data Mining 2020. p. 409-21
  • 59 Chen C, Dou Q, Chen H, Qin J, Heng PA. Unsupervised bidirectional cross-modality adaptation via deeply synergistic image and feature alignment for medical image segmentation. IEEE Trans Med Imaging 2020; 39 (07) 2494-505
  • 60 Rozantsev A, Salzmann M, Fua P. Beyond sharing weights for deep domain adaptation. IEEE transactions Trans Pattern Anal Mach Intell 2019; 41 (04) 801-14
  • 61 Sun B, Saenko K. Deep coral: Correlation alignment for deep domain adaptation. In: European conference on computer vision 2016. p. 443-50
  • 62 Bhushan Damodaran B, Kellenberger B, Flamary R, Tuia D, Courty N. Deepjdot: Deep joint distribution optimal transport for unsupervised domain adaptation. In: Proceedings of the European Conference on Computer Vision (ECCV) 2018. p. 447-63
  • 63 Sun B, Feng J, Saenko K. Return of frustratingly easy domain adaptation. In: Thirtieth AAAI Conference on Artificial Intelligence 2016 Mar 2
  • 64 Kang G, Jiang L, Yang Y, Hauptmann AG. Contrastive adaptation network for unsupervised domain adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2019. p. 4893-902
  • 65 Tzeng E, Hoffman J, Saenko K, Darrell T. Adversarial discriminative domain adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2017. p. 7167-76
  • 66 Ganin Y, Ustinova E, Ajakan H, Germain P, Larochelle H, Laviolette F. et al. Domain-adversarial training of neural networks. The Journal of Machine Learning Research 2016; 17 (01) 2096-30
  • 67 Tzeng E, Hoffman J, Zhang N, Saenko K, Darrell T. Deep domain confusion: Maximizing for domain invariance. arXiv preprint arXiv:1412.3474. 2014 Dec 10
  • 68 Tsai YH, Hung WC, Schulter S, Sohn K, Yang MH, Chandraker M. Learning to adapt structured output space for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2018. p. 7472-81
  • 69 Bousmalis K, Trigeorgis G, Silberman N, Krishnan D, Erhan D. Domain separation networks. In: Advances in neural information processing systems 2016. p. 343-51
  • 70 Ghifary M, Kleijn WB, Zhang M, Balduzzi D, Li W. Deep reconstruction-classification networks for unsupervised domain adaptation. In: European Conference on Computer Vision 2016. p. 597-613
  • 71 Zhu X, Thung KH, Adeli E, Zhang Y, Shen D. Maximum mean discrepancy based multiple kernel learning for incomplete multimodality neuroimaging data. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2017. p. 72-80
  • 72 Venkataramani R, Ravishankar H, Anamandra S. Towards Continuous Domain Adaptation For Medical Imaging. In: IEEE 16th International Symposium on Biomedical Imaging, ISBI 2019. p. 443-6
  • 73 Zhuang J, Chen Z, Zhang J, Zhang D, Cai Z. Domain adaptation for retinal vessel segmentation using asymmetrical maximum classifier discrepancy. In: Proceedings of the ACM Turing Celebration Conference-China 2019. p. 1-6
  • 74 Bermúdez-Chacón R, Altingövde O, Becker C, Salzmann M, Fua P. Visual Correspondences for Unsupervised Domain Adaptation on Electron Microscopy Images. IEEE Trans Med Imaging 2020; 39 (04) 1256-67
  • 75 Zhang Y, Chen H, Wei Y, Zhao P, Cao J, Fan X. et al. From whole slide imaging to microscopy: Deep microscopy adaptation network for histopathology cancer image classification. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2019. p. 360-8
  • 76 Lafarge MW, Pluim JP, Eppenhof KA, Moeskops P, Veta M. Domain-adversarial neural networks to address the appearance variability of histopathology images. In: Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support 2017. p. 83-91
  • 77 Kamnitsas K, Baumgartner C, Ledig C, Newcombe V, Simpson J, Kane A. et al. Unsupervised domain adaptation in brain lesion segmentation with adversarial networks. In: International conference on information processing in medical imaging 2017. p. 597-609
  • 78 Orbes-Arteaga M, Varsavsky T, Sudre CH, Eaton-Rosen Z, Haddow LJ, Sørensen L. , et al. Multi-domain adaptation in brain MRI through paired consistency and adversarial learning. In: Domain Adaptation and Representation Transfer and Medical Image Learning with Less Labels and Imperfect Data 2019. p. 54-62
  • 79 Shanis Z, Gerber S, Gao M, Enquobahrie A. Intramodality Domain Adaptation Using Self Ensembling and Adversarial Training. In: Domain Adaptation and Representation Transfer and Medical Image Learning with Less Labels and Imperfect Data 2019. p. 28-36
  • 80 Oliveira H, Ferreira E, Dos Santos JA. Truly Generalizable Radiograph Segmentation with Conditional Domain Adaptation. IEEE Access 2020 May 1
  • 81 Ouyang C, Kamnitsas K, Biffi C, Duan J, Rueckert D. Data efficient unsupervised domain adaptation for Cross-Modality image segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2019. p. 669-77
  • 82 Tellez D, Litjens G, Bándi P, Bulten W, Bokhorst JM, Ciompi F, van der Laak J. Quantifying the effects of data augmentation and stain color normalization in convolutional neural networks for computational pathology. Med Image Anal 2019 Dec 2019; Dec 1 58: 101544
  • 83 Peng X, Bai Q, Xia X, Huang Z, Saenko K, Wang B. Moment matching for multi-source domain adaptation. In: Proceedings of the IEEE International Conference on Computer Vision 2019. p. 1406-15
  • 84 Dong J, Cong Y, Sun G, Zhong B, Xu X. What Can Be Transferred: Unsupervised Domain Adaptation for Endoscopic Lesions Segmentation. arXiv preprint arXiv:2004.11500; 2020 Apr 24
  • 85 Saito K, Watanabe K, Ushiku Y, Harada T. Maximum classifier discrepancy for unsupervised domain adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2018. p. 3723-32
  • 86 Cao Z, Ma L, Long M, Wang J. Partial adversarial domain adaptation. In: Proceedings of the European Conference on Computer Vision, ECCV 2018. p. 135-50
  • 87 Panareda Busto P, Gall J. Open set domain adaptation. In: Proceedings of the IEEE International Conference on Computer Vision 2017. p. 754-63
  • 88 Goldberger J, Ben-Reuven E. Training deep neural-networks using a noise adaptation layer. In: International Conference on Learning Representations 2017
  • 89 Valindria VV, Lavdas I, Bai W, Kamnitsas K, Aboagye EO, Rockall AG. et al. Domain adaptation for MRI organ segmentation using reverse classification accuracy. arXiv preprint arXiv:1806.00363; 2018 Jun 1
  • 90 Arora S, Ge R, Liang Y, Ma T, Zhang Y. Generalization and equilibrium in generative adversarial nets (gans). In: Proceedings of the 34th International Conference on Machine Learning Volume 70 2017 Aug 6. p. 224-32
  • 91 Martin A, Lon B. Towards principled methods for training generative adversarial networks. In: NIPS 2016 Workshop on Adversarial Training
  • 92 Chu C, Zhmoginov A, Sandler M. Cyclegan, a master of steganography. arXiv preprint arXiv:1712.02950; 2017 Dec 8
  • 93 Shrivastava A, Adorno W, Ehsan L, Ali SA, Moore SR, Amadi BC. et al. Self-Attentive Adversarial Stain Normalization. arXiv preprint arXiv:1909.01963; 2019 Sep 4
  • 94 Wang Y, Zhong Z, Hua J. DeepOrganNet: On-the-Fly Reconstruction and Visualization of 3D/4D Lung Models from Single-View Projections by Deep Deformation Network. IEEE Trans Vis Comput Graph 2019; 26 (01) 960-70
  • 95 Jac Jr CR, Knopman DS, Jagust WJ, Petersen RC, Weiner MW, Aisen PS. et al. Update on hypothetical model of Alzheimer’s disease biomarkers. Lancet Neurology 2013; 12 (02) 207
  • 96 Mahmood F, Durr NJ. Deep learning and conditional random fields-based depth estimation and topographical reconstruction from conventional endoscopy. Med Image Anal 2018 Aug 2018; Aug 1 48: 230-43
  • 97 Shrivastava A, Pfister T, Tuzel O, Susskind J, Wang W, Webb R. Learning from simulated and unsupervised images through adversarial training. In: Proceedings of the IEEE conference on computer vision and pattern recognition 2017. p. 2107-16
  • 98 Ghorbani A, Natarajan V, Coz D, Liu Y. DermGAN: Synthetic Generation of Clinical Skin Images with Pathology. arXiv preprint arXiv:1911.0871
  • 99 Gholami A, Subramanian S, Shenoy V, Himthani N, Yue X, Zhao S. et al. A novel domain adaptation framework for medical image segmentation. In: International MICCAI Brainlesion Workshop 2018. p. 289-98
  • 100 Tang Y, Tang Y, Sandfort V, Xiao J, Summers RM. TUNA-Net: Task-oriented UNsupervised Adversarial Network for Disease Recognition in Cross-Domain Chest X-rays. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2019. p. 431-40
  • 101 Kapil A, Wiestler T, Lanzmich S, Silva A, Steele K, Rebelatto M. et al. DASGAN--Joint Domain Adaptation and Segmentation for the Analysis of Epithelial Regions in Histopathology PD-L1 Images. arXiv preprint arXiv:1906.11118; 2019 Jun 26
  • 102 Yang J, Dvornek NC, Zhang F, Chapiro J, Lin M, Duncan JS. Unsupervised domain adaptation via disentangled representations: Application to cross-modality liver segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2019. p. 255-63
  • 103 Dou Q, Ouyang C, Chen C, Chen H, Glocker B, Zhuang X. et al. Pnp-adanet: Plug-and-play adversarial domain adaptation network with a benchmark at cross-modality cardiac segmentation. arXiv preprint arXiv:1812.07907; 2018 Dec 19
  • 104 Dong N, Kampffmeyer M, Liang X, Wang Z, Dai W, Xing E. Unsupervised domain adaptation for automatic estimation of cardiothoracic ration. In: International Conference on Medical Image Computing and Computer-assisted intervention 2018. p. 544-52
  • 105 Javanmardi M, Tasdizen T. Domain adaptation for biomedical image segmentation using adversarial training. In: IEEE 15th International Symposium on Biomedical Imaging 2018. p. 554-8
  • 106 Novosad P, Fonov V, Collins DL. Unsupervised domain adaptation for the automated segmentation of neuroanatomy in MRI: a deep learning approach. bioRxiv preprint bioRxiv:845537 2019 Jan 1
  • 107 Ren J, Hacihaliloglu I, Singer EA, Foran DJ, Qi X. Adversarial domain adaptation for classification of prostate histopathology whole-slide images. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2018. p. 201-9
  • 108 Wang S, Yu L, Li K, Yang X, Fu CW, Heng PA. Boundary and Entropy-driven Adversarial Learning for Fundus Image Segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2019. p. 102-10
  • 109 Yan W, Wang Y, Xia M, Tao Q. Edge-Guided Output Adaptor: Highly Efficient Adaptation Module for Cross-Vendor Medical Image Segmentation. IEEE Signal Processing Letters 2019; 26 (11) 1593-7
  • 110 Hou X, Liu J, Xu B, Liu B, Chen X, Ilyas M. et al. Dual Adaptive Pyramid Network for Cross-Stain Histopathology Image Segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention 2019. p. 101-9
  • 111 Zhao H, Li H, Maurer-Stroh S, Gui Y, Deng Q, Cheng L. Supervised segmentation of un-annotated retinal fundus images by synthesis. IEEE Trans Med Imaging 2018; 38 (01) 46-56
  • 112 Tong L, Wu H, Wang MD. CAESNet: Convolutional AutoEncoder based Semi-supervised Network for improving multiclass classification of endomicroscopic images. J Am Med Inform Assoc 2019; Nov; 26 (11) 1286-96

Zoom Image
Fig. 1 a) Transfer learning (TL) and its different types; b) Overview of domain adaptation; c) Organization of this survey paper.
Zoom Image
Fig. 2 a) Summary of domain adaptation methodologies employed in medical imaging; b) Different scenarios encountered in cross-modality [16] [28] and single-modality [25] [29] domain adaptation.
Zoom Image
Fig. 3 Categorization of medical imaging DA publications as per a) imaging modality; b) anatomy; c) learning scenarios