Preprint
Article

This version is not peer-reviewed.

Dual Generators and Dynamically Fused Discriminators Adversarial Network to Create Synthetic Coronary Optical Coherence Tomography Images for Coronary Artery Disease Classification

Submitted:

06 March 2025

Posted:

07 March 2025

You are already at the latest version

Abstract
Deep neural networks have led to a substantial incursion for multifaceted classification tasks by making use of large-scale and diverse annotated datasets. However, diverse optical coherence tomography (OCT) datasets in the cardiovascular imaging remains an uphill task. This research focuses on improving the diversity and generalization ability of augmentation architectures while maintaining the baseline classification accuracy for coronary atrial plaques using a novel dual generators and dynamically fused discriminators conditional generative adversarial network (DGDFGAN). Our method is demonstrated on an augmented OCT dataset of 6900 images. With dual generators, our network provides the diverse outputs for the same input condition as each generator acts as a regularize for the other. In our model, this mutual regularization enhances the ability of both generators to generalize better across different features. The fusion discriminators use one discriminator for classification purposes hence avoiding the need for a separate deep architecture. A loss functional including the SSIM loss and FID scores confirm that perfect synthetic OCT image aliases are created. We optimize our model via Grey Wolf optimizer during model training. Furthermore, an inter-comparison and recorded SSID loss 0.9542±0.008 and FID score of 7 are suggestive of better diversity and generation characteristics that outperforms the performance of leading GANs architectures. We trust that our approach is practically viable and thus assist professionals for an informed decision making in clinical settings.
Keywords: 
;  ;  ;  

1. Introduction

Coronary artery disease (CAD) is a serious health challenge that is affecting substantial individuals and imposes burdens on healthcare systems worldwide [1,2]. CAD is an artery narrowing and blockage, leading to severe complications like heart attacks and heart failure [3,4]. Collaborative efforts by researchers and professionals resulted in advancements in treating and managing this disease. Evolving CAD prevention and management procedures embraces modern AI driven architectures and cutting-edge diagnostics including the Optical Coherence Tomography (OCT) [5,6,7]. OCT provides high-resolution images of the coronary artery walls for detailed analysis of plaque composition, and identification of vulnerable plaques [8]. With OCT, different plaque components can be distinguished based on their optical properties, including fibrous tissue, lipid cores and calcium deposits [9]. Fibrous plaque appears as a homogeneous, signal-rich region on OCT whereas lipid-rich plaque indicate as a low-signal region on OCT. Fibrocalcific plaque presents a heterogeneous appearance with sharp borders in OCT scans. The OCT imaging datasets suffers from key challenges including size, diversity, accurate manual annotations and significant variability which hinders their full potential utilization in modern data driven architectures [10,11,12,13,14,15].
Artificial Intelligence has emerged as a potential candidate when it comes to detection and characterization of coronary atrial plaques and is therefore being exploited to support improve the diagnosis via analyzing OCT scans. Multifarious deep learning architectures were reported in the literature regarding the detection and classification of coronary plaques using OCT scans [16,17,18,19,20,21,22,23,24,25,26,27]. In [28], a fully automated, two-step deep learning approach for characterizing coronary calcified plaque in intravascular optical coherence tomography (IVOCT) images is reported where a 3D convolutional neural network (CNN) was used with a SegNet deep learning model. Similarly, an automated atherosclerotic plaque characterization method that used a hybrid learning approach and self attention based U- Net architecture were reported elsewhere [29,30] for the better classification performance of coronary atrial plaques than existing methods.
Vision Transformers (ViTs) have also gained traction for coronary plaque detection especially in cases where global feature relationships are essential. In this regard [31] explored the use of ViTs for coronary plaque detection where ViTs outperformed CNN-based models for large datasets. Instead of relying on lumen segmentation, the proposed method identifies the bifurcation image using a ViT-based classification model and then estimate bifurcation ostium points by a ViT-based landmark detection model. The proposed ViT-based model are 2.54% and 16.08% higher than that of traditional non-deep learning methods [32]. For better generalization characteristics, a transformer-based pyramid network called AFS-TPNet for robust, end-to-end segmentation of CCP from OCT images [33]. Researchers also used physics-informed deep network QOCT-Net to recover pixel-level optical attenuation coefficients directly from standard IVOCT B-scan images [34]. While these cutting-edge architectures delivered satisfactory results but from a clinical standpoint, acquiring large and diverse datasets with patients across different disease stages is stimulating, shadowing real potential of proposed deep learning architectures. To address this issue, data augmentation (DA) techniques were deployed to augment limited medical imaging training data that can be further fed to deep learning algorithms for better insights. Generative Adversarial Networks (GANs) have offered promising solutions by generating high-quality images that closely resemble real OCT scans, thereby aiding in dataset augmentation, image enhancement, and training deep learning classifiers. In [35], pseudo labeling, using model predictions as labels for unlabeled data, has been employed as a data augmentation technique. This method has demonstrated improvements in model performance by increasing the effective size of the training dataset. The StyleGAN2 and Cyclic GANs frameworks were used to generate high resolution synthetic patches for data augmentation for an improved data augmentation performance in the case of low data, across three different OCT datasets encompassing a range of scanning parameters [36,37]. To improvise generalization across different datasets, sparsity constrained GANs with baseline accuracy is available in the literature [38].
However, the recent works show that despite high structural similarity between the synthetic data and the real images, a considerable distortion is observed in frequency domain, therefore dual and triple discriminator architectures including Fourier acquisitive GAN (DDFA-GAN) to generate more realistic OCT images were proposed [39,40,41]. By applying multiple discriminators, the proposed models were jointly trained with the Fourier and spatial details of the images and results were are compared with popular GANs including deep convolutional GAN (DCGAN), Wasserstein GAN with gradient penalized (WGAN-GP), and least square GAN (LS-GAN). In [42] a multi-stage and multi-discriminatory generative adversarial network (MultiSDGAN) specifically for super-resolution and segmentation of OCT scans was proposed. This resulted in an improved performance by satisfying all the discriminators at multiple scales and including perceptual loss function. While the use of multiple discriminators in GANs attracted much attention but due to overloading of a single generator in the network, the desired generalization and diversity in the generated images is yet an open problem.
The resolve of this study was to perform data augmentation on an OCT dataset using a novel dual generator multiple fusion discriminator network for synthesizing high quality coronary images. The use of dual generators exploited in this paper helps to achieve better generalization and generate diverse of images of coronary arterial plaques. In our model, the two generators G1 and G2 receives the same conditional input y but they generate different variations of the OCT images as each generator acts as a regularizer for the other. In our model, one generator’s output can serve as a reference for improving the other and this mutual regularization enhances the ability of both generators to generalize better across different features and conditions. Additionally, we skipped the need of a separate classification architecture, rather uses a discriminator within our presented GAN architecture for classification of coronary atrial plaques into three classes. The objective functions of the generator and the five discriminators were set in competition against each other. We derive a novel objective function for DGDFGAN and optimize it through Whales algorithm during model training. This makes the model training more stable and improving the quality and diversity of the generated images. Using assessments, we illustrated how populating real images with created instances during the training phase increased our confidence in the reliable label prediction. Experimental results demonstrate that DGCGAN achieves optimal results in terms of the similarity and generalization characteristics.
The main contributions of the paper are summarized as under;
  • A novel dual generator architecture using mutual regularization to achieve better generalization for OCT images of coronary arterial plaques. A discriminator (D1) within the network is used for classification alleviating the need for a separate classification architecture resulting in computational efficiency.
  • A novel architecture of dynamic multi-discriminator fusion is designed to play an adversarial game against two generators. We introduce a dynamic fusion mechanism to adjust the weighting of discriminators based on the image specific conditions.
  • Essentially, we incorporated adversarial loss, perceptual loss, L1 loss, cross entropy and diversity loss in our loss function for an enhanced realism in the generated images.

2. Methods and Materials

Generally, the cGAN takes two input data, namely the latent variable z and the conditional constraints into both generator G and discriminator D as illustrated in Figure 1. The G combines the initial noise input Z with label information to produce the generated output after reshaping to feature ma, transposed Convolutions, batch normalization and activation functions as presented in Figure 1. The final layer of the generator typically uses an activation function.

2.1. OCT Dataset and Its Pre-Processing

A dataset of 51 patients was created using a commercially available OCT system. The focus of the study was on vessels affected by stenosis, though for simplicity, cases involving serial stenosis, mixed plaques and bypass graft stenosis were excluded. Ethical approval for this research was granted by the Galway Clinical Research Ethics Committee (GCREC), and informed consent was obtained from all participants.
Three clinicians independently annotated the OCT images, but final labels were determined through consensus agreement. The classification task was structured by designating a specific label in contrast to the others. Prior to model input, pre- processing steps were applied to the raw OCT images. Vulnerable plaques meeting the predefined fibrous cap thickness criteria were excluded from the analysis. Plaque characterization was based on signal intensity relative to the lumen, classifying them into three classes namely; lipid plaques, calcified plaques and as “no plaque.”

2.2. Proposed Dual Generator Multi-Fusion GAN

In our configuration of the dual generators presented in Figure 2, each generator learns different plausible ways to generate an OCT image based on the same conditioning input y. The two generators G1 and G2 in Figure 2 learn to capture different aspects of the underlying OCT image structure. G1 focus on capturing fine details of the OCT image, while G2 emphasize on abstract features including overall tissue structure and patterns. With dual generators, the risk of mode collapse is also reduced because each generator learns a distinct aspect of the data. In case, if one generator fails to cover the diversity of the distribution, the other generator fills in missing modes which helps to ensure more realistic and varied synthetic OCT images. Likewise, the two generators learn different modes of the data distribution, they help the model to generalize better for unseen OCT data.
In the presented formulation, the role and parameters of each discriminator is illustrated in Figure 3. The ultimate aim of dynamic discriminators fusion is to improve quality, realism, and diversity of the generated OCT images. As presented in Figure 3, Discriminator 1 (D1) is used for the classification of coronary atrial plaques into three predefined classes. designed to focus on a different characteristic of the OCT image. D2 evaluates the temporal consistency of the generated MRI images and D3 ensures that the generated image matches a particular condition. The role of D4 is to ensure global realism of the generated images by taking into account spatial and temporal features. D5 embeds perceptual loss which is computed as the distance between the feature maps of the real and generated images. This loss is combined with the adversarial loss to guide the discriminators on high level perceptual differences.

2.3. Mathematical Formulation

We use dual- generator cGANs and multi-discriminators for OCT synthetic image generation. The generator takes a latent vector z (random noise) and a label y, and produces a synthetic MRI image x.
x= G (z, y)
where z is the latent vector (random noise), y is the label and x is the generated MRI image.
A latent vector sampled from a Gaussian distribution to introduce randomness and variability into the generated images. The conditioning information in the form of one-hot vector representing the class label of the OCT image is concatenated with random noise z that forms the initial input to the generators. Each discriminator receives the same generated image and label but has a slightly different focus in its evaluation. D1 is used for classification that to which class the generated image belongs using the expression given in Equation (2).
L 1 = E z ~ P d a t a z c = 1 3 y c l o g D 1 ( x ) E z ~ P z c = 1 3 l o g [ ( 1 D 1 G ( z ) ) ]
where Pdata is the real data distribution, Pz is the distribution over the latent space and yc is the one-hot encoding vector indicating the true class of image x.
D2 performs label-specific evaluation whether the generated image matches the expected distribution of the label using Equation (3).
L 2 = E z ~ P d a t a z l o g D 2 ( x ) E z ~ P z l o g [ ( 1 D 2 G ( z ) ) ]
where D2(x) is the probability that the image x is real.
Discriminator 3 evaluates perceptual quality, which compares the high-level features of the generated and real images. This can be modeled using a perceptual loss based on the difference in feature representations between real and generated images as given in Equation (4).
L 3 = E x ~ P data f ( x ) f ( G ( z ) ) 2 2
Discriminator 4 evaluates the quality of the image, which can be linked to the luminance or image gradient loss. It attempts to measure how sharp and high-quality the generated images using the L1 loss.
L 4 = E z ~ P d a t a x G ( z ) 1 E z ~ P z l o g [ ( 1 D 4 G ( z ) ) ]
where x−G(z) denotes the L1 normalized between the real image x and the generated image G(z).
Discriminator 5 evaluates the consistency loss between the generated image to maintain style over different transformations of the input.
L 5 = E z ~ P z G 1 ( G ( z ) ) z 1
where G−1(G(z)) represents the inverse transformation.
The total combined loss LT can be written as:
L T = λ 1 L 1 + λ 2 L 2 + λ 3 L 3 + λ 4 L 4 + λ 5 L 5
where λ1 controls the importance of the classification loss in discriminator 1, λ2 controls the importance of the realism/quality loss via discriminator 2, λ3 controls the importance of the perceptual loss in discriminator 3), λ4 controls the importance of the quality loss in discriminator 4 and λ5 controls the importance of the consistency loss via discriminator 5
L T = λ 1 ( E z ~ P d a t a z c = 1 3 y c l o g D 1 ( x ) E z ~ P z l o g c = 1 3 l o g [ ( 1 D 1 G ( z ) ) ] ) + λ 2 ( E z ~ P d a t a z l o g D 2 ( x ) E z ~ P z l o g [ ( 1 D 2 G ( z ) ) ] ) + λ 3 ( E z ~ P d a t a x G ( z ) 2 2 ) + λ 4 ( E z ~ P d a t a x G ( z ) 1 E z ~ P z l o g [ ( 1 D 4 G ( z ) ) ] ) + λ 5 ( E z ~ P z G 1 ( G ( z ) ) z 1
We use latent interpolation between the two generators to encourage diverse image generation. By interpolating between the latent space of G1 and G2, we force both generators to produce more diverse outputs, allowing them to explore a broader space of possibilities using Equation9.
z′=α⋅z1+(1−α)⋅z2
where z1 and z2 are two different latent codes, and α∈[0,1] is a blending factor that encourages diverse outputs. The generator now operates on the blended latent space as given in Equation (10).
G1(z′∣y) and G2(z′∣y)
Grey Wolf Optimization is used to explore the search space effectively for an optimal solution where each wolf represents a candidate solution in the optimization space. Each wolf represents a set of parameters of the neural network that is optimized by the GWO algorithm.

3. Results and Analysis

In this paper, the experiments are conducted according to the baseline architecture in Figure 3. Our original dataset contains 27 images (no plaque), 22 images (calcified) and 20 images of lipid plaques. After augmentation by a factor of 100 the dataset’s size was increased 6900 images with 2700 (no plaque), 22oo calcified images and 2000 lipid images. The samples generated by our model for each class is presented in Figure 4. Then, we applied cross validation (10-CV) to estimate the generalization performance of the compared segmentation models on the unseen data.
The DGDFGAN model is used for experimentation, and we tried different combinations where the multi-stage aspect or the multi-discriminatory aspect was treated incrementally. We analyzed the Structural Similarity Index Measure (SSIM) loss and the Fréchet Inception Distance (FID) score as an additional cost function. All of the experiments were run for 100 epochs for each fold. SSIM index is a metric that evaluates the degradation of images from a perceptual point of view and FID score underlines the diversity impact. We also included the L1 loss over 10 fold cross validation (10-CV). L1 loss measures the absolute difference between predicted and ground-truth values. It is used to enforce pixel-wise similarity in image generation tasks.
An inter-comparison our proposed model with the cutting edge Multi SDGAN model is presented in Table 1 in terms of SSIM, L1 and FID scores. It is evident that employing the dual generators and multi-discriminatory fusion module improves the performance in both similarity and diversity aspects. The dual generator helps in achieving an improved diversity in generated images as ach generator focus on different aspects of OCT image synthesis. This reduces mode collapse, where a single generator might otherwise get stuck producing similar images. Each generator interacts with different discriminators, ensuring that features are well-learned from multiple perspectives for an enhanced feature representation. By training on different discriminators, the generators learn robust representations that can generalize better to unseen OCT images. As illustrated in Table 2, the two generators share the load, leading to faster convergence resulting in lower FID (better realism) and higher SSIM (better structural similarity) when compared to other approaches including DDFAGAN, MHWGAN and MultiSDGAN.
We computed the SSIM loss which is a perceptual loss function to evaluate the similarity between the original and generated OCT images. It exploits contrast, and structure rather than just pixel-wise differences. The algorithm and related parameters description are presented in Figure 5.
The FID score is used to measure similarity between generated images and OCT real images by comparing their feature distributions. It exploits the mean and covariance of the feature embeddings. FID considers both realism and diversity for making it a better generative model. The used algorithm and the relevant variables are illustrated in Figure 6.
As illustrated in Table 2, an increase of 10. 81% in SSIM index observed as compared to leading SSIM measurements in case of Multi- SDGAN [42]. This is indicative of achieving better structural similarity between the original and generated OCT images. An improvement of 41. 66% in the FID score with baseline results of Multi- SDGAN confirms that our model significantly squeeze the difference between the feature distributions of real and generated images. L1 score reported in Table 2 is further suggestive of the fact that our model predictions are close to ground truth. The dual-generator setup mitigates the problem of biased learning, where a single generator struggle to capture all variations in the OCT dataset. By working in parallel, the generators foster a more comprehensive understanding of data distribution. Consequently, the multi-discriminator framework enhanced the FID scores by ensuring a richer adversarial signal to produce images with better realism and diversity.

4. Conclusions

In this work, we designed a dual generator architecture, entitled DGDFGAN which achieved the better diversity and generalization characteristics for generating synthetic OCT images. The two generators balance the load of multi- discriminators and helps in good generalization via mutual regularization. Further, we skipped the need of a separate classification model by using one discriminator within the DGDFGAN for classification, hence reducing computational effort. The multi-discriminatory modules are responsible for discriminating generators outputs using different loss functions. We computed SSIM loss, FID score and other losses for our model and results suggest that dual generator multi- discriminator set up leads to higher-quality image synthesis and diverse data augmentation which helps in improving diagnostic reliability of medical image datasets including OCT for clinical AI applications.

References

  1. Borrelli, N.; Merola, A.; Barracano, R.; Palma, M.; Altobelli, I.; Abbate, M.; Papaccioli, G.; Ciriello, G.D.; Liguori, C.; Sorice, D.; et al. The Unique Challenge of Coronary Artery Disease in Adult Patients with Congenital Heart Disease. J. Clin. Med. 2024. [CrossRef]
  2. https://www.cdc.gov/heart-disease/about/coronary-artery-disease.html.
  3. https://www.ncbi.nlm.nih.gov/books/NBK355309/.
  4. Młynarska, E.; Czarnik, W.; Fularski, P.; Hajdys, J.; Majchrowicz, G.; Stabrawa, M.; Rysz, J.; Franczyk, B. From Atherosclerotic Plaque to Myocardial Infarction—The Leading Cause of Coronary Artery Occlusion. Int. J. Mol. Sci. 2024. [CrossRef] [PubMed]
  5. Yang, D.; Ran, A.R.; Nguyen, T.X.; Lin, T.P.H.; Chen, H.; Lai, T.Y.Y.; Tham, C.C.; Cheung, C.Y. Deep Learning in Optical Coherence Tomography Angiography: Current Progress, Challenges, and Future Directions. Diagnostics 2023, 13, 326. [Google Scholar] [CrossRef] [PubMed]
  6. Zafar, H.; Zafar, J.; Sharif, F. Automated Clinical Decision Support for Coronary Plaques Characterization from Optical Coherence Tomography Imaging with Fused Neural Networks. Optics 2022, 3, 8–18. [Google Scholar] [CrossRef]
  7. Zafar, H.; Zafar, J.; Sharif, F. GANs-Based Intracoronary Optical Coherence Tomography Image Augmentation for Improved Plaques Characterization Using Deep Neural Networks. Optics 2023, 4, 288–299. [Google Scholar] [CrossRef]
  8. https://onlinelibrary.wiley.com/doi/full/10.1002/tbio.201900034.
  9. https://iopscience.iop.org/article/10.1088/2057-1976/aab640/meta.
  10. Araki, M. , Park, SJ., Dauerman, H.L. et al. Optical coherence tomography in coronary atherosclerosis assessment and intervention. Nat Rev Cardiol 19, 684–703 (2022). [CrossRef]
  11. Oosterveer, T.T.M. , van der Meer, S.M., Scherptong, R.W.C. et al. Optical Coherence Tomography: Current Applications for the Assessment of Coronary Artery Disease and Guidance of Percutaneous Coronary Interventions. Cardiol Ther 9, 307–321 (2020). [CrossRef]
  12. Sant Kumar, Miao Chu, Jordi Sans-Roselló et al., In-Hospital Heart Failure in Patients With Takotsubo Cardiomyopathy Due to Coronary Artery Disease: An Artificial Intelligence and Optical Coherence Tomography Study, Cardiovascular Revascularization Medicine,Volume 47, 2023, Pages 40-45. [CrossRef]
  13. Mintz, Gary S et al.Intravascular imaging in coronary artery disease, The Lancet, Volume 390, Issue 10096, 793 - 809. [CrossRef]
  14. Matthews, Stephen Daniel MD; Frishman, William H. MD, A Review of the Clinical Utility of Intravascular Ultrasound and Optical Coherence Tomography in the Assessment and Treatment of Coronary Artery Disease, Cardiology in Review 25(2):p 68-76, March/April 2017. [CrossRef]
  15. Aiko Shimokado, Yoshiki Matsuo, Takashi Kubo et al., In vivo optical coherence tomography imaging and histopathology of healed coronary plaques, Atherosclerosis, vol. 275, pp. 35-42, 2018. [CrossRef]
  16. Carpenter, H.J.; Ghayesh, M.H.; Zander, A.C.; Li, J.; Di Giovanni, G.; Psaltis, P.J. Automated Coronary Optical Coherence Tomography Feature Extraction with Application to Three-Dimensional Reconstruction. Tomography 2022, 8, 1307–1349. [Google Scholar] [CrossRef] [PubMed]
  17. Infact 15 Avital, Y., Madar, A., Arnon, S. et al. Identification of coronary calcifications in optical coherence tomography imaging using deep learning. Sci Rep 11, 11269 (2021). [CrossRef]
  18. Niioka, H. , Kume, T., Kubo, T. et al. Automated diagnosis of optical coherence tomography imaging on plaque vulnerability and its relation to clinical outcomes in coronary artery disease. Sci Rep 12, 14067 (2022). [CrossRef]
  19. Yoon M, Park JJ, Hur T, Hua CH, Hussain M, Lee S, Choi DJ. Application and Potential of Artificial Intelligence in Heart Failure: Past, Present, and Future. Int J Heart Fail. 2023 Nov 30;6(1):11-19. [CrossRef]
  20. Molenaar, M.A. , Selder, J.L., Nicolas, J. et al. Current State and Future Perspectives of Artificial Intelligence for Automated Coronary Angiography Imaging Analysis in Patients with Ischemic Heart Disease. Curr Cardiol Rep 24, 365–376 (2022). [CrossRef]
  21. Föllmer, B., Williams, M.C., Dey, D. et al. Roadmap on the use of artificial intelligence for imaging of vulnerable atherosclerotic plaque in coronary arteries. Nat Rev Cardiol 21, 51–64 (2024). [CrossRef]
  22. Seetharam K, Min JK. Artificial Intelligence and Machine Learning in Cardiovascular Imaging. Methodist Debakey Cardiovasc J. 2020 Oct-Dec;16(4):263-271. [CrossRef]
  23. Langlais, É.L., Thériault-Lauzier, P., Marquis-Gravel, G. et al. Novel Artificial Intelligence Applications in Cardiology: Current Landscape, Limitations, and the Road to Real-World Applications. J. of Cardiovasc. Trans. Res. 16, 513–525 (2023). [CrossRef]
  24. Thomas, T., Kurian, A.N. (2022). Artificial Intelligence of Things for Early Detection of Cardiac Diseases. In: Al-Turjman, F., Nayyar, A. (eds) Machine Learning for Critical Internet of Medical Things. Springer, Cham. [CrossRef]
  25. Chen, SF., Loguercio, S., Chen, KY. et al. Artificial Intelligence for Risk Assessment on Primary Prevention of Coronary Artery Disease. Curr Cardiovasc Risk Rep 17, 215–231 (2023). [CrossRef]
  26. Bozyel S, Şimşek E, Koçyiğit Burunkaya D, Güler A, Korkmaz Y, Şeker M, Ertürk M, Keser N. Artificial Intelligence-Based Clinical Decision Support Systems in Cardiovascular Diseases. Anatol J Cardiol. 2024 Jan 7;28(2):74–86. [CrossRef]
  27. Ke-Xin Tang, Yan-Lin Wu, Su-Kang Shan et al.,Advancements in the application of deep learning for coronary artery calcification, Meta-Radiology,vol.3, Issue 1, 100134. [CrossRef]
  28. J. Lee et al., “Segmentation of Coronary Calcified Plaque in Intravascular OCT Images Using a Two-Step Deep Learning Approach,” in IEEE Access, vol. 8, pp. 225581-225593, 2020. [CrossRef]
  29. Infact 15 Lee, J., Prabhu, D., Kolluru, C. et al. Fully automated plaque characterization in intravascular OCT images using hybrid convolutional and lumen morphology features. Sci Rep 10, 2596 (2020). [CrossRef]
  30. Infact 16 https://www.sciencedirect.com/science/article/abs/pii/S174680942300321X.
  31. X. Li et al., “Multi-Scale Reconstruction of Undersampled Spectral-Spatial OCT Data for Coronary Imaging Using Deep Learning,” in IEEE Transactions on Biomedical Engineering, vol. 69, no. 12, pp. 3667-3677, Dec. 2022. [CrossRef]
  32. Bifurcation detection in intravascular optical coherence tomography using vision transformer based deep learning. [CrossRef]
  33. Rongyang Zhu, Qingrui Li, Zhenyang Ding, Kun Liu, Qiutong Lin, Yin Yu, Yuanyao Li, Shanshan Zhou, Hao Kuang, Junfeng JiangShow full author list.
  34. Published 18 July 2024 • © 2024 Institute of Physics and Engineering in Medicine Physics in Medicine & Biology, Volume 69, Number 15Citation Rongyang Zhu et al. 2024 Phys. Med. Biol. 69 155009.
  35. Yiqing Liu, Farhad R. Nezami, Elazer R. Edelman,A transformer-based pyramid network for coronary calcified plaque segmentation in intravascular optical coherence tomography images, Computerized Medical Imaging and Graphics,Volume 113, 2024, 102347. [CrossRef]
  36. S. Zheng, W. Shuyan, H. Yingsa and S. Meichen, “QOCT-Net: A Physics-Informed Neural Network for Intravascular Optical Coherence Tomography Attenuation Imaging,” in IEEE Journal of Biomedical and Health Informatics, vol. 27, no. 8, pp. 3958-3969, 2023. [CrossRef]
  37. https://arxiv.org/abs/2310.05990.
  38. Kugelman, J. , Alonso-Caneiro, D., Read, S.A. et al. Enhancing OCT patch-based segmentation with improved GAN data augmentation and semi-supervised learning. Neural Comput & Applic 36, 18087–18105 (2024). [CrossRef]
  39. X. Li et al., “Cross-Platform Super-Resolution for Human Coronary Oct Imaging Using Deep Learning,” 2024 IEEE International Symposium on Biomedical Imaging (ISBI), Athens, Greece, 2024, pp. 1-5. [CrossRef]
  40. K. Zhou et al., “Sparse-Gan: Sparsity-Constrained Generative Adversarial Network for Anomaly Detection in Retinal OCT Image,” 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), Iowa City, IA, USA, 2020, pp. 1227-1231. [CrossRef]
  41. J. Ma, H. Xu, J. Jiang, X. Mei and X. -P. Zhang, “DDcGAN: A Dual-Discriminator Conditional Generative Adversarial Network for Multi-Resolution Image Fusion,” in IEEE Transactions on Image Processing, vol. 29, pp. 4980-4995, 2020.
  42. M. Tajmirriahi, R. Kafieh, Z. Amini and V. Lakshminarayanan, “A Dual-Discriminator Fourier Acquisitive GAN for Generating Retinal Optical Coherence Tomography Images,” in IEEE Transactions on Instrumentation and Measurement, vol. 71, pp. 1-8, 2022, Art no. 5015708. [CrossRef]
  43. C. Zhao et al., “MHW-GAN: Multidiscriminator Hierarchical Wavelet Generative Adversarial Network for Multimodal Image Fusion,” in IEEE Transactions on Neural Networks and Learning Systems, vol. 35, no. 10, pp. 13713-13727, 2024. [CrossRef]
  44. [42] P. Jeihouni, O. Dehzangi, A. Amireskandari, A. Rezai and N. M. Nasrabadi, “MultiSDGAN: Translation of OCT Images to Superresolved Segmentation Labels Using Multi-Discriminators in Multi-Stages,” in IEEE Journal of Biomedical and Health Informatics, vol. 26, no. 4, pp. 1614-1627, April 2022. [CrossRef]
Figure 1. A generic scheme of operations performed in CGANs for image augmentation.
Figure 1. A generic scheme of operations performed in CGANs for image augmentation.
Preprints 151546 g001
Figure 2. Dual generator with fused multi- discriminators set up for generating synthetic images.
Figure 2. Dual generator with fused multi- discriminators set up for generating synthetic images.
Preprints 151546 g002
Figure 3. A schematic of internal architectures within the dynamic fusion discriminators.
Figure 3. A schematic of internal architectures within the dynamic fusion discriminators.
Preprints 151546 g003
Figure 4. DGDFGAN generated synthetic images for different classes: (A) normal plaque, (B) calcium plaque, (C) lipid plaque.
Figure 4. DGDFGAN generated synthetic images for different classes: (A) normal plaque, (B) calcium plaque, (C) lipid plaque.
Preprints 151546 g004
Figure 5. An algorithm used to compute the SSIM loss and the description of involved parameters.
Figure 5. An algorithm used to compute the SSIM loss and the description of involved parameters.
Preprints 151546 g005
Figure 6. An algorithm used to compute the FID score and the description of involved parameters.
Figure 6. An algorithm used to compute the FID score and the description of involved parameters.
Preprints 151546 g006
Table 1. An inter-comparison of our proposed model with the leading architectures.
Table 1. An inter-comparison of our proposed model with the leading architectures.
Model Description SSIM FID Score L1
Generator with two SR Layer, MultiSD Module [42] 0.7933±0.0077 39 0.057±0.0028
MultiSDGAN Module with Attention mechanism [42] 0.8240±0.0028 16 0.032±0.0017
Proposed DGDFGAN 0.9542±0.008 7 0.010±0.0005
Table 2. An inter-comparison on the basis of number of discriminators to our proposed model with the leading architectures on our augmented OCT Dataset.
Table 2. An inter-comparison on the basis of number of discriminators to our proposed model with the leading architectures on our augmented OCT Dataset.
Model Description SSIM FID Score L1
DDFA-GAN model using dual disciminators [40] 0.7350±0.0064 67 0.054±0.0047
MHW-GAN for multimodal image fusion using triple discriminators [41] 0.7718±0.0039 46 0.054±0.0047
Multi- SDGAN with attention mechanism using maximum discriminators [42] 0.8611±0.0017 12 0.032±0.0017
Dual Generator proposed model 0.9452±0.008 7 0.010±0.0005
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated