Preprint
Article

This version is not peer-reviewed.

Deep Learning–Based Choroidal Boundary Detection in Geographic Atrophy Using Spectral-Domain Optical Coherence Tomography

Submitted:

28 January 2026

Posted:

29 January 2026

You are already at the latest version

Abstract
Background/Objectives: To evaluate the challenges and limitations of a deep learning model for automated choroidal boundary detection in eyes with geographic atrophy (GA) using Spectral-Domain OCT (SD-OCT), and to assess the workflow efficiency of an AI-assisted manual verification approach. Methods: In this retrospective study, total 5,723 scans (Heidelberg Spectralis) with GA were analyzed. A previously validated tool (NMI ChoroidAI) was used to segment the choroidal inner (CIB) and outer (COB) boundaries. We compared the "AI-assisted" workflow (automated segmentation followed by manual verification) against "manual segmentation only" in terms of accuracy and time consumption. Slice-wise boundary errors were graded as 0 (accurate), 1 (≤33% deviation), 2 (33–66% deviation), or 3 (>66% deviation). Outcomes included error rates and weighted F₁-score (and precision where applicable). Total time for manual-only segmentation versus AI-assisted verification was recorded .Inter-reader variability was assessed between the two readers using intraclass correlation coefficient. Results: For CIB, only 5.2% of B-scans showed any deviation (strictly accurate in 94.8%), with weighted F₁-score 0.97 and precision 1.00. COB was more error-prone: 19.0% of B-scans showed deviation, however, when minor deviations were considered acceptable, COB acceptability increased to 94.2% (i.e., 5.8% remained >33% deviated). Only 13.2% of B-scans required minor manual correction. For a 97-scan volume, processing time decreased from an average of 7 hours (manual only) to 45 minutes (AI + human verification), an approximate 90% reduction in manual effort. Inter-reader agreement was high (ICC 0.923 for CIB and 0.938 for COB). Conclusions: Although the deep learning model exhibits limitations in COB detection due to artifacts, it serves as a valuable assistive tool. Our model substantially reduces human effort, but mandatory human verification is required to correct boundary errors caused by hyper-transmission before use in clinical trials.
Keywords: 
;  ;  ;  ;  

1. Introduction

The role of the choroid in age-related macular degeneration (AMD) pathophysiology remains debated, including the sequencing of choriocapillaris loss and retinal pigment epithelium (RPE) degeneration. In geographic atrophy (GA), the advanced stage of AMD, this temporal relationship between choriocapillaris dropout and RPE degeneration is correlated to longitudinal OCT-based quantification of choroidal structure [1]. Growing evidence highlights the role of the choroid in AMD and GA, reflecting its critical contribution to retinal metabolic support. [2] Choroidal thinning has been consistently reported in eyes with GA, while a distinct subset, referred to as pachychoroid GA, has also been identified. [3] Similarly, other choroidal biomarkers in geographic atrophy have also been explored, including the choroidal vascularity index (CVI) [4] and total choroidal volume, both of which have been shown to be reduced in GA.
Manual segmentation of the choroidal inner boundary (CIB) and outer boundary (COB) on high-density optical coherence tomography (OCT) volumes is prohibitively time-consuming for both clinical trials and routine practice. Precise delineation of these boundaries forms the foundation for deriving key choroidal biomarkers, including volumetric and sectoral choroidal thickness, choroidal volume, choroidal vascularity index, choroidal contour [6] and three-dimensional reconstruction of choroidal vascular architecture. [7] Deep-learning models can delineate the choroid reliably in many healthy eyes, but GA changes the imaging problem itself. With RPE loss, SD-OCT often shows marked hyper-transmission that can wash out the choroidal–scleral interface (CSI) or create “tailing,” so the boundary is no longer cleanly expressed in the intensity profile [8,9]. In this setting, accuracy reported in healthy or non-GA AMD cohorts may not carry over, because the disease introduces GA-specific contrast shifts and local ambiguities, most notably for COB detection [10,11]. Recent GA imaging work echoes this point, emphasizing that RPE loss fundamentally reshapes OCT contrast at the CSI in ways that can mislead automated boundary estimation [12].
Currently, there is no automated algorithm available that can reliably segment the choroid in GA eyes without supervision. Given the persistence of these artifacts, the role of AI must be defined as assistive rather than autonomous. A "human-in-the-loop" approach is critical, where manual verification serves as a mandatory quality control step to correct errors caused by hyper-transmission, ensuring that efficiency does not come at the cost of clinical validity. Accordingly, no AI-based method is expected to be 100% error-free in this setting. In this study, we evaluated a deep learning pipeline for the automated choroidal segmentation of GA eyes imaged with SPECTRALIS® OCT system (Heidelberg Engineering, Heidelberg, Germany) [13,14]. Unlike lesion-detection work, we perform a "failure analysis" of boundary definition, hypothesizing that hyper-transmission artifacts affect COB more than CIB. We also quantify efficiency of an AI-assisted approach versus manual grading to see if time savings justify expert artifact correction.

2. Materials and Methods

Data and Imaging

This retrospective study included 5723 scans from 59 volumes (each 97 scans) with GA secondary to dry AMD. The study was conducted at the UPMC Vision Institute, from October 2023 to September 2024 following approval from the Institutional Review Board. The study adhered to the tenets of the Declaration of Helsinki. A “waiver of informed consent” was obtained considering the retrospective nature of the study.

Imaging Acquisition

Volumetric SD-OCT imaging was performed using the SPECTRALIS® OCT system (Heidelberg Engineering, Heidelberg, Germany). Each volume covered a 6x6 mm area and consisted of 97 horizontal B-scans covering macular regions. It can scan up to 40,000 A-scans per second, has a wavelength of 870nm and an axial resolution of about 3.9 µm in tissue. The SD-OCT software’s integrated scoring system called the quality score or Q-score, evaluated the quality of the scans. Only scans with a quality score of 25 or higher on a scale of 0-40 met the criteria for inclusion in the analysis. For each volume, the SPECTRALIS® system also acquired a co-registered infrared (IR) fundus image used for scan localization during acquisition. Scan quality was reviewed, and volumes with significant motion artifacts or signal loss were excluded from analysis.
Two trained readers (SL and JW) independently assessed the OCT images and graded the degree of inaccuracy for each B-scan, while also recording the number of completely accurate B-scans within each volume. Manual correction was performed in a subset of volumes, and the time required for full manual segmentation was compared with the time required for manual correction following automated segmentation. (Figure 1)

Segmentation Model and Pipeline

We used the previously validated NMI-ChoroidAI algorithm (Netramind Innovations Inc., Pittsburgh, PA), a deep convolutional neural network based on a ResUNet (Residual U-Net) architecture, to segment the choroidal inner and outer boundaries in the OCT volumes. [15] The segmentation of choroid layer boundaries involves identifying two critical interfaces: the RPE–choroidal interface (CIB) and the choroid–sclera interface (COB/CSI). Manual delineation and correction of CIB and COB was performed using the manual segmentation feature on the NMI ChoroidAI tool.
Shadow compensation was applied to each B-scan to enhance visualization of the choroidal region. The preprocessed B-scan was then passed to the ResUNet, which produced pixel-wise probability maps for the CIB and COB. These probability maps were converted into continuous boundaries by selecting the highest-probability pixel for each boundary, followed by minimal post-processing (e.g. Robust Locally Estimated Scatterplot Smoothing). For visualization and quality control, the predicted CIB and COB lines were overlaid on the original B-scans.

Performance Evaluation and Classification Metrics

Performance was assessed by comparing automated segmentations against manual ground truth with slice-wise error analysis. Each B-scan’s boundary deviation was graded: 0 = accurate, 1 = minor (≤33% of the B-scan deviated), 2 = moderate (33–66%), 3 = major (>66% or missed) (Figure 2). The percentage quantified the portion of each B-scan where the automated segmentation diverged from the ground-truth boundary. Manual segmentation paid close attention to distinguishing the true scleral interface from hyper-transmission artifacts. We also identified error causes such as low image quality, motion, or lesion morphology.
We computed confusion matrices and derived accuracy, sensitivity, precision, and F1 score for CIB and COB. Specificity and negative predictive value were not reported since every B-scan contains true CIB/COB boundaries. Metrics were calculated separately for CIB and COB to assess performance under different error tolerance levels. Inter-reader agreement for identifying number of accurate scans from each volume was assessed using intraclass coefficient (ICC; two-way mixed-effects model, absolute agreement).

Workflow Efficiency Assessment

To evaluate the clinical utility of the AI model despite its limitations, we compared the time required for two workflows in three volumes. "Manual-Only" segmentation; the time taken for an expert to segment the full choroidal volume (CIB and COB) from scratch. "AI-Assisted" workflow; the time taken for the model to generate initial boundaries plus the time required for a human expert to verify and correct all segmentation errors (specifically correcting COB displacements in areas of signal washout). Timing was recorded from the start of each workflow to completion and summarized across volumes.

Statistical Analysis

Normality was tested with the Shapiro–Wilk test. The ANOVA test was used for normally distributed data and the Kruskal-Wallis test for non-normally distributed data. The Chi-square test was employed to compare the two methods. A confidence interval (CI) of 95% and a p-value of ≤0.05 were considered statistically significant. All statistical procedures were executed using IBM SPSS Statistics software, version 26.

3. Results

Choroidal Boundary Segmentation Performance and Failure Analysis

Across 5,723 B-scans, CIB was highly robust, with a strict accuracy (Grade 0) of 94.8%, a weighted F₁-score of 0.97, and a precision of 1.0. The remaining were minor deviations at 2.5%, and moderate and major errors at 1.01% and 1.64%, respectively.
In contrast, COB accuracy dropped to 81.0% (weighted F₁-score: 0.89). There were minor deviations in 13.2% of scans and moderate-to-major errors in about 6% (Categories 2 and 3). Confusion matrices (Figure 3) indicate many more Grade-3 errors for COB (202 scans) compared to CIB (94 scans). However, considering minor deviations (≤33% deviation is considered acceptable) increased COB sensitivity to 94.2%, suggesting many errors consisted of minor spatial deviation rather than outright failures. Inter-grader reliability for identifying the number of accurate B-scans per volume was excellent, with ICCs of 0.938 (95% CI: 0.874–0.970, p < 0.001) for COB and 0.923 (95% CI: 0.847–0.962, p < 0.001) for CIB. Inter-grader reliability for identifying the number of accurate B-scans per volume was excellent, with ICCs of 0.938 (95% CI: 0.874–0.970, p < 0.001) for COB and 0.923 (95% CI: 0.847–0.962, p < 0.001) for CIB.

Failure Mode Characterization

The causes of inaccuracy are detailed in Table 1. While CIB errors were primarily driven by image quality issues, COB failures were largely attributed to hypertransmission artifacts specific to GA area. Specifically, signal washout and hyper-transmission defects accounted for approximately 40% of COB errors, where the increased light penetration through the atrophic RPE obscured the choroidal-scleral interface.

Workflow Efficiency and Clinical Utility

Despite the limitations in COB accuracy requiring human verification, the AI-assisted workflow demonstrated a dramatic improvement in efficiency compared to manual grading. Manual-only segmentation of the full choroid (CIB and COB) across a complete 97-slice volume was extremely labor-intensive, requiring an average of approximately 7 hours per volume by a single experienced ophthalmologist. In contrast, the "AI-assisted" workflow, comprising automated inference followed by expert correction of artifacts, took an average of 45 minutes per volume by a single experienced ophthalmologist. Although the model frequently misidentified the COB in areas of hyper-transmission (necessitating human correction), the AI provided a foundational structure that reduced the total human effort by approximately 90%.

4. Discussion

In this study, we performed a failure analysis of automated choroidal boundary detection in GA eyes using a validated a choroid segmentation tool (NMI-CHoroidAI), focusing on the markedly different behavior of the model when estimating the CIB versus the COB. While the CIB remained highly detectable even in advanced GA (94.8% grade-0 accuracy), COB definition consistently deteriorated in areas of hyper-transmission, reaching only 81.0% strict accuracy.
Compared with prior studies, which have predominantly evaluated choroidal boundary detection in healthy eyes or non-GA AMD, our findings demonstrate a critical gap in literature that is current automated choroidal segmentation methods rarely account for GA-specific optical artifacts. Several high-performing deep-learning models have reported strong choroidal and retinal boundary accuracy under normal imaging conditions [10,11], yet these results do not necessarily translate to GA eyes, where loss of the RPE results in marked hyper-transmission and signal washout. Consistent with this, recent research on GA imaging has emphasized that RPE loss fundamentally alters OCT contrast characteristics [12] which result in producing structural ambiguities at the CSI that complicate automated COB estimation. Our strict COB accuracy of 81% reflects a pathology-driven limitation rather than a model flaw alone. However, when scans with minor inaccuracies were included, the model demonstrated improved performance, achieving an overall accuracy of 94.2%. Only 13.2% of scans required minor manual corrections, substantially reducing the time required and overall workload.
Accurate and reproducible delineation of the choroidal inner and outer boundaries is clinically crucial; as choroidal thickness and vascular metrics are increasingly investigated as imaging biomarkers and potential exploratory endpoints in AMD and GA research, as well as interventional trials [2,16,17]. Manual slice-by-slice segmentation is prohibitive in volume, explaining why AI-assisted workflows are appealing. In this study, AI-assisted processing reduced the total time to about 45 minutes per volume, representing nearly a 90% reduction in human effort. This pattern is consistent with prior AI-based OCT pipelines, which have demonstrated substantial time savings in clinical trial settings while still relying on human verification rather than fully autonomous deployment [18,19,20]. In this context, our model should therefore be considered as a workflow accelerator, providing a plausible anatomical estimate whose COB accuracy requires expert verification before use in trials.
The strong performance on the CIB and the weaker performance on the COB arise from fundamentally different signal characteristics within SD-OCT volumes. The CIB corresponds to a high-contrast RPE–Bruch’s membrane transition, which remains anatomically recognizable even in eyes with advanced cRORA [12,21]. This structural stability allows deep networks to detect the CIB reliably across varying image quality and patient-dependent fluctuations [10]. In contrast, the COB does not represent a discrete optical interface in GA eyes, but rather a gradual attenuation zone whose appearance varies with the degree of RPE loss, choroidal vascular density, and local scleral reflectivity. As a result, the model frequently encounters regions where no clear boundary exists at all, making errors unavoidable regardless of architectural complexity.
Another factor influencing performance is the directionality of OCT signal propagation. The hyper-transmission that accompanies RPE loss increases light penetration into deeper tissues but simultaneously reduces the local contrast between the choroid and sclera. This “signal dilution” effect decreases the sharpness of the true boundary and increases the probability that the model converges on an anatomically invalid but visually dominant tailing artifact. This behavior is consistent with histopathologic and imaging observations describing the fading of the CSI in cRORA [12]. These effects can also influence boundary consistency across adjacent B-scans: while slice-to-slice stabilization steps may improve the continuity of the CIB, they offer limited benefit for the COB because its visibility fluctuates along the scan. Some preprocessing strategies have been reported to stabilize retinal and choroidal boundaries in automated OCT analysis [10,11]. In regions where hyper-transmission abruptly increases or posterior ciliary vessels locally dominate the intensity profile, preprocessing cannot reconstruct a boundary that is physically absent.
NMI-ChoroidAI tool integrates manual editing within the automated segmentation workflow, allowing users to refine results in real time while viewing OCT images. Importantly, corrections are required only for boundary segments that deviate from the true anatomical layer, rather than redrawing the entire contour. This selective adjustment approach greatly improves efficiency and enables rapid generation of high-quality segmentations.
Furthermore, architectural elements such as residual connections and attention mechanisms help the model focus on coherent, layer-specific patterns rather than isolated noise. Residual learning has been shown to improve detection of fine structural transitions in medical image segmentation [22,23], while attention mechanisms enhance the network’s ability to prioritize relevant anatomical features and suppress background clutter [24,25,26]. While this improves CIB detection, it also means the model tends to ignore faint or discontinuous deep signals, which in GA lesions are often the only remaining cues for COB localization. Thus, architecture inadvertently reinforces the performance gap: robust for structures with preserved reflectivity, fragile for structures whose visibility deteriorates under disease-specific optical conditions.

Limitations

Several limitations of this study should be acknowledged. First, the analysis was performed exclusively on eyes with established GA that met Classification of atrophy meetings (CAM) criteria for cRORA. Earlier stages such as iRORA, which exhibit only partial RPE-photoreceptor disruption, were not included. Because the optical visibility of the CSI degrades progressively along the iRORA to cRORA continuum, the present model may not generalize to earlier atrophic stages where the boundary is partially preserved and the pattern of hyper-transmission differs [21].
Second, although the model demonstrated high reliability for the CIB, COB performance was strongly constrained by OCT physics rather than by network architecture. In regions of pronounced hyper-transmission, the choroid–sclera interface becomes attenuated or disappears entirely, resulting in true signal ambiguity. As described in structural and histopathologic studies of GA, this washout of the deeper interface is an inherent consequence of RPE loss [12]. In such regions, even a theoretically ideal model would lack sufficient reflectivity cues to localize the COB consistently, limiting the achievable performance ceiling.
Third, the dataset consisted of SD-OCT volumes from a single imaging platform (Heidelberg Spectralis), all with relatively high acquisition quality. Device-specific differences such as axial resolution, sensitivity roll-off, and noise characteristics may affect the appearance of the COB. Previous work has shown that automated OCT algorithms may degrade when transferred across devices without retraining [9]. Broader validation across additional devices, including swept-source OCT with deeper penetration profiles, will therefore be required.
Finaly, all segmentations were evaluated in a curated dataset without the full range of real-world image artifacts (motion, low SNR, shadowing from media opacities). While registration and curvature normalization can reduce inter-scan variability, they cannot reconstruct a boundary that is not visible due to washout. Consequently, automated COB estimation in lower-quality OCT scans may be more error-prone.

Prospective Trajectories

Future work should focus on addressing the optical limitations that underline COB failures. Because the CSI progressively fades along the CAM-defined iRORA to cRORA spectrum, incorporating stage-specific annotations may help models learn how boundary visibility deteriorates with advancing atrophy in GA patients [21]. Additional progress will require artifact-aware or physics-informed approaches capable of distinguishing true boundary loss from hyper-transmission–related washout [12]. Broader training across multiple OCT platforms is also needed, as automated algorithms often show reduced performance when transferred between devices [9]. Together, these steps may reduce analysis time further, but human verification will still be required for clinical-grade outputs.

5. Conclusions

This study provides a failure analysis of automated choroidal boundary detection in GA eyes imaged with Spectralis SD-OCT. While the model reliably identified the CIB, COB detection was substantially limited by GA-related optical artifacts, particularly hyper-transmission in cRORA lesions. Despite these inaccuracies, the AI-assisted workflow reduced segmentation time from approximately seven hours to forty-five minutes per volume (1/10th), underscoring its value as an efficiency tool rather than an autonomous solution. Improved handling of deep signal loss, stage-specific modeling across the iRORA–cRORA spectrum, and broader cross-device training will be required before automated COB measurements can be used reliably in future GA studies.

Author Contributions

Conceptualization, ER and ES.; methodology, ER and NH.; software, AG, SCB and KKR.; validation, NH and JC.; formal analysis, ER.; investigation, JW, RS and SL.; resources, JC, MJS and KKV.; data curation, CS JW, RS and SL.; writing—original draft preparation, ER; writing—review and editing, ER, NH and JC..; supervision, JC.;. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by NIH CORE Grant P30 EY08098 to the Department of Ophthalmology, the Eye and Ear Foundation of Pittsburgh, and an unrestricted grant from Research to Prevent Blindness, New York, NY.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki. Ethical review and approval were waived for this study due to retrospective nature of the study.

Informed Consent Statement

Patient consent was waived due to retrospective nature of the study.

Data Availability Statement

Data availability may be provided upon reasonable request to the corresponding author.

Conflicts of Interest

SCB, KKV, and JC are owners of NetraMind Innovations Inc. This affiliation had no influence on the study design, data analysis, interpretation, or manuscript preparation. Other authors declare no competing interests.

Abbreviations

The following abbreviations are used in this manuscript:
GA Geographic atrophy
SD-OCT Spectral domain optical coherence tomography
CIB Choroidal inner boundary
COB Choroidal outer boundary
NMI NetraMind Innovations
AI Artificial Intelligence
AMD Age related macular degeneration
RPE Retinal pigment epithelium
CSI choroidoscleral interface
ICC Intraclass coefficient
cRORA complete RPE and outer retinal atrophy
CAM Classification of Atrophy Meetings

References

  1. Wong, W.L.; Su, X.; Li, X.; Cheung, C.M.; Klein, R.; Cheng, C.Y.; et al. Global prevalence of age-related macular degeneration and disease burden projection for 2020 and 2040: A systematic review and meta-analysis. Lancet Glob Health 2014, 2(2), e106–16. [Google Scholar] [CrossRef] [PubMed]
  2. Sadeghi, E.; Valsecchi, N.; Rahmanipour, E.; Ejlalidiz, M.; Hasan, N.; Vupparaboina, K.K.; et al. Choroidal biomarkers in age-related macular degeneration. Survey of Ophthalmology 2025, 70(2), 167–183. [Google Scholar] [CrossRef] [PubMed]
  3. Takahashi, A.; Ooto, S.; Yamashiro, K.; Tamura, H.; Oishi, A.; Miyata, M.; et al. Pachychoroid Geographic Atrophy: Clinical and Genetic Characteristics. Ophthalmol Retina 2018, 2(4), 295–305. [Google Scholar] [CrossRef] [PubMed]
  4. Sacconi, R.; Battista, M.; Borrelli, E.; Senni, C.; Tombolini, B.; Grosso, D.; et al. CHOROIDAL VASCULARITY INDEX IS ASSOCIATED WITH GEOGRAPHIC ATROPHY PROGRESSION. Retina 2022, 42(2), 381–387. [Google Scholar] [CrossRef]
  5. Querques, G.; Costanzo, E.; Miere, A.; Capuano, V.; Souied, E.H. Choroidal Caverns: A Novel Optical Coherence Tomography Finding in Geographic Atrophy. Investigative Ophthalmology & Visual Science 2016, 57(6), 2578–2582. [Google Scholar] [CrossRef]
  6. Arora, S.; Singh, S.R.; Rosario, B.; Ibrahim, M.N.; Selvam, A.; Zarnegar, A.; et al. Three-dimensional choroidal contour mapping in healthy population. Scientific Reports 2024, 14(1), 6210. [Google Scholar] [CrossRef]
  7. Valsecchi, N.; Sadeghi, E.; Davis, E.; Ibrahim, M.N.; Hasan, N.; Bollepalli, S.C.; et al. Assessment of choroidal vessels in healthy eyes using 3-dimensional vascular maps and a semi-automated deep learning approach. Scientific Reports 2025, 15(1), 714. [Google Scholar] [CrossRef]
  8. Al-khersan, H.; Sodhi, S.K.; Cao, J.A.; Saju, S.M.; Pattathil, N.; Zhou, A.W.; et al. Deep Learning-Based Segmentation of Geographic Atrophy: A Multi-Center, Multi-Device Validation in a Real-World Clinical Cohort. Diagnostics [Internet] 2025, 15(20). [Google Scholar] [CrossRef]
  9. Zhang, G.; Fu, D.J.; Liefers, B.; Faes, L.; Glinton, S.; Wagner, S.; et al. Clinically relevant deep learning for detection and quantification of geographic atrophy from optical coherence tomography: A model development and external validation study. Lancet Digit Health 2021, 3(10), e665–e675. [Google Scholar] [CrossRef]
  10. Mazzaferri, J.; Beaton, L.; Hounye, G.; Sayah, D.N.; Costantino, S. Open-source algorithm for automatic choroid segmentation of OCT volume reconstructions. Scientific Reports 2017, 7(1), 42112. [Google Scholar] [CrossRef]
  11. Niu, S.; de Sisternes, L.; Chen, Q.; Leng, T.; Rubin, D.L. Automated geographic atrophy segmentation for SD-OCT images using region-based CV model via local similarity factor. Biomedical optics express 2016, 7(2), 581–600. [Google Scholar] [CrossRef] [PubMed]
  12. Fleckenstein, M.; Mitchell, P.; Freund, K.B.; Sadda, S.; Holz, F.G.; Brittain, C.; et al. The Progression of Geographic Atrophy Secondary to Age-Related Macular Degeneration. Ophthalmology 2018, 125(3), 369–390. [Google Scholar] [CrossRef] [PubMed]
  13. Derradji, Y.; Mosinska, A.; Apostolopoulos, S.; Ciller, C.; De Zanet, S.; Mantel, I. Fully-automated atrophy segmentation in dry age-related macular degeneration in optical coherence tomography. Scientific Reports 2021, 11(1), 21893. [Google Scholar] [CrossRef] [PubMed]
  14. Klein, R.; Meuer, S.M.; Knudtson, M.D.; Klein, B.E. The epidemiology of progression of pure geographic atrophy: The Beaver Dam Eye Study. Am J Ophthalmol. 2008, 146(5), 692–699. [Google Scholar] [CrossRef]
  15. Vupparaboina, K.K.; et al. Accurate Detection of 3D Choroidal Vasculature Using Swept-Source OCT Volumetric Scans Based on Phansalkar Thresholding. 2023 IEEE EMBS International Conference on Biomedical and Health Informatics (BHI), Pittsburgh, October 15 2023; pp. 1–4. [Google Scholar]
  16. Sadda, S.R.; Chakravarthy, U.; Birch, D.G.; Staurenghi, G.; Henry, E.C.; Brittain, C. Clinical endpoints for the study of geographic atrophy secondary to age-related macular degeneration. Retina 2016, 36(10), 1806–1822. [Google Scholar] [CrossRef]
  17. Vallino, V.; Berni, A.; Coletto, A.; Serafino, S.; Bandello, F.; Reibaldi, M.; et al. Structural OCT and OCT angiography biomarkers associated with the development and progression of geographic atrophy in AMD. Graefe’s Archive for Clinical and Experimental Ophthalmology 2024, 262(11), 3421–3436. [Google Scholar] [CrossRef]
  18. Enzendorfer, M.L.; Tratnig-Frankl, M.; Eidenberger, A.; Schrittwieser, J.; Kuchernig, L.; Schmidt-Erfurth, U. Rethinking Clinical Trials in Age-Related Macular Degeneration: How AI-Based OCT Analysis Can Support Successful Outcomes. Pharmaceuticals 2025, 18(3), 284. [Google Scholar] [CrossRef]
  19. Fu, D.J.; Glinton, S.; Lipkova, V.; Faes, L.; Liefers, B.; Zhang, G.; et al. Deep-learning automated quantification of longitudinal OCT scans demonstrates reduced RPE loss rate, preservation of intact macular area and predictive value of isolated photoreceptor degeneration in geographic atrophy patients receiving C3 inhibition treatment. British Journal of Ophthalmology 2024, 108(4), 536–545. [Google Scholar]
  20. Rahmanipour, E.; Afazel, S.; Ashrafi, S.; Sadeghi, E.; Ghorbani, M.; Zarranz-Ventura, J.; et al. Artificial intelligence in ophthalmology clinical trials: A narrative review. In Graefe’s Archive for Clinical and Experimental Ophthalmology; 2025. [Google Scholar]
  21. Guymer, R.H.; Rosenfeld, P.J.; Curcio, C.A.; Holz, F.G.; Staurenghi, G.; Freund, K.B.; et al. Incomplete Retinal Pigment Epithelial and Outer Retinal Atrophy in Age-Related Macular Degeneration: Classification of Atrophy Meeting Report 4. Ophthalmology 2020, 127(3), 394–409. [Google Scholar] [CrossRef]
  22. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition; He, K., Zhang, X., Ren, S., Sun, J., Eds.; 2016. [Google Scholar]
  23. Sahoo, J.; Saini, S.K.; singh, S.; Saxena, A.K.; Sharma, S.; Awasthi, A.; et al. Residual learning for segmentation of the medical images in healthcare. Measurement: Sensors 2024, 32, 100998. [Google Scholar] [CrossRef]
  24. Cao, G.; Wu, Y.; Peng, Z.; Zhou, Z.; Dai, C. Self-attention CNN for retinal layer segmentation in OCT. Biomedical Optics Express 2024, 15(3), 1605–1617. [Google Scholar] [CrossRef]
  25. Soni, T.; Gupta, S.; Almogren, A.; Altameem, A.; Rehman, A.U.; Hussen, S.; et al. ARCUNet: Enhancing skin lesion segmentation with residual convolutions and attention mechanisms for improved accuracy and robustness. Sci Rep. 2025, 15(1), 9262. [Google Scholar] [CrossRef]
  26. Xie, Y.; Yang, B.; Guan, Q.; Zhang, J.; Wu, Q.; Xia, Y. Attention mechanisms in medical image segmentation: A survey. arXiv 2023, arXiv:2305.17937. [Google Scholar] [CrossRef]
Figure 1. An 82-year-old male with fovea-involving geographic atrophy (GA). (A) Infrared reflectance image demonstrating the extent of GA and the location of the OCT scan. (B) Corresponding OCT B-scan showing diffuse hypertransmission and the presence of outer retinal tubulations, consistent with GA. (C) OCT B-scan with automated segmentation of the choroid, demonstrating largely accurate choroidal boundary delineation with a minor inaccuracy (black circle) in the inner choroidal boundary (red line) at the temporal edge (black circle). (D) Manually corrected segmentation with adjustment of the previously noted inaccuracy.
Figure 1. An 82-year-old male with fovea-involving geographic atrophy (GA). (A) Infrared reflectance image demonstrating the extent of GA and the location of the OCT scan. (B) Corresponding OCT B-scan showing diffuse hypertransmission and the presence of outer retinal tubulations, consistent with GA. (C) OCT B-scan with automated segmentation of the choroid, demonstrating largely accurate choroidal boundary delineation with a minor inaccuracy (black circle) in the inner choroidal boundary (red line) at the temporal edge (black circle). (D) Manually corrected segmentation with adjustment of the previously noted inaccuracy.
Preprints 196561 g001
Figure 2. Representative OCT B-scans illustrating varying degrees of inaccuracy in automated segmentation of the choroidal outer boundary (COB) (blue line). (A) Minor inaccuracy (<33%), beginning at the junction of the hypertransmission area. (B) Moderate inaccuracy (33–66%), observed in regions of marked choroidal thinning starting from the entry of a posterior ciliary vessel. (C) Major inaccuracy (>66%), starting at the junction of the hypertransmission area. (D) Near-complete segmentation failure (approximately 100% inaccuracy), associated with severe choroidal thinning and poor image quality.
Figure 2. Representative OCT B-scans illustrating varying degrees of inaccuracy in automated segmentation of the choroidal outer boundary (COB) (blue line). (A) Minor inaccuracy (<33%), beginning at the junction of the hypertransmission area. (B) Moderate inaccuracy (33–66%), observed in regions of marked choroidal thinning starting from the entry of a posterior ciliary vessel. (C) Major inaccuracy (>66%), starting at the junction of the hypertransmission area. (D) Near-complete segmentation failure (approximately 100% inaccuracy), associated with severe choroidal thinning and poor image quality.
Preprints 196561 g002
Figure 3. Confusion matrices for choroidal boundary grading. (A) Choroidal inner boundary (CIB) and (B) choroidal outer boundary (COB) using a four-level error scale: Accurate, Minor (≤33% of the B-scan deviated), Moderate (33–66%), Major (>66% or missed). Rows denote ground-truth class and columns denote model predictions. Because a true anatomical boundary exists on every slice, all reference labels fall in the “Accurate” row. Counts shown in each cell indicate the number of B-scans assigned to each predicted category (CIB: 5,427 Accurate, 144 Minor, 58 Moderate, 94 Major; COB: 4,636 Accurate, 755 Minor, 130 Moderate, 202 Major), illustrating that most slices were classified as error-free and that most residual errors were minor.
Figure 3. Confusion matrices for choroidal boundary grading. (A) Choroidal inner boundary (CIB) and (B) choroidal outer boundary (COB) using a four-level error scale: Accurate, Minor (≤33% of the B-scan deviated), Moderate (33–66%), Major (>66% or missed). Rows denote ground-truth class and columns denote model predictions. Because a true anatomical boundary exists on every slice, all reference labels fall in the “Accurate” row. Counts shown in each cell indicate the number of B-scans assigned to each predicted category (CIB: 5,427 Accurate, 144 Minor, 58 Moderate, 94 Major; COB: 4,636 Accurate, 755 Minor, 130 Moderate, 202 Major), illustrating that most slices were classified as error-free and that most residual errors were minor.
Preprints 196561 g003
Table 1. Causes of inaccuracy in each boundary.
Table 1. Causes of inaccuracy in each boundary.
CIB
Low image quality/signal loss ~45 % Degraded scan contrast, media opacities, or defocus led to blurred definition of the RPE–choroid interface.
Sudden change in RPE curvature ~30 % Abrupt topographic variations of the RPE, particularly in parafoveal regions or near drusen, caused local mis-registration of the CIB line.
Fixation change/motion artifact ~20 % Horizontal eye movement between adjacent B-scans slightly displaced the inner choroidal contour, particularly near the disc margin.
Combined effects (e.g., curvature + signal drop-off) ~5 % Compounded quality issues or irregular RPE slopes caused limited regional deviations.
COB
Signal washout/Hyper-transmission ~40 % Hyper-transmission artifacts beneath the atrophic RPE caused the choroidal-scleral interface to fade or become indistinguishable from the sclera.
Fixation change/motion artifact ~25 % Minor eye movement between scans distorted local curvature of the choroid–sclera interface.
Presence of large posterior ciliary vessels ~20 % These vessels create local discontinuities or high-contrast streaks that confuse the model’s estimate of the scleral border.
Combined low quality + motion ~15 % Concurrent factors caused larger regional deviations (> 66 %).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated