Preprint
Article

This version is not peer-reviewed.

Semi-Automatic Refinement of Myocardial Segmentations for Better LVNC Detection

A peer-reviewed article of this preprint also exists.

Submitted:

26 November 2024

Posted:

27 November 2024

You are already at the latest version

Abstract
Accurate segmentation of the left ventricular myocardium in cardiac MRI is essential for developing reliable deep learning models to diagnose Left Ventricular Non-Compaction Cardiomyopathy (LVNC). This work focuses on improving the segmentation database used to train these models, enhancing the quality of myocardial segmentation for more precise model training. We present a semi-automatic framework that refines segmentations through three fundamental approaches: 1) Combining neural network outputs with expert-driven corrections, 2) Implementing a blob-selection method to correct segmentation errors and neural network hallucinations, and \textcolor{black}{3) Employing a cross-validation process using the baseline U-Net model. Applied to datasets from three hospitals, these methods demonstrate improved segmentation accuracy, with the blob-selection technique boosting the Dice coefficient for the trabecular zone up to 0.06 in certain populations. Our approach enhances the dataset’s quality, providing a more robust foundation for future LVNC diagnostic models.
Keywords: 
;  ;  ;  

1. Motivation and Objectives

Cardiovascular diseases remain the leading cause of mortality worldwide, accounting for approximately 32% of all deaths globally [1,2]. An early and accurate detection of cardiac anomalies is crucial for improving patient outcomes. Among these illnesses, Left Ventricular Non-Compaction Cardiomyopathy (LVNC) is a severe but rare cardiac disorder characterized by excessive trabeculations and deep recesses in the left ventricular myocardium due to incomplete myocardial compaction during embryonic development. LVNC can lead to a spectrum of symptoms ranging from fatigue and dyspnea to heart failure and is associated with an increased risk of arrhythmias and thromboembolic events.
Magnetic Resonance Imaging (MRI) is the preferred modality for diagnosing LVNC, providing high-resolution images that allow detailed visualization of myocardial structures. Quantifying the degree of trabeculation is essential for diagnosing LVNC [3,4,5,6] and assessing its severity, typically achieved by calculating the percentage of trabeculated myocardial volume (see Section 2.1).
Accurate segmentation of the left ventricular myocardium is crucial for the reliable quantification of trabeculated myocardial volume. Although manual segmentation methods are precise, they are labor-intensive, time-consuming, and susceptible to inter-observer variability. There has been significant interest in developing automated segmentation techniques utilizing computer vision algorithms. Then, we proposed QLTVHC [7], a semi-automated software tool leveraging medical expertise, which was developed to quantify the extent of non-compaction using cardiac MRI. This tool employs conventional computer vision techniques and requires cardiologists to adjust manually. Subsequently, we proposed an automatic tool called SOST [8] to complement the previous one by applying a least squares adjustment.
Another way to automatically segment images is using deep neural networks (DNNs). Due to their ability to learn complex patterns directly from data, DNNs have been successfully applied across various scientific domains, including medical imaging [9,10,11,12]. DNNs are now considered foundational tools in modern Artificial Intelligence (AI) [13]. We also explored this approach. In previous work, we showed that the U-Net model can accurately segment the left ventricle for detecting LVNC [14]. Our approach utilized a fully automatic segmentation method like DL-LVTQ [15]. DL-LVTQ is a novel automated method based on a U-Net architecture [15] designed specifically for diagnosing LVNC. However, the segmentations produced, while promising, are not flawless, limiting our ability to optimize the performance of our models.
Then, to improve segmentation accuracy in myocardial images, in previous papers [14,16] we increased the dataset size, tested more advanced neural network architectures (MSA-UNet, AttUNet, and U-Net++), and implemented a clustering algorithm to reduce hallucinations. However, our best results reached a Dice score of 0.87 in the trabecular zone on a particular subset of the dataset (P).
Building upon these advancements, we recognize the potential of semi-automatic segmentation methods in improving both the diagnosis of LVNC and the segmentation images used for training these models. Given the complexity of the left ventricular trabeculations and the limitations of fully automatic methods for creating the training datasets, an approach that combines the strengths of automatic segmentation with expert validation can enhance accuracy and efficiency.
In this work, we propose a semi-automatic method to further enhance the quality of these segmentations. Combining user-friendly interfaces with predictions from our existing neural network models allows for the refinement of segmentations through minimal manual input, offering a balance between efficiency and accuracy.
The main objective of this paper is to enhance the accuracy and reliability of myocardial image segmentation by improving the quality of the training datasets used for deep learning models. We have developed a semi-automatic image correction framework designed to refine and correct segmentations within the training data. The framework tackles common issues such as incomplete border coloring, hallucinated structures, and general inconsistencies by combining multiple neural network outputs with user intervention. To achieve this, this work comprehends:
  • A Semi-Automatic Image Fixing Method: We have created a cross-validation framework that utilizes multiple neural network models to identify and correct errors in image segmentation. We aim to propose a fixer for each image by training several models and comparing their outputs.
  • An Image Fixer Program: We have designed a user-friendly software tool for interactive image segmentation correction. The program displays the original segmentation alongside the corrected version, highlighting differences and enabling clinical experts to apply targeted adjustments efficiently.
  • An Image Selector Tool: We have developed an application that presents multiple segmentation options for a given image and allows users to select the most accurate one. This tool facilitates subjective evaluation and selection, ensuring that the final segmentation used for diagnosis is of the highest quality.
Through these initiatives, we seek to refine existing segmentation methods by leveraging automated correction techniques and expert validation. By improving the training dataset to reflect the true myocardial segmentation better, we develop more reliable models for LVNC diagnosis and contribute to better patient outcomes.

2. Background

2.1. Left Ventricular Non-Compaction Cardiomyopathy (LVNC)

Left Ventricular Non-Compaction Cardiomyopathy (LVNC) is characterized by excessive trabeculations and deep recesses in the left ventricular myocardium. LVNC is usually diagnosed using MRI, which provides high-resolution images that allow for detailed visualization of myocardial structures.
LVNC is diagnosed by quantifying the degree of trabeculation, which is essential for even assessing its severity. To calculate the percentage of trabeculated myocardial volume, we use the equation:
V T % = T Z T Z + E L · 100
where TZ represents the volume of the trabeculated zone and EL denotes the volume of the compact external layer. A threshold of 27.4% is commonly used as a diagnostic criterion [3], with values above this indicating significant hyper trabeculation consistent with LVNC.

2.2. Semi-automatic segmentation methods

Semi-automatic segmentation methods have gained widespread use in medical imaging. For example, a semi-automatic approach to facial wrinkle detection reduces manual effort while enhancing detection accuracy by focusing on intricate textures [17]. In bladder cancer diagnosis, semi-automatic segmentation achieves results comparable to manual methods, significantly reducing processing time [18]. Similarly, in dental imaging, a semi-automatic coarse-to-fine segmentation method utilizing intuitive single clicks effectively segments complex 3D tooth models, improving efficiency and accuracy over fully automatic approaches [19]. The use of these methods in rectal cancer tumor segmentation further highlights the growing preference for semi-automatic approaches, which streamline workflows without compromising precision [20].

3. Materials and Methods

We develop a semi-automatic correction method to address segmentation errors in a database of images, including issues such as incomplete border coloring, hallucinated trabeculae in empty regions, and general inconsistencies.
Although the images generated from the proposed method here are automatic, we call it semi-automatic since many of the changes provided by these methods are not desired and should, therefore, be sorted out. For this reason, we have developed a user-friendly application to refine and compare these segmentations efficiently.

3.1. Cross-validation method

The cross-validation method splits the dataset into five equal parts (called folds). For each fold, one part is used as the validation set, while the remaining four are used to train the neural network model. Figure 1 shows this method. This process is repeated five times, with each part used as the validation set precisely once, resulting in a total of 5 neural nets being trained.
For each image, we look at the four outputs of neural nets that have used that image in the training dataset. For each pixel of that image, we look at the four outputs, and whether three of them contain the same output class, that will be the new value for that pixel, if not the value will remain the same as the original.
Modifying only images in the training dataset allows the model to learn that image in depth making it unlikely to do any big changes, leading to more stable results than if we did a normal cross-validation.
We use the baseline U-Net in this method due to the lengthy training times of neural networks, with the same architecture applied across all training sets, as shown in Figure 2. As the original segmentations were made at 800x800, we upscaled our images to 800x800 but then cropped them to 384x384 to focus on the left ventricle. Our U-Net takes these 384x384 images as input, as the images are cropped to focus on the left ventricle, eliminating the need for additional downsampling. Each step of the encoder path includes two 3x3 convolutions with instance normalization and leaky ReLU (0.01 negative slope), followed by a stride-two convolution to reduce spatial dimensions. The resulting network has different filters in each convolutional block, ranging from 64 in the first layer to 1024 in the bottom layer. The U-Net then outputs a segmentation map with the same 384x384 resolution, with four channels representing each class’s probabilities.
This automated framework minimizes the need for manual segmentation adjustments by offering modification suggestions to clinical experts, enabling them to identify patterns that might otherwise go unnoticed.

3.2. Neural nets used for testing

After fixing the images, we perform cross-validation with U-Net on train/val/test sets, training on full-sized 800x800 images to allow a fair comparison to the U-Net model from our previous work [14]. We adopt the same architecture as the prior study to handle these larger images effectively, incorporating additional layers designed to process 800x800 inputs. This ensures both consistency in architecture and the capacity to manage the increased resolution. First, images are downsampled to 200x200 using two 3x3 convolutions with instance normalization and leaky ReLU activations (negative slope 0.01), each with a stride of two, effectively halving each dimension. The model then processes this 200x200 input and outputs at the same resolution.
To upscale back to 800x800 before the final output, we add a decoding layer comprising two transposed convolutions (stride of two), each followed by two 3x3 convolutions with instance normalization and leaky ReLU. This configuration restores the spatial resolution to 800x800, preparing the model for the final segmentation output.

3.3. Image fixer

The goal of this application’s part is to obtain a segmentation for our image that is closest to what we believe to be as correct as possible. To facilitate this, we use two different segmentation methods.
This program presents two images as shown in Figure 3. The image on the left is obtained from our previous method (QLVTHC), while the one on the right is obtained from our neural networks (SubSection 3.1). The final output of our program is the segmentation on the right.
We use a muted color scheme to make the colors distinguishable for colorblind people. In the image on the right, the External Layer is olive, the Internal Cavity is cyan, and the Trabecular Zone is rose.
On the figure on the left, we mark the differences between both segmentations. This way, where there is an External layer on the left and something else on the right, we mark it with green. For additional Internal Cavity, we mark it with blue; for Trabeculae, we mark it with wine; for Background, we mark it with purple.
The differences between both images can be leveraged for easy transformations of the output image. The transformation is applied to the image on the right by simply clicking one of the differences. For example, if we click a blob that is colored green on the left (meaning an additional External Layer), an External Layer will appear in that zone of the image on the right. Some of these blobs are very small, so you can select them by right-clicking and dragging over them for ease of use.
Painting directly on the output image (image on the right) is also possible. For this, we select either BG (Background), EL (External Layer), IC (Internal Cavity) or TZ (Trabecular Zone). Then, simply left-clicking on where we want to paint (shown in Figure 4).
When we finish the image, we just save the image and go to the next image automatically. However, if we want to restart the editing of the image, we can refresh it and erase the modifications made.
Finally, we can also toggle the segmentation which helps view the borders’ coloring. It is important to note that you can still paint when the image is toggled to the raw image.

3.4. Image Selector

This tool allows the user to select the most accurate segmentation for a given image from three different segmentations. The process is designed as a blind test to ensure an unbiased evaluation, allowing the user to choose the segmentation that most closely represents reality.
The program presents four images in a grid, as shown in Figure 5. The raw image is presented on the top-left, while the other three images are segmentations obtained differently. These three images are randomly placed in each iteration for this to be a blind test.
On top of each image, we show the percentage of image trabeculation for the given segmentation (Eq. 1).
Again, we use a muted color scheme to make the colors distinguishable for colorblind people. The External Layer is olive, the Internal Cavity is cyan, and the Trabecular Zone is rose.
To select an image, you have to left-click on the image, and then a green box appears around the selected image (see Figure 6). On the right, we can choose a mark of 1-5, indicating how good the image is. Table 1 shows the subjective evaluation scale proposed in [21]. Finally, we can save it, which automatically brings up the next batch.
Additionally, we can give feedback in the textbox on the left. We can also indicate whether the quality of the image is bad by selecting the checkbox at the center top.
It is possible to zoom into an image using the mouse wheel. It is also possible to toggle between viewing each image as the raw image for easy comparison between borders.
The three images presented (see Figure 5 and Figure 6) are obtained using the following methods:
  • Original targets: QLVTHC output [8].
  • Blob-selection method: We compare the cross-validation method (Section 3.1) for improving images to the QLVTHC method on the image fixer. We improve the images only by choosing the best blobs provided by their differences.
  • Manually-fixed method: We manually fix 100 images by coloring on the image fixer (left-click). We make the cross-validation method on these 100 fixed images and continue manually fixing 400 more images with these outputs as a new base.

3.5. Datasets

The datasets used in this study are derived from three hospitals: Virgen de la Arrixaca of Murcia (HVAM), Mesa del Castillo of Murcia (HMCM), and Universitari Vall d’Hebron of Barcelona (HUVHB). These hospitals provide the medical imaging data for the analysis and contribute a variety of patient profiles that enrich the study.
HVAM operates two scanners, one from Philips and one from General Electric, with a field strength of 1.5 T. The acquisition matrices for these scanners are 256 × 256 pixels and 224 × 224 pixels, respectively, with pixel spacing of 1.5 × 1.5 × 0.8 mm and 1.75 × 1.75 × 0.8 mm. HMCM uses a General Electric model scanner identical to HVAM’s, while HUVHB utilizes a 1.5 T Siemens Avanto scanner with an acquisition matrix of 224 × 224 pixels. For all institutions, the images were captured using balanced steady-state free precision (b-SSFP) sequences. The primary parameters for the scans, including echo time (1.7 ms), flip angle (60º), slice thickness (8 mm), slice gap (2 mm), and 20 imaging phases, were consistent across all hospitals. All patients underwent the scans while in apnea, synchronized with ECG, and without using contrast agents.
The original dataset comprises data from three subsets: P, X, and H. Set P, from HVAM, consists of 293 patients (2381 slices) with hypertrophic cardiomyopathy (HCM). Set X, from HMCM, includes 58 patients (467 slices) with various heart conditions, including HCM. Finally, set H, from HUVHB, comprises 28 patients (196 slices) diagnosed with left ventricular non-compaction cardiomyopathy (LVNC) according to the Petersen criteria.
Given the time-intensive nature of manual segmentation methods, we use a representative subset of 545 modified segmentations from the original dataset. This subset includes 355 slices from the P dataset, 75 from the X dataset, and 115 from the H dataset. The selection process ensures that the larger dataset’s diversity of heart conditions and image characteristics is still adequately represented in the smaller subset.
To enable comparison with the results of our previous U-Net model, we train our new models using a train/validation/test split. From our dataset of 545 images, we retain the 113 images used as a test in our prior work [14]. The remaining 432 images were divided into 5 non-overlapping folds, ensuring no patient data was shared between folds, allowing us to create a cross-validation dataset. For the creation of these training folds, 297 images were from P, 53 from X, and 82 from H, while for testing, 58 were from P, 22 from X, and 33 were from H.

4. Results and Discussion

This section presents the results obtained by applying the blob-selection and manual correction methods to the segmentation tasks. First, we compare the performance of these methods against the original segmentation targets. Then, we analyze the differences between the methods and discuss their efficiency.

4.1. Comparison Between Models Generated by the Datasets

To determine whether modifying the dataset has led to improvements in the test Dice scores, we evaluated the cross-validation U-Nets from [14] and compared them to cross-validation U-Nets trained on the adjusted datasets. Specifically, we trained five U-Nets for the blob-selection and manually-fixed methods, evaluating each model on test images from their respective datasets. The results of these evaluations are shown in Table 2, Table 3, Table 4.
As shown in Table 2, the baseline U-Net model achieves Dice coefficients from 0.82-0.87 for the trabecular zone. Comparing this to the blob-selection and manually-fixed methods in Table 3, Table 4, there is a noticeable improvement in the trabecular Dice coefficient, increasing up to 0.06 Dice score for the H population. Both methods also increase the Dice coefficients in P by 0.02. Interestingly, the Dice coefficient for the X trabecular zone does not seem to increase, but this may be due to a lower number of X images in the 432 images used throughout the training sets.
The blob-selection method consistently outperforms both the baseline and the manually-fixed method, though the difference with the latter is slight. This is because the blob-selection method only allows changes proposed by the cross-validation method, which tends to homogenize results. While this approach reduces variability, it may also exclude changes that would increase accuracy by correctly capturing variations.
The baseline model was trained with 7 times more data per neural network than the adjusted models, yet it still performs worse. This highlights the effectiveness of the adjustments made in the blob-selection and manual methods, as they yield higher stability and accuracy despite having less training data. This is particularly evident in regions like the trabecular zone, where the adjusted models achieve improved Dice scores, underscoring the quality of segmentation produced by the new methods.

4.2. Dataset Comparison with Original Segmentation Targets

We evaluate the performance of both the blob-selection and manual correction methods by comparing their resulting segmentations with the original segmentation targets using the Dice coefficient. Table 5 and Table 6 show the average Dice coefficients for each method across the three populations: P, X, and H.
Table 5 shows that the blob-selection method achieves high Dice coefficients across all populations, with average values around 0.94. In contrast, the manual correction method yields lower Dice coefficients, as shown in Table 6, with average values ranging from 0.80 to 0.89.
The higher Dice coefficients for the blob-selection method indicate a closer similarity to the original segmentation targets. This is expected because the blob-selection method makes more conservative adjustments, while the manual correction method allows for more significant modifications that may deviate further from the original targets.

4.3. Comparison Between Blob-Selection and Manual Correction Methods

To assess how closely the blob-selection method approximates the manual corrections, we directly compare the segmentations from both methods, as shown in Table 7 with the corresponding Dice coefficients from this comparison.
The results in Table 7 indicate a slight improvement in Dice coefficients for the blob-selection method compared to those between the manual correction method and the original targets (Table 6). This slight increase of approximately 0.03 suggests that the blob-selection method aligns segmentations closer to the desired outcomes. However, the difference may not be substantial enough to fully correct the images on its own.
In addition to accuracy, we compare the time efficiency of both methods. Modifying an image using the blob-selection method takes approximately 20 to 30 seconds, while manual correction requires around 2 to 3 minutes per image. This significant difference indicates that the blob-selection method is much more efficient in terms of time.
Given the balance between accuracy and efficiency, a mixed-method approach could benefit larger datasets. Initially, a subset of images could be manually corrected to create a robust foundational model, enhancing cross-validation predictions. Subsequently, the remaining images could primarily utilize the blob-selection method, with manual adjustments as necessary. This approach not only ensures data quality and optimizes resource allocation but also leverages the strengths of both methods: the automatic blob-selection method provides a rapid and reliable baseline, while expert manual adjustments further refine the dataset, maintaining both high accuracy and flexibility for future modifications.

5. Conclusions

This paper enhances the approach to diagnosing Left Ventricular Non-Compaction Cardiomyopathy (LVNC) by introducing a semi-automatic framework designed to improve the quality of segmentations in training datasets, ultimately yielding more robust models. Utilizing cross-validation with multiple neural networks, the framework includes tools such as the Image Fixer and Image Selector to refine segmentation quality within the training dataset. This approach addresses a primary bottleneck in model effectiveness: the quality of input segmentations.
Our results demonstrate that improving segmentation quality in the training data substantially impacts the effectiveness of neural network models. Notably, despite being trained on datasets with seven times fewer images per neural network than the baseline model, the models trained on adjusted datasets using the blob-selection and manual correction methods achieved superior performance. Specifically, we observed an increase of up to 0.06 in the Dice coefficient for the trabecular zone in the H population subset when comparing models trained on the original dataset to those trained on manually improved segmentations. This improvement underscores the effectiveness of our adjustments and highlights the paramount importance of high-quality annotations in training data, especially for complex regions like the trabecular zone.
The fast image modification method (blob-selection method) led to an alignment improvement of approximately 0.03 average Dice coefficient. This leaves significant room for improvement, making it necessary to first create good segmentations via the proposed manual method.
We intend to provide clinicians with the Image Fixer tool, allowing them to refine the automatically generated segmentation immediately after receiving it, as needed. This could potentially improve both the efficiency and accuracy of future model training datasets. Once we apply these methods to all the images, we anticipate a significantly improved model, as previous papers have shown that more data leads to better, more robust models.

Funding

Grant TED2021-129221B-I00 funded by MCIN/AEI/10.13039/501100011033 and by the “European Union NextGenerationEU/PRTR”.

Data Availability Statement

The dataset used to support the findings of this study was approved by the local ethics committee and so cannot be made freely available. Requests for access to these data should be made to the corresponding author, Gregorio Bernabé, gbernabe@um.es.

Conflicts of Interest

The authors declare that there is no conflict of interest.

References

  1. World Health Organization. Cardiovascular diseases. https://www.who.int/health-topics/cardiovascular-diseases#tab=tab_1, 2022. Accessed: November 27, 2024.
  2. Namara, K.M.; Alzubaidi, H.; Jackson, J.K. Cardiovascular disease as a leading cause of death: how are pharmacists getting involved? Integrated pharmacy research & practice 2019, 8, 1–11. [Google Scholar] [CrossRef]
  3. Petersen, S.E.; Selvanayagam, J.B.; Wiesmann, F.; Robson, M.D.; Francis, J.M.; Anderson, R.H.; Watkins, H.; Neubauer, S. Left Ventricular Non-Compaction: Insights From Cardiovascular Magnetic Resonance Imaging. JACC 2005, 46, 101–105. [Google Scholar] [CrossRef] [PubMed]
  4. Jacquier, A.; Thuny, F.; Jop, B.; Giorgi, R.; Cohen, F.; Gaubert, J.Y.; Vidal, V.; Bartoli, J.M.; Habib, G.; Moulin, G. Measurement of trabeculated left ventricular mass using cardiac magnetic resonance imaging in the diagnosis of left ventricular non-compaction. European Heart Journal 2010, 31, 1098–1104. [Google Scholar] [CrossRef] [PubMed]
  5. Captur, G.; Muthurangu, V.; Cook, C.; Flett, A.S.; Wilson, R.; Barison, A.; Sado, D.M.; Anderson, S.; McKenna, W.J.; Mohun, T.J.; Elliott, P.M.; Moon, J.C. Quantification of left ventricular trabeculae using fractal analysis. Journal of Cardiovascular Magnetic Resonance 2013, 15, 36. [Google Scholar] [CrossRef] [PubMed]
  6. Choi, Y.; Kim, S.M.; Lee, S.C.; Chang, S.A.; Jang, S.Y.; Choe, Y.H. Quantification of left ventricular trabeculae using cardiovascular magnetic resonance for the diagnosis of left ventricular non-compaction: evaluation of trabecular volume and refined semi-quantitative criteria. Journal of cardiovascular magnetic resonance : official journal of the Society for Cardiovascular Magnetic Resonance 2016, 18, 24–24. [Google Scholar] [CrossRef] [PubMed]
  7. Bernabé, G.; González-Carrillo, J.; Cuenca, J.; Rodríguez, D.; Saura, D.; Gimeno, J.R. Performance of a new software tool for automatic quantification of left ventricular trabeculations. Revista Española de Cardiología (English Edition) 2017, 70, 405–407. [Google Scholar] [CrossRef] [PubMed]
  8. Bernabé, G.; Casanova, J.D.; Cuenca, J.; Carrillo, J.G. A self-optimized software tool for quantifying the degree of left ventricle hyper-trabeculation. The Journal of Supercomputing 2018, 75, 1625–1640. [Google Scholar] [CrossRef]
  9. Corsi, C.; Ghaderzadeh, M.; Eshraghi, M.A.; Asadi, F.; Hosseini, A.; Jafari, R.; Bashash, D.; Abolghasemi, H. Efficient Framework for Detection of COVID-19 Omicron and Delta Variants Based on Two Intelligent Phases of CNN Models. Computational and Mathematical Methods in Medicine 2022, 2022. Article ID 4838009. [CrossRef]
  10. Hosseini, A.; Eshraghi, M.A.; Taami, T.; Sadeghsalehi, H.; Hoseinzadeh, Z.; Ghaderzadeh, M.; Rafiee, M. A mobile application based on efficient lightweight CNN model for classification of B-ALL cancer from non-cancerous cells: A design and implementation study. Informatics in Medicine Unlocked 2023, 39, 101244. [Google Scholar] [CrossRef]
  11. Ghaderzadeh, M. Clinical Decision Support System for Early Detection of Prostate Cancer from Benign Hyperplasia of Prostate. Studies in health technology and informatics 2013, 192:928. [Google Scholar] [PubMed]
  12. Sadoughi, F.; Ghaderzadeh, M. A hybrid particle swarm and neural network approach for detection of prostate cancer from benign hyperplasia of prostate. Studies in health technology and informatics 2014, 205:481-5. [Google Scholar] [PubMed]
  13. Sze, V.; Chen, Y.H.; Yang, T.J.; Emer, J.S. Efficient Processing of Deep Neural Networks: A Tutorial and Survey. Proceedings of the IEEE 2017, 105, 2295–2329. [Google Scholar] [CrossRef]
  14. Barón, J.R.; Bernabé, G.; González-Férez, P.; García, J.M.; Casas, G.; González-Carrillo, J. Improving a Deep Learning Model to Accurately Diagnose LVNC. Journal of Clinical Medicine 2023, 12. [Google Scholar] [CrossRef] [PubMed]
  15. Rodríguez-de-Vera, J.M.; Bernabé, G.; García, J.M.; Saura, D.; González-Carrillo, J. Left ventricular non-compaction cardiomyopathy automatic diagnosis using a deep learning approach. Comput. Methods Programs Biomed 2022, 214, 106548. [Google Scholar] [CrossRef] [PubMed]
  16. Bernabé, G.; González-Férez, P.; García, J.M.; Casas, G.; González-Carrillo, J. Expanding the deep-learning model to diagnosis LVNC: limitations and trade-offs. Computer Methods in Biomechanics and Biomedical Engineering: Imaging & Visualization 2024, 12, 2314566. [Google Scholar] [CrossRef]
  17. Kim, S.; Yoon, H.; Lee, J.; Yoo, S. Semi-automatic Labeling and Training Strategy for Deep Learning-based Facial Wrinkle Detection. 2022 IEEE 35th International Symposium on Computer-Based Medical Systems (CBMS), 2022, pp. 383–388. [CrossRef]
  18. Ye, Y.; Luo, Z.; Qiu, Z.; Cao, K.; Huang, B.; Deng, L.; Zhang, W.; Liu, G.; Zou, Y.; Zhang, J.; Li, J. Radiomics Prediction of Muscle Invasion in Bladder Cancer Using Semi-Automatic Lesion Segmentation of MRI Compared with Manual Segmentation. Bioengineering 2023, 10. [Google Scholar] [CrossRef] [PubMed]
  19. Jiang, X.; Xu, B.; Wei, M.; Wu, K.; Yang, S.; Qian, L.; Liu, N.; Peng, Q. C2F-3DToothSeg: Coarse-to-fine 3D tooth segmentation via intuitive single clicks. CComputers & Graphics 2022, 102, 601–609. [Google Scholar] [CrossRef]
  20. Knuth, F.; Groendahl, A.R.; Winter, R.M.; Torheim, T.; Negård, A.; Holmedal, S.H.; Bakke, K.M.; Meltzer, S.; Futsæther, C.M.; Redalen, K.R. Semi-automatic tumor segmentation of rectal cancer based on functional magnetic resonance imaging. Physics and Imaging in Radiation Oncology 2022, 22, 77–84. [Google Scholar] [CrossRef] [PubMed]
  21. Gibson, D.; Spann, M.; Woolley, S. A wavelet-based region of interest encoder for the compression of angiogram video sequences. IEEE Transactions on Information Technology in Biomedicine 2004, 8, 103–113. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Cross-validation method.
Figure 1. Cross-validation method.
Preprints 140896 g001
Figure 2. U-Net from [15] adapted to our use case.
Figure 2. U-Net from [15] adapted to our use case.
Preprints 140896 g002
Figure 3. Initial interface for Image fixer.
Figure 3. Initial interface for Image fixer.
Preprints 140896 g003
Figure 4. Painting trabecular zone on output image.
Figure 4. Painting trabecular zone on output image.
Preprints 140896 g004
Figure 5. Image selector interface.
Figure 5. Image selector interface.
Preprints 140896 g005
Figure 6. Image selector with the bottom-left image selected.
Figure 6. Image selector with the bottom-left image selected.
Preprints 140896 g006
Table 1. Subjective evaluation scale proposed in [21].
Table 1. Subjective evaluation scale proposed in [21].
Score Segmentation quality
5.0 Exact match: there are no noticeable differences
4.5
4.0 Noticeable differences: they are not diagnostically significant
3.5
3.0 Small diagnostically significant differences
2.5
2.0 Significant diagnostic information is lost
1.5
1.0 Large diagnostically significant differences
Table 2. Dice coefficients from the U-Net in [14] on its images from our test set.
Table 2. Dice coefficients from the U-Net in [14] on its images from our test set.
Population Dice EL Dice IC Dice TZ Average Dice
P 0.90 ± 0.02 0.963 ± 0.002 0.87 ± 0.01 0.91 ± 0.01
X 0.88 ± 0.02 0.967 ± 0.003 0.87 ± 0.003 0.90 ± 0.01
H 0.89 ± 0.01 0.937 ± 0.003 0.82 ± 0.01 0.88 ± 0.01
Table 3. Dice coefficients from the U-Net trained on blob-selection images on its test set
Table 3. Dice coefficients from the U-Net trained on blob-selection images on its test set
Population EL IC TZ Media
P 0.928 ± 0.004 0.97 ± 0.01 0.895 ± 0.004 0.930 ± 0.004
X 0.90 ± 0.01 0.97 ± 0.01 0.87 ± 0.01 0.91 ± 0.01
H 0.90 ± 0.01 0.95 ± 0.01 0.87 ± 0.02 0.91 ± 0.01
Table 4. Dice coefficients from the U-Net trained on manually-fixed images on its test set
Table 4. Dice coefficients from the U-Net trained on manually-fixed images on its test set
Population EL IC TZ Media
P 0.91 ± 0.01 0.966 ± 0.004 0.89 ± 0.02 0.92 ± 0.01
X 0.88 ± 0.02 0.97 ± 0.01 0.87 ± 0.02 0.90 ± 0.01
H 0.87 ± 0.02 0.93 ± 0.01 0.88 ± 0.02 0.89 ± 0.02
Table 5. Dice coefficients comparing original segmentation targets with those obtained using the blob-selection method.
Table 5. Dice coefficients comparing original segmentation targets with those obtained using the blob-selection method.
Population Dice EL Dice IC Dice TZ Average Dice
P 0.94 ± 0.02 0.98 ± 0.02 0.91 ± 0.04 0.94 ± 0.04
X 0.93 ± 0.02 0.98 ± 0.01 0.90 ± 0.03 0.94 ± 0.03
H 0.93 ± 0.02 0.95 ± 0.03 0.86 ± 0.05 0.92 ± 0.05
Table 6. Dice coefficients comparing original segmentation targets with those obtained through manual correction.
Table 6. Dice coefficients comparing original segmentation targets with those obtained through manual correction.
Population Dice EL Dice IC Dice TZ Average Dice
P 0.89 ± 0.03 0.94 ± 0.03 0.85 ± 0.05 0.89 ± 0.03
X 0.84 ± 0.05 0.95 ± 0.02 0.82 ± 0.05 0.87 ± 0.04
H 0.85 ± 0.04 0.88 ± 0.05 0.75 ± 0.05 0.80 ± 0.10
Table 7. Dice coefficients comparing the blob-selection method with the manual correction method.
Table 7. Dice coefficients comparing the blob-selection method with the manual correction method.
Population Dice EL Dice IC Dice TZ Average Dice
P 0.91 ± 0.03 0.95 ± 0.02 0.87 ± 0.02 0.90 ± 0.03
X 0.88 ± 0.04 0.96 ± 0.02 0.85 ± 0.02 0.90 ± 0.03
H 0.87 ± 0.03 0.90 ± 0.05 0.80 ± 0.06 0.84 ± 0.08
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated