Preprint
Article

This version is not peer-reviewed.

Feasibility of Proton Dosimetry Overriding Planning CT with Daily CBCT Elaborated through Generative Artificial Intelligence Tools

A peer-reviewed article of this preprint also exists.

Submitted:

19 April 2023

Posted:

20 April 2023

Read the latest preprint version here

Abstract
In radiotherapy, cone-beam computed tomography (CBCT) is mainly used for patient positioning and treatment monitoring. CBCT is deemed to be secure for patients, making it suitable for the delivery of fractional doses. However, because of a narrow field of view (FOV), beam hardening and scattered radiation artifacts, loss of anatomical information and variability in pixel intensity occur that prevent the use of raw CBCT images for dose recalculation during treatment. To address this issue, reliable correction techniques are mandatory to remove artifacts and remap pixel intensity into Hounsfield Unit (HU) values. The present study proposes a deep-learning framework to calibrate CBCT images acquired with narrow FOV systems and demonstrate the potential use in proton treatment planning updates. Cycle-consistent GAN processes raw CBCT to reduce scatter and remap HU. Monte Carlo simulation enables a fair comparison between planning CT and calibrated CBCT dosimetry. Tests were performed on a publicly available dataset of 40 patients who received ablative radiation therapy for locally advanced pancreatic cancer. The CBCT calibration led to a difference in proton dosimetry of less than 2%, compared to the planning CT. The potential toxicity effect on the organs at risk (bowel and stomach) decreased from about 50% (uncalibrated CBCT) up the 2% (calibrated CBCT). These results may confirm that generative artificial intelligence brings the use of CBCT images incrementally closer to clinical translation in proton therapy.
Keywords: 
;  ;  ;  ;  ;  ;  ;  ;  
Subject: 
Engineering  -   Bioengineering

1. Introduction

Medical imaging plays a crucial role in oncology, particularly in radiotherapy. Computed tomography (CT) is the primary modality used in radiation therapy for high-resolution patient geometry and accurate dose calculations [1]. However, CT is associated with high patient exposure to ionizing radiation. Cone-beam computed tomography (CBCT) has the potential to provide faster imaging and reduce patient exposure to non-therapeutic radiation, making it a valuable imaging modality used for patient positioning and monitoring in radiotherapy. CBCT is currently used to monitor and detect changes in patient anatomy throughout the treatment; it is also compatible with fractional dose delivery, making it a patient-safe imaging modality with less additional non-therapeutic dose than traditional CT. However, this modality can introduce scattered radiation image artifacts like shading, cupping, and beam-hardening [2,3]. The artifacts resulting from scattered radiation in CBCT images can cause fluctuations in pixel values, making it difficult to use these images directly for dose calculation. Consequently, CBCT images cannot be used directly for dose calculations unless correction methods are applied. Reliable correction techniques for calibrating CBCT images to Hounsfield Unit (HU) values used by CT scanners would expand the clinical usage of CBCT in treatment planning and evaluation of tumor shrinkage and organ shift [4,5,6]. In recent years, traditional approaches, such as anti-scatter grid, partial beam blockers, and scattering estimators [7,8,9], have been joined to deep learning-based methods, which have been showing interesting potential to improve CBCT quality [10]. Such methods, leveraging mainly convolutional neural networks (CNN) and generative adversarial networks (GAN), were investigated to map the physical model of the x-ray interaction with matter disregarding the underlying complex analytics and avoiding the use of explicit statistical approaches such as Monte Carlo. Aimed at removing scatter and correcting HU units in CBCT scans, many authors explored various types of CNN, ranging from UNet trained with a supervised training approach [11,12,13,14,15] to the more complex cycle-consistent Generative Adversarial Network (cGAN), based on an unsupervised training approach [16,17,18,19,20,21]. cGAN model consists of two subnetworks, the generator and the discriminator, with opposite roles. While the generator tries to learn how to convert one dataset to another, the discriminator distinguishes between real and synthetic images. This process creates a cycle-consistent loop that improves the generator’s ability to produce synthetic images that look just like real ones. Focusing on CBCT-to-CT mapping, Xie et al. proposed a scatter artifact removal CNN based on a contextual loss function trained on the pelvis region of 11 subjects to correct the CBCT artifacts in the pelvic area [22]. Another research focused on a cGAN model to calibrate CBCT HU values in the pelvis region. The model was trained on 49 patients with unpaired data and tested on nine independent subjects, and the authors claimed the method kept the anatomical structure of CBCT images unchanged [18]. Exploring the use of deep residual neural networks in this field, a study demonstrated the capability of such architectures by proposing an iterative tuning-based training, where images with increasing resolutions are used at each step [23]. Likewise, our group recently reported that cGAN has better capability than CNN trained with pure supervised techniques to preserve anatomical coherence [24]. All these contributions, however, did not address the consistency of the treatment planning performed with the corrected CBCT. Conversely, Zhang et al. [25] reported the test of pelvis treatment planning in proton therapy performed on CBCT corrected with CNN. However, they summarized that the dose distribution calculated for traditional photon-based treatment outperformed the one computed for proton therapy. CBCT corrected with a cGAN was applied to evaluate the quality of the proton therapy planning in cancer treatment across different datasets with satisfactory results [13,20]. All the mentioned research works focused on the problem of CBCT-to-CT HU conversion exploiting CBCT with a wide field of view (FOV). However, some systems present in clinical practice have a limited FOV, not sufficient to contain the entire volume of the patient, e.g., in the presence of large regions such as the pelvis or with obese patients [26]. Considering the current use of CBCT for patient positioning purposes, small FOV CBCT systems could be preferred due to their reduced imaging dose, shorter computation time, and increased resolution over the treatment region of interest [27]. However, the limited FOV also causes a truncation problem during reconstruction. Consequently, the non-uniqueness of the solution for the iterative reconstruction causes additional bright-band effects that add to the previously mentioned CBCT artifacts [28]. Even in the case of optimal HU calibration and scatter reduction, a CBCT, acquired in a narrow FOV cannot be used for adaptive dose planning. Especially, narrow FOV CBCT lacks important anatomical information (e.g., the air/skin interface) necessary for correctly calculating the beam path. The present work aimed to propose a deep-learning framework that elaborates the CBCT to calibrate the HU, remove artifacts due to the conic geometry acquisition, and handle narrow FOV issues to demonstrate the potential use of the corrected CBCT in the context of proton treatment planning updates. The work is part of a larger study carried out in collaboration with the Italian National Center of Hadrontherapy (CNAO, Pavia, Italy) that aims to explore the possibility to use the in-house narrow FOV CBCT system not only for patient positioning but also for dosimetric evaluation without hardware modifications [29]. The deep-learning framework took its root from the CBCT-to-CT mapping model based on cGAN proposed in [24] that was here extended to address the case of narrow FOV. Tests were carried out on a public dataset of planning CT scans of 40 oncological patients affected by pancreatic cancer. To handle narrow FOV, synthetic raw CBCT volumes were properly generated from CT scans throughout the Monte Carlo simulation. This enabled us to dump anatomical variations usually present in real CBCT with respect to the corresponding planning CT. As the dataset provided annotation data about the segmented lesion and organs at risk, particle beam dosimetry was computed in the original planning CT and the corrected CBCT volume, verifying the coherency between the two dose distributions. The main contributions of this paper may therefore be summarized as:
  • capability of the cGAN to correct CBCT (scatter reduction and HU remapping) when applied to small FOV;
  • consistency of the proton dosimetry computed on corrected CBCT with respect to the original planning CT.

2. Materials and Methods

2.1. Dataset Description

A publicly available dataset obtained from the Cancer Imaging Archive, called Pancreatic-CT-CBCT-SEG [30], was exploited in this work. The dataset contained CT acquisition from 40 patients who received ablative radiation therapy for locally advanced pancreatic cancer at Memorial Sloan Kettering Cancer Center. Each CT was acquired during a deep inspiration breath-hold verified with an external respiratory monitor. The dataset also contains manual segmentations of a region of interest (ROI) defined by expanding the dose planning target volume by 1 cm. Along with the ROI, each scan provides contours of some organ at risk (OAR), namely: i) the stomach with the first two segments of the duodenum, ii) the remainder of the small bowel, and iii) both lungs. The authors reported that the segmentations were performed independently by six radiation oncologists and reviewed by two trained medical physicists. While the dataset provided two CBCT scans of each subject, these were not considered here because they were obtained at different times with respect to the corresponding CT scan, which could lead to potential changes in patient anatomy between acquisitions. Simulated CBCT scans were considered instead by generating them directly from the corresponding planning CT (implementation is detailed below in Section 2.1.1). This way, perfect alignment and anatomical correspondence between the two volumes were both ensured, avoiding the need for additional registration steps (rigid or deformable). To summarize, using simulated CBCT scans allowed the study to focus solely on the algorithm’s ability to reduce artifacts and cupping effects without considering intra-patient longitudinal variability.

2.1.1. CBCT simulation

Synthetic CBCTs were generated from the original available CTs following the approach documented in [31]. Specifically, Monte Carlo (MC) simulations were conducted to generate primary ( P M C ) and scatter ( S M C ) X-ray images for each CT scan. All simulations were performed using the GATE open-source software v9.2 (based on Geant4 v11) [32] with fixed forced detection, a variance reduction technique aimed at minimizing computation time. The energy-dependent detector efficiency was based on the design specifications provided by the manufacturer for the Paxscan 4030D (Varian Medical System, Palo Alto, CA). The X-ray fluence spectrum was computed using the open-source software SpekPy [33], employing 3.2 mm Al filtration at 100 kVp. The A-277 X-ray tube (Varian Medical System, Palo Alto, CA), chosen for this work, features a 7 rhenium-tungsten molybdenum target. Images were produced according to a CBCT scan of 220 and with projection matching the size of the Paxscan 4030D detector (isometric pixel spacing 0.388 mm, detector size 768 × 1024 pixels). The source-to-detector and source-to-isocenter distances were set to 1600 mm and 1100 mm, respectively. A further acceleration of scatter calculation was achieved by downsampling resolution 8-fold and simulating S M C at 5 steps with a statistical uncertainty < 5 % . S M C images were then upsampled and interpolated at the required points to match the corresponding P M C images. Lastly, the final projections S M C + P M C were normalized by the simulated flat field image. CBCT scans were then reconstructed using open-source software RTK [34] at a 1 × 1 × 1 mm resolution, with a 220 × 220 × 220 size in pixel and masked to the axial field-of-view of diameter equal to 204 mm. Some examples of planning CT and simulated CBCT axial slices are visible in Figure 1, along with the intensity profile of the central pixel row. The cupping effect is evident as a shaded portion in the middle of the CBCT and confirmed by the concavity in the intensity profile.

2.2. CBCT-to-CT Correction

2.2.1. Neural Network Architecture and Main Processing Layers

The network implemented for CBCT correction was based on the cycle Generative Adversarial Network (cGAN) [35]. This architecture is based on four concurrent subnetworks, two generators and two discriminators, which work in opposition. While the generators try to learn the mapping to convert CBCT to CT (or CT to CBCT), the discriminator’s objective is to distinguish between authentic and network-generated images. This generator-discriminator cycle-consistent loop is designed to improve the generators’ ability to produce synthetic images that reproduce with high fidelity the characteristic of the destination modality (e.g., generate a calibrated synthetic CT starting from a scattered CBCT). The network’s fundamental processing unit was referred to as ConvBlocks (Figure 2), which were built using a 2D convolution with a 3 × 3 kernel, followed by an instance normalization layer and a swish activation function. The other basic processing unit for cGAN structure was the InceptionBlock, consisting of four parallel ConvBlocks, each with an increasing kernel size of dimensions 1 × 1 , 3 × 3 , 7 × 7 , and 11 × 11 , which processed the same input simultaneously with multiple receptive fields. The output of each branch of InceptionBlock was then combined, and the complete set of feature maps was produced as output. The primary objective of this processing block was to execute multi-scale feature extraction from the initial image. The extracted multi-scale features, varying from small to large receptive fields, can produce improved outcomes for image synthesis. The general design of the generator was then carried out as a modified version of the commonly used U-Net architecture. The U-Net model is usually utilized for solving pixel-by-pixel classification challenges in image segmentation [36]. Still, it can also be used to solve image-to-image conversion problems with minor changes. The overall generator structure, depicted in Figure 2, was composed of a contracting and an expanding path. The upper two processing layers of the generator were based on InceptionBlocks, while the deeper two exploited ConvBlocks. Consequently, the network can be broadly top-bottom divided into two segments, each serving distinct functions: i) the inception part (upper layers) extracted global contextual information, whereas ii) the traditional 2D convolution part (bottom layers) was responsible for capturing more intricate context and precise localization. On the other hand, the CNN utilized as the discriminator was responsible for image classification and relied on the PatchGAN architecture [37]. Its architecture was based on four sequential ConvBlocks, each with a kernel size of 4 × 4 . In the initial three ConvBlocks, the convolution was set with stride 2, leading to an output tensor with half the size and twice the features map. In contrast, the last ConvBlock had stride one and maintained the size and the number of feature maps unchanged. A sigmoid activation function was applied to the last layer, generating a 32 × 32 map used to classify the input image as real or fake.

2.2.2. Model Training

As previously stated, the cGAN overall training routine involved two generators and two discriminators, competing against one another to solve the CBCT-to-CT conversion problem. The subnetworks were referred to as generator CT (G C T ), generator CBCT (G C B C T ), discriminator CT (D C T ), and discriminator CBCT (D C B C T ). G C T and g C B C T were used to produce generated CT from CBCT, and generated CBCT from CT, respectively, while D C T and D C B C T were used to distinguish the original CT and CBCT from their generated counterparts. The training routine can be divided into two main phases that occur simultaneously. In the first phase, called the generative phase, G C T (G C B C T ) took a 2D axial slice of a CBCT (CT) as input and produced a generated CT (generated CBCT) as output. Then, G C T (G C B C T ) took the generated CT (generated CBCT) as input and produced a cyclic CBCT (cyclic CT), which was supposed to be equal to the original CBCT (CT). At the same time, during the second phase, called the classification phase, D C T (D C B C T ) tried to distinguish between real CT (CBCT) and generated CT (generated CBCT). The entire cGAN was implemented in Python, using Keras [38] and TensorFlow [39] frameworks. All the technical details about the network implementation can be found in a previous work of our group [24]. The whole dataset was divided into training, validation, and test sets in proportions of 70%, 15%, and 15%, containing 5698, 1221, and 1221 2D axial slices, respectively.

2.2.3. Performance Metrics for Model Evaluation

The network performances were quantitatively evaluated using the original CT as the ground truth reference. In particular, the metrics evaluated were i) peak signal-to-noise ratio (PSNR), ii) structural similarity index measure (SSIM), and iii) mean absolute error (MAE) [40]. The PSNR quantifies the quality of images by comparing the mean square error of the images being compared to the maximum possible signal power [41]. It is measured in decibels, and its value increases towards infinity as the difference between the calibrated CBCT and ground-truth CT decreases. Therefore, a larger PSNR value indicates better image quality, while a lower value indicates the opposite. SSIM evaluates the resemblance between two images by analyzing their luminance, contrast, and structure [42]. Compared to PSNR, SSIM is considered a more human-like measure of similarity. The SSIM score ranges from 0 to 1, with a value of 1 indicating the highest level of similarity between the images. MAE was used to quantitatively assess the accuracy of Hounsfield Units (HU) between the generated CBCT and the original CT. The lower value corresponds to the higher level of HU accuracy between the two images. The significance ( p < 0 . 05 ) of the statistical difference between CBCT slices prior to and following calibration was verified using Kruskall-Wallis non-parametric test.

2.2.4. Synthetic CT Generation Pipeline

Despite the better quality of calibrated CBCT in terms of HU density values, these volumes cannot yet be used for adaptive dose planning due to their limited FOV. In fact, these CBCT acquisitions lack important anatomical information (e.g., the air/skin interface) necessary for the correct calculation of the beam path. In order to overcome this intrinsic limit, the original planning CT was used to provide the missing information. Therefore, synthetic CT (sCT) is defined in this work as an updated version of the original planning CT overridden with the calibrated voxels from daily CBCT acquired during the treatment. Starting from a scattered CBCT, the following procedure was followed in an axial slice-by-slice approach. At first, each pixel in the slice was clipped between values [ 1000 ; 3000 ] and then normalized in the [ 0 ; 1 ] range. This step is fundamental because the neural network needs value in this range to operate properly. Then, the generator G C T processed the normalized CBCT, producing a corrected version of the same axial slice. It is important to remember that G C T is the only cGAN subnetwork used after completing the training. After neural network processing, the previous pixel clipping guarantees that the normalization can be reversed back to Hounsfield Unit. The last step involved overriding the planning CT pixels with the region acquired with the cone beam modality. Every pixel outside the CBCT FOV belonged to the original CT. The entire pipeline is summarized in Figure 3. In order to evaluate the effective improvement obtained by the corrected CBCT in terms of treatment planning, two versions of sCT were generated for each subject. The first, called sCT corrected (sCTc) was obtained following the mentioned procedure, while the second, called sCT uncorrected (sCTu), was obtained simply by overriding the original CBCT volume without any kind of processing.

2.3. Dosimetric Analysis

2.3.1. Proton-Based Treatment Planning

The treatment plan for each subject was computed with the matRad package [43], an open-source radiation treatment planning toolkit written in Matlab. In particular, the planning was developed using protons as the radiation mode and optimized using the constant relative biological effectiveness times dose method, which accounts for the varying biological effectiveness of different radiation types and energies. A total of 30 fractions were scheduled for the treatment, with two beams used at gantry angles of 0 degrees (anterior direction) and 270 degrees (right lateral direction). Several constraints were chosen in the planning definition to ensure the safety and efficacy of the treatment. For the bowel and stomach regions, squared overdosing and maximum dose volume histogram constraints were used to limit the radiation dose received by these sensitive areas. The lung regions were also subject to squared overdosing constraints to limit the dose delivered to that areas. Finally, the ROI was subject to squared deviation constraints, which aim to keep the dose distribution as close as possible to the prescribed dose (30 fractions of 2 Gy equivalent) [44]. The reference dose planning was first computed directly on the pCT and used as the ground truth in further comparison. Then, this reference plan was updated, giving either corrected or uncorrected sCT as the new volume. As each CBCT was generated from its corresponding pCT, the alignment was intrinsically guaranteed, enabling the comparison of the calculated dose plans without considering any organ deformations between the two fractions.

2.3.2. Dose Evaluation

To evaluate the suitability of corrected CBCT scans, various metrics were used, including dose difference pass rate (DPR), dose–volume histogram (DVH) metrics, and Gamma pass rates (GPR). The treatment dose computed for the original planning CT was considered as the prescribed ground truth dose [40]. DPR measures the percentage of pixels that meet a certain dose difference threshold, DVH compares cumulative dose to different structures in relation to volume, while GPR assesses the similarity of two dose distributions based on dose difference and distance-to-agreement criteria. The significance ( p < 0 . 05 ) of the statistical difference in GPR distributions between sCTu and sCTc was verified using Kruskall-Wallis non-parametric test. Mean doses, D5, and D95, measured on ROI, bowel, and stomach were considered to assess the treatment quality and the toxicity control on the organ at risk.

3. Results

At a qualitative inspection, the corrected overwritten CBCT sCT scans were more similar to pCTs with respect to their uncorrected counterpart. The darker region within the CBCT FOV characterized by the cupping artifacts was no longer noticeable (Figure 4). Intensity profiles also confirmed this, showing that the concave shape observable in the unprocessed lines disappeared in the processed ones, now matching the HU values range of the reference pCT and confirming that non-linearity present in the CBCT tissue density was corrected.

3.1. cGAN Model Evaluation

The results from evaluating the model’s performance metrics (Figure 5) showed promising improvements in the quality of the CBCT slices. The original images had a median PSNR of 24.60 dB (IQR 1.40 dB), while the processed images had a median PSNR of 33.41 dB (IQR 3.36 dB), resulting in a relative gain of approximately 37%. In terms of the SSIM score, the original images had a median of 0.90 (IQR 0.03), and the processed CBCTs showed a median of 0.95 (IQR 0.02), which corresponded to a relative enhancement of around 5%. Furthermore, the median MAE for the original images was 148.96 HU (IQR 31.24 HU), whereas the median MAE for the processed images was 43.47 HU (IQR 14.82 HU). These results demonstrate the effectiveness of the cGAN approach in improving CBCT image quality. For all three metrics, a statistical difference was found ( p < 0 . 01 ).

3.2. Treatment Planning Evaluation

The qualitative comparison between the treatment plans computed for each modality confirmed an ameliorated similarity between sCTc and pCT with respect to their uncorrected counterpart. An example of this can be seen in Figure 6 upper row. It is visible how the beam path computed on sCTu exceeded the ROI releasing a more significant amount of dose in the following tissues. Moreover, it could also be seen that high dose values (red pixels) break over ROI boundaries, indicating that a portion of surrounding healthy areas received overexposure to radiation, as also confirmed by the dose difference computed with respect to the pCT reference plan (cfr. Figure 6 bottom row). Conversely, the sCTc treatment plan corrected that pattern, showing a more similar beam path and a reduced difference with respect to the pCT plan. Likewise, the dose-volume histogram computed for the same test subject confirmed and enforced such a consideration (Figure 7). Observing the ROI lines, the sCTc (yellow dotted line) followed the profile of pCT (yellow solid line) more closely than the sCTu (yellow dashed line). The treatment plan calculated on sCTc and pCT showed a steep slope, indicating that 2 Gy was the dose delivered to almost all the ROI, while sCTu showed a smoother slope, a sign of overdosing in a portion of this area. Concerning the organs at risk (green and light blue lines), this consideration was even more evident, with an overdosing in the order of about two times with respect to the reference plan that is incompatible with clinical practice.
The GPR results for the entire dataset were computed for different gamma criteria and summarized in Table 1 as median (interquartile range, IQR). The median gamma pass rates for the 1%/1 mm, 2%/2 mm, 3%/2 mm, and 3%/3 mm criteria were consistently higher for the sCTc the sCTu, with the most significant improvement observed for the 3%/3 mm criterion (92.82% vs. 57.57%).
DPR at 1% was also found to be significantly higher for the processed sCT than the unprocessed ones (93.97% vs. 79.76%), indicating a 14.21% improvement in dose accuracy with the use of sCTc. In regard to mean dose distribution across the overall dataset, the advantage of sCTc was evident with respect to the overdosing of the sCTu in the ROI (Table 2). The relative percentage error decreased from 6% for sCTu up to 2% for sCTc. A greater advantage was achieved in terms of unwanted doses distributed at the bowel as 97% against 2%, with respect to the nominal toxicity in the pCT. Likewise, the relative toxicity in the stomach decreased from 49% up the 2%. For D5, the correction was effective in reducing the overexposure found in the unprocessed sCT. For D95, the correction underestimated the dose of about 5%. As far as OAR is concerned, the correction was again effective in ensuring a low dose, very similar to that one obtained in the planning CT, at both bowel and stomach. Interestingly, the IQR range of 0.12 Gy (D95) for the stomach potentially delivered when using the sCTu was completely zeroed by the correction.

4. Discussion

This work proposed a novel image-processing framework for generating synthetic CT scans, which combines the original planning CT with routine CBCT scans, usable to update the dosimetry plan in proton therapy. The core of the framework was represented by a deep learning model, namely a cycle-consistency GAN, to correct the scatter artifacts in the CBCT images and calibrate intensity values, in the proper HU range. Especially, the framework was shown to properly address CBCT equipment scanning narrow FOV [40]. To the aim, the public dataset of CT scans, named Pancreatic-CT-CBCT-SEG [30], was exploited because of clinically consistent segmentation of ROI and OAR across all the considered patients. This allowed us to generate physically-consistent CBCT, by means of Monte Carlo simulation, featuring also narrow FOV. The available segmentations granted replicating the particle beam planning on the original CT and corrected sCT for straightforward comparison. The obtained results confirmed, both qualitatively and quantitatively, improvements in the CBCT images. As shown, the intensity profiles were properly rectified as a result of cupping and truncation artifact removal (cfr. Figure 4). A sensible increase in PSNR, SSIM, and MAE metrics testified to the effectiveness of the correction methodology (cfr. Figure 5). Dosimetry computed on uncorrected sCT featured overdosing, especially at OAR (cfr. Table 1 and Table 2). An inaccurate assessment of tissue densities was made due to the lack of HU calibration. The difference in grayscale values between CBCT and CT caused a discontinuity in the volume (cfr. Figure 4), leading to errors in particle beam path computation. Conversely, the dosimetric plan computed on the corrected sCT confirmed the consistency of the plan computed on the corresponding CT scan (cfr. Table 1). The American Association of Physicists in Medicine (AAPM) Task Group 218 defined acceptance criteria for tolerance and action limits as exceeding 95% and falling below 90%, respectively, for a 3%/2 mm GPR standard [48]. While not completely in agreement with the upper threshold, the 90.26% found in this work (cfr. Table 1) is to be deemed reasonable. Nonetheless, such value, overcoming the lower 90% action limit threshold, makes the methodology promising for clinical application. Comparison with works in the literature using deep learning to correct CBCT and test proton dosimetry showcased the consistency of the methodology in terms of GPR 2%/2 mm results, even though slightly smaller values obtained in the present work (Table 3). Nonetheless, we remark that our study applied the correction to small FOV (204 mm diameter) while the considered works in the literature dealt with mainly wider FOV (about 480 mm diameter on average). All the works in this comparison involved large anatomical sites (e.g., pelvis, thorax, and abdomen) and the patient cohort had a similar size. Remarkably, the comparison highlighted the superiority of the generative adversarial network with respect to traditional UNet. In order to avoid confounding factors induced by organ deformation, this work did use simulated CBCT by means of the Monte Carlo approach, so that differences between the images were exclusively due to artifacts rather than anatomy. This allowed us to use the same lesion and OAR segmentation used to compute the reference treatment plan. In any case, this aspect should not be considered a limitation in the usability of the framework with real CBCTs. For instance, a rigid registration step guided by the large bone structures in the pelvic area should be considered to ensure macro alignment between the CBCT and the overwritten pCT. It is fundamental to recall that the volume of interest is entirely contained in the CBCT and that the tissues coming from the pCT are only needed to calculate the beam path. In addition, this makes it mandatory to update the segmentation mask. CBCT with narrow FOV does not represent an uncommon situation. In this work, the FOV was particularly reduced to demonstrate the feasibility of the methodology. In general, the proposed framework can be adopted in all cases in which the district of interest is too large to fit into a single acquisition. Furthermore, the proposed method extends the use of CBCT systems currently used for patient positioning without additional hardware. This method will be retrospectively evaluated in a real offline therapy context at the CNAO facility in future work [29]. Correction of CBCTs obtained day-by-day during treatment will be used to assess the evolution of the dose plan without the need to acquire additional CTs and administer further toxicity to the patient [6,20]. Moreover, no additional hardware will be needed to add to the patient positioning setup in order to increase the FOV for dosimetric evaluation.

5. Conclusions

The present study proposed a generative artificial intelligence tool to correct CBCT scans, acquired with narrow FOV systems, enabling the reduction of scatter and the remap pixel intensity in HU. The methodology made feasible treatment planning updates, which brings the use of CBCT images incrementally closer to clinical translation in proton therapy.

Author Contributions

Conceptualization: P.C.; data curation: M.R. and G.Be.; formal analysis: M.R.; funding acquisition: G.Ba. and P.C.; methodology: M.R. and P.C.; project administration: G.Ba. and P.C.; resources: G.Ba. and P.C.; software: M.R. and G.Be.; supervision: P.C.; validation: M.R.; visualization: M.R.; writing—original draft: M.R.; writing—review and editing: M.R. and P.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by CNAO Foundation (Pavia, Italy) and EU Por-FESR Lombardia 2014-2020 program.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in The Cancer Imaging Archive (TCIA) at 10.7937/TCIA.ESHQ-4D90. These collections are freely available to browse, download, and use for commercial, scientific and educational purposes as outlined in the Creative Commons Attribution 3.0 Unported License.

Acknowledgments

The authors would like to thank Fabio Casaccio for his support in data retrieval.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CBCT Cone-Beam Computed Tomography
cGAN cycle-consistent Generative Adversarial Network
CNN Convolutional Neural Network
CT Computed Tomography
D C B C T Discriminator CBCT
D C T Discriminator CT
DPR Dose Difference Pass Rate
DVH Dose–Volume Histogram
FOV Field of View
G C B C T Generator CBCT
G C T Generator CT
GPR Gamma Pass Rate
IQR Interquartile Range
MAE Mean Absolute Error
MC Monte Carlo
OAR Organ at Risk
pCT planning CT
PSNR Peak Signal-to-Noise Ratio
ROI Region of Interest
sCT synthetic CT
sCTc corrected sCT
sCTu uncorrected sCT
SSIM Structural Similarity Index Measure

References

  1. Bortfeld, T. IMRT: A review and preview. Physics in Medicine and Biology 2006, 51, R363–R379. [CrossRef]
  2. Joseph, P.M.; Spital, R.D. The effects of scatter in x-ray computed tomography. Medical Physics 1982, 9, 464–472. [CrossRef]
  3. Schulze, R.; Heil, U.; Groβ, D.; Bruellmann, D.D.; Dranischnikow, E.; Schwanecke, U.; Schoemer, E. Artefacts in CBCT: A review. Dentomaxillofacial Radiology 2011, 40, 265–273. [CrossRef]
  4. Kurz, C.; Kamp, F.; Park, Y.K.; Zöllner, C.; Rit, S.; Hansen, D.; Podesta, M.; Sharp, G.C.; Li, M.; Reiner, M.; et al. Investigating deformable image registration and scatter correction for CBCT-based dose calculation in adaptive IMPT. Medical Physics 2016, 43, 5635–5646. [CrossRef]
  5. Thing, R.S.; Bernchou, U.; Mainegra-Hing, E.; Hansen, O.; Brink, C. Hounsfield unit recovery in clinical cone beam CT images of the thorax acquired for image guided radiation therapy. Physics in Medicine and Biology 2016, 61, 5781–5802. [CrossRef]
  6. Giacometti, V.; Hounsell, A.R.; McGarry, C.K. A review of dose calculation approaches with cone beam CT in photon and proton therapy. Physica Medica 2020, 76, 243–276. [CrossRef]
  7. Sun, M.; Star-Lack, J.M. Improved scatter correction using adaptive scatter kernel superposition. Physics in Medicine and Biology 2010, 55, 6695–6720. [CrossRef]
  8. Sisniega, A.; Zbijewski, W.; Badal, A.; Kyprianou, I.S.; Stayman, J.W.; Vaquero, J.J.; Siewerdsen, J.H. Monte Carlo study of the effects of system geometry and antiscatter grids on cone-beam CT scatter distributions. Medical Physics 2013, 40, 051915. [CrossRef]
  9. Stankovic, U.; Ploeger, L.S.; van Herk, M.; Sonke, J.J. Optimal combination of anti-scatter grids and software correction for CBCT imaging. Medical Physics 2017, 44, 4437–4451. [CrossRef]
  10. Rusanov, B.; Hassan, G.M.; Reynolds, M.; Sabet, M.; Kendrick, J.; Rowshanfarzad, P.; Ebert, M. Deep learning methods for enhancing cone-beam CT image quality toward adaptive radiation therapy: A systematic review. Medical Physics 2022, 49, 6019–6054. [CrossRef]
  11. Kida, S.; Nakamoto, T.; Nakano, M.; Nawa, K.; Haga, A.; Kotoku, J.; Yamashita, H.; Nakagawa, K. Cone Beam Computed Tomography image quality improvement using a deep convolutional neural network. Cureus 2018. [CrossRef]
  12. Jiang, Y.; Yang, C.; Yang, P.; Hu, X.; Luo, C.; Xue, Y.; Xu, L.; Hu, X.; Zhang, L.; Wang, J.; et al. Scatter correction of cone-beam CT using a deep residual convolution neural network (DRCNN). Physics in Medicine & Biology 2019, 64, 145003. [CrossRef]
  13. Landry, G.; Hansen, D.; Kamp, F.; Li, M.; Hoyle, B.; Weller, J.; Parodi, K.; Belka, C.; Kurz, C. Comparing Unet training with three different datasets to correct CBCT images for prostate radiotherapy dose calculations. Physics in Medicine & Biology 2019, 64, 035011. [CrossRef]
  14. Chen, L.; Liang, X.; Shen, C.; Jiang, S.; Wang, J. Synthetic CT generation from CBCT images via deep learning. Medical Physics 2020, 47, 1115–1125. [CrossRef]
  15. Rossi, M.; Belotti, G.; Paganelli, C.; Pella, A.; Barcellini, A.; Cerveri, P.; Baroni, G. Image-based shading correction for narrow-FOV truncated pelvic CBCT with deep convolutional neural networks and transfer learning. Medical physics 2021, 48, 7112–7126. [CrossRef]
  16. Kida, S.; Kaji, S.; Nawa, K.; Imae, T.; Nakamoto, T.; Ozaki, S.; Ohta, T.; Nozawa, Y.; Nakagawa, K. Visual enhancement of Cone-beam CT by use of CycleGAN. Medical Physics 2020, 47, 998–1010. [CrossRef]
  17. Eckl, M.; Hoppen, L.; Sarria, G.R.; Boda-Heggemann, J.; Simeonova-Chergou, A.; Steil, V.; Giordano, F.A.; Fleckenstein, J. Evaluation of a cycle-generative adversarial network-based cone-beam CT to synthetic CT conversion algorithm for adaptive radiation therapy. Physica Medica 2020, 80, 308–316. [CrossRef]
  18. Dong, G.; Zhang, C.; Liang, X.; Deng, L.; Zhu, Y.; Zhu, X.; Zhou, X.; Song, L.; Zhao, X.; Xie, Y. A Deep Unsupervised Learning Model for Artifact Correction of Pelvis Cone-Beam CT. Frontiers in Oncology 2021, 11. [CrossRef]
  19. Sun, H.; Fan, R.; Li, C.; Lu, Z.; Xie, K.; Ni, X.; Yang, J. Imaging Study of Pseudo-CT Synthesized From Cone-Beam CT Based on 3D CycleGAN in Radiotherapy. Frontiers in Oncology 2021, 11. [CrossRef]
  20. Uh, J.; Wang, C.; Acharya, S.; Krasin, M.J.; ho Hua, C. Training a deep neural network coping with diversities in abdominal and pelvic images of children and young adults for CBCT-based adaptive proton therapy. Radiotherapy and Oncology 2021, 160, 250–258. [CrossRef]
  21. Zhao, J.; Chen, Z.; Wang, J.; Xia, F.; Peng, J.; Hu, Y.; Hu, W.; Zhang, Z. MV CBCT-Based Synthetic CT Generation Using a Deep Learning Method for Rectal Cancer Adaptive Radiotherapy. Frontiers in Oncology 2021, 11. [CrossRef]
  22. Xie, S.; Liang, Y.; Yang, T.; Song, Z. Contextual loss based artifact removal method on CBCT image. Journal of Applied Clinical Medical Physics 2020, 21, 166–177. [CrossRef]
  23. Wu, W.; Qu, J.; Cai, J.; Yang, R. Multiresolution residual deep neural network for improving pelvic CBCT image quality. Medical Physics 2022, 49, 1522–1534. [CrossRef]
  24. Rossi, M.; Cerveri, P. Comparison of Supervised and Unsupervised Approaches for the Generation of Synthetic CT from Cone-Beam CT. Diagnostics (Basel, Switzerland) 2021, 11. [CrossRef]
  25. Zhang, Y.; Yue, N.; Su, M.Y.; Liu, B.; Ding, Y.; Zhou, Y.; Wang, H.; Kuang, Y.; Nie, K. Improving CBCT quality to CT level using deep learning with generative adversarial network. Medical Physics 2021, 48, 2816–2826. [CrossRef]
  26. Landry, G.; Hua, C.h. Current state and future applications of radiological image guidance for particle therapy. Medical Physics 2018, 45, e1086–e1095. [CrossRef]
  27. Lu, W.; Yan, H.; Zhou, L.; Cervino, L.; Jiang, S.; Jia, X. TU-G-141-05: Limited Field-Of-View Cone-Beam CT Reconstruction for Adaptive Radiotherapy. Medical Physics 2013, 40, 457–457. [CrossRef]
  28. Clackdoyle, R.; Defrise, M. Tomographic Reconstruction in the 21st Century. IEEE Signal Processing Magazine 2010, 27, 60–80. [CrossRef]
  29. Fattori, G.; Riboldi, M.; Pella, A.; Peroni, M.; Cerveri, P.; Desplanques, M.; Fontana, G.; Tagaste, B.; Valvo, F.; Orecchia, R.; et al. Image guided particle therapy in CNAO room 2: Implementation and clinical validation. Physica Medica 2015, 31, 9–15. [CrossRef]
  30. Hong, J.; Reyngold, M.; Crane, C.; Cuaron, J.; Hajj, C.; Mann, J.; Zinovoy, M.; Yorke, E.; LoCastro, E.; Apte, A.P.; et al. Breath-hold CT and cone-beam CT images with expert manual organ-at-risk segmentations from radiation treatments of locally advanced pancreatic cancer (Pancreatic-CT-CBCT-SEG), 2021. [CrossRef]
  31. Poludniowski, G.; Evans, P.M.; Hansen, V.N.; Webb, S. An efficient Monte Carlo-based algorithm for scatter correction in keV cone-beam CT. Physics in Medicine and Biology 2009, 54, 3847–3864. [CrossRef]
  32. Jan, S.; Santin, G.; Strul, D.; Staelens, S.; Assié, K.; Autret, D.; Avner, S.; Barbier, R.; Bardiès, M.; Bloomfield, P.M.; et al. GATE: A simulation toolkit for PET and SPECT. Physics in Medicine and Biology 2004, 49, 4543–4561. [CrossRef]
  33. Poludniowski, G.; Omar, A.; Bujila, R.; Andreo, P. Technical Note: SpekPy v2.0—a software toolkit for modeling x-ray tube spectra. Medical Physics 2021, 48, 3630–3637. [CrossRef]
  34. Rit, S.; Oliva, M.V.; Brousmiche, S.; Labarbe, R.; Sarrut, D.; Sharp, G.C. The Reconstruction Toolkit (RTK), an open-source cone-beam CT reconstruction toolkit based on the Insight Toolkit (ITK). Journal of Physics: Conference Series 2014, 489, 012079. [CrossRef]
  35. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV). IEEE, 2017. [CrossRef]
  36. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation, 2015. [CrossRef]
  37. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, 2017. [CrossRef]
  38. Chollet, F.; et al. Keras. https://keras.io, 2015.
  39. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems, 2015. Software available from tensorflow.org.
  40. Spadea, M.F.; Maspero, M.; Zaffino, P.; Seco, J. Deep learning based synthetic-CT generation in radiotherapy and PET: A review. Medical Physics 2021, 48, 6537–6566. [CrossRef]
  41. Hore, A.; Ziou, D. Image quality metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition. IEEE, 2010. [CrossRef]
  42. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Transactions on Image Processing 2004, 13, 600–612. [CrossRef]
  43. Wieser, H.P.; Cisternas, E.; Wahl, N.; Ulrich, S.; Stadler, A.; Mescher, H.; Müller, L.R.; Klinge, T.; Gabrys, H.; Burigo, L.; et al. Development of the open-source dose calculation and optimization toolkit matRad. Medical Physics 2017, 44, 2556–2568. [CrossRef]
  44. Dreher, C.; Habermehl, D.; Ecker, S.; Brons, S.; El-Shafie, R.; Jäkel, O.; Debus, J.; Combs, S.E. Optimization of carbon ion and proton treatment plans using the raster-scanning technique for patients with unresectable pancreatic cancer. Radiation Oncology 2015, 10. [CrossRef]
  45. Hansen, D.C.; Landry, G.; Kamp, F.; Li, M.; Belka, C.; Parodi, K.; Kurz, C. ScatterNet: A convolutional neural network for cone-beam CT intensity correction. Medical Physics 2018, 45, 4916–4926. [CrossRef]
  46. Thummerer, A.; Oria, C.S.; Zaffino, P.; Meijers, A.; Marmitt, G.G.; Wijsman, R.; Seco, J.; Langendijk, J.A.; Knopf, A.C.; Spadea, M.F.; et al. Clinical suitability of deep learning based synthetic CTs for adaptive proton therapy of lung cancer. Medical Physics 2021, 48, 7673–7684. [CrossRef]
  47. Kurz, C.; Maspero, M.; Savenije, M.H.F.; Landry, G.; Kamp, F.; Pinto, M.; Li, M.; Parodi, K.; Belka, C.; van den Berg, C.A.T. CBCT correction using a cycle-consistent generative adversarial network and unpaired training to enable photon and proton dose calculation. Physics in Medicine & Biology 2019, 64, 225004. [CrossRef]
  48. Miften, M.; Olch, A.; Mihailidis, D.; Moran, J.; Pawlicki, T.; Molineu, A.; Li, H.; Wijesooriya, K.; Shi, J.; Xia, P.; et al. Tolerance limits and methodologies for IMRT measurement-based verification QA: Recommendations of AAPM Task Group No. 218. Medical Physics 2018, 45. [CrossRef]
Figure 1. Examples of two CT axial slices with their corresponding CBCT. The intensity profiles of the central row (marked as a line in both images) are plotted in the right panel. Each image is displayed with Window = 1300, Level = 0.
Figure 1. Examples of two CT axial slices with their corresponding CBCT. The intensity profiles of the central row (marked as a line in both images) are plotted in the right panel. Each image is displayed with Window = 1300, Level = 0.
Preprints 71369 g001
Figure 2. Schematic of the Generator model architecture.
Figure 2. Schematic of the Generator model architecture.
Preprints 71369 g002
Figure 3. Schematic of sCT generation pipeline.
Figure 3. Schematic of sCT generation pipeline.
Preprints 71369 g003
Figure 4. Examples of two CT axial slices with their corresponding sCT generated overriding the uncorrected CBCT (sCTu) and the corrected ones (sCTc). The intensity profiles of the central row (marked as a line in both images) are plotted in the right panel. Each image is displayed with Window = 1300, Level = 0.
Figure 4. Examples of two CT axial slices with their corresponding sCT generated overriding the uncorrected CBCT (sCTu) and the corrected ones (sCTc). The intensity profiles of the central row (marked as a line in both images) are plotted in the right panel. Each image is displayed with Window = 1300, Level = 0.
Preprints 71369 g004
Figure 5. Performance metrics for cGAN model evaluation, computed on the axial slices of the test set before and after model processing.
Figure 5. Performance metrics for cGAN model evaluation, computed on the axial slices of the test set before and after model processing.
Preprints 71369 g005
Figure 6. Example of dose planning for an axial slice of a subject for the original pCT, sCTu, and sCTc (upper row). The difference between both sCT treatment plans with respect to the original pCT plan is shown in the bottom row.
Figure 6. Example of dose planning for an axial slice of a subject for the original pCT, sCTu, and sCTc (upper row). The difference between both sCT treatment plans with respect to the original pCT plan is shown in the bottom row.
Preprints 71369 g006
Figure 7. Example of DVH computed for ROI and two organs at risk (bowel and stomach). Solid line: pCT, dashed line: sCTu, dotted line: sCTc.
Figure 7. Example of DVH computed for ROI and two organs at risk (bowel and stomach). Solid line: pCT, dashed line: sCTu, dotted line: sCTc.
Preprints 71369 g007
Table 1. Gamma pass rate results for different criteria. Statistical difference in gamma score distributions, between uncorrected and corrected sCT, was found ( p < 0 . 001 ).
Table 1. Gamma pass rate results for different criteria. Statistical difference in gamma score distributions, between uncorrected and corrected sCT, was found ( p < 0 . 001 ).
Gamma Criterion sCTu sCTc
1%/1 mm 44.68 (6.91) 74.37 (4.63)
2%/2 mm 51.72 (8.15) 87.30 (6.32)
3%/2 mm 53.78 (8.53) 90.26 (5.70)
3%/3 mm 57.57 (7.49) 92.82 (5.94)
Table 2. Mean doses, D5, and D95, measured on ROI, bowel, and stomach. Values are expressed as median(IQR).
Table 2. Mean doses, D5, and D95, measured on ROI, bowel, and stomach. Values are expressed as median(IQR).
Mean dose D5 D95
pCT 1.98 (0.01) 2.04 (0.01) 1.86 (0.07)
ROI sCTu 2.10 (0.09) 2.61 (0.28) 1.64 (0.10)
sCTc 1.93 (0.08) 2.09 (0.05) 1.54 (0.28)
pCT 0.47 (0.44) 2.01 (0.04) 0.00 (0.00)
Bowel sCTu 0.93 (0.52) 2.06 (0.31) 0.00 (0.00)
sCTc 0.48 (0.40) 2.00 (0.13) 0.00 (0.00)
pCT 0.65 (0.27) 2.01 (0.03) 0.00 (0.00)
Stomach sCTu 0.97 (0.35) 2.14 (0.24) 0.00 (0.12)
sCTc 0.64 (0.29) 2.02 (0.11) 0.00 (0.00)
Table 3. Comparison of dosimetry results with literature outcomes in terms of GPR2%/2 mm in the domain of proton therapy.
Table 3. Comparison of dosimetry results with literature outcomes in terms of GPR2%/2 mm in the domain of proton therapy.
Work Model Anatomic Site axial FOV [mm] Patient cohort GPR 2%/2 mm
Hansen et al. [45] Unet Pelvis 410 30 53%
Landry et al. [13] Unet Pelvis 410 42 85%
Thummerer et al. [46] UNet Thorax 500 33 90.7%
Kurz et al. [47] cGAN Pelvis 550 33 96%
Uh et al. [20] cGAN Abdomen/Pelvis 530 50 98.5%
This work cGAN Pelvis 204 40 87.3%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated