Preprint
Article

This version is not peer-reviewed.

HDNLS: Hybrid Deep-Learning and Non-Linear Least Squares-Based Method for Fast Multi-Component T1ρ Mapping in the Knee Joint

A peer-reviewed article of this preprint also exists.

Submitted:

11 December 2024

Posted:

11 December 2024

You are already at the latest version

Abstract
Non-linear least squares (NLS) methods are commonly used for quantitative magnetic resonance imaging (MRI), especially for multi-exponential T1ρ mapping. It provides precise parameter estimation for different relaxation models in tissues, such as mono-exponential (ME), bi-exponential (BE), and stretched-exponential (SE). However, NLS may suffer from problems like sensitivity to initial guesses, slow convergence speed, and high computational cost. While deep learning (DL)-based T1ρ fitting methods offer faster alternatives, they often face challenges such as noise sensitivity and reliance on NLS-generated reference data for training. To address these limitations of both approaches, we propose the HDNLS, a hybrid model for fast multi-component parameter mapping, particularly targeted for T1ρ mapping in the knee joint. HDNLS combines voxel-wise DL, trained with synthetic data, with a few iterations of NLS to accelerate the fitting process, thus eliminating the need for reference MRI data for training. Due to the inverse-problem nature of the parameter mapping, certain parameters in a specific model may be more sensitive to noise, such as the short component in the BE model. To address this, the number of NLS iterations in HDNLS can act as a regularization, stabilizing the estimation to get meaningful solutions. Thus, in this work, we conducted a comprehensive analysis of the impact of NLS iterations on HDNLS performance and proposed four variants that balance estimation accuracy and computational speed. These variants are Ultrafast-NLS, Superfast-HDNLS, HDNLS, and Relaxed-HDNLS. These methods allow users to select a suitable configuration based on their specific speed and performance requirements. Among these, HDNLS emerges as the optimal trade-off between performance and fitting time. Extensive experiments on synthetic data demonstrate that HDNLS achieves comparable performance to NLS and regularized-NLS (RNLS) with a minimum of 13-fold improvement in speed. HDNLS is just a little slower than DL-based methods, however, it significantly improves estimation quality, offering a solution for T1ρ fitting that is fast and reliable.
Keywords: 
;  ;  ;  ;  

1. Introduction

Quantitative magnetic resonance imaging (qMRI) is essential for accurately evaluating tissue properties and detecting early pathological changes across various anatomical regions before the development of morphological changes [1,2,3]. Traditional qualitative MRI imaging, used in clinical practice, often lacks the sensitivity and specificity required for such early detection [4]. qMRI overcomes this problem by acquiring multi-contrast images with controlled variations in MRI pulse sequence parameters [5]. It enables the estimation of tissue relaxation properties, and the non-linear least squares (NLS) methods are usually the reference approaches to compute quantitative parameters [6,7,8].
NLS methods are fundamental tools in inverse problems and statistical estimations, particularly in parameter fitting. These methods extend linear regression to handle problems where the variables have a non-linear relationship. They aim to minimize the sum of squared differences between observed and predicted values [9,10]. Theoretical support and soundness of NLS methods such as Jacobian-based approaches (like Gauss-Newton and Levenberg-Marquardt) and derivative-free algorithms (like Nelder-Mead) are strong, making them the preferred choice in multiple applications, but they are generally slow for large multi-dimensional datasets, among other problems [11,12,13].
Gauss-Newton methods are efficient for those problems where models are nearly linear. To achieve better convergence speed, it utilizes the Jacobian matrix to approximate the Hessian [9,10]. The Levenberg-Marquardt algorithm refines the Gauss-Newton method by using the concept of a damping factor. Thus, it enhances convergence stability and performance in more complex non-linear problems when the Hessian is ill-conditioned [11,12]. Powell’s Dog Leg method integrates Gauss-Newton and the steepest descent directions to effectively manage substantial curvature in the error surface. In contrast, hybrid approaches that combine the Levenberg–Marquardt algorithm with Quasi-Newton methods use adaptive step sizes and approximate Hessians to achieve robustness and efficiency in large-scale optimization problems [9,10,11]. Approximation-based versions of these methods, (i.e., adaptations of the Levenberg–Marquardt and Dog Leg approaches) replace exact derivative computations with approximations [12,13]. This substitution reduces computational complexity and improves efficiency [14,15]. Trust Region Conjugate Gradient (TRCG) [9,45] combines trust-region methods with iterative inversion of Hessian or its approximation, being a good choice for parameter fitting. Therefore, these approximation-based methods may perform better on high-dimensional and large-scale optimization problems.
However, these methods come with significant computational demands and the risk of converging to local minima, especially in complex parameter spaces. The accuracy of these methods is significantly dependent on the quality of initial parameter estimates. Thus, poor initialization potentially results in suboptimal solutions.
From the literature, it is found that NLS methods are effective for T1ρ mapping but are hampered by slow computational speeds. As a result, NLS is typically applied only to specific regions of interest (ROIs), such as cartilage voxels in the knee joint. NLS can also be unstable when dealing with multi-component relaxation. To address these issues, researchers have developed variants like non-negative least squares (NNLS) [16,17,18] and regularized NLS (RNLS) [18,19].
NNLS adds a non-negativity constraint to stabilize estimations, ensuring real-valued, monotonically decreasing relaxations [16,17,18]. RNLS extends NLS by incorporating regularization penalties to improve stability and accommodate independent (in-voxel) or multi-voxel (spatial) regularization functions [18,19]. Thus, RNLS is more suitable for recovering all voxels in imaging applications. Improved stability makes RNLS somewhat faster than NLS, but it still requires many iterations.

1.1. Research Motivation

Considering an MR image slice of size 256 × 256 = 65,536 voxels, NLS fitting requires approximately 1 millisecond per voxel on high-performance computing (HPC) systems. Thus, fitting all 65,536 voxels would take about 65.54 seconds for a single slice. Extending this to 4D-MRI data or higher-dimensional scans, such as 512 × 512 or 1024 × 1024, results in exponential increases in computational time. Figure 1 illustrates the time required by our NLS fitting implementation for varying numbers of MRI slices on HPC systems and NVIDIA GPUs. For a single slice, the theoretical time on HPC is 1.09 minutes, while the NVIDIA GPU takes 2.24 minutes. Assuming a linear increase in time with the number of slices in a dataset, one can easily reach the HPC requirement of 1092.3 minutes, and NVIDIA GPU requirement of 2240 minutes for 1000 images. This data highlights the substantial computational demands of NLS fitting.
Due to the time-consuming nature of NLS methods, they are often limited to knee cartilage ROIs. However, recent research highlights the importance of the relaxation map of the full knee joint MRI for assessing osteoarthritis (OA) progression and detecting subtle changes in tissue properties beyond just the cartilage [20]. Therefore, researchers have started exploring artificial intelligence (AI) models as faster alternatives to NLS.

1.2. AI-Based Exponential Fitting

Recent advancements in AI offer suitable alternatives to traditional NLS methods. AI models, such as random forest regression, neural networks (NNs), and deep NNs have been extensively utilized to enhance exponential model fitting [21,22,23,24]. These models benefit from rapid training and deployment, thus providing stable and accurate curve fitting that is usually faster than existing NLS methods [26,27,28]. By utilizing advanced function approximation abilities, AI models can efficiently address the issues of slow convergence and suboptimal solutions associated with traditional NLS methods [23,24,25,26].
By incorporating AI into the fitting process, researchers have achieved significant improvements in parameter mapping [29,30,31,32]. Additionally, AI models can be trained in an end-to-end manner, i.e., learning optimal mappings from raw k-space or undersampled multi-contrast MRI data to T1ρ maps without the need for explicit MRI reconstruction and NLS fitting [33].
Therefore, researchers are actively exploring faster methods for obtaining quantitative MRI values of the knee joint, including deep learning (DL) approaches [34,35,36,37,49,50,51,52]. However, most supervised DL methods [34,35,36,37] rely on reference values produced by NLS methods. This requires applying NLS to thousands of voxels in each 3D image, and this process must be repeated for every image in a large dataset.
Moreover, the integration of AI-based models with traditional NLS methods has led to the development of hybrid approaches that utilize the strengths of both methods. For example, DL models can provide initial parameter estimates or refine results from NLS methods, thus, enhancing accuracy and robustness in T1ρ estimation.
Table 1 presents a description of existing AI-based exponential fitting methods. The table highlights that the majority of AI-based solutions predominantly focus on ME fitting, leaving SE and BE relaxation models relatively underexplored. This limitation creates a significant gap in addressing the diverse relaxation dynamics present in different tissues. Additionally, most existing approaches are often sensitive to noise and prone to inaccuracies in parameter estimation, not providing similar goodness of fit properties as NLS-based techniques.
This gap underscores the importance of hybrid approaches, which integrate the strengths of DL and NLS. Such integration ensures improved parameter estimation, enhanced robustness to noise, and better handling of complex multi-component T1ρ mapping tasks. Thus, Table 1 further emphasizes the need for methods that go beyond ME fitting and utilize hybrid strategies to overcome the limitations of AI-based or traditional approaches.

1.3. Proposed Solution

Although AI-based exponential fitting models have demonstrated remarkable performance, most researchers have primarily focused on mono-exponential (ME) fitting, with limited exploration of bi-exponential (BE) and stretched-exponential (SE) fitting. Additionally, many AI-based exponential fitting methods require actual MRI data as a reference for training. To address these gaps, this paper proposes the HDNLS, a hybrid model for fast multi-component fitting, dedicated in this work for T1ρ mapping of the knee joint. In HDNLS, we initially designed a DL model that uses synthetic data for training, thus, eliminating the need for reference MRI data. However, the estimated map may be sensitive to noise, particularly in the short component of the BE model. To address this, the results from the proposed DL model are further refined using NLS with an optimized number of iterations. To optimize the number of iterations, this paper presents an ablation study in which we analyze the trade-off between the number of NLS iterations and the mean normalized absolute difference (MNAD), as well as between the NLS iterations and normalized root mean squared residuals (NRMSR).

1.4. Contributions

The key contributions of the paper are as follows:
a) Initially, a DL-based multi-component T1ρ mapping method is proposed. This approach utilizes synthetic data for training, thereby eliminating the need for reference MRI data.
b) The HDNLS method is proposed for fast multi-component T1ρ mapping in the knee joint. This method integrates DL and NLS. It effectively addresses key limitations of NLS, including sensitivity to initial guesses, poor convergence speed, and high computational cost.
c) Since HDNLS's performance depends on the number of NLS iterations, it becomes a Pareto optimization problem. To address this, a comprehensive analysis of the impact of NLS iterations on HDNLS performance is conducted.
d) Four variants of HDNLS are suggested that balance accuracy and computational speed. These variants are Ultrafast-NLS, Superfast-HDNLS, HDNLS, and Relaxed-HDNLS. These methods enable users to select a suitable configuration based on their specific speed and performance requirements.

2. Proposed Methodology

The measured signal s of the relaxation of a voxel in position x, at TSL specified by t, can be expressed as:
s x , t = f x , t , θ + v x , t ,
where v is the noise, and f x , t , θ is the relaxation model with parameters θ . For simplicity, we will omit the spatial position x, assuming it corresponds to the relaxation process in one voxel. Note that the parameters θ are also related to that specific voxel.

2.1. Multi-Component Relaxation Models

Generally, ME, BE, and SE models are commonly used for fitting the T1ρ map. Table 2 presents the mathematical formulations, associated parameters, descriptions, and applicable ranges for these three models. The ME model assumes a homogeneous tissue environment, while the BE model introduces two compartments with distinct relaxation times, and the SE model accounts for heterogeneous environments with distributed relaxation rates.
Table 3 highlights the key features, advantages, and disadvantages of each multi-component relaxation model. The choice of the specific model depends on the tissue characteristics and the requirements of the analysis. While the ME model is simple and suitable for homogeneous tissues, the BE model provides a more detailed representation of tissues with two relaxation components but is computationally more demanding. The SE model offers a balance by modeling complex tissue environments with fewer parameters and greater stability, though its physical interpretation can be less intuitive.
Figure 2 illustrates the working of multi-component relaxation models and their respective outputs. It includes T1 ρ weighted images with six TSLs and voxel-wise relaxation measurements. It presents the expected T1ρ maps obtained from ME, BE (long component, fraction of long component, and short component), and SE models (SE-T1ρ map and Beta (β) parameter).
The NLS approach fits the parameters voxel-wise minimizing the squared Euclidean norm of the residual, using:
θ ^ = arg min θ t = 1 T s t f t , θ 2 ,
where T is the number of TSLs. This is solved using iterative algorithms such as in [9].

2.2. Deep Learning-Based Multi-Component Relaxation Models

Figure 3 shows the diagram of the voxel-by-voxel working of the DL model for fitting multi-component T1ρ relaxation models. Typically, NNs are trained by minimizing the squared error of the target parameters. However, to achieve data consistency, as in NLS estimations, we used a customized loss function, designed as:
L o s s = i γ s s i f ^ i 2 + γ θ θ i θ ^ i 2 ,
where i is the index of the data elements in the training dataset, θ ^ i = F w s i , being s = [ s 1 s T ] T , a column vector with the measured relaxation signal, F w the fitting network with learning parameters w , and f ^ i is the relaxation signal produced with the estimated parameters θ ^ i , using any of the equations presented in Table 2. Target parameters θ i are sampled from a known distribution Θ , and s i are synthetically generated using equation (1) and equations presented in Table 2 and the sampled parameters.
The first term of the loss function tries to reduce the residue, as in NLS approaches, producing a relaxation curve consistent with the synthetically generated measured values at the voxel. The second term tries to reduce the error against the ground truth values. This term also acts as a regularization in the NLS sense. We used γ s = 10 and γ θ = 1 .
The NN architecture comprises 7 repeated blocks of fully connected layers, each containing 512 intermediate elements followed by a nonlinear activation function and dropouts. Thereafter, we utilize customized layers to evaluate the relaxation values for the customized loss, as described in equation (3). Finally, the regression layer is used to predict the T1ρ maps. This network estimates the relaxation parameters voxel-wise, like any of the NLS-based approaches.

2.3. Sensitivity Analysis of DL Model

We performed a sensitivity analysis of the proposed DL model to determine the optimal configuration of hyperparameters. Initially, multiple trial-and-error experiments were conducted using various hyperparameter values commonly reported in the literature. The final configuration comprised 7 fully connected layers, each containing 512 intermediate elements. Five different activation functions are also evaluated such as Gaussian Error Linear Unit (GELU), Leaky ReLU, hyperbolic tangent (Tanh), Swish, and ReLU. The model was trained using the Adam optimizer, with a learning rate initialized at 1.0e-3, over a maximum of 100 epochs. Each epoch processed mini-batches size of 200 samples. We selected γ s = 10 and γ θ = 1 as the optimal weights for the proposed customized loss function. During the sensitivity analysis, we varied only the specific parameter under evaluation while keeping all other parameters fixed.
Figure 4 illustrates the impact of (a) optimizer selection (Adam, SGD, and RMSProp), (b) number of epochs (20 to 200), (c) batch size (50 to 1000), and (d) number of layers (4 to 11). From Figure 4 (a), it is evident that the RMSE shows a decreasing trend as the number of epochs increases. It eventually stabilized around 100 epochs. Adam outperformed SGD and RMSProp by achieving significantly lower RMSE values (see Figure 4 (b)). Smaller batch sizes achieved lower RMSE values with performance stabilizing beyond a batch size of 200 (see Figure 4 (c)). Finally, seven fully connected layers achieved noticeably lower RMSE values compared to other configurations (see Figure 4 (d)).
Figure 5 presents the sensitivity analysis of activation functions (ReLU, Leaky ReLU, GELU, Tanh, and Swish) for the DL model. The analysis evaluates these functions across three optimizers such as Adam, SGD, and RMSProp. It shows RMSE variations for ME, SE, and BE components under different activation function-optimizer combinations. Among these, the Adam optimizer combined with the ReLU activation function achieves significantly lower RMSE values across all ME, SE, and BE components.
Finally, the sensitivity analysis is conducted to evaluate the impact of varying γ s and γ θ in the customized loss function on validation data. For better analysis, we evaluated the average RMSE values across all ME, SE, and BE components. Figure 6 (a) shows that increasing γ θ from 1 to 10, with γ s = 1, results in a rise in RMSE. This emphasizes the importance of keeping γ θ relatively low. In contrast, Figure 6 (b) demonstrates a decreasing trend in RMSE as γ s increases from 1 to 15, with γ θ = 1. This indicates the benefits of assigning a higher weight to γ s . However, continually increasing γ s in the loss function places an increasingly higher emphasis on minimizing the discrepancy between the measured relaxation signal s i and the model-predicted signal f ^ i . While this might improve the consistency of the relaxation curve with the synthetic measurements, it can increase the error in the estimated parameters. The model prioritizes signal fitting at the expense of accurately estimating the parameters θ i , as the second term in the loss function becomes relatively low, diminished the impact of the regularization term in stabilizing the parameter estimation. As shown in Figure 6 (b), RMSE increases when γ s exceeds 10, indicating a poor balance between signal consistency and parameter accuracy. Based on these observations, we selected γ s = 10 and γ θ = 1 as the optimal weights for our loss function.
and γ θ on validation data. (a) Impact of γ θ on RMSE with γ s fixed at 1, demonstrating an increase in RMSE as γ θ rises from 1 to 10. (b) Influence of γ s on RMSE with γ θ fixed at 1, showing decreasing RMSE values as γ s increases from 1 to 15. However, RMSE starts to increase beyond γ s = 10.

2.4. Training and Validation Analysis

To ensure a robust evaluation of the proposed DL model, the synthetic dataset was divided into three subsets: training, validation, and test data, in an 80:10:10 ratio. The training subset (80% of the data) was used to train the DL model and optimize its parameters. To prevent overfitting, the validation subset (10% of the data) was employed to monitor the model’s performance during training and fine-tune hyperparameters.
Figure 7 illustrates the training and validation loss curves for 10 and 100 epochs, respectively. In Figure 7 (a), the difference between training and validation loss values is more apparent, reflecting the initial stages of model training. In contrast, Figure 7 (b) shows that, with 100 epochs, it is difficult to distinguish between the loss values for both training and validation converge. This convergence indicates that the DL model does not suffer from overfitting.
This partitioning strategy also ensures efficient training of the DL model while providing a fair and unbiased assessment of its generalization capability. Data for each subset was randomly selected in a stratified manner. Therefore, the used data also preserves the distribution of key features across all subsets. This approach guarantees consistency and reliability in the evaluation process.
Finally, the test subset, comprising 10% of the data, was reserved for the final evaluation of the model’s performance on unseen data (see analysis in Figure 9 and Figure 10 and Table 7).

2.5. HDNLS

Figure 8 illustrates the proposed HDNLS-based fitting of multi-component T1ρ relaxation models. The estimated map obtained from the DL model may be sensitive to noise, particularly in the short component of BE. To mitigate this issue, the results from the DL model are further refined using NLS iterations. The DL results serve as an initial guess for NLS. This helps to address both the initial guess dependency and poor convergence speed associated with NLS.
In this paper, TRCG [9,45] is used to solve the NLS problem. At each iteration k , the parameters are estimated using:
θ k + 1 N L S _ U p d a t e s , θ k
This update aims to minimize the residuals as:
r k + 1 s f θ k + 1 2
Here, s represents the observed data, and f ( θ ( k + 1 ) ) denotes the model predictions based on the updated parameters. TRCG ensures that each step effectively refines the parameter estimates. Thus, it enhances the convergence of the optimization process to achieve an optimal fit to the data.

2.6. T1ρ -MRI Data Acquisition and Fitting Methods

All the experiments were performed on a 3T MRI scanner with the magnetization-prepared angle-modulated partitioned k-space spoiled gradient echo sequence snapshots (MAPSS) [44], using six spin-lock times (TSL) values 0.05, 0.6, 1.6, 6, 16, and 36 ms. We scanned a total of 12 healthy volunteers, with a mean age of 38 and SD=12.4 years old (n=7 males and n=5 females). The MRI scans have a voxel size of 0.7mm×0.7mm×4mm, with an FOV of 180mm×180mm×96mm. This study was approved by the institutional review board (IRB) of New York University Langone Health and was compliant with the Health Insurance Portability and Accountability Act (HIPAA). All volunteers provided their consent before MRI scanning.
For NLS-based fitting, we used 2000 iterations of the TRCG method [45]. The ME model assumes the T1ρ values can be in the range of 1-300 ms. The SE model assumes the T1ρ values can be in the range of 1-300 ms and β can be in the range of 0.1-1. Because we observed no ME relaxation below 5ms, we assume that the long component of the BE model can be in the range of 5-300 ms, the short component can be in the range of 0.5-4 ms, and the BE fraction can be in the range of 0.01-0.99.
The DL model is trained using the Adam optimizer with a maximum of 100 epochs, and each epoch processes mini-batches of 200 samples. The initial learning rate is set to 1.0e-3. The dropout value is set to be 0.1. The proposed model is tested on 3D-T1ρ MRI data from five different healthy subjects. The DL model was trained on 5 healthy data sets.

2.7. Performance Metrics

For comparative analysis, four performance metrics are employed: the median of normalized absolute difference (MNAD), normalized root mean squared residuals (NRMSR), fitting time, and normalized root mean squared error (NRMSE). Table 4 provides detailed descriptions and mathematical formulations of these metrics.

3. Optimal Selection of NLS Iterations for HDNLS

The main objective of this section is to evaluate and compare the performance metrics of NLS and HDNLS on synthetic data by varying the number of NLS iterations. Specifically, the focus is on assessing the convergence speed of both methods and determining whether HDNLS can achieve similar performance to NLS with improved speed.
Figure 9 shows the comparison between NLS and HDNLS on synthetic data in terms of MNAD across three different fitting models such as ME, SE, and BE. Row 1 shows the trade-off between fitting time (blue) and MNAD (red) for NLS as the number of NLS iterations increases. For each fitting model, MNAD decreases initially with more iterations but eventually shows a minor reduction, while the fitting time increases at a rapid rate. Row 2 shows the performance trade-off between fitting time and MNAD for the HDNLS approach. It also shows that as the number of NLS iterations increases, the HDNLS achieves significantly lesser MNAD values. However, after 500 iterations, the improvement in MNAD becomes marginal compared to the increase in fitting time. Row 3 provides direct comparisons between NLS and HDNLS. It shows that HDNLS takes significantly less time compared to NLS when the number of NLS iterations is less than 300. Thereafter, both NLS and HDNLS show almost similar MNAD values with no significant difference.
Similarly, Figure 10 presents the performance of NLS and HDNLS across ME, SE, and BE fitting models on synthetic data in terms of NRMSR. The yellow line represents the NRMSR of noise only. The primary goal of NLS and HDNLS is to reduce the NRMSR to levels just below the noise-level. It is found that in the case of HDNLS, there are little or no benefits in exceeding 500 iterations, except for the SE component. Please note that due to the significant variations in the scale of values for both NLS and HDNLS, it becomes difficult to observe changes in NLS when the number of iterations exceeds 100, 500, and 1000, respectively for ME, BE, and SE components. To better illustrate this, we present the comparison of NRMSR between NLS and HDNLS in row 3, clearly demonstrating the improved convergence speed of HDNLS over NLS. Additionally, the difference in NRMSR for HDNLS after 500 iterations is minimal and can be disregarded.
Figure 9. Comparison of NLS and HDNLS across mono-exponential, bi-exponential, and stretched-exponential fitting models on synthetic data. Row 1 illustrates the trade-off between fitting time and MNAD for NLS. Row 2 shows the trade-off between fitting time and MNAD for HDNLS. Row 3 shows the convergence of NLS and HDNLS in terms of MNAD.
Figure 9. Comparison of NLS and HDNLS across mono-exponential, bi-exponential, and stretched-exponential fitting models on synthetic data. Row 1 illustrates the trade-off between fitting time and MNAD for NLS. Row 2 shows the trade-off between fitting time and MNAD for HDNLS. Row 3 shows the convergence of NLS and HDNLS in terms of MNAD.
Preprints 142530 g009
From Row 3 of both Figure 9 and Figure 10, it is evident that HDNLS converges significantly faster than standard NLS in terms of MNAD and NRMSR, respectively. This makes HDNLS an efficient choice for determining the number of NLS iterations based on specific requirements. For scenarios where speed is the highest priority and small performance compromises are acceptable, 10 iterations, referred to as Ultrafast-NLS, provide quick results with minimal computational cost. Superfast-HDNLS, with 50 NLS iterations, strikes a balance between speed and performance, offering a faster alternative with acceptable accuracy. HDNLS with 200 iterations achieves an optimal middle ground, delivering both rapid convergence and higher performance. However, for SE, we observed that the results at 200 iterations are not as desirable as those for ME and BE; therefore, we selected HDNLS with 300 iterations for SE map fitting only.
For those who prioritize performance over speed, 500 iterations, termed Relaxed-HDNLS, provide the best results while tolerating slower computation times. Thus, HDNLS allows users to select a suitable configuration based on their specific speed and performance requirements.
Figure 10. Performance metrics comparison of NLS and HDNLS in terms of fitting time and normalized root mean squared residual (NRMSR) across ME, SE, and BE fitting models. Row 1 shows the trade-off between fitting time and NRMSR for NLS. Row 2 shows the trade-off between fitting time and NRMSR for HDNLS. Row 3 presents the convergence of both NLS and HDNLS with respect to log (NRMSR).
Figure 10. Performance metrics comparison of NLS and HDNLS in terms of fitting time and normalized root mean squared residual (NRMSR) across ME, SE, and BE fitting models. Row 1 shows the trade-off between fitting time and NRMSR for NLS. Row 2 shows the trade-off between fitting time and NRMSR for HDNLS. Row 3 presents the convergence of both NLS and HDNLS with respect to log (NRMSR).
Preprints 142530 g010
To check the performance of these HDNLS variants, we further tested them on real MRI data (see Section 2.4 for dataset details). Due to the unavailability of ground truth T1ρ maps, we used the NLS-based estimated T1ρ map as the ground truth. The estimated ME and SE T1ρ maps of the whole knee from NLS and HDNLS variants are presented in Figure 11, along with their respective squared error compared to the NLS-based ME T1ρ map. Similarly, Figure 12 shows the ME and SE T1ρ maps for the knee cartilage only. Both figures demonstrate that HDNLS and Relaxed-HDNLS achieve significantly better results, with notably lower errors, compared to Ultra-HDNLS and Super-HDNLS.
Although it is challenging to visually distinguish differences between the obtained T1ρ maps, the evaluated squared error differences provide a clear distinction. Higher squared error values indicate poorer performance. From Figure 11 and Figure 12, the squared error maps for HDNLS and Relaxed-HDNLS show significantly lower errors. This highlights their superior performance and closer alignment with the NLS-based T1ρ maps.
Figure 13 presents the estimated BE long T1ρ component, the fraction of the long T1ρ component, and short T1ρ maps of the whole knee from NLS and HDNLS variants, along with their respective squared errors compared to the ground truth, i.e., the NLS-based ME T1ρ map. Similarly, Figure 14 shows the estimated BE long T1ρ component, the fraction of the long T1ρ component, and short T1ρ maps for knee cartilage only. Both figures demonstrate that HDNLS and Relaxed HDNLS achieve significantly better results with much lower errors compared to Ultra-HDNLS and Super-HDNLS.
Table 5 and Table 6 provide a quantitative analysis of HDNLS variants, i.e., Ultra-HDNLS, Super-HDNLS, HDNLS, and Relaxed-HDNLS, for estimating ME, SE, and BE T1ρ maps using real data for the full knee and knee cartilage, respectively. The Relaxed-HDNLS variant outperforms the others by achieving the minimum MNAD and NRMSR values. In contrast, Ultra-HDNLS provides the fastest fitting time for the estimation of all ME, SE, and BE T1ρ maps, but has higher MNAD and NRMSR values. Therefore, Table 5 and Table 6 highlight a trade-off in selecting the best approach. Faster fitting times come at the cost of performance. This necessitates careful consideration of whether to prioritize speed or accuracy when estimating T1ρ maps.

4. Performance Analysis

Based on the visual and quantitative analysis presented in Section 3, we have identified a trade-off in selecting the best approach. Faster fitting times come at the cost of performance, necessitating careful consideration of whether to prioritize speed or accuracy when estimating T1ρ maps. Therefore, we have selected an optimized HDNLS variant named HDNLS. For ME and BE, it will use 200 NLS iterations, while for SE, we have chosen 300 iterations. The idea is to select a solution that is around 30 times faster than NLS with significantly fewer errors on synthetic data. For ME and BE, when NLS iterations are 200, we achieve around 34 times better fitting speed with comparable performance to NLS. However, with 200 iterations for SE fitting, we observed slightly more errors compared to when NLS iterations were set to 300. Due to this, for SE fitting, HDNLS (NLS iterations=300) shows a speed improvement of approximately 27 times better than NLS, with similar results compared to NLS.
The main objective of this section is to compare the performance of HDNLS against NLS, RNLS, and DL.
Figure 15 and Figure 16 present the estimated ME and SE T1ρ maps for the full MRI joint and knee cartilage, obtained using NLS, RNLS, DL, and HDNLS, respectively. The error is also computed between NLS and DL, and NLS and HDNLS for the full knee joint and knee cartilage ROIs. The analysis shows that the error between NLS and HDNLS is lower compared to that between NLS and DL. This indicates that, compared to DL, the proposed HDNLS method can more effectively serve as an alternative to NLS and RNLS for estimating ME T1ρ, SE T1ρ, and SE beta maps.
The higher error observed in the BE short component is due to its susceptibility to noise caused by low signal intensity and rapid decay dynamics. This fast decay can lead to insufficient signal sampling, negatively affecting accurate fitting. Furthermore, short components exhibit increased sensitivity to minor variations in model parameters, resulting in escalated errors during the fitting process.
Figure 17 and Figure 18 show the predicted BE long T1ρ maps, the fraction of the long BE T1ρ component, and the short T1ρ maps for the full knee joint and knee cartilage ROIs, generated using NLS, RNLS, DL, and HDNLS, respectively. The error between NLS and DL, and between NLS and HDNLS, is also calculated for both the full knee joint and knee cartilage ROIs. The analysis demonstrates that the error between NLS and HDNLS is lower than that between NLS and DL. Thus, it is revealed that, compared to DL, the HDNLS method is a more effective alternative to NLS and RNLS for estimating BE components.
Table 7 presents a quantitative analysis of NLS, RNLS, DL, and HDNLS methods for multi-component T1ρ mapping (ME, SE, and BE) on synthetic data in terms of MNAD, fitting time (in seconds), and NRMSR. It indicates that though, NLS outperforms RNLS, DL, and HDNLS, but compared to RNSL and HDNLS this difference is not significant. Additionally, DL achieved better speed compared to the other approaches, but it also achieved the poorest performance. Overall, the HDNLS achieved comparable performance to NLS and RNLS and significantly better performance over DL.
Table 7. Quantitative analysis of NLS, RNLS, DL, and HDNLS for ME, SE, and BE T1ρ maps on synthetic data.
Table 7. Quantitative analysis of NLS, RNLS, DL, and HDNLS for ME, SE, and BE T1ρ maps on synthetic data.
Method Metric NLS RNLS DL HDNLS
ME MNAD (%) 3.65 3.69 4.29 3.70
Fitting time (sec) 37.48 32.83 0.79 4.94
NRMSR 0.02 0.02 0.03 0.02
SE MNAD (%) 6.18 6.27 7.61 6.32
Fitting time (sec) 41.9 37.75 0.61 4.17
NRMSR 0.15 0.15 0.18 0.16
BE MNAD (%) 16.19 16.36 18.37 16.47
Fitting time (sec) 11.73 11.16 0.52 2.48
NRMSR 0.02 0.02 0.02 0.02
Table 8 presents a comparison of MNAD and NRMSE for DL, and HDNLS methods for multi-component T1ρ mapping (ME, SE, and BE) on real data for both the full knee joint and knee cartilage ROI. The results show that HDNLS significantly outperforms DL by achieving lower MNAD and NRMSE values across all ME, SE, and BE categories. This indicates that HDNLS is a better alternative to NLS than DL for T1ρ mapping.
Table 9 presents a computational speed (in seconds) analysis of NLS, RNLS, DL, and HDNLS methods for multi-component T1ρ mapping (ME, SE, and BE) on real data for both the full knee and knee cartilage ROI. The results indicate that HDNLS significantly outperforms the traditional methods (NLS and RNLS) in terms of computational speed, although it is slower than DL. Though DL demonstrates substantial speed improvements over NLS, RNLS, and HDNLS, it also results in greater error compared to HDNLS (see Table 8). As stated in Section 3, our objective was to find a suitable trade-off between performance and computational speed. Therefore, while HDNLS takes more time than DL, it also shows significant performance improvements over DL. Thus, HDNLS achieves comparable performance to NLS and RNLS. Compared to NLS, it provides approximately 13 to 14 times better speed for each ME, SE, and BE component for the full knee joint. Additionally, for the knee cartilage ROI, it offers about 11, 12, and 11 times better speed over NLS, respectively.

5. Discussion, Limitations, and Future Directions

5.1. Discussion

The proposed HDNLS method effectively addresses the inherent limitations in traditional NLS [6,7,8,9,10,11,12,13,14,15], NNLS [16,17,18], RNLS [18,19], and AI-based [21,22,23,24,25,26] T1ρ fitting methods by combining their strengths. Since NLS, NNLS, and RNLS approaches are often slow and necessitate careful initialization, therefore, HDNLS employs a DL model to provide an initial guess for the NLS fitting process. Additionally, HDNLS eliminates the dependency on reference data, as it was trained on synthetic data. By utilizing DL predictions for NLS initialization, HDNLS achieves superior parameter estimation compared to AI-based models [21,22,23,24,25,26], while avoiding the high computational costs typically associated with conventional methods such as NLS [6,7,8,9,10,11,12,13,14,15], NNLS [16,17,18], RNLS [18,19]. Additionally, HDNLS extends its capabilities beyond end-to-end DL models [33,34,35,36,37], which are primarily designed for ME fitting, accommodating various multi-component fitting scenarios such as BE and SE.
As we can see in Table 7, HDNLS significantly outperforms DL by achieving lower MNAD and NRMSE values across all categories for ME, SE, and BE on real data from both the full knee joint and knee cartilage ROI. Thus, HDNLS is a more robust and superior alternative to DL for accurate T1ρ mapping. However, HDNLS has poorer computational speed than DL due to the use of NLS iterations, as indicated in Table 8. Nonetheless, HDNLS markedly surpasses traditional methods like NLS and RNLS in speed. The rapid speed of DL comes at the cost of accuracy, as reflected in the higher MNAD and NRMSE values compared to HDNLS. Our aim, as outlined in Section 3, was to strike a balance between performance metrics and computational efficiency.
Therefore, we prefer HDNLS over DL, even though it requires more time than DL. HDNLS provides significant performance enhancements, achieving results comparable to those of NLS and RNLS. Specifically, HDNLS achieves approximately 13 to 14 times the speed of NLS across ME, SE, and BE components for full knee joint analysis and offers about 11 to 12 times better speed for the knee cartilage ROI. Therefore, HDNLS stands out as an efficient solution for fast multi-component T1ρ mapping, particularly in knee joint imaging, striking an optimal balance between efficiency and high precision.

5.2. Limitations of HDNLS

Although the proposed HDNLS model demonstrates promising performance in T1ρ mapping, it does have certain limitations, which include:
a)
The analysis of the BE short component reveals significant challenges due to its susceptibility to noise and rapid decay dynamics. These factors contribute to increased errors in fitting processes, making accurate estimation of BE short relaxation times particularly difficult. The sensitivity of these components to minor parameter variations further complicates the issue.
b)
Despite its computational efficiency compared to traditional NLS and RNLS methods, the iterative nature of HDNLS makes it more computationally demanding than alternative DL approaches.
c)
Although HDNLS demonstrates comparable performance in T1ρ mapping, its applicability to other quantitative MRI parameters, such as T2 mapping or diffusion tensor imaging (DTI), remains unexplored.
d)
The current implementation of HDNLS does not incorporate self-supervised learning techniques. As a result, it may not fully utilize undersampled multi-contrast MRI data.

5.3. Future Directions

Although HDNLS has demonstrated remarkable performance, several challenges remain that need to be addressed in the future.
a)
Further studies are essential to evaluate the performance of HDNLS across diverse patient populations and imaging protocols. Testing the model with data from various MRI scanners is crucial. This will help identify potential limitations and improve generalizability in clinical settings.
b)
While HDNLS primarily targets T1ρ mapping, future research could adapt its methodology for other quantitative MRI parameters, such as T2 mapping and diffusion tensor imaging (DTI). This broader application could provide valuable insights into tissue characterization and significantly enhance diagnostic capabilities.
c)
Self-supervised learning techniques could minimize reliance on labeled and synthetic data. This would enhance the training efficiency of HDNLS. Additionally, this approach may improve HDNLS's ability to learn from raw or undersampled multi-contrast MRI data. Consequently, we can develop an end-to-end HDNLS model for parameter fitting.
d)
Optimizing the DL architecture or incorporating advanced techniques like attention mechanisms or transformer networks could enhance HDNLS's ability to capture complex data patterns. This improvement would reduce the number of NLS iterations required. As a result, the overall process would become faster. Additionally, using advanced regularization techniques could make HDNLS more robust to noise and outliers. This would ensure more accurate parameter estimation in challenging scenarios.
e)
The BE short component is inherently challenging to estimate due to its rapid decay dynamics and low signal intensity, making it more susceptible to noise. We propose the following steps to address this issue in the near future (i) Implement regularization techniques, such as spatial smoothness constraints and multi-voxel regularization, to stabilize parameter estimation across neighboring voxels. (ii) Integrate advanced noise-aware DL architectures, incorporating uncertainty quantification, into the DL component of HDNLS. These architectures will explicitly account for input data variability, improving robustness and reliability under noisy conditions. (iii) Explore adaptive weighting schemes in the loss function during training for BE fitting to assign appropriate emphasis to the short component. This approach aims to reduce the sensitivity of the short component to noise while maintaining overall model performance.
f)
HDNLS can be extended using a pre-trained model on similar domains, such as quantitative MRI or exponential model fitting tasks. This may accelerate convergence, improve generalization, and potentially expand the range of applicability for voxel-wise multi-component T1ρ fitting.
g)
As part of future studies, we will evaluate the HDNLS framework in clinical workflows to assess its utility and impact on patient diagnosis and treatment planning.

6. Conclusion

This paper introduced HDNLS, a hybrid model that combined DL and NLS for efficient multi-component T1ρ mapping in the knee joint. HDNLS effectively addressed key limitations of NLS such as sensitivity to initial guesses, poor convergence speed, and high computational costs. By utilizing synthetic data for training of DL model, we eliminated the need for reference MRI data and enhanced noise robustness through NLS refinements. We conducted a comprehensive analysis of NLS iterations on FastNL and proposed four variants. These variants were Ultrafast-NLS, Superfast-HDNLS, HDNLS, and Relaxed HDNLS. Among these, HDNLS achieved an optimal balance between accuracy and speed. HDNLS achieved performance comparable to NLS and RNLS with a minimum 13-times increase in fitting speed. Although HDNLS is slower than DL, it significantly outperformed in terms of MNAD and NRMSE. Therefore, HDNLS serves as a robust and faster alternative to NLS for multi-component T1ρ fitting.
Author Contribution: Conceptualization, methodology, and writing—original draft preparation D.S.; visualization, supervision, project administration, and funding acquisition M.V.W.Z and R.R.R

Funding

This study was supported by NIH grants, R01-AR076328-01A1, R01-AR076985-01A1, and R01-AR078308-01A1 and was performed under the rubric of the Center of Advanced Imaging Innovation and Research (CAI2R), an NIBIB Biomedical Technology Resource Center (NIH P41- EB017183).

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Institutional Review Board (or Ethics Committee) of NYU Langone Health (protocol code i21-00710 and date of approval March 13th, 2022).

Informed Consent Statement

This study was approved by the institutional review board (IRB) of New York University Langone Health and was compliant with the Health Insurance Portability and Accountability Act (HIPAA). All volunteers provided their consent before MRI scanning.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author. The data are not publicly available due to privacy and ethical policy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Vergeldt, Frank J., Alena Prusova, Farzad Fereidouni, Herbert van Amerongen, Henk Van As, Tom WJ Scheenen, and Arjen N. Bader. "Multi-component quantitative magnetic resonance imaging by phasor representation." Scientific reports 7, no. 1 (2017): 861.
  2. Rosenkrantz, Andrew B., Mishal Mendiratta-Lala, Brian J. Bartholmai, Dhakshinamoorthy Ganeshan, Richard G. Abramson, Kirsteen R. Burton, J. Yu John-Paul et al. "Clinical utility of quantitative imaging." Academic radiology 22, no. 1 (2015): 33-49.
  3. Keenan, Kathryn E., Joshua R. Biller, Jana G. Delfino, Michael A. Boss, Mark D. Does, Jeffrey L. Evelhoch, Mark A. Griswold et al. "Recommendations towards standards for quantitative MRI (qMRI) and outstanding needs." Journal of magnetic resonance imaging: JMRI 49, no. 7 (2019): e26.
  4. Richter, Ryan H., Douglas Byerly, Donald Schultz, and Liem T. Mansfield. "Challenges in the interpretation of MRI examinations without radiographic correlation: pearls and pitfalls to avoid." Cureus 13, no. 7 (2021).
  5. Cercignani, Mara, Nicholas G. Dowell, and Paul S. Tofts. Quantitative MRI of the brain: principles of physical measurement. Boca Raton, Florida, 2018.
  6. Brown, Robert W., Y-C. Norman Cheng, E. Mark Haacke, Michael R. Thompson, and Ramesh Venkatesan. Magnetic resonance imaging: physical principles and sequence design. John Wiley & Sons, 2014.
  7. Kuperman, Vadim. Magnetic resonance imaging: physical principles and applications. Elsevier, 2000.
  8. Jelescu, Ileana O., Jelle Veraart, Els Fieremans, and Dmitry S. Novikov. "Degeneracy in model parameter estimation for multi-compartmental diffusion in neuronal tissue." NMR in Biomedicine 29, no. 1 (2016): 33-47.
  9. Madsen, Kaj, Hans Bruun Nielsen, and Ole Tingleff. "Methods for non-linear least squares problems." (2004).
  10. Dong, Guozhi, Moritz Flaschel, Michael Hintermüller, Kostas Papafitsoros, Clemens Sirotenko, and Karsten Tabelow. "Data-driven methods for quantitative imaging. arXiv:2404.07886 (2024).
  11. Levenberg, Kenneth. "A method for the solution of certain non-linear problems in least squares." Quarterly of applied mathematics 2, no. 2 (1944): 164-168.
  12. Gavin, Henri P. "The Levenberg-Marquardt algorithm for nonlinear least squares curve-fitting problems." Department of Civil and Environmental Engineering Duke University August 3 (2019).
  13. Shterenlikht, A. , and N. A. Alexander. "Levenberg–Marquardt vs Powell’s dogleg method for Gurson–Tvergaard–Needleman plasticity model." Computer Methods in Applied Mechanics and Engineering 237 (2012): 1-9.
  14. Marquardt, Donald W. "An algorithm for least-squares estimation of nonlinear parameters." Journal of the society for Industrial and Applied Mathematics 11, no. 2 (1963): 431-441.
  15. Nelder, John A., and Roger Mead. "A simplex method for function minimization." The computer journal 7, no. 4 (1965): 308-313.
  16. Hwang, D., Du, Y.P.: Improved myelin water quantification using spatially regularized non-negative least squares algorithm. Journal of Magnetic Resonance Imaging 30(1), 203-208 (2009).
  17. Fenrich, F.R.E., Beaulieu, C., Allen, P.S.: Relaxation times and microstructures. NMR in Biomedicine 14(2), 133-139 (2001).
  18. Zibetti, M.V.W., Helou, E.S., Sharafi, A., Regatte, R.R.: Fast multicomponent 3D-T1ρ relaxometry. NMR in Biomedicine 33(12), e4318 (2020).
  19. Eriksson, J.: Optimization and Regularization of Nonlinear Least Squares Problems. Jerusalem: Verlag nicht ermittelbar (1996).
  20. Dilbag Singh, Ravinder R. Regatte, and Marcelo V. W. Zibetti, “Self-Supervised Deep-Learning Networks for Mono and Bi-exponential T1ρ Fitting in the Knee Joint,” ISMRM, 2024, Singapore. [Available Online:] https://submissions.mirasmart.com/ISMRM2024/Itinerary/?
  21. Nedjati-Gilani, G. L., Schneider, T., Hall, M. G., Cawley, N., Hill, I., Ciccarelli, O., Drobnjak, I., Gandini Wheeler-Kingshott, C. A. M., & Alexander, D. C. (2017). Machine learning based compartment models with permeability for white matter microstructure imaging. NeuroImage, 150, 119–135.
  22. Barbieri, S., Gurney-Champion, O. J., Klaassen, R., & Thoeny, H. C. (2020). Deep learning how to fit an intravoxel incoherent motion model to diffusion-weighted MRI. Magnetic Resonance in Medicine, 83(1), 312–321.
  23. Sjölund, J., Eklund, A., Özarslan, E., Herberthson, M., Bänkestad, M., & Knutsson, H. (2018). Bayesian uncertainty quantification in linear models for diffusion MRI. NeuroImage, 175, 272–285.
  24. Liu, H., & Li, D. K. B. (2020). Myelin water imaging data analysis in less than one minute. NeuroImage, 210, 116551.
  25. Rafati, Jacob, and Roummel F. Marcia. "Improving L-BFGS initialization for trust-region methods in deep learning." In 2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA), pp. 501-508. IEEE, 2018.
  26. Bishop, Christopher M., and C. M. Roach. "Fast curve fitting using neural networks." Review of scientific instruments 63, no. 10 (1992): 4450-4456.
  27. Gözcü, Baran, Rabeeh Karimi Mahabadi, Yen-Huan Li, Efe Ilıcak, Tolga Cukur, Jonathan Scarlett, and Volkan Cevher. "Learning-based compressive MRI." IEEE transactions on medical imaging 37, no. 6 (2018): 1394-1406.
  28. Jung, Soozy, Hongpyo Lee, Kanghyun Ryu, Jae Eun Song, Mina Park, Won-Jin Moon, and Dong-Hyun Kim. "Artificial neural network for multi-echo gradient echo–based myelin water fraction estimation." Magnetic resonance in medicine 85, no. 1 (2021): 380-389.
  29. Zibetti, Marcelo VW, Gabor T. Herman, and Ravinder R. Regatte. "Fast data-driven learning of parallel MRI sampling patterns for large scale problems." Scientific Reports 11, no. 1 (2021): 19312.
  30. Bertleff, Marco, Sebastian Domsch, Sebastian Weingärtner, Jascha Zapp, Kieran O'Brien, Markus Barth, and Lothar R. Schad. "Diffusion parameter mapping with the combined intravoxel incoherent motion and kurtosis model using artificial neural networks at 3 T." NMR in Biomedicine 30, no. 12 (2017): e3833.
  31. Domsch, Sebastian, Bettina Mürle, Sebastian Weingärtner, Jascha Zapp, Frederik Wenz, and Lothar R. Schad. "Oxygen extraction fraction mapping at 3 Tesla using an artificial neural network: a feasibility study." Magnetic resonance in medicine 79, no. 2 (2018): 890-899.
  32. Balasubramanyam, C., M. S. Ajay, K. R. Spandana, Amogh B. Shetty, and K. N. Seetharamu. "Curve fitting for coarse data using artificial neural network." WSEAS Transaction on Mathematics 13 (2014): 406-415.
  33. Liu, Fang, Li Feng, and Richard Kijowski. "MANTIS: model-augmented neural network with incoherent k-space sampling for efficient MR parameter mapping." Magnetic resonance in medicine 82, no. 1 (2019): 174-188.
  34. Fu, Z., Mandava, S., Keerthivasan, M.B., Li, Z., Johnson, K., Martin, D.R., Altbach, M.I., Bilgin, A.: A multi-scale residual network for accelerated radial MR parameter mapping. Magnetic Resonance Imaging 73, 152-162 (2020).
  35. Li, H., Yang, M., Kim, J.H., Zhang, C., Liu, R., Huang, P., Liang, D., Zhang, X., Li, X., Ying, L.: SuperMAP: Deep ultra-fast MR relaxometry with joint spatiotemporal undersampling. Magnetic Resonance in Medicine 89(1), 64-76 (2023).
  36. Liu, F., Samsonov, A., Chen, L., Kijowski, R., Feng, L.: SANTIS: Sampling-augmented neural network with incoherent structure for MR image reconstruction. Magnetic Resonance in Medicine 82(5), 1890-1904 (2019).
  37. Liu, F., Kijowski, R., El Fakhri, G., Feng, L.: Magnetic resonance parameter mapping using model-guided self-supervised deep learning. Magnetic Resonance in Medicine 85(6), 3211-3226 (2021).
  38. H. Shao et al., “UTE bi-component analysis of T2* relaxation in articular cartilage,” Osteoarthr. Cartil., vol. 24, no. 2, pp. 2016. [CrossRef]
  39. Sharafi, D. Xia, G. Chang, and R. R. Regatte, “Biexponential T1ρ relaxation mapping of human knee cartilage in vivo at 3 T,” NMR Biomed., vol. 30, no. 10, Oct. 2017. [CrossRef]
  40. J. G. Bakker and J. Vriend, “Multi-exponential water proton spin-lattice relaxation in biological tissues and its implications for quantitative NMR imaging,” Phys. Med. Biol., vol. 29, no. 5, pp. 509–518, 1984. [CrossRef]
  41. A. Reiter, R. L. Magin, W. Li, J. J. Trujillo, M. Pilar Velasco, and R. G. Spencer, “Anomalous T2 relaxation in normal and degraded cartilage,” Magn. Reson. Med., vol. 76, no. 3, pp. 953–962, Sep. 2016. [CrossRef]
  42. R. Wilson, L. Bowen, W. Kim, D. Reiter, and C. Neu, “Stretched-Exponential Modeling of Anomalous T1ρ and T2 Relaxation in the Intervertebral Disc In Vivo,” bioRxiv. p. 2020.05.21.109785, 2020. [CrossRef]
  43. D. C. Johnston, “Stretched exponential relaxation arising from a continuous sum of exponential decays,” Phys. Rev. B, vol. 74, no. 18, p. 184430, Nov. 2006. [CrossRef]
  44. Li, X., Han, E.T., Busse, R.F., Majumdar, S.: In vivo T1ρ mapping in cartilage using 3D magnetization-prepared angle-modulated partitioned k-space spoiled gradient echo snapshots (3D MAPSS). Magnetic Resonance in Medicine 59(2), 298-307 (2008).
  45. T. Steihaug, “The conjugate gradient method and trust regions in large scale optimization,” SIAM J. Numer. Anal., vol. 20, no. 3, pp. 626–637, Jun. 1983. [CrossRef]
  46. Gabrielle Blumenkrantz, B., Jesus Lozano, Carballido-Gamio Julio, M. Ries, and Sharmila Majumdar. "In vivo T1Rho and T2 mapping of articular cartilage in osteoarthritis of the knee using 3 tesla MRI." Osteoarthritis Cartilage 15, no. 7 (2007): 789-797.
  47. Milford, David, Nicolas Rosbach, Martin Bendszus, and Sabine Heiland. "Mono-exponential fitting in T2-relaxometry: relevance of offset and first echo." PloS one 10, no. 12 (2015): e0145255.
  48. Souza, Richard B., Brian T. Feeley, Zinta A. Zarins, Thomas M. Link, Xiaojuan Li, and Sharmila Majumdar. "T1rho MRI relaxation in knee OA subjects with varying sizes of cartilage lesions." The knee 20, no. 2 (2013): 113-119.
  49. Liu, F., Kijowski, R., Feng, L. and El Fakhri, G.: High-performance rapid MR parameter mapping using model-based deep adversarial learning. Magnetic resonance imaging, 74, 152-160 (2020).
  50. Feng, R., Zhao, J., Wang, H., Yang, B., Feng, J., Shi, Y., Zhang, M., Liu, C., Zhang, Y., Zhuang, J. and Wei, H.: MoDL-QSM: Model-based deep learning for quantitative susceptibility mapping. NeuroImage, 240, 118376 (2021).
  51. Jun, Y., Shin, H., Eo, T., Kim, T. and Hwang, D.: Deep model-based magnetic resonance parameter mapping network (DOPAMINE) for fast T1 mapping using variable flip angle method. Medical Image Analysis, 70, 102017 (2021).
  52. Bian, W., Jang, A. and Liu, F.: Improving quantitative MRI using self-supervised deep learning with model reinforcement: Demonstration for rapid T1 mapping. Magnetic Resonance in Medicine 92(1), 98-111 (2024).
Figure 1. Computational time for NLS fitting across different MRI slice counts on HPC (theoretical) and NVIDIA GPU systems.
Figure 1. Computational time for NLS fitting across different MRI slice counts on HPC (theoretical) and NVIDIA GPU systems.
Preprints 142530 g001
Figure 2. Representative visualization of multi-component T1ρ relaxation models in the knee joint.
Figure 2. Representative visualization of multi-component T1ρ relaxation models in the knee joint.
Preprints 142530 g002
Figure 3. Voxel-by-voxel framework of the DL model for fitting multi-component T1 ρ relaxation models with a customized loss function.
Figure 3. Voxel-by-voxel framework of the DL model for fitting multi-component T1 ρ relaxation models with a customized loss function.
Preprints 142530 g003
Figure 4. Sensitivity analysis of the DL model: (a) optimizer selection, (b) epochs, (c) batch size, and (d) number of layers, on RMSE.
Figure 4. Sensitivity analysis of the DL model: (a) optimizer selection, (b) epochs, (c) batch size, and (d) number of layers, on RMSE.
Preprints 142530 g004
Figure 5. Sensitivity analysis of the combination of activation functions (ReLU, Leaky ReLU, GELU, Tanh, and Swish) with optimizers (Adam, SGD, and RMSProp) for the DL model: (a) ME component, (b) SE component, and (c) BE component.
Figure 5. Sensitivity analysis of the combination of activation functions (ReLU, Leaky ReLU, GELU, Tanh, and Swish) with optimizers (Adam, SGD, and RMSProp) for the DL model: (a) ME component, (b) SE component, and (c) BE component.
Preprints 142530 g005
Figure 6. Sensitivity analysis of the loss function with varying γ s
Figure 6. Sensitivity analysis of the loss function with varying γ s
Preprints 142530 g006
Figure 7. Training and validation loss analysis: (a) For 10 epochs and (b) For 100 epochs.
Figure 7. Training and validation loss analysis: (a) For 10 epochs and (b) For 100 epochs.
Preprints 142530 g007
Figure 8. Proposed HDNLS-based fitting of multi-component T1ρ relaxation models in the knee joint.
Figure 8. Proposed HDNLS-based fitting of multi-component T1ρ relaxation models in the knee joint.
Preprints 142530 g008
Figure 11. Estimated HDNLS variants-based ME and SE-T1ρ maps for the full knee joint.
Figure 11. Estimated HDNLS variants-based ME and SE-T1ρ maps for the full knee joint.
Preprints 142530 g011
Figure 12. Estimated HDNLS variants-based ME and SE T1ρ maps for the knee cartilage ROIs only.
Figure 12. Estimated HDNLS variants-based ME and SE T1ρ maps for the knee cartilage ROIs only.
Preprints 142530 g012
Figure 13. Estimated HDNLS variants-based BE long T1ρ, the fraction of the long T1ρ, and short T1ρ maps for the full knee joint.
Figure 13. Estimated HDNLS variants-based BE long T1ρ, the fraction of the long T1ρ, and short T1ρ maps for the full knee joint.
Preprints 142530 g013
Figure 14. Estimated HDNLS variants-based BE long T1ρ, the fraction of the long T1ρ, and short T1ρ maps for knee cartilage ROIs only.
Figure 14. Estimated HDNLS variants-based BE long T1ρ, the fraction of the long T1ρ, and short T1ρ maps for knee cartilage ROIs only.
Preprints 142530 g014
Figure 15. Estimated NLS, RNLS, DL, and HDNLS-based ME and SE T1ρ maps for the full knee joint.
Figure 15. Estimated NLS, RNLS, DL, and HDNLS-based ME and SE T1ρ maps for the full knee joint.
Preprints 142530 g015
Figure 16. Estimated NLS, RNLS, DL, and HDNLS-based ME and SE T1ρ maps for the knee cartilage ROIs only.
Figure 16. Estimated NLS, RNLS, DL, and HDNLS-based ME and SE T1ρ maps for the knee cartilage ROIs only.
Preprints 142530 g016
Figure 17. Estimated NLS, RNLS, DL, and HDNLS-based BE long T1ρ component, the fraction of the long T1ρ component, and short T1ρ maps for the full knee joint.
Figure 17. Estimated NLS, RNLS, DL, and HDNLS-based BE long T1ρ component, the fraction of the long T1ρ component, and short T1ρ maps for the full knee joint.
Preprints 142530 g017
Figure 18. Estimated NLS, RNLS, DL, and HDNLS-based BE long T1ρ component, the fraction of the long T1ρ component, and short T1ρ maps for knee cartilage ROIs only.
Figure 18. Estimated NLS, RNLS, DL, and HDNLS-based BE long T1ρ component, the fraction of the long T1ρ component, and short T1ρ maps for knee cartilage ROIs only.
Preprints 142530 g018
Table 1. Comparative analysis of the existing AI-based exponential fitting methods.
Table 1. Comparative analysis of the existing AI-based exponential fitting methods.
Ref. Year Method Features Open Challenges
[21] 2017 Monte Carlo simulations and a random forest regressor -Established a robust mapping from diffusion-weighted MR signal features to the underlying microstructure parameters of brain tissue.
-Estimated water residence time in brain white matter.
-Did not estimate multi-component T1ρ map.
-Did not provide uncertainty estimates for predictions.
-Used a very basic representation of white matter.
[22] 2019 Deep Neural Network (DNN) - Employed non-supervised DNNs to fit the Intravoxel Incoherent Motion (IVIM).
-Does not rely on priors
-Parameter maps were smooth and consistent within homogeneous tissues.
-DNN may require retraining for different acquisition protocols.
-Training and testing on same data.
-Did not improve Bayesian convergence in low SNR regions.
[24] 2020 DNN -Used a DNN to directly estimate myelin water fraction (MWF).
-Can produce whole-brain MWF maps in approximately 33 seconds.
-Better voxel-wise agreement was observed between DNN-based MWF estimates and NNLS-based ground truths.
-DNN relied on NNLS-generated ground truths, so it inherited the NNLS method’s vulnerability to noisy data
-DNN are sensitive to noise and artifacts in the input data
[33] 2019 MANTIS - Integrated a model-augmented neural network with incoherent k-space sampling for efficient MR parameter mapping.
-Combined physics-based modeling with DL for accelerated image reconstruction.
- Reduced noise and artifacts in reconstructed parameter maps
-Generalization to diverse datasets may need further validation due to dependency on sampling patterns.
-Performance depends on the quality of its training data, which could lead to overfitting and limited generalizability.
[49] 2020 Model-based deep adversarial learning -Employed a generative adversarial network (GAN) for enhanced reconstruction quality.
-Achieved rapid parameter mapping with high fidelity and robustness.
-Effectively removed noise and artifacts.
-Adversarial training is computationally demanding and sensitive to hyperparameter tuning.
-GANs may generate outputs that appear realistic but deviate from the true underlying signal.
[50] 2021 MoDL-QSM -Combined model-based reconstruction with DL for quantitative susceptibility mapping (QSM).
-Used iterative model unrolling with DL modules to improve QSM accuracy.
-Produced high-quality susceptibility maps with reduced streaking artifacts.
-Model unrolling introduces computational overhead.
-Requires careful hyperparameter optimization to ensure convergence.
[51] 2021 DOPAMINE integrated model-based and deep learning techniques for fast T1 mapping
Used variable flip angle (VFA) method for efficient data acquisition
Reduced dependency on large training datasets by incorporating physics-based modeling.
Handled noise and artifacts effectively
-Dependent on specific acquisition protocols (VFA), limiting flexibility for other methods.
-Computationally intensive due to the integration of model-based techniques.
[37] 2021 RELAX -Utilized two physical models, such as the MR imaging model and the quantitative imaging model, in network training to fit the parameters.
-Eliminated the need for fully sampled reference datasets
-Directly reconstructs MR parameter maps from undersampled k-space data.
-Tested only on simulated and coil-combined data
-Only a residual U-Net was used
-RELAX's performance relies on the precision of its physical models, limiting its robustness to inaccuracies.
[35] 2023 SuperMAP -Achieved mapping with acceleration factors up to R = 32.
-Generate relaxation maps from a single scan using optimized data acquisition.
-Directly convert undersampled parameter-weighted images into quantitative maps.
-Performed well on both retrospectively undersampled and prospective data, demonstrating adaptability.
-Performance relies on the quality and diversity of retrospectively undersampled training data.
-Heavily reliant on the quality of the undersampling patterns, suboptimal patterns may degrade performance.
[52] 2024 RELAX-MORE -Used self-supervised learning, eliminating the need for large, labeled datasets.
-Unrolled an iterative model-based qMRI reconstruction into a DL framework.
-Produced accurate parameter maps that correct image artifacts, remove noise, and recover features even in regions with imperfect imaging conditions.
-The unrolling of iterative processes and subject-specific training could still be computationally intensive.
-Subject-specific training might require more careful convergence monitoring, as the self-supervised framework relies on single-subject data.
Table 2. Multi-component relaxation models: Mono-Exponential (ME), Bi-Exponential (BE), and Stretched-Exponential (SE).
Table 2. Multi-component relaxation models: Mono-Exponential (ME), Bi-Exponential (BE), and Stretched-Exponential (SE).
Model Equation Parameters   θ Description Range
ME[20,46,47,48] f t , θ = A   exp t T M E A , T M E Assumes tissue is homogeneous with a single relaxation time. A is the complex-valued amplitude and T M E is the real-valued relaxation time. T M E > 0   m s (for knee cartilage ≥ 5 ms)
BE [38,39,40] f t , θ = A ( 1 f ) exp t T s + f exp t T l A , f , T s , T l Accounts for two compartments with distinct relaxation times T s (short) and T l (long). f is the fraction of the long compartment. 0 f 1 T s < 5   m s T l 5   m s
SE [41,42,43] f t , θ = A exp t T S E β A , T S E , β Models heterogeneous environments with distributed relaxation rates. T S E is the relaxation time, and β is the stretching exponent. 0 < β 1 T S E > 0   m s
Table 3. Features, advantages, and disadvantages of different relaxation models (ME, BE and SE).
Table 3. Features, advantages, and disadvantages of different relaxation models (ME, BE and SE).
Model Features Advantages Disadvantages
ME • Fundamental model for single-component relaxation. • Simple and easy to implement.
• Suitable for homogeneous tissues with a single relaxation component.
• Limited in handling tissues with multiple relaxation components.
BE • Extends the ME model by introducing two compartments (short and long). • Represents two-component decay systems.
• Accounts for two non-exchanging compartments in tissues, each with distinct short and long relaxation times.
• Better representation of tissues with two relaxation components, e.g., free water and water bonded to macromolecules. • Requires more parameters and is computationally expensive.
• May face convergence issues during fitting.
• The short component is usually sensitive to noise.
SE • Reduces to ME when β=1. • Can be used as an alternative to BE for describing tissue heterogeneity.
• Can approximate a broad distribution of relaxation rates with fewer parameters, simplifying complex tissue modeling.
• Suitable for heterogeneous tissues.
• More stable and requires fewer parameters than BE.
• Can describe multi-component relaxation.
• Less straightforward physical interpretation compared to BE.
Table 4. Mathematical Formulations of Performance Metrics Used for Comparative Analysis.
Table 4. Mathematical Formulations of Performance Metrics Used for Comparative Analysis.
Metric Description Equation Symbols
MNAD Measures the median of the normalized absolute difference between estimated and reference T1ρ parameters across voxels. M N A D ( Φ ) = median i I , p P θ ¯ i , p θ i , p θ ¯ i , p θ i , p and θ ¯ i , p are the estimated and reference parameters p of the voxel indexed by i that belongs to the region of interest or set of voxels I, and Φ = θ i i I
NRMSR Assesses the size of the residual between MR measurements and predicted MR signals. N R M S R ( F ) = m e a n i I s i f ^ i 2 s i 2 s i is the measured MR signal at voxel indexed by i, and f ^ i isThe predicted MR signal at the same voxel with θ i , and F = f ^ i i I
NRMSE Assesses the normalized value of the root mean squared error (RMSE) across the reference parameters. N R M S E Φ = m e a n i I , p P θ ¯ i , p θ i , p 2 m e a n i I , p P θ ¯ i , p 2 # θ i , p and θ ¯ i , p are the estimated and reference parameters p of the voxel indexed by i that belongs to the region of interest or set of voxels I, and Φ = θ i i I .
Fitting Time Represents the time required by models (e.g., NLS, RNLS, HDNLS, DL) to compute the T1ρ map for one MRI slice, measured on the same hardware. F i t t i n g   T i m e = t o c t i c t i c : Start time of the fitting process. t o c : End time of the fitting process
Table 5. Quantitative analysis of HDNLS variants for ME, SE, and BE T1ρ maps on real data for full knee joint. Bold quantitative values indicate the outperforming parameter for the specific model.
Table 5. Quantitative analysis of HDNLS variants for ME, SE, and BE T1ρ maps on real data for full knee joint. Bold quantitative values indicate the outperforming parameter for the specific model.
Method Metric Ultra-HDNLS Super-HDNLS HDNLS Relaxed-HDNLS
ME MNAD (%) 4.68 4.22 4.03 3.82
Fitting time (sec) 0.99 1.76 5.13 10.96
NRMSR 0.03 0.03 0.02 0.02
SE MNAD (%) 13.49 8.92 7.18 6.75
Fitting time (sec) 1.12 1.85 3.42 7.12
NRMSR 1.09 0.70 0.18 0.15
BE MNAD (%) 21.38 20.26 19.87 19.24
Fitting time (sec) 1.22 2.14 3.96 7.73
NRMSR 0.03 0.03 0.02 0.02
Table 6. Quantitative analysis of HDNLS variants for ME, SE, and BE T1ρ maps on real data for knee cartilage ROI only. Bold quantitative values indicate the outperforming parameter for the specific model.
Table 6. Quantitative analysis of HDNLS variants for ME, SE, and BE T1ρ maps on real data for knee cartilage ROI only. Bold quantitative values indicate the outperforming parameter for the specific model.
Method Metric Ultra-HDNLS Super-HDNLS HDNLS Relaxed-HDNLS
ME MNAD (%) 4.36 4.16 3.86 3.74
Fitting time (sec) 0.52 0.95 2.38 5.73
NRMSR 0.03 0.02 0.02 0.02
SE MNAD (%) 13.24 8.73 7.12 6.67
Fitting time (sec) 0.51 0.87 2.28 3.71
NRMSR 1.09 0.70 0.18 0.14
BE MNAD (%) 20.34 19.77 19.28 18.94
Fitting time (sec) 0.46 0.87 1.85 3.26
NRMSR 0.03 0.02 0.02 0.02
Table 8. Quantitative analysis of NLS, RNLS, DL, and HDNLS for ME, SE, and BE T1ρ maps on real data for both full knee joint and knee cartilage ROI only.
Table 8. Quantitative analysis of NLS, RNLS, DL, and HDNLS for ME, SE, and BE T1ρ maps on real data for both full knee joint and knee cartilage ROI only.
Method Metric Full Knee Knee Cartilage
DL HDNLS DL HDNLS
ME MNAD (%) 22.36 18.93 19.32 17.20
NRMSE (%) 24.05 21.25 21.20 19.31
SE MNAD (%) 26.38 23.94 23.38 21.64
NRMSE (%) 25.81 22.94 22.46 19.85
BE MNAD (%) 17.70 13.02 14.09 12.24
NRMSE (%) 20.16 16.36 18.37 14.85
Table 9. Computational speed analysis of NLS, RNLS, DL, and HDNLS for ME, SE, and BE T1ρ maps on real data for both full knee joint and knee cartilage only.
Table 9. Computational speed analysis of NLS, RNLS, DL, and HDNLS for ME, SE, and BE T1ρ maps on real data for both full knee joint and knee cartilage only.
Method Metric NLS RNLS DL HDNLS
ME Full Knee 274.68 234.22 1.35 19.32
Knee Cartilage 76.37 65.82 0.92 6.78
SE Full Knee 293.49 262.92 1.78 22.50
Knee Cartilage 91.12 88.39 1.67 7.51
BE Full Knee 257.38 229.13 1.21 18.14
Knee Cartilage 70.26 67.14 0.84 6.13
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated