1.Introduction1.1.Background and MotivationComputed tomography (CT) imaging has become an important tool in modern healthcare, playing a critical role in quantitative diagnosis, evidence-based treatment planning, and patient-centric care across a wide spectrum of disease pathways.1 The quality and speed of CT image reconstruction directly influence the accuracy of diagnosis, and the effectiveness of treatment plans, hence the overall efficiency of clinical workflows.2 Traditional reconstruction methods such as filtered backprojection (FBP) and algebraic reconstruction technique (ART) often face challenges when noise or disturbance is present in the data. These approaches struggle to effectively balance image quality and computational efficiency, particularly in scenarios where such noise, artifacts, or disturbances exist. In recent years, there has been a marked increase in the application of deep learning techniques to enhance various aspects of CT reconstruction.3–5 These approaches have shown promising results in addressing longstanding issues such as noise reduction and artifact removal.6–10 The potential of deep learning in CT reconstruction lies in its ability to learn complex, nonlinear mappings between input data and high-quality reconstructions, often outperforming traditional methods.11,12 1.2.Loss Functions in Deep Learning-Based CT ReconstructionAt the core of deep learning approaches lies the optimization process, which relies on minimizing a specific objective function, known as the loss function, through backpropagation. The design of this loss function is critical in determining the quality and characteristics of the reconstructed CT images. It serves as the primary guide for the neural network during training, influencing what features are emphasized or suppressed in the final reconstruction. The most commonly used loss function in CT image reconstruction is the pixel-wise mean squared error (MSE). Its popularity stems from its computational simplicity and its alignment with Gaussian noise models, which are often assumed in CT imaging.13 However, MSE has significant limitations when it comes to capturing the nuances of human visual perception. Research has shown that MSE may not accurately reflect human-perceived image quality,14 as it fails to account for the varying perception of noise based on image content, such as luminance and contrast.15 This discrepancy between MSE-based optimization and human visual assessment can lead to reconstructions that, while numerically accurate, may not be optimal for clinical interpretation. For instance, MSE tends to overly smooth images, potentially obscuring fine details that could be crucial for accurate diagnosis. Moreover, it does not adequately penalize structural distortions that might be visually significant but contribute little to the overall pixel-wise error.16 1.3.Alternative Loss Functions and Their LimitationsRecognizing the limitations of MSE, researchers have proposed various alternative loss functions.17–26 These can be broadly categorized into three main types, each with its own strengths and limitations:
Stylistic deviations and mitigation: Perceptual loss can bias reconstructed images towards the style of its training data,28 which is particularly problematic for CT images where such deviations can obscure important diagnostic details. To address this issue, transfer learning techniques can be employed. Pre-training the perceptual loss model on CT data can mitigate style biases,29 helping to align the perceptual model’s feature space with the characteristics of CT images.30 Computational considerations: The use of perception-driven loss functions introduces significant computational challenges. The pre-trained network needs to process the images and calculate the loss for each iteration during training and validation, substantially increasing the time and resources required compared with simpler loss functions such as MSE. In addition, the complex nature of these networks reduces interpretability, which is crucial in medical applications. These limitations underscore the need for more specialized and efficient loss functions that can capture the perceptual quality of CT images without introducing style biases or excessive computational burden. Future research could focus on developing lightweight perceptual models specifically tailored for CT imaging or exploring hybrid approaches that balance perceptual quality with computational efficiency.31
Each of these approaches offers certain advantages over MSE, but none fully addresses all the challenges inherent in CT image reconstruction.35,36 The ideal loss function would need to balance perceptual quality, structural accuracy, and computational efficiency while also being adaptable to the specific characteristics of CT images. 1.4.ContributionsOur work makes several contributions to the field of CT image reconstruction:
This wide-ranging evaluation highlights Eagle-Loss’s adaptability across diverse CT imaging paradigms and reconstruction challenges.
By introducing Eagle-Loss and demonstrating its effectiveness across multiple scenarios, our work paves the way for improved CT image reconstruction methods that can potentially enhance diagnostic accuracy and patient care. 1.5.Paper StructureThe paper is structured as follows: Sec. 2 provides an in-depth discussion of the motivation behind Eagle-Loss as well as its mathematical foundations. Section 3 outlines the experimental setup and evaluation framework. Section 4 presents and analyzes the results of our experiments across various CT reconstruction scenarios and performs a comprehensive ablation study on the key components and hyperparameters of Eagle-Loss. Finally, Sec. 5 summarizes the main findings and contributions of our work, discussing implications and future directions. 2.MethodologyTo address the limitations of existing loss functions and the specific challenges of CT image reconstruction, we introduce a loss function termed “Eagle-Loss.” This approach is inspired by the observation that image blurring typically reduces the variance in gradient map patches.18 Eagle-Loss leverages this insight to develop a more effective method for assessing image quality during the reconstruction process. The key innovations of Eagle-Loss are
To the best of our knowledge, this is the first application of frequency analysis to localized features within gradient-derived variance maps for CT image reconstruction. By combining spatial localization with frequency domain analysis, Eagle-Loss aims to capture both fine-grained texture information and larger-scale structural features, potentially overcoming the limitations of existing methods. Our method builds upon and significantly extends the work of Abrahamyan et al.,18 who pioneered the use of gradient variance for image generation tasks. Figure 1 provides a comprehensive overview of the Eagle-Loss process, illustrating the steps involved in computing the magnitude spectrum from an input image. This approach is motivated by a key observation: blurring in reconstructed images typically manifests as reduced variance in the patches of their gradient maps.18 By focusing our analysis on the high-frequency components of these variance maps, we can more effectively capture and preserve fine details in reconstructed images, maintain sharp edges and intricate textures, and enhance overall image quality and fidelity. Fig. 1Illustration of steps involved in computing the magnitude spectrum and from an image . In this figure, larger patches are used for better visualization. ![]() Before explaining the specifics of Eagle-Loss, it is important to understand two key concepts: gradient maps and variance maps. A gradient map represents the directional change in intensity or color in an image. It highlights edges and textures, where pixel values change rapidly. Gradient maps are crucial for detecting features and understanding image structure, typically computed using convolution operations with specific kernels designed to detect changes in horizontal and vertical directions. A variance map quantifies the local contrast within an image. It is created by calculating the statistical variance of pixel values within small neighborhoods or patches of the image. High variance indicates areas with significant intensity changes (such as edges or textures), whereas low variance suggests more uniform regions. Variance maps can reveal subtle textures and structures that might not be immediately apparent in the original image. Eagle-Loss leverages these concepts through a four-step process:
and are the patches extracted from patchified gradient maps and along the -axis and -axis. We then calculate the variance: The variance for a patch is computed as where represents each element in the patch and is the mean value of the patch
The term calculates the radial distance from the origin in the frequency domain, effectively measuring how “high” the frequency is. The parameter is the cutoff frequency, which determines the threshold between low and high frequencies. It acts as a control for the filter’s behavior:
A larger value of allows more low frequencies to pass through, resulting in less aggressive filtering, while a smaller value of creates a more stringent high-pass filter that emphasizes finer details and edges more strongly. This Gaussian high-pass filter design ensures a smooth transition between attenuated and preserved frequencies, avoiding abrupt changes that could introduce artifacts in the filtered image.
This formulation of Eagle-Loss allows us to comprehensively assess the quality of reconstructed CT images by comparing both local and global features in the frequency domain, potentially leading to more accurate and visually pleasing reconstructions. 3.Experiments3.1.Synthetic, Animal, and Clinical DataOur experiments utilized three diverse datasets comprising synthetic, animal, and clinical data to evaluate the performance of Eagle-Loss across different CT reconstruction tasks comprehensively:
3.2.Experiment SetupOur experiment framework was implemented based on Python 3.10 and PyTorch 2.0. For optimization, we employed the Adam optimizer (, ), starting with an initial learning rate of . A dynamic learning rate was implemented using a OneCycle learning rate scheduler, varying between and . All models were trained over 100 epochs on an NVIDIA RTX A6000 GPU. A hybrid loss function, integrating MSE and our proposed Eagle-Loss [refer to Eq. (8)], was employed for training the models. This combination leverages the strengths of both components: MSE ensures global image fidelity, provides stability in training, and helps in noise reduction, whereas Eagle-Loss focuses on preserving high-frequency details and structural information critical in CT imaging. The loss function is formally defined as where denotes the weight coefficient for Eagle-Loss, empirically determined to balance the contributions of both components. This value ensures that Eagle-Loss significantly influences the optimization process without overshadowing the stabilizing effect of MSE. Based on our studies, we chose a patch size of [see Eq. (2)] and a cutoff frequency of [see Eq. (7)] for Eagle-Loss. The small patch size allows for detailed analysis of local structures, while the selected value in the high-pass filter increases the model’s sensitivity to high-frequency components. This selection balances computational efficiency and the ability to capture important image features in CT reconstruction.It is important to note that while these hyperparameters are not ideal for matrix calculations, they produced the sharpest images with well-defined edges and fine details, which aligns with our study’s goals. We provide a thorough analysis of how different hyperparameter choices affect the results in Sec. 4.4. Combining with MSE is important not only for Eagle-Loss but also for our competitive method. For instance, total variation (TV)44 is a regularizer that cannot be used alone for training, and in our experiment, Gaussian edge-enhanced (GEE) also needs to be combined with MSE for stable training or it will induce a vanishing gradient problem. For the perceptual-driven approaches, because the backbone model is not trained on CT data, MSE serves as a strong constraint for optimizing the model with minimal style artifacts. Figure 2 presents the outcomes from TF-FBP models trained using individual loss functions. The results show that both Eagle-Loss and learned perceptual image patch similarity (LPIPS)45 loss introduce some artifacts in the reconstructed images. Fig. 2Comparison of results from TF-FBP models trained with different loss functions without combining with MSE for low-dose CT reconstruction. ![]() To ensure fair comparisons, all comparison methods use hybrid loss functions. We establish a systematic rule for selecting appropriate values across different loss functions to maintain fairness and optimize results. The rule works as follows: we evaluate both loss functions during the first training epoch, where always takes the form . We adjust to ensure that the MSE and the weighted loss components (such as TV, perceptual loss, and others) are of the same order of magnitude. For example, if and , we would set to obtain , resulting in , which matches the order of magnitude of . For each dataset, we employed different models:
Fig. 3Illustration of models used for low-dose CT reconstruction. (a) The structure of TF-FBP. Here, represents the trainable filter and represents the differentiable backprojection operator, whereas and denote the inverse and forward DFT. The differentiable backprojection operator was implemented using PYRO-NN.47 (b) RED-CNN with its encoder-decoder structure. The input of the encoder is the reconstructed image using FBP. ![]() 3.3.Evaluation MetricsWe used three primary metrics to evaluate the performance of various loss functions:
These metrics collectively provide a comprehensive assessment of reconstruction quality, structural preservation, and computational efficiency across our diverse experimental scenarios. 4.Results and DiscussionsIn this section, we present the results of comparing our proposed Eagle-Loss with other loss functions: GEE,33 TV,44 gradient variance (GV),18 SSIM, perceptual loss,19 and LPIPS.45 GEE compares the high-frequency difference of images in the frequency domain using a high-pass filter. TV penalizes gradients to enhance smoothness, whereas GV compares the variance between image gradients. SSIM utilizes the structural similarity index as a loss function. The perceptual loss compares the MSE between different image features using an ImageNet52 pre-trained VGG-1953 model. Finally, LPIPS is a perceptual image similarity metric that leverages deep features extracted from ImageNet52 pre-trained AlexNet54 and employs learned weights to compare images. 4.1.Low-Dose CT ReconstructionEagle-Loss demonstrates superior performance in low-dose CT reconstruction, as evidenced by both qualitative and quantitative evaluations. Figures 4 and 5 provide a visual comparison of two reconstruction samples using different loss functions. Fig. 4Comparison of low-dose CT reconstruction results using different loss functions on the LoDoPaB-CT dataset, implemented with RED-CNN. Although the differences between methods are subtle, they are still discernible: (1) MSE, MSE+TV, and MSE + SSIM show a slight tendency to oversmooth, particularly in bone structures and soft tissue boundaries. (2) The proposed Eagle-Loss (rightmost column) achieves a marginally better balance between detail preservation and noise suppression. Note the subtle improvements in edge definition and texture preservation in the Eagle-Loss results compared with other methods, though the distinctions are less pronounced than in the TF-FBP implementation. ![]() Fig. 5Comparison of low-dose CT reconstruction results using different loss functions on the LoDoPaB-CT dataset, implemented with TF-FBP. The differences between methods are more pronounced in this case: (1) MSE, MSE + TV, and MSE + SSIM clearly oversmooth the image, resulting in significant loss of fine details, especially in bone structures and soft tissue boundaries. (2) The proposed Eagle-Loss (rightmost column) demonstrates a superior balance between detail preservation and noise suppression, maintaining sharpness in bone structures while avoiding oversmoothing in soft tissue areas. Note the markedly improved edge definition and texture preservation in the Eagle-Loss results. Below each zoomed-in image, its corresponding difference map is displayed, further highlighting the performance disparities between the methods. ![]() The images reveal that MSE + Eagle preserves fine details and structural integrity more effectively than other methods, which tend to oversmooth the images. This is particularly noticeable in bone structures and soft tissue boundaries, where Eagle-Loss maintains sharpness while avoiding oversmoothing. The RED-CNN model generally produces better images compared to TF-FBP, regardless of the loss function used. The qualitative observations are corroborated by quantitative metrics presented in Table 1. Table 1Performance comparison of loss functions for low-dose CT reconstruction, evaluated on 200 samples. Metrics include SSIM, PSNR, and training speed for TF-FBP and RED-CNN models. Paired t tests are calculated55 to compare each method with MSE + Eagle.
Note: Numbers shown in bold represent the best performance values within each matrix, making it easier for readers to identify the top results. In terms of SSIM, the combination of MSE and Eagle-Loss outperforms other configurations, achieving the highest values for both TF-FBP (0.958) and RED-CNN (0.972) models. This represents a 0.24% and 1.66% improvement over the baseline MSE, respectively, indicating superior structural preservation. Regarding PSNR, MSE alone yields the highest value for TF-FBP (34.860 dB), whereas MSE + TV achieves the best for RED-CNN (37.296 dB). Interestingly, MSE + Eagle shows lower PSNR values (33.621 dB for TF-FBP and 36.233 dB for RED-CNN). This is attributed to its focus on preserving high-frequency details rather than oversmoothing the image, a trade-off often observed in image reconstruction tasks where methods that preserve more details may have lower PSNR due to the retention of some noise. In terms of computational efficiency, the baseline MSE loss function exhibits the fastest training times for both models (8210.00 s for TF-FBP and 8306.20 s for RED-CNN). The MSE + Eagle combination is about 7.3% slower for TF-FBP and 6.4% slower for RED-CNN compared with MSE alone. However, it remains efficient compared with more complex losses such as MSE + perceptual, which is slower for TF-FBP and 16.1% slower for RED-CNN. During the last training epoch, MSE accounted for 41.86% and 50.26% of our proposed hybrid loss in the two models, respectively. These percentages indicate that both MSE and Eagle-Loss make significant contributions to the overall loss function and final reconstructed image. The effectiveness of Eagle-Loss extends beyond neural network-based approaches. Figure 6 illustrates the integration of Eagle-Loss as a regularization term into the ART algorithm, comparing it with TV regularization and no regularization. Fig. 6Comparison of regularization techniques in ART reconstruction using two samples from the LoDoPaB-CT dataset. Perception-based regularizers (perceptual and LPIPS) introduce strong grid-like artifacts, whereas TV oversmooths the image. SSIM exhibits artifacts in high-contrast regions, and GV over-enhances edge contrast. By contrast, our proposed Eagle Loss achieves the best visual quality, successfully preserving fine details while avoiding noticeable artifacts, thus striking an optimal balance between noise reduction and detail preservation. GEE is not included in this comparison because of its vanishing gradient without a combination of MSE. ![]() The results demonstrate that Eagle-Loss significantly improves the sharpness and fidelity of CT reconstruction compared with no regularization and other regularization methods. Although TV regularization performs better than no regularization, it tends to oversmooth some details that Eagle-Loss successfully preserves. Other methods such as perception-based regularizers (perceptual and LPIPS) introduce strong grid-like artifacts, SSIM exhibits issues in high-contrast regions, and GV tends to over-enhance edge contrast. These findings indicate that Eagle-Loss is a promising tool for low-dose CT reconstruction, offering enhanced image quality, particularly in terms of structural preservation and detail retention. Although there is a modest increase in computational cost, the improved image quality justifies this trade-off in many clinical scenarios where diagnostic accuracy is paramount. Eagle-Loss strikes an optimal balance between noise reduction and detail preservation, addressing the limitations observed in other regularization techniques and potentially enhancing the diagnostic value of low-dose CT images. 4.2.CT FOV ExtensionTable 2 summarizes the performance metrics for various loss functions. The combination of MSE and Eagle-Loss (MSE + Eagle) demonstrates superior performance across key metrics. In terms of structural similarity, it achieves the highest SSIM of 0.966, a 1.09% improvement over baseline MSE. For image quality, it attains the best PSNR of 30.964 dB, surpassing MSE by 1.790 dB. Regarding computational efficiency, with a training time of 37602.32 s, it is only 1.15% slower than the fastest method (MSE + GEE) and 4.32% faster than MSE alone. These metrics indicate that Eagle-Loss enhances structural preservation and overall image quality without significantly increasing computational cost. During the last training epoch, MSE accounted for 48.13% of our proposed hybrid loss. This percentage indicates that both MSE and Eagle-Loss make important contributions to the overall loss function and final reconstructed image. Table 2Performance comparison of loss functions for CT FOV Extension, evaluated on 52 samples. Metrics include SSIM, PSNR, and training speed for U-Net. Paired t tests55 are calculated to compare each method with MSE + Eagle.
Note: Numbers shown in bold represent the best performance values within each matrix, making it easier for readers to identify the top results. Figure 7 provides visual comparisons of FOV extension results. The images demonstrate that MSE + Eagle produces clearer structural boundaries compared with other methods. The transition between original and extended FOV appears more seamless with Eagle-Loss, and fine anatomical details are better preserved. By contrast, other methods, especially MSE alone, tend to generate blurrier extended regions. Fig. 7Visualization of CT FOV extension results on the SMIR dataset using the U-Net model. The blue-dotted circle denotes the original FOV boundary. The input images show significant truncation artifacts outside the original FOV. Note the improved detail preservation and reduced artifacts in our method, particularly visible in the extended regions of the head and chest. The red arrows in our result highlight areas of notable improvement in structural definition and edge preservation compared with other methods. Below each zoomed-in picture, its corresponding difference map is displayed. ![]() Although Eagle-Loss shows overall superior performance, it exhibits some limitations in addressing streaky artifacts. This is possibly due to the high-pass filter interpreting these artifacts as low-frequency features. This observation points to potential areas for future improvements, such as incorporating additional terms specifically designed to mitigate such artifacts. The ability of Eagle-Loss to maintain clear structural boundaries and seamlessly integrate the extended region is crucial for accurate interpretation and diagnosis in clinical settings. Its improved performance in both quantitative metrics and qualitative assessments suggests that Eagle-Loss could be a valuable tool for CT FOV extension tasks in practical applications. In conclusion, Eagle-Loss demonstrates a compelling balance of improved image quality, structural preservation, and computational efficiency in CT FOV extension. Although there is room for further refinement, particularly in artifact reduction, the current results indicate its strong potential for enhancing CT imaging capabilities in clinical practice. 4.3.PCCT Super-ResolutionTable 3 presents the quantitative results for the PCCT super-resolution task. In terms of SSIM, MSE + Eagle achieves the highest value (0.998), showing a 0.17% improvement over the baseline MSE. For PSNR, MSE + Eagle again achieves the highest value (45.105 dB), a substantial 3.217 dB improvement over MSE alone. These results demonstrate the exceptional performance of Eagle-Loss in preserving structural details and enhancing overall image quality in the super-resolution task and during the last training epoch, MSE accounted for 40.57% of our proposed hybrid loss. This percentage indicates that both MSE and Eagle-Loss make important contributions to the overall loss function and final reconstructed image. Table 3Performance comparison of loss functions for PCCT super-resolution, evaluated on 135 samples. Metrics include SSIM, PSNR, and training speed for NinaSR. Paired t tests55 are calculated to compare each method with MSE + Eagle.
Note: Numbers shown in bold represent the best performance values within each matrix, making it easier for readers to identify the top results. Regarding computational efficiency, MSE + GV has the fastest training time (5558.84 s), with MSE + Eagle being highly competitive (5595.89 s), only 0.67% slower than MSE + GV, and 0.50% slower than MSE alone. This demonstrates that the superior image quality achieved by Eagle-Loss comes with minimal additional computational cost. Figure 8 provides a visual comparison of the PCCT super-resolution results using different loss functions. The images demonstrate that MSE + Eagle produces results with clearer and more defined structural boundaries. Fine details, particularly in areas of complex tissue structure, are better preserved with MSE + Eagle. By contrast, other loss functions, including MSE alone, tend to produce slightly blurrier results with less distinct edges. Fig. 8Visualization of two PCCT super-resolution samples on the private PCCT dataset using the NinaSR model.49 This figure compares the effectiveness of different loss functions on NinaSR in achieving 8× super-resolution. Note the improved detail preservation in our method, particularly visible in the fine structures of the small bones and surrounding soft tissues. Red arrows in our result highlight areas of notable improvement in edge definition and structural clarity compared with other methods. Below each zoomed-in picture, its corresponding difference map is displayed. ![]() The exceptional performance of Eagle-Loss in the PCCT super-resolution task is particularly noteworthy. The substantial improvement in both SSIM and PSNR, coupled with the visual enhancement of fine details and structural boundaries, suggests that Eagle-Loss could be a powerful tool for improving the resolution and quality of PCCT images. This could have significant implications for medical imaging, potentially enabling more accurate diagnoses and better visualization of subtle anatomical features. The consistent superior performance of Eagle-Loss across all three tasks: low-dose CT reconstruction, CT FOV extension, and PCCT super-resolution, highlights its versatility and effectiveness in various medical imaging applications. Its ability to preserve structural integrity and fine details while maintaining computational efficiency makes it a promising tool for improving the quality of medical images across different modalities and tasks. 4.4.Ablation StudyAn ablation study was conducted to analyze the effect of varying cutoff frequencies, denoted as , within the Gaussian high-pass filter [refer to Eq. (7)] on the quality of reconstructed images. The TF-FBP model was employed for this analysis due to its compact parameter space, where only 255 parameters define the outcome. In TF-FBP, the quality of the reconstructed images is directly linked to the characteristics of the learned filter. Figure 9 illustrates the resultant filters and corresponding reconstructed images for different values of . A notable observation from this study is that higher values of enhance the sharpness of the reconstructed image while retaining more of the high-frequency noise. This demonstrates the direct relationship between the cutoff frequency and the enhancement of high-frequency components during image reconstruction, providing insights into the mechanism by which Eagle-Loss achieves its superior performance in preserving fine structural details. Fig. 9Illustration of filters, line profiles, and reconstructed images for different values in the TF-FBP model. The red lines on the CT images indicate the locations of the line profiles. This figure demonstrates how the cutoff frequency directly affects the enhancement of high-frequency details during image reconstruction. ![]() We also investigated the impact of the weighting coefficient in our hybrid loss function [Eq. (9)] using the TF-FBP model with a fixed patch size of . Figure 10 shows visual results, whereas Table 4 provides quantitative analysis. Fig. 10Illustration of reconstructed images for different values in the TF-FBP model with a fixed patch size of [refer to Eq. (2)]. ![]() Table 4Performance of different λ values in SSIM, PSNR, and LPIPS. Here, the model is TFFBP with a fixed patch size of n=3 [refer to Eq. (2)].
Note: Numbers shown in bold represent the best performance values within each matrix, making it easier for readers to identify the top results. Figure 10 demonstrates improved edge definition and image sharpness as increases, with showing the sharpest image. Our analysis reveals that SSIM is the highest at , but varies minimally across all values. The discrepancy between optimal for SSIM versus PSNR and LPIPS highlights the complexity of image quality assessment. In our investigation of Eagle-Loss, our primary objective is to enhance image sharpness rather than merely optimizing evaluation metrics such as SSIM or PSNR. To explore the influence of patch size on Eagle-Loss, we conducted experiments using the TF-FBP model, fixing at [see Eq. (9)] while varying from 3 to 15. Figure 11 illustrates that as decreases, there are notable improvements in edge definition and fine detail preservation. This visual enhancement aligns with our goal of improving image sharpness. Although Table 5 presents quantitative results showing that traditional metrics like SSIM and PSNR improve with larger patch sizes, these metrics do not fully capture the enhancements in image sharpness achieved with smaller patches. Our comprehensive ablation studies on key parameters—weighting parameter , cutoff frequency , and patch size —revealed a consistent trend: image sharpness increased with larger values, higher cutoff frequencies, and smaller patch sizes. Specifically, larger values amplified the influence of Eagle-Loss, preserving more high-frequency details, higher cutoff frequencies allowed more high-frequency components to pass through the Gaussian high-pass filter, and smaller patch sizes enabled more localized feature analysis, effectively capturing fine structures. Although these parameter choices maximized sharpness, they did not always optimize traditional metrics, highlighting a trade-off between image clarity and metric-based quality. These findings underscore Eagle-Loss’s flexibility and its potential to significantly enhance CT image reconstruction, particularly in scenarios that prioritize image sharpness and detail preservation over conventional quantitative metrics. They also emphasize the importance of careful parameter tuning based on specific application requirements, paving the way for future research into adaptive parameter selection methods and variant forms of Eagle-Loss tailored to different imaging modalities or reconstruction challenges. Fig. 11Illustration of reconstructed images for different values in the TF-FBP model with a fixed [refer to Eq. (9)]. ![]() Table 5Performance of different patch sizes n in SSIM, PSNR, and LPIPS. Here, the model is TF-FBP with a fixed λ=8×10−4 [refer to Eq. (9)].
Note: Numbers shown in bold represent the best performance values within each matrix, making it easier for readers to identify the top results. 5.ConclusionIn this paper, we propose Eagle-Loss, a new loss function specifically designed to improve the quality of reconstructed images by emphasizing high-frequency details crucial for accurate edge and texture representation. Eagle-Loss leverages localized feature analysis within gradient maps and applies frequency analysis to capture and preserve essential image features. Our experimental results demonstrate that Eagle-Loss effectively reduces image blur and enhances edge sharpness, leading to reconstructed images that exhibit superior fidelity to the ground truth. In addition, Eagle-Loss maintains computational efficiency in both speed and quality across all tasks. However, Eagle-Loss is sensitive to hyperparameter settings, and the optimal selection of these parameters varies across reconstruction tasks. Future efforts will focus on devising an automated hyperparameter optimization strategy to enhance the versatility and efficacy of Eagle-Loss in diverse imaging contexts. Our findings suggest that Eagle-Loss holds significant promise for CT image reconstruction and has the potential for broader applications in other fields. DisclosuresThe authors declare no conflicts of interest related to this work. No financial, personal, or professional affiliations influenced the research presented in this paper. All funding sources supporting this study are acknowledged within the paper. Any affiliations, financial involvement, or collaborations that could potentially influence the interpretation or outcome of this research have been disclosed. The development of Eagle-Loss was conducted with strict adherence to ethical guidelines and standards to ensure unbiased and objective results. Code and Data AvailabilityThe code and data for this study are available at https://github.com/sypsyp97/Eagle_Loss. AcknowledgmentsThis research was financed by the “Verbundprojekt 05D2022 - KI4D4E: Ein KI-basiertes Framework für die Visualisierung und Auswertung der massiven Datenmengen der 4D-Tomographie für Endanwender von Beamlines. Teilprojekt 5” (Grant No. 05D23WE1). Part of this work was carried out at the Research Campus M2OLIE and funded by the German Federal Ministry of Education and Research (BMBF) within the framework “Forschungscampus: Public–Private Partnership for Innovations” under the funding code 13GW0388A. We also acknowledge the use of Claude 3.5 Sonnet for language and grammar clean-up of this paper. ReferencesW. A. Kalender, Computed Tomography: Fundamentals, System Technology, Image Quality, Applications, John Wiley & Sons(
(2011). Google Scholar
T. M. Buzug,
“Computed tomography,”
Springer Handbook of Medical Technology, 311
–342 Springer Berlin Heidelberg, Berlin, Heidelberg
(2011). Google Scholar
L. Wang, J. A. Sahel and S. Pi,
“Sub2Full: split spectrum to boost optical coherence tomography despeckling without clean data,”
Opt. Lett., 49
(11), 3062
–3065 https://doi.org/10.1364/OL.518906
(2024).
Google Scholar
Y. Chen et al.,
“Self-supervised neuron segmentation with multi-agent reinforcement learning,”
in Proc. Thirty-Second Int. Joint Conf. Artif. Intell.,
609
–617
(2023). Google Scholar
G. Wang, J. C. Ye and B. De Man,
“Deep learning for tomographic image reconstruction,”
Nat. Mach. Intell., 2
(12), 737
–748 https://doi.org/10.1038/s42256-020-00273-z
(2020).
Google Scholar
B. Yu et al.,
“Ea-GANs: edge-aware generative adversarial networks for cross-modality MR image synthesis,”
IEEE Trans. Med. Imaging, 38
(7), 1750
–1762 https://doi.org/10.1109/TMI.2019.2895894 ITMID4 0278-0062
(2019).
Google Scholar
W. Weimin et al.,
“Enhancing liver segmentation: a deep learning approach with eas feature extraction and multi-scale fusion,”
Int. J. Innov. Res. Comput. Sci. Technol., 12
(1), 26
–34
(2024).
Google Scholar
J. Chi et al.,
“Low-dose CT image super-resolution network with dual-guidance feature distillation and dual-path content communication,”
Lect. Notes Comput. Sci., 14229 98
–108 https://doi.org/10.1007/978-3-031-43999-5_10 LNCSD9 0302-9743
(2023).
Google Scholar
J. Pan et al.,
“Motion-compensated MR CINE reconstruction with reconstruction-driven motion estimation,”
IEEE Trans. Med. Imaging, 43
(7), 2420
–2433 https://doi.org/10.1109/TMI.2024.3364504
(2024).
Google Scholar
Z. A. Balogh, Z. Barna and E. Majoros,
“Comparison of iterative reconstruction implementations for multislice helical CT,”
Zeitschr. Med. Phys., https://doi.org/10.1016/j.zemedi.2024.04.001
(2024).
Google Scholar
A. Maier et al.,
“A gentle introduction to deep learning in medical image processing,”
Zeitschr. Med. Phys., 29
(2), 86
–101 https://doi.org/10.1016/j.zemedi.2018.12.003
(2019).
Google Scholar
Y. LeCun, Y. Bengio and G. Hinton,
“Deep learning,”
Nature, 521
(7553), 436
–444 https://doi.org/10.1038/nature14539
(2015).
Google Scholar
Z. Wang et al.,
“Image quality assessment: from error visibility to structural similarity,”
IEEE Trans. Image Process., 13
(4), 600
–612 https://doi.org/10.1109/TIP.2003.819861 IIPRE4 1057-7149
(2004).
Google Scholar
L. Zhang et al.,
“A comprehensive evaluation of full reference image quality assessment algorithms,”
in 19th IEEE Int. Conf. Image Process.,
1477
–1480
(2012). https://doi.org/10.1109/ICIP.2012.6467150 Google Scholar
H. Zhao et al.,
“Loss functions for image restoration with neural networks,”
IEEE Trans. Comput. Imaging, 3
(1), 47
–57 https://doi.org/10.1109/TCI.2016.2644865
(2016).
Google Scholar
Y. Song et al.,
“Deep learning enables accurate diagnosis of novel coronavirus (COVID-19) with CT images,”
IEEE/ACM Trans. Comput. Biol. Bioinf., 18
(6), 2775
–2780 https://doi.org/10.1109/TCBB.2021.3065361
(2021).
Google Scholar
B. Stimpel et al.,
“Projection-to-projection translation for hybrid X-ray and magnetic resonance imaging,”
Sci. Rep., 9
(1), 18814 https://doi.org/10.1038/s41598-019-55108-8
(2019).
Google Scholar
L. Abrahamyan et al.,
“Gradient variance loss for structure-enhanced image super-resolution,”
in ICASSP 2022-2022 IEEE Int. Conf. Acoust. Speech and Signal Process. (ICASSP),
3219
–3223
(2022). https://doi.org/10.1109/ICASSP43922.2022.9747387 Google Scholar
J. Johnson, A. Alahi and L. Fei-Fei,
“Perceptual losses for real-time style transfer and super-resolution,”
Lect. Notes Comput. Sci., 9906 694
–711 https://doi.org/10.1007/978-3-319-46475-6_43 LNCSD9 0302-9743
(2016).
Google Scholar
G. Seif and D. Androutsos,
“Edge-based loss function for single image super-resolution,”
in IEEE Int. Conf. Acoust. Speech and Signal Process. (ICASSP),
1468
–1472
(2018). https://doi.org/10.1109/ICASSP.2018.8461664 Google Scholar
L. Jiang et al.,
“Focal frequency loss for image reconstruction and synthesis,”
in Proc. IEEE/CVF Int. Conf. Comput. Vision,
13919
–13929
(2021). https://doi.org/10.1109/ICCV48922.2021.01366 Google Scholar
L. Ge and L. Dou,
“G-Loss: a loss function with gradient information for super-resolution,”
Optik, 280 170750 https://doi.org/10.1016/j.ijleo.2023.170750 OTIKAJ 0030-4026
(2023).
Google Scholar
C. Ledig et al.,
“Photo-realistic single image super-resolution using a generative adversarial network,”
in Proc. IEEE Conf. Comput. Vision and Pattern Recognit.,
4681
–4690
(2017). https://doi.org/10.1109/CVPR.2017.19 Google Scholar
B. Benjdiraa, A. M. Alia and A. Koubaa,
“Guided frequency loss for image restoration,”
(2023). Google Scholar
Z. Fu et al.,
“Edge-aware deep image deblurring,”
Neurocomputing, 502 37
–47 https://doi.org/10.1016/j.neucom.2022.06.051 NRCGEO 0925-2312
(2022).
Google Scholar
Z. Wang et al.,
“Depth upsampling based on deep edge-aware learning,”
Pattern Recognit., 103 107274 https://doi.org/10.1016/j.patcog.2020.107274 PTNRA8 0031-3203
(2020).
Google Scholar
S. Ramanathan and M. Ramasundaram,
“Low-dose CT image reconstruction using vector quantized convolutional autoencoder with perceptual loss,”
Sädhanä, 48 43
(2023).
Google Scholar
C. Ma et al.,
“Structure-preserving super resolution with gradient guidance,”
in Proc. IEEE/CVF Conf. Comput. Vision and Pattern Recognit.,
7769
–7778
(2020). https://doi.org/10.1109/CVPR42600.2020.00779 Google Scholar
M. Han, H. Shim and J. Baek,
“Perceptual CT loss: implementing ct image specific perceptual loss for cnn-based low-dose ct denoiser,”
IEEE Access, 10 62412
–62422 https://doi.org/10.1109/ACCESS.2022.3182821
(2022).
Google Scholar
F. Piri, N. Karimi and S. Samavi,
“Enhanced segmentation in abdominal CT images: leveraging hybrid CNN-transformer architectures and compound loss function,”
in IEEE World AI IoT Congr. (AIIoT),
363
–369
(2024). Google Scholar
V. K. Tanwar, V. Anand and B. Sharma,
“A deep-learning model based on channel and spatial attention for cervical fractures prediction using CT images,”
in 5th Int. Conf. Emerg. Technol. (INCET),
1
–5
(2024). Google Scholar
J. Si and S. Kim,
“Defect detection of wood grain images with image gradient applied to ae-based generation,”
in IEEE 6th Int. Conf. Knowl. Innov. and Invent. (ICKII),
335
–337
(2023). https://doi.org/10.1109/ICKII58656.2023.10332785 Google Scholar
Y. Sun et al.,
“Data-driven filter design in FBP: transforming CT reconstruction with trainable Fourier series,”
(2024). Google Scholar
Q. Zhang et al.,
“Wavefront coding image reconstruction via physical prior and frequency attention,”
Opt. Express, 31
(20), 32875
–32886 https://doi.org/10.1364/OE.503026 OPEXFF 1094-4087
(2023).
Google Scholar
Z. Li et al.,
“Low-dose CT image denoising with improving wgan and hybrid loss function,”
Comput. Math. Methods Med., 2021
(1), 2973108 https://doi.org/10.1155/2021/2973108
(2021).
Google Scholar
Y. Kim and H. Kudo,
“Nonlocal total variation using the first and second order derivatives and its application to ct image reconstruction,”
Sensors, 20
(12), 3494 https://doi.org/10.3390/s20123494 SNSRES 0746-9462
(2020).
Google Scholar
H. Scharr,
“Optimal operators in digital image processing dissertation,”
(2000). Google Scholar
A. Dogra and P. Bhalla,
“Image sharpening by Gaussian and butterworth high pass filter,”
Biomed. Pharmacol. J., 7
(2), 707
–713 https://doi.org/10.13005/bpj/545
(2014).
Google Scholar
M.-A. Parseval,
“Mémoire sur les séries et sur l’intégration complète d’une équation aux différences partielles linéaires du second ordre, à coefficients constants,”
Mém. prés. par divers savants, Acad. des Sci. Paris (1), 1 638
–648
(1806).
Google Scholar
J. Leuschner et al.,
“Lodopab-CT, a benchmark dataset for low-dose computed tomography reconstruction,”
Sci. Data, 8
(1), 109 https://doi.org/10.1038/s41597-021-00893-z
(2021).
Google Scholar
M. Kistler et al.,
“The virtual skeleton database: an open access repository for biomedical research and collaboration,”
J. Med. Internet Res., 15
(11), e245 https://doi.org/10.2196/jmir.2930
(2013).
Google Scholar
Y. Huang et al.,
“Data extrapolation from learned prior images for truncation correction in computed tomography,”
IEEE Trans. Med. Imaging, 40
(11), 3042
–3053 https://doi.org/10.1109/TMI.2021.3072568 ITMID4 0278-0062
(2021).
Google Scholar
O. Ronneberger, P. Fischer and T. Brox,
“U-Net: convolutional networks for biomedical image segmentation,”
Lect. Notes Comput. Sci., 9351 234
–241 https://doi.org/10.1007/978-3-319-24574-4_28 LNCSD9 0302-9743
(2015).
Google Scholar
L. A. Gatys, A. S. Ecker and M. Bethge,
“Image style transfer using convolutional neural networks,”
in Proc. IEEE Conf. Comput. Vision and Pattern Recognit.,
2414
–2423
(2016). https://doi.org/10.1109/CVPR.2016.265 Google Scholar
R. Zhang et al.,
“The unreasonable effectiveness of deep features as a perceptual metric,”
in Proc. IEEE Conf. Comput. Vision and Pattern Recognit.,
586
–595
(2018). https://doi.org/10.1109/CVPR.2018.00068 Google Scholar
H. Chen et al.,
“Low-dose CT with a residual encoder-decoder convolutional neural network,”
IEEE Trans. Med. Imaging, 36
(12), 2524
–2535 https://doi.org/10.1109/TMI.2017.2715284 ITMID4 0278-0062
(2017).
Google Scholar
C. Syben et al.,
“PYRO-NN: python reconstruction operators in neural networks,”
Med. Phys., 46
(11), 5110
–5115 https://doi.org/10.1002/mp.13753 MPHYA6 0094-2405
(2019).
Google Scholar
G. Gouvine,
“Super-resolution networks for PyTorch,”
(2021). https://github.com/Coloquinte/torchSR Google Scholar
G. Gouvine,
“NinaSR: efficient small and large ConvNets for super-resolution,”
(2021). https://github.com/Coloquinte/torchSR/blob/main/doc/NinaSR.md Google Scholar
A. Hore and D. Ziou,
“Image quality metrics: PSNR vs. SSIM,”
in 20th Int. Conf. Pattern Recognit.,
2366
–2369
(2010). https://doi.org/10.1109/ICPR.2010.579 Google Scholar
I. Goodfellow, Y. Bengio and A. Courville,
“Deep feedforward networks,”
Deep Learn.,
(2016).
Google Scholar
J. Deng et al.,
“ImageNet: a large-scale hierarchical image database,”
in IEEE Conf. Comput. Vision and Pattern Recognit.,
248
–255
(2009). https://doi.org/10.1109/CVPR.2009.5206848 Google Scholar
K. Simonyan and A. Zisserman,
“Very deep convolutional networks for large-scale image recognition,”
(2014). Google Scholar
A. Krizhevsky, I. Sutskever and G. E. Hinton,
“ImageNet classification with deep convolutional neural networks,”
in Adv. Neural Inf. Process. Syst. 25,
(2012). Google Scholar
H. Hsu and P. A. Lachenbruch,
“Paired t test,”
(2014). Google Scholar
BiographyYipeng Sun is a PhD student at the Pattern Recognition Lab, University of Erlangen-Nuremberg, focusing on deep learning techniques for image reconstruction. He began his PhD in July 2023, following an MSc in medical engineering from FAU (2020–2023) and his BEng in measurement and control technology and instruments from Nanjing University of Science and Technology, China (2015–2019). Yixing Huang received his BE degree in 2013 from the Department of Biomedical Engineering, Peking University, China. He received his MS degree and PhD in 2016 and 2020, respectively, from Pattern Recognition Lab, Department of Computer Science, University of Erlangen-Nuremberg. In April 2021, he became a leading computer science researcher with the Department of Radiation Oncology, University Hospital Erlangen, Germany. Since September 2024, he has become an assistant professor at the Institute of Medical Technology, Peking University, China. His current research interests include machine learning in medical imaging and radiation oncology. Zeyu Yang is a PhD student at the Computer-Assisted Clinical Medicine, Medical Faculty Mannheim, Heidelberg University, focusing on deep learning techniques for photon counting CT reconstruction and postprocessing. He began his PhD in September 2022, following his MSc in electrical engineering from KIT (2019–2022) and his BEng in telecommunication technology from Beijing University of Post- and Telecommunications (2014–2018). Linda-Sophie Schneider is a PhD student at the Pattern Recognition Lab, University of Erlangen-Nuremberg, focusing on deep learning and optimization-based techniques for CT trajectory optimization and CT reconstruction. She began her PhD in August 2021, following her MSc in applied mathematics from FAU (2018–2021). Her current research interests include machine learning in CT imaging and combinatorial optimization. Siyuan Mei is a PhD student at the Pattern Recognition Lab, University of Erlangen-Nuremberg, focusing on deep learning techniques for image reconstruction. He began his PhD in September 2023, following his MSc in medical engineering from FAU (2020–2023) and his BEng in automation from Zhejiang University of Science and Technology (2016–2020). Andreas Maier received his bachelor’s degree and PhD in computer science from the University of Erlangen-Nuremberg, Erlangen, in 2005 and 2009, respectively. From 2005 to 2009, he was with the Pattern Recognition Laboratory, Computer Science Department, University of Erlangen-Nuremberg. From 2009 to 2010, he worked on flat-panel C-arm CT as a postdoctoral fellow with the Radiological Sciences Laboratory, Department of Radiology, Stanford University, Stanford, CA, United States. From 2011 to 2012, he joined Siemens Healthcare, Erlangen, Germany, as an Innovation Project Manager. In 2012, he returned to the University of Erlangen-Nuremberg as the Leader of the Medical Reconstruction Group, Pattern Recognition Laboratory, where he became a professor and the head of the Pattern Recognition Laboratory, in 2015. |
Computed tomography
CT reconstruction
Education and training
Image restoration
Image quality
Visualization
Image sharpness