Dual-branch residual encoder-decoder convolutional neural network (DB-REDCNN): a computed tomography-integrated multimodal network for positron emission tomography denoising
Original Article

Dual-branch residual encoder-decoder convolutional neural network (DB-REDCNN): a computed tomography-integrated multimodal network for positron emission tomography denoising

Yang Liu1, Guanglu Zou1 ORCID logo, Tao Li1, Haojia Liu2

1School of Electronics and Information, Zhengzhou University of Light Industry, Zhengzhou, China; 2Department of Medical Imaging, The Fifth Affiliated Hospital of Zhengzhou University, Zhengzhou, China

Contributions: (I) Conception and design: Y Liu, G Zou; (II) Administrative support: T Li, Y Liu; (III) Provision of study materials or patients: H Liu; (IV) Collection and assembly of data: G Zou; (V) Data analysis and interpretation: G Zou; (VI) Manuscript writing: All authors; (VII) Final approval of manuscript: All authors.

Correspondence to: Tao Li, PhD. School of Electronics and Information, Zhengzhou University of Light Industry, No. 136 Kexue Avenue, Zhengzhou 450000, China. Email: susiecash1220@gmail.com; Haojia Liu, M.Eng. Department of Medical Imaging, The Fifth Affiliated Hospital of Zhengzhou University, No.3 Kangfuqian Street, Zhengzhou 450052, China. Email: 2798617484@qq.com.

Background: Positron emission tomography (PET) requires injection of radioactive tracers, which entails concerns regarding radiation exposure. Thus, there is a need for low-dose PET (LDPET) imaging that maintains full-dose PET (FDPET) quality while reducing tracer dosage. However, dose reduction often degrades image quality. Although deep learning methods have shown promise, denoising small lesions remains challenging. To address this, we propose a dual-branch network that incorporates structural information from computed tomography (CT) to enhance LDPET quality, preserve fine details, and improve small-lesion imaging.

Methods: We propose the dual-branch residual encoder-decoder convolutional neural network (DB-REDCNN), a dual-branch multimodal network that leverages structural priors from paired CT images to enhance edge detail in PET reconstructions. Specifically, the network architecture consists of two parallel branches that independently extract modality-specific features from PET and CT inputs, followed by an effective fusion mechanism for reconstructing high-quality PET images. The overall architecture adopts a residual encoder-decoder design, with dedicated encoder and decoder modules at the input and output stages, respectively. It is important to note that although minor misalignments between CT and PET images may occur due to respiratory motion or patient movement, our dual-branch framework is inherently robust to such inconsistencies. For data acquisition, LDPET images were generated with an acquisition time of 30 seconds per bed position, while FDPET images were acquired over 150 seconds. For performance evaluation, we conducted extensive comparisons with several state-of-the-art deep learning methods. Moreover, the standardized uptake value (SUV) was included as an additional clinical indicator.

Results: The proposed DB-REDCNN demonstrated superior performance across multiple evaluation metrics. In quantitative analysis, it achieved the largest improvements compared with LDPET images in root mean square error (0.0023±0.0075), peak signal-to-noise ratio (PSNR) (1.2314±4.8354), and the structural similarity index measure (0.0078±0.0153), with a significantly higher PSNR than REDCNN and consistent advantages across all three metrics (P<0.05). Edge sharpness assessments further confirmed the superiority of DB-REDCNN, which yielded the highest |K| values across five tumor slices with statistically significant improvements over competing models (P<0.01). In SUV evaluation, although the SUV_mean error of the DB-REDCNN was slightly higher than that of LDPET, the proposed method markedly reduced SUV_max error and outperformed all other approaches.

Conclusions: The proposed DB-REDCNN model, by integrating structural priors from CT through a multibranch architecture, enhances the synthesis of FDPET from LDPET. It achieves superior quantitative performance, improves edge sharpness in lesion regions, and preserves peak SUV information, thereby offering a clinically valuable approach for radiation dose reduction in PET imaging.

Keywords: Positron emission tomography (PET); computed tomography (CT); low-dose; multibranch; deep learning


Submitted Jun 09, 2025. Accepted for publication Sep 18, 2025. Published online Nov 21, 2025.

doi: 10.21037/qims-2025-1321


Introduction

The high sensitivity of positron emission tomography (PET) in detecting metabolic abnormalities and tumor activity has led to its widespread application in clinical oncology, neurology, and cardiovascular disease diagnosis. However, acquiring high-quality PET images typically requires the administration of relatively large doses of radioactive tracers [such as 18F-fluorodeoxyglucose (18F-FDG)], raising concerns regarding the potential carcinogenic risks associated with radiation exposure (1). Although the risk from a single PET scan is relatively low, the increasing utilization of high-dose imaging modalities such as PET-computed tomography (PET/CT) has contributed to a cumulative radiation burden at the population level, which may constitute a significant public health concern (2). Therefore, the development of imaging technologies that can substantially reduce radiation exposure without compromising diagnostic image quality has become a critical focus in contemporary nuclear medicine research.

To enhance image quality while reducing tracer dosage or acquisition time, previous studies have explored the simulation of full-dose PET (FDPET) images using low-dose PET (LDPET) scans. In the early stages, efforts primarily focused on non-deep learning approaches, such as total variation (TV) regularization (3), sparse representation techniques (4), and optimization models incorporating structural constraints (5). These methods typically rely on handcrafted priors—such as edge preservation and image smoothing—to guide the reconstruction process. Despite being theoretically beneficial, they often face limitations of high computational cost, poor generalizability, and difficulty in parameter tuning. Furthermore, their dependence on fixed priors frequently results in oversmoothing and the loss of fine anatomical details, potentially compromising diagnostic sensitivity.

In recent years, the rapid advancement of deep learning has brought about significant breakthroughs in medical image processing, catalyzing the development of numerous neural network-based models for enhancing LDPET and propelling the field forward. A prominent example is the residual encoder-decoder convolutional neural network (REDCNN) model, which introduced the use of residual autoencoders for denoising low-dose CT (LDCT) images, effectively reducing noise while preserving critical structural information (6); this architectural approach has since been adapted for PET image enhancement. For PET denoising, a variety of network architectures have been proposed, including three-dimensional conditional generative adversarial networks (3D-cGANs) (7), CycleWGAN (8), and SubtlePET (9), all of which leverage adversarial training and perceptual loss to maintain lesion contrast and ensure consistency in standardized uptake value (SUV), achieving promising results. Notably, SubtlePET has demonstrated subjectively comparable image quality to FDPET and exhibited high lesion detection consistency in large-scale clinical validation, further underscoring the clinical utility of deep learning in PET imaging.

Despite recent advancements, most existing approaches remain constrained to single-modality reconstruction, relying solely on information contained within PET images. This limitation hampers the recovery of structural details, particularly under ultralow-dose or short acquisition conditions. In contrast, CT images, which exhibit high resolution and clear anatomical structures, offer valuable complementary information. When effectively integrated into the PET reconstruction pipeline, they have the potential to improve edge fidelity and structural interpretability. However, the primary value of CT lies in providing structural priors and anatomical mapping. For small lesions with low contrast or those that are not visible on CT, its direct contribution is limited; nonetheless, the overall enhancement in PET image quality indirectly improves the detectability of such lesions. Consequently, multimodal reconstruction strategies have garnered increased attention. Some studies have incorporated CT images alongside LDPET through image-level registration to train neural networks (5), while others have employed structurally constrained frameworks, such as anatomy-guided networks (10), to enforce anatomical priors to promote structural consistency. Additionally, bidirectional mapping GAN (11) employs cross-modal latent space alignment to facilitate PET-magnetic resonance imaging (MRI) fusion, and multimodal super-resolved q-space deep learning (MSR-q-DL) (12) leverages attention mechanisms to guide intermodal information flow. Collectively, these approaches underscore the pivotal role of structural imaging in enhancing PET reconstruction quality under low-dose scenarios.

The study of multimodal approaches extends beyond theoretical modeling, demonstrating strong transferability across various clinical application scenarios. In LDPET imaging, for instance, deep learning methods have been shown to outperform traditional filters such as Gaussian and nonlocal means in maintaining perceptual image quality (13). In LDCT imaging, the implementation of Wasserstein GANs (WGANs) has provided valuable technical insights for cross-modal enhancement, effectively balancing noise suppression with structural detail preservation (14). Moreover, recent progress in brain structure segmentation and image registration tasks has underscored the importance of incorporating anatomical priors from CT or MRI to improve the overall quality of medical image processing (15).

The REDCNN network is recognized for its strong performance in terms of conventional evaluation metrics such as the peak signal-to-noise ratio (PSNR) and the structural similarity index measure (SSIM). However, it often produces overly smoothed outputs, which may obscure critical anatomical details in the reconstructed images. To address this limitation, this work introduces an enhanced architecture, termed dual-branch REDCNN (DB-REDCNN), which incorporates co-registered CT images as an auxiliary input branch to provide structural priors during training. The incorporation of CT images as complementary information substantially enhances the network’s ability to reconstruct high-quality PET images from LDPET inputs. This enhancement is primarily attributed to the high-resolution anatomical information provided by CT, which compensates for the blurred structures and low signal-to-noise ratio (SNR) typically observed in LDPET images due to insufficient tracer dosage (16).

CT images are primarily used to depict normal anatomical structures. By incorporating CT data, the network can be expected to be better equipped to identify structural features such as tissue boundaries and organ contours, effectively guiding the generation of FDPET images with consistent morphology and coherent spatial distribution. This, in turn, enhances the clarity and credibility of the reconstructed outputs (17). Additionally, leveraging CT as a structural prior improves the model’s generalization ability, allowing it to produce more stable and reliable predictions, particularly in preserving high-frequency details and edge features (18). Furthermore, the complementary characteristics of CT and LDPET enable the network to learn more comprehensive feature representations, thereby further enhancing image quality and increasing the diagnostic value of the generated outputs in clinical applications (19).

Recognizing the limited utility of CT in detecting lesions with low radiotracer uptake or small size, we designed our network architecture and training strategy to treat PET as the primary source of functional information, with CT being exclusively used for structural guidance. To further evaluate the model’s ability to recover small lesions, we specifically analyze changes in SUV within lesion regions in LDPET images, thereby assessing the model’s effectiveness in preserving the functional characteristics of these lesions.

The proposed framework adopts a dual-branch architecture that independently extracts modality-specific features from PET and CT images, while a shared encoder-decoder network is employed to capture long-range dependencies. To further enhance structural representation, a structural similarity constraint has been introduced, reinforcing the boundary information guided by CT. This design enables the model to preserve functional characteristics while significantly improving anatomical detail depiction. Evaluation results demonstrate that DB-REDCNN outperforms existing methods in both conventional image quality metrics and SUV-based assessments. In light of recent advances in research and clinical needs, the proposed approach offers a practical and efficient solution for LDPET imaging, with strong scalability for extension to other multimodal applications such as brain and cardiac imaging.


Methods

Network architecture

The proposed DB-REDCNN includes a dual-branch architecture consisting of a main branch and an auxiliary branch. Each branch comprises five convolutional blocks, with each block containing a convolutional layer, a batch normalization layer, and a rectified linear unit (ReLU) activation function. All convolutional layers consist of 96 two-dimensional kernels with a size of 5×5 and a stride of 1. Following independent feature extraction by the two branches, the extracted features are concatenated and passed through five deconvolutional blocks, each replicating the structure of the convolutional blocks, including a convolutional layer, batch normalization, and ReLU activation (20,21). The full architecture of the network is illustrated in Figure 1.

Figure 1 Architecture of the proposed DB-REDCNN. CONV-BN, convolution batch normalization; CONVTrans, convolutional transpose; CT, computed tomography; DB-REDCNN, dual-branch residual encoder-decoder convolutional neural network; FDPET, full-dose positron emission tomography; PET, positron emission tomography; ReLU, rectified linear unit.

Denoising model

By constraining the image denoising problem to the image domain (22), the image denoising problem can be simplified in the context of deep learning as follows: let XRm×n denote the LDPET image, YRm×n the FDPET image, and C the corresponding CT image. The relationship among them can be formulated as follows:

X+δ(C)=σ(Y)

Here, δ(C) denotes a structural correction term extracted from the corresponding CT image C, while σ(Y) represents the degradation process mapping FDPET (denoted as Y) to LDPET (denoted as X). This mapping may be ambiguous due to the loss of structural information resulting from dose reduction. The auxiliary bias δ(C), due to the anatomical priors derived from the CT image being embedded, helps the network preserve sharp tissue boundaries and the global anatomical structure during reconstruction. Although CT images cannot directly reveal small, low-contrast lesions, their primary value lies in providing anatomical guidance for structural recovery in PET imaging. Essentially, δ(C) functions as a learnable, spatially variant offset that modulates σ(Y), thereby enabling a more accurate approximation of LDPET (denoted as X). The objective then becomes to find a function f such that:

argfminf[X+δ(C)]Y22

Here, f is treated as the optimal approximation of σ.

Residual encoding network

A symmetric encoder-decoder architecture, composed of convolutional and deconvolutional layers (23), as illustrated in Figure 2, was adopted in this study. To better preserve fine structural details in PET images and minimize the risk of irreversible degradation during deep feature extraction, a residual compensation mechanism is integrated into the PET branch. As opposed to the sole reliance on stacked layers for learning the mapping from input to output, a residual learning strategy (24) is employed to enhance feature representation and improve reconstruction quality.

Figure 2 Shortcut in the residual compensation structure. CONVTrans, convolutional transpose.

With the input e denoted as x, the residual mapping can be defined as follows:

F(x)=H(x)x

where H(x) represents the desired underlying mapping, and F(x) is the residual function. By reformulating the equation, the following is obtained:

H(x)=F(x)+x

Loss function

Since the primary objective of the loss function is to evaluate the discrepancy between the predicted and the ground truth values, L1 loss is used as the core optimization criterion. L1 loss directly penalizes the absolute difference between predictions and the corresponding ground truth, thereby encouraging the model to produce outputs that closely approximate the reference. Smaller differences yield lower loss values, while larger deviations incur proportionally greater penalties.

Let y^l denote the predicted value and yi the corresponding ground truth. L1 loss is defined as follows:

LL1=1ni=1n|yiy^i|

where n represents the total number of pixels in each image.

Evaluation metrics

To quantitatively evaluate the effectiveness of our method, we employed three widely used evaluation metrics: root mean square error (RMSE) (25), PSNR (26), and SSIM (27). Together, these metrics provide a comprehensive framework for assessing both pixel-level accuracy and perceptual similarity between the synthesized PET images and the corresponding full-dose references. These metrics are discussed in detail in the following sections.

RMSE

RMSE measures the average pixel-wise deviation between the reconstructed image and the ground truth. A lower RMSE value indicates a closer match to the reference image, reflecting higher reconstruction accuracy.

RMSE=1MNi=1Nj=1M[u(i,j)v(i,j)]2

Here, u(i,j) and v(i,j) denote the pixel intensities of the reconstructed and reference images respectively, while M and N are the image dimensions.

PSNR

The PSNR assesses the quality of image reconstruction, with a particular focus on preserving pixel intensity fidelity. Higher PSNR values indicate that the reconstructed image is more similar to the reference image in terms of both brightness accuracy and overall visual quality.

PSNR=20×log10(MAXRMSE)

Here, MAX is the maximum possible pixel value of the image. Since the PSNR is inversely related to RMSE, improving pixel accuracy directly contributes to a higher PSNR.

SSIM

SSIM evaluates perceptual similarity by comparing local patterns of pixel intensities that have been normalized for luminance and contrast. It is particularly effective in assessing structural information, such as lesion boundaries and anatomical features, which are critical for clinical interpretation in PET imaging.

SSIM(x,y)=(2μxμy+c1)(2σxy+c2)(μx2+μy2+c1)(σx2+σy2+c2)

In this equation, μx and μy are the mean intensities of images x and y, respectively; σx2 and σy2 are the variances; σxy is the covariance; and C1 and C2 are small constants used to stabilize the division.

SUV evaluation metrics

To assess quantitative accuracy, volumes of interest (VOIs) were manually delineated within lesion regions by an experienced nuclear medicine physician, using the corresponding CT images as anatomical references. The same VOIs were consistently applied across all PET images to ensure the spatial alignment and reproducibility of SUV measurements. Both SUVmean and SUVmax were extracted for each VOI; subsequently, the relative errors, referred to as SUVmeanerror and SUVmaxerror, were calculated with the following formulae:

SUVmeanerror=|SUVmeanpredSUVmeantrue|SUVmeantrue

SUVmaxerror=|SUVmaxpredSUVmaxtrue|SUVmaxtrue

To comprehensively evaluate the ability of different models to maintain the quantitative accuracy of PET imaging, the errors of SUVmean and SUVmax within the VOIs were compared across various methods.

Edge sharpness index

To analyze the model’s performance specifically at the lesion boundaries, an edge evaluation was incorporated. For this purpose, slices from the test set containing visibly identifiable small lesions were selected. A novel evaluation metric, the edge sharpness index (|K|) (28), was introduced to quantitatively assess tumor boundary delineation performance. The formula used to compute |K| is as follows:

|K|=max(I)σI

where I is the image gradient map, max(I) is the maximum gradient magnitude in the image, and σI is the standard deviation of the gradient values. A higher |K| value corresponds to sharper and cleaner image edges.

Data acquisition

This study employed whole-body PET images acquired with 18F-FDG and corresponding CT images obtained with a Vereos PET/CT system (Philps, Amsterdam, the Netherlands). No restrictions were imposed regarding patient sex, age, or disease type; however, cases exhibiting artifacts, registration errors, or lacking apparent lesions were excluded. Registration errors were primarily identified through manual inspection by experienced radiologists, who assessed the spatial alignment of key anatomical structures—such as the lungs, liver, and heart—on both PET and CT images. Misalignment of organ boundaries or significant overlap discrepancies was considered indicative of registration errors. To minimize the misalignment potentially caused by respiratory motion, standard breath-hold instructions were provided during image acquisition, and rigid registration was applied during preprocessing to further enhance alignment accuracy. In total, data from 33 patients were included. Each patient received an intravenous injection of 18F-FDG at an activity of 3.7 MBq/kg based on body weight, which was followed by a resting period of 40 to 90 minutes in a quiet waiting area prior to scanning.

All scans were performed in the three-dimensional acquisition mode, with a duration of 2.5 minutes per bed position. Image reconstruction was conducted with the ordered subsets expectation maximization (OSEM) algorithm (29), with both time-of-flight (TOF) and point-spread-function (PSF) modeling techniques being applied (30). The reconstruction matrix was set to 288×288, with two iterations and six subsets per iteration. Slice thickness and interslice spacing were both maintained at 2 mm. The number of PSF iterations and the regularization factor were set to 1 and 6, respectively. Attenuation correction for PET was performed with an LDCT scan (120 kV and 150 mAs) (31).

To simulate LDPET images, a retrospective reconstruction approach was employed with the acquisition time being reduced to 30 seconds and the same reconstruction parameters as those used for FDPET images acquired over 150 seconds being applied. Attenuation, scatter, and random coincidence corrections, as well as the image reconstruction, were performed with vendor-provided software.

In the dataset, each “pair” of images referred to a PET image slice and its corresponding CT image slice. To prevent potential bias caused by interslice correlation within the same patient, data partitioning was conducted on a per-patient basis. Specifically, 10,568 image pairs were allocated for training, 3,656 for validation, and 864 for testing, with no patient overlap across these subsets. This strategy ensured the independence and robustness of the evaluation results. The number of patients and images in the training, test, and validation sets is summarized in Table 1. This study was conducted in accordance with the Declaration of Helsinki and its subsequent amendments, and the study protocol was approved by the Ethics Committee of The Fifth Affiliated Hospital of Zhengzhou University (approval No. KY2025044). Patient consent was waived due to the retrospective nature of the data analysis and the use of anonymized data.

Table 1

Number of patients and images in the training, test, and validation sets

Items Training, n Test, n Validation, n
Patients 23 2 8
Images 10,568 864 3,656

Data preprocessing and experimental settings

Before training on the dataset comprising 33 cases, several preprocessing steps were performed. All PET images were strictly registered to their corresponding CT images during the data preprocessing stage via a voxel similarity-based registration algorithm. The alignment accuracy was further verified by experienced radiologists through visual inspection of key anatomical landmarks—such as the ventricles and liver boundaries—to ensure consistency across modalities. Subsequently, both CT and PET images were resampled to a resolution of 256×256 via bilinear interpolation. As illustrated in Figure 3, the overlay image was employed to demonstrate the registration performance of representative PET/CT pairs from the dataset.

Figure 3 Visualization of the image generated by overlaying the PET image onto the CT image via pseudocolor representation. CT, computed tomography; PET, positron emission tomography.

To reduce overfitting and improve the network’s generalization capability, data augmentation techniques were applied, including random horizontal and vertical flips as well as random rotations of up to 45 degrees. As a result, a total of 31,704 image pairs were generated.

In addition, all images were normalized through the of scaling pixel intensities to the range [0, 1], ensuring a consistent numerical scale across inputs, accelerating network convergence, and mitigating issues such as gradient explosion or vanishing gradients caused by extreme intensity values.

To further enhance generalization, increase robustness, and encourage the model to focus on local features, a patch-based training strategy was employed. This involved randomly cropping smaller patches from the input images in line with strategies adopted in previous studies (32-34).

To initialize network parameters in a manner compatible with ReLU activation, the Kaiming initialization method (35) was employed. Model training was conducted with the Adam optimizer (36), which adaptively adjusts the learning rate of each parameter with decay rates set to β1 =0.9 and β2 =0.999. Additional training configurations, including the number of epochs, batch size, patch size, and learning rate, are summarized in Table 2.

Table 2

Hyperparameters during training

Items Epochs, n Learning rate Batch size, n Patches, n Patch size
Value 50 1e−4 10 10 64

The selection of hyperparameters was guided by empirical experience from prior experiments and informed by preliminary results on the validation set. Specifically, the number of training epochs was chosen to ensure model convergence while minimizing the risk of overfitting. Batch size and patch size were determined by balancing training stability with Graphics Processing Unit (GPU) memory limitations. The initial learning rate was selected to facilitate rapid convergence during the early stages of training. A range of hyperparameter combinations was tested—for example, batch sizes ranging from 4 to 16, patch sizes from 32 to 64, and learning rates from 1e−3 to 1e−5. The final configuration was selected based on optimal performance in terms of validation loss and PSNR.

All experiments were conducted via Python 3.8 (Python Software Foundation, Wilmington, DE, USA) with the PyTorch framework (37) on a Dell Workstation (Dell Technologies, Round Rock, TX, USA) equipped with a 12 vCPU Intel(R) Xeon(R) Platinum 8352V CPU and a vGPU (32 GB) (Intel, Santa Clara, CA, USA).

Models used in the ablation study

To evaluate the effectiveness of the dual-branch architecture in the DB-REDCNN network, a series of ablation experiments was conducted. All experiments were performed on the same dataset and computing platform to ensure consistent and fair comparisons of both quantitative and qualitative performance. Specifically, as illustrated in Figure 4, we compared three models: (I) the original REDCNN (REDCNN uni), which uses only PET images as input; (II) a modified version (REDCNN multi), in which PET and CT images are concatenated along the channel dimension; and (III) the proposed DB-REDCNN with a dual-branch architecture.

Figure 4 The input image modalities of the three models. (A) REDCNN model with a single PET image used as input. (B) REDCNN multimodel, in which PET and CT images are concatenated along the vertical dimension as input. (C) Our proposed dual-branch network architecture. CT, computed tomography; PET, positron emission tomography; REDCNN, residual encoder-decoder convolutional neural network.

Results

Comparison of loss functions

We conducted comparative experiments using L1 loss and other classical loss functions, including SSIM loss and mean square error (MSE) loss. The results demonstrated that L1 loss yielded the best performance in our model. Based on the data in Table 3, after simple numerical calculations, the changes in the evaluation metrics (RMSE, PSNR, SSIM) for the model using L1 loss relative to the 30s-PET images were 0.0023±0.0075, 1.2314±4.8354, and 0.0078±0.0153. These results were superior to those of the model using SSIM loss (–0.0072±0.0333, –0.5547±6.6596, and –0.0037±0.0427) and the model using MSE loss (0.0014±0.0079, 0.7131±4.6028, and –0.0004±0.0177). Here, the values in parentheses represent, respectively, the differences in the mean RMSE, PSNR, and SSIM between the PET images generated by models using different loss functions and the 30s-PET images.

Table 3

Evaluation of the loss effect

Loss RMSE PSNR SSIM
30-s PET 0.0195±0.0083 35.1759±4.4807 0.9626±0.0180
L1 0.0172±0.0075 36.4073±4.8354 0.9704±0.0153
SSIM 0.0267±0.0333 34.6212±6.6596 0.9589±0.0427
MSE 0.0181±0.0079 35.8890±4.6028 0.9622±0.0177

The number before the “±” sign represents the mean value calculated from 864 test samples, while the number after the “±” sign indicates the standard deviation. MSE, mean square error; PET, positron emission tomography; PSNR, peak signal-to-noise ratio; RMSE, root mean square error; SSIM, structural similarity index measure.

Quantitative evaluation

The other models evaluated in this study included REDCNN, SubtlePET, and UNET_RA (38). REDCNN is a unimodal network that exclusively utilizes LDPET images as input, while SubtlePET and UNET_RA are multimodal architectures that integrate both LDPET and the corresponding CT images as inputs.

Qualitative visual assessments of FDPET, LDPET, and the PET images reconstructed by the models are provided in the axial, coronal, and sagittal views in Figure 5. Based on visual inspection, all models improved the quality of LDPET images, with our proposed method and REDCNN yielding relatively better perceptual results. However, visual inspection alone is insufficient for rigorously demonstrating the superiority of our approach; Therefore, the quantitative results are presented in Table 4 and Figure 6 to support these observations.

Figure 5 Sample images of LDPET (30 s), FDPET, and estimated PET images from the proposed, REDCNN, SubtlePET, and UNET_RA methods. The arrows in the figure indicate the location of the small lesions. Rows 1, 2, and 3 are PET images in the axial, coronal, and sagittal directions, respectively. The grayscale bar on the top indicates the SUV (g/mL). FDPET, full-dose PET; LDPET, low-dose PET; PET, positron emission tomography; REDCNN, residual encoder-decoder convolutional neural network; SUV, standardized uptake value.

Table 4

Evaluation of quantitative indicators

Method RMSE PSNR SSIM
30-s PET 0.0195±0.0083 35.1759±4.4807 0.9626±0.0180
REDCNN 0.0185±0.0089 35.9540±5.2702 0.9714±0.0157
UNET_RA 0.0187±0.0080 35.5428±4.4837 0.9464±0.0205
SubtlePET 0.0192±0.0083 35.3041±4.4202 0.9068±0.0262
Proposed 0.0172±0.0075 36.4073±4.8354 0.9704±0.0153

The number before the “±” sign represents the mean value calculated from 864 test samples, while the number after the “±” sign indicates the standard deviation. PET, positron emission tomography; PSNR, peak signal-to-noise ratio; REDCNN, residual encoder-decoder convolutional neural network; RMSE, root mean square error; SSIM, structural similarity index measure.

Figure 6 Box plots for the evaluation of quantitative indicators (data from Table 4). The improvement of model-estimated PET images in RMSE, PSNR, and SSIM with respect to the original LDPET images. LDPET, low-dose PET; PET, positron emission tomography; PSNR, peak signal-to-noise ratio; RMSE, root mean square error; SSIM, structural similarity index measure.

Compared with those of LDPET images, the improvements of model-estimated PET images in terms of RMSE, PSNR, and SSIM were, respectively, 0.001±0.0089, 0.7781±5.2702, and 0.0088±0.0157 for REDCNN; 0.0008±0.0080, 0.3669±4.4837, and –0.0162±0.0205 for UNET_RA; 0.0003±0.0083, 0.1282±4.4202, and –0.0558±0.0262 for SubtlePET; and 0.0023±0.0075, 1.2314±4.8354, and 0.0078±0.0153 for the proposed model. All estimated PET images outperformed the original low-dose data in terms of RMSE and PSNR. However, UNET_RA and SubtlePET showed lower SSIM values than did the original LDPET images. The proposed approach demonstrated consistently favorable results across all three metrics, achieving a notably higher PSNR than REDCNN, although its SSIM was slightly below that of REDCNN.

To assess the statistical significance of the performance differences across the methods, significance tests were conducted for all methods. As shown in Table 5, all models—except SubtlePET—showed statistically significant differences (P<0.05) across all metrics when compared to the 30-s PET data. SubtlePET did not exhibit statistically significant differences in RMSE or PSNR values (P>0.05). Notably, the proposed method demonstrated the most significant improvements in RMSE and PSNR values, further validating its superior performance.

Table 5

Statistical significance (P values) of the differences between each model and 30-s PET

Method RMSE PSNR SSIM
REDCNN 5.42838e−09 3.70264e−15 3.61526e−72
UNET_RA 0.000195392 0.000146802 2.34925e−72
SubtlePET 0.294267589 0.210342034 1.61363e−72
Proposed 1.07e−35 4.19352e−56 2.34465e−72

PET, positron emission tomography; PSNR, peak signal-to-noise ratio; REDCNN, residual encoder-decoder convolutional neural network; RMSE, root mean square error; SSIM, structural similarity index measure.

Figure 7 displays the radioactivity concentration (AC; Bq/mL) distribution curves across the lesion region, with a zoomed-in view focusing on the peak area to more clearly illustrate the variations in AC within this region. Compared with the other models—REDCNN (green), SubtlePET (red), and UNET_RA (purple)—the model we developed (orange) aligned most closely with the full-dose curve (brown).

Figure 7 AC contrast curve covering the lesion area. The inset in the upper left corner is a magnified image near the peak AC value. The image in the lower right corner shows the pixel region selected (covering the lesion area). For visualization purposes, AC values were proportionally scaled during plotting. AC, radioactivity concentration; PET, positron emission tomography; REDCNN, residual encoder-decoder convolutional neural network.

SUV evaluation

As illustrated in Figure 8, although the SUVmeanerror of the proposed approach was slightly higher than that of the original LDPET (30 s) image, it surpassed all other compared techniques—including LDPET, REDCNN, SubtlePET, and UNET_RA—in all remaining metrics, exhibiting a particularly pronounced advantage in reducing SUVmaxerror (39).

Figure 8 SUVmean error and SUVmax error in lesion VOIs. SUV, standardized uptake value; VOI, volume of interest.

Ablation study

Figure 9 illustrates both the cross-sectional and whole-body PET images, highlighting the impact of different network branch configurations on the REDCNN model. In these experiments, REDCNN uni was considered to be a single-branch network that processes only PET images as input, whereas REDCNN multi was considered to be a network that takes PET and CT images concatenated along the channel dimension as input.

Figure 9 Sample images of LDPET (30 s), FDPET, and estimated PET images from the proposed, REDCNN uni, and REDCNN multi methods. The arrows in the figure indicate the locations of the small lesions. Rows 1, 2, and 3 are PET images in the axial, coronal, and sagittal direction, respectively. The grayscale bar on the top indicates the SUV (g/mL). FDPET, full-dose positron emission tomography; LDPET, low-dose positron emission tomography; PET, positron emission tomography; REDCNN, residual encoder-decoder convolutional neural network; SUV, standardized uptake value.

Table 6 summarizes the RMSE, PSNR, and SSIM values for each model, reflecting their respective improvements over the LDPET images: the RMSE, PSNR, and SSIM values for the proposed model were 0.0023±0.0075, 1.2314±4.8354, and 0.0078±0.0153, respectively; those for REDCNN uni were 0.001±0.0089, 0.7781±5.2702, and 0.0088±0.0157, respectively; and those for REDCNN multi were 0.0011±0.0087, 0.8000±5.2093, and 0.0091±0.0155, respectively. Although REDCNN multi exhibited slight improvements over REDCNN uni in RMSE, PSNR, and SSIM, the significance tests (Table 7) indicated that the P values for RMSE and PSNR were both greater than 0.05, indicating that these differences were not statistically significant. Only the improvement in SSIM reached statistical significance (P<0.05). In contrast, the proposed model yielded P values well below 0.05 across all metrics, confirming that its quantitative gains are statistically significant and supporting the effectiveness of the multibranch design strategy.

Table 6

Evaluation of quantitative indicators

Methods RMSE PSNR SSIM
30-s PET 0.0195±0.0083 35.1759±4.4807 0.9626±0.0180
Proposed 0.0172±0.0075 36.4073±4.8354 0.9704±0.0153
REDCNN uni 0.0185±0.0089 35.9540±5.2702 0.9714±0.0157
REDCNN multi 0.0184±0.0087 35.9759±5.2093 0.9717±0.0155

The number before the “±” sign represents the mean value calculated from 864 test samples, while the number after the “±” sign indicates the standard deviation. PET, positron emission tomography; PSNR, peak signal-to-noise ratio; REDCNN, residual encoder-decoder convolutional neural network; RMSE, root mean square error; SSIM, structural similarity index measure.

Table 7

Statistical significance (P values) of the differences between each model and REDCNN uni

Method RMSE PSNR SSIM
REDCNN multi 0.104880231 0.145233327 1.53e−28
Proposed 8.676e−18 3.01201e−13 5.71875e−50

PSNR, peak signal-to-noise ratio; REDCNN, residual encoder-decoder convolutional neural network; RMSE, root mean square error; SSIM, structural similarity index measure.

Edge evaluation

As shown in Figure 10, Sobel operators were first applied to the selected PET slices containing lesion regions to generate corresponding gradient images; lesion areas are highlighted with red circles. Compared to other models, the proposed model produced more distinct lesion boundaries, indicating superior performance in noise suppression and detail preservation—closely approximating the quality of FDPET images.

Figure 10 The gradient images of the lesion area slices providing a more intuitive comparison of contour clarity. The lesion regions are highlighted with red circles. PET, positron emission tomography; REDCNN, residual encoder-decoder convolutional neural network.

Subsequently, regions of interest (ROIs) were delineated on the selected slices, as indicated by the rectangular box on the PET image in the upper right corner of Figure 11. The same ROI mask was applied across all reconstructed images, and the corresponding |K| values were computed. These values were then visualized in the form of a histogram (Figure 11), indicating that all other models yielded lower |K| values than the proposed model, indicating that our model achieved the best edge sharpness performance.

Figure 11 Histogram of edge sharpness index |K| value. The white box indicates the ROI area used for calculating the |K| value. PET, positron emission tomography; REDCNN, residual encoder-decoder convolutional neural network; ROI, region of interest.

To ensure the generalizability of the observed improvements, five distinct tumor slices were selected, and the |K| values for each model were computed, as shown in Table 8. A paired t-test was performed, and the results (Table 9) demonstrated that the proposed DB-REDCNN yielded statistically significant improvements (P<0.01) over other models, indicating superior performance in enhancing tumor edge sharpness.

Table 8

The |K| values of five distinct tumor ROIs

Tumor Full dose Proposed REDCNN SubtlePET UNET_RA
1 5.2367 5.1468 4.9822 4.8796 4.9024
2 6.5289 6.3452 6.0231 5.9874 6.0073
3 4.8105 4.7865 4.5379 4.5078 4.5190
4 5.9162 5.7864 5.5840 5.5342 5.5267
5 3.8421 3.7523 3.6219 3.5924 3.6134

REDCNN, residual encoder-decoder convolutional neural network; ROI, region of interest.

Table 9

Statistical significance (P values) of differences between each model and proposed

Method P value (vs. proposed)
REDCNN 0.003103**
SubtlePET 0.001139**
UNET_RA 0.001459**

**, P<0.01. PET, positron emission tomography; REDCNN, residual encoder-decoder convolutional neural network.


Discussion

In this study, we developed a novel DB-REDCNN model, an extension of the REDCNN architecture. Through its multibranch network design, the model integrates CT images as structural priors to guide the synthesis of FDPET from LDPET inputs, thereby producing outputs that more closely resemble FDPET in quality.

Regarding the choice of loss function, experimental results indicated that L1 loss consistently outperformed the other functions across three evaluation metrics, whereas SSIM loss yielded the poorest performance. In the comparative experiments, DB-REDCNN achieved competitive results on standard quantitative metrics, surpassing other models in terms of numerical performance.

To further assess the model’s ability to preserve edge sharpness, we used |K| to evaluate tumor boundary delineation within the ROIs. A series of experimental results demonstrated that our model is more effective in enhancing the representation of tumor edge contours.

In addition, we assessed the SUV, a clinically meaningful semantic-level metric, to further enhance the clinical interpretability of the proposed method. The results confirmed that DB-REDCNN not only effectively suppresses noise and restores image details but also more accurately preserves peak radiotracer uptake information within high-activity regions.

Limitation

Despite the promising results achieved in our study, several noteworthy limitations should be noted. First, the experiments did not account for the potential influence of varying CT image quality on the model’s output performance. Second, to enhance the generalizability of the proposed method, future research could incorporate list-mode data reconstructed under different conditions—such as varying numbers of iterations, matrix sizes, and PSF configurations—or adopt domain adaptation techniques (40) and invariance learning strategies (41). Third, the reliance on PET/CT hardware and manual registration by clinicians cannot fully eliminate the inherent misalignment between PET and CT images. Thus, integrating more advanced registration algorithms during preprocessing, along with introducing quantitative metrics to assess registration errors, is required. Fourth, CT-guided imaging is inherently unsuited to the restoration small lesions, particularly in regions with low radiotracer uptake. The integration of functional imaging modalities or multimodal fusion techniques may help improve the reconstruction of such lesions. Finally, although the proposed model performed well across the dataset, variations in uptake time postinjection (ranging from 40 to 90 minutes) may introduce inconsistencies in SUV distribution, potentially affecting the reliability of training and evaluation. Future work should consider enforcing stricter control of uptake times or systematically assessing the impact of uptake time variation on model performance.


Conclusions

We developed a novel DB-REDCNN model that leverages a multibranch architecture to incorporate structural priors from CT images, thereby enhancing the synthesis of FDPET from LDPET. The model provides superior performance across quantitative evaluation metrics and significantly improves edge sharpness in lesion regions. It effectively preserves peak SUV information and offers clear advantages in both visual quality and clinical applicability.


Acknowledgments

None.


Footnote

Data Sharing Statement: Available at https://qims.amegroups.com/article/view/10.21037/qims-2025-1321/dss

Funding: This work was supported by the Henan Provincial Science and Technology Research Project (No. 252102310453).

Conflicts of Interest: All authors have completed the ICMJE uniform disclosure form (available at https://qims.amegroups.com/article/view/10.21037/qims-2025-1321/coif). The authors have no conflicts of interest to declare.

Ethical Statement: The authors are accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved. The study was conducted in accordance with the Declaration of Helsinki and its subsequent amendments. The study protocol was approved by the Ethics Committee of The Fifth Affiliated Hospital of Zhengzhou University (approval No. KY2025044). Patient consent was waived due to the retrospective nature of the data analysis and the use of anonymized data.

Open Access Statement: This is an Open Access article distributed in accordance with the Creative Commons Attribution-NonCommercial-NoDerivs 4.0 International License (CC BY-NC-ND 4.0), which permits the non-commercial replication and distribution of the article with the strict proviso that no changes or edits are made and the original work is properly cited (including links to both the formal publication through the relevant DOI and the license). See: https://creativecommons.org/licenses/by-nc-nd/4.0/.


References

  1. Brenner DJ, Hall EJ. Computed tomography--an increasing source of radiation exposure. N Engl J Med 2007;357:2277-84. [Crossref] [PubMed]
  2. Hosono M, Takenaka M, Monzen H, Tamura M, Kudo M, Nishimura Y. Cumulative radiation doses from recurrent PET-CT examinations. Br J Radiol 2021;94:20210388. [Crossref] [PubMed]
  3. Wang C, Hu Z, Shi P, Liu H. Low dose PET reconstruction with total variation regularization. 2014 36th Annual International Conference of the IEEE Engineering in Medicine and Biology Society; Chicago, IL, USA. 2014:1917-20.
  4. Wang Y, Ma G, An L, Shi F, Zhang P, Lalush DS, Wu X, Pu Y, Zhou J, Shen D. Semisupervised Tripled Dictionary Learning for Standard-Dose PET Image Prediction Using Low-Dose PET and Multimodal MRI. IEEE Trans Biomed Eng 2017;64:569-79. [Crossref] [PubMed]
  5. Xie Z, Li T, Zhang X, Qi W, Asma E, Qi J. Anatomically aided PET image reconstruction using deep neural networks. Med Phys 2021;48:5244-58. [Crossref] [PubMed]
  6. Chen H, Zhang Y, Kalra MK, Lin F, Chen Y, Liao P, Zhou J, Wang G, Low-Dose CT. With a Residual Encoder-Decoder Convolutional Neural Network. IEEE Trans Med Imaging 2017;36:2524-35. [Crossref] [PubMed]
  7. Wang Y, Yu B, Wang L, Zu C, Lalush DS, Lin W, Wu X, Zhou J, Shen D, Zhou L. 3D conditional generative adversarial networks for high-quality PET image estimation at low dose. Neuroimage 2018;174:550-62. [Crossref] [PubMed]
  8. Zhou L, Schaefferkoetter JD, Tham IWK, Huang G, Yan J. Supervised learning with cyclegan for low-dose FDG PET image denoising. Med Image Anal 2020;65:101770. [Crossref] [PubMed]
  9. Weyts K, Lasnon C, Ciappuccini R, Lequesne J, Corroyer-Dulmont A, Quak E, Clarisse B, Roussel L, Bardet S, Jaudet C. Artificial intelligence-based PET denoising could allow a two-fold reduction in [18F]FDG PET acquisition time in digital PET/CT. Eur J Nucl Med Mol Imaging 2022;49:3750-60.
  10. Oktay O, Ferrante E, Kamnitsas K, Heinrich M, Bai W, Caballero J, Cook SA, de Marvao A, Dawes T, O'Regan DP, Kainz B, Glocker B, Rueckert D. Anatomically Constrained Neural Networks (ACNNs): Application to Cardiac Image Enhancement and Segmentation. IEEE Trans Med Imaging 2018;37:384-95. [Crossref] [PubMed]
  11. Hu S, Lei B, Wang S, Wang Y, Feng Z, Shen Y. Bidirectional Mapping Generative Adversarial Networks for Brain MR to PET Synthesis. IEEE Trans Med Imaging 2022;41:145-57. [Crossref] [PubMed]
  12. Qin Y, Li Y, Zhuo Z, Liu Z, Liu Y, Ye C. Multimodal super-resolved q-space deep learning. Med Image Anal 2021;71:102085. [Crossref] [PubMed]
  13. Amirrashedi M, Sarkar S, Mamizadeh H, Ghadiri H, Ghafarian P, Zaidi H, Ay MR. Leveraging deep neural networks to improve numerical and perceptual image quality in low-dose preclinical PET imaging. Comput Med Imaging Graph 2021;94:102010. [Crossref] [PubMed]
  14. Yang Q, Yan P, Zhang Y, Yu H, Shi Y, Mou X, Kalra MK, Zhang Y, Sun L, Wang G, Low-Dose CT. Image Denoising Using a Generative Adversarial Network With Wasserstein Distance and Perceptual Loss. IEEE Trans Med Imaging 2018;37:1348-57. [Crossref] [PubMed]
  15. Kleesiek J, Urban G, Hubert A, Schwarz D, Maier-Hein K, Bendszus M, Biller A. Deep MRI brain extraction: A 3D convolutional neural network for skull stripping. Neuroimage 2016;129:460-9. [Crossref] [PubMed]
  16. Pinton NJ, Bousse A, Cheze-Le-Rest C, Visvikis D. Multi-Branch Generative Models for Multichannel Imaging With an Application to PET/CT Synergistic Reconstruction. IEEE Transactions on Radiation and Plasma Medical Sciences 2025;9:654-6.
  17. Chen X, Zhou B, Xie H, Guo X, Zhang J, Sinusas AJ, Onofrey JA, Liu C. Dual-branch squeeze-fusion-excitation module for cross-modality registration of cardiac SPECT and CT. In: Wang L, Dou Q, Fletcher PT, Speidel S, Li S, editors. International Conference on Medical Image Computing and Computer-Assisted Intervention. Cham: Springer Nature Switzerland; 2022:46-55.
  18. Pinton NJ, Bousse A, Wang Z, Cheze-Le-Rest C, Maxim V, Comtat C, Sureau F, Visvikis D. Synergistic PET/CT reconstruction using a joint generative model. arXiv:2411.07339 [preprint]. 2024. Available online: https://arxiv.org/abs/2411.07339
  19. Ding Z, Zhou D, Nie R, Hou R, Liu Y. Brain Medical Image Fusion Based on Dual-Branch CNNs in NSST Domain. Biomed Res Int 2020;2020:6265708. [Crossref] [PubMed]
  20. Cao H, Liu H, Song E, Hung CC, Ma G, Xu X, Jin R, Lu J. Dual-branch residual network for lung nodule segmentation. Applied Soft Computing 2020;86:105934.
  21. Nair V, Hinton GE. Rectified linear units improve restricted boltzmann machines. Available online: https://www.cs.toronto.edu/~hinton/absps/reluICML.pdf
  22. Chen H, Zhang Y, Zhang W, Liao P, Li K, Zhou J, Wang G. Low-dose CT via convolutional neural network. Biomed Opt Express 2017;8:679-94. [Crossref] [PubMed]
  23. Mao X, Shen C, Yang YB. Image restoration using very deep convolutional encoder-decoder networks with symmetric skip connections. Advances in neural information processing systems 2016. Avaliable online: https://arxiv.org/abs/1603.09056
  24. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition. 2016: 770-8.
  25. Hodson TO. Root mean square error (RMSE) or mean absolute error (MAE): When to use them or not. Geoscientific Model Development Discussions 2022;15:5481-7.
  26. Hore A, Ziou D. Image quality metrics: PSNR vs. SSIM. 2010 20th international conference on pattern recognition. IEEE, 2010:2366-9.
  27. Wang Z, Bovik AC, Sheikh HR, Simoncelli EP. Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process 2004;13:600-12. [Crossref] [PubMed]
  28. Ahmad R, Ding Y, Simonetti OP. Edge Sharpness Assessment by Parametric Modeling: Application to Magnetic Resonance Imaging. Concepts Magn Reson Part A Bridg Educ Res 2015;44:138-49. [Crossref] [PubMed]
  29. Yao R, Seidel J, Johnson CA, Daube-Witherspoon ME, Green MV, Carson RE. Performance characteristics of the 3-D OSEM algorithm in the reconstruction of small animal PET images. IEEE transactions on medical imaging 2000;19:798-804. [Crossref] [PubMed]
  30. Joshi N, Szeliski R, Kriegman DJ. PSF estimation using sharp edge prediction. 2008 IEEE conference on computer vision and pattern recognition. IEEE; 2008:1-8.
  31. Kinahan PE, Townsend DW, Beyer T, Sashin D. Attenuation correction for a combined 3D PET/CT scanner. Med Phys 1998;25:2046-53. [Crossref] [PubMed]
  32. Dong C, Loy CC, He K, Tang X. Image Super-Resolution Using Deep Convolutional Networks. IEEE Trans Pattern Anal Mach Intell 2016;38:295-307. [Crossref] [PubMed]
  33. Xie J, Xu L, Chen E. Image denoising and inpainting with deep neural networks. Advances in neural information processing systems, 2012. Avaliable online: https://dl.acm.org/doi/10.5555/2999134.2999173
  34. Jain V, Seung S. Natural image denoising with convolutional networks. Advances in neural information processing systems, 2008. Avaliable online: https://dl.acm.org/doi/10.5555/2981780.2981876
  35. He K, Zhang X, Ren S, Sun J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. Proceedings of the IEEE international conference on computer vision. 2015:1026-34.
  36. Diederik K. Adam: A method for stochastic optimization. 2014. Avaliable online: https://arxiv.org/abs/1412.6980
  37. Paszke A, Gross S, Chintala S, et al. Automatic differentiation in pytorch. 2017. Avaliable online: https://openreview.net/forum?id=BJJsrmfCZ
  38. Chen KT, Gong E, de Carvalho Macruz FB, Xu J, Boumis A, Khalighi M, Poston KL, Sha SJ, Greicius MD, Mormino E, Pauly JM, Srinivas S, Zaharchuk G. Ultra-Low-Dose (18)F-Florbetaben Amyloid PET Imaging Using Deep Learning with Multi-Contrast MRI Inputs. Radiology 2019;290:649-56. [Crossref] [PubMed]
  39. Berghmans T, Dusart M, Paesmans M, Hossein-Foucher C, Buvat I, Castaigne C, Scherpereel A, Mascaux C, Moreau M, Roelandts M, Alard S, Meert AP, Patz EF Jr, Lafitte JJ, Sculier JPEuropean Lung Cancer Working Party for the IASLC Lung Cancer Staging Project. Primary tumor standardized uptake value (SUVmax) measured on fluorodeoxyglucose positron emission tomography (FDG-PET) is of prognostic value for survival in non-small cell lung cancer (NSCLC): a systematic review and meta-analysis (MA) by the European Lung Cancer Working Party for the IASLC Lung Cancer Staging Project. J Thorac Oncol 2008;3:6-12. [Crossref] [PubMed]
  40. Ganin Y, Lempitsky V. Unsupervised domain adaptation by backpropagation. Proceedings of the 32nd International Conference on Machine Learning, PMLR 2015;37:1180-9.
  41. Kwon J, Kim J, Park H, Choi IK. Asam: Adaptive sharpness-aware minimization for scale-invariant learning of deep neural networks. Proceedings of the 38th International Conference on Machine Learning, PMLR 2021;139:5905-14.
Cite this article as: Liu Y, Zou G, Li T, Liu H. Dual-branch residual encoder-decoder convolutional neural network (DB-REDCNN): a computed tomography-integrated multimodal network for positron emission tomography denoising. Quant Imaging Med Surg 2025;15(12):11778-11794. doi: 10.21037/qims-2025-1321

Download Citation