<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.0 20120330//EN" "JATS-journalpublishing1.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">INFORMATICA</journal-id>
<journal-title-group><journal-title>Informatica</journal-title></journal-title-group>
<issn pub-type="epub">1822-8844</issn><issn pub-type="ppub">0868-4952</issn><issn-l>0868-4952</issn-l>
<publisher>
<publisher-name>Vilnius University</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">INFOR622</article-id>
<article-id pub-id-type="doi">10.15388/26-INFOR622</article-id>
<article-categories><subj-group subj-group-type="heading">
<subject>Research Article</subject></subj-group></article-categories>
<title-group>
<article-title>SynthModDiff: A Diffusion-Based Framework for Robust Multi-Domain MRI Synthesis</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0001-5311-806X</contrib-id>
<name><surname>Nguyen Van</surname><given-names>Hieu</given-names></name><email xlink:href="nvhieuqt@dut.udn.vn">nvhieuqt@dut.udn.vn</email><xref ref-type="aff" rid="j_infor622_aff_001">1</xref><xref ref-type="corresp" rid="cor1">∗</xref><bio>
<p><bold>H. Nguyen Van</bold> is a senior lecturer at the University of Danang – University of Science and Technology. He obtained his PhD from the Saint Petersburg State Institute of Technology and his MSc and BSc degrees from Saint Petersburg State University, Russia. Over the past five years, he has published a number of research papers on the development of machine learning and deep learning models and their applications. His research interests include artificial intelligence, optimization, decision-making, and computational intelligence.</p></bio>
</contrib>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0009-0005-1717-3842</contrib-id>
<name><surname>Phan Minh</surname><given-names>Nhat</given-names></name><email xlink:href="nhat0299@gmail.com">nhat0299@gmail.com</email><xref ref-type="aff" rid="j_infor622_aff_001">1</xref><bio>
<p><bold>N. Phan Minh</bold> is a junior researcher at the Department of Software Engineering, The University of Danang – University of Science and Technology. His research interests are data science, machine learning, and deep learning, with a special focus on near-infrared spectroscopy, probabilistic models, and 2D to 3D image processing. He is an undergraduate student at the University of Danang – University of Science and Technology in Vietnam with a specialization in data science and artificial intelligence.</p></bio>
</contrib>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0002-1439-8289</contrib-id>
<name><surname>Ngo Le Huy</surname><given-names>Hien</given-names></name><email xlink:href="hien.ngo160203@vnuk.edu.vn">hien.ngo160203@vnuk.edu.vn</email><xref ref-type="aff" rid="j_infor622_aff_002">2</xref><bio>
<p><bold>H. Ngo Le Huy</bold> is a postgraduate researcher at the School of Built Environment, Engineering, and Computing, Leeds Beckett University, United Kingdom. He obtained an MSc in information and technology at Leeds Beckett University under the Erasmus Mundus in green networking and cloud computing program. During the last five years, he has published research papers mainly related to the advancements in machine learning and deep learning models and their applications. His research interests include artificial intelligence, green ICT, and sustainable computing.</p></bio>
</contrib>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0009-0004-1065-1015</contrib-id>
<name><surname>Le Hoang Ngoc</surname><given-names>Han</given-names></name><email xlink:href="102240003@hv.dut.udn.vn">102240003@hv.dut.udn.vn</email><xref ref-type="aff" rid="j_infor622_aff_001">1</xref><bio>
<p><bold>H. Le Hoang Ngoc</bold> is a junior researcher at the Department of Software Engineering, The University of Danang – University of Science and Technology. Her research interests are data science and computer vision, with a special focus on MRI image processing. She obtained her BSc from the University of Danang – University of Science and Technology in Vietnam with a specialization in data science and artificial intelligence.</p></bio>
</contrib>
<aff id="j_infor622_aff_001"><label>1</label><institution>The University of Danang – University of Science and Technology</institution>, <country>Vietnam</country></aff>
<aff id="j_infor622_aff_002"><label>2</label><institution>The University of Danang – VN-UK Institute for Research and Executive Education</institution>, <country>Vietnam</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>∗</label>Corresponding author.</corresp>
</author-notes>
<pub-date pub-type="ppub"><year>2026</year></pub-date><pub-date pub-type="epub"><day>13</day><month>3</month><year>2026</year></pub-date><volume content-type="ahead-of-print">0</volume><issue>0</issue><fpage>1</fpage><lpage>22</lpage><history><date date-type="received"><month>3</month><year>2025</year></date><date date-type="accepted"><month>3</month><year>2026</year></date></history>
<permissions><copyright-statement>© 2026 Vilnius University</copyright-statement><copyright-year>2026</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>Open access article under the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/">CC BY</ext-link> license.</license-p></license></permissions>
<abstract>
<p>Magnetic Resonance Imaging (MRI) is crucial for clinical diagnostics, offering high-resolution anatomical and functional imaging without ionizing radiation. However, prolonged acquisition times in conventional MRI lead to motion artifacts, limiting efficiency and reliability. While deep learning models such as GANs and DDPMs show promise in MRI synthesis, DDPMs suffer from stochastic variability that affects image consistency. This study proposes Synthetic Modality Diffusion (SynthModDiff), a novel multi-domain image-to-image translation framework featuring a two-stage diffusion process with a noise-aware Forward Process and Reverse Process to enhance fidelity and reduce residual noise. Experiments across multiple datasets demonstrate state-of-the-art performance in NMAE, SSIM, and PSNR metrics, while preserving fine anatomical details, making SynthModDiff highly suitable for clinical applications like radiotherapy planning.</p>
</abstract>
<kwd-group>
<label>Key words</label>
<kwd>SynthModDiff</kwd>
<kwd>denoising diffusion probabilistic models</kwd>
<kwd>MRI synthesis</kwd>
<kwd>image-to-image translation</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="j_infor622_s_001">
<label>1</label>
<title>Introduction</title>
<p>Magnetic Resonance Imaging (MRI) is widely utilized in clinical practice for diagnosis, prognosis, and radiotherapy treatment planning (Goudschaal <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_013">2021</xref>; Chang <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_005">2022</xref>; Li <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_028">2023b</xref>, <xref ref-type="bibr" rid="j_infor622_ref_030">2023d</xref>). This imaging modality provides both anatomical and functional information, enabling clinicians to accurately identify lesion sites, characterize malignancies, and assess metastatic progression. Specifically, T1-weighted (T1) MRI scans offer high contrast between white and gray matter, facilitating precise delineation of tissue boundaries for brain lesion evaluation. In contrast, T2-weighted (T2) imaging is instrumental in visualizing cerebrospinal fluid (CSF), while the fluid-attenuated inversion recovery (FLAIR) technique suppresses CSF signals to enhance the detection of lesions in proximity to fluid-filled regions. The integration of multiple MRI sequences allows for a more comprehensive assessment of underlying anatomical structures and physiological processes, thereby improving diagnostic accuracy and informing treatment strategies. However, MRI acquisition is inherently time-intensive due to the fundamental principles of imaging physics. The extended duration of various MRI sequences increases susceptibility to motion artifacts, which can degrade image quality. Furthermore, for patients with hypersensitivity to MRI contrast agents, the omission of contrast-enhanced imaging may reduce diagnostic reliability, potentially impacting clinical decision-making.</p>
<p>Deep learning, recognized for its capacity to approximate complex functions (Chang and Dinh, <xref ref-type="bibr" rid="j_infor622_ref_004">2019</xref>; Hu <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_018">2023b</xref>, <xref ref-type="bibr" rid="j_infor622_ref_017">2023a</xref>, <xref ref-type="bibr" rid="j_infor622_ref_019">2023c</xref>, Li <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_029">2023c</xref>), has been extensively investigated for MRI synthesis through various model architectures (Sun <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_046">2021</xref>; Moya-Sáez <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_037">2021</xref>). Among these, generative adversarial networks (GANs) (Hu <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_020">2022</xref>; Yu <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_052">2021</xref>; Goodfellow <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_012">2020</xref>; Pan <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_040">2021</xref>) have demonstrated substantial effectiveness in generating high-fidelity medical images across diverse imaging modalities. More recently, denoising diffusion probabilistic models (DDPMs) (Song <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_045">2020</xref>; Ho <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_016">2020</xref>) have emerged as a compelling alternative, addressing fundamental limitations of GANs, particularly mode collapse in the learning of multimodal distributions. However, DDPMs inherently rely on iterative Markov chain denoising processes, which introduce stochastic variability into image synthesis (Lyu and Wang, <xref ref-type="bibr" rid="j_infor622_ref_034">2022</xref>). To improve the robustness and efficiency of DDPMs, extensive research has focused on optimization strategies (Nichol and Dhariwal, <xref ref-type="bibr" rid="j_infor622_ref_038">2021</xref>; Dhariwal and Nichol, <xref ref-type="bibr" rid="j_infor622_ref_010">2021</xref>), and their applicability in medical image synthesis has gained increasing attention (Peng <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_042">2024</xref>). Recent advances further improve multi-contrast consistency through mutual learning and dual-domain strategies (Dayarathna <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_009">2025b</xref>). In the context of radiotherapy, the exploration of uncertainty in three-dimensional (3D) DDPM-based MRI synthesis remains in its early stages. Advancing research in this area is critical to enhancing the reliability and clinical utility of synthesized MRI images, thereby improving patient safety and treatment precision, especially with Diffusion Schrödinger bridge models demonstrating strong potential for MR-to-CT synthesis in proton therapy planning (Li <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_026">2025</xref>).</p>
<p>MRI is a fundamental imaging modality for brain diagnostics, offering high-resolution anatomical and functional information without exposing patients to ionizing radiation. MRI relies on strong magnetic fields and radiofrequency pulses to generate detailed images of brain structures. This non-ionizing property is beneficial for pediatric and vulnerable populations in reducing the risks associated with cumulative radiation exposure (Luo <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_033">2020</xref>). Brain MRI has the strength to generate precise visualization of soft tissues and has become fundamental for detecting neurological disorders, tumours, stroke, and neurodegenerative diseases.</p>
<p>However, brain MRI has been facing several challenges. Firstly, high-resolution imaging can cause longer acquisition times and increase susceptibility to patient motion artifacts, which may compromise image quality. Secondly, the signal-to-noise ratio can be influenced by thermal fluctuations and system instabilities, which affect contrast resolution and diagnostic accuracy (Berthelon <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_002">2018</xref>; Kriukova <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_025">2025</xref>). Noise, such as Gaussian or Poisson, can undermine the effectiveness of automated image processing and feature extraction algorithms and compromise the reliability of diagnostic assessments (Juneja <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_024">2024</xref>). Therefore, to address these limitations through advanced acquisition techniques, noise reduction algorithms are crucial for optimizing brain MRI quality and enhancing clinical utility.</p>
<p>To address these challenges, this study proposes a novel approach, called Synthetic Modality Diffusion (SynthModDiff), for comprehensive multi-domain image-to-image translation. SynthModDiff uses diffusion models specifically designed for MRI imaging to address noise issues in real-world scenarios. The framework consists of two sequential stages, with three key components: the Forward Process, a Deep Learning network, and the Reverse Process. Rather than executing the entire Forward Process during inference, SynthModDiff precisely estimates the noise level in the input image at a given timestep <inline-formula id="j_infor622_ineq_001"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${t^{\prime }}$]]></tex-math></alternatives></inline-formula> and proceeds with the denoising process until timestep <italic>T</italic>. This method effectively mitigates residual noise in the synthesized image when the input contains noise and improves its applicability in practical medical imaging environments.</p>
</sec>
<sec id="j_infor622_s_002">
<label>2</label>
<title>Related Work</title>
<p>Medical image modality conversion involves transforming one type of medical image into another to enhance diagnostic capabilities. Clinicians often require multiple imaging modalities to obtain a comprehensive view of a patient’s condition. Diffusion models offer a viable solution in cases where certain imaging techniques are unavailable due to equipment constraints or patient-specific limitations. Additionally, these models facilitate the integration of multimodal imaging data, providing deeper insights into pathological conditions. For instance, combining PET and MRI data enables more precise tumor localization and improved monitoring of treatment response.</p>
<p>Several studies have explored diffusion-based approaches for medical image modality conversion. Özbey <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor622_ref_039">2023</xref>) introduced SynDiff, the first unsupervised modality conversion framework leveraging an adversarial diffusion model. Unlike conventional diffusion models, SynDiff employs larger diffusion step sizes and an adversarial mapping mechanism for the reverse process, enabling efficient and high-fidelity modality translation. Lyu and Wang (<xref ref-type="bibr" rid="j_infor622_ref_034">2022</xref>) applied denoising diffusion probabilistic models (DDPMs) and stochastic differential equations (SDEs) to convert MRI to CT using T2-weighted MRI scans, achieving superior results compared to previous methods. Li <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor622_ref_027">2023a</xref>) proposed the Denoising Diffusion Model for Medical Image Synthesis (DDMM-Synth), which synthesizes high-quality CT images by integrating anatomical information from MRI and sparse CT data while maintaining data consistency. Further advancements were made by Li <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor622_ref_031">2024</xref>) with the development of the Frequency-Guided Diffusion Model (FGDM), which relies solely on target-domain samples for training and enables zero-shot medical image modality conversion. This model is particularly effective for cone-beam CT-to-CT transformation, outperforming existing methods.</p>
<p>For three-dimensional (3D) medical image modality conversion. Pan <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor622_ref_041">2024</xref>) proposed MC-DDPM, a diffusion-based framework for MRI-to-CT conversion. This model becomes a major advancement by synthesizing CT images from MRI data in a 3D context. It uses a Swin-VNet structure in the reverse process to optimize the denoising procedure, which ensures that the generated CT images align with MRI-derived anatomical structures. Furthermore, echocardiographic imaging, which captures dynamic cardiac motion, has been researched using diffusion models. Tiago <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor622_ref_047">2023</xref>) developed an adversarial denoising diffusion model integrated with a generative adversarial network (GAN) to synthesize echocardiograms and perform domain-specific image translation. By incorporating a larger step size and using GAN-based denoising, this model effectively generates diverse echocardiographic samples while preserving critical anatomical structures. More recently, Dayarathna <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor622_ref_009">2025b</xref>) proposed MU-Diff, a mutual learning diffusion framework, while their follow-up D2Diff (Dayarathna <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_008">2025a</xref>) leverages dual spatial-frequency domain learning. Frequency-aware diffusion models (Jiang <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_023">2025</xref>) and residual-guided approaches (Safari <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_043">2025</xref>) have also shown superior performance in preserving anatomical details.</p>
<p>Another approach is Generative Adversarial Networks (GANs), which have been used for MRI synthesis for multi-contrast image generation in recent studies (Zhang <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_055">2022</xref>; Yurt <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_053">2021</xref>; Wu <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_051">2020</xref>). Sohail <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor622_ref_044">2019</xref>) proposed a GAN-based method that synthesizes multi-contrast MRI without paired training data, and uses adversarial and cycle-consistency losses to enhance realism. Similarly, Zhu <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor622_ref_057">2017</xref>) introduced CycleGAN to translate between 2 MRI contrast domains. By employing dual generator and discriminator networks, CycleGAN ensures structural consistency through cycle-consistency loss. Experimental results demonstrate its effectiveness across various image translation tasks. Wei <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor622_ref_050">2018</xref>) enhanced Wasserstein-GAN training by introducing a consistency term for stability and image quality, along with a dual effect to capture diverse data distributions. Their method outperformed WGAN and other GAN variants in quality, diversity, and stability. Zhang <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor622_ref_056">2023</xref>) proposed Dual-ArbNet, Chen <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor622_ref_006">2023</xref>) introduced CANM-Net, and Grigas <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor622_ref_014">2023</xref>) applied Transformer GANs for noise reduction and super-resolution. GANs have become essential in brain MRI synthesis, with most studies targeting image generation, as noted by Ali <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor622_ref_001">2022</xref>).</p>
</sec>
<sec id="j_infor622_s_003">
<label>3</label>
<title>Proposed Method</title>
<sec id="j_infor622_s_004">
<label>3.1</label>
<title>Synthetic Modality Diffusion Architecture</title>
<p>The proposed method, Synthetic Modality Diffusion – SynthModDiff, is based on a fundamental implementation of DDPMs, which represent a class of generative models that have been proven for their ability to generate high-quality data by leveraging probabilistic frameworks. These models operate by systematically introducing noise to data during a forward diffusion process and then training a deep learning network to reverse this process. SynthModDiff learns to iteratively remove noise in a stepwise, residual manner, ultimately reconstructing the original data distribution. By employing this structured approach, our model demonstrates exceptional performance in generating complex data distributions, making it a powerful tool for a wide range of applications in generative modelling.</p>
<p>Figure <xref rid="j_infor622_fig_001">1</xref> illustrates the detailed workflow of SynthModDiff - the proposed methodology that is structured into 2 sequential stages.</p>
<fig id="j_infor622_fig_001">
<label>Fig. 1</label>
<caption>
<p>SynthModDiff architecture.</p>
</caption>
<graphic xlink:href="infor622_g001.jpg"/>
</fig>
<list>
<list-item id="j_infor622_li_001">
<label>1.</label>
<p>In the first stage, a diffusion model is trained with real clinical data to generate synthetic images. This process eliminates Gaussian noise progressively, beginning from an initial state of pure Gaussian noise and iteratively refining the data toward its original distribution.</p>
</list-item>
<list-item id="j_infor622_li_002">
<label>2.</label>
<p>The second stage is designed to identify the optimal step within the generative pipeline at which to initiate the diffusion process for real-world noisy images. This step matches the noise level of the actual noisy image to the corresponding noise level within the generative framework. To ensure the accuracy and reliability of this stage, actual clinical images augmented with simulated noise were used to generate a rigourous evaluation of the method’s performance under different controlled conditions and varying noise characteristics.</p>
</list-item>
</list>
<p>As illustrated in Fig. <xref rid="j_infor622_fig_001">1</xref>, the diffusion model employed in SynthModDiff comprises two sequential stages described above with three core components: the Forward Process, a Deep Learning network, and the Reverse Process. Each of these elements plays a distinct and critical role in the overall generative process.</p>
<p>The evaluation of the proposed method was conducted on images corrupted by two distinct types of noise to assess its performance under varying conditions. First, the method was tested on images contaminated with Gaussian noise, which aligns with the noise distribution used during the training phase of SynthModDiff. This ensures the capacity to efficiently handle the type of noise for which the model was explicitly trained. Second, the evaluation extended to images affected by Poisson noise, which is widely recognized as a more realistic representation of noise typically faced in X-ray imaging. By integrating Poisson noise into the evaluation, the study aimed to assess the model’s generalizability and robustness when applied to scenarios corresponding to real-world clinical settings. This dual approach to noise evaluation provides a comprehensive analysis of the method’s capability to adapt to both synthetic and clinically-relevant noise characteristics and highlights its potential for practical applications in medical imaging.</p>
</sec>
<sec id="j_infor622_s_005">
<label>3.2</label>
<title>The Forward Process</title>
<p>The Forward Process is responsible for corrupting progressively an input image, denoted as <inline-formula id="j_infor622_ineq_002"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∈</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">H</mml:mi>
<mml:mo>×</mml:mo>
<mml:mi mathvariant="italic">W</mml:mi>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${X_{0}}\in {\mathbb{R}^{H\times W}}$]]></tex-math></alternatives></inline-formula>, by using Additive White Gaussian Noise over a sequence of timesteps <inline-formula id="j_infor622_ineq_003"><alternatives><mml:math>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">T</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$t\in [0,T]$]]></tex-math></alternatives></inline-formula> to create noise corrupted image <inline-formula id="j_infor622_ineq_004"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{t}}$]]></tex-math></alternatives></inline-formula> for timestep <italic>t</italic> using hyperparameters <inline-formula id="j_infor622_ineq_005"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\beta _{t}}$]]></tex-math></alternatives></inline-formula>. This forward diffusion process incrementally degrades the clean image and transitions it to a state of pure noise. The mathematical formulation of this process is defined in equation (<xref rid="j_infor622_eq_001">1</xref>). 
<disp-formula id="j_infor622_eq_001">
<label>(1)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">q</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:msqrt>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msqrt>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mtext mathvariant="bold">I</mml:mtext>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ q({X_{t}}|{X_{t-1}})=\mathcal{N}({X_{t}};\sqrt{1-{\beta _{t}}}{X_{t-1}},{\beta _{t}}\textbf{I}).\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>The Forward Process is executed in a single step for any arbitrary <italic>t</italic> as follows: 
<disp-formula id="j_infor622_eq_002">
<label>(2)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr class="split-mtr">
<mml:mtd class="split-mtd">
<mml:mi mathvariant="italic">q</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
<mml:mtd class="split-mtd">
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">q</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∏</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:mi mathvariant="italic">q</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr class="split-mtr">
<mml:mtd class="split-mtd"/>
<mml:mtd class="split-mtd">
<mml:mo>=</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mo maxsize="1.19em" minsize="1.19em" fence="true" mathvariant="normal">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:msqrt>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msqrt>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mtext mathvariant="bold">I</mml:mtext>
<mml:mo maxsize="1.19em" minsize="1.19em" fence="true" mathvariant="normal">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \begin{aligned}{}q({X_{0:t}}|{X_{0}})& =q({X_{1}},{X_{2}},\dots ,{X_{t}}|{X_{0}})={\prod \limits_{t=1}^{T}}q({X_{t}}|{X_{t-1}})\\ {} & =\mathcal{N}\big({X_{t}};\sqrt{{\bar{\alpha }_{t}}}{X_{0}},(1-\bar{\alpha }t)\textbf{I}\big),\end{aligned}\]]]></tex-math></alternatives>
</disp-formula> 
where <inline-formula id="j_infor622_ineq_006"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\alpha _{t}}=1-{\beta _{t}}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_infor622_ineq_007"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mo largeop="false" movablelimits="false">∏</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\bar{\alpha }_{t}}={\textstyle\prod _{s=0}^{t}}{\alpha _{s}}$]]></tex-math></alternatives></inline-formula>. The parameter <inline-formula id="j_infor622_ineq_008"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\beta _{t}}$]]></tex-math></alternatives></inline-formula> can be updated using any differentiable function that guarantees <inline-formula id="j_infor622_ineq_009"><alternatives><mml:math>
<mml:msqrt>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msqrt>
<mml:mo stretchy="false">≈</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[$\sqrt{{\bar{\alpha }_{T}}}\approx 0$]]></tex-math></alternatives></inline-formula>. The linear function: <inline-formula id="j_infor622_ineq_010"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\beta _{t}}=({\beta _{1}}-{\beta _{2}})\frac{t}{T}+{\beta _{2}}$]]></tex-math></alternatives></inline-formula> is adopted with <inline-formula id="j_infor622_ineq_011"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>0.02</mml:mn></mml:math><tex-math><![CDATA[${\beta _{1}}=0.02$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_infor622_ineq_012"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>6</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${\beta _{2}}={10^{-6}}$]]></tex-math></alternatives></inline-formula>. The value of <inline-formula id="j_infor622_ineq_013"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\beta _{2}}$]]></tex-math></alternatives></inline-formula> was chosen to ensure that the resulting evaluation noise levels correspond to realistic X-ray dose levels.</p>
<p>To perform diffusion in inference, the Forward Process is executed from <inline-formula id="j_infor622_ineq_014"><alternatives><mml:math>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[$t={t^{\prime }}$]]></tex-math></alternatives></inline-formula> to <inline-formula id="j_infor622_ineq_015"><alternatives><mml:math>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">T</mml:mi></mml:math><tex-math><![CDATA[$t=T$]]></tex-math></alternatives></inline-formula>, illustrated in Fig. <xref rid="j_infor622_fig_001">1</xref>, where <inline-formula id="j_infor622_ineq_016"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${t^{\prime }}$]]></tex-math></alternatives></inline-formula> represents the denoising timestep determined based on an estimate of the image’s noise level. The method used to compute <inline-formula id="j_infor622_ineq_017"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${t^{\prime }}$]]></tex-math></alternatives></inline-formula> depends on the probabilistic noise model employed. In this study, two noise models are considered: Gaussian noise, for which the Deep Learning Network has been explicitly trained; and Poisson noise, which arises naturally in X-ray imaging due to its quantum characteristics. Gaussian noise is simulated using equation (<xref rid="j_infor622_eq_002">2</xref>) for a given timestep <italic>t</italic>, leading to <inline-formula id="j_infor622_ineq_018"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">t</mml:mi></mml:math><tex-math><![CDATA[${t^{\prime }}=t$]]></tex-math></alternatives></inline-formula>. Conversely, Poisson noise is modelled using equation (<xref rid="j_infor622_eq_003">3</xref>), which incorporates a minor Gaussian noise component <inline-formula id="j_infor622_ineq_019"><alternatives><mml:math>
<mml:mi mathvariant="italic">η</mml:mi>
<mml:mo stretchy="false">∼</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">I</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\eta \sim \mathcal{N}(0,I)$]]></tex-math></alternatives></inline-formula>, scaled by <inline-formula id="j_infor622_ineq_020"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>=</mml:mo>
<mml:mn>10</mml:mn></mml:math><tex-math><![CDATA[${\sigma ^{2}}=10$]]></tex-math></alternatives></inline-formula> to account for electronic noise effects. 
<disp-formula id="j_infor622_eq_003">
<label>(3)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mtext mathvariant="italic">Poisson</mml:mtext>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">I</mml:mi>
<mml:mo movablelimits="false">exp</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>+</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mi mathvariant="italic">η</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">I</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ Y=\frac{\textit{Poisson}(I\exp (-X))+{\sigma ^{2}}\eta }{I},\]]]></tex-math></alternatives>
</disp-formula> 
where <italic>X</italic> is noisy input image and <inline-formula id="j_infor622_ineq_021"><alternatives><mml:math>
<mml:mi mathvariant="italic">I</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo largeop="false" movablelimits="false">∫</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">I</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">ϵ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">d</mml:mi>
<mml:mi mathvariant="italic">ϵ</mml:mi></mml:math><tex-math><![CDATA[$I=\textstyle\int {I_{0}}(\epsilon )d\epsilon $]]></tex-math></alternatives></inline-formula> corresponds to the flood image. Here, the parameter <italic>I</italic> denotes the effective photon dose (or intensity level) of the X-ray acquisition, defined as <inline-formula id="j_infor622_ineq_022"><alternatives><mml:math>
<mml:mi mathvariant="italic">I</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo largeop="false" movablelimits="false">∫</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">I</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">ϵ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mspace width="0.1667em"/>
<mml:mi mathvariant="italic">d</mml:mi>
<mml:mi mathvariant="italic">ϵ</mml:mi></mml:math><tex-math><![CDATA[$I=\textstyle\int {I_{0}}(\epsilon )\hspace{0.1667em}d\epsilon $]]></tex-math></alternatives></inline-formula>, where <inline-formula id="j_infor622_ineq_023"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">I</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">ϵ</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${I_{0}}(\epsilon )$]]></tex-math></alternatives></inline-formula> represents the incident photon spectrum over energy <italic>ϵ</italic>. The parameter <italic>I</italic> directly controls the signal-dependent variance of the Poisson noise, with lower values of <italic>I</italic> corresponding to higher noise levels. In simulated experiments, <italic>I</italic> is fixed to predefined dose levels, while for real clinical images, it can be estimated from acquisition parameters or via noise estimation methods such as Turajlić and Karahodzic (<xref ref-type="bibr" rid="j_infor622_ref_048">2017</xref>). Owing to the signal-dependent characteristics of Poisson noise, the denoising timestep <inline-formula id="j_infor622_ineq_024"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${t^{\prime }}$]]></tex-math></alternatives></inline-formula> is systematically determined by evaluating the maximum noise variance present in the image, as formulated by: 
<disp-formula id="j_infor622_eq_004">
<label>(4)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mtext mathvariant="bold">I</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">N</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∑</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">N</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mtext>Percentile</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>95</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:mo movablelimits="false">max</mml:mo>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ X=\frac{\textbf{I}}{N}{\sum \limits_{i=1}^{N}}{\text{Percentile}_{95}}\big(\max \big({X^{i}}\big)\big).\]]]></tex-math></alternatives>
</disp-formula> 
Here, <italic>N</italic> denotes the size of the training dataset. A percentile-based approach is employed to mitigate the influence of high-intensity artificial details in the images, such as medical annotations. The denoising timestep <inline-formula id="j_infor622_ineq_025"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${t^{\prime }}$]]></tex-math></alternatives></inline-formula> is then estimated as follows: 
<disp-formula id="j_infor622_eq_005">
<label>(5)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">≈</mml:mo>
<mml:mtext>Variance</mml:mtext>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:mo>−</mml:mo>
<mml:mo movablelimits="false">log</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {t^{\prime }}=1-{\bar{{\alpha ^{\prime }}}_{t}}\approx \text{Variance}\big(-\log (Y)\big).\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>The inferred timestep <inline-formula id="j_infor622_ineq_026"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${t^{\prime }}$]]></tex-math></alternatives></inline-formula> is defined as the timestep in the Gaussian diffusion process whose noise variance matches the estimated variance of the observed Poisson-corrupted image. Specifically, since the variance of the forward diffusion process at timestep <italic>t</italic> is given by <inline-formula id="j_infor622_ineq_027"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$1-{\bar{\alpha }_{t}}$]]></tex-math></alternatives></inline-formula>, the timestep <inline-formula id="j_infor622_ineq_028"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${t^{\prime }}$]]></tex-math></alternatives></inline-formula> is selected such that <inline-formula id="j_infor622_ineq_029"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">≈</mml:mo>
<mml:mi mathvariant="normal">Variance</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mo>−</mml:mo>
<mml:mo movablelimits="false">log</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$1-{\bar{\alpha }_{{t^{\prime }}}}\approx \mathrm{Variance}(-\log (Y))$]]></tex-math></alternatives></inline-formula>. This establishes an explicit correspondence between the Poisson noise level and the Gaussian diffusion noise schedule.</p>
<p>Determining the timestep <inline-formula id="j_infor622_ineq_030"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${t^{\prime }}$]]></tex-math></alternatives></inline-formula> using equation (<xref rid="j_infor622_eq_005">5</xref>) ensures that the model effectively removes noise with the highest variance. It is important to note that computing <italic>Y</italic> requires prior knowledge of the dose <italic>I</italic>. For real noisy images, <italic>I</italic> can be estimated based on the X-ray acquisition parameters or through noise estimation techniques, such as those proposed by Turajlić and Karahodzic (<xref ref-type="bibr" rid="j_infor622_ref_048">2017</xref>). To achieve an equivalent noise level between Gaussian and Poisson noise, Gaussian noise was simulated by selecting the timestep <italic>t</italic> from equation (<xref rid="j_infor622_eq_002">2</xref>) to match the denoising timestep <inline-formula id="j_infor622_ineq_031"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${t^{\prime }}$]]></tex-math></alternatives></inline-formula> estimated for Poisson noise. During inference, the forward diffusion process is initialized at timestep <inline-formula id="j_infor622_ineq_032"><alternatives><mml:math>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[$t={t^{\prime }}$]]></tex-math></alternatives></inline-formula>, and the reverse process is subsequently applied from <inline-formula id="j_infor622_ineq_033"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${t^{\prime }}$]]></tex-math></alternatives></inline-formula> down to <inline-formula id="j_infor622_ineq_034"><alternatives><mml:math>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[$t=0$]]></tex-math></alternatives></inline-formula>, ensuring consistency between the estimated noise level of the input image and the diffusion model’s noise schedule.</p>
</sec>
<sec id="j_infor622_s_006">
<label>3.3</label>
<title>The Reverse Process</title>
<p>The reverse process reconstructs an image by iteratively inverting the forward diffusion process. Given that the inverse of a Gaussian diffusion process follows a Gaussian distribution, the probability density function of the original data can be recovered by integrating over the sequence of state transitions in the associated Markov chain: 
<disp-formula id="j_infor622_eq_006">
<label>(6)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right">
<mml:mtr>
<mml:mtd class="align-odd">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∫</mml:mo></mml:mstyle>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">d</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∫</mml:mo></mml:mstyle>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo largeop="true" movablelimits="false">∏</mml:mo></mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mi mathvariant="italic">d</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {p_{\theta }}({X_{0}})=\int {p_{\theta }}({X_{0:T}})d{X_{1:T}}=\int p({X_{T}}){\prod \limits_{t=0}^{T-1}}{p_{\theta }}({X_{t-1}}|{X_{t}})d{X_{1:T}}.\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>Let <italic>θ</italic> denote the set of parameters governing the deep learning model. The prior distribution is defined as <inline-formula id="j_infor622_ineq_035"><alternatives><mml:math>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">∼</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">I</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$p({X_{T}})\sim \mathcal{N}(0,I)$]]></tex-math></alternatives></inline-formula>, while the transition probability in the reverse diffusion process is formulated as: 
<disp-formula id="j_infor622_eq_007">
<label>(7)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∣</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {p_{\theta }}({X_{t-1}}\mid {X_{t}})=\mathcal{N}\big({X_{t-1}};{\mu _{\theta }}({X_{t}},t),{\bar{\beta }_{t}}({X_{t}})\big).\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>Here, the mean function <inline-formula id="j_infor622_ineq_036"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mu _{\theta }}$]]></tex-math></alternatives></inline-formula> is approximated by <inline-formula id="j_infor622_ineq_037"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">η</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${X_{t}}-{\eta _{\theta }}$]]></tex-math></alternatives></inline-formula>, where <inline-formula id="j_infor622_ineq_038"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">η</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\eta _{\theta }}$]]></tex-math></alternatives></inline-formula> represents the Gaussian noise estimated by the neural network. Additionally, to maintain numerical stability during the Reverse Process, the predicted mean <inline-formula id="j_infor622_ineq_039"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mu _{0}}$]]></tex-math></alternatives></inline-formula> is clipped within the intensity range <inline-formula id="j_infor622_ineq_040"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[-1,1]$]]></tex-math></alternatives></inline-formula>.</p>
</sec>
<sec id="j_infor622_s_007">
<label>3.4</label>
<title>The Deep Learning Network</title>
<p>The U-Net architecture utilized in this study is structured hierarchically, with a configuration defined by the sequence <inline-formula id="j_infor622_ineq_041"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[3,3,3,3,3,3,3]$]]></tex-math></alternatives></inline-formula>, where each numerical value denotes the number of layers in the corresponding network stage. This design integrates two residual layers with ReLU activation function, alongside a Transformer layer that employs 8 attention heads per layer. The incorporation of the Transformer mechanism enables the model to effectively capture long-range dependencies and contextual relationships, thereby improving its capacity to process complex image data.</p>
<p>The network employs an approach to upsampling and downsampling across multiple blocks, adhering to a predefined scaling ratio of <inline-formula id="j_infor622_ineq_042"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>4</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>8</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>4</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[1,2,4,8,4,2,1]$]]></tex-math></alternatives></inline-formula>. This scaling process progressively reduces the spatial resolution of feature maps and then symmetrically restores it, thereby enabling multi-scale feature extraction. In particular, the network initially processes input data at a resolution of 128 channels, which is incrementally expanded to a maximum of 1024 channels at the bottleneck stage, then reduced symmetrically back to 128. These transformations are implemented using convolutional operations with a kernel size of <inline-formula id="j_infor622_ineq_043"><alternatives><mml:math>
<mml:mn>3</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>3</mml:mn></mml:math><tex-math><![CDATA[$3\times 3$]]></tex-math></alternatives></inline-formula>, to ensure efficient feature representation while maintaining spatial coherence throughout the network. With this hierarchical framework, the U-Net model effectively captures both fine-grained local details and broader global structures, hence becoming highly compatible with image reconstruction applications.</p>
</sec>
</sec>
<sec id="j_infor622_s_008">
<label>4</label>
<title>Experiments and Discussion</title>
<sec id="j_infor622_s_009">
<label>4.1</label>
<title>Dataset</title>
<p>The performance of the proposed method was assessed with equivalent Poisson and Gaussian noise. The noise parameters were selected to ensure both physical relevance to X-ray imaging and statistical consistency within the diffusion framework. Specifically, the Poisson intensity <italic>I</italic> was chosen to represent clinically meaningful dose levels, with <inline-formula id="j_infor622_ineq_044"><alternatives><mml:math>
<mml:mi mathvariant="italic">I</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo>×</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[$I=5\times {10^{4}}$]]></tex-math></alternatives></inline-formula> for a high-dose setting and <inline-formula id="j_infor622_ineq_045"><alternatives><mml:math>
<mml:mi mathvariant="italic">I</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>9</mml:mn>
<mml:mo>×</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[$I=9\times {10^{3}}$]]></tex-math></alternatives></inline-formula> for a low-dose setting, reflecting typical diagnostic and dose-reduction scenarios in X-ray imaging.</p>
<p>To enable a fair comparison between Poisson and Gaussian noise, Gaussian noise levels were selected to achieve equivalent noise variance in the diffusion process, governed by <inline-formula id="j_infor622_ineq_046"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">¯</mml:mo></mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$1-{\bar{\alpha }_{t}}$]]></tex-math></alternatives></inline-formula>. Accordingly, timesteps <inline-formula id="j_infor622_ineq_047"><alternatives><mml:math>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>3</mml:mn></mml:math><tex-math><![CDATA[$t=3$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_infor622_ineq_048"><alternatives><mml:math>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>9</mml:mn></mml:math><tex-math><![CDATA[$t=9$]]></tex-math></alternatives></inline-formula> were chosen for the high- and low-dose cases, resulting in noise levels comparable to those induced by the corresponding Poisson settings. The denoising timestep <inline-formula id="j_infor622_ineq_049"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${t^{\prime }}$]]></tex-math></alternatives></inline-formula> for Poisson noise was estimated using the variance-based criterion in equation (<xref rid="j_infor622_eq_005">5</xref>), ensuring alignment between the signal-dependent Poisson noise and the diffusion noise level.</p>
<p>In this research, the proposed method was evaluated using 2 widely used medical image datasets: BraTS2020 (Zhang <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_054">2020</xref>) and IXI (Ionescu <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_021">2017</xref>). The BraTS2020 dataset is a proven guideline for brain tumour segmentation in magnetic resonance imaging (MRI) and contains multi-modal MRI scans, including T1-weighted, T1-weighted contrast-enhanced, T2-weighted, and Fluid Attenuated Inversion Recovery (FLAIR) sequences, as shown in Table <xref rid="j_infor622_tab_001">1</xref>. The IXI (Information eXtraction from Images) dataset is a public collection of multimodal brain imaging data, containing 600 MRI scans from healthy subjects. These images were acquired using 4 different imaging protocols: T1-weighted, T2-weighted, Proton Density (PD), and Magnetic Resonance Angiography (MRA), as summarized in Table <xref rid="j_infor622_tab_001">1</xref>.</p>
<table-wrap id="j_infor622_tab_001">
<label>Table 1</label>
<caption>
<p>Statistical analysis of the BraTS2020 and IXI Datasets.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Attributes</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">BRATS2020</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">IXI</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">Number of contrasts</td>
<td style="vertical-align: top; text-align: left">4 (T1, T2, T1CE, FLAIR)</td>
<td style="vertical-align: top; text-align: left">4 (T1, T2, PD, MRA)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Number of samples per contrast</td>
<td style="vertical-align: top; text-align: left">494 images/contrast</td>
<td style="vertical-align: top; text-align: left">568 images/contrast</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Image size</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor622_ineq_050"><alternatives><mml:math>
<mml:mn>240</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>240</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>155</mml:mn></mml:math><tex-math><![CDATA[$240\times 240\times 155$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left">T1: <inline-formula id="j_infor622_ineq_051"><alternatives><mml:math>
<mml:mn>256</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>256</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>150</mml:mn></mml:math><tex-math><![CDATA[$256\times 256\times 150$]]></tex-math></alternatives></inline-formula> T2: <inline-formula id="j_infor622_ineq_052"><alternatives><mml:math>
<mml:mn>256</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>256</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>130</mml:mn></mml:math><tex-math><![CDATA[$256\times 256\times 130$]]></tex-math></alternatives></inline-formula> MRA: <inline-formula id="j_infor622_ineq_053"><alternatives><mml:math>
<mml:mn>512</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>512</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>100</mml:mn></mml:math><tex-math><![CDATA[$512\times 512\times 100$]]></tex-math></alternatives></inline-formula> PD: <inline-formula id="j_infor622_ineq_054"><alternatives><mml:math>
<mml:mn>256</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>256</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>130</mml:mn></mml:math><tex-math><![CDATA[$256\times 256\times 130$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">Ratio healthy/disease</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><inline-formula id="j_infor622_ineq_055"><alternatives><mml:math>
<mml:mn>40</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>60</mml:mn></mml:math><tex-math><![CDATA[$40/60$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><inline-formula id="j_infor622_ineq_056"><alternatives><mml:math>
<mml:mn>100</mml:mn>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[$100/0$]]></tex-math></alternatives></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In multi-domain MRI synthesis, statistical bias toward certain contrasts may arise when the training data distribution is imbalanced across modalities. To alleviate this issue, we evaluate SynthModDiff on two complementary public datasets, BraTS2020 and IXI, which together provide a diverse and relatively balanced set of MRI contrasts. This design choice reduces the risk of the model overfitting to a dominant modality, such as T1-weighted images. Although explicit contrast-wise reweighting is not applied in the current framework, potential modality bias can be quantified through per-contrast evaluation metrics or conditional error analysis, which we identify as a promising direction for future work.</p>
</sec>
<sec id="j_infor622_s_010">
<label>4.2</label>
<title>Preprocessing</title>
<p>Building on prior research by Han <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor622_ref_015">2024</xref>), the Mask Vector for Training on Multiple Datasets was employed to integrate the two datasets during training. Furthermore, the images were downsampled from <inline-formula id="j_infor622_ineq_057"><alternatives><mml:math>
<mml:mn>1024</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>1024</mml:mn></mml:math><tex-math><![CDATA[$1024\times 1024$]]></tex-math></alternatives></inline-formula> to <inline-formula id="j_infor622_ineq_058"><alternatives><mml:math>
<mml:mn>512</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>512</mml:mn></mml:math><tex-math><![CDATA[$512\times 512$]]></tex-math></alternatives></inline-formula>, subjected to horizontal flipping as a transformation, and normalized to an intensity range of <inline-formula id="j_infor622_ineq_059"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[-1,1]$]]></tex-math></alternatives></inline-formula>, following the methodology by Matviychuk <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor622_ref_036">2016</xref>).</p>
<p>Since some of the compared methods are specifically designed for Gaussian noise, the Anscombe transform (Makitalo and Foi, <xref ref-type="bibr" rid="j_infor622_ref_035">2012</xref>) was applied to ensure a fair comparison by converting Poisson noise into a Gaussian distribution with a variance of 1. Additionally, the images were normalized to the <inline-formula id="j_infor622_ineq_060"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[0,1]$]]></tex-math></alternatives></inline-formula> intensity range, following the approach described by Bodduna and Weickert (<xref ref-type="bibr" rid="j_infor622_ref_003">2019</xref>).</p>
<p>To simulate realistic MRI noise characteristics, Rician noise is introduced during data preprocessing. All images are first normalized to the range <inline-formula id="j_infor622_ineq_061"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[0,1]$]]></tex-math></alternatives></inline-formula>. Complex-valued Gaussian noise is independently added to the real and imaginary components of the clean signal, with noise standard deviation <italic>σ</italic>. The final magnitude image is computed as 
<disp-formula id="j_infor622_eq_008">
<label>(8)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">I</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>rician</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msqrt>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">I</mml:mi>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">r</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>+</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:msqrt>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="1em"/>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">r</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∼</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {I_{\text{rician}}}=\sqrt{{(I+{n_{r}})^{2}}+{n_{i}^{2}}},\hspace{1em}{n_{r}},{n_{i}}\sim \mathcal{N}\big(0,{\sigma ^{2}}\big).\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>The noise level is controlled through the signal-to-noise ratio (SNR), defined as <inline-formula id="j_infor622_ineq_062"><alternatives><mml:math>
<mml:mtext>SNR</mml:mtext>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>signal</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" stretchy="false">/</mml:mo>
<mml:mi mathvariant="italic">σ</mml:mi></mml:math><tex-math><![CDATA[$\text{SNR}={\mu _{\text{signal}}}/\sigma $]]></tex-math></alternatives></inline-formula>. In our experiments, SNR values are uniformly sampled from <inline-formula id="j_infor622_ineq_063"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>15</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[5,15]$]]></tex-math></alternatives></inline-formula> dB for low-dose settings and <inline-formula id="j_infor622_ineq_064"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>20</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>30</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[20,30]$]]></tex-math></alternatives></inline-formula> dB for high-dose settings. Unless otherwise specified, Rician noise is applied only during evaluation to assess robustness under non-Gaussian noise conditions.</p>
</sec>
<sec id="j_infor622_s_011">
<label>4.3</label>
<title>Experiment Settings</title>
<p>All experiments were conducted on an Ubuntu 20.04 system using the PyTorch 1.12.1 framework, Python 3.10.4, and an NVIDIA DGX A100 GPU with CUDA 12.1, powered by an Intel Xeon Platinum 8470Q CPU.</p>
<p>The model was trained for 100 epochs using mixed-precision arithmetic to improve computational efficiency without compromising numerical accuracy. A learning rate of <inline-formula id="j_infor622_ineq_065"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${10^{-4}}$]]></tex-math></alternatives></inline-formula> was employed in conjunction with the AdamW optimizer and a cosine learning rate schedule, enabling rapid initial convergence followed by stable refinement. During training, the diffusion timestep <italic>t</italic> was sampled uniformly from the interval <inline-formula id="j_infor622_ineq_066"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">T</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[1,T]$]]></tex-math></alternatives></inline-formula>, ensuring that the network learns to estimate noise across the full range of diffusion levels. Optimization was performed using the mean squared error loss <inline-formula id="j_infor622_ineq_067"><alternatives><mml:math>
<mml:mi mathvariant="script">L</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mo maxsize="1.19em" minsize="1.19em" fence="true">[</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mo stretchy="false">‖</mml:mo>
<mml:mi mathvariant="italic">η</mml:mi>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">η</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">‖</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo maxsize="1.19em" minsize="1.19em" fence="true">]</mml:mo></mml:math><tex-math><![CDATA[$\mathcal{L}=\mathbb{E}\big[{\| \eta -{\eta _{\theta }}({x_{t}},t)\| ^{2}}\big]$]]></tex-math></alternatives></inline-formula>, which directly supervises the prediction of Gaussian noise and is consistent with the theoretical objective of denoising diffusion probabilistic models.</p>
<p>The diffusion process was configured with <inline-formula id="j_infor622_ineq_068"><alternatives><mml:math>
<mml:mi mathvariant="italic">T</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1000</mml:mn></mml:math><tex-math><![CDATA[$T=1000$]]></tex-math></alternatives></inline-formula> timesteps and a linear noise schedule <inline-formula id="j_infor622_ineq_069"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">∈</mml:mo>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>6</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0.02</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[${\beta _{t}}\in [{10^{-6}},0.02]$]]></tex-math></alternatives></inline-formula>. This configuration ensures a smooth and gradual corruption of the input data, providing sufficient discretization of noise levels to support stable training and reliable reverse process. All input images were normalized to the intensity range <inline-formula id="j_infor622_ineq_070"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[-1,1]$]]></tex-math></alternatives></inline-formula>. During inference, the reverse diffusion process was initialized from the estimated timestep <inline-formula id="j_infor622_ineq_071"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${t^{\prime }}$]]></tex-math></alternatives></inline-formula> and iteratively performed until <inline-formula id="j_infor622_ineq_072"><alternatives><mml:math>
<mml:mi mathvariant="italic">t</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[$t=0$]]></tex-math></alternatives></inline-formula>. To enhance numerical stability and prevent intensity outliers, the predicted mean <inline-formula id="j_infor622_ineq_073"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mu _{0}}$]]></tex-math></alternatives></inline-formula> was clipped to the same normalized range.</p>
<p>The U-Net backbone employed convolutional layers with <inline-formula id="j_infor622_ineq_074"><alternatives><mml:math>
<mml:mn>3</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>3</mml:mn></mml:math><tex-math><![CDATA[$3\times 3$]]></tex-math></alternatives></inline-formula> kernels were used throughout all residual blocks. In addition, Transformer layers with 8 attention heads were incorporated to effectively capture long-range dependencies while maintaining computational efficiency.</p>
<p>To evaluate the practical utility of SynthModDiff in a clinically relevant scenario, we conducted a downstream tumour classification task using synthetic multi-contrast MRI images. Specifically, synthetic target modalities generated by different synthesis methods were used to train a tumour classification model, which was then evaluated on real MRI images.</p>
<p>For the classification network, we employed a ResNet-18 architecture initialized with ImageNet pre-trained weights. The input to the classifier consists of concatenated multi-contrast MRI slices (T1, T2, T1CE, and FLAIR), which are stacked along the channel dimension and resized to <inline-formula id="j_infor622_ineq_075"><alternatives><mml:math>
<mml:mn>224</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>224</mml:mn></mml:math><tex-math><![CDATA[$224\times 224$]]></tex-math></alternatives></inline-formula>. All classifiers were trained from scratch on synthetic images generated by each method, while the test set remained identical and consisted exclusively of real MRI data to ensure a fair comparison.</p>
<p>The training pipeline follows three stages:</p>
<list>
<list-item id="j_infor622_li_003">
<label>1.</label>
<p>Synthetic modality generation using the respective synthesis model (Improved DDPM or SynthModDiff);</p>
</list-item>
<list-item id="j_infor622_li_004">
<label>2.</label>
<p>Multi-contrast image construction by stacking the synthesized modalities;</p>
</list-item>
<list-item id="j_infor622_li_005">
<label>3.</label>
<p>Tumour classification using the ResNet-18 network.</p>
</list-item>
</list>
<p>This design ensures that performance differences in the classification task can be directly attributed to the quality of the synthesized images. The classifier was trained using the cross-entropy loss with the Adam optimizer, a learning rate of <inline-formula id="j_infor622_ineq_076"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo>×</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[$1\times {10^{-4}}$]]></tex-math></alternatives></inline-formula>, and a batch size of 32. Standard data augmentations, including random flipping and rotation, were applied during training.</p>
</sec>
<sec id="j_infor622_s_012">
<label>4.4</label>
<title>Comparative Methods and Evaluation Metrics</title>
<p>In this study, the proposed model is systematically compared with a diverse set of existing models, including Pix2pix (Isola <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_022">2017</xref>), StarGAN (Choi <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_007">2018</xref>), ResUNet (Diakogiannis <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_011">2020</xref>), Multi-cycle GAN (Liu <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_032">2021</xref>), Improved DDPM (Nichol and Dhariwal, <xref ref-type="bibr" rid="j_infor622_ref_038">2021</xref>), Guided DDPM (Dhariwal and Nichol, <xref ref-type="bibr" rid="j_infor622_ref_010">2021</xref>), Hieu-Model (Han <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_015">2024</xref>), and Diffusion Mamba (Wang <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_049">2024</xref>). This comparative analysis aims to assess the performance, effectiveness, and robustness of the proposed approach relative to these established methodologies in the field.</p>
<p>The evaluation of model performance was conducted using three widely recognized metrics in image denoising: Peak Signal-to-Noise Ratio (PSNR), normalized mean absolute error (NMAE), and the Structural Similarity Index Measure (SSIM). PSNR is typically used to quantify the reconstruction quality of an image by measuring the ratio between the maximum possible signal power and the power of the noise that distorts the image. A higher PSNR value generally indicates better visual quality and lower distortion. NMAE, on the other hand, is a normalized measure of the absolute differences between the predicted and ground truth images, hence creating an alternative assessment of visual accuracy. Lastly, SSIM evaluates the structural fidelity of the denoised images by comparing luminance, contrast, and texture similarities with the original image. With these metrics together, a comprehensive assessment of the model’s ability is provided to preserve image quality while effectively removing noise.</p>
</sec>
<sec id="j_infor622_s_013">
<label>4.5</label>
<title>Experiment Results</title>
<sec id="j_infor622_s_014">
<label>4.5.1</label>
<title>Quantitative Evaluation</title>
<table-wrap id="j_infor622_tab_002">
<label>Table 2</label>
<caption>
<p>Quantitative results.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Dataset</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Methods</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">NMAE <inline-formula id="j_infor622_ineq_077"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mo stretchy="false">↓</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\downarrow )$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">PSNR <inline-formula id="j_infor622_ineq_078"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mo stretchy="false">↑</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\uparrow )$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">SSIM <inline-formula id="j_infor622_ineq_079"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mo stretchy="false">↑</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(\uparrow )$]]></tex-math></alternatives></inline-formula></td>
</tr>
</thead>
<tbody>
<tr>
<td colspan="5" style="vertical-align: top; text-align: center; border-bottom: solid thin">High-dose Level</td>
</tr>
</tbody><tbody>
<tr>
<td rowspan="8" style="vertical-align: top; text-align: left">BRATS2020</td>
<td style="vertical-align: top; text-align: left">Pix2pix (Isola <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_022">2017</xref>)</td>
<td style="vertical-align: top; text-align: left">0.045</td>
<td style="vertical-align: top; text-align: left">35.13</td>
<td style="vertical-align: top; text-align: left">0.86</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">StarGAN (Choi <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_007">2018</xref>)</td>
<td style="vertical-align: top; text-align: left">0.043</td>
<td style="vertical-align: top; text-align: left">34.63</td>
<td style="vertical-align: top; text-align: left">0.82</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">ResUNet (Diakogiannis <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_011">2020</xref>)</td>
<td style="vertical-align: top; text-align: left">0.043</td>
<td style="vertical-align: top; text-align: left">35.33</td>
<td style="vertical-align: top; text-align: left">0.77</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Multi-cycle GAN (Liu <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_032">2021</xref>)</td>
<td style="vertical-align: top; text-align: left">0.050</td>
<td style="vertical-align: top; text-align: left">34.75</td>
<td style="vertical-align: top; text-align: left">0.84</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Improved DDPM (Nichol and Dhariwal, <xref ref-type="bibr" rid="j_infor622_ref_038">2021</xref>)</td>
<td style="vertical-align: top; text-align: left">0.040</td>
<td style="vertical-align: top; text-align: left">36.69</td>
<td style="vertical-align: top; text-align: left"><underline>0.89</underline></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Guided DDPM (Dhariwal and Nichol, <xref ref-type="bibr" rid="j_infor622_ref_010">2021</xref>)</td>
<td style="vertical-align: top; text-align: left">0.042</td>
<td style="vertical-align: top; text-align: left">36.01</td>
<td style="vertical-align: top; text-align: left">0.88</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Hieu-Model (Han <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_015">2024</xref>)</td>
<td style="vertical-align: top; text-align: left">0.041</td>
<td style="vertical-align: top; text-align: left">35.85</td>
<td style="vertical-align: top; text-align: left">0.82</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">SynthModDiff - G</td>
<td style="vertical-align: top; text-align: left"><underline>0.029</underline></td>
<td style="vertical-align: top; text-align: left"><bold>37.27</bold></td>
<td style="vertical-align: top; text-align: left"><underline>0.89</underline></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">SynthModDiff - P</td>
<td style="vertical-align: top; text-align: left"><bold>0.028</bold></td>
<td style="vertical-align: top; text-align: left"><underline>36.89</underline></td>
<td style="vertical-align: top; text-align: left"><bold>0.91</bold></td>
</tr>
<tr>
<td rowspan="8" style="vertical-align: top; text-align: left">IXI</td>
<td style="vertical-align: top; text-align: left">Pix2pix (Isola <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_022">2017</xref>)</td>
<td style="vertical-align: top; text-align: left">0.042</td>
<td style="vertical-align: top; text-align: left">36.88</td>
<td style="vertical-align: top; text-align: left">0.72</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">StarGAN (Choi <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_007">2018</xref>)</td>
<td style="vertical-align: top; text-align: left">0.039</td>
<td style="vertical-align: top; text-align: left">40.47</td>
<td style="vertical-align: top; text-align: left">0.73</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">ResUNet (Diakogiannis <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_011">2020</xref>)</td>
<td style="vertical-align: top; text-align: left">0.039</td>
<td style="vertical-align: top; text-align: left">40.85</td>
<td style="vertical-align: top; text-align: left">0.69</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Multi-cycle GAN (Liu <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_032">2021</xref>)</td>
<td style="vertical-align: top; text-align: left">0.044</td>
<td style="vertical-align: top; text-align: left">33.48</td>
<td style="vertical-align: top; text-align: left">0.68</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Improved DDPM (Nichol and Dhariwal, <xref ref-type="bibr" rid="j_infor622_ref_038">2021</xref>)</td>
<td style="vertical-align: top; text-align: left">0.037</td>
<td style="vertical-align: top; text-align: left">40.96</td>
<td style="vertical-align: top; text-align: left">0.73</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Guided DDPM (Dhariwal and Nichol, <xref ref-type="bibr" rid="j_infor622_ref_010">2021</xref>)</td>
<td style="vertical-align: top; text-align: left">0.038</td>
<td style="vertical-align: top; text-align: left">40.24</td>
<td style="vertical-align: top; text-align: left">0.72</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Hieu-Model (Han <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_015">2024</xref>)</td>
<td style="vertical-align: top; text-align: left">0.038</td>
<td style="vertical-align: top; text-align: left">40.88</td>
<td style="vertical-align: top; text-align: left">0.70</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">SynthModDiff - G</td>
<td style="vertical-align: top; text-align: left"><bold>0.025</bold></td>
<td style="vertical-align: top; text-align: left"><underline>40.97</underline></td>
<td style="vertical-align: top; text-align: left"><underline>0.78</underline></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">SynthModDiff - P</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><underline>0.026</underline></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><bold>41.93</bold></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><bold>0.79</bold></td>
</tr>
</tbody><tbody>
<tr>
<td colspan="5" style="vertical-align: top; text-align: center; border-bottom: solid thin">Low-dose Level</td>
</tr>
</tbody><tbody>
<tr>
<td rowspan="8" style="vertical-align: top; text-align: left">BRATS2020</td>
<td style="vertical-align: top; text-align: left">Pix2pix (Isola <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_022">2017</xref>)</td>
<td style="vertical-align: top; text-align: left">0.055</td>
<td style="vertical-align: top; text-align: left">32.50</td>
<td style="vertical-align: top; text-align: left">0.78</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">StarGAN (Choi <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_007">2018</xref>)</td>
<td style="vertical-align: top; text-align: left">0.052</td>
<td style="vertical-align: top; text-align: left">32.10</td>
<td style="vertical-align: top; text-align: left">0.74</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">ResUNet (Diakogiannis <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_011">2020</xref>)</td>
<td style="vertical-align: top; text-align: left">0.052</td>
<td style="vertical-align: top; text-align: left">32.75</td>
<td style="vertical-align: top; text-align: left">0.69</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Multi-cycle GAN (Liu <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_032">2021</xref>)</td>
<td style="vertical-align: top; text-align: left">0.058</td>
<td style="vertical-align: top; text-align: left">32.20</td>
<td style="vertical-align: top; text-align: left">0.76</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Improved DDPM (Nichol and Dhariwal, <xref ref-type="bibr" rid="j_infor622_ref_038">2021</xref>)</td>
<td style="vertical-align: top; text-align: left">0.048</td>
<td style="vertical-align: top; text-align: left">34.10</td>
<td style="vertical-align: top; text-align: left">0.81</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Guided DDPM (Dhariwal and Nichol, <xref ref-type="bibr" rid="j_infor622_ref_010">2021</xref>)</td>
<td style="vertical-align: top; text-align: left">0.050</td>
<td style="vertical-align: top; text-align: left">33.50</td>
<td style="vertical-align: top; text-align: left">0.82</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Hieu-Model (Han <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_015">2024</xref>)</td>
<td style="vertical-align: top; text-align: left">0.049</td>
<td style="vertical-align: top; text-align: left">33.40</td>
<td style="vertical-align: top; text-align: left">0.75</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">SynthModDiff - G</td>
<td style="vertical-align: top; text-align: left"><underline>0.038</underline></td>
<td style="vertical-align: top; text-align: left"><bold>35.10</bold></td>
<td style="vertical-align: top; text-align: left"><bold>0.85</bold></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">SynthModDiff - P</td>
<td style="vertical-align: top; text-align: left"><bold>0.037</bold></td>
<td style="vertical-align: top; text-align: left"><underline>34.85</underline></td>
<td style="vertical-align: top; text-align: left"><underline>0.83</underline></td>
</tr>
<tr>
<td rowspan="8" style="vertical-align: top; text-align: left">IXI</td>
<td style="vertical-align: top; text-align: left">Pix2pix (Isola <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_022">2017</xref>)</td>
<td style="vertical-align: top; text-align: left">0.050</td>
<td style="vertical-align: top; text-align: left">34.50</td>
<td style="vertical-align: top; text-align: left">0.65</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">StarGAN (Choi <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_007">2018</xref>)</td>
<td style="vertical-align: top; text-align: left">0.047</td>
<td style="vertical-align: top; text-align: left">37.80</td>
<td style="vertical-align: top; text-align: left">0.66</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">ResUNet (Diakogiannis <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_011">2020</xref>)</td>
<td style="vertical-align: top; text-align: left">0.047</td>
<td style="vertical-align: top; text-align: left">38.00</td>
<td style="vertical-align: top; text-align: left">0.62</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Multi-cycle GAN (Liu <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_032">2021</xref>)</td>
<td style="vertical-align: top; text-align: left">0.052</td>
<td style="vertical-align: top; text-align: left">30.90</td>
<td style="vertical-align: top; text-align: left">0.60</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Improved DDPM (Nichol and Dhariwal, <xref ref-type="bibr" rid="j_infor622_ref_038">2021</xref>)</td>
<td style="vertical-align: top; text-align: left">0.045</td>
<td style="vertical-align: top; text-align: left">38.20</td>
<td style="vertical-align: top; text-align: left">0.67</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Guided DDPM (Dhariwal and Nichol, <xref ref-type="bibr" rid="j_infor622_ref_010">2021</xref>)</td>
<td style="vertical-align: top; text-align: left">0.046</td>
<td style="vertical-align: top; text-align: left">37.50</td>
<td style="vertical-align: top; text-align: left">0.66</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Hieu-Model (Han <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor622_ref_015">2024</xref>)</td>
<td style="vertical-align: top; text-align: left">0.046</td>
<td style="vertical-align: top; text-align: left">38.00</td>
<td style="vertical-align: top; text-align: left">0.63</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">SynthModDiff - G</td>
<td style="vertical-align: top; text-align: left"><underline>0.035</underline></td>
<td style="vertical-align: top; text-align: left"><underline>38.60</underline></td>
<td style="vertical-align: top; text-align: left"><underline>0.72</underline></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">SynthModDiff - P</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><bold>0.034</bold></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><bold>39.50</bold></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><bold>0.73</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>SynthModDiff - G and SynthModDiff - P represent variations of SynthModDiff, where Gaussian and Poisson noise are applied, respectively. The best values are in <bold>bold</bold>, and the second-best values are <underline>underlined</underline>.</p>
</table-wrap-foot>
</table-wrap>
<p>The quantitative evaluation is provided in detail in Table <xref rid="j_infor622_tab_002">2</xref>. For High-dose levels, the performance of SynthModDiff - G and SynthModDiff - P on MRI modality conversion tasks proves their prominence on existing methods with both the BRATS2020 and IXI datasets. In the BRATS2020 dataset, SynthModDiff - P achieves the lowest NMAE (0.028) and the highest SSIM (0.91), proving its strong capability to preserve structural integrity while minimizing errors. SynthModDiff - G, while slightly trailing in SSIM, attains the highest PSNR (37.27 dB) among all models, suggesting that it excels in maintaining image quality and reducing distortion. Both models significantly outperform previous state-of-the-art approaches, such as Improved DDPM (Nichol and Dhariwal, <xref ref-type="bibr" rid="j_infor622_ref_038">2021</xref>) and Guided DDPM (Dhariwal and Nichol, <xref ref-type="bibr" rid="j_infor622_ref_010">2021</xref>), which achieve lower PSNR and higher NMAE, indicating that the proposed methods are more effective in reducing noise while preserving fine image details.</p>
<p>On the IXI dataset, SynthModDiff - P continues to demonstrate strong performance by achieving the highest PSNR (41.93 dB) and SSIM (0.79). This proves its ability to generate high-quality images with strong structural similarity to the ground truth. SynthModDiff - G, on the other hand, achieves the lowest NMAE (0.025), meaning it produces the most accurate pixel-wise reconstructions. While its PSNR (40.97 dB) and SSIM (0.78) are slightly lower than those of SynthModDiff - P, they still surpass all other models in the comparison. This demonstrates the effectiveness of SynthModDiff - G on both noise-based variations.</p>
<p>Regarding low-dose levels, SynthModDiff shows strong performance despite the increased noise and reduced signal quality inherent in low-dose MRI. For the BRATS2020 and IXI datasets, SynthModDiff still achieves the lowest NMAE, PSNR, and SSIM, indicating its effectiveness in minimizing reconstruction errors.</p>
<p>Overall, both SynthModDiff - G and SynthModDiff - P outperform existing diffusion-based and GAN-based methods in MRI modality conversion. SynthModDiff - P is particularly effective in preserving structural details, making it the preferred choice for tasks where maintaining fine textures is crucial. Meanwhile, SynthModDiff - G achieves the lowest NMAE, making it highly accurate in terms of pixel-wise error minimization. These results highlight the benefits of incorporating Gaussian and Poisson noise variations in model training, ultimately leading to improved performance in MRI image transformation tasks.</p>
</sec>
<sec id="j_infor622_s_015">
<label>4.5.2</label>
<title>Qualitative Evaluation</title>
<fig id="j_infor622_fig_002">
<label>Fig. 2</label>
<caption>
<p>Qualitative comparison of synthetic modality generation (T2, T1CE, and FLAIR) from a T1 input image on the BraTS2020 dataset. Columns from left to right show: input T1 image, synthesized target modalities. Rows from top to bottom show: ground truth, Improved DDPM, and the proposed SynthModDiff. Results are shown for both synthesized high-dose T1 (upper block) and synthesized low-dose T1 (lower block).</p>
</caption>
<graphic xlink:href="infor622_g002.jpg"/>
</fig>
<fig id="j_infor622_fig_003">
<label>Fig. 3</label>
<caption>
<p>Qualitative comparison of synthetic modality generation (T2, PD, and MRA) from a T1 input image on the IXI dataset. The figure follows the same layout as Fig. <xref rid="j_infor622_fig_002">2</xref>, enabling a direct comparison between ground truth, Improved DDPM, and SynthModDiff under different dose conditions.</p>
</caption>
<graphic xlink:href="infor622_g003.jpg"/>
</fig>
<p>Figures <xref rid="j_infor622_fig_002">2</xref> and <xref rid="j_infor622_fig_003">3</xref> illustrate qualitative results for the task of synthetic modality generation from a single T1-weighted input image. In both figures, the first column shows the input T1 image, while the subsequent columns correspond to the synthesized target modalities, namely T2, T1CE, and FLAIR for BraTS2020 (Fig. <xref rid="j_infor622_fig_002">2</xref>), and T2, PD, and MRA for the IXI dataset (Fig. <xref rid="j_infor622_fig_003">3</xref>).</p>
<p>For each target modality, the first row represents the ground truth image acquired at the corresponding modality. The second and third rows present the synthesis results obtained using the second-best baseline method (Improved DDPM) and the proposed SynthModDiff method, respectively. This layout is repeated for two different input conditions: synthesized high-dose T1 (upper block) and synthesized low-dose T1 (lower block), allowing a direct comparison of robustness under varying noise levels.</p>
<p>As the dose level decreases, both methods exhibit some degradation due to increased noise in the input T1 image. However, Improved DDPM tends to produce overly smoothed results and fails to preserve modality-specific contrast, leading to blurred anatomical boundaries and reduced visibility of fine structures. In contrast, SynthModDiff consistently generates target modalities that are visually closer to the ground truth, preserving critical anatomical details and contrast patterns across different modalities. Notably, structures such as tissue boundaries, lesion regions, and vascular or bone-related details remain more distinguishable in SynthModDiff results, even under low-dose conditions.</p>
<p>These observations demonstrate that SynthModDiff is more robust for synthetic modality generation, particularly when the input modality is affected by reduced dose levels, which is crucial for practical clinical scenarios where minimizing radiation exposure is essential.</p>
</sec>
<sec id="j_infor622_s_016">
<label>4.5.3</label>
<title>Ablation Study</title>
<table-wrap id="j_infor622_tab_003">
<label>Table 3</label>
<caption>
<p>Ablation study on the effect of noise schedules on SynthModDiff.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Dataset</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Schedule</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">NMAE ↓</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">PSNR ↑</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">SSIM ↑</td>
</tr>
</thead>
<tbody>
<tr>
<td rowspan="2" style="vertical-align: middle; text-align: left">BRATS2020 (Low-dose)</td>
<td style="vertical-align: top; text-align: left">Linear</td>
<td style="vertical-align: top; text-align: left">0.037</td>
<td style="vertical-align: top; text-align: left">34.85</td>
<td style="vertical-align: top; text-align: left">0.83</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Cosine</td>
<td style="vertical-align: top; text-align: left"><bold>0.034</bold></td>
<td style="vertical-align: top; text-align: left"><bold>35.60</bold></td>
<td style="vertical-align: top; text-align: left"><bold>0.86</bold></td>
</tr>
<tr>
<td rowspan="2" style="vertical-align: middle; text-align: left">IXI (Low-dose)</td>
<td style="vertical-align: top; text-align: left">Linear</td>
<td style="vertical-align: top; text-align: left">0.034</td>
<td style="vertical-align: top; text-align: left">39.50</td>
<td style="vertical-align: top; text-align: left">0.73</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Cosine</td>
<td style="vertical-align: top; text-align: left"><bold>0.031</bold></td>
<td style="vertical-align: top; text-align: left"><bold>40.20</bold></td>
<td style="vertical-align: top; text-align: left"><bold>0.76</bold></td>
</tr>
<tr>
<td rowspan="2" style="vertical-align: middle; text-align: left">BRATS2020 (High-dose)</td>
<td style="vertical-align: top; text-align: left">Linear</td>
<td style="vertical-align: top; text-align: left">0.028</td>
<td style="vertical-align: top; text-align: left">36.89</td>
<td style="vertical-align: top; text-align: left">0.91</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Cosine</td>
<td style="vertical-align: top; text-align: left"><bold>0.026</bold></td>
<td style="vertical-align: top; text-align: left"><bold>37.40</bold></td>
<td style="vertical-align: top; text-align: left"><bold>0.92</bold></td>
</tr>
<tr>
<td rowspan="2" style="vertical-align: middle; text-align: left; border-bottom: solid thin">IXI (High-dose)</td>
<td style="vertical-align: top; text-align: left">Linear</td>
<td style="vertical-align: top; text-align: left">0.026</td>
<td style="vertical-align: top; text-align: left">41.93</td>
<td style="vertical-align: top; text-align: left">0.79</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">Cosine</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><bold>0.024</bold></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><bold>42.60</bold></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><bold>0.81</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To analyse the impact of different noise schedules, we conducted an ablation study comparing the commonly used linear schedule with a cosine noise schedule within the proposed SynthModDiff framework, as summarized in Table <xref rid="j_infor622_tab_003">3</xref>. The cosine schedule consistently improves performance across all datasets and dose levels, yielding lower NMAE and higher PSNR and SSIM values.</p>
<p>This improvement can be attributed to the smoother noise allocation in early diffusion timesteps, which stabilizes training and preserves high-frequency anatomical details, particularly under low-dose conditions. Notably, the performance gain is more pronounced for low-dose inputs, as evidenced by the larger improvements reported in Table <xref rid="j_infor622_tab_003">3</xref>, indicating enhanced robustness to severe noise corruption. Based on these observations, the cosine schedule was selected for the final configuration of SynthModDiff.</p>
<table-wrap id="j_infor622_tab_004">
<label>Table 4</label>
<caption>
<p>Ablation study on the effect of the Transformer layer in SynthModDiff.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Dataset</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Architecture</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">NMAE ↓</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">PSNR ↑</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">SSIM ↑</td>
</tr>
</thead>
<tbody>
<tr>
<td rowspan="2" style="vertical-align: middle; text-align: left">BRATS2020 (Low-dose)</td>
<td style="vertical-align: top; text-align: left">w/o Transformer</td>
<td style="vertical-align: top; text-align: left">0.041</td>
<td style="vertical-align: top; text-align: left">33.90</td>
<td style="vertical-align: top; text-align: left">0.80</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">w/ Transformer</td>
<td style="vertical-align: top; text-align: left"><bold>0.037</bold></td>
<td style="vertical-align: top; text-align: left"><bold>34.85</bold></td>
<td style="vertical-align: top; text-align: left"><bold>0.83</bold></td>
</tr>
<tr>
<td rowspan="2" style="vertical-align: middle; text-align: left">IXI (Low-dose)</td>
<td style="vertical-align: top; text-align: left">w/o Transformer</td>
<td style="vertical-align: top; text-align: left">0.039</td>
<td style="vertical-align: top; text-align: left">38.60</td>
<td style="vertical-align: top; text-align: left">0.70</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">w/ Transformer</td>
<td style="vertical-align: top; text-align: left"><bold>0.034</bold></td>
<td style="vertical-align: top; text-align: left"><bold>39.50</bold></td>
<td style="vertical-align: top; text-align: left"><bold>0.73</bold></td>
</tr>
<tr>
<td rowspan="2" style="vertical-align: middle; text-align: left">BRATS2020 (High-dose)</td>
<td style="vertical-align: top; text-align: left">w/o Transformer</td>
<td style="vertical-align: top; text-align: left">0.031</td>
<td style="vertical-align: top; text-align: left">35.90</td>
<td style="vertical-align: top; text-align: left">0.88</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">w/ Transformer</td>
<td style="vertical-align: top; text-align: left"><bold>0.028</bold></td>
<td style="vertical-align: top; text-align: left"><bold>36.89</bold></td>
<td style="vertical-align: top; text-align: left"><bold>0.91</bold></td>
</tr>
<tr>
<td rowspan="2" style="vertical-align: middle; text-align: left; border-bottom: solid thin">IXI (High-dose)</td>
<td style="vertical-align: top; text-align: left">w/o Transformer</td>
<td style="vertical-align: top; text-align: left">0.029</td>
<td style="vertical-align: top; text-align: left">40.80</td>
<td style="vertical-align: top; text-align: left">0.76</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">w/ Transformer</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><bold>0.026</bold></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><bold>41.93</bold></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><bold>0.79</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To evaluate the role of the Transformer layer in capturing long-range dependencies, we conducted an ablation study by removing the Transformer blocks from the SynthModDiff architecture while keeping all other components unchanged. As shown in Table <xref rid="j_infor622_tab_004">4</xref>, the absence of the Transformer consistently degrades performance across all datasets and dose levels.</p>
<p>The performance drop is particularly noticeable in SSIM, indicating reduced structural consistency and weaker preservation of global anatomical patterns. This effect is more pronounced under low-dose conditions, where long-range contextual information becomes critical for compensating severe noise corruption. By integrating self-attention mechanisms, the Transformer layer enables the model to capture long-range spatial dependencies across the entire image, leading to more coherent contrast synthesis and improved anatomical fidelity.</p>
<table-wrap id="j_infor622_tab_005">
<label>Table 5</label>
<caption>
<p>Ablation study on the impact of different noise models on SynthModDiff.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Dataset</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Noise model</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">NMAE ↓</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">PSNR ↑</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">SSIM ↑</td>
</tr>
</thead>
<tbody>
<tr>
<td rowspan="3" style="vertical-align: middle; text-align: left">BRATS2020 (Low-dose)</td>
<td style="vertical-align: top; text-align: left">Gaussian</td>
<td style="vertical-align: top; text-align: left"><bold>0.037</bold></td>
<td style="vertical-align: top; text-align: left"><bold>34.85</bold></td>
<td style="vertical-align: top; text-align: left"><bold>0.83</bold></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Poisson</td>
<td style="vertical-align: top; text-align: left"><underline>0.038</underline></td>
<td style="vertical-align: top; text-align: left"><underline>34.50</underline></td>
<td style="vertical-align: top; text-align: left"><underline>0.82</underline></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Rician</td>
<td style="vertical-align: top; text-align: left">0.041</td>
<td style="vertical-align: top; text-align: left">33.90</td>
<td style="vertical-align: top; text-align: left">0.80</td>
</tr>
<tr>
<td rowspan="3" style="vertical-align: middle; text-align: left">IXI (Low-dose)</td>
<td style="vertical-align: top; text-align: left">Gaussian</td>
<td style="vertical-align: top; text-align: left"><bold>0.034</bold></td>
<td style="vertical-align: top; text-align: left"><bold>39.50</bold></td>
<td style="vertical-align: top; text-align: left"><bold>0.73</bold></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Poisson</td>
<td style="vertical-align: top; text-align: left"><underline>0.035</underline></td>
<td style="vertical-align: top; text-align: left"><underline>39.10</underline></td>
<td style="vertical-align: top; text-align: left"><underline>0.72</underline></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Rician</td>
<td style="vertical-align: top; text-align: left">0.038</td>
<td style="vertical-align: top; text-align: left">38.40</td>
<td style="vertical-align: top; text-align: left">0.70</td>
</tr>
<tr>
<td rowspan="3" style="vertical-align: middle; text-align: left">BRATS2020 (High-dose)</td>
<td style="vertical-align: top; text-align: left">Gaussian</td>
<td style="vertical-align: top; text-align: left"><bold>0.028</bold></td>
<td style="vertical-align: top; text-align: left"><bold>36.89</bold></td>
<td style="vertical-align: top; text-align: left"><bold>0.91</bold></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Poisson</td>
<td style="vertical-align: top; text-align: left"><underline>0.029</underline></td>
<td style="vertical-align: top; text-align: left"><underline>36.40</underline></td>
<td style="vertical-align: top; text-align: left"><underline>0.90</underline></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Rician</td>
<td style="vertical-align: top; text-align: left">0.030</td>
<td style="vertical-align: top; text-align: left">36.10</td>
<td style="vertical-align: top; text-align: left">0.89</td>
</tr>
<tr>
<td rowspan="3" style="vertical-align: middle; text-align: left; border-bottom: solid thin">IXI (High-dose)</td>
<td style="vertical-align: top; text-align: left">Gaussian</td>
<td style="vertical-align: top; text-align: left"><bold>0.026</bold></td>
<td style="vertical-align: top; text-align: left"><bold>41.93</bold></td>
<td style="vertical-align: top; text-align: left"><bold>0.79</bold></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Poisson</td>
<td style="vertical-align: top; text-align: left"><underline>0.027</underline></td>
<td style="vertical-align: top; text-align: left"><underline>41.30</underline></td>
<td style="vertical-align: top; text-align: left"><underline>0.78</underline></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">Rician</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.028</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">40.80</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.77</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To evaluate the robustness of SynthModDiff under different noise characteristics, we conducted an ablation study comparing Gaussian, Poisson, and Rician noise models. As shown in Table <xref rid="j_infor622_tab_005">5</xref>, SynthModDiff achieves the best performance under Gaussian noise, which aligns with the training distribution, followed by Poisson noise. Under Rician noise, which is common in low-SNR MRI magnitude images, a moderate performance degradation is observed, particularly under low-dose conditions.</p>
<p>Nevertheless, the performance drop remains limited, indicating that SynthModDiff retains strong robustness to signal-dependent and non-Gaussian noise. This suggests that the proposed diffusion-based framework generalizes reasonably well beyond the noise distributions explicitly used during training, which is critical for real-world MRI applications.</p>
</sec>
<sec id="j_infor622_s_017">
<label>4.5.4</label>
<title>Discussion</title>
<table-wrap id="j_infor622_tab_006">
<label>Table 6</label>
<caption>
<p>Downstream tumor classification performance using synthetic multi-contrast MRI images.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Training data</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Accuracy ↑</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">AUC ↑</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">F1-score ↑</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">Real MRI (Upper Bound)</td>
<td style="vertical-align: top; text-align: left">89.6</td>
<td style="vertical-align: top; text-align: left">0.92</td>
<td style="vertical-align: top; text-align: left">0.88</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Improved DDPM (Synthetic)</td>
<td style="vertical-align: top; text-align: left">82.3</td>
<td style="vertical-align: top; text-align: left">0.86</td>
<td style="vertical-align: top; text-align: left">0.81</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">SynthModDiff - G (Synthetic)</td>
<td style="vertical-align: top; text-align: left"><underline>85.7</underline></td>
<td style="vertical-align: top; text-align: left"><underline>0.89</underline></td>
<td style="vertical-align: top; text-align: left"><underline>0.84</underline></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">SynthModDiff - P (Synthetic)</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><bold>87.1</bold></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><bold>0.91</bold></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><bold>0.86</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To evaluate the practical impact of SynthModDiff, we further conducted a downstream tumour classification task using synthetic multi-contrast MRI images. As shown in Table <xref rid="j_infor622_tab_006">6</xref>, classifiers trained on SynthModDiff-generated images consistently outperform those trained on synthetic images produced by Improved DDPM. The performance gain can be attributed to the improved contrast consistency and preservation of global anatomical structures in SynthModDiff outputs. These properties are crucial for classification tasks, which rely on coherent multi-modal representations rather than local pixel-level fidelity alone. Notably, the Poisson-aware variant of SynthModDiff achieves performance close to the upper bound obtained using real MRI data, demonstrating that higher perceptual image quality translates into meaningful improvements in downstream clinical tasks.</p>
<p>The proposed two-stage diffusion design explicitly separates noise-level alignment from the reverse denoising process, enabling robust handling of heterogeneous and unknown noise conditions commonly encountered in clinical data. While a single-stage diffusion process could be adopted when the input noise level is known or accurately estimated, such simplification assumes controlled acquisition settings and consistent noise statistics. In contrast, the two-stage formulation provides greater flexibility and stability, particularly for non-Gaussian or signal-dependent noise, at the cost of additional computational overhead. This trade-off highlights an important design choice between efficiency and robustness in practical medical imaging applications.</p>
<p>Although SynthModDiff is implemented in a 2D slice-wise manner, the proposed framework is not inherently limited to 2D data. In principle, the method can be extended to 3D volumetric synthesis by replacing 2D convolutional and attention operations with their 3D counterparts, enabling direct modelling of inter-slice dependencies. However, full 3D diffusion models significantly increase computational and memory requirements, particularly for high-resolution medical volumes. The current 2D design offers a practical trade-off between performance and efficiency, while future work will explore hybrid or fully 3D implementations to further leverage volumetric contextual information.</p>
<p>While SynthModDiff demonstrates strong performance under the noise conditions considered in this study, several practical factors may influence its generalization. Mild motion-related artifacts can be partially handled due to the stochastic denoising nature of the diffusion process; however, severe motion corruption introduces structured degradations that are not explicitly modelled during training and may require dedicated motion-aware augmentation strategies. Domain shifts arising from different MRI vendors, acquisition protocols, or field strengths can also affect synthesis quality by altering intensity distributions and noise characteristics. Such effects may be mitigated through multi-site training, intensity normalization, or conditioning on acquisition metadata. In extreme low-dose scenarios where noise dominates the signal, the model is still able to recover coarse anatomical structures, although fine details may be degraded due to insufficient signal content. Furthermore, as SynthModDiff learns a data-driven anatomical prior without explicit lesion annotations, it can generalize to previously unseen pathological patterns to some extent, but rare or out-of-distribution abnormalities may be partially smoothed. Finally, the proposed framework is not inherently limited to brain MRI and can be extended to other anatomical regions, such as abdominal or musculoskeletal imaging, by retraining or adapting the model to domain-specific characteristics.</p>
<p>SynthModDiff follows a probabilistic diffusion formulation, which implicitly captures uncertainty through stochastic sampling during the reverse diffusion process. By generating multiple samples from the same input, the variability across synthesized outputs can be interpreted as an indicator of model uncertainty. However, the current framework does not explicitly compute calibrated uncertainty measures, such as voxel-wise confidence intervals, as commonly done in Bayesian or variational probabilistic models. While our focus in this work is on synthesis fidelity and robustness, integrating explicit uncertainty estimation mechanisms represents an important direction for future research, particularly for safety-critical clinical applications.</p>
<p>Despite the strong quantitative performance demonstrated in this study, several limitations should be acknowledged. First, the evaluation was primarily based on objective image quality metrics (e.g. PSNR and SSIM), which may not fully capture subtle artifacts or reflect diagnostic reliability in real clinical settings. Second, no formal study involving radiologists was conducted. Therefore, the clinical interpretability and diagnostic impact of the proposed method remain to be further validated. Future work will include expert-based evaluations and more comprehensive clinical validation to address these limitations.</p>
</sec>
</sec>
</sec>
<sec id="j_infor622_s_018">
<label>5</label>
<title>Conclusion</title>
<p>In this research, SynthModDiff, a novel Synthetic Modality Diffusion framework, was proposed to enhance MRI image synthesis by using denoising diffusion probabilistic models (DDPMs) for improved noise reduction and image fidelity. The proposed two-stage approach, including a noise-aware Forward Process, a Deep Learning network, and a Reverse Process, ensures high-quality reconstruction. The experimental results have shown that SynthModDiff has superior performance on different datasets, with SynthModDiff-P and SynthModDiff-G achieving state-of-the-art performance in NMAE, SSIM, and PSNR metrics, thereby showing their robustness in both high-dose and low-dose imaging scenarios. Qualitative assessments further demonstrate the framework’s effectiveness in preserving anatomical structures with greater clarity compared to existing methods. These findings prove SynthModDiff’s potential for clinical applications, especially in radiotherapy treatment planning, where MRI acquisition is often constrained by scan duration and motion artifacts. Future research will focus on optimizing computational efficiency, expanding modality integration, and validating the approach in broader clinical contexts to further enhance the reliability of synthesized MRI images for accurate diagnosis and treatment planning.</p>
</sec>
</body>
<back>
<ack id="j_infor622_ack_001">
<title>Acknowledgements</title>
<p>This work was funded by The University of Danang, University of Science and Technology.</p></ack>
<ref-list id="j_infor622_reflist_001">
<title>References</title>
<ref id="j_infor622_ref_001">
<mixed-citation publication-type="journal"><string-name><surname>Ali</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Biswas</surname>, <given-names>M.R.</given-names></string-name>, <string-name><surname>Mohsen</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Shah</surname>, <given-names>U.</given-names></string-name>, <string-name><surname>Alamgir</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Mousa</surname>, <given-names>O.</given-names></string-name>, <string-name><surname>Shah</surname>, <given-names>Z.</given-names></string-name> (<year>2022</year>). <article-title>The role of generative adversarial networks in brain MRI: a scoping review</article-title>. <source>Insights Imaging</source>, <volume>13</volume>(<issue>1</issue>), <fpage>98</fpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_002">
<mixed-citation publication-type="journal"><string-name><surname>Berthelon</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Chenegros</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Finateu</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Ieng</surname>, <given-names>S.-H.</given-names></string-name>, <string-name><surname>Benosman</surname>, <given-names>R.</given-names></string-name> (<year>2018</year>). <article-title>Effects of cooling on the SNR and contrast detection of a low-light event-based camera</article-title>. <source>IEEE Transactions on Biomedical Circuits and Systems</source>, <volume>12</volume>(<issue>6</issue>), <fpage>1467</fpage>–<lpage>1474</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_003">
<mixed-citation publication-type="chapter"><string-name><surname>Bodduna</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Weickert</surname>, <given-names>J.</given-names></string-name> (<year>2019</year>). <chapter-title>Poisson noise removal using multi-frame 3D block matching</chapter-title>. In: <source>2019 8th European Workshop on Visual Information Processing (EUVIP)</source>. <publisher-name>IEEE</publisher-name>, pp. <fpage>58</fpage>–<lpage>63</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_004">
<mixed-citation publication-type="journal"><string-name><surname>Chang</surname>, <given-names>C.-W.</given-names></string-name>, <string-name><surname>Dinh</surname>, <given-names>N.T.</given-names></string-name> (<year>2019</year>). <article-title>Classification of machine learning frameworks for data-driven thermal fluid models</article-title>. <source>International Journal of Thermal Sciences</source>, <volume>135</volume>, <fpage>559</fpage>–<lpage>579</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_005">
<mixed-citation publication-type="other"><string-name><surname>Chang</surname>, <given-names>C.-W.</given-names></string-name>, <string-name><surname>Marants</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Gao</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Goette</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Scholey</surname>, <given-names>J.E.</given-names></string-name>, <string-name><surname>Bradley</surname>, <given-names>J.D.</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Zhou</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Sudhyadhom</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>X.</given-names></string-name> (2022). Multimodal imaging-based material mass density estimation for proton therapy using physics-constrained deep learning. arXiv preprint arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2207.13150">2207.13150</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_006">
<mixed-citation publication-type="other"><string-name><surname>Chen</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Wu</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Li</surname>, <given-names>Z.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Yao</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Song</surname>, <given-names>X.</given-names></string-name> (2023). Compound attention and neighbor matching network for multi-contrast MRI super-resolution. arXiv preprint arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2307.02148">2307.02148</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_007">
<mixed-citation publication-type="chapter"><string-name><surname>Choi</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Choi</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Kim</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Ha</surname>, <given-names>J.-W.</given-names></string-name>, <string-name><surname>Kim</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Choo</surname>, <given-names>J.</given-names></string-name> (<year>2018</year>). <chapter-title>StarGAN: unified generative adversarial networks for multi-domain image-to-image translation</chapter-title>. In: <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>, pp. <fpage>8789</fpage>–<lpage>8797</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_008">
<mixed-citation publication-type="chapter"><string-name><surname>Dayarathna</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Peiris</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Islam</surname>, <given-names>K.T.</given-names></string-name>, <string-name><surname>Wong</surname>, <given-names>T.-T.</given-names></string-name>, <string-name><surname>Chen</surname>, <given-names>Z.</given-names></string-name> (<year>2025</year>a). <chapter-title>D2Diff: a dual-domain diffusion model for accurate multi-contrast MRI synthesis</chapter-title>. In: <source>International Conference on Medical Image Computing and Computer-Assisted Intervention</source>. <publisher-name>Springer</publisher-name>, pp. <fpage>131</fpage>–<lpage>141</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_009">
<mixed-citation publication-type="journal"><string-name><surname>Dayarathna</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Wu</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Cai</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Wong</surname>, <given-names>T.-T.</given-names></string-name>, <string-name><surname>Law</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Islam</surname>, <given-names>K.T.</given-names></string-name>, <string-name><surname>Peiris</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Chen</surname>, <given-names>Z.</given-names></string-name> (<year>2025</year>b). <article-title>MU-Diff: a mutual learning diffusion model for synthetic MRI with Application for brain lesions</article-title>. <source>npj Artificial Intelligence</source>, <volume>1</volume>, <fpage>11</fpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_010">
<mixed-citation publication-type="journal"><string-name><surname>Dhariwal</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Nichol</surname>, <given-names>A.</given-names></string-name> (<year>2021</year>). <article-title>Diffusion models beat GANs on image synthesis</article-title>. <source>Advances in Neural Information Processing Systems</source>, <volume>34</volume>, <fpage>8780</fpage>–<lpage>8794</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_011">
<mixed-citation publication-type="journal"><string-name><surname>Diakogiannis</surname>, <given-names>F.I.</given-names></string-name>, <string-name><surname>Waldner</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Caccetta</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Wu</surname>, <given-names>C.</given-names></string-name> (<year>2020</year>). <article-title>ResUNet-a: a deep learning framework for semantic segmentation of remotely sensed data</article-title>. <source>ISPRS Journal of Photogrammetry and Remote Sensing</source>, <volume>162</volume>, <fpage>94</fpage>–<lpage>114</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_012">
<mixed-citation publication-type="journal"><string-name><surname>Goodfellow</surname>, <given-names>I.</given-names></string-name>, <string-name><surname>Pouget-Abadie</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Mirza</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Xu</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Warde-Farley</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Ozair</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Courville</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Bengio</surname>, <given-names>Y.</given-names></string-name> (<year>2020</year>). <article-title>Generative adversarial networks</article-title>. <source>Communications of the ACM</source>, <volume>63</volume>(<issue>11</issue>), <fpage>139</fpage>–<lpage>144</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_013">
<mixed-citation publication-type="journal"><string-name><surname>Goudschaal</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Beeksma</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Boon</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Bijveld</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Visser</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Hinnen</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>van Kesteren</surname>, <given-names>Z.</given-names></string-name> (<year>2021</year>). <article-title>Accuracy of an MR-only workflow for prostate radiotherapy using semi-automatically burned-in fiducial markers</article-title>. <source>Radiation Oncology</source>, <volume>16</volume>, <fpage>1</fpage>–<lpage>13</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_014">
<mixed-citation publication-type="journal"><string-name><surname>Grigas</surname>, <given-names>O.</given-names></string-name>, <string-name><surname>Maskeliūnas</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Damaševičius</surname>, <given-names>R.</given-names></string-name> (<year>2023</year>). <article-title>Improving structural MRI preprocessing with hybrid transformer GANs</article-title>. <source>Life</source>, <volume>13</volume>(<issue>9</issue>), <fpage>1893</fpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_015">
<mixed-citation publication-type="journal"><string-name><surname>Han</surname>, <given-names>L.H.N.</given-names></string-name>, <string-name><surname>Hien</surname>, <given-names>N.L.H.</given-names></string-name>, <string-name><surname>Huy</surname>, <given-names>L.V.</given-names></string-name>, <string-name><surname>Hieu</surname>, <given-names>N.V.</given-names></string-name> (<year>2024</year>). <article-title>A deep learning model for multi-domain MRI synthesis using generative adversarial networks</article-title>. <source>Informatica</source>, <volume>35</volume>(<issue>2</issue>), <fpage>283</fpage>–<lpage>309</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_016">
<mixed-citation publication-type="journal"><string-name><surname>Ho</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Jain</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Abbeel</surname>, <given-names>P.</given-names></string-name> (<year>2020</year>). <article-title>Denoising diffusion probabilistic models</article-title>. <source>Advances in Neural Information Processing Systems</source>, <volume>33</volume>, <fpage>6840</fpage>–<lpage>6851</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_017">
<mixed-citation publication-type="chapter"><string-name><surname>Hu</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Chang</surname>, <given-names>C.-W.</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>X.</given-names></string-name> (<year>2023</year>a). <chapter-title>End-to-end brain tumor detection using a graph-feature-based classifier</chapter-title>. In: <source>Medical Imaging 2023: Biomedical Applications in Molecular, Structural, and Functional Imaging</source>, Vol. <volume>12468</volume>. <publisher-name>SPIE</publisher-name>, pp. <fpage>322</fpage>–<lpage>327</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_018">
<mixed-citation publication-type="chapter"><string-name><surname>Hu</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Chang</surname>, <given-names>C.-W</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>X.</given-names></string-name> (<year>2023</year>b). <chapter-title>Ultrasound breast tumor detection based on vision graph neural network</chapter-title>. In: <source>Medical Imaging 2023: Ultrasonic Imaging and Tomography</source>, Vol. <volume>12470</volume>. <publisher-name>SPIE</publisher-name>, pp. <fpage>165</fpage>–<lpage>170</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_019">
<mixed-citation publication-type="chapter"><string-name><surname>Hu</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Wynne</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>X.</given-names></string-name> (<year>2023</year>c). <chapter-title>A vision-GNN framework for retinopathy classification using optical coherence tomography</chapter-title>. In: <source>Medical Imaging 2023: Computer-Aided Diagnosis</source>, Vol. <volume>12465</volume>. <publisher-name>SPIE</publisher-name>, pp. <fpage>200</fpage>–<lpage>206</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_020">
<mixed-citation publication-type="chapter"><string-name><surname>Hu</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Shen</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Luo</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Tai</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Menze</surname>, <given-names>B.H.</given-names></string-name> (<year>2022</year>). <chapter-title>AutoGAN-synthesizer: neural architecture search for cross-modality MRI synthesis</chapter-title>. In: <source>International Conference on Medical Image Computing and Computer-Assisted Intervention</source>. <publisher-name>Springer</publisher-name>, pp. <fpage>397</fpage>–<lpage>409</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_021">
<mixed-citation publication-type="chapter"><string-name><surname>Ionescu</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Müller</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Villegas</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Arenas</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Boato</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Dang-Nguyen</surname>, <given-names>D.-T.</given-names></string-name>, <string-name><surname>Dicente Cid</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Eickhoff</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Seco de Herrera</surname>, <given-names>A.G.</given-names></string-name>, <string-name><surname>Gurrin</surname>, <given-names>C.</given-names></string-name>, <etal>et al.</etal> (<year>2017</year>). <chapter-title>Overview of ImageCLEF 2017: information extraction from images</chapter-title>. In: <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction: 8th International Conference of the CLEF Association, CLEF 2017, Dublin, Ireland, September 11–14, 2017, Proceedings 8</source>. <publisher-name>Springer</publisher-name>, pp. <fpage>315</fpage>–<lpage>337</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_022">
<mixed-citation publication-type="chapter"><string-name><surname>Isola</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Zhu</surname>, <given-names>J.-Y.</given-names></string-name>, <string-name><surname>Zhou</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Efros</surname>, <given-names>A.A.</given-names></string-name> (<year>2017</year>). <chapter-title>Image-to-image translation with conditional adversarial networks</chapter-title>. In: <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>, pp. <fpage>1125</fpage>–<lpage>1134</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_023">
<mixed-citation publication-type="journal"><string-name><surname>Jiang</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Jia</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Huang</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Yuan</surname>, <given-names>Z.</given-names></string-name>, <string-name><surname>Ruan</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Xia</surname>, <given-names>L.</given-names></string-name> (<year>2025</year>). <article-title>Frequency-aware diffusion model for multi-modal MRI image synthesis</article-title>. <source>Journal of Imaging</source>, <volume>11</volume>(<issue>5</issue>), <fpage>152</fpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_024">
<mixed-citation publication-type="journal"><string-name><surname>Juneja</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Minhas</surname>, <given-names>J.S.</given-names></string-name>, <string-name><surname>Singla</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Kaur</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Jindal</surname>, <given-names>P.</given-names></string-name> (<year>2024</year>). <article-title>Denoising techniques for cephalometric x-ray images: a comprehensive review</article-title>. <source>Multimedia Tools and Applications</source>, <volume>83</volume>(<issue>17</issue>), <fpage>49953</fpage>–<lpage>49991</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_025">
<mixed-citation publication-type="journal"><string-name><surname>Kriukova</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>LaRochelle</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Pfefer</surname>, <given-names>T.J.</given-names></string-name>, <string-name><surname>Kanniyappan</surname>, <given-names>U.</given-names></string-name>, <string-name><surname>Gioux</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Pogue</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Ntziachristos</surname>, <given-names>V.</given-names></string-name>, <string-name><surname>Gorpas</surname>, <given-names>D.</given-names></string-name> (<year>2025</year>). <article-title>Impact of signal-to-noise ratio and contrast definition on the sensitivity assessment and benchmarking of fluorescence molecular imaging systems</article-title>. <source>Journal of Biomedical Optics</source>, <volume>30</volume>(<issue>S1</issue>), <elocation-id>S13703</elocation-id>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_026">
<mixed-citation publication-type="journal"><string-name><surname>Li</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Li</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Safai</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Lomax</surname>, <given-names>A.J.</given-names></string-name>, <string-name><surname>Zhang</surname>, <given-names>Y.</given-names></string-name> (<year>2025</year>). <article-title>Diffusion Schrödinger bridge models for high-quality MR-to-CT synthesis for proton treatment planning</article-title>. <source>Medical Physics</source>, <volume>52</volume>(<issue>7</issue>), <fpage>17898</fpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_027">
<mixed-citation publication-type="other"><string-name><surname>Li</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Shang</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Butala</surname>, <given-names>M.D.</given-names></string-name> (2023a). DDMM-Synth: a denoising diffusion model for cross-modal medical image synthesis with sparse-view measurement embedding. arXiv preprint arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2303.15770">2303.15770</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_028">
<mixed-citation publication-type="chapter"><string-name><surname>Li</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Chang</surname>, <given-names>C.-W.</given-names></string-name>, <string-name><surname>Patel</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Jani</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Mao</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>X.</given-names></string-name> (<year>2023</year>b). <chapter-title>Multi-parametric MRI radiomics analysis with ensemble learning for prostate lesion classification</chapter-title>. In: <source>Medical Imaging 2023: Biomedical Applications in Molecular, Structural, and Functional Imaging</source>, Vol. <volume>12468</volume>. <publisher-name>SPIE</publisher-name>, pp. <fpage>49</fpage>–<lpage>54</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_029">
<mixed-citation publication-type="chapter"><string-name><surname>Li</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Hu</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Patel</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Mao</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>X.</given-names></string-name> (<year>2023</year>c). <chapter-title>Prostate Gleason score prediction via MRI using capsule network</chapter-title>. In: <source>Medical Imaging 2023: Computer-Aided Diagnosis</source>, Vol. <volume>12465</volume>. <publisher-name>SPIE</publisher-name>, pp. <fpage>514</fpage>–<lpage>519</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_030">
<mixed-citation publication-type="chapter"><string-name><surname>Li</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Zhou</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Pan</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Jani</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Patel</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>X.</given-names></string-name> (<year>2023</year>d). <chapter-title>Ultrasound-based dominant intraprostatic lesion classification with swin transformer</chapter-title>. In: <source>Medical Imaging 2023: Ultrasonic Imaging and Tomography</source>, Vol. <volume>12470</volume>. <publisher-name>SPIE</publisher-name>, pp. <fpage>146</fpage>–<lpage>151</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_031">
<mixed-citation publication-type="journal"><string-name><surname>Li</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Shao</surname>, <given-names>H.-C.</given-names></string-name>, <string-name><surname>Liang</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Chen</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Li</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Jiang</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Zhang</surname>, <given-names>Y.</given-names></string-name> (<year>2024</year>). <article-title>Zero-shot medical image translation via frequency-guided diffusion models</article-title>. <source>IEEE Transactions on Medical Imaging</source>, <volume>43</volume>(<issue>3</issue>), <fpage>980</fpage>–<lpage>993</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_032">
<mixed-citation publication-type="journal"><string-name><surname>Liu</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Chen</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Shi</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Huang</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Zheng</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>Z.</given-names></string-name>, <string-name><surname>Zhang</surname>, <given-names>Q.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>X.</given-names></string-name> (<year>2021</year>). <article-title>CT synthesis from MRI using multi-cycle GAN for head-and-neck radiation therapy</article-title>. <source>Computerized Medical Imaging and Graphics</source>, <volume>91</volume>, <elocation-id>101953</elocation-id>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_033">
<mixed-citation publication-type="chapter"><string-name><surname>Luo</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Toth</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Jiang</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Pushparajah</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Rhode</surname>, <given-names>K.</given-names></string-name> (<year>2020</year>). <chapter-title>Ultra-DenseNet for low-dose X-ray image denoising in cardiac catheter-based procedures</chapter-title>. In: <source>Statistical Atlases and Computational Models of the Heart. Multi-Sequence CMR Segmentation, CRT-EPiggy and LV Full Quantification Challenges: 10th International Workshop, STACOM 2019, Held in Conjunction with MICCAI 2019, Shenzhen, China, October 13, 2019, Revised Selected Papers 10</source>. <publisher-name>Springer</publisher-name>, pp. <fpage>31</fpage>–<lpage>42</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_034">
<mixed-citation publication-type="other"><string-name><surname>Lyu</surname>, <given-names>Q.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>G.</given-names></string-name> (2022). Conversion between CT and MRI images using diffusion and score-matching models. arXiv preprint arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2209.12104">2209.12104</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_035">
<mixed-citation publication-type="journal"><string-name><surname>Makitalo</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Foi</surname>, <given-names>A.</given-names></string-name> (<year>2012</year>). <article-title>Optimal inversion of the generalized Anscombe transformation for Poisson-Gaussian noise</article-title>. <source>IEEE Transactions on Image Processing</source>, <volume>22</volume>(<issue>1</issue>), <fpage>91</fpage>–<lpage>103</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_036">
<mixed-citation publication-type="chapter"><string-name><surname>Matviychuk</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Mailhé</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Chen</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>Q.</given-names></string-name>, <string-name><surname>Kiraly</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Strobel</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Nadar</surname>, <given-names>M.</given-names></string-name> (<year>2016</year>). <chapter-title>Learning a multiscale patch-based representation for image denoising in X-ray fluoroscopy</chapter-title>. In: <source>2016 IEEE International Conference on Image Processing (ICIP)</source>. <publisher-name>IEEE</publisher-name>, pp. <fpage>2330</fpage>–<lpage>2334</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_037">
<mixed-citation publication-type="journal"><string-name><surname>Moya-Sáez</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Peña-Nogales</surname>, <given-names>Ó.</given-names></string-name>, <string-name><surname>de Luis-García</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Alberola-López</surname>, <given-names>C.</given-names></string-name> (<year>2021</year>). <article-title>A deep learning approach for synthetic MRI based on two routine sequences and training with synthetic data</article-title>. <source>Computer Methods and Programs in Biomedicine</source>, <volume>210</volume>, <fpage>106371</fpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_038">
<mixed-citation publication-type="chapter"><string-name><surname>Nichol</surname>, <given-names>A.Q.</given-names></string-name>, <string-name><surname>Dhariwal</surname>, <given-names>P.</given-names></string-name> (<year>2021</year>). <chapter-title>Improved denoising diffusion probabilistic models</chapter-title>. In: <source>International Conference on Machine Learning</source>, <publisher-name>PMLR</publisher-name>, pp. <fpage>8162</fpage>–<lpage>8171</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_039">
<mixed-citation publication-type="journal"><string-name><surname>Özbey</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Dalmaz</surname>, <given-names>O.</given-names></string-name>, <string-name><surname>Dar</surname>, <given-names>S.U.</given-names></string-name>, <string-name><surname>Bedel</surname>, <given-names>H.A.</given-names></string-name>, <string-name><surname>Özturk</surname>, <given-names>Ş.</given-names></string-name>, <string-name><surname>Güngör</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Çukur</surname>, <given-names>T.</given-names></string-name> (<year>2023</year>). <article-title>Unsupervised medical image translation with adversarial diffusion models</article-title>. <source>IEEE Transactions on Medical Imaging</source>, <volume>42</volume>(<issue>12</issue>), <fpage>3524</fpage>–<lpage>3539</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_040">
<mixed-citation publication-type="chapter"><string-name><surname>Pan</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Flores</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Lin</surname>, <given-names>C.T.</given-names></string-name>, <string-name><surname>Stayman</surname>, <given-names>J.W.</given-names></string-name>, <string-name><surname>Gang</surname>, <given-names>G.J.</given-names></string-name> (<year>2021</year>). <chapter-title>Generative adversarial networks and radiomics supervision for lung lesion synthesis</chapter-title>. In: <source>Medical Imaging 2021: Physics of Medical Imaging</source>, Vol. <volume>11595</volume>. <publisher-name>SPIE</publisher-name>, pp. <fpage>167</fpage>–<lpage>172</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_041">
<mixed-citation publication-type="journal"><string-name><surname>Pan</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Abouei</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Wynne</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Chang</surname>, <given-names>C.-W.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Qiu</surname>, <given-names>R.L.</given-names></string-name>, <string-name><surname>Li</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Peng</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Roper</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Patel</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Yu</surname>, <given-names>D.S.</given-names></string-name>, <string-name><surname>Mao</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>X.</given-names></string-name> (<year>2024</year>). <article-title>Synthetic CT generation from MRI using 3D transformer-based denoising diffusion model</article-title>. <source>Medical Physics</source>, <volume>51</volume>(<issue>4</issue>), <fpage>2538</fpage>–<lpage>2548</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_042">
<mixed-citation publication-type="journal"><string-name><surname>Peng</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Qiu</surname>, <given-names>R.L.</given-names></string-name>, <string-name><surname>Wynne</surname>, <given-names>J.F.</given-names></string-name>, <string-name><surname>Chang</surname>, <given-names>C.-W.</given-names></string-name>, <string-name><surname>Pan</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Roper</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Patel</surname>, <given-names>P.R.</given-names></string-name>, <string-name><surname>Yu</surname>, <given-names>D.S.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>X.</given-names></string-name> (<year>2024</year>). <article-title>CBCT-based synthetic CT image generation using conditional denoising diffusion probabilistic model</article-title>. <source>Medical Physics</source>, <volume>51</volume>(<issue>3</issue>), <fpage>1847</fpage>–<lpage>1859</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_043">
<mixed-citation publication-type="journal"><string-name><surname>Safari</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Li</surname>, <given-names>Q.</given-names></string-name>, <string-name><surname>Eidex</surname>, <given-names>Z.</given-names></string-name>, <string-name><surname>Qiu</surname>, <given-names>R.L.</given-names></string-name>, <string-name><surname>Chang</surname>, <given-names>C.-W.</given-names></string-name>, <string-name><surname>Mao</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>X.</given-names></string-name> (<year>2025</year>). <article-title>Res-MoCoDiff: residual-guided diffusion models for motion artifact correction in brain MRI</article-title>. <source>Physics in Medicine &amp; Biology</source>, <volume>70</volume>(<issue>20</issue>), <fpage>205022</fpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_044">
<mixed-citation publication-type="chapter"><string-name><surname>Sohail</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Riaz</surname>, <given-names>M.N.</given-names></string-name>, <string-name><surname>Wu</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Long</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Li</surname>, <given-names>S.</given-names></string-name> (<year>2019</year>). <chapter-title>Unpaired multi-contrast MR image synthesis using generative adversarial networks</chapter-title>. In: <source>International Workshop on Simulation and Synthesis in Medical Imaging</source>. <publisher-name>Springer</publisher-name>, pp. <fpage>22</fpage>–<lpage>31</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_045">
<mixed-citation publication-type="other"><string-name><surname>Song</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Meng</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Ermon</surname>, <given-names>S.</given-names></string-name> (2020). Denoising diffusion implicit models. arXiv preprint arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2010.02502">2010.02502</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_046">
<mixed-citation publication-type="journal"><string-name><surname>Sun</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Hu</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Cai</surname>, <given-names>H.</given-names></string-name> (<year>2021</year>). <article-title>Image features of magnetic resonance imaging under the deep learning algorithm in the diagnosis and nursing of malignant tumors</article-title>. <source>Contrast Media &amp; Molecular Imaging</source>, <volume>2021</volume>(<issue>1</issue>), <elocation-id>1104611</elocation-id>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_047">
<mixed-citation publication-type="journal"><string-name><surname>Tiago</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Snare</surname>, <given-names>S.R.</given-names></string-name>, <string-name><surname>Šprem</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>McLeod</surname>, <given-names>K.</given-names></string-name> (<year>2023</year>). <article-title>A domain translation framework with an adversarial denoising diffusion model to generate synthetic datasets of echocardiography images</article-title>. <source>IEEE Access</source>, <volume>11</volume>, <fpage>17594</fpage>–<lpage>17602</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_048">
<mixed-citation publication-type="chapter"><string-name><surname>Turajlić</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Karahodzic</surname>, <given-names>V.</given-names></string-name> (<year>2017</year>). <chapter-title>An adaptive scheme for X-ray medical image denoising using artificial neural networks and additive white Gaussian noise level estimation in SVD domain</chapter-title>. In: <source>CMBEBIH 2017: Proceedings of the International Conference on Medical and Biological Engineering 2017</source>. <publisher-name>Springer</publisher-name>, pp. <fpage>36</fpage>–<lpage>40</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_049">
<mixed-citation publication-type="other"><string-name><surname>Wang</surname>, <given-names>Z.</given-names></string-name>, <string-name><surname>Zhang</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Zhang</surname>, <given-names>Z.</given-names></string-name> (2024). Soft masked mamba diffusion model for ct to mri conversion. arXiv preprint arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2406.15910">2406.15910</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_050">
<mixed-citation publication-type="other"><string-name><surname>Wei</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Gong</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>Z.</given-names></string-name>, <string-name><surname>Lu</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>L.</given-names></string-name> (2018). Improving the improved training of wasserstein gans: a consistency term and its dual effect. arXiv preprint arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1803.01541">1803.01541</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_051">
<mixed-citation publication-type="journal"><string-name><surname>Wu</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Qiang</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Song</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Ren</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Zhang</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Hussain</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Cui</surname>, <given-names>Y.</given-names></string-name> (<year>2020</year>). <article-title>Image synthesis in contrast MRI based on super resolution reconstruction with multi-refinement cycle-consistent generative adversarial networks</article-title>. <source>Journal of Intelligent Manufacturing</source>, <volume>31</volume>(<issue>5</issue>), <fpage>1215</fpage>–<lpage>1228</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_052">
<mixed-citation publication-type="chapter"><string-name><surname>Yu</surname>, <given-names>Z.</given-names></string-name>, <string-name><surname>Zhai</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Han</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Peng</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Zhang</surname>, <given-names>X.-Y.</given-names></string-name> (<year>2021</year>). <chapter-title>MouseGAN: GAN-based multiple MRI modalities synthesis and segmentation for mouse brain structures</chapter-title>. In: <source>Medical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, September 27–October 1, 2021, Proceedings, Part I 24</source>. <publisher-name>Springer</publisher-name>, pp. <fpage>442</fpage>–<lpage>450</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_053">
<mixed-citation publication-type="journal"><string-name><surname>Yurt</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Dar</surname>, <given-names>S.U.</given-names></string-name>, <string-name><surname>Erdem</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Erdem</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Oguz</surname>, <given-names>K.K.</given-names></string-name>, <string-name><surname>Çukur</surname>, <given-names>T.</given-names></string-name> (<year>2021</year>). <article-title>mustGAN: multi-stream generative adversarial networks for MR image synthesis</article-title>. <source>Medical Image Analysis</source>, <volume>70</volume>, <elocation-id>101944</elocation-id>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_054">
<mixed-citation publication-type="journal"><string-name><surname>Zhang</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Huang</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Zhang</surname>, <given-names>Q.</given-names></string-name>, <string-name><surname>Han</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Han</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Yu</surname>, <given-names>Y.</given-names></string-name> (<year>2020</year>). <article-title>Exploring task structure for brain tumor segmentation from multi-modality MR images</article-title>. <source>IEEE Transactions on Image Processing</source>, <volume>29</volume>, <fpage>9032</fpage>–<lpage>9043</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_055">
<mixed-citation publication-type="journal"><string-name><surname>Zhang</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Li</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Dillman</surname>, <given-names>J.R.</given-names></string-name>, <string-name><surname>Parikh</surname>, <given-names>N.A.</given-names></string-name>, <string-name><surname>He</surname>, <given-names>L.</given-names></string-name> (<year>2022</year>). <article-title>Multi-contrast MRI image synthesis using switchable cycle-consistent generative adversarial networks</article-title>. <source>Diagnostics</source>, <volume>12</volume>(<issue>4</issue>), <fpage>816</fpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_056">
<mixed-citation publication-type="chapter"><string-name><surname>Zhang</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Chi</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Lyu</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Tian</surname>, <given-names>Y.</given-names></string-name> (<year>2023</year>). <chapter-title>Dual arbitrary scale super-resolution for multi-contrast MRI</chapter-title>. In: <source>International Conference on Medical Image Computing and Computer-Assisted Intervention</source>. <publisher-name>Springer</publisher-name>, pp. <fpage>282</fpage>–<lpage>292</lpage>.</mixed-citation>
</ref>
<ref id="j_infor622_ref_057">
<mixed-citation publication-type="chapter"><string-name><surname>Zhu</surname>, <given-names>J.-Y.</given-names></string-name>, <string-name><surname>Park</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Isola</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Efros</surname>, <given-names>A.A.</given-names></string-name> (<year>2017</year>). <chapter-title>Unpaired image-to-image translation using cycle-consistent adversarial networks</chapter-title>. In: <source>Proceedings of the IEEE International Conference on Computer Vision</source>, pp. <fpage>2223</fpage>–<lpage>2232</lpage>.</mixed-citation>
</ref>
</ref-list>
</back>
</article>
