<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.0 20120330//EN" "JATS-journalpublishing1.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">INFORMATICA</journal-id>
<journal-title-group><journal-title>Informatica</journal-title></journal-title-group>
<issn pub-type="epub">1822-8844</issn><issn pub-type="ppub">0868-4952</issn><issn-l>0868-4952</issn-l>
<publisher>
<publisher-name>Vilnius University</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">INFOR556</article-id>
<article-id pub-id-type="doi">10.15388/24-INFOR556</article-id>
<article-categories><subj-group subj-group-type="heading">
<subject>Research Article</subject></subj-group></article-categories>
<title-group>
<article-title>A Deep Learning Model for Multi-Domain MRI Synthesis Using Generative Adversarial Networks</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Han</surname><given-names>Le Hoang Ngoc</given-names></name><email xlink:href="hanlehngoc080601@gmail.com">hanlehngoc080601@gmail.com</email><xref ref-type="aff" rid="j_infor556_aff_001">1</xref><bio>
<p><bold>L.H.N. Han</bold> is a junior researcher at the Department of Software Engineering, The University of Danang – University of Science and Technology. Her research interests are data science and computer vision, with a special focus on MRI image processing. She obtained her BSc from the University of Danang – University of Science and Technology in Vietnam with a specialization in data science and artificial intelligence.</p></bio>
</contrib>
<contrib contrib-type="author">
<name><surname>Hien</surname><given-names>Ngo Le Huy</given-names></name><email xlink:href="ngolehuyhien@gmail.com">ngolehuyhien@gmail.com</email><xref ref-type="aff" rid="j_infor556_aff_002">2</xref><bio>
<p><bold>N.L.H. Hien</bold> is a postgraduate researcher at the School of Built Environment, Engineering, and Computing, Leeds Beckett University, United Kingdom. He obtained an MSc in information and technology at Leeds Beckett University under the Erasmus Mundus in Green Networking and Cloud Computing program. During the last six years, he has published research papers mainly related to the development of machine learning and deep learning models and their applications. His research interests include artificial intelligence, green ICT, and sustainable computing.</p></bio>
</contrib>
<contrib contrib-type="author">
<name><surname>Huy</surname><given-names>Luu Van</given-names></name><email xlink:href="lvhuy@dut.udn.com">lvhuy@dut.udn.com</email><xref ref-type="aff" rid="j_infor556_aff_001">1</xref><bio>
<p><bold>L.V. Huy</bold> is a teaching assistant at the University of Danang – University of Science and Technology. His research interests include computer vision and natural language processing. He obtained his MSc and BSc degrees from the University of Danang – University of Science and Technology in Vietnam.</p></bio>
</contrib>
<contrib contrib-type="author">
<name><surname>Hieu</surname><given-names>Nguyen Van</given-names></name><email xlink:href="nvhieuqt@dut.udn.com">nvhieuqt@dut.udn.com</email><xref ref-type="aff" rid="j_infor556_aff_001">1</xref><xref ref-type="corresp" rid="cor1">∗</xref><bio>
<p><bold>N.V. Hieu</bold> is a full senior lecturer at the University of Danang – University of Science and Technology. His research interests include deep learning, modelling, optimization, decision making and computational impacts. He obtained his PhD from Saint Petersburg State Institute of Technology and his MSc and BSc degrees from Saint Petersburg State University in Russia.</p></bio>
</contrib>
<aff id="j_infor556_aff_001"><label>1</label>Faculty of Information Technology, <institution>The University of Danang – University of Science and Technology</institution>, <country>Vietnam</country></aff>
<aff id="j_infor556_aff_002"><label>2</label>School of Built Environment, Engineering, and Computing – <institution>Leeds Beckett University</institution>, <country>United Kingdom</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>∗</label>Corresponding author.</corresp>
</author-notes>
<pub-date pub-type="ppub"><year>2024</year></pub-date><pub-date pub-type="epub"><day>29</day><month>4</month><year>2024</year></pub-date><volume>35</volume><issue>2</issue><fpage>283</fpage><lpage>309</lpage><history><date date-type="received"><month>7</month><year>2023</year></date><date date-type="accepted"><month>4</month><year>2024</year></date></history>
<permissions><copyright-statement>© 2024 Vilnius University</copyright-statement><copyright-year>2024</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>Open access article under the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/">CC BY</ext-link> license.</license-p></license></permissions>
<abstract>
<p>In recent years, Magnetic Resonance Imaging (MRI) has emerged as a prevalent medical imaging technique, offering comprehensive anatomical and functional information. However, the MRI data acquisition process presents several challenges, including time-consuming procedures, prone motion artifacts, and hardware constraints. To address these limitations, this study proposes a novel method that leverages the power of generative adversarial networks (GANs) to generate multi-domain MRI images from a single input MRI image. Within this framework, two primary generator architectures, namely ResUnet and StarGANs generators, were incorporated. Furthermore, the networks were trained on multiple datasets, thereby augmenting the available data, and enabling the generation of images with diverse contrasts obtained from different datasets, given an input image from another dataset. Experimental evaluations conducted on the IXI and BraTS2020 datasets substantiate the efficacy of the proposed method compared to an existing method, as assessed through metrics such as Structural Similarity Index (SSIM), Peak Signal-to-Noise Ratio (PSNR) and Normalized Mean Absolute Error (NMAE). The synthesized images resulting from this method hold substantial potential as invaluable resources for medical professionals engaged in research, education, and clinical applications. Future research gears towards expanding experiments to larger datasets and encompassing the proposed approach to 3D images, enhancing medical diagnostics within practical applications.</p>
</abstract>
<kwd-group>
<label>Key words</label>
<kwd>medical imaging</kwd>
<kwd>MRI</kwd>
<kwd>synthesis</kwd>
<kwd>deep learning</kwd>
<kwd>GANs</kwd>
</kwd-group>
<funding-group><funding-statement>This work was funded by The University of Danang, University of Science and Technology and the Master, PhD Scholarship Programme of Vingroup Innovation Foundation (VINIF), the code number of Project: VINIF.2023.Ths.045.</funding-statement></funding-group>
</article-meta>
</front>
<body>
<sec id="j_infor556_s_001">
<label>1</label>
<title>Introduction</title>
<p>Magnetic Resonance Imaging (MRI) plays a crucial role in modern medical diagnostics, providing non-invasive visualization of internal body structures and aiding in the diagnosis and monitoring of various diseases (Katti <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_015">2011</xref>). However, acquiring high-quality MRI data can be a time-consuming and costly process. Additionally, the availability of diverse pulse sequences for MRI imaging is limited, which can restrict the ability to capture specific anatomical or functional information (Geethanath and Vaughan, <xref ref-type="bibr" rid="j_infor556_ref_008">2019</xref>).</p>
<p>The aim of MRI synthesis is to provide clinicians and researchers with a broader range of imaging options without the need for additional scanning (Liang and Lauterbur, <xref ref-type="bibr" rid="j_infor556_ref_018">2000</xref>). By synthesizing multiple pulse sequences from a single input image, medical professionals can study various aspects of the same anatomical region or analyse the effects of different imaging parameters without subjecting patients to multiple scans (Wang <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_024">2021</xref>). This approach has the potential to save time, reduce costs, and minimize patient discomfort.</p>
<p>In recent years, several state-of-the-art MRI synthesis techniques have been developed to address the challenges of multimodal MRI, such as long scan times and artifacts (Liu <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_019">2019</xref>; Li <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_017">2019</xref>). One of the techniques is the pyramid transformer network (PTNet3D), a new framework for 3D MRI synthesis (Zhang <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_028">2022</xref>). PTNet3D uses attention mechanisms through transformer and performer layers, and it consistently outperforms convolutional neural network-based generative adversarial networks (GANs) in terms of synthesis accuracy and generalization on high resolution datasets. However, PTNet3D requires a large amount of training data and is computationally intensive. Another study by Massa <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor556_ref_020">2020</xref>) explored the use of different clinical MRI sequences as inputs for deep CT synthesis pipelines. Nevertheless, while deep learning-based CT synthesis from MRI is possible with various input types, no single image type performed best in all cases. Additionally, Hatamizadeh <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor556_ref_010">2021</xref>) proposed the Swin UNETR model, which uses Swin Transformers for brain tumour segmentation in MRI images. They emphasized the need for multiple complementary 3D MRI modalities to highlight different tissue properties and areas of tumour spread. However, automated medical image segmentation techniques are still in their early stages, more research is needed to improve their accuracy.</p>
<p>It can be seen from the abovementioned MRI synthesis models that they remain to face limitations in relying on large training datasets (which must be synchronized and adhere to a standardized template) and high computational complexity. Generalization across different datasets and imaging modalities can be challenging. Therefore, to address these research gaps, this study proposes a novel method that utilizes a single generator and a single discriminator for the entire multi-domain image-to-image translation task. Instead of employing separate generators and discriminators for each target domain, a unified architecture is adopted, enabling the generation and discrimination of multiple classes utilizing a single set of model parameters. This approach not only simplifies the overall model architecture but also facilitates efficient training and inference processes.</p>
<p>The contributions of this paper are threefold: 
<list>
<list-item id="j_infor556_li_001">
<label>1.</label>
<p>Firstly, a deep learning based approach for MRI synthesis is proposed, incorporating GANs based architecture and new generation loss, enabling the generation of multiple MRI images with different contrasts from a single input image.</p>
</list-item>
<list-item id="j_infor556_li_002">
<label>2.</label>
<p>Secondly, the effectiveness of the proposed method is asserted through training on multiple MRI datasets, including IXI and BraTS2020, which have different contrasts, showcasing the ability to generate realistic and matching images.</p>
<p>
<fig id="j_infor556_fig_001">
<label>Fig. 1</label>
<caption>
<p>Illustration of multi-contrast MRI synthesis on BraTS2020 and IXI datasets.</p>
</caption>
<graphic xlink:href="infor556_g001.jpg"/>
</fig>
<fig id="j_infor556_fig_002">
<label>Fig. 2</label>
<caption>
<p>‘Glomerular hemorrhage’ was detected in T2 (right) instead of T1 (left).</p>
</caption>
<graphic xlink:href="infor556_g002.jpg"/>
</fig>
<fig id="j_infor556_fig_003">
<label>Fig. 3</label>
<caption>
<p>‘Parotid pleomorphic adenoma’ detected in T1CE (right) instead of T1 (left).</p>
</caption>
<graphic xlink:href="infor556_g003.jpg"/>
</fig>
</p>
</list-item>
<list-item id="j_infor556_li_003">
<label>3.</label>
<p>Thirdly, both qualitative and quantitative results are generated on multicontrast MRI synthesis tasks using the ResUnet generator and StarGAN generators.</p>
</list-item>
</list> 
The aim of this research is to generate MRI images with diverse contrasts solely from a single MRI image. For instance, given an input MRI image acquired with a T1 pulse sequence, the model can generate MRI images with T2, MRA, FLAIR contrasts, and more (as depicted in Fig. <xref rid="j_infor556_fig_001">1</xref>). This innovative approach has the potential to yield substantial benefits, including cost savings, time efficiency, and reduced patient discomfort, by eliminating the necessity for multiple MRI scans that require patients to remain motionless for prolonged periods. Moreover, taking the same anatomical structure in a different pulse sequence can assist diagnosticians in detecting underlying pathologies or injuries, such as brain tumours or oedema, as illustrated in Figs. <xref rid="j_infor556_fig_002">2</xref>, <xref rid="j_infor556_fig_003">3</xref>, and <xref rid="j_infor556_fig_004">4</xref>.</p>
<fig id="j_infor556_fig_004">
<label>Fig. 4</label>
<caption>
<p>‘Simple ovarian cyst’ detected in T2 (left) instead of T1 (right).</p>
</caption>
<graphic xlink:href="infor556_g004.jpg"/>
</fig>
<p>The remainder of this paper is structured as follows. Section <xref rid="j_infor556_s_002">2</xref> provides an overview of related work in MRI synthesis using deep learning. The methodology and architecture of the proposed GAN-based approach are described in Section <xref rid="j_infor556_s_005">3</xref>. Then, the experimental setup and results are detailed in Section <xref rid="j_infor556_s_015">4</xref>. Finally, Section <xref rid="j_infor556_s_026">6</xref> concludes the paper and discusses potential future directions for research in MRI synthesis using deep learning models.</p>
</sec>
<sec id="j_infor556_s_002">
<label>2</label>
<title>Related Works</title>
<sec id="j_infor556_s_003">
<label>2.1</label>
<title>Medical Image-to-Image Transformation</title>
<p>Recent studies have examined the application of deep learning models for MRI synthesis, aiming to generate diverse MRI images with different contrasts. These approaches have leveraged various techniques and architectures to achieve accurate and realistic synthesis results. One prominent approach in MRI synthesis is the utilization of generative adversarial networks (GANs). Sohail <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor556_ref_023">2019</xref>) introduced a method for synthesizing multi-contrast MR images without paired training data. The proposed approach utilizes GANs to learn the mapping between input images and target contrasts. Adversarial and cycle-consistency losses are employed to ensure the generation of high-quality and realistic multi-contrast MR images. In another study, Zhu <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor556_ref_029">2017</xref>) suggested an approach that bears similarities to the method introduced in Sohail <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor556_ref_023">2019</xref>). However, this method utilizes cycle-consistent adversarial networks (CycleGAN) to learn mappings between only two different image domains. By employing two separate generator networks and two discriminator networks, the model performs both forward and backward translations between the domains. The cycle-consistency loss ensures that the reconstructed images from the translated images remain close to their original counterparts. Experimental results demonstrate the effectiveness of CycleGAN in various image translation tasks, including style transfer, object transfiguration, and season translation.</p>
<p>Wei <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor556_ref_025">2018</xref>) proposed a method to enhance the training of Wasserstein-GANs by introducing a consistency term. The consistency term ensures that the generated samples are consistent under small perturbations in the input noise space. This term helps stabilize the training process and improve the quality of generated samples. They also introduced the concept of the dual effect allowing the generator to capture more diverse modes of data distribution. Experimental results demonstrate that the proposed method outperforms the original WGAN and other state-of-the-art GAN variants in terms of image quality, diversity, and stability.</p>
<p>In recent advancements in medical imaging, several methods have emerged to enhance MRI image resolution such as Zhang <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor556_ref_027">2023</xref>) using Dual-ArbNet and Chen <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor556_ref_004">2023</xref>) with CANM-Net. Additionally, another study has been conducted to increase the resolution and remove noise from 2D images using Transformer GANs by Grigas <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor556_ref_009">2023</xref>). In general, although the task is named super-resolution, it can still be considered a sub-category of image synthesis. It can be observed that in recent years, GANs have played a crucial role and have been widely applied in tasks related to brain MRI images, especially with the majority of studies aiming at MRI image synthesis applications, according to statistics from Ali <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor556_ref_001">2022</xref>).</p>
<p>However, it can be seen that most of the above methods have limitations to provide single-input multi-output (SIMO) MRI image synthesis, which could be extremely beneficial, especially under circumstances where only a single modality is available but multiple contrasts are necessary for diagnosis and treatment planning. Moreover, it is imperative to develop an approach that facilitates training on multiple datasets to enhance the diversity of learning data and improve the synthesizing performance.</p>
</sec>
<sec id="j_infor556_s_004">
<label>2.2</label>
<title>Multi-Domain Image-to-Image Transformation</title>
<p>StarGANs represent one of the pioneering attempts to address multi-domain image-to-image translation within a single dataset (Nelson and Menon, <xref ref-type="bibr" rid="j_infor556_ref_021">2022</xref>). This innovation allows the generator <italic>G</italic> to transform a single image <italic>X</italic> to an output image <italic>Y</italic> conditioned by the expected target domain <inline-formula id="j_infor556_ineq_001"><alternatives><mml:math>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="italic">G</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">X</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">→</mml:mo>
<mml:mi mathvariant="italic">Y</mml:mi></mml:math><tex-math><![CDATA[$c:G(X,c)\to Y$]]></tex-math></alternatives></inline-formula>. To achieve this, the input image is combined with the target domain information to form the input for the generator (Choi <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_005">2018</xref>).</p>
<p>By incorporating the target domain information during the synthesis process, StarGANs effectively enable the generator to adapt its output based on the specified domain. This conditioning mechanism facilitates the translation of images across various domains within a single training dataset, eliminating the need for separate models or training processes for each specific domain. The generator, equipped with the ability to utilize the target domain information, can effectively learn the mapping between the input image and the desired output in a domain-specific manner (Choi <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_006">2020</xref>).</p>
<p>While StarGan has shown impressive results in multi-domain image-to-image translation, the original research did not explore its application in the medical domain. Furthermore, the original architecture simply consisted of a straightforward CNN with a few layers. As a result, several subsequent efforts have focused on modifying the architecture and adapting the StarGan approach to medical datasets. Building upon it, this study also aims to further enhance the method and prove its capacity in the medical domain.</p>
</sec>
</sec>
<sec id="j_infor556_s_005" sec-type="methods">
<label>3</label>
<title>Methodology</title>
<p>The design methodology of this research is built upon a set of fundamental principles. Firstly, simplicity and generality are given utmost priority in the model architecture, while also ensuring its robustness to capture intricate features and facilitate efficient training. Secondly, emphasis is placed on leveraging information from various aspects during the optimization process to enhance the model’s generalization capabilities. This section entails how those principles converge to the proposed modification to the original design.</p>
<sec id="j_infor556_s_006">
<label>3.1</label>
<title>Generator</title>
<sec id="j_infor556_s_007">
<label>3.1.1</label>
<title>StarGANs-Based Generator Architecture</title>
<p>The first approach in this study is StarGANs, based on the architecture of CycleGAN, utilizing a generator network with two downsampling convolutional layers, six residual blocks, and two upsampling transposed convolutional layers. Instance Normalization is applied to the generator, while no normalization is used for the discriminator (Bharti <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_003">2022</xref>). The detailed architecture is shown in Fig. <xref rid="j_infor556_fig_005">5</xref>.</p>
<fig id="j_infor556_fig_005">
<label>Fig. 5</label>
<caption>
<p>Illustration of the StarGANs generator architecture.</p>
</caption>
<graphic xlink:href="infor556_g005.jpg"/>
</fig>
<list>
<list-item id="j_infor556_li_004">
<label>•</label>
<p>Downsampling Convolutional Layers: The two downsampling convolutional layers are responsible for reducing the spatial dimensions of the input image (Hien <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_013">2021</xref>). Convolutions are applied with a stride size of two, which effectively downsamples the image and captures the coarse features.</p>
</list-item>
<list-item id="j_infor556_li_005">
<label>•</label>
<p>Residual Blocks: These parts play a crucial role in capturing fine-grained details and enhancing the gradient flow through the network. Each residual block consists of two convolutional layers with batch normalization and ReLU activation functions. The residual connection allows the network to learn residual mappings, which helps in alleviating the vanishing gradient problem and enables the network to learn more complex representations.</p>
</list-item>
<list-item id="j_infor556_li_006">
<label>•</label>
<p>Upsampling Transposed Convolutional Layers: The two upsampling transposed convolutional layers perform the opposite operation of the downsampling layers. They increase the spatial dimensions of the features, allowing the generator to reconstruct the output image with higher resolution (Hien <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_012">2020</xref>). These layers use transposed convolutions with a stride size of two, effectively upsampling the features.</p>
</list-item>
<list-item id="j_infor556_li_007">
<label>•</label>
<p>Instance Normalization: Instance normalization is applied to the generator. It normalizes the activations within each instance (sample) in the batch, making the network more robust to variations in instance statistics. By applying instance normalization, the generator can effectively capture and preserve the style and texture information of the input image during the translation process.</p>
</list-item>
</list>
</sec>
<sec id="j_infor556_s_008">
<label>3.1.2</label>
<title>ResUNet Architecture</title>
<p>It is observed that the shallow design of the standard architecture might struggle to capture fine-grained information at multiple scales. To address that concern, a ResUNet generator is utilized, which incorporates feature fusion at multiple scales for the image-to-image translation task. The ResUNet network is constructed by stacking multiple modified residual blocks (ResBlock) in an encoder-decoder fashion (U-Net). In line with the U-Net design, this approach also maintains a symmetric architecture, as depicted in Fig. <xref rid="j_infor556_fig_006">6</xref>. The ResUNet generator starts with 2 downsampling ResBlocks, followed by an encoder ResBlock, an upsampling ResBlock, and a final ResBlock called Decoder. The generation of the output is facilitated by a convolutional layer. The subsequent sections provide a detailed exploration of the architecture and its functionality:</p>
<fig id="j_infor556_fig_006">
<label>Fig. 6</label>
<caption>
<p>Illustration of the ResUNet architecture as a generator.</p>
</caption>
<graphic xlink:href="infor556_g006.jpg"/>
</fig>
<list>
<list-item id="j_infor556_li_008">
<label>•</label>
<p>Encoder: Positioned on the left side of the dashed line, the encoder consists of Downsampling Blocks, each composed of Residual Blocks. Within each Residual Block are two convolutional layers accompanied by batch normalization and ReLU activation functions. Notably, there’s an additional AvgPool layer at the conclusion of the Residual Block, tasked with diminishing the spatial dimensions of the input image.</p>
</list-item>
<list-item id="j_infor556_li_009">
<label>•</label>
<p>Decoder: The right-hand side of the dashed line consists of upsampling blocks. Similar to the Encoder, each upsampling block is composed of residual blocks; however, contrary to downsampling, at the end of each Residual Block, there is a deconvolution layer to upscale the generated image, along with skip connections from the Encoder. These skip connections fuse the high-resolution features from the encoder with the upsampled features, enabling the reconstruction of detailed and high-quality images.</p>
</list-item>
<list-item id="j_infor556_li_010">
<label>•</label>
<p>Output Layer: This is the final layer of the generator and typically employs a suitable activation function, such as sigmoid or tanh, to map the output values to the desired range. This layer generates the translated image in the target domain based on the learned features and the input image.</p>
</list-item>
</list>
</sec>
<sec id="j_infor556_s_009">
<label>3.1.3</label>
<title>Proposed Generator Architecture</title>
<p>After conducting experiments with the ResUNet architecture, a drawback was observed, which is that the generated MRI images, although exhibiting relatively good structural features, do not accurately represent the colours and components of the brain. It is speculated that the extensive use of Residual Blocks throughout the network and concatenation of all the encoder outputs to the decoder input causes the model to learn excessive colour features from the source image and introduce colour discrepancies in the generated images. Meanwhile, images generated by StarGANs exhibit high colour accuracy, addressing the shortcomings of ResUNet, yet face challenges in ensuring fine-grained details and high resolution.</p>
<p>Therefore, the image generation capability of the StarGANs network and the feature extraction capability of the ResUnet network will be leveraged in this research. Specifically, the main components of the ResUnet network, including ResBlock, Encoder, and Decoder, are retained. Additionally, two <inline-formula id="j_infor556_ineq_002"><alternatives><mml:math>
<mml:mn>7</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>7</mml:mn></mml:math><tex-math><![CDATA[$7\times 7$]]></tex-math></alternatives></inline-formula> convolutional layers are added followed by instance normalization layers and a ReLU activation function at the two ends of the network (Hien <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_011">2021</xref>). This research aims to reduce the gradient of the network and allow the model to retain features while still flexibly generating images, ensuring that the colours of the generated images differ from the input images (e.g. image T1 has contrasting colours with image T2). Therefore, the resblock block in the bottleneck section is replaced with 6 Residual Blocks, adapting from Choi <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor556_ref_005">2018</xref>). This modification aims to make the network less deep and reduce the dominance of dense concatenation, which may have been contributing to the colour discrepancies. The detailed architecture summary is described in the Appendix section, in Table <xref rid="j_infor556_tab_006">6</xref>.</p>
<fig id="j_infor556_fig_007">
<label>Fig. 7</label>
<caption>
<p>Illustration of the proposed generator architecture.</p>
</caption>
<graphic xlink:href="infor556_g007.jpg"/>
</fig>
</sec>
</sec>
<sec id="j_infor556_s_010">
<label>3.2</label>
<title>Discriminator</title>
<p>The discriminator is responsible for classifying whether an input image is real or fake, and also predicting the class label of the image in the case of conditional GANs. It follows the PatchGAN architecture (Choi <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_005">2018</xref>; Zhu <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_029">2017</xref>), which operates on local image patches rather than the whole image.</p>
<fig id="j_infor556_fig_008">
<label>Fig. 8</label>
<caption>
<p>Illustration of the discriminator architecture.</p>
</caption>
<graphic xlink:href="infor556_g008.jpg"/>
</fig>
<p>The main part of the discriminator contains the stacked convolutional layers, accumulating the number of channels to capture more complex features. The output of these layers is then passed through two separate convolutional layers, as shown in Fig. <xref rid="j_infor556_fig_007">7</xref> and Fig. <xref rid="j_infor556_fig_008">8</xref>.</p>
<p>The structure of the Discriminator in StarGAN is similar to a typical Discriminator, but it includes two separate convolutional layers near the final layer to generate two distinct outputs: the first output represents the probability of the image being real or fake, while the second predicts the class label. The discriminator is designed to handle images of varying sizes, and its architecture promotes the learning of local image details for accurate discrimination. The detailed architecture summary is described at Table <xref rid="j_infor556_tab_007">7</xref>.</p>
</sec>
<sec id="j_infor556_s_011">
<label>3.3</label>
<title>Loss Functions</title>
<p>The main purpose of GANs in general and this study approach, in particular, is to make the generated images indistinguishable from real images, an adversarial loss of typical GANs (Bharti <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_003">2022</xref>) is adopted, defined 
<disp-formula id="j_infor556_eq_001">
<label>(1)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">adv</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">[</mml:mo>
<mml:mo movablelimits="false">log</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mi mathvariant="italic">r</mml:mi>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">]</mml:mo>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">[</mml:mo>
<mml:mo movablelimits="false">log</mml:mo>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mi mathvariant="italic">r</mml:mi>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:mi mathvariant="italic">G</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">]</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {L_{\textit{adv}}}={E_{x}}\big[\log {D_{src}}(x)\big]+{E_{x,c}}\big[\log \big(1-{D_{src}}\big(G(x,c)\big)\big)\big],\]]]></tex-math></alternatives>
</disp-formula> 
where <italic>G</italic> creates an image <inline-formula id="j_infor556_ineq_003"><alternatives><mml:math>
<mml:mi mathvariant="italic">G</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$G(x,c)$]]></tex-math></alternatives></inline-formula> based on both the input image <italic>x</italic> and the specified target domain label <italic>c</italic>, while <italic>D</italic> aims to differentiate between real and generated images. Throughout this paper, the probability distribution over sources provided by <italic>D</italic> is denoted as <inline-formula id="j_infor556_ineq_004"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mi mathvariant="italic">r</mml:mi>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${D_{src}}$]]></tex-math></alternatives></inline-formula>. The primary aim of the generator <italic>G</italic> is to minimize this objective, while the discriminator <italic>D</italic> aims to maximize it.</p>
<sec id="j_infor556_s_012">
<label>3.3.1</label>
<title>Multi-Domain Image-to-Image Translation</title>
<p><bold>Adversarial loss:</bold> Instead of utilizing the adversarial loss in typical GANs (Eq. (<xref rid="j_infor556_eq_001">1</xref>)), which has been observed to encounter challenges such as mode collapse, vanishing gradients, and hyper-parameter sensitivity, he regularized Wasserstein GAN with gradient penalty WGAN-GP is employed. This choice ensures stable training for both the generator and discriminator networks while also improving the overall image quality of the generated outputs. The WGAN-GP loss function is defined as follows: 
<disp-formula id="j_infor556_eq_002">
<label>(2)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">adv</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext mathvariant="italic">WGAN</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="italic">C</mml:mi>
<mml:mi mathvariant="italic">T</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>″</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {L_{\textit{adv}}}={L_{{\textit{WGAN}_{gp}}}}+{\lambda _{ct}}CT{|_{{x^{\prime }},{x^{\prime\prime }}}}.\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>In the above formulation, the initial term represents the WGAN-GP loss, while the second term introduces a consistency term to regularize this loss. The WGAN-GP loss is expressed as follows: 
<disp-formula id="j_infor556_eq_003">
<label>(3)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext mathvariant="italic">WGAN</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">E</mml:mi>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">[</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mi mathvariant="italic">r</mml:mi>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">]</mml:mo>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">[</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mi mathvariant="italic">r</mml:mi>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:mi mathvariant="italic">G</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">]</mml:mo>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">[</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo maxsize="1.19em" minsize="1.19em" stretchy="true">‖</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo>∇</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mi mathvariant="italic">r</mml:mi>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo maxsize="1.19em" minsize="1.19em" stretchy="true">‖</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">]</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {L_{{\textit{WGAN}_{gp}}}}=E\big[{D_{src}}(x)\big]+{E_{x,c}}\big[{D_{src}}\big(G(x,c)\big)\big]-{\lambda _{gp}}{E_{\hat{x}}}\big[{\big({\big\| {\nabla _{\hat{x}}}{D_{src}}(\hat{x})\big\| _{2}}-1\big)^{2}}\big],\]]]></tex-math></alternatives>
</disp-formula> 
where <inline-formula id="j_infor556_ineq_005"><alternatives><mml:math><mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">ˆ</mml:mo></mml:mover></mml:math><tex-math><![CDATA[$\hat{x}$]]></tex-math></alternatives></inline-formula> is sampled uniformly along a straight line between a pair of real and generated images. Specifically, Gulrajani <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor556_ref_014">2017</xref>) presented a gradient penalty approach, highlighting that a differentiable discriminator <inline-formula id="j_infor556_ineq_006"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
<mml:mi mathvariant="italic">r</mml:mi>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${D_{src}}$]]></tex-math></alternatives></inline-formula> maintains 1-Lipschitz continuity when the norm of its gradients remains at or below 1 universally. 
<disp-formula id="j_infor556_eq_004">
<label>(4)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">C</mml:mi>
<mml:mi mathvariant="italic">T</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>″</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo stretchy="false">∼</mml:mo>
<mml:mi mathvariant="italic">P</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">[</mml:mo>
<mml:mo movablelimits="false">max</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">d</mml:mi>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:mi mathvariant="italic">D</mml:mi>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">D</mml:mi>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo movablelimits="false">″</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo>+</mml:mo>
<mml:mn>0.1</mml:mn>
<mml:mo>·</mml:mo>
<mml:mi mathvariant="italic">d</mml:mi>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:mi mathvariant="italic">D</mml:mi>
<mml:mtext>_</mml:mtext>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">D</mml:mi>
<mml:mtext>_</mml:mtext>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo movablelimits="false">″</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo>−</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">]</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ CT{|_{{x^{\prime }},{x^{\prime\prime }}}}={E_{x\sim P}}\big[\max (0,d\big(D\big({x^{\prime }}\big),D\big({x^{\prime\prime }}\big)\big)+0.1\cdot d\big(D\text{\_}\big({x^{\prime }}\big),D\text{\_}\big({x^{\prime\prime }}\big)-{M^{\prime }}\big)\big].\]]]></tex-math></alternatives>
</disp-formula> 
In particular, <inline-formula id="j_infor556_ineq_007"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${x^{\prime }}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_infor556_ineq_008"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>″</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${x^{\prime\prime }}$]]></tex-math></alternatives></inline-formula> represent virtual data points that are in proximity to <italic>x</italic>. <inline-formula id="j_infor556_ineq_009"><alternatives><mml:math>
<mml:mi mathvariant="italic">D</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$D({x^{\prime }})$]]></tex-math></alternatives></inline-formula> represents the discriminator’s output when provided with input <italic>x</italic>, while <inline-formula id="j_infor556_ineq_010"><alternatives><mml:math>
<mml:mi mathvariant="italic">D</mml:mi>
<mml:mtext>_</mml:mtext></mml:math><tex-math><![CDATA[$D\text{\_}$]]></tex-math></alternatives></inline-formula> refers to the output of the second-to-last layer discriminator. And <italic>d</italic> represents the distance between two data points. <inline-formula id="j_infor556_ineq_011"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${M^{\prime }}$]]></tex-math></alternatives></inline-formula> in the above formula is a constant by which <italic>d</italic> is bounded as assumed by Wei <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor556_ref_025">2018</xref>)</p>
<p><bold>Contrast classification loss:</bold> In order to achieve the desired image translation, an auxiliary classifier is introduced on top of the discriminator <italic>D</italic>. This involves optimizing <italic>D</italic> using a domain classification loss for real images and optimizing <italic>G</italic> using a domain classification loss for fake images. The following are defined for real images: 
<disp-formula id="j_infor556_eq_005">
<label>(5)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">cls</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">r</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">[</mml:mo>
<mml:mo>−</mml:mo>
<mml:mo movablelimits="false">log</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">cls</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">]</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {L_{\textit{cls}}^{r}}={E_{x,{c^{\prime }}}}\big[-\log {D_{\textit{cls}}}\big({c^{\prime }}|x\big)\big],\]]]></tex-math></alternatives>
</disp-formula> 
where the expression <inline-formula id="j_infor556_ineq_012"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">cls</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${D_{\textit{cls}}}({c^{\prime }}|x)$]]></tex-math></alternatives></inline-formula> denotes a probability distribution over domain labels determined by <italic>D</italic>. Through minimizing this objective, <italic>D</italic> is trained to classify a real image <italic>x</italic> to its original domain <inline-formula id="j_infor556_ineq_013"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${c^{\prime }}$]]></tex-math></alternatives></inline-formula>. It is assumed that the training data provides the input image and its associated domain label pair <inline-formula id="j_infor556_ineq_014"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(x,{c^{\prime }})$]]></tex-math></alternatives></inline-formula>. On the flip side, the below formula is the loss function for distinguishing fake images within the domain: 
<disp-formula id="j_infor556_eq_006">
<label>(6)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">cls</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">f</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">[</mml:mo>
<mml:mo>−</mml:mo>
<mml:mo movablelimits="false">log</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">cls</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">G</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">]</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {L_{\textit{cls}}^{f}}={E_{x,{c^{\prime }}}}\big[-\log {D_{\textit{cls}}}\big({c^{\prime }}|G(x,c)\big)\big].\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>In particular, <italic>x</italic> and <inline-formula id="j_infor556_ineq_015"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${c^{\prime }}$]]></tex-math></alternatives></inline-formula> represent a real image and its original label, respectively. <inline-formula id="j_infor556_ineq_016"><alternatives><mml:math>
<mml:mi mathvariant="italic">G</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$G(x,c)$]]></tex-math></alternatives></inline-formula> and <italic>c</italic> denote a generated (fake) image and the desired target contrast.</p>
<p>For this experiment, Softmax Cross Entropy is utilized to compute the classification loss.</p>
<p><bold>Cycle consistency loss:</bold> To enforce a reliable one-to-one mapping between the source and target image domains, a cycle consistency loss is utilized in the compound loss. This loss term ensures that the generated images can faithfully reconstruct the original source image when translated back to the source domain. By incorporating the cycle consistency loss, the model becomes more robust and capable of producing consistent and accurate image translations. In this experiment, L1-norm is applied to compute cycle consistency loss: 
<disp-formula id="j_infor556_eq_007">
<label>(7)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">[</mml:mo>
<mml:mo maxsize="1.19em" minsize="1.19em" stretchy="true">‖</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">G</mml:mi>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:mi mathvariant="italic">G</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo maxsize="1.19em" minsize="1.19em" stretchy="true">‖</mml:mo>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">]</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {L_{cyc}}={E_{x,c,{c^{\prime }}}}\big[\big\| x-G\big(G(x,c),{c^{\prime }}\big)\big\| \big].\]]]></tex-math></alternatives>
</disp-formula>
</p>
</sec>
<sec id="j_infor556_s_013">
<label>3.3.2</label>
<title>New Generation Loss Functions with DSSIM and LPIPS</title>
<p>For the generator, the L1-norm has already been applied for cycle-consistent loss, but the current L1 loss disregards patch-level dissimilarity among images, limiting the information available for the generator. To overcome this limitation, two additional terms are introduced in the generation loss, promoting small-scale dissimilarity between real and reconstructed images at the patch level. This enhances the generator’s available information.</p>
<p>Motivated by the effectiveness of measuring structural similarity between images using the Structural Similarity (SSIM) metric in a patch-wise manner, the Structural Dissimilarity Loss (DSSIM) is adopted as an extension of SSIM: 
<disp-formula id="j_infor556_eq_008">
<label>(8)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">DSSIM</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" maxsize="2.03em" minsize="2.03em">[</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>−</mml:mo>
<mml:mtext mathvariant="italic">SSIM</mml:mtext>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">G</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">G</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo fence="true" maxsize="2.03em" minsize="2.03em">]</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {L_{\textit{DSSIM}}}={E_{x,c,{c^{\prime }}}}\bigg[\frac{(1-\textit{SSIM}(x-G(G(x,c),{c^{\prime }})))}{2}\bigg].\]]]></tex-math></alternatives>
</disp-formula> 
In particular, SSIM is formulated as: 
<disp-formula id="j_infor556_eq_009">
<label>(9)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mtext mathvariant="italic">SSIM</mml:mtext>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>2</mml:mn>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>+</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>2</mml:mn>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>+</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>+</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \textit{SSIM}(x,y)=\frac{(2{\mu _{x}}{\mu _{y}}+{C_{1}})+(2{\sigma _{x}}y+{C_{2}})}{({\mu _{x}^{2}}+{\mu _{y}^{2}}+{C_{1}})({\sigma _{x}^{2}}+{\sigma _{y}^{2}}+{C_{2}})},\]]]></tex-math></alternatives>
</disp-formula> 
with:</p>
<p><inline-formula id="j_infor556_ineq_017"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mu _{x}}$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_infor556_ineq_018"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\mu _{y}}$]]></tex-math></alternatives></inline-formula>: the pixel sample mean of <italic>x</italic> and <italic>y</italic>, respectively;</p>
<p><inline-formula id="j_infor556_ineq_019"><alternatives><mml:math>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup></mml:math><tex-math><![CDATA[${\sigma _{x}^{2}}$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_infor556_ineq_020"><alternatives><mml:math>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup></mml:math><tex-math><![CDATA[${\sigma _{y}^{2}}$]]></tex-math></alternatives></inline-formula>: the variance of <italic>x</italic> and <italic>y</italic>, respectively;</p>
<p><inline-formula id="j_infor556_ineq_021"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="italic">y</mml:mi></mml:math><tex-math><![CDATA[${\sigma _{x}}y$]]></tex-math></alternatives></inline-formula>: the covariance of <italic>x</italic> and <italic>y</italic>;</p>
<p><inline-formula id="j_infor556_ineq_022"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${C_{1}}={({k_{1}}L)^{2}}$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_infor556_ineq_023"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="italic">L</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${C_{2}}={({k_{2}}L)^{2}}$]]></tex-math></alternatives></inline-formula>: two variables to stabilize the division with weak denominator;</p>
<p><italic>L</italic>: the dynamic range of the pixel-values (typically this is <inline-formula id="j_infor556_ineq_024"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">#</mml:mi>
<mml:mtext mathvariant="italic">bits per pixel</mml:mtext>
</mml:mrow>
</mml:msup>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[${2^{\mathrm{\# }\textit{bits per pixel}}}-1$]]></tex-math></alternatives></inline-formula>);</p>
<p><inline-formula id="j_infor556_ineq_025"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>0.01</mml:mn></mml:math><tex-math><![CDATA[${k_{1}}=0.01$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_infor556_ineq_026"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">k</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>0.03</mml:mn></mml:math><tex-math><![CDATA[${k_{2}}=0.03$]]></tex-math></alternatives></inline-formula> by default.</p>
<p>Additionally, to ensure that the generator continuously generates images that closely match the target contrast, the Learned Perceptual Image Patch Similarity (LPIPS) metric is incorporated, which has been recently proposed: 
<disp-formula id="j_infor556_eq_010">
<label>(10)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">LPIPS</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">[</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">G</mml:mi>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:mi mathvariant="italic">G</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">]</mml:mo>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {L_{\textit{LPIPS}}}={E_{x,c,{c^{\prime }}}}\big[x-G\big(G(x,c),c\big)\big].\]]]></tex-math></alternatives>
</disp-formula> 
The ultimate reconstruction loss combines the desirable aspects of all three terms: 
<disp-formula id="j_infor556_eq_011">
<label>(11)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">rec</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">DSSIM</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">DSSIM</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">LPIPS</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">LPIPS</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {L_{\textit{rec}}}={\lambda _{cyc}}{L_{cyc}}+{\lambda _{\textit{DSSIM}}}{L_{\textit{DSSIM}}}+{\lambda _{\textit{LPIPS}}}{L_{\textit{LPIPS}}}.\]]]></tex-math></alternatives>
</disp-formula>
</p>
</sec>
<sec id="j_infor556_s_014">
<label>3.3.3</label>
<title>Full Objective</title>
<p>In summary, the proposed loss function for the generator and discriminator networks can be formulated as follows: <disp-formula-group id="j_infor556_dg_001">
<disp-formula id="j_infor556_eq_012">
<label>(12)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">D</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">adv</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">cls</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">r</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}& {L_{D}}=-{L_{\textit{adv}}}+{L_{\textit{cls}}^{r}},\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
<disp-formula id="j_infor556_eq_013">
<label>(13)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">G</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">adv</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">cls</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">cls</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">f</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">rec</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">rec</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}& {L_{G}}={L_{\textit{adv}}}+{\lambda _{\textit{cls}}}{L_{\textit{cls}}^{f}}+{\lambda _{\textit{rec}}}{L_{\textit{rec}}}.\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
</disp-formula-group></p>
</sec>
</sec>
</sec>
<sec id="j_infor556_s_015">
<label>4</label>
<title>Experimental Results</title>
<sec id="j_infor556_s_016">
<label>4.1</label>
<title>Dataset Detail and Data Preprocessing</title>
<p>To evaluate the method, 2 standard medical image datasets, namely BraTS2020 and IXI, are adopted in this study.</p>
<sec id="j_infor556_s_017">
<label>4.1.1</label>
<title>Dataset Detail</title>
<p><bold>BraTS2020:</bold> The BraTS2020 dataset is a widely used benchmark dataset for brain tumor segmentation in magnetic resonance imaging (MRI). It consists of multi-modal MRI scans, including T1-weighted, T1-weighted contrast-enhanced, T2-weighted, and Fluid Attenuated Inversion Recovery (FLAIR) images (indicated in Table <xref rid="j_infor556_tab_001">1</xref>).</p>
<p><bold>IXI dataset:</bold> The IXI (Information eXtraction from Images) dataset is a comprehensive and publicly available collection of multimodal brain imaging data. 600 MR images from normal, healthy subjects are collected with 4 acquisition protocols: T1-weighted, T2-weighted, Proton Density (PD), and Magnetic Resonance Angiography (MRA) (indicated in Table <xref rid="j_infor556_tab_002">2</xref>).</p>
<table-wrap id="j_infor556_tab_001">
<label>Table 1</label>
<caption>
<p>Detailed information on the BraTS2020 dataset.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Attributes</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Detailed information</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">Number of contrasts</td>
<td style="vertical-align: top; text-align: left">4 (T1, T2, T1CE, FLAIR)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Number of samples per contrast</td>
<td style="vertical-align: top; text-align: left">494 images/contrast</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Image size (3D image)</td>
<td style="vertical-align: top; text-align: left">Already registered</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">All images: <inline-formula id="j_infor556_ineq_027"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>240</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>240</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>155</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(240\times 240\times 155)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">Ratio healthy/disease (tumour)</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">40/60</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="j_infor556_tab_002">
<label>Table 2</label>
<caption>
<p>Detailed information on the IXI dataset.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Attributes</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Detailed information</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">Number of contrasts</td>
<td style="vertical-align: top; text-align: left">4 (T1, T2, PD, MRA)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Number of samples per contrast</td>
<td style="vertical-align: top; text-align: left">568 images/contrast</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Image size (3D image)</td>
<td style="vertical-align: top; text-align: left">Not registered:</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">T1: <inline-formula id="j_infor556_ineq_028"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>256</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>256</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>150</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(256\times 256\times 150)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">T2: <inline-formula id="j_infor556_ineq_029"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>256</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>256</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>130</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(256\times 256\times 130)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">MRA: <inline-formula id="j_infor556_ineq_030"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>512</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>512</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>100</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(512\times 512\times 100)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">PD: <inline-formula id="j_infor556_ineq_031"><alternatives><mml:math>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mn>256</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>256</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>130</mml:mn>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$(256\times 256\times 130)$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">Ratio healthy/disease (tumour)</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">100/0</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="j_infor556_s_018">
<label>4.1.2</label>
<title>Data Preprocessing</title>
<p>For both datasets, the nibabel library is employed to manage the loading and saving of MRI image files in .nii format.</p>
<p><bold>BraTS2020:</bold> Since all of the images had already been registered to the same template with size (240, 240, 155), there is no need to apply any registering method for this dataset. The middle axial slices of 4 contrasts are shown in Fig. <xref rid="j_infor556_fig_009">9</xref>.</p>
<fig id="j_infor556_fig_009">
<label>Fig. 9</label>
<caption>
<p>Illustration of axial slices of the BraTS2020 dataset.</p>
</caption>
<graphic xlink:href="infor556_g009.jpg"/>
</fig>
<p>The original BraTS2020 dataset has already been divided into a training set and a validation set, as presented in Table <xref rid="j_infor556_tab_003">3</xref>. Since the images of the BraTS2020 dataset provide better resolution in the axial plane, axial slices of all images were taken. For this experiment, 10 slices per 3D image are taken.</p>
<p><bold>IXI dataset:</bold> Upon reviewing and analysing the entire IXI dataset, it was observed that there are 2 contrasts that exist in different templates from the remaining. The middle original slices of the 4 contrasts are shown in Fig. <xref rid="j_infor556_fig_010">10</xref>:</p>
<table-wrap id="j_infor556_tab_003">
<label>Table 3</label>
<caption>
<p>Number of samples per contrast in the BraTS2020 dataset.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Train</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">FLAIR</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">3690</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">T1</td>
<td style="vertical-align: top; text-align: left">3690</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">T1CE</td>
<td style="vertical-align: top; text-align: left">3690</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">T2</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">3690</td>
</tr>
</tbody>
</table>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Test</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">FLAIR</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">1250</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">T1</td>
<td style="vertical-align: top; text-align: left">1250</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">T1CE</td>
<td style="vertical-align: top; text-align: left">1250</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">T2</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">1250</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="j_infor556_fig_010">
<label>Fig. 10</label>
<caption>
<p>Illustration of axial slices of IXI dataset before registration.</p>
</caption>
<graphic xlink:href="infor556_g010.jpg"/>
</fig>
<p>As depicted in Fig. <xref rid="j_infor556_fig_010">10</xref>, it is evident that MRA images are in shapes (512, 512, 100), while T1 images are in completely different templates (size and plane). However, it is noteworthy that there are still PD and T2 images in the same template and similar to BraTS2020 images. In light of this observation, the decision is made to apply some registering transformations to MRA and T1 images using the common template synthesized from PD and T2. For this experiment, an affine transformation is employed from the AntsPy package (Avants <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_002">2009</xref>) which is a Python library that wraps the C++ biomedical image processing library ANTs. First, a common template is extracted from two classes: PD and T2. Subsequently, a mask is derived from this template and subsequently the T1 and MRA images are registered onto the mask using affine transformation, as mentioned in Sohail <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor556_ref_023">2019</xref>). The preprocessing process for the IXI dataset is depicted in Figs. <xref rid="j_infor556_fig_011">11</xref>, <xref rid="j_infor556_fig_012">12</xref>, and <xref rid="j_infor556_fig_013">13</xref>.</p>
<fig id="j_infor556_fig_011">
<label>Fig. 11</label>
<caption>
<p>The axial slices of a common template synthesized from PD and T2 images.</p>
</caption>
<graphic xlink:href="infor556_g011.jpg"/>
</fig>
<fig id="j_infor556_fig_012">
<label>Fig. 12</label>
<caption>
<p>The axial slices of MRA/T1 when embedded in the common template.</p>
</caption>
<graphic xlink:href="infor556_g012.jpg"/>
</fig>
<fig id="j_infor556_fig_013">
<label>Fig. 13</label>
<caption>
<p>The axial slices of MRA/T1 after applied affine transformation.</p>
</caption>
<graphic xlink:href="infor556_g013.jpg"/>
</fig>
<p>This provides 568 images of the same size and position from which 68 were randomly selected for testing while the remaining 500 were used for training. Since the MRA images of the IXI dataset provide better resolution in the axial plane, axial slices of all images were taken (as shown in Fig. <xref rid="j_infor556_fig_014">14</xref>). For the IXI dataset, only 5 middle slices per 3D image are selected. This decision is made due to the presence of missing or noisy images in some contrasts (as shown in Figs. <xref rid="j_infor556_fig_015">15</xref> and <xref rid="j_infor556_fig_016">16</xref>), particularly in MRA, when attempting to take 8 or 10 slices.</p>
<fig id="j_infor556_fig_014">
<label>Fig. 14</label>
<caption>
<p>Illustration of axial slices of IXI dataset after registration.</p>
</caption>
<graphic xlink:href="infor556_g014.jpg"/>
</fig>
<fig id="j_infor556_fig_015">
<label>Fig. 15</label>
<caption>
<p>Missing images in MRA contrast.</p>
</caption>
<graphic xlink:href="infor556_g015.jpg"/>
</fig>
<p>For data preprocessing, transformations including horizontal flipping, resizing to <inline-formula id="j_infor556_ineq_032"><alternatives><mml:math>
<mml:mn>256</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>256</mml:mn></mml:math><tex-math><![CDATA[$256\times 256$]]></tex-math></alternatives></inline-formula>, and normalization are applied. Table <xref rid="j_infor556_tab_004">4</xref> and Fig. <xref rid="j_infor556_fig_017">17</xref> indicate the number of samples in the IXI dataset and visualization of the distribution of actual collected data from both datasets.</p>
<fig id="j_infor556_fig_016">
<label>Fig. 16</label>
<caption>
<p>Blur images in MRA contrast.</p>
</caption>
<graphic xlink:href="infor556_g016.jpg"/>
</fig>
<table-wrap id="j_infor556_tab_004">
<label>Table 4</label>
<caption>
<p>Number of samples per contrast in the IXI dataset.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Train</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">MRA</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">2430</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">PD</td>
<td style="vertical-align: top; text-align: left">2430</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">T1</td>
<td style="vertical-align: top; text-align: left">2430</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">T2</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">2430</td>
</tr>
</tbody>
</table>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Test</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">MRA</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">335</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">PD</td>
<td style="vertical-align: top; text-align: left">335</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">T1</td>
<td style="vertical-align: top; text-align: left">335</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">T2</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">335</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="j_infor556_fig_017">
<label>Fig. 17</label>
<caption>
<p>Distribution of data from both datasets divided by datasets (BraTS2020/IXI).</p>
</caption>
<graphic xlink:href="infor556_g017.jpg"/>
</fig>
<fig id="j_infor556_fig_018">
<label>Fig. 18</label>
<caption>
<p>Distribution of data from both datasets divided by brain’s condition (healthy/diseases).</p>
</caption>
<graphic xlink:href="infor556_g018.jpg"/>
</fig>
<p>Regarding the number of samples per contrast, as depicted in Fig. <xref rid="j_infor556_fig_017">17</xref>, a notable imbalance is observed, with a higher number of samples in T1 and T2 compared to the other contrasts. However, this disparity falls within an acceptable range. In Fig. <xref rid="j_infor556_fig_018">18</xref>, there is a significant disparity in the number of samples between the two brain conditions (healthy/diseased). However, as the main purpose of this experiment is to focus on synthesizing MRI in general, this problem would be resolved in future approaches.</p>
<p>In Fig. <xref rid="j_infor556_fig_018">18</xref>, a substantial disparity in the number of samples between the two brain conditions (healthy/diseased) is evident. Nonetheless, it is important to note that the primary objective of this experiment is to focus on synthesizing MRI in general. The issue of sample disparity will be acknowledged and addressed in future approaches.</p>
</sec>
</sec>
<sec id="j_infor556_s_019">
<label>4.2</label>
<title>Experimental Results</title>
<sec id="j_infor556_s_020">
<label>4.2.1</label>
<title>Implementation Details</title>
<p>All experiments were conducted using the PyTorch framework. To accommodate computational limitations, the image slices were resized to dimensions of <inline-formula id="j_infor556_ineq_033"><alternatives><mml:math>
<mml:mn>256</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>256</mml:mn></mml:math><tex-math><![CDATA[$256\times 256$]]></tex-math></alternatives></inline-formula>. During training, the input images and target contrast were randomly selected in an unpaired manner to ensure fairness in comparison (Xiang <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_026">2018</xref>). Both the default StarGAN model and the proposed models used the same set of hyperparameter values. The training was carried out for 200,000 iterations with a batch size of 4. Through numerous hands-on trials, this study observed that the loss began to rise when the iteration count was configured to exceed 200,000. Furthermore, due to hardware constraints, the batch size could only be set to a value below 4 to prevent training process crashes. All models are trained using Adam optimizer with <inline-formula id="j_infor556_ineq_034"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>0.5</mml:mn></mml:math><tex-math><![CDATA[${\beta _{1}}=0.5$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_infor556_ineq_035"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">β</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>0.999</mml:mn></mml:math><tex-math><![CDATA[${\beta _{2}}=0.999$]]></tex-math></alternatives></inline-formula>, as experienced by Kingma and Ba (<xref ref-type="bibr" rid="j_infor556_ref_016">2014</xref>). All models are trained with a learning rate of 0.0001 for the initial 100,000 iterations, followed by a linear decay of the learning rate to 0 over the next 100,000 iterations. In this study, adjustments were made to the learning rate, resulting in a marginal fluctuation of the loss value, which proved challenging to converge. The training process takes about one day when conducted on a single NVIDIA RTX-3060 GPU.</p>
<p>During the training process, the following parameters are employed: where <inline-formula id="j_infor556_ineq_036"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\lambda _{ct}}$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_infor556_ineq_037"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${\lambda _{gp}}$]]></tex-math></alternatives></inline-formula> are hyper-parameters that control the relative importance of consistent regularization and gradient penalty, respectively. Within it, consistent regularization plays a supportive role in complementing and improving the gradient penalty. Specifically, <inline-formula id="j_infor556_ineq_038"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[${\lambda _{ct}}=1$]]></tex-math></alternatives></inline-formula> (Eq. (<xref rid="j_infor556_eq_002">2</xref>)) and <inline-formula id="j_infor556_ineq_039"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">g</mml:mi>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>10</mml:mn></mml:math><tex-math><![CDATA[${\lambda _{gp}}=10$]]></tex-math></alternatives></inline-formula> (Eq. (<xref rid="j_infor556_eq_003">3</xref>)) were selected after experimenting with neighbouring value ranges. In the training of this image generator, prioritization is given to image quality over classification. The reason is that image classification is relatively straightforward and dependent on the task of synthesizing images; the closer the generated images are to the ground truth, the higher the accuracy of the classification task will be. Additionally, image reconstruction is as important as image generation, leading to the setting of <inline-formula id="j_infor556_ineq_040"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">cls</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[${\lambda _{\textit{cls}}}=1$]]></tex-math></alternatives></inline-formula> (Eq. (<xref rid="j_infor556_eq_013">13</xref>)) and <inline-formula id="j_infor556_ineq_041"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">rec</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[${\lambda _{\textit{rec}}}=1$]]></tex-math></alternatives></inline-formula> (Eq. (<xref rid="j_infor556_eq_013">13</xref>)). As stated previously, when <inline-formula id="j_infor556_ineq_042"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${M^{\prime }}$]]></tex-math></alternatives></inline-formula> varies within the range of 0 to 0.3, the experiments yield similar outcomes. Consequently, <inline-formula id="j_infor556_ineq_043"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn></mml:math><tex-math><![CDATA[${M^{\prime }}=0$]]></tex-math></alternatives></inline-formula> (Eq. (<xref rid="j_infor556_eq_004">4</xref>)) and <inline-formula id="j_infor556_ineq_044"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">DSSIM</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">λ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">LPIPS</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mn>10</mml:mn></mml:math><tex-math><![CDATA[${\lambda _{cyc}}={\lambda _{\textit{DSSIM}}}={\lambda _{\textit{LPIPS}}}=10$]]></tex-math></alternatives></inline-formula> (Eq. (<xref rid="j_infor556_eq_011">11</xref>)) are set uniformly across all experiments.</p>
</sec>
<sec id="j_infor556_s_021">
<label>4.2.2</label>
<title>Mask Vector for Training on Multiple Datasets</title>
<p>The aim of this experiment is to integrate multiple datasets during model training to mitigate the limitations posed by data scarcity in medical image synthesis. However, a challenge arises when incorporating multiple datasets, as each dataset only holds partial knowledge regarding the label information.</p>
<p>In the specific scenario of using datasets like BraTS2020 and IXI, a limitation arises where one dataset (BraTS2020) includes labels for contrasts: T1, T2, T1CE, and FLAIR but lacks labels for other contrasts such as PD and MRA, while the other dataset (IXI) is the opposite (lack of T1CE and FLAIR). This poses a challenge because the reconstruction of the input image <italic>x</italic> from the translated image <inline-formula id="j_infor556_ineq_045"><alternatives><mml:math>
<mml:mi mathvariant="italic">G</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$G(x,c)$]]></tex-math></alternatives></inline-formula> requires complete information on the label vector <inline-formula id="j_infor556_ineq_046"><alternatives><mml:math>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>′</mml:mo>
</mml:mrow>
</mml:msup></mml:math><tex-math><![CDATA[${c^{\prime }}$]]></tex-math></alternatives></inline-formula>.</p>
<p>To address this issue, a mask vector is adopted, denoted as <italic>m</italic>, which was introduced in the StarGAN framework. The purpose of the mask vector is to allow the model to ignore unspecified labels and focus only on the explicitly known label provided by a particular dataset. In the context of the paper, <italic>m</italic> is an n-dimensional one-hot vector, where <italic>n</italic> is the number of datasets. Each element of the mask vector indicates whether the corresponding dataset label is specified or not. 
<disp-formula id="j_infor556_eq_014">
<label>(14)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:munder accentunder="false">
<mml:mrow>
<mml:mi mathvariant="italic">c</mml:mi>
</mml:mrow>
<mml:mo accent="true">_</mml:mo></mml:munder>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">c</mml:mi>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \underline{c}=[c1,\dots ,cn,m]\]]]></tex-math></alternatives>
</disp-formula>
</p>
<p>For example, if <inline-formula id="j_infor556_ineq_047"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>2</mml:mn></mml:math><tex-math><![CDATA[$n=2$]]></tex-math></alternatives></inline-formula> as mentioned in the paper (representing two datasets, BraTS2020 and IXI), and if a sample belongs to the BraTS2020 dataset, then the mask vector <italic>m</italic> can be represented as <inline-formula id="j_infor556_ineq_048"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[1,0]$]]></tex-math></alternatives></inline-formula>, indicating that the first dataset label (BraTS2020) is specified while the second dataset label (IXI) is unspecified. This approach helps the model to handle multiple datasets efficiently, focusing on relevant labels while ignoring irrelevant ones.</p>
<p>In the defined expression, [·] denotes concatenation, and represents a vector containing the labels for the <italic>i</italic>th dataset. The vector, which represents the known labels, can take the form of a binary vector for binary attributes or a one-hot vector for categorical attributes. The remaining <inline-formula id="j_infor556_ineq_049"><alternatives><mml:math>
<mml:mi mathvariant="italic">n</mml:mi>
<mml:mo>−</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$n-1$]]></tex-math></alternatives></inline-formula> unknown labels are assigned zero values. In the specific experiments conducted, the BraTS2020 and IXI datasets are utilized, resulting in a value of n equal to two.</p>
</sec>
<sec id="j_infor556_s_022">
<label>4.2.3</label>
<title>Qualitative Evaluation</title>
<fig id="j_infor556_fig_019">
<label>Fig. 19</label>
<caption>
<p>Synthesis of MRA, PD-weighted and T1-weighted images using a single T2-weighted image as input (IXI dataset).</p>
</caption>
<graphic xlink:href="infor556_g019.jpg"/>
</fig>
<fig id="j_infor556_fig_020">
<label>Fig. 20</label>
<caption>
<p>Synthesis of T1-weighted, T2-weighted, and FLAIR images using a single T1CE-weighted image as input (BraTS2020 dataset).</p>
</caption>
<graphic xlink:href="infor556_g020.jpg"/>
</fig>
<p>Figures <xref rid="j_infor556_fig_019">19</xref> and <xref rid="j_infor556_fig_020">20</xref> provide a qualitative comparison between the proposed method and Star-GAN in terms of multi-contrast synthesis. The results demonstrate that Star-GAN partly fails to capture structural and perceptual similarities for small anatomical regions, which are effectively captured by the proposed method. In Fig. <xref rid="j_infor556_fig_019">19</xref>, the synthesis of MRA from a T2-weighted image shows that Star-GAN struggles to reproduce the accurate colour of the image, while the proposed method successfully generates an image that is nearly identical to the real one. Similarly, Fig. <xref rid="j_infor556_fig_020">20</xref> with FLAIR contrast.</p>
</sec>
<sec id="j_infor556_s_023">
<label>4.2.4</label>
<title>Quantitative Evaluation</title>
<p>In order to evaluate the performance of the proposed model trained on multiple datasets compared to Star-GAN and the training experiment conducted on a single dataset, well-established metrics such as peak signal-to-noise ratio (PSNR), structural similarity index (SSIM) [21], and normalized mean absolute error (NMAE) are utilized [22]. The average results on BraTS2020 and IXI datasets are presented in Table <xref rid="j_infor556_tab_005">5</xref>. A higher PSNR and SSIM, as well as a lower NMAE, indicate better quality in the generated images. The experimental evaluation encompasses StarGANs (SG), ResUNet (RU), and the proposed methods: With two training strategies (training on a single dataset and multiple datasets); With new generation losses and new generator architecture. The proposed approach has acquired better performance than that of Star-GAN in evaluating PSNR and NMAE, demonstrating superior performance.</p>
<table-wrap id="j_infor556_tab_005">
<label>Table 5</label>
<caption>
<p>Quantitative results.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Metric</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">SG single<sup>1</sup></td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">RU single<sup>2</sup></td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">SG both<sup>3</sup></td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">RU both<sup>4</sup></td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Proposed method<sup>∗</sup></td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">BraTS2020</td>
<td style="vertical-align: top; text-align: left">SSIM</td>
<td style="vertical-align: top; text-align: left">0.8166</td>
<td style="vertical-align: top; text-align: left">0.7672</td>
<td style="vertical-align: top; text-align: left"><bold>0.8238</bold></td>
<td style="vertical-align: top; text-align: left">0.7734</td>
<td style="vertical-align: top; text-align: left">0.8167</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">PSNR</td>
<td style="vertical-align: top; text-align: left">34.578</td>
<td style="vertical-align: top; text-align: left"><bold>35.959</bold></td>
<td style="vertical-align: top; text-align: left">34.632</td>
<td style="vertical-align: top; text-align: left">35.327</td>
<td style="vertical-align: top; text-align: left">35.847</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">NMAE</td>
<td style="vertical-align: top; text-align: left">0.0426</td>
<td style="vertical-align: top; text-align: left"><bold>0.0412</bold></td>
<td style="vertical-align: top; text-align: left">0.0431</td>
<td style="vertical-align: top; text-align: left">0.0427</td>
<td style="vertical-align: top; text-align: left">0.0414</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">IXI</td>
<td style="vertical-align: top; text-align: left">SSIM</td>
<td style="vertical-align: top; text-align: left">0.6883</td>
<td style="vertical-align: top; text-align: left">0.6854</td>
<td style="vertical-align: top; text-align: left">0.7310</td>
<td style="vertical-align: top; text-align: left">0.6913</td>
<td style="vertical-align: top; text-align: left"><bold>0.6972</bold></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">PSNR</td>
<td style="vertical-align: top; text-align: left">38.289</td>
<td style="vertical-align: top; text-align: left">40.214</td>
<td style="vertical-align: top; text-align: left">40.469</td>
<td style="vertical-align: top; text-align: left">40.848</td>
<td style="vertical-align: top; text-align: left"><bold>40.882</bold></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">NMAE</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.0437</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.0399</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.0392</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.0387</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><bold>0.0384</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
 <p><sup>1</sup>StarGANs model trained on 1 dataset.</p> 
<p><sup>2</sup>Model with ResUnet generator trained on 1 dataset.</p> 
<p><sup>3</sup>StarGANs model trained on multiple datasets.</p> 
<p><sup>4</sup>Model with ResUnet generator trained on multiple datasets.</p> 
</table-wrap-foot>
</table-wrap>
<p>Regarding the comparison of architecture between the two models, it is observed that the StarGANs-based with ResUNet (RU) and proposed generator outperformed default StarGANs (SG) in most of the metric values, particularly in terms of metrics PSNR and NMAE, for both training strategies. Table <xref rid="j_infor556_tab_005">5</xref> reveals that model RU single yields the best results in 2 metrics on the BraTS2020 dataset with PSNR = 35.959 and NMAE = 0.0412, followed by the second-best performance from the proposed model. However, because RU single is trained solely on the BraTS2020 dataset, which contains both healthy normal brain MRI images and brains with tumours or abnormalities, instead of being trained on both BraTS2020 and IXI datasets, it’s plausible that it may achieve superior performance compared to a model trained on both datasets. This is because the IXI dataset only consists of healthy normal brain MRI images. Upon close examination, the difference between the top-performing outcome and the second-best outcome is minimal. Additionally, the proposed model exhibited the best performance across all three metrics on the IXI dataset. This underscores the model’s robustness and its capability to adapt effectively when trained on diverse datasets.</p>
<p>On the other hand, when comparing the metrics between training on a single dataset and training on multiple datasets on the same generator architecture, results show that training on multiple datasets yielded comparable or sometimes even better results compared to training on individual datasets. This approach can be applied to reduce time and computational costs without compromising the model’s performance (results in Table <xref rid="j_infor556_tab_005">5</xref>). After applying the proposed generator architecture, new generation losses, and training on both datasets. The results (*) are shown in the last column in Table <xref rid="j_infor556_tab_005">5</xref>. demonstrate superior effectiveness compared to previous methods, as indicated by higher SSIM and PSNR values and lower NMAE.</p>
</sec>
</sec>
<sec id="j_infor556_s_024">
<label>4.3</label>
<title>Experience in Training</title>
<p>Throughout the training process, a rapid decrease in classification loss was observed, particularly within the first 1000 iterations for both datasets. However, with the IXI dataset, the classification loss remained stable at a value of zero. By contrast, for the BraTS2020 dataset, the classification loss occasionally fluctuated away from zero. This can be understood as the MRI images in the BraTS2020 dataset having similar colour characteristics, making it challenging for the model to learn these specific features and smaller details such as tumours and subtle brain oedema. On the other hand, the IXI dataset consists of healthy, normal brain images with relatively different colour characteristics between contrasts, providing more distinguishable features for the model to learn.</p>
</sec>
</sec>
<sec id="j_infor556_s_025">
<label>5</label>
<title>Demonstration</title>
<p>A website has been developed as a platform to showcase the conducted experiments. This website serves as a repository of the research findings and provides a convenient means for accessing the experimental data, results, and related materials. The repository can be accessed via <uri>https://github.com/hanahh080601/Generate-MRI-FastAP</uri>. As illustrated in Fig. <xref rid="j_infor556_fig_021">21</xref>, the website offers a user-friendly interface, allowing visitors to navigate through the MRI Synthesis experiments, explore the data, and gain insights into the research process. It serves as a valuable resource for researchers, practitioners, and interested individuals to access and review the experimental work conducted in this study.</p>
<fig id="j_infor556_fig_021">
<label>Fig. 21</label>
<caption>
<p>Illustration of the user interface for the demo website.</p>
</caption>
<graphic xlink:href="infor556_g021.jpg"/>
</fig>
<p>Specifically, users can select the input contrast and output contrast they desire for image synthesis upon accessing the website. They can then choose to upload files from their local device (the uploaded image must match the selected input contrast). Alternatively, users may only select the input and output contrast, after which the system will generate image using input data from the BraTS2020 and IXI datasets. This feature is used to qualitatively assess the quality of the generated images compared to the original ones.</p>
<p>Based on observation in Fig. <xref rid="j_infor556_fig_021">21</xref>, T1 is selected as the input contrast, T2 as the output contrast, and the IXI dataset is the chosen dataset. The system automatically chose a random T1 image from the test set of the IXI dataset and sent a request for the model to generate the T2 image. Additionally, the system also selected the ground truth T2 image from the dataset and displayed both for user comparison.</p>
<fig id="j_infor556_fig_022">
<label>Fig. 22</label>
<caption>
<p>Illustration of the testing model with low-resolution input image.</p>
</caption>
<graphic xlink:href="infor556_g022.jpg"/>
</fig>
<fig id="j_infor556_fig_023">
<label>Fig. 23</label>
<caption>
<p>Illustration of the testing model with input image T2 exhibiting abnormal fluid with the same colour as cerebral spinal fluid.</p>
</caption>
<graphic xlink:href="infor556_g023.jpg"/>
</fig>
<p>Moreover, testing was carried out to assess the model’s performance when presented with noisy or partially noisy input images. The experimentation was carried out on MRI images from the test datasets (BraTS2020 and IXI), as well as random images sourced from the Internet. Notable cases include input images corrupted with noise and T2 input images exhibiting abnormal fluid with the same colour as cerebral spinal fluid. Observing Fig. <xref rid="j_infor556_fig_022">22</xref> with an input image of a noisy, low-resolution MRI T1 randomly obtained from the Internet, it can be seen that the generated image maintains the same quality (low resolution) as the input image, ensuring accuracy in colour contrast: cerebral spinal fluid, appearing dark in T1, manifests as bright in T2. In Fig. <xref rid="j_infor556_fig_023">23</xref>, with an input MRI T2 image from the BraTS2020 test dataset, the synthesized MRI FLAIR image exhibits relatively high image quality similar to the input image. Particularly, the model can learn and distinguish between cerebral spinal fluid (dark) and abnormal fluid (bright) in the generated FLAIR image while both appear bright in the T2 image. Furthermore, in reality, MRI images scanned by MRI machines may suffer from poor quality, noise, or misalignment due to objective and subjective factors, as illustrated by the REAL FLAIR image in Fig. <xref rid="j_infor556_fig_023">23</xref>. In such cases, the model-generated images rely solely on the input image without consideration of the MRI scanning process or subjective factors from the patient.</p>
<p>Therefore, this system can be utilized to synthesize multiple MRI images across various contrasts from just a single MRI image. This application can assist medical imaging specialists or diagnosing physicians by providing additional references in cases where they are uncertain about the original MRI image or desire to observe additional anatomical structures in a different contrast without requiring the patient to undergo additional MRI scans. This could help minimize costs, time, and partly support in making diagnoses.</p>
</sec>
<sec id="j_infor556_s_026">
<label>6</label>
<title>Conclustion</title>
<p>In this research, a novel method is proposed based on StarGANs, utilizing a ResUNet generator and a novel generation loss, for synthesizing multi-contrast MRI images. Despite substantial efforts made in prior studies (Bharti <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_003">2022</xref>; Dai <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_007">2020</xref>; Shen <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor556_ref_022">2020</xref>), the proposed approach only requires a single generator and discriminator. The results showcase that both qualitative and quantitative demonstrate the superiority of the proposed method over the default Star-GAN. Moreover, this study method overcomes the limitation of training multiple networks for multi-contrast image synthesis, which is crucial for deep learning methods reliant on multi-contrast data during training. The synthesized images produced by this method possess significant potential as highly valuable assets for medical professionals involved in research, education, and clinical applications.</p>
<p>Future work gears towards expanding the experiments to larger datasets in size and the number of contrasts, avoiding disparity in the proportion of healthy and diseased brain images. Additionally, efforts will be directed toward training the proposed approach with larger batch sizes and developing it into 3D images, which can significantly contribute to medical diagnosis in practical applications.</p>
</sec>
</body>
<back>
<app-group> 
<app id="j_infor556_app_001"><label>A</label>
<title>Network Architecture Details</title>
<table-wrap id="j_infor556_tab_006">
<label>Table 6</label>
<caption>
<p>Proposed generator architecture summary.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Layer</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Type</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Details</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">Input</td>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(13, 64, kernel_size=(7, 7), stride=(1, 1), padding=(3, 3), bias=False)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">InstanceNorm2d</td>
<td style="vertical-align: top; text-align: left">(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">ReLU</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">(inplace=True)</td>
</tr>
</tbody><tbody>
<tr>
<td style="vertical-align: top; text-align: left">down_block1</td>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(64, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">(ResBlock)</td>
<td style="vertical-align: top; text-align: left">InstanceNorm2d</td>
<td style="vertical-align: top; text-align: left">(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">ReLU</td>
<td style="vertical-align: top; text-align: left">(inplace=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(128, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">InstanceNorm2d</td>
<td style="vertical-align: top; text-align: left">(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">AvgPool2d</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">(kernel_size=2, stride=2, padding=0)</td>
</tr>
</tbody><tbody>
<tr>
<td style="vertical-align: top; text-align: left">down_block2</td>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(128, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">(ResBlock)</td>
<td style="vertical-align: top; text-align: left">InstanceNorm2d</td>
<td style="vertical-align: top; text-align: left">(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">ReLU</td>
<td style="vertical-align: top; text-align: left">(inplace=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">InstanceNorm2d</td>
<td style="vertical-align: top; text-align: left">(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">AvgPool2d</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">(kernel_size=2, stride=2, padding=0)</td>
</tr>
</tbody><tbody>
<tr>
<td style="vertical-align: top; text-align: left">bottle_neck</td>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1), bias=False)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">(6 x ResBlocks)</td>
<td style="vertical-align: top; text-align: left">InstanceNorm2d</td>
<td style="vertical-align: top; text-align: left">(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Each ResBlock →</td>
<td style="vertical-align: top; text-align: left">ReLU</td>
<td style="vertical-align: top; text-align: left">(inplace=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1), bias=False)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">InstanceNorm2d</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">(256, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)</td>
</tr>
</tbody><tbody>
<tr>
<td style="vertical-align: top; text-align: left">encoder</td>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(256, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">(ResBlock)</td>
<td style="vertical-align: top; text-align: left">InstanceNorm2d</td>
<td style="vertical-align: top; text-align: left">(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">ReLU</td>
<td style="vertical-align: top; text-align: left">(inplace=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(128, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">InstanceNorm2d</td>
<td style="vertical-align: top; text-align: left">(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">ConvTranspose2d</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">(128, 64, kernel_size=(2, 2), stride=(2, 2))</td>
</tr>
</tbody><tbody>
<tr>
<td style="vertical-align: top; text-align: left">up_block1</td>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(320, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">(ResBlock)</td>
<td style="vertical-align: top; text-align: left">InstanceNorm2d</td>
<td style="vertical-align: top; text-align: left">(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">ReLU</td>
<td style="vertical-align: top; text-align: left">(inplace=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(128, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">InstanceNorm2d</td>
<td style="vertical-align: top; text-align: left">(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">ConvTranspose2d</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">(128, 64, kernel_size=(2, 2), stride=(2, 2))</td>
</tr>
</tbody><tbody>
<tr>
<td style="vertical-align: top; text-align: left">up_block2</td>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">(ResBlock)</td>
<td style="vertical-align: top; text-align: left">InstanceNorm2d</td>
<td style="vertical-align: top; text-align: left">(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">ReLU</td>
<td style="vertical-align: top; text-align: left">(inplace=True)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(64, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">InstanceNorm2d</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)</td>
</tr>
</tbody><tbody>
<tr>
<td style="vertical-align: top; text-align: left">Output</td>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(64, 3, kernel_size=(7, 7), stride=(1, 1), padding=(3, 3), bias=False)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">Tanh</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">()</td>
</tr>
</tbody>
</table>
</table-wrap> 
<table-wrap id="j_infor556_tab_007">
<label>Table 7</label>
<caption>
<p>Discriminator architecture summary.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Layer</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Type</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Details</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">(main)</td>
<td style="vertical-align: top; text-align: left">Sequential</td>
<td style="vertical-align: top; text-align: left"/>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(3, 64, kernel_size=(4, 4), stride=(2, 2), padding=(1, 1))</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">LeakyReLU</td>
<td style="vertical-align: top; text-align: left">(negative_slope=0.01)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(64, 128, kernel_size=(4, 4), stride=(2, 2), padding=(1, 1))</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">LeakyReLU</td>
<td style="vertical-align: top; text-align: left">(negative_slope=0.01)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(128, 256, kernel_size=(4, 4), stride=(2, 2), padding=(1, 1))</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">LeakyReLU</td>
<td style="vertical-align: top; text-align: left">(negative_slope=0.01)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(256, 512, kernel_size=(4, 4), stride=(2, 2), padding=(1, 1))</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">LeakyReLU</td>
<td style="vertical-align: top; text-align: left">(negative_slope=0.01)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(512, 1024, kernel_size=(4, 4), stride=(2, 2), padding=(1, 1))</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">LeakyReLU</td>
<td style="vertical-align: top; text-align: left">(negative_slope=0.01)</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">Conv2d</td>
<td style="vertical-align: top; text-align: left">(1024, 2048, kernel_size=(4, 4), stride=(2, 2), padding=(1, 1))</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">LeakyReLU</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">(negative_slope=0.01)</td>
</tr>
</tbody><tbody>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">conv1</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">Conv2d</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">(2048, 1, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1), bias=False)</td>
</tr>
</tbody><tbody>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">conv2</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">Conv2d</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">(2048, 4, kernel_size=(2, 2), stride=(1, 1), bias=False)</td>
</tr>
</tbody>
</table>
</table-wrap>
</app></app-group>
<ref-list id="j_infor556_reflist_001">
<title>References</title>
<ref id="j_infor556_ref_001">
<mixed-citation publication-type="journal"><string-name><surname>Ali</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Biswas</surname>, <given-names>M.R.</given-names></string-name>, <string-name><surname>Mohsen</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Shah</surname> <given-names>U.</given-names></string-name>, <string-name><surname>Alamgir</surname> <given-names>A.</given-names></string-name>, <string-name><surname>Mousa</surname> <given-names>O.</given-names></string-name>, <string-name><surname>Shah</surname> <given-names>Z.</given-names></string-name> (<year>2022</year>). <article-title>The role of generative adversarial networks in brain MRI: a scoping review</article-title>. <source>Insights Imaging</source>, <volume>13</volume>, <fpage>98</fpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1186/s13244-022-01237-0" xlink:type="simple">https://doi.org/10.1186/s13244-022-01237-0</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_002">
<mixed-citation publication-type="journal"><string-name><surname>Avants</surname>, <given-names>B.B.</given-names></string-name>, <string-name><surname>Tustison</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Song</surname>, <given-names>G.</given-names></string-name>, (<year>2009</year>). <article-title>Advanced normalization tools (ANTS)</article-title>. <source>Insight Journal</source>, <volume>2</volume>(<issue>365</issue>), <fpage>1</fpage>–<lpage>35</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_003">
<mixed-citation publication-type="chapter"><string-name><surname>Bharti</surname>, <given-names>S.K.</given-names></string-name>, <string-name><surname>Inani</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Gupta</surname>, <given-names>R.K.</given-names></string-name> (<year>2022</year>). <chapter-title>Smart photo editor using generative adversarial network: a machine learning approach</chapter-title>. In: <source>2022 IEEE 2nd International Symposium on Sustainable Energy, Signal Processing and Cyber Security (ISSSC)</source>, pp. <fpage>1</fpage>–<lpage>6</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_004">
<mixed-citation publication-type="other"><string-name><surname>Chen</surname> <given-names>W.</given-names></string-name>, <string-name><surname>Wu</surname> <given-names>S.</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>S.</given-names></string-name>, <string-name><surname>Li</surname> <given-names>Z.</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>J.</given-names></string-name>, <string-name><surname>Yao</surname> <given-names>H.</given-names></string-name>, <string-name><surname>Song</surname> <given-names>X.</given-names></string-name> (2023). <italic>Compound attention and neighbor matching network for multi-contrast MRI super-resolution</italic>. arXiv preprint arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2307.02148">2307.02148</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_005">
<mixed-citation publication-type="chapter"><string-name><surname>Choi</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Choi</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Kim</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Ha</surname>, <given-names>J.-W.</given-names></string-name>, <string-name><surname>Kim</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Choo</surname>, <given-names>J.</given-names></string-name> (<year>2018</year>). <chapter-title>StarGAN: unified generative adversarial networks for multi-domain image-to-image translation</chapter-title>. In: <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>, pp. <fpage>8789</fpage>–<lpage>8797</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_006">
<mixed-citation publication-type="chapter"><string-name><surname>Choi</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Uh</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Yoo</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Ha</surname>, <given-names>J.-W.</given-names></string-name> (<year>2020</year>). <chapter-title>StarGAN v2: diverse image synthesis for multiple domains</chapter-title>. In: <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source>, pp. <fpage>8188</fpage>–<lpage>8197</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_007">
<mixed-citation publication-type="journal"><string-name><surname>Dai</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Lei</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Fu</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Curran</surname>, <given-names>W.J.</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Mao</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>X.</given-names></string-name> (<year>2020</year>). <article-title>Multimodal MRI synthesis using unified generative adversarial networks</article-title>. <source>Medical Physics</source>, <volume>47</volume>(<issue>12</issue>), <fpage>6343</fpage>–<lpage>6354</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_008">
<mixed-citation publication-type="journal"><string-name><surname>Geethanath</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Vaughan</surname> <suffix>Jr.</suffix>, <given-names>J.T.</given-names></string-name> (<year>2019</year>). <article-title>Accessible magnetic resonance imaging: a review</article-title>. <source>Journal of Magnetic Resonance Imaging</source>, <volume>49</volume>(<issue>7</issue>), <fpage>65</fpage>–<lpage>77</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_009">
<mixed-citation publication-type="journal"><string-name><surname>Grigas</surname>, <given-names>O.</given-names></string-name>, <string-name><surname>Maskeliūnas</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Damaševičius</surname>, <given-names>R.</given-names></string-name> (<year>2023</year>). <article-title>Improving structural MRI preprocessing with hybrid transformer GANs</article-title>. <source>Life</source>, <volume>13</volume>(<issue>9</issue>), <fpage>1893</fpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.3390/life13091893" xlink:type="simple">https://doi.org/10.3390/life13091893</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_010">
<mixed-citation publication-type="chapter"><string-name><surname>Hatamizadeh</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Nath</surname>, <given-names>V.</given-names></string-name>, <string-name><surname>Tang</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Roth</surname>, <given-names>H.R.</given-names></string-name>, <string-name><surname>Xu</surname>, <given-names>D.</given-names></string-name> (<year>2021</year>). <chapter-title>Swin UNETR: swin transformers for semantic segmentation of brain tumors in MRI images</chapter-title>. In: <source>International MICCAI Brainlesion Workshop</source>. <publisher-name>Springer</publisher-name>, <publisher-loc>Cham</publisher-loc>, pp. <fpage>272</fpage>–<lpage>284</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_011">
<mixed-citation publication-type="journal"><string-name><surname>Hien</surname>, <given-names>N.L.H.</given-names></string-name>, <string-name><surname>Hoang</surname>, <given-names>H.M.</given-names></string-name>, <string-name><surname>Tien</surname>, <given-names>N.V.</given-names></string-name>, <string-name><surname>Hieu</surname>, <given-names>N.V.</given-names></string-name> (<year>2021</year>). <article-title>Keyphrase extraction model: a new design and application on tourism information</article-title>. <source>Informatica</source>, <volume>45</volume>(<issue>4</issue>), <fpage>563</fpage>–<lpage>569</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_012">
<mixed-citation publication-type="journal"><string-name><surname>Hien</surname>, <given-names>N.L.H.</given-names></string-name>, <string-name><surname>Tien</surname>, <given-names>T.Q.</given-names></string-name>, <string-name><surname>Hieu</surname>, <given-names>N.V.</given-names></string-name> (<year>2020</year>). <article-title>Web crawler: design and implementation for extracting article-like contents</article-title>. <source>Cybernetics and Physics</source>, <volume>9</volume>(<issue>3</issue>), <fpage>144</fpage>–<lpage>151</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_013">
<mixed-citation publication-type="journal"><string-name><surname>Hien</surname>, <given-names>N.L.H.</given-names></string-name>, <string-name><surname>Van Huy</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Van Hieu</surname>, <given-names>N.</given-names></string-name> (<year>2021</year>). <article-title>Artwork style transfer model using deep learning approach</article-title>. <source>Cybernetics and Physics</source>, <volume>10</volume>, <fpage>127</fpage>–<lpage>137</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_014">
<mixed-citation publication-type="other"><string-name><surname>Gulrajani</surname>, <given-names>I.</given-names></string-name>, <string-name><surname>Ahmed</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Arjovsky</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Dumoulin</surname>, <given-names>V.</given-names></string-name>, <string-name><surname>Courville</surname>, <given-names>A.</given-names></string-name> (2017). <italic>Improved training of wasserstein GANs</italic>. arXiv preprint arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1704.00028">1704.00028</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_015">
<mixed-citation publication-type="journal"><string-name><surname>Katti</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Ara</surname>, <given-names>S.A.</given-names></string-name>, <string-name><surname>Shireen</surname>, <given-names>A.</given-names></string-name> (<year>2011</year>). <article-title>Magnetic resonance imaging (MRI)– a review</article-title>. <source>International Journal of Dental Clinics</source>, <volume>3</volume>(<issue>1</issue>), <fpage>65</fpage>–<lpage>70</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_016">
<mixed-citation publication-type="other"><string-name><surname>Kingma</surname>, <given-names>D.P.</given-names></string-name>, <string-name><surname>Ba</surname>, <given-names>J.L.</given-names></string-name> (2014). <italic>Adam: a method for stochastic optimization</italic>. arXiv preprint arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1412.6980">1412.6980</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_017">
<mixed-citation publication-type="chapter"><string-name><surname>Li</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Paetzold</surname>, <given-names>J.C.</given-names></string-name>, <string-name><surname>Sekuboyina</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Kofler</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Zhang</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Kirschke</surname>, <given-names>J.S.</given-names></string-name>, <string-name><surname>Wiestler</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Menze</surname>, <given-names>B.</given-names></string-name> (<year>2019</year>). <chapter-title>DiamondGAN: unified multi-modal generative adversarial networks for MRI sequences synthesis</chapter-title>. In: <source>Medical Image Computing and Computer Assisted Intervention–MICCAI 2019: 22nd International Conference, Shenzhen, China, October 13–17, 2019, Proceedings, Part IV 22</source>, pp. <fpage>795</fpage>–<lpage>803</lpage>. <comment>Springer</comment>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_018">
<mixed-citation publication-type="book"><string-name><surname>Liang</surname>, <given-names>Z.-P.</given-names></string-name>, <string-name><surname>Lauterbur</surname>, <given-names>P.C.</given-names></string-name> (<year>2000</year>). <source>Principles of Magnetic Resonance Imaging</source>. <publisher-name>SPIE Optical Engineering Press, Belllingham, WA</publisher-name>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_019">
<mixed-citation publication-type="journal"><string-name><surname>Liu</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Lei</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Shafai-Erfani</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Tian</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Patel</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Jani</surname>, <given-names>A.B.</given-names></string-name>, <string-name><surname>McDonald</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Curran</surname>, <given-names>W.J.</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Zhou</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>X.</given-names></string-name> (<year>2019</year>). <article-title>Evaluation of a deep learning-based pelvic synthetic CT generation technique for MRI-based prostate proton treatment planning</article-title>. <source>Physics in Medicine &amp; Biology</source>, <volume>64</volume>(<issue>20</issue>), <elocation-id>205022</elocation-id>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_020">
<mixed-citation publication-type="journal"><string-name><surname>Massa</surname>, <given-names>H.A.</given-names></string-name>, <string-name><surname>Johnson</surname>, <given-names>J.M.</given-names></string-name>, <string-name><surname>McMillan</surname>, <given-names>A.B.</given-names></string-name> (<year>2020</year>). <article-title>Comparison of deep learning synthesis of synthetic CTs using clinical MRI inputs</article-title>. <source>Physics in Medicine &amp; Biology</source>, <volume>65</volume>(<issue>23</issue>), <elocation-id>23N03</elocation-id>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_021">
<mixed-citation publication-type="journal"><string-name><surname>Nelson</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Menon</surname>, <given-names>R.</given-names></string-name> (<year>2022</year>). <article-title>Bijective-constrained cycle-consistent deep learning for optics-free imaging and classification</article-title>. <source>Optica</source>, <volume>9</volume>(<issue>1</issue>), <fpage>26</fpage>–<lpage>31</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_022">
<mixed-citation publication-type="journal"><string-name><surname>Shen</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Zhu</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Xing</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Pauly</surname>, <given-names>J.M.</given-names></string-name>, <string-name><surname>Turkbey</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Harmon</surname>, <given-names>S.A.</given-names></string-name>, <string-name><surname>Sanford</surname>, <given-names>T.H.</given-names></string-name>, <string-name><surname>Mehralivand</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Choyke</surname>, <given-names>P.L.</given-names></string-name>, <string-name><surname>Wood</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Xu</surname>, <given-names>D.</given-names></string-name> (<year>2020</year>). <article-title>Multi-domain image completion for random missing input data</article-title>. <source>IEEE Transactions on Medical Imaging</source>, <volume>40</volume>(<issue>4</issue>), <fpage>1113</fpage>–<lpage>1122</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_023">
<mixed-citation publication-type="chapter"><string-name><surname>Sohail</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Riaz</surname>, <given-names>M.N.</given-names></string-name>, <string-name><surname>Wu</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Long</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Li</surname>, <given-names>S.</given-names></string-name> (<year>2019</year>). <chapter-title>Unpaired multi-contrast MR image synthesis using generative adversarial networks</chapter-title>. In: <source>International Workshop on Simulation and Synthesis in Medical Imaging</source>. <publisher-name>Springer</publisher-name>, <publisher-loc>Cham</publisher-loc>, pp. <fpage>22</fpage>–<lpage>31</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_024">
<mixed-citation publication-type="chapter"><string-name><surname>Wang</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Lei</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Curran</surname>, <given-names>W.J.</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>X.</given-names></string-name> (<year>2021</year>). <chapter-title>Contrast-enhanced MRI synthesis from non-contrast MRI using attention CycleGAN</chapter-title>. In: <source>Proceedings od SPIE 11600, Medical Imaging 2021: Biomedical Applications in Molecular, Structural, and Functional Imaging</source>, pp. <fpage>388</fpage>–<lpage>393</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_025">
<mixed-citation publication-type="other"><string-name><surname>Wei</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Gong</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>Z.</given-names></string-name>, <string-name><surname>Lu</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>L.</given-names></string-name> (2018). <italic>Improving the improved training of wasserstein gans: a consistency term and its dual effect</italic>. arXiv preprint arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1803.01541">1803.01541</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_026">
<mixed-citation publication-type="chapter"><string-name><surname>Xiang</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Li</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Lin</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>Q.</given-names></string-name>, <string-name><surname>Shen</surname>, <given-names>D.</given-names></string-name> (<year>2018</year>). <chapter-title>Unpaired deep cross-modality synthesis with fast training</chapter-title>. In: <string-name><surname>Stoyanov</surname>, <given-names>D.</given-names></string-name> <etal>et al.</etal> (Eds.), <source>Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, DLMIA ML-CDS 2018</source>. <series><italic>Lecture Notes in Computer Science</italic></series>, Vol. <volume>11045</volume>. <publisher-name>Springer</publisher-name>, <publisher-loc>Cham</publisher-loc>, pp. <fpage>155</fpage>–<lpage>164</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_027">
<mixed-citation publication-type="other"><string-name><surname>Zhang</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Chi</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Lyu</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Tian</surname>, <given-names>Y.</given-names></string-name> (2023). <italic>Dual arbitrary scale super-resolution for multi-contrast MRI</italic>. arXiv preprint arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2307.02334">2307.02334</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_028">
<mixed-citation publication-type="journal"><string-name><surname>Zhang</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>He</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Guo</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Ettehadi</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Aw</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Semanek</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Posner</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Laine</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>Y.</given-names></string-name> (<year>2022</year>). <article-title>PTNet3d: a 3D high-resolution longitudinal infant brain MRI synthesizer based on transformers</article-title>. <source>IEEE Transactions on Medical Imaging</source>, <volume>41</volume>(<issue>10</issue>), <fpage>2925</fpage>–<lpage>2940</lpage>.</mixed-citation>
</ref>
<ref id="j_infor556_ref_029">
<mixed-citation publication-type="chapter"><string-name><surname>Zhu</surname>, <given-names>J.-Y.</given-names></string-name>, <string-name><surname>Park</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Isola</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Efros</surname>, <given-names>A.A.</given-names></string-name> (<year>2017</year>). <chapter-title>Unpaired image-to-image translation using cycle-consistent adversarial networks</chapter-title>. In: <source>Proceedings of the IEEE International Conference on Computer Vision</source>, pp. <fpage>2223</fpage>–<lpage>2232</lpage>.</mixed-citation>
</ref>
</ref-list>
</back>
</article>
