Multi-modality MRI fusion with patch complementary pre-training for internet of medical things-based smart healthcare (2024)

research-article

Authors: Jun Lyu, Xiudong Chen, Salman A. AlQahtani, and M. Shamim Hossain

Published: 02 July 2024 Publication History

  • 0citation
  • 0
  • Downloads

Metrics

Total Citations0Total Downloads0

Last 12 Months0

Last 6 weeks0

  • Get Citation Alerts

    New Citation Alert added!

    This alert has been successfully added and will be sent to:

    You will be notified whenever a record that you have chosen has been cited.

    To manage your alert preferences, click on the button below.

    Manage my Alerts

    New Citation Alert!

    Please log in to your account

      • View Options
      • References
      • Media
      • Tables
      • Share

    Abstract

    Magnetic Resonance Imaging (MRI) is a pivotal neuroimaging technique capable of generating images with various contrasts, known as multi-modal images. The integration of these diverse modalities is essential for improving model performance across various tasks. However, in real clinical scenarios, acquiring MR images for all modalities is frequently hindered by factors such as patient comfort and scanning costs. Therefore, effectively fusing different modalities to synthesize missing modalities has become a research hot-spot in the field of smart healthcare, particularly in the context of the Internet of Medical Things (IoMT). In this study, we introduce a multi-modal coordinated fusion network (MCF-Net) with Patch Complementarity Pre-training. This network leverages the complementarity and correlation between different modalities to make the fusion of multi-modal MR images, addressing challenges in the IoMT. Specifically, we first employ a Patch Complementarity Mask Autoencoder (PC-MAE) for self-supervised pre-training. The complementarity learning mechanism is introduced to align masks and visual annotations between two modalities. Subsequently, a dual-branch MAE architecture and a shared encoder–decoder are adopted to facilitate cross-modal interactions within mask tokens. Furthermore, during the fine-tuning phase, we incorporate an Attention-Driven Fusion (ADF) module into the MCF-Net. This module synthesizes missing modal images by fusion of multi-modal features from the pre-trained PC-MAE encoder. Additionally, we leverage the pre-trained encoder to extract high-level features from both synthetic and corresponding real images, ensuring consistency throughout the training process. Our experimental findings showcase a notable enhancement in performance across various modalities with our fusion method, outperforming state-of-the-art techniques.

    Highlights

    Propose a multi-modal medical image fusion framework with patch complementary pre-training

    Design a novel masking alignment strategy to learn complementary information between modalities.

    Introduce an attention-driven fusion module to aggregate multi-modal features.

    References

    [1]

    Zhou Tao, Liu Mingxia, Fu Huazhu, Wang Jun, Shen Jianbing, Shao Ling, Shen Dinggang, Deep multi-modal latent representation learning for automated dementia diagnosis, in: International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, 2019, pp. 629–638.

    [2]

    Muhammad Ghulam, Alshehri Fatima, Karray Fakhri, El Saddik Abdulmotaleb, Alsulaiman Mansour, Falk Tiago H, A comprehensive survey on multimodal medical signals fusion for smart healthcare systems, Inf. Fusion 76 (2021) 355–375.

    Digital Library

    [3]

    Zhang Pengfei, Li Tianrui, Wang Guoqiang, Wang Dexian, Lai Pei, Zhang Fan, A multi-source information fusion model for outlier detection, Inf. Fusion 93 (2023) 192–208.

    [4]

    Zhang Hao, Xu Han, Tian Xin, Jiang Junjun, Ma Jiayi, Image fusion meets deep learning: A survey and perspective, Inf. Fusion 76 (2021) 323–336.

    Digital Library

    [5]

    Sevetlidis Vasileios, Giuffrida Mario Valerio, Tsaftaris Sotirios A, Whole image synthesis using a deep encoder-decoder network, in: Simulation and Synthesis in Medical Imaging: First International Workshop, SASHIMI 2016, Held in Conjunction with MICCAI 2016, Athens, Greece, October 21, 2016, Proceedings 1, Springer, 2016, pp. 127–137.

    [6]

    Bavirisetti Durga Prasad, Dhuli Ravindra, Two-scale image fusion of visible and infrared images using saliency detection, Infrared Phys. Technol. 76 (2016) 52–64.

    [7]

    Dar Salman UH, Yurt Mahmut, Karacan Levent, Erdem Aykut, Erdem Erkut, Cukur Tolga, Image synthesis in multi-contrast MRI with conditional generative adversarial networks, IEEE Trans. Med. Imaging 38 (10) (2019) 2375–2388.

    [9]

    Wen Wei, Emilie Poirion, Benedetta Bodini, Stanley Durrleman, Olivier Colliot, Bruno Stankoff, Nicholas Ayache, FLAIR MR image synthesis by using 3D fully convolutional networks for multiple sclerosis, in: ISMRM-ESMRMB 2018-Joint Annual Meeting, 2018, pp. 1–6.

    [10]

    Wang Guanhua, Gong Enhao, Banerjee Suchandrima, Martin Dann, Tong Elizabeth, Choi Jay, Chen Huijun, Wintermark Max, Pauly John M, Zaharchuk Greg, Synthesize high-quality multi-contrast magnetic resonance imaging from multi-echo acquisition using multi-task deep generative model, IEEE Trans. Med. Imaging 39 (10) (2020) 3089–3099.

    [11]

    Jun-Yan Zhu, Taesung Park, Phillip Isola, Alexei A Efros, Unpaired image-to-image translation using cycle-consistent adversarial networks, in: Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 2223–2232.

    [12]

    Yang Xin, Lin Yi, Wang Zhiwei, Li Xin, Cheng Kwang-Ting, Bi-modality medical image synthesis using semi-supervised sequential generative adversarial networks, IEEE J. Biomed. Health Inform. 24 (3) (2019) 855–865.

    [13]

    Olut Sahin, Sahin Yusuf H., Demir Ugur, Unal Gozde, Generative adversarial training for MRA image synthesis using multi-contrast MRI, in: PRedictive Intelligence in MEdicine: First International Workshop, PRIME 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, September 16, 2018, Proceedings 1, Springer, 2018, pp. 147–154.

    [14]

    Zhou Tao, Fu Huazhu, Chen Geng, Shen Jianbing, Shao Ling, Hi-net: Hybrid-fusion network for multi-modal MR image synthesis, IEEE Trans. Med. Imaging 39 (9) (2020) 2772–2781.

    Digital Library

    [15]

    Hossain M.S., et al., Smart healthcare monitoring: A voice pathology detection paradigm for smart cities, Multimedia Syst. 25 (5) (2019) 565–575.

    [16]

    Umirzakova Sabina, Ahmad Shabir, Khan Latif U, Whangbo Taegkeun, Medical image super-resolution for smart healthcare applications: A comprehensive survey, Inf. Fusion (2023).

    [17]

    Tao Xiaohui, Velasquez Juan D., Multi-source information fusion for smart health with artificial intelligence, Inf. Fusion 83 (2022) 93–95.

    [18]

    Goodfellow Ian, Pouget-Abadie Jean, Mirza Mehdi, Xu Bing, Warde-Farley David, Ozair Sherjil, Courville Aaron, Bengio Yoshua, Generative adversarial nets, Advances in Neural Information Processing Systems, vol. 27, 2014.

    [19]

    Dar Salman UH, Yurt Mahmut, Karacan Levent, Erdem Aykut, Erdem Erkut, Cukur Tolga, Image synthesis in multi-contrast MRI with conditional generative adversarial networks, IEEE Trans. Med. Imaging 38 (10) (2019) 2375–2388.

    [20]

    Wolterink Jelmer M, Dinkla Anna M, Savenije Mark HF, Seevinck Peter R, van den Berg Cornelis AT, Išgum Ivana, Deep MR to CT synthesis using unpaired data, in: Simulation and Synthesis in Medical Imaging: Second International Workshop, SASHIMI 2017, Held in Conjunction with MICCAI 2017, QuÉBec City, QC, Canada, September 10, 2017, Proceedings 2, Springer, 2017, pp. 14–23.

    [21]

    Jun-Yan Zhu, Taesung Park, Phillip Isola, Alexei A Efros, Unpaired image-to-image translation using cycle-consistent adversarial networks, in: Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 2223–2232.

    [22]

    Wen Wei, Emilie Poirion, Benedetta Bodini, Stanley Durrleman, Olivier Colliot, Bruno Stankoff, Nicholas Ayache, FLAIR MR image synthesis by using 3D fully convolutional networks for multiple sclerosis, in: ISMRM-ESMRMB 2018-Joint Annual Meeting, 2018, pp. 1–6.

    [23]

    Huang Yawen, Shao Ling, Frangi Alejandro F., DOTE: Dual convolutional filter learning for super-resolution and cross-modality synthesis in MRI, in: Medical Image Computing and Computer Assisted Intervention- MICCAI 2017: 20th International Conference, Quebec City, QC, Canada, September 11-13, 2017, Proceedings, Part III 20, Springer, 2017, pp. 89–98.

    [24]

    Shen Liyue, Zhu Wentao, Wang Xiaosong, Xing Lei, Pauly John M, Turkbey Baris, Harmon Stephanie Anne, Sanford Thomas Hogue, Mehralivand Sherif, Choyke Peter L, et al., Multi-domain image completion for random missing input data, IEEE Trans. Med. Imaging 40 (4) (2020) 1113–1122.

    [25]

    Wang Yan, Zhou Luping, Yu Biting, Wang Lei, Zu Chen, Lalush David S, Lin Weili, Wu Xi, Zhou Jiliu, Shen Dinggang, 3D auto-context-based locality adaptive multi-modality GANs for PET synthesis, IEEE Trans. Med. Imaging 38 (6) (2018) 1328–1339.

    [26]

    Vaswani Ashish, Shazeer Noam, Parmar Niki, Uszkoreit Jakob, Jones Llion, Gomez Aidan N, Kaiser Łukasz, Polosukhin Illia, Attention is all you need, Advances in Neural Information Processing Systems, vol. 30, 2017.

    [27]

    Chen Jieneng, Lu Yongyi, Yu Qihang, Luo Xiangde, Adeli Ehsan, Wang Yan, Lu Le, Yuille Alan L, Zhou Yuyin, Transunet: Transformers make strong encoders for medical image segmentation, 2021, arXiv preprint arXiv:2102.04306.

    [28]

    Chen Junyu, He Yufan, Frey Eric C., Li Ye, Du Yong, Vit-v-net: Vision transformer for unsupervised volumetric medical image registration, 2021, arXiv preprint arXiv:2104.06468.

    [29]

    Dalmaz Onat, Yurt Mahmut, Çukur Tolga, ResViT: Residual vision transformers for multimodal medical image synthesis, IEEE Trans. Med. Imaging 41 (10) (2022) 2598–2614.

    [30]

    Peng Bo, Liu Bingzheng, Bin Yi, Shen Lili, Lei Jianjun, Multi-modality mr image synthesis via confidence-guided aggregation and cross-modality refinement, IEEE J. Biomed. Health Inf. 26 (1) (2021) 27–35.

    [31]

    Yan Shouang, Wang Chengyan, Chen Weibo, Lyu Jun, Swin transformer-based GAN for multi-modal medical image translation, Front. Oncol. 12 (2022).

    [32]

    Liu Jiang, Pasumarthi Srivathsa, Duffy Ben, Gong Enhao, Datta Keshav, Zaharchuk Greg, One model to synthesize them all: Multi-contrast multi-scale transformer for missing data imputation, IEEE Trans. Med. Imaging (2023).

    [33]

    Li Yonghao, Zhou Tao, He Kelei, Zhou Yi, Shen Dinggang, Multi-scale transformer network with edge-aware pre-training for cross-modality MR image synthesis, IEEE Trans. Med. Imaging (2023).

    [34]

    Li Yan, Xu Sisi, Chen Haibin, Sun Ying, Bian Jing, Guo Shuanshuan, Lu Yao, Qi Zhenyu, CT synthesis from multi-sequence MRI using adaptive fusion network, Comput. Biol. Med. 157 (2023).

    [35]

    Li Yan, Xu Sisi, Lu Yao, Qi Zhenyu, CT synthesis from MRI with an improved multi-scale learning network, Front. Phys. 11 (2023).

    [36]

    Hardoon David R., Szedmak Sandor, Shawe-Taylor John, Canonical correlation analysis: An overview with application to learning methods, Neural Comput. 16 (12) (2004) 2639–2664.

    Digital Library

    [37]

    Lin Yen-Yu, Liu Tyng-Luh, Fuh Chiou-Shann, Multiple kernel learning for dimensionality reduction, IEEE Trans. Pattern Anal. Mach. Intell. 33 (6) (2010) 1147–1160.

    Digital Library

    [38]

    Zhou Tao, Thung Kim-Han, Zhu Xiaofeng, Shen Dinggang, Effective feature learning and fusion of multimodality data using stage-wise deep neural network for dementia diagnosis, Hum. Brain Mapp. 40 (3) (2019) 1001–1016.

    [39]

    Hou Ming, Tang Jiajia, Zhang Jianhai, Kong Wanzeng, Zhao Qibin, Deep multimodal multilinear fusion with high-order polynomial pooling, Adv. Neural Inf. Process. Syst. 32 (2019).

    [40]

    Yu Biting, Zhou Luping, Wang Lei, Fripp Jurgen, Bourgeat Pierrick, 3D cGAN based cross-modality MR image synthesis for brain tumor segmentation, in: 2018 IEEE 15th International Symposium on Biomedical Imaging, ISBI 2018, IEEE, 2018, pp. 626–630.

    [41]

    Wei Wen, Poirion Emilie, Bodini Benedetta, Durrleman Stanley, Colliot Olivier, Stankoff Bruno, Ayache Nicholas, Fluid-attenuated inversion recovery MRI synthesis from multisequence MRI using three-dimensional fully convolutional networks for multiple sclerosis, J. Med. Imaging 6 (1) (2019) 014005.

    [42]

    Menze Bjoern H, Jakab Andras, Bauer Stefan, Kalpathy-Cramer Jayashree, Farahani Keyvan, Kirby Justin, Burren Yuliya, Porz Nicole, Slotboom Johannes, Wiest Roland, et al., The multimodal brain tumor image segmentation benchmark (BRATS), IEEE Trans. Med. Imaging 34 (10) (2014) 1993–2024.

    [43]

    Zhang Xuzhe, He Xinzi, Guo Jia, Ettehadi Nabil, Aw Natalie, Semanek David, Posner Jonathan, Laine Andrew, Wang Yun, PTNet: A high-resolution infant MRI synthesizer based on transformer, 2021, arXiv preprint arXiv:2105.13993.

    [44]

    Yu Biting, Zhou Luping, Wang Lei, Shi Yinghuan, Fripp Jurgen, Bourgeat Pierrick, Ea-GANs: Edge-aware generative adversarial networks for cross-modality MR image synthesis, IEEE Trans. Med. Imaging 38 (7) (2019) 1750–1762.

    Recommendations

    • A comprehensive survey on multimodal medical signals fusion for smart healthcare systems

      Highlights

      • A comprehensive survey on IoMT-based fusion for healthcare is presented.
      • ...

      Abstract

      Smart healthcare is a framework that utilizes technologies such as wearable devices, the Internet of Medical Things (IoMT), sophisticated machine learning algorithms, and wireless communication technology to seamlessly access health ...

      Read More

    • Multi-modal medical image fusion in NSST domain for internet of medical things

      Abstract

      The Internet of Medical Things (IoMT) has included a new layer for development and smart infrastructure growth in the medical field. Besides, the medical data on IoMT systems are constantly expanding due to the rising peripherals in the health ...

      Read More

    • A survey of multimodal information fusion for smart healthcare: Mapping the journey from data to wisdom

      Abstract

      Multimodal medical data fusion has emerged as a transformative approach in smart healthcare, enabling a comprehensive understanding of patient health and personalized treatment plans. In this paper, a journey from data to information to knowledge ...

      Read More

    Comments

    Information & Contributors

    Information

    Published In

    Multi-modality MRI fusion with patch complementary pre-training for internet of medical things-based smart healthcare (1)

    Information Fusion Volume 107, Issue C

    Jul 2024

    522 pages

    ISSN:1566-2535

    Issue’s Table of Contents

    Elsevier B.V.

    Publisher

    Elsevier Science Publishers B. V.

    Netherlands

    Publication History

    Published: 02 July 2024

    Author Tags

    1. Multi-modal medical image fusion
    2. Smart healthcare
    3. Internet of medical things (IoMT)
    4. Magnetic resonance imaging (MRI)

    Qualifiers

    • Research-article

    Contributors

    Multi-modality MRI fusion with patch complementary pre-training for internet of medical things-based smart healthcare (2)

    Other Metrics

    View Article Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Total Citations

    • Total Downloads

    • Downloads (Last 12 months)0
    • Downloads (Last 6 weeks)0

    Other Metrics

    View Author Metrics

    Citations

    View Options

    View options

    Get Access

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    Get this Publication

    Media

    Figures

    Other

    Tables

    Multi-modality MRI fusion with patch complementary pre-training for internet of medical things-based smart healthcare (2024)
    Top Articles
    Latest Posts
    Article information

    Author: Terence Hammes MD

    Last Updated:

    Views: 6488

    Rating: 4.9 / 5 (69 voted)

    Reviews: 92% of readers found this page helpful

    Author information

    Name: Terence Hammes MD

    Birthday: 1992-04-11

    Address: Suite 408 9446 Mercy Mews, West Roxie, CT 04904

    Phone: +50312511349175

    Job: Product Consulting Liaison

    Hobby: Jogging, Motor sports, Nordic skating, Jigsaw puzzles, Bird watching, Nordic skating, Sculpting

    Introduction: My name is Terence Hammes MD, I am a inexpensive, energetic, jolly, faithful, cheerful, proud, rich person who loves writing and wants to share my knowledge and understanding with you.