Enhancing Robust Liver Cancer Diagnosis: A Contrastive Multi-Modality Learner with Lightweight Fusion and Effective Data Augmentation

Pei Xuan Li, Hsun Ping Hsieh, Yang Fan-Chiang, Ding You Wu, Ching Chung Ko

Research output: Contribution to journalArticlepeer-review

Abstract

This article explores the application of self-supervised contrastive learning in the medical domain, focusing on classification of multi-modality Magnetic Resonance (MR) images. To address the challenges of limited and hard-to-annotate medical data, we introduce multi-modality data augmentation (MDA) and cross-modality group convolution (CGC). In the pre-training phase, we leverage Simple Siamese networks to maximize the similarity between two augmented MR images from a patient, without a handcrafted pretext task. Our approach also combines 3D and 2D group convolution with a channel shuffle operation to efficiently incorporate different modalities of image features. Evaluation on liver MR images from a well-known hospital in Taiwan demonstrates a significant improvement over previous methods. This work contributes to advancing multi-modality contrastive learning, particularly in the context of medical imaging, offering enhanced tools for analyzing complex image data.

Original languageEnglish
JournalACM Transactions on Computing for Healthcare
Volume5
Issue number2
DOIs
Publication statusPublished - 2024 Apr 22

All Science Journal Classification (ASJC) codes

  • Software
  • Medicine (miscellaneous)
  • Information Systems
  • Biomedical Engineering
  • Computer Science Applications
  • Health Informatics
  • Health Information Management

Fingerprint

Dive into the research topics of 'Enhancing Robust Liver Cancer Diagnosis: A Contrastive Multi-Modality Learner with Lightweight Fusion and Effective Data Augmentation'. Together they form a unique fingerprint.

Cite this