Simple Science

Cutting edge science explained simply

# Computer Science# Computer Vision and Pattern Recognition

Improving Medical Image Segmentation with Limited Data

A new method enhances medical image segmentation using semi-supervised learning.

― 5 min read


Advancing Medical ImageAdvancing Medical ImageAnalysis Techniquesaccuracy with limited labeled data.New approach boosts segmentation
Table of Contents

Medical Image Segmentation is a crucial step for diagnosing diseases and planning treatments. However, getting enough labeled data for this task can be very challenging, as collecting these images is costly and time-consuming. Since unlabeled data is often more available, Semi-supervised Learning (SSL) has emerged as a method that makes use of both limited labeled data and a larger number of unlabeled images.

SSL methods generally focus on using the abundant unlabeled images to improve learning after understanding some basic information from the limited labeled data. However, many existing methods often fail to effectively learn from the limited labeled data. To address this, recent advancements in using general segmentation models, like the Segment Anything Model (SAM), offer a potential solution by using less labeled data to adapt to new tasks.

Overview of Segment Anything Model (SAM)

The Segment Anything Model is a powerful tool that has shown great adaptability in various tasks. It learns from large, natural datasets, making it capable of transferring its knowledge to new tasks with only a small amount of additional labeled data. However, many approaches using SAM for medical image segmentation focus primarily on training it fully with plenty of labeled data. This is not the best use of its capabilities, especially when the amount of labeled data is limited.

Some recent efforts have attempted to use SAM in a semi-supervised context, but they often treat SAM as a separate element to generate pseudo labels for medical images. This can lead to performance issues due to the differences between natural and medical images.

Proposed Method

To improve how SAM can be used for medical image segmentation with limited labeled data, we propose a new method we call Cross Prompting Consistency with Segment Anything Model (CPC-SAM). Our approach leverages SAM's unique prompt design and creates a system where two branches interact to guide learning. By doing so, it efficiently learns from both the limited labeled data and the vast amount of unlabeled data.

Cross Prompting Strategy

The heart of our method lies in the cross prompting strategy, which involves two branches of the model. Each branch generates prompts for the other based on their outputs. The basic idea is that when one branch provides an output without prompts, the second branch can use that output to create prompts that will guide the first branch. This bidirectional prompting allows for more effective learning.

However, since unlabeled data can lead to noisy prompts, we need a way to ensure the outputs are still reliable. To address this, we incorporated a prompt consistency regularization strategy. This means we want to make sure that the outputs from SAM remain consistent even when we use different prompts, minimizing the sensitivity of SAM to where prompts are positioned.

Experimental Evaluation

To validate our method, we tested CPC-SAM on two public medical image datasets: one for breast cancer and another for cardiac structures. We found that our approach outperformed current state-of-the-art semi-supervised methods, achieving significant improvements in segmentation accuracy.

Dataset Details

  1. Breast Cancer Dataset: This dataset consists of ultrasound images used to identify benign and malignant tumors.
  2. Cardiac Structures Dataset: This dataset includes MRI scans of the heart, focusing on different regions that need to be accurately segmented.

In our tests, we only used a small number of labeled images and relied heavily on the unlabeled ones to demonstrate how well our method can work even with little initial information.

Results

The results showed that our method achieved over 9% improvement in Dice Coefficient on the breast cancer segmentation task while also demonstrating superior performance on the cardiac dataset. This confirms that CPC-SAM can effectively utilize unlabeled data to enhance the learning process.

Advantages of the Proposed Method

The key strengths of our approach can be summarized as follows:

  1. Effective Use of Limited Labeled Data: By leveraging the capabilities of SAM, we can achieve good learning outcomes with only a small initial labeled dataset.

  2. Bidirectional Learning: The cross prompting strategy allows the model to continually refine its understanding by using outputs from both branches to inform one another.

  3. Stability through Regularization: The prompt consistency regularization helps ensure that outputs remain reliable, addressing the potential noise introduced by the unlabeled data prompts.

  4. Robust Performance: Our method shows consistent improvements over various existing methods, especially when labeled data is scarce.

Conclusion and Future Work

In conclusion, our proposed CPC-SAM framework effectively bridges the gap between using limited labeled data and abundant unlabeled data in medical image segmentation tasks. It demonstrates that with the right strategies, even small labeled datasets can lead to robust models that perform well on complex tasks.

Looking ahead, we plan to explore additional strategies to select prompts that will lead to even more reliable outputs. We also want to investigate the use of other types of prompts or different modalities to further improve segmentation outcomes. Our ongoing research will continue to focus on making semi-supervised learning more efficient and effective in the medical imaging field.

Original Source

Title: Cross Prompting Consistency with Segment Anything Model for Semi-supervised Medical Image Segmentation

Abstract: Semi-supervised learning (SSL) has achieved notable progress in medical image segmentation. To achieve effective SSL, a model needs to be able to efficiently learn from limited labeled data and effectively exploiting knowledge from abundant unlabeled data. Recent developments in visual foundation models, such as the Segment Anything Model (SAM), have demonstrated remarkable adaptability with improved sample efficiency. To harness the power of foundation models for application in SSL, we propose a cross prompting consistency method with segment anything model (CPC-SAM) for semi-supervised medical image segmentation. Our method employs SAM's unique prompt design and innovates a cross-prompting strategy within a dual-branch framework to automatically generate prompts and supervisions across two decoder branches, enabling effectively learning from both scarce labeled and valuable unlabeled data. We further design a novel prompt consistency regularization, to reduce the prompt position sensitivity and to enhance the output invariance under different prompts. We validate our method on two medical image segmentation tasks. The extensive experiments with different labeled-data ratios and modalities demonstrate the superiority of our proposed method over the state-of-the-art SSL methods, with more than 9% Dice improvement on the breast cancer segmentation task.

Authors: Juzheng Miao, Cheng Chen, Keli Zhang, Jie Chuai, Quanzheng Li, Pheng-Ann Heng

Last Update: 2024-07-07 00:00:00

Language: English

Source URL: https://arxiv.org/abs/2407.05416

Source PDF: https://arxiv.org/pdf/2407.05416

Licence: https://creativecommons.org/licenses/by/4.0/

Changes: This summary was created with assistance from AI and may have inaccuracies. For accurate information, please refer to the original source documents linked here.

Thank you to arxiv for use of its open access interoperability.

More from authors

Similar Articles