• Open Access
    Original Article

    MUCM-Net: a Mamba powered UCM-Net for skin lesion segmentation

    Chunyu Yuan 1*
    Dongfang Zhao 2
    Sos S. Agaian 1,3

    Explor Med. 2024;5:694–708 DOI: https://doi.org/10.37349/emed.2024.00250

    Received: June 10, 2024 Accepted: August 27, 2024 Published: October 25, 2024

    Academic Editor: Lindsay A. Farrer, Boston University School of Medicine, USA

    Abstract

    Aim:

    Skin lesion segmentation is critical for early skin cancer detection. Challenges in automatic segmentation from dermoscopic images include variations in color, texture, and artifacts of indistinct lesion boundaries. This study aims to develop and evaluate MUCM-Net, a lightweight and efficient model for skin lesion segmentation, leveraging Mamba state-space models integrated with UCM-Net architecture optimized for mobile deployment and early skin cancer detection.

    Methods:

    MUCM-Net combines Convolutional Neural Networks (CNNs), multi-layer perceptions (MLPs), and Mamba elements into a hybrid feature learning module.

    Results:

    The model was trained and tested on the International Skin Imaging Collaboration (ISIC) 2017 and ISIC2018 datasets, consisting of 2,000 and 2,594 dermoscopic images, respectively. Critical metrics for evaluation included Dice Similarity Coefficient (DSC), sensitivity (SE), specificity (SP), and accuracy (ACC). The model’s computational efficiency was also assessed by measuring Giga Floating-point Operations Per Second (GFLOPS) and the number of parameters. MUCM-Net demonstrated superior performance in skin lesion segmentation with an average DSC of 0.91 on the ISIC2017 dataset and 0.89 on the ISIC2018 dataset, outperforming existing models. It achieved high SE (0.93), SP (0.95), and ACC (0.92) with low computational demands (0.055–0.064 GFLOPS).

    Conclusions:

    The model’s innovative Mamba-UCM layer significantly enhanced feature learning while maintaining efficiency that is suitable for mobile devices. MUCM-Net establishes a new standard in lightweight skin lesion segmentation, balancing exceptional ACC with efficient computational performance. Its ability to perform well on mobile devices makes it a scalable tool for early skin cancer detection in resource-limited settings. The open-source availability of MUCM-Net supports further research and collaboration, promoting advances in mobile health diagnostics and the fight against skin cancer. MUCM-Net source code will be posted on https://github.com/chunyuyuan/MUCM-Net.

    Keywords

    Medical image segmentation, lightweight model, mobile health, machine learning, skin cancer

    Introduction

    Skin cancer is primarily classified into two major types: melanoma and non-melanoma. Despite melanoma accounting for only 1% of skin cancer cases, it is responsible for the majority of skin cancer-related fatalities due to its aggressive behavior. In 2022, melanoma led to approximately 7,800 deaths in the United States, with an estimated 98,000 new cases anticipated in 2023 [1]. The lifetime risk for Americans developing skin cancer is alarmingly high, with current statistics showing that one in five individuals will be affected. This underscores the critical need for effective diagnostic and treatment approaches. The financial impact is also considerable, with skin cancer treatment in the U.S. exceeding $8.1 billion [2]. Malignant melanoma, in particular, is notorious for its rapid progression and high mortality rate, making early and accurate detection essential for improving patient outcomes. Dermatoscopy and dermoscopy play a vital role in the clinical evaluation of skin lesions, helping dermatologists to identify malignant characteristics [35]. However, manual interpretation can be time-consuming and error-prone, dependent on the clinician’s expertise. Recent advancements have introduced machine learning-driven techniques into clinical practice to improve diagnosis accuracy and efficiency. These techniques are particularly beneficial in computationally constrained environments like mobile health applications [6, 7].

    Manual interpretation can be time-consuming, error-prone, and heavily dependent on the clinician’s expertise. Furthermore, certain medical samples present considerable challenges [8]. These challenges include indistinct boundaries where lesions merge seamlessly with the surrounding skin, making it difficult to delineate them clearly. Variations in lighting can alter the appearance of lesions, affecting the consistency of their visual features. Obstructions such as hair and bubbles may obscure lesion edges, complicating accurate segmentation. Additionally, variations in lesion size and shape, along with inconsistencies in imaging conditions and resolutions, can lead to inaccuracies in the segmentation process. Age-related skin changes can impact texture and appearance, further complicating detection. Complex backgrounds and differences in skin tone, influenced by factors such as race and environmental conditions, also contribute to the difficulty of achieving precise segmentation. Addressing these issues requires advanced techniques that can adapt to these varying conditions and ensure accurate diagnosis. Figure 1 presents some representative complex skin lesion samples.

    Complex skin lesion samples. (A) Artifacts arising from the image acquisition process; (B) lesions obstructed by hair; (C) subtle differences between lesion and skin color; (D) low contrast between the wound and surrounding skin

    Note. Source from International Skin Imaging Collaboration (ISIC) datasets [1821]

    Recent advancements in artificial intelligence (AI) have increasingly integrated computer-aided tools into clinical practice to enhance skin cancer diagnosis [9, 10]. A crucial technique in this process is skin cancer segmentation, which accurately delineates the boundaries of skin lesions in medical images. This segmentation is vital for accurately assessing lesion characteristics, monitoring their progression, and guiding treatment decisions. With rapid advancements in AI techniques and the widespread adoption of smart devices, such as point-of-care ultrasound (POCUS) devices or smartphones [1113], AI-driven approaches for skin cancer detection have become popular.

    Patients now enjoy enhanced access to medical information, remote monitoring, and tailored care, which has improved their overall satisfaction with healthcare services. Despite these positive changes, certain obstacles remain, particularly in medical diagnostics. A notable issue is the precise and efficient segmentation of skin lesions, which is critical for diagnosis but challenging to implement on devices with limited computational resources. Most AI-driven medical applications rely on deep learning techniques described in detail by [14]. These methods typically require significant computational power and extensive learning parameters to deliver accurate predictions, posing a challenge for integration into devices with constrained hardware capabilities [15, 16].

    In this study, we extend our previous method UCM-Net [17], and introduce MUCM-Net, a lightweight, robust, and Mamba-powered approach for skin lesion segmentation. MUCM-Net leverages a new novel hybrid module that combines Convolutional Neural Networks (CNNs), multi-layer perceptions (MLPs), and Mamba to enhance feature learning. Utilizing new proposed group loss functions, our method surpasses existing Mamba-based techniques in skin lesion segmentation. Figure 2 presents the visualization of comparative experimental results on the International Skin Imaging Collaboration (ISIC) 2017 dataset [18, 19]. More results will be provided in the following “Results and Discussion” sessions.

    Comparative results on the International Skin Imaging Collaboration (ISIC) 2017 dataset. The X-axis shows the Dice Similarity Coefficient (DSC, higher values indicate better performance), and the Y-axis represents Giga Floating-point Operations Per Second (GFLOPS, lower values indicate greater efficiency). LightM-UNet’s result is out of range in the zoom-in window due to GFLOPS

    Key contributions of MUCM-Net include:

    • Hybrid feature learning: The MUCM-Net block integrates CNN, MLP, and Mamba elements, enhancing the learning of complex and distinct lesion features. We explore the application of the Mamba structure to increase the feature learning ability.

    • Computational efficiency: MUCM-Net’s design, based on Mamba-UCM blocks and UCM-Net, prioritizes accuracy and efficiency. It achieves high prediction performance with low computational demands [approx. 0.055–0.064 Giga Floating-point Operations Per Second (GFLOPS)], making it suitable for various deployment scenarios.

    • Enhanced loss function: A novel loss function integrates output and internal stage losses, ensuring efficient learning during the model’s training process.

    • Superior results: MUCM-Net achieves exceptional results on the ISIC2017 [18, 19] and ISIC2018 datasets [20, 21], outperforming previous Mamba-based methods on metrics like Dice similarity, sensitivity, specificity, and accuracy.

    Related works

    Tiny machine learning (TinyML) for healthcare: Biomedical imaging segmentation, which involves accurately delineating anatomical structures and pathological regions from medical images, is essential for precise diagnostics. Recent advances in AI have significantly improved segmentation methods, enhancing both their accuracy and efficiency. A burgeoning area of research in this field is the integration of TinyML into healthcare, particularly for applications like lesion segmentation, offering exciting possibilities for both research and practical use.

    TinyML refers to deploying machine learning models on low-power, compact hardware. This technology holds the potential to revolutionize healthcare by enabling advanced analytical capabilities directly at the point of care. It facilitates real-time, on-device processing, making sophisticated medical image analysis feasible even in settings with limited traditional computing resources or in mobile healthcare environments. For instance, using TinyML for lesion segmentation could provide immediate diagnostic feedback during patient consultations or in remote locations, significantly reducing dependence on extensive infrastructure typically needed for detailed analyses.

    The incorporation of TinyML into medical devices is expected to enhance diagnostic accuracy, improve patient outcomes, and broaden access to advanced medical technologies in underserved regions. To optimize the deployment of TinyML in these critical applications, researchers are exploring advanced methods such as hyper-structure optimization [22] and leveraging efficient techniques like binary neural networks [23].

    Hyper-structure optimization focuses on reducing the model’s parameter count without sacrificing performance, ensuring the models remain both practical and lightweight for use on miniature devices. Moreover, implementing binary neural networks helps streamline computations, further enhancing the practicality of TinyML applications in resource-constrained settings. As we delve into optimizing and applying models like MUCM-Net in healthcare. This research not only underscores the transformative possibilities of TinyML but also guides future explorations in deploying compact, efficient AI solutions in medical settings.

    Supervised methods of segmentation: As AI technology continues to advance, the approaches for medical image segmentation have evolved significantly. Initially, the field heavily relied on CNNs such as U-Net, which is essential in medical image segmentation [24], and its attention-enhanced variant, Att-UNet [25], which incorporates attention mechanisms to further refine the segmentation accuracy by focusing on relevant features within the images. The development of hybrid architectures marks a further evolution in segmentation techniques. There are some hybrid-based UNets for medical image segmentation: (1) Transformers-related: such as TransUNet [26], TransFuse [27], and SANet [28]; (2) MLP-related: such as ConvNeXts [29], UNeXt [30], MALUNet [31] and its extended version EGE-UNet [32]. Recently, as Vision Mamba [33]’s image processing ability with fewer parameters and lower computations, Mamba-based hybrid structure UNets are becoming popular such as VM-UNet [34], VM-UNet V2 [35], LightM-UNet [36], and UltraLight VM-UNet [37].

    State-space models (SSMs): SSMs have recently been recognized for their linear complexity concerning input size and memory usage, establishing them as fundamental components for lightweight model architectures [38]. SSMs are particularly effective at capturing long-range dependencies, offering a critical solution to the convolution challenge of processing information across extensive distances. With the advantage of SSMs, Mamba [39] has been proven to handle textual data with fewer parameters than Transformers. Similarly, the advent of Vision Mamba has advanced the application of SSMs in image processing, demonstrating a significant memory reduction, all without relying on traditional attention mechanisms. This pioneering research bolsters confidence in Mamba’s potential as a critical lightweight model component in future technological advancements.

    In this paper, we extend our previous work UCM-Net to propose a new hybrid work MUCM-Net which engages the Mamba’s features learning ability and maintains fewer parameters and lower computations.

    Materials and methods

    MUCM-Net

    Network structure design: Figure 3 presents a detailed overview of the structural framework of MUCM-Net, a cutting-edge architecture featuring a unique U-shape design, building upon UCM-Net. MUCM-Net consists of a down-sampling encoder and an up-sampling decoder, forming a robust network for skin lesion segmentation. The network is composed of six encoder-decoder stages, with channel capacities set at {8, 16, 24, 32, 48, 64}. The first stage in the encoder-decoder sequence is a convolutional block, responsible for extracting and capturing critical features. The subsequent stages incorporate our novel UCM-Net blocks, enhancing the network’s performance.

    MUCM-Net pipeline. Conv2D: 2-dimensional convolution

    Convolution block: In our design, the first encoder-decoder stage employs a standard convolution layer with a 3 × 3 filter. This convolution block utilizes a 3 × 3 kernel, which is widely used to capture spatial relationships within input features. This kernel size is especially beneficial in the initial layers of the network, where maintaining the spatial integrity of feature maps is crucial for decoding complex input patterns. From the 2nd to the 6th stage, we incorporate a 1 × 1 filter convolution layer, which supports the subsequent Mamba-UCM block. This adjustment significantly reduces the number of learnable parameters and the overall computational load, optimizing the network’s efficiency.

    Mamba-UCM block: The 2nd–6th stages mainly use the Mamba-UCM block for feature learning. The Mamba-UCM block showcases an advanced strategy that merges UCM-Net block, which contains CNNs with MLPs and Mamba block, to enhance feature learning. There are two advantages in the hybrid structure: (1) CNNs module can enable the model to extract the spatial feature; (2) MLPs and SSM can power the model’s pattern recognition capabilities. The process begins by reshaping the initial input feature map to meet the distinct requirements of CNNs, MLPs, and Mamba. This adaptation involves converting a four-dimensional tensor suitable for CNN processing into a three-dimensional tensor appropriate for MLP and Mamba operations. Inspired by UltraLight VM-UNet and Vision Mamba, we proposed four versions of MUCM-Net with different patch processing. Figure 4 presents the visible differences between the UCM-Net block and the two visions of Mamba-UCM blocks. The PyTorch-style pseudocode in Figure 5 presents our defined sequence of operations, which is how we combine the UCM block and the Mamba block for feature learning.

    MUCM-Net Structure. (A) UCM-Net pipeline; (B) MUCM-Net (1-patch) pipeline; (C) MUCM-Net (2-patch) pipeline. SSM: state-space model; Norm: normalization; Conv2D: 2-dimensional convolution

    Mamba-UCM block pseudocode

    Loss functions

    In our solution, we designed a new group loss function similar to those used in TransFuse, EGE-UNet, and our previous work, UCM-Net. However, different from theirs, our proposed base loss function is calculated from Binary Cross-Entropy (BCE) loss (1) and Dice loss (2) components and square Dice loss (3) components to calculate the loss from the scaled layer masks in different stages compared with the ground truth masks.

    The advantages of using different loss functions are as follows:

    • BCE loss: Widely used in classification tasks, including biomedical image segmentation, these losses are highly effective for pixel-level segmentation.

    • Dice loss: Commonly used in biomedical image segmentation, Dice loss addresses class imbalance by focusing on the overlap between predicted and true regions.

    • Square Dice loss: Further enhances the Dice loss by emphasizing the contribution of well-predicted regions, thereby improving stability and performance.

    B C E = 1 N i = 1 N [ y i l o g ( p i ) + ( 1 y i ) l o g ( 1 p i ) ]

    where N is the total number of pixels (for image segmentation) or elements (for other tasks), yi is the ground truth value, and pi is the predicted probability for the i-th element.

     Dice Loss  = 1 2 × i = 1 N ( p i y i ) + smooth i = 1 N p i + i = 1 N y i + smooth

    where we used a smooth constant, which equals 1e-5, to enhance numerical stability.

     Square Dice Loss  = 1 2 × ( i = 1 N ( p i y i ) ) 2 + smooth ( i = 1 N p i ) 2 + ( i = 1 N y i ) 2 + smooth

    where the equation represents an improvement over the standard Dice loss by placing greater emphasis on the squared terms of intersections and unions.

    Base_loss=BCE+Dice Loss+Square Dice Loss

    Equations 1, 2, and 3 define the base loss function (Equation 4) for our proposed model, incorporating the Dice loss, and square Dice loss components. λi represents the weight assigned to different stages. We explore adding the difference between the up-sample from the stage with ground truth to the final loss. Because the large stage’s result should be close to the final prediction, we set λi to 0.1, 0.2, 0.3, 0.4, and 0.5 based on the i-th stage, as illustrated in Figure 3.

    Equation 7 defines our proposed group loss function, which calculates the loss based on the scaled layer masks at various stages compared to the ground truth masks. Equations 5 and 6 detail the stage loss for different intermediate layers and the output loss for the final layer, respectively.

    L o s s S t a g e = Base_loss ( S t a g e P r e d , T a r g e t )
    L o s s O u t p u t = Base_loss ( O u t p u t P r e d , T a r g e t )
    G r o u p_L o s s = L o s s O u t p u t + i = 1 5 λ i × L o s s S t a g e i

    To prove the effectiveness of our proposed methods, we did an ablation experiment of various base loss functions and the experimental results are reported in the following Discussion section.

    Results

    Datasets: To evaluate the efficiency and performance of our proposed model with other published models, we picked the two public skin segmentation datasets from the ISIC, namely ISIC2017 and ISIC2018. The ISIC2017 dataset contains 2,000 dermoscopy images, while the ISIC2018 dataset includes 2,594 images. The ISIC2017 dataset was randomly divided into 1,250 for training, 150 for validation, and 600 for testing. The ISIC2018 dataset was randomly divided into 1,815 for training, 259 for validation, and 520 for testing.

    Evaluation setting: Our MUCM-Net is implemented with the PyTorch [40] framework. All experiments are conducted on the instance node at Lambda [41] that has a single NVIDIA RTX A6000 GPU (24 GB), 14vCPUs, 46 GiB RAM, and 512 GiB SSD. We normalized and resized the images to 256 × 256 pixels. To enhance data sample diversity, we applied basic data augmentations, including horizontal and vertical flipping, as well as random rotation. This is because, although there are various and effective image augmentation techniques in the current literature [42], they may lead to an increase in the computational cost. We selected AdamW [43] for the optimizer, initialized with a learning rate of 0.001 and a weight decay of 0.01. The CosineAnnealingLR scheduler [44] is used with a maximum number of iterations set to 50 and a minimum learning rate of 1e-5. The model is trained for a total of 200 epochs with a training batch size of 8 and a testing batch size of 1.

    Evaluation metrics: The model’s performance is evaluated using the Dice Similarity Coefficient (DSC), sensitivity (SE), specificity (SP), and accuracy (ACC). Furthermore, the model’s memory consumption is assessed based on the number of parameters and GFLOPS. DSC measures the degree of similarity between the ground truth and the predicted segmentation map. SE is used to measure the percentage of true positives in relation to the sum of true positives and false negatives. SP measures the percentage of true negatives in relation to the sum of true negatives and false positives. ACC measures the overall percentage of correct classifications. The formulas used are as follows:

    DSC=2×TP2×TP+FP+FN
    ACC=TP + TNTP + TN + FP + FN
    SE=TPTP+FN
    SP=TNTN+FP

    where TP denotes true positive, TN denotes true negative, FP denotes false positive, and FN denotes false negative. In our benchmark tests, we evaluate the effectiveness of our approach and contrast it with other existing high-performing models.

    Experiments results: Tables 1 and 2 present the segmentation prediction performance, while Table 3 provides an overview of the models’ physical performance.

    Comparative prediction results on the ISIC2017 dataset

    DatasetModelsYearDSC↑SE↑SP↑ACC↑
    ISIC2017U-Net*20150.89890.87930.98120.9613
    SCR-Net*20210.88980.84970.98530.9588
    C2SDG*20210.89380.88590.97650.9588
    ATTENTION SWIN U-NET*20220.88590.84920.98470.9591
    MALUNet*20220.88960.88240.97620.9583
    UNeXt-S#20220.90170.88940.98060.9633
    EGE-UNet*20230.90730.89310.98160.9642
    VM-UNet*20240.90700.88370.98420.9645
    VM-UNet v2*20240.90450.87680.98490.9637
    LightM-UNet*20240.90800.88390.98460.9649
    UltraLight VM-UNet*20240.90910.90530.97900.9646
    UltraLight VM-UNet#20240.90970.90420.98040.9660
    UCM-Net (Baseline)20240.91200.88240.98770.9678
    MUCM-Net (1-patch)20240.91600.90900.98690.9689
    MUCM-Net (2-patch)20240.91260.90080.98290.9679
    MUCM-Net (4-patch)20240.91330.88710.98700.9681
    MUCM-Net (8-patch)20240.91850.90140.98570.9697
    Display full size

    * the results cited from UltraLight VM-Net; # the results tested by us. ISIC: International Skin Imaging Collaboration; DSC: Dice Similarity Coefficient; SE: sensitivity; SP: specificity; ACC: accuracy. ↑: the higher the value, the better the performance

    Comparative prediction results on the ISIC2018

    DatasetModelsYearDSC↑SE↑SP↑ACC↑
    ISIC2018U-Net*20150.88510.87350.97440.9539
    SCR-Net*20210.88860.88920.97140.9547
    C2SDG*20210.88060.89700.96430.9506
    ATTENTION SWIN U-NET*20220.85400.80570.98260.9480
    MALUNet*20220.89310.88900.97250.9548
    UNeXt-S#20220.89100.85770.98180.9555
    EGE-UNet*20230.88190.90090.96380.9510
    VM-UNet*20240.88910.88090.97430.9554
    VM-UNet v2*20240.89020.89590.97020.9551
    LightM-UNet*20240.88980.88290.97650.9555
    UltraLight VM-UNet*20240.89400.86800.97810.9558
    UltraLight VM-UNet#20240.89050.87240.97660.9545
    UCM-Net (Baseline)20240.90600.90410.97530.9602
    MUCM-Net (1-patch)20240.91110.89530.98110.9629
    MUCM-Net (2-patch)20240.90400.91510.97060.9588
    MUCM-Net (4-patch)20240.90580.87520.98460.9614
    MUCM-Net (8-patch)20240.90950.90460.97720.9618
    Display full size

    * the results cited from UltraLight VM-Net; # the results tested by us. ISIC: International Skin Imaging Collaboration; DSC: Dice Similarity Coefficient; SE: sensitivity; SP: specificity; ACC: accuracy. ↑: the higher the value, the better the performance

    Comparative performance results on models’ computations and the number of parameters

    ModelsYearParams (millions)↓GFLOPS↓
    U-Net20152.0093.224
    SCR-Net20210.8011.567
    C2SDG202122.0017.972
    ATTENTION SWIN U-NET202246.9114.181
    MALUNet20220.1750.083
    UNeXt-S20220.2530.104
    EGE-UNet20230.0530.072
    VM-UNet202427.4274.112
    VM-UNet v2202422.7714.400
    LightM-UNet20240.4030.391
    UltraLight VM-UNet20240.0490.065
    UCM-Net (Baseline)20240.0470.045
    MUCM-Net (1-patch)20240.1390.064
    MUCM-Net (2-patch)20240.1000.059
    MUCM-Net (4-patch)20240.0810.057
    MUCM-Net (8-patch)20240.0710.055
    Display full size

    GFLOPS: Giga Floating-point Operations Per Second. ↓: the lower the value, the better the performance

    Discussion

    Experimental result analysis

    Tables 1 and 2 provide a thorough evaluation of the prediction performance of MUCM-Net, our novel Mamba-based skin lesion segmentation model, in comparison to established models. This assessment is conducted using the widely recognized ISIC2017 and ISIC2018 datasets. Introduced in 2024, MUCM-Net proves to be a robust and highly competitive solution in this domain. The key takeaway from these Tables is MUCM-Net’s ability to outperform all previous models, establishing a new state-of-the-art for lightweight skin lesion segmentation. Our model presents exceptional performance across various prediction metrics, underscoring its advancement in the field and potential to redefine the standard for accurate skin lesion delineation. To assess the physical performance of our proposed model, we compared its computational aspects and number of parameters with those of different segmentation models, as shown in Table 3. Remarkably, MUCM-Net (8-patch) operates with lower GFLOPS compared to other Mamba-based models. This efficiency does not come at the cost of performance, as MUCM-Net maintains high accuracy and robustness in segmentation tasks with the Mamba structure.

    Ablation experiment on loss functions

    To prove the effects of our proposed loss function, we conducted an ablation experiment using different base loss functions to train our model, MUCM-Net (1-patch). Table 4 presents the experimental results on the ISIC2017 dataset, providing a comparative analysis of the performance metrics for each loss function. The base loss functions evaluated include BCE loss (A), Dice loss (B), and square Dice loss (C). Additionally, combinations of these loss functions were tested: A + B, A + C, B + C, and A + B + C. The performance metrics considered were the DSC, SE, SP, and ACC. The training settings and hardware environment are the same as the benchmarks in Tables 1 and 2.

    Ablation experiments on loss functions

    DatasetBase loss functionDSC↑SE↑SP↑ACC↑
    ISIC2017BCE loss (A)0.91300.88780.98670.9680
    Dice loss (B)0.91520.90080.98140.9681
    Square Dice loss (C)0.91550.90510.98310.9684
    A + B0.91470.90780.98200.9680
    A + C0.91580.90100.98580.9677
    B + C0.91010.88540.98590.9669
    A + B + C (our proposed)0.91600.90890.98680.9690
    Display full size

    ISIC: International Skin Imaging Collaboration; BCE: Binary Cross-Entropy; DSC: Dice Similarity Coefficient; SE: sensitivity; SP: specificity; ACC: accuracy. ↑: the higher the value, the better the performance

    Table 4 reveals that individual loss functions and their combinations exhibit varying performance on the ISIC2017 dataset. Among the individual loss functions, BCE loss (A) demonstrates a solid overall performance with a high SP of 0.9867 and ACC of 0.9680. However, its SE is relatively lower at 0.8878. Dice loss (B) improves SE to 0.9008 and maintains high ACC (0.9681), while square Dice loss (C) further enhances SE to 0.9051 and retains high SP (0.9831) and ACC (0.9684). Combinations of these loss functions generally yield better results, with A + B and A + C showing improvements in DSC and SE. The proposed combination of A + B + C achieves the best overall performance, with the highest DSC (0.9160), SE (0.9089), SP (0.9868), and ACC (0.9690). This indicates that integrating all three loss functions leverages their strengths, resulting in a more robust model performance across all evaluation metrics.

    In conclusion, this paper introduces MUCM-Net, an innovative and efficient solution that integrates CNN, MLP, and Mamba to offer robust feature learning while keeping parameter counts low and computational demands minimal. Applied to the challenging task of skin lesion segmentation, MUCM-Net has been rigorously tested across various evaluation metrics, demonstrating superior performance compared to other recent lightweight or Mamba-based models.

    Looking ahead, we plan to extend the application of MUCM-Net to other critical medical imaging tasks, aiming to advance the field and explore its potential across a broader range of healthcare applications. Additionally, we plan to explore how MUCM-Net can be effectively integrated with established hand-crafted segmentation methods (e.g., from [45, 46]) to leverage their complementary strengths and potentially achieve even higher segmentation accuracy. We will also investigate methods to address adversarial noise attacks on skin cancer segmentation models [47, 48], enhancing MUCM-Net’s robustness against potential manipulations that could compromise its performance. Our goal is to push the boundaries of deep learning in healthcare applications, making significant contributions to medical imaging as well as other areas of healthcare.

    Abbreviations

    ACC:

    accuracy

    AI:

    artificial intelligence

    BCE:

    Binary Cross-Entropy

    CNNs:

    Convolutional Neural Networks

    DSC:

    Dice Similarity Coefficient

    GFLOPS:

    Giga Floating-point Operations Per Second

    ISIC:

    International Skin Imaging Collaboration

    MLPs:

    multi-layer perceptions

    SE:

    sensitivity

    SP:

    specificity

    SSMs:

    state-space models

    TinyML:

    tiny machine learning

    Supplementary materials

    The supplementary figures for this article are available at: https://www.explorationpub.com/uploads/Article/file/1001250_sup_1.pdf.

    Declarations

    Author contributions

    CY: Conceptualization, Investigation, Writing—original draft, Writing—review & editing. DZ: Validation, Writing—review & editing. SSA: Validation, Writing—review & editing, Supervision. All authors read and approved the submitted version.

    Conflicts of interest

    The authors declare that they have no conflicts of interest.

    Ethical approval

    Not applicable.

    Consent to participate

    Not applicable.

    Consent to publication

    Not applicable.

    Availability of data and materials

    The datasets analyzed for this study can be found in ISIC17 Dataset: https://challenge.isic-archive.com/data/#2017; ISIC18 Dataset: https://challenge.isic-archive.com/data/#2018; Code Repo: https://github.com/chunyuyuan/MUCM-Net.

    Funding

    Not applicable.

    Copyright

    © The Author(s) 2024.

    References

    Brown SL, Fisher P, Hope-Stone L, Damato B, Heimann H, Hussain R, et al. Fear of cancer recurrence and adverse cancer treatment outcomes: predicting 2- to 5-year fear of recurrence from post-treatment symptoms and functional problems in uveal melanoma survivors. J Cancer Surviv. 2023;17:18796. [DOI] [PubMed]
    Siegel RL, Miller KD, Wagle NS, Jemal A. Cancer statistics, 2023. CA Cancer J Clin. 2023;73:1748. [DOI] [PubMed]
    Marks R. An overview of skin cancers. Cancer. 1995;75:60712. [DOI] [PubMed]
    Lehmann TM, Gönner C, Spitzer K. Survey: interpolation methods in medical image processing. IEEE Trans Med Imaging. 1999;18:104975. [DOI] [PubMed]
    Goyal R, Husain S, Wilson K, Chopra H, Pahwa R, Loganathan M, et al. Recent advancements in skin cancer treatment: a critical review. Explor Med. 2023;4:782812. [DOI]
    Nasir M, Attique Khan M, Sharif M, Lali IU, Saba T, Iqbal T. An improved strategy for skin lesion detection and classification using uniform segmentation and feature selection based approach. Microsc Res Tech. 2018;81:52843. [DOI] [PubMed]
    Barata C, Celebi ME, Marques JS. Explainable skin lesion diagnosis using taxonomies. Pattern Recognit. 2021;110:107413. [DOI]
    Hosny KM, Elshoura D, Mohamed ER, Vrochidou E, Papakostas GA. Deep Learning and Optimization-Based Methods for Skin Lesions Segmentation: A Review. IEEE Access. 2023;11:8546788. [DOI]
    Huang HY, Hsiao YP, Karmakar R, Mukundan A, Chaudhary P, Hsieh SC, et al. A Review of Recent Advances in Computer-Aided Detection Methods Using Hyperspectral Imaging Engineering to Detect Skin Cancer. Cancers (Basel). 2023;15:5634. [DOI] [PubMed] [PMC]
    Sanchez I, Agaian S. A new system of computer-aided diagnosis of skin lesions. In: Egiazarian KO, Agaian SS, Gotchev AP, Recker J, Wang G, editors. Image Processing: Algorithms and Systems X; and Parallel Processing for Imaging Applications II. IS&T/SPIE Electronic Imaging; 2012 Jan 22-26; Burlingame, United States. 2012. [DOI]
    de Carvalho TM, Noels E, Wakkee M, Udrea A, Nijsten T. Development of Smartphone Apps for Skin Cancer Risk Assessment: Progress and Promise. JMIR Dermatol. 2019;2:e13376. [DOI]
    Butterfly iQ+TM [Internet]. Butterfly Network, inc; c2024 [cited 2024 Jul 30]. Available from: https://www.butterflynetwork.com/iq-plus
    Bui P, Liu Y. Using AI to help find answers to common skin conditions [Internet]. [cited 2024 Jul 30]. Available from: https://blog.google/technology/health/ai-dermatology-preview-io-2021/
    Esteva A, Robicquet A, Ramsundar B, Kuleshov V, DePristo M, Chou K, et al. A guide to deep learning in healthcare. Nat Med. 2019;25:249. [DOI] [PubMed]
    Chen C, Zhang P, Zhang H, Dai J, Yi Y, Zhang H, et al. Deep Learning on Computational-Resource-Limited Platforms: A Survey. Mobile Inf Syst. 2020;2020:8454327. [DOI]
    Thompson NC, Greenewald K, Lee K, Manso GF. The Computational Limits of Deep Learning. arXiv:2007.05558v1 [Preprint]. 2020 [cited 2024 Jul 30]: [46 p.]. Available from: https://doi.org/10.48550/arXiv.2007.05558
    Yuan C, Zhao D, Agaian SS. UCM-Net: A lightweight and efficient solution for skin lesion segmentation using MLP and CNN. Biomed Signal Process Control. 2024;96:106573. [DOI]
    ISIC Challenge [dataset]. 2017 [cited 2024 Jul 30]. Available from: https://challenge.isic-archive.com/data/#2017
    Berseth M. ISIC 2017 - Skin Lesion Analysis Towards Melanoma Detection. arXiv:1703.00523 [Preprint]. 2017 [cited 2024 Jul 30]: [4 p.]. Available from: https://doi.org/10.48550/arXiv.1703.00523
    Tschandl P, Rosendahl C, Kittler H. The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions. Sci Data. 2018;5:180161. [DOI] [PubMed] [PMC]
    Codella N, Rotemberg V, Tschandl P, Celebi ME, Dusza S, Gutman D, et al. Skin Lesion Analysis Toward Melanoma Detection 2018: A Challenge Hosted by the International Skin Imaging Collaboration (ISIC). arXiv:1902.03368 [Preprint]. 2019 [cited 2024 Jul 30]: [12 p.]. Available from: https://doi.org/10.48550/arXiv.1902.03368
    Kim H, Khan MUK, Kyung CM. Efficient Neural Network Compression. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Conference on Computer Vision and Pattern Recognition (CVPR); 2019 Jun 15-20; Long Beach, USA. IEEE; 2019. pp. 12561–9. [DOI]
    Yuan C, Agaian SS. A comprehensive review of Binary Neural Network. Artif Intell Rev. 2023;56:129493013. [DOI]
    Yang Z, Chen Y, Huangfu H, Ran M, Wang H, Li X, et al. Dynamic Corrected Split Federated Learning With Homomorphic Encryption for U-Shaped Medical Image Networks. IEEE J Biomed Health Inform. 2023;27:594657. [DOI] [PubMed]
    Oktay O, Schlemper J, Folgoc LL, Lee M, Heinrich M, Misawa K, et al. Attention U-Net: Learning Where to Look for the Pancreas. arXiv:1804.03999 [Preprint]. 2018 [cited 2024 Jul 30]: [10 p.]. Available from: https://doi.org/10.48550/arXiv.1804.03999
    Chen J, Lu Y, Yu Q, Luo X, Adeli E, Wang Y, et al. TransUNet: Transformers Make Strong Encoders for Medical Image Segmentation. arXiv:2102.04306 [Preprint]. 2021 [cited 2024 Jul 30]: [13 p.]. Available from: https://doi.org/10.48550/arXiv.2102.04306
    Zhang Y, Liu H, Hu Q. TransFuse: Fusing Transformers and CNNs for Medical Image Segmentation. In: de Bruijne M, Cattin PC, Cotin S, Padoy N, Speidel S, Zheng Y, et al., editors. Medical Image Computing and Computer Assisted Intervention – MICCAI 2021. International Conference on Medical Image Computing and Computer-Assisted Intervention; 2021 Sep 27-Oct 1; Strasbourg, France. Cham: Springer International Publishing; 2021. pp. 14–24. [DOI]
    Cao H, Wang Y, Chen J, Jiang D, Zhang X, Tian Q, et al. Swin-Unet: Unet-Like Pure Transformer for Medical Image Segmentation. In: Karlinsky L, Michaeli T, Nishino K, editors. Computer Vision – ECCV 2022 Workshops. European Conference on Computer Vision; 2022 Oct 23-27; Tel Aviv, Israel. Cham: Springer Nature Switzerland; 2023. pp. 205–18. [DOI]
    Liu Z, Mao H, Wu CY, Feichtenhofer C, Darrell T, Xie S. A ConvNet for the 2020s. In: 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Conference on Computer Vision and Pattern Recognition (CVPR); 2022 Jun 18-24; New Orleans, USA. IEEE; 2022. pp. 11966–76. [DOI]
    Valanarasu JMJ, Patel VM. UNeXt: MLP-Based Rapid Medical Image Segmentation Network. In: Wang L, Dou Q, Fletcher PT, Speidel S, Li S, editors. Medical Image Computing and Computer Assisted Intervention – MICCAI 2022. International Conference on Medical Image Computing and Computer-Assisted Intervention; 2022 Sep 18-22; Singapore, Singapore. Cham: Springer Nature Switzerland; 2022. pp. 23–33. [DOI]
    Ruan J, Xiang S, Xie M, Liu T, Fu Y. MALUNet: A Multi-Attention and Light-weight UNet for Skin Lesion Segmentation. In: 2022 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). IEEE International Conference on Bioinformatics and Biomedicine (BIBM); 2022 Dec 6-8; Las Vegas, USA. IEEE; 2022. pp. 1150–6. [DOI]
    Ruan J, Xie M, Gao J, Liu T, Fu Y. EGE-UNet: an Efficient Group Enhanced UNet for skin lesion segmentation. arXiv:2307.08473 [Preprint]. 2023 [cited 2024 Jul 30]: [10 p.]. Available from: https://doi.org/10.48550/arXiv.2307.08473
    Zhu L, Liao B, Zhang Q, Wang X, Liu W, Wang X. Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model. arXiv:2401.09417 [Preprint]. 2024 [cited 2024 Jul 30]: [11 p.]. Available from: https://doi.org/10.48550/arXiv.2401.09417
    Ruan J, Xiang S. VM-UNet: Vision Mamba UNet for Medical Image Segmentation. arXiv:2402.02491 [Preprint]. 2024 [cited 2024 Jul 30]: [12 p.]. Available from: https://doi.org/10.48550/arXiv.2402.02491
    Zhang M, Yu Y, Gu L, Lin T, Tao X. VM-UNET-V2 Rethinking Vision Mamba UNet for Medical Image Segmentation. arXiv:2403.09157 [Preprint]. 2024 [cited 2024 Jul 30]: [12 p.]. Available from: https://doi.org/10.48550/arXiv.2403.09157
    Liao W, Zhu Y, Wang X, Pan C, Wang Y, Ma L. LightM-UNet: Mamba Assists in Lightweight UNet for Medical Image Segmentation. arXiv:2403.05246 [Preprint]. 2024 [cited 2024 Jul 30]: [10 p.]. Available from: https://doi.org/10.48550/arXiv.2403.05246
    Wu R, Liu Y, Liang P, Chang Q. UltraLight VM-UNet: Parallel Vision Mamba Significantly Reduces Parameters for Skin Lesion Segmentation. arXiv:2403.20035 [Preprint]. 2024 [cited 2024 Jul 30]: [16 p.]. Available from: https://doi.org/10.48550/arXiv.2403.20035
    Wang X, Wang S, Ding Y, Li Y, Wu W, Rong Y, et al. State Space Model for New-Generation Network Alternative to Transformers: A Survey. arXiv:2404.09516 [Preprint]. 2024 [cited 2024 Jul 30]: [33 p.]. Available from: https://doi.org/10.48550/arXiv.2404.09516
    Gu A, Dao T. Mamba: Linear-Time Sequence Modeling with Selective State Spaces. arXiv:2312.00752v1 [Preprint]. 2023 [cited 2024 Jul 30]: [37 p.]. Available from: https://doi.org/10.48550/arXiv.2312.00752
    Paszke A, Gross S, Massa F, Lerer A, Bradbury J, Chanan G, et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In: Wallach H, Larochelle H, Beygelzimer A, d’Alché-Buc F, Fox E, Garnett R, editors. Advances in Neural Information Processing Systems. Curran Associates, Inc.; 2019.
    lambdalabs.com [Internet]. c2024 [cited 2024 Aug 10]. Available from: https://lambdalabs.com/
    Goceri E. Medical image data augmentation: techniques, comparisons and interpretations. Artif Intell Rev. 2023;56:12561605. [DOI] [PubMed] [PMC]
    Loshchilov I, Hutter F. Decoupled Weight Decay Regularization. arXiv:1711.05101 [Preprint]. 2017 [cited 2024 Jul 30]: [19 p.]. Available from: https://doi.org/10.48550/arXiv.1711.05101
    Loshchilov I, Hutter F. SGDR: Stochastic Gradient Descent with Warm Restarts. arXiv:1608.03983 [Preprint]. 2016 [cited 2024 Jul 30]: [16 p.]. Available from: https://doi.org/10.48550/arXiv.1608.03983
    Sanchez I, Agaian S. Computer aided diagnosis of lesions extracted from large skin surfaces. In: 2012 IEEE International Conference on Systems, Man, and Cybernetics (SMC). IEEE International Conference on Systems, Man and Cybernetics; 2012 Oct 14-17; Seoul, Korea (South). IEEE; 2012. pp. 2879–84. [DOI]
    Frants V, Agaian S. Dermoscopic image segmentation based on modified GrabCut with octree color quantization. In: Agaian SS, Asari VK, DelMarco SP, Jassim SA, editors. Mobile Multimedia/Image Processing, Security, and Applications 2020. SPIE Defense + Commercial Sensing; 2020 Apr 27-May 9; California, United States. 2020. [DOI]
    Liew A, Agaian S, Zhao L. Mitigation of adversarial noise attacks on skin cancer detection via ordered statistics binary local features. In: Agaian SS, Asari VK, DelMarco SP, editors. Multimodal Image Exploitation and Learning 2023. SPIE Defense + Commercial Sensing; 2023 Apr 30-May 5; Orlando, United States. 2023. [DOI]
    Goceri E. Evaluation of denoising techniques to remove speckle and Gaussian noise from dermoscopy images. Comput Biol Med. 2023;152:106474. [DOI] [PubMed]