
Please use this identifier to cite or link to this item:
http://dspace.bits-pilani.ac.in:8080/jspui/handle/123456789/19260
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Chalapathi, G.S.S. | - |
dc.date.accessioned | 2025-08-29T04:20:23Z | - |
dc.date.available | 2025-08-29T04:20:23Z | - |
dc.date.issued | 2025-07 | - |
dc.identifier.uri | https://arxiv.org/abs/2507.00511 | - |
dc.identifier.uri | http://dspace.bits-pilani.ac.in:8080/jspui/handle/123456789/19260 | - |
dc.description.abstract | In this paper, we present the VMSE U-Net and VM-Unet CBAM+ model, two cutting-edge deep learning architectures designed to enhance medical image segmentation. Our approach integrates Squeeze-and-Excitation (SE) and Convolutional Block Attention Module (CBAM) techniques into the traditional VM U-Net framework, significantly improving segmentation accuracy, feature localization, and computational efficiency. Both models show superior performance compared to the baseline VM-Unet across multiple datasets. Notably, VMSEUnet achieves the highest accuracy, IoU, precision, and recall while maintaining low loss values. It also exhibits exceptional computational efficiency with faster inference times and lower memory usage on both GPU and CPU. Overall, the study suggests that the enhanced architecture VMSE-Unet is a valuable tool for medical image analysis. These findings highlight its potential for real-world clinical applications, emphasizing the importance of further research to optimize accuracy, robustness, and computational efficiency. | en_US |
dc.language.iso | en | en_US |
dc.subject | EEE | en_US |
dc.subject | Medical image segmentation | en_US |
dc.subject | Deep learning | en_US |
dc.title | Medical image segmentation using advanced UNETt: VMSE-Unet and VM-Unet CBAM+ | en_US |
dc.type | Preprint | en_US |
Appears in Collections: | Department of Electrical and Electronics Engineering |
Files in This Item:
There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.