CIFFormer: A Contextual Information Flow Guided Transformer for colorectal polyp segmentation DOI
Cunlu Xu, Long Lin, Bin Wang

et al.

Neurocomputing, Journal Year: 2025, Volume and Issue: unknown, P. 130413 - 130413

Published: May 1, 2025

Language: Английский

MTANet: Multi-Task Attention Network for Automatic Medical Image Segmentation and Classification DOI
Yating Ling, Yuling Wang, Wenli Dai

et al.

IEEE Transactions on Medical Imaging, Journal Year: 2023, Volume and Issue: 43(2), P. 674 - 685

Published: Sept. 19, 2023

Medical image segmentation and classification are two of the most key steps in computer-aided clinical diagnosis. The region interest were usually segmented a proper manner to extract useful features for further disease classification. However, these methods computationally complex time-consuming. In this paper, we proposed one-stage multi-task attention network (MTANet) which efficiently classifies objects an while generating high-quality mask each medical object. A reverse addition module was designed task fusion areas global map boundary cues high-resolution features, bottleneck used feature fusion. We evaluated performance MTANet with CNN-based transformer-based architectures across three imaging modalities different tasks: CVC-ClinicDB dataset polyp segmentation, ISIC-2018 skin lesion our private ultrasound liver tumor Our model outperformed state-of-the-art models on all datasets superior 25 radiologists

Language: Английский

Citations

45

Medical Image Segmentation Based on Transformer and HarDNet Structures DOI Creative Commons
Tongping Shen, Huanqing Xu

IEEE Access, Journal Year: 2023, Volume and Issue: 11, P. 16621 - 16630

Published: Jan. 1, 2023

Medical image segmentation is a crucial way to assist doctors in the accurate diagnosis of diseases. However, accuracy medical needs further improvement due problems many noisy images and high similarity between background target regions. The current mainstream networks, such as TransUnet, have achieved segmentation. Still, encoders networks do not consider local connection adjacent chunks lack interaction inter-channel information during upsampling decoder. To address above problems, this paper proposed dual-encoder network, including HarDNet68 Transformer branch, which can extract features global feature input image, allowing network learn more information, thus improving effectiveness In paper, realize fusion different dimensions two stages encoding decoding, we propose adaptation module fuse channel multi-level channels, then improve accuracy. experimental results on CVC-ClinicDB, ETIS-Larib, COVID-19 CT datasets show that model performs better four evaluation metrics, Dice, Iou, Prec, Sens, achieves both internal filling edge prediction images. Accurate making critical cancerous regions advance, ensure cancer patients receive timely targeted treatment, their survival quality.

Language: Английский

Citations

22

DeepPoly: Deep Learning-Based Polyps Segmentation and Classification for Autonomous Colonoscopy Examination DOI Creative Commons
Md Shakhawat Hossain, Md Mahmudur Rahman, M. M. Mahbubul Syeed

et al.

IEEE Access, Journal Year: 2023, Volume and Issue: 11, P. 95889 - 95902

Published: Jan. 1, 2023

Colorectal cancer (CRC) is the third most common cause of cancer-related deaths in United States and anticipated to another 52,580 2023. The standard medical procedure for screening treating colorectal disease a colonoscopy. By effectively examining colonoscopy identify precancerous polyps early remove them before they become cancerous, CRC mortality can be lowered significantly. Manual examination detection time-consuming, tedious, prone human error. Automatic segmentation could fast practical; however, existing automated methods fail attain adequate accuracy segmentation. Moreover, these do not assess risk detected polyps. In this paper, we proposed an autonomous method detect their potential threats. utilized DoubleU-Net Vision Transformer (ViT) classifying based on risks. has achieved mean dice-coefficient 0.834 0.956 Endotech challenge Kvasir-SEG dataset, accordingly, outperforming state-of-the-art Then, classified segmented as hyper-plastic or adenomatous with 99% test accuracy.

Language: Английский

Citations

19

Advancing Ocular Imaging: A Hybrid Attention Mechanism-Based U-Net Model for Precise Segmentation of Sub-Retinal Layers in OCT Images DOI Creative Commons
Prakash Kumar Karn, Waleed H. Abdulla

Bioengineering, Journal Year: 2024, Volume and Issue: 11(3), P. 240 - 240

Published: Feb. 28, 2024

This paper presents a novel U-Net model incorporating hybrid attention mechanism for automating the segmentation of sub-retinal layers in Optical Coherence Tomography (OCT) images. OCT is an ophthalmology tool that provides detailed insights into retinal structures. Manual these time-consuming and subjective, calling automated solutions. Our proposed combines edge spatial mechanisms with architecture to improve accuracy. By leveraging mechanisms, focuses selectively on image features. Extensive evaluations using datasets demonstrate our outperforms existing approaches, making it valuable medical professionals. The study also highlights model's robustness through performance metrics such as average Dice score 94.99%, Adjusted Rand Index (ARI) 97.00%, Strength Agreement (SOA) classifications like "Almost Perfect", "Excellent", "Very Strong". advanced predictive shows promise expediting processes enhancing precision ocular imaging real-world applications.

Language: Английский

Citations

9

BGLE-YOLO: A Lightweight Model for Underwater Bio-Detection DOI Creative Commons
Hua Zhao, Chao Xu, Jiaxing Chen

et al.

Sensors, Journal Year: 2025, Volume and Issue: 25(5), P. 1595 - 1595

Published: March 5, 2025

Due to low contrast, chromatic aberration, and generally small objects in underwater environments, a new fish detection model, BGLE-YOLO, is proposed investigate automated methods dedicated accurately detecting images. The model has parameters computational effort suitable for edge devices. First, an efficient multi-scale convolutional EMC module introduced enhance the backbone network capture dynamic changes targets environment. Secondly, global local feature fusion (BIG) integrated into neck preserve more information, reduce error information higher-level features, increase model’s effectiveness targets. Finally, prevent accuracy impact due excessive lightweighting, lightweight shared head (LSH) constructed. reparameterization technique further improves without additional cost. Experimental results of BGLE-YOLO on datasets DUO (Detection Underwater Objects) RUOD (Real-World Object Detection) show that achieves same as benchmark with ultra-low cost 6.2 GFLOPs parameter 1.6 MB.

Language: Английский

Citations

1

DGBL-YOLOv8s: An Enhanced Object Detection Model for Unmanned Aerial Vehicle Imagery DOI Creative Commons
Chonghao Wang, Huaian Yi

Applied Sciences, Journal Year: 2025, Volume and Issue: 15(5), P. 2789 - 2789

Published: March 5, 2025

Unmanned aerial vehicle (UAV) imagery often suffers from significant object scale variations, high target density, and varying distances due to shooting conditions environmental factors, leading reduced robustness low detection accuracy in conventional models. To address these issues, this study adopts DGBL-YOLOv8s, an improved model tailored for UAV perspectives based on YOLOv8s. First, a Dilated Wide Residual (DWR) module is introduced replace the C2f backbone network of YOLOv8, enhancing model’s capability capture fine-grained features contextual information. Second, neck structure redesigned by incorporating Global-to-Local Spatial Aggregation (GLSA) combined with Bidirectional Feature Pyramid Network (BiFPN), which strengthens feature fusion. Third, lightweight shared convolution head proposed, batch normalization techniques. Additionally, further improve small detection, dedicated small-object introduced. Results experiments VisDrone dataset reveal that DGBL-YOLOv8s enhances 8.5% relative baseline model, alongside 34.8% reduction parameter count. The overall performance exceeds most current models, confirms advantages proposed improvement.

Language: Английский

Citations

1

MGCBFormer: The multiscale grid-prior and class-inter boundary-aware transformer for polyp segmentation DOI
Yang Xia,

Haijiao Yun,

Yanjun Liu

et al.

Computers in Biology and Medicine, Journal Year: 2023, Volume and Issue: 167, P. 107600 - 107600

Published: Oct. 20, 2023

Language: Английский

Citations

14

A new segmentation algorithm for peripapillary atrophy and optic disk from ultra-widefield Photographs DOI
Cheng Wan, Jiyi Fang,

Kunke Li

et al.

Computers in Biology and Medicine, Journal Year: 2024, Volume and Issue: 172, P. 108281 - 108281

Published: March 13, 2024

Language: Английский

Citations

5

BMCS-Net: A Bi-directional multi-scale cascaded segmentation network based on transformer-guided feature Aggregation for medical images DOI
Bicao Li, Jing Wang, Bei Wang

et al.

Computers in Biology and Medicine, Journal Year: 2024, Volume and Issue: 180, P. 108939 - 108939

Published: July 29, 2024

Language: Английский

Citations

5

Improving Skin Lesion Segmentation with Self-Training DOI Open Access
Aleksandra Dzieniszewska, Piotr Garbat, Ryszard Piramidowicz

et al.

Cancers, Journal Year: 2024, Volume and Issue: 16(6), P. 1120 - 1120

Published: March 11, 2024

Skin lesion segmentation plays a key role in the diagnosis of skin cancer; it can be component both traditional algorithms and end-to-end approaches. The quality directly impacts accuracy classification; however, attaining optimal necessitates substantial amount labeled data. Semi-supervised learning allows for employing unlabeled data to enhance results machine model. In case medical image segmentation, acquiring detailed annotation is time-consuming costly requires skilled individuals so utilization significant mitigation manual efforts. This study proposes novel approach semi-supervised using self-training with Noisy Student. utilizing large amounts available images. It consists four steps—first, training teacher model on only, then generating pseudo-labels model, student pseudo-labeled data, lastly, student* generated this work, we implemented DeepLabV3 architecture as models. As final result, achieved mIoU 88.0% ISIC 2018 dataset 87.54% PH2 dataset. evaluation proposed shows that Student improves performance neural networks task while only small

Language: Английский

Citations

4