
Abstract Deep learning models are used to minimize the number of polyps that goes unnoticed by the experts and to accurately segment the detected polyps during interventions. Although state‐of‐the‐art models are proposed, it remains a challenge to define representations that are able to generalize well and that mediate between capturing low‐level features and higher‐level semantic details without being redundant. Another challenge with these models is that they are computation and memory intensive, which can pose a problem with real‐time applications. To address these problems, PlutoNet is proposed for polyp segmentation which requires only 9 FLOPs and 2,626,537 parameters, less than 10% of the parameters required by its counterparts. With PlutoNet, a novel decoder consistency training approach is proposed that consists of a shared encoder, the modified partial decoder , which is a combination of the partial decoder and full‐scale connections that capture salient features at different scales without redundancy, and the auxiliary decoder which focuses on higher‐level semantic features. The modified partial decoder and the auxiliary decoder are trained with a combined loss to enforce consistency, which helps strengthen learned representations. Ablation studies and experiments are performed which show that PlutoNet performs significantly better than the state‐of‐the‐art models, particularly on unseen datasets.
FOS: Computer and information sciences, Computer Vision and Pattern Recognition (cs.CV), Special Issue: Augmented Environments for Computer‐assisted Interventions (Ae‐cai) 2024, Image and Video Processing (eess.IV), Computer Science - Computer Vision and Pattern Recognition, Electrical Engineering and Systems Science - Image and Video Processing, medical image processing, computer vision, neural nets, convolutional neural nets, Medical technology, FOS: Electrical engineering, electronic engineering, information engineering, learning (artificial intelligence), R855-855.5, image segmentation
FOS: Computer and information sciences, Computer Vision and Pattern Recognition (cs.CV), Special Issue: Augmented Environments for Computer‐assisted Interventions (Ae‐cai) 2024, Image and Video Processing (eess.IV), Computer Science - Computer Vision and Pattern Recognition, Electrical Engineering and Systems Science - Image and Video Processing, medical image processing, computer vision, neural nets, convolutional neural nets, Medical technology, FOS: Electrical engineering, electronic engineering, information engineering, learning (artificial intelligence), R855-855.5, image segmentation
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 1 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
