We propose a novel Attentional Scale Sequence Fusion based You Only Look Once (YOLO) framework (ASF-YOLO) which combines spatial and scale features for accurate and fast cell instance segmentation. Built on the YOLO segmentation framework, we employ the Scale Sequence Feature Fusion (SSFF) module to enhance the multi-scale information extraction capability of the network, and the Triple Feature Encoder (TFE) module to fuse feature maps of different scales to increase detailed information. We further introduce a Channel and Position Attention Mechanism (CPAM) to integrate both the SSFF and TPE modules, which focus on informative channels and spatial position-related small objects for improved detection and segmentation performance. Experimental validations on two cell datasets show remarkable segmentation accuracy and speed of the proposed ASF-YOLO model. It achieves a box mAP of 0.91, mask mAP of 0.887, and an inference speed of 47.3 FPS on the 2018 Data Science Bowl dataset, outperforming the state-of-the-art methods. The source code is available at https://github.com/mkang315/ASF-YOLO.
翻译:我们提出了一种新颖的基于注意力尺度序列融合的YOLO框架(ASF-YOLO),该框架结合空间和尺度特征以实现准确快速的细胞实例分割。在YOLO分割框架的基础上,我们采用尺度序列特征融合(SSFF)模块增强网络的多尺度信息提取能力,并利用三重特征编码器(TFE)模块融合不同尺度的特征图以增加细节信息。进一步引入通道与位置注意力机制(CPAM)来整合SSFF与TFE模块,使其聚焦于信息丰富的通道和与空间位置相关的小目标,从而提升检测与分割性能。在两个细胞数据集上的实验验证表明,所提出的ASF-YOLO模型具有显著的分割精度和速度。在2018年Data Science Bowl数据集上,该模型取得了0.91的边界框mAP、0.887的掩码mAP以及47.3 FPS的推理速度,超越了现有最优方法。源代码已开源至https://github.com/mkang315/ASF-YOLO。