In this paper, we propose two modified neural networks based on dual path multi-scale fusion networks (SFANet) and SegNet for accurate and efficient crowd counting. Inspired by SFANet, the first model, which is named M-SFANet, is attached with atrous spatial pyramid pooling (ASPP) and context-aware module (CAN). The encoder of M-SFANet is enhanced with ASPP containing parallel atrous convolutional layers with different sampling rates and hence able to extract multi-scale features of the target object and incorporate larger context. To further deal with scale variation throughout an input image, we leverage the CAN module which adaptively encodes the scales of the contextual information. The combination yields an effective model for counting in both dense and sparse crowd scenes. Based on the SFANet decoder structure, M-SFANet's decoder has dual paths, for density map and attention map generation. The second model is called M-SegNet, which is produced by replacing the bilinear upsampling in SFANet with max unpooling that is used in SegNet. This change provides a faster model while providing competitive counting performance. Designed for high-speed surveillance applications, M-SegNet has no additional multi-scale-aware module in order to not increase the complexity. Both models are encoder-decoder based architectures and are end-to-end trainable. We conduct extensive experiments on five crowd counting datasets and one vehicle counting dataset to show that these modifications yield algorithms that could improve state-of-the-art crowd counting methods. Codes are available at https://github.com/Pongpisit-Thanasutives/Variations-of-SFANet-for-Crowd-Counting.
翻译:在本文中,我们提出基于双路径多级聚合网络(SFANet)和SegNet的两套修改的神经网络,用于准确和高效的人群计数。受SFANet(第一个模型,名为M-SFANet)的启发,第一个模型(名为M-SFANet)与原始空间金字塔集合(ASP)和背景觉悟模块(CAN)相连。M-SFANet的编码由ASPP(MSFANet)组成,包含不同取样率的平行微变变相层,因此能够提取目标对象对象的多级变异特性,并纳入更大的背景。为了进一步处理一个输入图像图像图像中的规模变异,我们利用CNCAN(CAN)模块,该模块适应性化了背景信息尺度的编码。根据SFANet的解码器结构,M-SANet的解码器有双重路径,用于密度地图的生成。第二个模型称为M-SegawaNet(M-SO-SODO-ODO-O-ODOO-s-s-s-serviewdalmodalmodal-modalmodal-modal-modal-modal-modationalmodal-mod-mod-modations)) 正在取代Sal-modromodal-modational-s-dal-dal-dal-modal-modal-modal-modal-modal-modal-mod-modal-modal-modal-mod-mod-mod-mods-moxmusdal-mods-mods-mod-modal-modal-mod-mod-modal-modal-modal-mocal-mocal-mocal-modal-modal-modal-mods-mod-mod-mod-mod-mod-mod-mod-mod-modal-mod-mod-mod-mod-mod-mod-mod-mod-mod-