Background modeling and subtraction is a promising research area with a variety of applications for video surveillance. Recent years have witnessed a proliferation of effective learning-based deep neural networks in this area. However, the techniques have only provided limited descriptions of scenes' properties while requiring heavy computations, as their single-valued mapping functions are learned to approximate the temporal conditional averages of observed target backgrounds and foregrounds. On the other hand, statistical learning in imagery domains has been a prevalent approach with high adaptation to dynamic context transformation, notably using Gaussian Mixture Models (GMM) with its generalization capabilities. By leveraging both, we propose a novel method called CDN-MEDAL-net for background modeling and subtraction with two convolutional neural networks. The first architecture, CDN-GM, is grounded on an unsupervised GMM statistical learning strategy to describe observed scenes' salient features. The second one, MEDAL-net, implements a light-weighted pipeline of online video background subtraction. Our two-stage architecture is small, but it is very effective with rapid convergence to representations of intricate motion patterns. Our experiments show that the proposed approach is not only capable of effectively extracting regions of moving objects in unseen cases, but it is also very efficient.
翻译:图像建模和减法是一个很有希望的研究领域,有各种视频监视应用。近些年来,在这方面出现了基于有效学习的深心神经网络的扩散。然而,这些技术仅提供了有限的场景特性描述,而需要大量计算,因为其单价绘图功能的单一价值绘图功能可以接近观测目标背景和前景背景的有条件时间平均数。另一方面,图像领域的统计学习是一个普遍做法,高度适应动态环境变化,特别是使用高山混合模型(GMMM)及其通用能力。通过利用这两种手段,我们提出了一种名为CDN-MEDAL-net的新颖方法,用于背景模型和两个革命性神经网络的减缩。第一个结构,即CDN-GM,以非超超强的GMM统计学习战略为基础,以描述观测到的场景色特征。第二个是MEDAL-net,采用轻量的在线视频背景减法管道。我们的两阶段结构很小,但非常有效,与复杂运动图案的立体快速融合。我们的实验还表明,在移动图案中,我们所提议得力的方法是有效的。