Group re-identification (G-ReID) is an important yet less-studied task. Its challenges not only lie in appearance changes of individuals which have been well-investigated in general person re-identification (ReID), but also derive from group layout and membership changes. So the key task of G-ReID is to learn representations robust to such changes. To address this issue, we propose a Transferred Single and Couple Representation Learning Network (TSCN). Its merits are two aspects: 1) Due to the lack of labelled training samples, existing G-ReID methods mainly rely on unsatisfactory hand-crafted features. To gain the superiority of deep learning models, we treat a group as multiple persons and transfer the domain of a labeled ReID dataset to a G-ReID target dataset style to learn single representations. 2) Taking into account the neighborhood relationship in a group, we further propose learning a novel couple representation between two group members, that achieves more discriminative power in G-ReID tasks. In addition, an unsupervised weight learning method is exploited to adaptively fuse the results of different views together according to result patterns. Extensive experimental results demonstrate the effectiveness of our approach that significantly outperforms state-of-the-art methods by 11.7\% CMC-1 on the Road Group dataset and by 39.0\% CMC-1 on the DukeMCMT dataset.
翻译:团体再定位是一项重要但研究较少的任务,其挑战不仅在于在一般个人再身份(ReID)中经过充分调查的个人的外貌变化,而且在于群体布局和成员变化。因此,G-ReID的关键任务是学习对此类变化的有力表现。为了解决这一问题,我们提议建立一个单一和双重代表制学习网络(TSCN),其优点有两个方面:(1)由于缺乏标签培训样本,现有的G-ReID方法主要依赖不令人满意的手工艺特征。为了获得深层次学习模式的优越性,我们把一个群体作为多人对待,并将一个标签的重信息数据集的域转移到G-ReID目标数据集的域,以学习单一的表达方式。(2)考虑到一个集团的邻里关系,我们进一步提议在两个集团成员之间学习新的一对一代表制,在G-REID任务中实现更具有歧视性的力量。此外,一种不超超重重的学习方法主要用于将不同观点的结果整合在一起。我们通过C-MMC第11项数据模式,大规模实验性结果展示了我们“C-MMM”系统的数据方法的实效。