Obtaining demographics information from video is valuable for a range of real-world applications. While approaches that leverage facial features for gender inference are very successful in restrained environments, they do not work in most real-world scenarios when the subject is not facing the camera, has the face obstructed or the face is not clear due to distance from the camera or poor resolution. We propose a weakly-supervised method for learning gender information of people based on their manner of walking. We make use of state-of-the art facial analysis models to automatically annotate front-view walking sequences and generalise to unseen angles by leveraging gait-based label propagation. Our results show on par or higher performance with facial analysis models with an F1 score of 91% and the ability to successfully generalise to scenarios in which facial analysis is unfeasible due to subjects not facing the camera or having the face obstructed.
翻译:从视频获取人口信息对于一系列现实世界应用来说是有价值的。虽然在受限环境中利用面部特征进行性别推断的方法非常成功,但在大多数现实世界情景中,当对象不面对相机、脸部受到阻碍或脸部因距离相机不清晰或分辨率不清晰而不清楚时,这些方法并不奏效。我们建议了一种薄弱的监控方法,根据人们的行走方式来学习其性别信息。我们利用最新水平的面部分析模型,通过利用以格子为基础的标签传播,自动对前视行走序列进行注注解,并概括到看不见的角度。我们的成果显示,与面部分析模型的相同或更高性能,面部分析模型的F1分为91%,而且能够成功地概括到面部分析因不面对相机或面部受到阻碍而不可行的情景。