In supervised learning, low quality annotations lead to poorly performing classification and detection models, while also rendering evaluation unreliable. This is particularly apparent on temporal data, where annotation quality is affected by multiple factors. For example, in the post-hoc self-reporting of daily activities, cognitive biases are one of the most common ingredients. In particular, reporting the start and duration of an activity after its finalisation may incorporate biases introduced by personal time perceptions, as well as the imprecision and lack of granularity due to time rounding. Here we propose a method to model human biases on temporal annotations and argue for the use of soft labels. Experimental results in synthetic data show that soft labels provide a better approximation of the ground truth for several metrics. We showcase the method on a real dataset of daily activities.
翻译:在受监督的学习中,低质量说明导致分类和检测模型工作不良,同时也使评价不可靠,这在时间数据中特别明显,因为时间数据说明的质量受到多种因素的影响。例如,在对日常活动的事后自我报告中,认知偏差是最常见的因素之一,特别是,在活动完成后报告活动的开始和持续时间可能包含个人时间感的偏差,以及不精确和由于时间四舍五入而缺乏颗粒。我们在这里提出了一个方法,用以模拟人类对时间说明的偏差,并论证使用软标签。合成数据的实验结果表明,软标签为若干计量标准提供了更好的地面真相近似。我们用真实的日常活动数据集展示了这种方法。