Over the past decades the machine and deep learning community has celebrated great achievements in challenging tasks such as image classification. The deep architecture of artificial neural networks together with the plenitude of available data makes it possible to describe highly complex relations. Yet, it is still impossible to fully capture what the deep learning model has learned and to verify that it operates fairly and without creating bias, especially in critical tasks, for instance those arising in the medical field. One example for such a task is the detection of distinct facial expressions, called Action Units, in facial images. Considering this specific task, our research aims to provide transparency regarding bias, specifically in relation to gender and skin color. We train a neural network for Action Unit classification and analyze its performance quantitatively based on its accuracy and qualitatively based on heatmaps. A structured review of our results indicates that we are able to detect bias. Even though we cannot conclude from our results that lower classification performance emerged solely from gender and skin color bias, these biases must be addressed, which is why we end by giving suggestions on how the detected bias can be avoided.
翻译:在过去几十年里,机器和深层学习界在图像分类等具有挑战性的任务中庆祝了巨大的成就。人工神经网络的深层结构以及大量可用数据使得能够描述高度复杂的关系。然而,仍然无法充分捕捉深层次学习模式所学到的知识,并核实它是否公平运作,特别是在关键任务方面,特别是在医疗领域产生的那些任务方面。这种任务的一个例子是在面部图像中发现不同的面部表情,称为“行动股”。考虑到这一具体任务,我们的研究目的是在偏见方面提供透明度,特别是在性别和肤色方面。我们训练一个神经网络,以行动单位分类,并根据它的准确性和质量根据热测图从数量上分析其性能。对我们的结果进行结构化审查后显示,我们能够发现偏见。尽管我们无法从我们的结果中得出结论,较低分类性的工作表现完全来自性别和肤色的偏差,但必须解决这些偏见,这就是为什么我们最后就如何避免所发现的偏见提出建议。