Logical theories in the form of ontologies and similar artefacts in computing and IT are used for structuring, annotating, and querying data, among others, and therewith influence data analytics regarding what is fed into the algorithms. Algorithmic bias is a well-known notion, but what does bias mean in the context of ontologies that provide a structuring mechanism for an algorithm's input? What are the sources of bias there and how would they manifest themselves in ontologies? We examine and enumerate types of bias relevant for ontologies, and whether they are explicit or implicit. These eight types are illustrated with examples from extant production-level ontologies and samples from the literature. We then assessed three concurrently developed COVID-19 ontologies on bias and detected different subsets of types of bias in each one, to a greater or lesser extent. This first characterisation aims contribute to a sensitisation of ethical aspects of ontologies primarily regarding representation of information and knowledge.
翻译:在计算和信息技术中,以本体学和类似手工艺品为形式的逻辑理论用于构建、注解和查询数据,从而影响关于算法中所含内容的数据分析。 算法偏差是一个众所周知的概念,但在为算法输入提供结构化机制的本体学理论中,偏见意味着什么? 偏见的来源是什么,它们如何在本体学中表现出来? 我们研究和列举与本体有关的偏见类型,以及它们是明示的还是隐含的。这八种类型的数据以来自原生生产水平的本体学和文献样本的实例加以说明。然后,我们同时评估了三种关于偏向的COVID-19研究,并发现每一种偏向类型的不同类别,其程度或多或少。第一个特征化的目的是帮助对主要涉及信息和知识表述的本体学的伦理问题进行敏锐化。