Spatial models of preference, in the form of vector embeddings, are learned by many deep learning systems including recommender systems. Often these models are assumed to approximate a Euclidean structure, where an individual prefers alternatives positioned closer to their "ideal point", as measured by the Euclidean metric. However, Bogomolnaia and Laslier (2007) showed that there exist ordinal preference profiles that cannot be represented with this structure if the Euclidean space has two fewer dimensions than there are individuals or alternatives. We extend this result, showing that there are realistic situations in which almost all preference profiles cannot be represented with the Euclidean model, and derive a theoretical lower bound on the information lost when approximating non-representable preferences with the Euclidean model. Our results have implications for the interpretation and use of vector embeddings, because in some cases close approximation of arbitrary, true preferences is possible only if the dimensionality of the embeddings is a substantial fraction of the number of individuals or alternatives.
翻译:许多深层学习系统,包括推荐系统,都学习了以矢量嵌入为形式的空间偏好模型,这些模型通常被假定为接近欧几里德结构,其中个人更喜欢以欧几里德衡量的接近其“理想点”的替代物。然而,Bogomolnaia和Lasierer(2007年)表明,如果欧几里德空间比个人或替代物少两个维度,则这种结构中无法代表的偏好特征特征特征特征特征特征。我们扩展了这一结果,表明存在现实的情况,即几乎所有的偏爱特征都无法与欧几里德模式相代表,并且从理论上上看,在与欧几里德模式相近非可呈现的偏好时,信息损失范围较低。我们的结果对矢量嵌入的诠释和使用具有影响,因为在某些情况下,由于任意性接近,只有嵌入的维度是个人或替代物数量的一大部分,才可能实现真正的偏好。