Online advertising platforms have commonly focused on maximizing total advertiser value (or welfare) to attract advertiser traffic and spend, and have resorted to machine learning predictions on advertiser values (also known as machine-learned advice) to improve ad auction designs and thus total welfare of advertisers. Yet, such improvements could come at the cost of individual bidders' welfare, consequently eroding fairness of ad platforms, and do not shed light on how particular advertiser bidding strategies impact individual fairness. Motivated by this, we present a novel fairness metric that measures an individual bidder's welfare loss, and also uncovers how advertiser strategies relate to such losses. Under this metric, we then study how ad platforms can utilize ML advice to improve welfare guarantees and fairness on the individual bidder level. We first motivate a simple approach that directly sets such ML advice as personalized reserve prices when the platform consists of \textit{autobidders} who maximize value while respecting a return-on-ad spent (ROAS) constraint. Under parallel VCG auctions with such ML advice-based reserves, we utilize our fairness metric to present a worst-case welfare lower-bound guarantee for an individual autobidder, and show that platform fairness is positively correlated with ML advice quality as well the scale of bids induced by the autobidder's bidding strategies. Further, we prove an impossibility result showing that no truthful, and possibly randomized mechanism with anonymous allocations can achieve universally better fairness guarantees than VCG, in presence of personalized reserves based on ML-advice of equal quality. Finally, we extend our fairness guarantees to generalized first price (GFP) and generalized second price (GSP) auctions.
翻译:在线广告平台通常侧重于最大限度地提高广告商总价值(或福利)以吸引广告商流量和支出,并采用广告商价值(又称机器学习建议)的机器学习预测来改进拍卖设计,从而改善广告商的总体福利。然而,这些改进可能以个体投标人的福利为代价,从而削弱广告平台的公平性,而不能揭示特定广告商投标战略如何影响个人公平。为此,我们提出了一个新的公平度量度,以衡量个体投标人的福利损失,并披露广告商战略如何与此类损失相关。在此度下,我们随后研究广告平台如何利用ML建议来改善个体投标人的福利保障和公平性。我们首先鼓励一种简单的方法,当平台由textit{utobirders}组成时,将ML建议直接设定为个性化储备价格,从而在尊重成本回报(ROAS)限制的同时,我们提出了一个新的公平性价价比我们基于一般价价的储备,我们利用我们的公平度指标来提出最差的福利低质量建议。