We present a data-driven algorithm that advertisers can use to automate their digital ad-campaigns at online publishers. The algorithm enables the advertiser to search across available target audiences and ad-media to find the best possible combination for its campaign via online experimentation. The problem of finding the best audience-ad combination is complicated by a number of distinctive challenges, including (a) a need for active exploration to resolve prior uncertainty and to speed the search for profitable combinations, (b) many combinations to choose from, giving rise to high-dimensional search formulations, and (c) very low success probabilities, typically just a fraction of one percent. Our algorithm (designated LRDL, an acronym for Logistic Regression with Debiased Lasso) addresses these challenges by combining four elements: a multiarmed bandit framework for active exploration; a Lasso penalty function to handle high dimensionality; an inbuilt debiasing kernel that handles the regularization bias induced by the Lasso; and a semi-parametric regression model for outcomes that promotes cross-learning across arms. The algorithm is implemented as a Thompson Sampler, and to the best of our knowledge, it is the first that can practically address all of the challenges above. Simulations with real and synthetic data show the method is effective and document its superior performance against several benchmarks from the recent high-dimensional bandit literature.
翻译:我们展示了一种数据驱动算法,广告商可以在网上出版商上将其数字广告活动自动化。该算法使广告商能够在现有目标受众和媒体之间搜索,通过在线实验找到最佳组合。找到最佳受众-受众组合的问题因若干特殊挑战而变得复杂,包括:(a) 需要积极探索,以解决先前的不确定性,并加快寻找有利可图的组合;(b) 许多组合可以选择,产生高维搜索配方,以及(c) 成功概率非常低,通常只占1%。我们的算法(名为LRDL,即物流回流缩缩缩略词,与贬低的Lasso一起)通过结合四个要素来应对这些挑战:积极探索的多臂条纹框架;处理高维度的拉索惩罚功能;处理由Lasso引起的正规化偏差的内嵌入式内嵌入骨架;以及(c)促进跨武器交叉学习结果的半维度回归模型。我们算法的模型(称为LRDRDL,即“LRD”缩略图)通过四个要素来应对这些挑战,并展示我们最佳的合成文件的高级性标准。它,其高端标准是其高水平,从高端文件,从高层次中可以显示最佳的成绩,从高层次上展示最佳的成绩。