Gradient-based deep-learning algorithms exhibit remarkable performance in practice, but it is not well-understood why they are able to generalize despite having more parameters than training examples. It is believed that implicit bias is a key factor in their ability to generalize, and hence it has been widely studied in recent years. In this short survey, we explain the notion of implicit bias, review main results and discuss their implications.
翻译:基于渐进的深层学习算法在实践中表现出显著的绩效,但人们并不清楚为什么尽管有比培训实例更多的参数,它们还是能够一概而论。 人们相信,隐含的偏见是它们能够一概而论的一个关键因素,因此近年来已经对此进行了广泛研究。 在本次简短的调查中,我们解释了隐含的偏见的概念,审查了主要结果并讨论了其影响。