These are the notes for the lectures that I was giving during Fall 2020 at the Moscow Institute of Physics and Technology (MIPT) and at the Yandex School of Data Analysis (YSDA). The notes cover some aspects of initialization, loss landscape, generalization, and a neural tangent kernel theory. While many other topics (e.g. expressivity, a mean-field theory, a double descent phenomenon) are missing in the current version, we plan to add them in future revisions.
翻译:这些笔记是我在莫斯科物理与技术研究所(MIPT)和Yandex数据分析学院(YSDA)2020年秋天期间在莫斯科物理与技术研究所(MIPT)和Yandex数据分析学院(YSDA)举办的讲座的笔记。 这些笔记涵盖了初始化、损失地貌、概括化和神经相近内核理论的某些方面。 虽然目前版本中缺少许多其他主题(如表达性、平均场理论、双血缘现象 ), 我们计划在未来的修改中加入它们。