The learning process of deep learning methods usually updates the model's parameters in multiple iterations. Each iteration can be viewed as the first-order approximation of Taylor's series expansion. The remainder, which consists of higher-order terms, is usually ignored in the learning process for simplicity. This learning scheme empowers various multimedia based applications, such as image retrieval, recommendation system, and video search. Generally, multimedia data (e.g., images) are semantics-rich and high-dimensional, hence the remainders of approximations are possibly non-zero. In this work, we consider the remainder to be informative and study how it affects the learning process. To this end, we propose a new learning approach, namely gradient adjustment learning (GAL), to leverage the knowledge learned from the past training iterations to adjust vanilla gradients, such that the remainders are minimized and the approximations are improved. The proposed GAL is model- and optimizer-agnostic, and is easy to adapt to the standard learning framework. It is evaluated on three tasks, i.e., image classification, object detection, and regression, with state-of-the-art models and optimizers. The experiments show that the proposed GAL consistently enhances the evaluated models, whereas the ablation studies validate various aspects of the proposed GAL. The code is available at \url{https://github.com/luoyan407/gradient_adjustment.git}.
翻译:深层次学习方法的学习过程通常在多个迭代中更新模型参数。 每种迭代可被视为泰勒系列扩展的第一阶近似值。 其余部分由更高层次的术语组成, 在简单化的学习过程中通常被忽略。 这个学习计划赋予各种多媒体应用程序, 如图像检索、 推荐系统和视频搜索等。 一般而言, 多媒体数据( 如图像) 是精度丰富和高维的, 因此近点的剩余部分可能不是零。 我们认为, 这项工作的剩余部分是信息化的, 并研究它如何影响学习进程。 为此, 我们提出一种新的学习方法, 即梯度调整学习( GAL), 以利用从以往培训迭代学学到的知识来调整香草梯度, 使其余部分最小化, 近似于改进。 拟议的 GAL 是模型和最优化, 容易适应标准学习框架。 我们评估了三项任务, 即图像分类、 对象检测和回归过程, 并持续地展示了GA- L 最佳模型和拟议的最佳模型。