Deep learning models are trained and deployed in multiple domains. Increasing usage of deep learning models alarms the usage of memory consumed while computation by deep learning models. Existing approaches for reducing memory consumption like model compression, hardware changes are specific. We propose a generic analysis of memory consumption while training deep learning models in comparison with hyperparameters used for training. Hyperparameters which includes the learning rate, batchsize, number of hidden layers and depth of layers decide the model performance, accuracy of the model. We assume the optimizers and type of hidden layers as a known values. The change in hyperparamaters and the number of hidden layers are the variables considered in this proposed approach. For better understanding of the computation cost, this proposed analysis studies the change in memory consumption with respect to hyperparameters as main focus. This results in general analysis of memory consumption changes during training when set of hyperparameters are altered.
翻译:深层学习模型在多个领域得到培训和部署。越来越多的深层学习模型的使用提醒人们在深层学习模型的计算过程中使用内存消耗。现有的减少内存消耗的方法,如模型压缩,硬件变化是具体的。我们提议对内存消耗进行一般性分析,同时与用于培训的超参数相比,对深层学习模型进行培训,包括学习率、批量、隐藏层数和层深层数在内的超参数决定模型的性能、准确性。我们假设隐藏层的优化和类型为已知值。超光谱器的变化和隐藏层的数量是这一拟议方法中考虑的变量。为了更好地了解计算成本,本拟议分析研究与超光谱仪有关的内存消耗变化,将其作为主要重点。这导致在调整多参数时对培训期间的内存消耗变化进行总体分析。