A multilayer perceptron (MLP) is typically made of multiple fully connected layers with nonlinear activation functions. There have been several approaches to make them better (e.g. faster convergence, better convergence limit, etc.). But the researches lack in more structured ways to test them. We test different MLP architectures by carrying out the experiments on the age and gender datasets. We empirically show that by whitening inputs before every linear layer and adding skip connections, our proposed MLP architecture can result in better performance. Since the whitening process includes dropouts, it can also be used to approximate Bayesian inference. We have open sourced our code released models and docker images at https://github.com/tae898/age-gender/.
翻译:多层透视器(MLP)通常由与非线性激活功能完全相连的多个层组成,有几种方法可以使其更好(例如更快的趋同、更好的趋同限制等),但研究缺乏更有条理的测试方法。我们通过在年龄和性别数据集上进行实验来测试不同的多层透视器结构。我们从经验上表明,通过在每个线性层之前进行输入白化和增加跳过连接,我们提议的MLP结构可以产生更好的性能。由于白化过程包括了辍学者,它也可以用来接近Bayesian的推断。我们已经在 https://github.com/tae898/age-gender/ 上打开了我们的代码发布模型和 docker 图像的来源。