Soft-thresholding has been widely used in neural networks. Its basic network structure is a two-layer convolution neural network with soft-thresholding. Due to the network's nature of nonlinearity and nonconvexity, the training process heavily depends on an appropriate initialization of network parameters, resulting in the difficulty of obtaining a globally optimal solution. To address this issue, a convex dual network is designed here. We theoretically analyze the network convexity and numerically confirm that the strong duality holds. This conclusion is further verified in the linear fitting and denoising experiments. This work provides a new way to convexify soft-thresholding neural networks.
翻译:软阈值在神经网络中被广泛应用。基本的网络结构是一个双层卷积神经网络,其中采用了软阈值。由于网络具有非线性和非凸性的特性,训练过程非常依赖于网络参数的适当初始化,导致难以获得全局最优解。为了解决这个问题,本文设计了一个凸对偶网络。我们从理论上分析了网络的凸性,并通过线性拟合和去噪实验进行数值验证。该工作提供了一种将软阈值神经网络凸化的新方法。