Fully connected layers are a primary source of memory and computational overhead in deep neural networks due to their dense, often redundant parameterization. While various compression techniques exist, they frequently introduce complex engineering trade-offs or degrade model performance. We propose the Parametrized Random Projection (PRP) layer, a novel approach that decouples feature mixing from adaptation by utilizing a fixed random matrix modulated by lightweight, learnable element-wise parameters. This architecture drastically reduces the trainable parameter count to a linear scale while retaining reliable accuracy across various benchmarks. The design serves as a stable, computationally efficient solution for architectural scaling and deployment in resource-limited settings.
翻译:全连接层因其密集且通常冗余的参数化,成为深度神经网络中内存和计算开销的主要来源。尽管存在多种压缩技术,但这些方法常引入复杂的工程权衡或导致模型性能下降。我们提出参数化随机投影(PRP)层,这是一种新颖的方法,通过利用由轻量级可学习的元素级参数调制的固定随机矩阵,将特征混合与自适应过程解耦。该架构将可训练参数数量大幅减少至线性规模,同时在多种基准测试中保持可靠的准确性。该设计为资源受限环境中的架构扩展与部署提供了一种稳定且计算高效的解决方案。