Bottleneck problems are an important class of optimization problems that have recently gained increasing attention in the domain of machine learning and information theory. They are widely used in generative models, fair machine learning algorithms, design of privacy-assuring mechanisms, and appear as information-theoretic performance bounds in various multi-user communication problems. In this work, we propose a general family of optimization problems, termed as complexity-leakage-utility bottleneck (CLUB) model, which (i) provides a unified theoretical framework that generalizes most of the state-of-the-art literature for the information-theoretic privacy models, (ii) establishes a new interpretation of the popular generative and discriminative models, (iii) constructs new insights to the generative compression models, and (iv) can be used in the fair generative models. We first formulate the CLUB model as a complexity-constrained privacy-utility optimization problem. We then connect it with the closely related bottleneck problems, namely information bottleneck (IB), privacy funnel (PF), deterministic IB (DIB), conditional entropy bottleneck (CEB), and conditional PF (CPF). We show that the CLUB model generalizes all these problems as well as most other information-theoretic privacy models. Then, we construct the deep variational CLUB (DVCLUB) models by employing neural networks to parameterize variational approximations of the associated information quantities. Building upon these information quantities, we present unified objectives of the supervised and unsupervised DVCLUB models. Leveraging the DVCLUB model in an unsupervised setup, we then connect it with state-of-the-art generative models, such as variational auto-encoders (VAEs), generative adversarial networks (GANs), as well as the Wasserstein GAN (WGAN), Wasserstein auto-encoder (WAE), and adversarial auto-encoder (AAE) models through the optimal transport (OT) problem. We then show that the DVCLUB model can also be used in fair representation learning problems, where the goal is to mitigate the undesired bias during the training phase of a machine learning model. We conduct extensive quantitative experiments on colored-MNIST and CelebA datasets, with a public implementation available, to evaluate and analyze the CLUB model.
翻译:瓶装问题是一个重要的优化问题类别, 最近在机器学习和信息理论领域日益引起人们的关注。 它们被广泛用于基因模型、 公平的机器学习算法、 隐私保障机制的设计, 并被看成信息理论性能约束在多种多用户交流问题中。 在这项工作中, 我们提出一个总体的优化问题, 被称为复杂- 泄漏- 工具瓶颈( CLUB) 模式。 它 (一) 提供了一个统一的理论框架, 将大多数最新动态文献的变异性 用于信息- 感官隐私模型, (二) 建立对流行型变异性与歧视模式的新解释, (三) 构建基因化压缩模型, (四) 我们首先将CLUB模型设计成一个复杂- 隐私优化问题( 以复杂- 调频( IB) 的变异性( BIF), 不确定性 IB (DI B) 的变异性变异性( ), 最老的变异性ODLILIF 模式, 以我们的变异(CUDIF) 展示其他的变现。