Scientific and engineering problems often require the use of artificial intelligence to aid understanding and the search for promising designs. While Gaussian processes (GP) stand out as easy-to-use and interpretable learners, they have difficulties in accommodating big datasets, categorical inputs, and multiple responses, which has become a common challenge for a growing number of data-driven design applications. In this paper, we propose a GP model that utilizes latent variables and functions obtained through variational inference to address the aforementioned challenges simultaneously. The method is built upon the latent variable Gaussian process (LVGP) model where categorical factors are mapped into a continuous latent space to enable GP modeling of mixed-variable datasets. By extending variational inference to LVGP models, the large training dataset is replaced by a small set of inducing points to address the scalability issue. Output response vectors are represented by a linear combination of independent latent functions, forming a flexible kernel structure to handle multiple responses that might have distinct behaviors. Comparative studies demonstrate that the proposed method scales well for large datasets with over 10^4 data points, while outperforming state-of-the-art machine learning methods without requiring much hyperparameter tuning. In addition, an interpretable latent space is obtained to draw insights into the effect of categorical factors, such as those associated with building blocks of architectures and element choices in metamaterial and materials design. Our approach is demonstrated for machine learning of ternary oxide materials and topology optimization of a multiscale compliant mechanism with aperiodic microstructures and multiple materials.
翻译:科学和工程问题往往需要使用人工智能来帮助理解和寻找有希望的设计。虽然Gossian进程(GP)作为容易使用和可解释的学习者而突出,但它们难以容纳大数据集、绝对投入和多种反应,这已成为越来越多的数据驱动的设计应用的共同挑战。在本文件中,我们提议了一个GP模型,利用通过变式推断获得的潜在变量和功能,同时应对上述挑战。该方法建立在潜伏可变的Gausian选择(LVGP)模型上,其中将绝对因素映入连续的潜在空间,以便能够对混合可变数据集进行GP建模。通过向LVGP模型扩展变异性推断、绝对投入投入和多种反应,大型培训数据集被一套小的导出点取代,以解决可缩化问题。输出响应矢量由独立潜伏功能的线性组合代表,形成一种灵活的内核结构结构,处理多种反应,可能具有不同的行为。比较研究表明,拟议的方法在10-4级的大型数据集结构中,需要10-4级数据结构的大型数据集的模型,而不需要对机型结构进行最精确的系统化的模型进行深度解释。