Shape-constrained symbolic regression (SCSR) allows to include prior knowledge into data-based modeling. This inclusion allows to ensure that certain expected behavior is better reflected by the resulting models. The expected behavior is defined via constraints, which refer to the function form e.g. monotonicity, concavity, convexity or the models image boundaries. In addition to the advantage of obtaining more robust and reliable models due to defining constraints over the functions shape, the use of SCSR allows to find models which are more robust to noise and have a better extrapolation behavior. This paper presents a mutlicriterial approach to minimize the approximation error as well as the constraint violations. Explicitly the two algorithms NSGA-II and NSGA-III are implemented and compared against each other in terms of model quality and runtime. Both algorithms are capable of dealing with multiple objectives, whereas NSGA-II is a well established multi-objective approach performing well on instances with up-to 3 objectives. NSGA-III is an extension of the NSGA-II algorithm and was developed to handle problems with "many" objectives (more than 3 objectives). Both algorithms are executed on a selected set of benchmark instances from physics textbooks. The results indicate that both algorithms are able to find largely feasible solutions and NSGA-III provides slight improvements in terms of model quality. Moreover, an improvement in runtime can be observed using the many-objective approach.
翻译:受限制的符号回归(SCSR)允许将先前的知识纳入基于数据的模型中。 包含此功能可以确保某些预期行为得到更好的反映, 并通过制约来界定预期行为, 这些制约是指功能形式, 如单调、 调和、 调和或模型图像界限; 除了由于界定功能形状的制约而获得更稳健和可靠的模型的优势外, 使用 SSR 还可以找到对噪音更强、更具有更好的推断行为的模型。 本文展示了将近似错误和限制违反行为降至最低的模范标准化方法。 明显地实施了两种算法 NSGA- II 和 NSGA- III, 在模型质量和运行时间方面相互比较。 两种算法都能够应对多重目标, 而NSGA- II 的使用非常成熟的多的多目标。 NSGA- III 的算法是用来处理“ many- many ” 目标的“ many many ” 和“ many realal commal ” lavelrial real real ress sup lavels sup labal lactions sups sup labs sals sals sup laveals