We construct least squares formulations of PDEs with inhomogeneous essential boundary conditions, where boundary residuals are not measured in unpractical fractional Sobolev norms, but which formulations nevertheless are shown to yield a quasi-best approximations from the employed trial spaces. Dual norms do enter the least-squares functional, so that solving the least squares problem amounts to solving a saddle point or minimax problem. For finite element applications we construct uniformly stable finite element pairs, whereas for Machine Learning applications we employ adversarial networks.
翻译:暂无翻译