Machine Learning in Adaptive FETI-DP

Reducing the Effort in Sampling

Conference Paper (2021)
Author(s)

Alexander Heinlein (Universität zu Köln)

Axel Klawonn (Universität zu Köln)

Martin Lanser (Universität zu Köln)

Janine Weber (Universität zu Köln)

Affiliation
External organisation
DOI related publication
https://doi.org/10.1007/978-3-030-55874-1_58 Final published version
More Info
expand_more
Publication Year
2021
Language
English
Affiliation
External organisation
Pages (from-to)
593-603
Publisher
Springer
ISBN (print)
9783030558734
Event
European Conference on Numerical Mathematics and Advanced Applications, ENUMATH 2019 (2019-09-30 - 2019-10-04), Egmond aan Zee, Netherlands
Downloads counter
137

Abstract

The convergence rate of classic domain decomposition methods in general deteriorates severely for large discontinuities in the coefficient functions of the considered partial differential equation. To retain the robustness for such highly heterogeneous problems, the coarse space can be enriched by additional coarse basis functions. These can be obtained by solving local generalized eigenvalue problems on subdomain edges. In order to reduce the number of eigenvalue problems and thus the computational cost, we use a neural network to predict the geometric location of critical edges, i.e., edges where the eigenvalue problem is indispensable. As input data for the neural network, we use function evaluations of the coefficient function within the two subdomains adjacent to an edge. In the present article, we examine the effect of computing the input data only in a neighborhood of the edge, i.e., on slabs next to the edge. We show numerical results for both the training data as well as for a concrete test problem in form of a microsection subsection for linear elasticity problems. We observe that computing the sampling points only in one half or one quarter of each subdomain still provides robust algorithms.