Skip to main content
Erschienen in: Computational Mechanics 4/2023

Open Access 27.12.2022 | Original Paper

Numerical upscaling of parametric microstructures in a possibilistic uncertainty framework with tensor trains

verfasst von: Martin Eigel, Robert Gruhlke, Dieter Moser, Lars Grasedyck

Erschienen in: Computational Mechanics | Ausgabe 4/2023

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

A fuzzy arithmetic framework for the efficient possibilistic propagation of shape uncertainties based on a novel fuzzy edge detection method is introduced. The shape uncertainties stem from a blurred image that encodes the distribution of two phases in a composite material. The proposed framework employs computational homogenisation to upscale the shape uncertainty to a effective material with fuzzy material properties. For this, many samples of a linear elasticity problem have to be computed, which is significantly sped up by a highly accurate low-rank tensor surrogate. To ensure the continuity of the underlying mapping from shape parametrisation to the upscaled material behaviour, a diffeomorphism is constructed by generating an appropriate family of meshes via transformation of a reference mesh. The shape uncertainty is then propagated to measure the distance of the upscaled material to the isotropic and orthotropic material class. Finally, the fuzzy effective material is used to compute bounds for the average displacement of a non-homogenized material with uncertain star-shaped inclusion shapes.
Hinweise
Robert Gruhlke and Dieter Moser have contributed equally to this work.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

1 Introduction

Composite materials are ubiquitous in many applications. Whether they are formed by chance (like with unfavourable impurities in metal) or by design (like pebbles in concrete), the mechanical properties of the whole material are determined by the resulting composite structure [1]. The different types of composite materials—e.g. metal matrix composites [2], fiber-reinforced polymers [3], composite wood [4] and other advanced composite materials [5]—find their use in a wide range of applications, e.g. masonry, aerospace industry, wind power plants and sports equipment. In many applications, the added composite material improves a base matrix material in terms of wear resistance, damping properties and mechanical strength, while keeping the same weight. However, such improved composite materials are a result of empirical studies, experimentation and chance. It hence is obvious that a thorough understanding of composite materials via theoretical and accurate computational models is desirable to predict and systematically improve the properties and applicability of composites.
What makes this task challenging are the influence of material behaviour on the micro- and macro scales, as well as the multifaceted sources of uncertainties. For instance, the involved length scales may span up to ten orders of magnitudes, i.e. the size of a nano particle of \(10^{-6}\,\textrm{m}\) is embedded in a material with a length scale of \(10^{-1}{-}10^{1}\,\textrm{m}\) [2]. The same holds for the material constants if the constituents are fundamentally different as in metals and polymers. A standard finite element approach becomes very costly in such a setting since the details of the composite have to be resolved by the mesh for accurate simulations. Additionally, one has to handle the uncertainty of the material constants such as the position, form and size of the inclusions. This increases the costs even further since many realisations are required to correctly capture the uncertainty statistically. Each new realisations requires a costly (possible automatic) remeshing of the computational domain. The mapping from geometric parameters—namely position and shape of each inclusion—onto the generated meshes are usually discontinuous. This introduces unwanted effects that we tackle in Sect. 4.1.
The first challenge—the span of scales—is tackled by homogenisation methods, a term introduced by Babuška in [6]. In some form or another, these methods incorporate micro-scale behaviour into an adjusted macro-scale model, replacing the full composite material model by a corrected homogeneous one, obtained by an asymptotic limit of an assumed (periodic) domain. The central idea is to derive equations that describe the effective material properties analytically. Alternatively, computational approaches have been devised to solve particular micro-scale problems to deduce the adjusted macro-scale behaviour numerically [7, 8]. Some notable examples are stochastic homogenisation, see [912], projection based homogenisation [13] and (stochastic) representative volume element methods [14].
The second challenge—the uncertainty of the material—is sometimes neglected [15, 16] by only considering deterministic material properties. If the probability distribution of the material constants is known, the uncertainty can be modelled with precise probabilities [17]. Ignoring the uncertainty seems valid if it has little influence on the system’s response or the model data is known sufficiently accurate and is free of inherent fluctuations. Using precise probabilities is valid if the distributions of the material constants are known precisely. However, as was pointed out by Motamed and Babuška [18], stochastic models based on precise probabilities are not always able to model the uncertainty in composite materials. Instead, they propose a model based on an imprecise probability theory. Examples for this are evidence theory [19], random set theory [20], possibility theory [21] and—more recently—optimal Uncertainty Quantification (UQ) [22]. In comparison to precise probability, imprecise probability methods are able to provide estimates of uncertainties based only on a small set of data and few assumptions, which would be far too limited for a probabilistic method. For an illuminating work that dissects the difference between precise and imprecise probabilities, we refer the reader to [23]. As a motivation the approach developed in what follows, we especially point to the false confidence theorem, which qualitatively states that “probability dilution is a symptom of a fundamental deficiency in probabilistic representations of statistical inference, in which there are propositions that will consistently be assigned a high degree of belief, regardless of whether or not they are true” [23].
In this work, we consider a practical problem that is supposed to illustrate when imprecise probabilities are appropriate (in particular more so than probabilistic methods), the challenges performing a non-probabilistic uncertainty propagation and quantification, and ways to overcome the computational difficulties. We assume a material consisting of two phases, namely a soft matrix phase and a hard inclusion phase. Both phases are linear elastic with precisely known Young’s modulus and Poisson ratio. The inclusions repeat periodically in a checkerboard fashion. Consequently, a classical numerical homogenisation method would yield a homogenised (globally constant) macro-scale material. Considering real applications, the main problem stems from the unknown shape of the inclusions. Often, the shape is retrieved with a non-intrusive imaging process, e.g. computed tomography scans or magnetic resonance imaging. To illustrate this, Fig. 1 shows a computed tomography scan of an imperfect adhesive bond within a Henkel beam, cf. [24]. Apparently, the image is noisy, blurred, pixelated and exhibits artefacts. While it is possible to de-noise and de-blurr the image with image post-processing, which might make the identification of inclusions possible, different numerical methods yield different shapes [25]. This reveals the shape uncertainty inherent to the image. Here, we just focus on one blurred image per inclusion, assuming that the artefacts and noise were successfully removed already, see Fig. 3 for an example. Opposite to this, in the case of multiple CT scans of the same inclusion, the additional information can be employed to reduce the shape uncertainty and thus refine the epistemic uncertainty model until the uncertainty becomes irreducible and hereby aleatoric uncertainty models are preferable.
To tackle a problem in this setting, we introduce a novel fuzzy edge detection based on possibility theory, presented in Sect. 2.1 and a restriction of possible interface boundaries via bounded total variation in Sect. 3. This yields a fuzzy model of the boundary which in turn introduces a computational challenge in terms of a numerical homogenisation problem as discussed in Sect. 2.2. For this, the micro-model has to be solved very often in order to propagate the uncertainty of the boundary to the homogenised material model. To alleviate this expensive task, we introduce a highly accurate rank adaptive low-rank tensor surrogate in Sect. 4. In the numerical experiments Sect. 5, the surrogate model is validated numerically. Moreover, we measure the distance of the homogenised material to the class of isotropic and orthotropic elastic tensors and eventually use the homogenised material to perform a worst/best case analysis for a full matrix composite model with 64 inclusions, i.e., with help of the homogenised material we will find bounds for the average displacement of the non-homogenised material.

2 Basics

This section serves as a brief introduction of three fundamental topics that form the basis of this work. First, fuzzy set theory which is used to model the uncertainty ingrained in the blurred image. Second, the computational up-scaling method (numerical homogenisation). Third, classes of constitutive tensors to measure the distance of the up-scaled material to the isotropic and orthotropic material are discussed.

2.1 Fuzzy set theory

We introduce a possibilistic framework with fuzzy sets based on 4 central definitions, see e.g. [26] or more recently [2729].
Fuzzy sets are common sets that are equipped with a membership function which assigns each element in the set a value between zero and one. This values only task is to communicate a degree of belongingness to the set. The meaning of this value depends on the problem and—more importantly—on the community using this value to formalise uncertainty. Arguably, this ambiguity is a feature and not a short-coming of this theory since it avoids assumptions that are usually made by other approaches. In probability theory for instance, one has to assume a prior distribution before updating the posterior distribution with new samples. The following definition introduces the terminology of fuzzy sets.
Definition 1
(Fuzzy set/variable, \(\alpha \)-cuts and interactivity) Let \(Z\ne \emptyset \) be a set and \(\mu :Z\rightarrow [0,1]\) be a map such that there exists \(z\in Z\) with \(\mu (z)=1\). The map \(\mu \) is called (normalised) membership function. We define a (normalised) fuzzy set \(\tilde{z}\) on Z by
$$\begin{aligned} \tilde{z}:=\{ (z,\mu (z))\,\mid \, z\in Z \}. \end{aligned}$$
(1)
If \(\mu (Z) =\{0,1\}\) with unique \(z^*\in Z\) and \(\mu (z^*)=1\) then \(\tilde{z}\) is called a crisp set. Furthermore, we denote by \(\mathcal {F}(Z)\) the set of all fuzzy sets on Z. We thus simply write \(\tilde{z}\in \mathcal {F}(Z)\). For each fuzzy set \(\tilde{z}\in \mathcal {F}(Z)\) we denote the associated membership function by \(\mu _{\tilde{z}}\). If \(Z\subset \mathbb {R}^N\) for \(N\in \mathbb {N}\), we call \(\tilde{z}\in \mathcal {F}(Z)\) a fuzzy variable \((N=1)\) or vectorial fuzzy variable \((N>1)\) described by a (joint) membership function \(\mu _{\tilde{z}}\). The support of \(\mu _{\tilde{z}}\) is defined as
$$\begin{aligned} \textrm{supp}({\tilde{z}}) = \{ z \in Z \mid \mu _{{\tilde{z}}}(z) > 0 \}. \end{aligned}$$
For \(\alpha \in [0,1]\), the \(\alpha \)-cut \(C_\alpha \) of \(\mu _{\tilde{z}}\) is defined as
$$\begin{aligned} C_\alpha [\tilde{z}]:= {\left\{ \begin{array}{ll} \{z\in Z: \mu _{\tilde{z}}(z)\ge \alpha \}, &{}\quad \text {for } \alpha >0, \\ \overline{{\text {supp}}(\tilde{z})}, &{}\quad \text {for } \alpha = 0. \end{array}\right. } \end{aligned}$$
(2)
Let \(\tilde{z}_i\in \mathcal {F}(Z_i)\) for sets \(Z_i\) with \(i=1,\ldots ,M<\infty \), https://static-content.springer.com/image/art%3A10.1007%2Fs00466-022-02261-z/MediaObjects/466_2022_2261_IEq32_HTML.gif and \(\tilde{z}=(\tilde{z}_1,\ldots , \tilde{z}_M)\). If the joint membership function associated with \(\tilde{z}\) has the form \(\mu _{\tilde{z}}=\min _i \mu _{\tilde{z}_i}\) then \(\tilde{z}\) is called non-interactive and interactive otherwise.
Fuzzy sets are quite versatile since there are no restrictions on the (type and structure of the) used sets. However, for numerical methods to become efficient, certain assumptions are beneficial. The first restriction mimics numbers and vectors.
Definition 2
(Fuzzy number/vector) Let \(\tilde{z}\in \mathcal {F}(Z)\) with \(Z\subset \mathbb {R}^N\) for some \(N\in \mathbb {N}\) such that Z is bounded and convex and the (joint) membership function \(\mu _{\tilde{z}}\) is upper semi-continuous, i.e.
$$\begin{aligned} \limsup \limits _{z\rightarrow z_0}\mu (z)\le \mu (z_0),\quad \forall z_0\in Z \end{aligned}$$
(3)
and quasi-concave, i.e.
$$\begin{aligned} \mu (\lambda z_1+(1-\lambda )z_2)\ge \min (\mu (z_1),\mu (z_2)), \quad \forall z_1,z_2 \in Z. \end{aligned}$$
(4)
If there exists a unique \(z^*\in Z\) such that \(\mu _{\tilde{z}}(z^*)=1\) then we call \(\tilde{z}\) a fuzzy number for \(n=1\) and a fuzzy vector for \(n>1\).
This notion is easily extended to intervals and domains.
Definition 3
(Fuzzy interval/domain) With the same assumptions as in Definition 2, there exists a subset \(S\subset Z\) such that \(\mu _{\tilde{z}}(z) = 1\) for all \(z\in S\). Then \(\tilde{z}\) is called a fuzzy interval for \(n=1\) and fuzzy domain for \(n>1\) .
Note that the quasi-concavity of the membership function implies convexity of any \(\alpha \)-cut \(C_\alpha \). In particular, given \(S\subset Z\) as in Definition 3, the convex hull \({\text {conv}}(S)\) is a proper subset of \(C_1\), too. The quasi-concavity also leads to the nestedness property of \(\alpha \)-cuts from Definitions 2 and 3, i.e.
$$\begin{aligned} C_\alpha [\tilde{z}] \subset C_\beta [\tilde{z}],\quad \forall \alpha \ge \beta . \end{aligned}$$
(5)
This property is essential for the \(\alpha \)-cut propagation method, see Theorem 2. In the following example, we introduce the most common fuzzy structures. Namely, the fuzzy trapezoidal interval which we will use throughout this work.
Example 1
A particular class of fuzzy sets is the trapezoidal fuzzy set \(\tilde{z}\), forming a subset of \(\mathcal {F}(Z)\) with \(Z\subset \mathbb {R}\). The respective membership function \(\mu _{\tilde{z}}\) is described by two upper semi-continuous functions \(f_L:(-\infty ,0]\rightarrow [0,1]\) and \(f_R:[0,\infty )\rightarrow [0,1]\). Here, \(f_L(0) = f_R(0)=1\) with \(f_L\) \(( f_R )\) is monotonously increasing (decreasing) and \(\lim _{z\rightarrow -\infty }f_L(z)=0\) \(( \lim _{z\rightarrow \infty }f_R(z)=0 )\) such that there exist \(\ell ^*,r^*\in Z\) with \(z_\ell ^*\le z_r^*\) and
$$\begin{aligned} \mu _{\tilde{z}}(z) = \left\{ \begin{array}{ll} f_L(z^*-z),&{} z \le z_\ell ^*, \\ 1,&{} z \in (z_\ell ^*,z_r^*),\\ f_R(z-z^*),&{} z \ge r^*, \end{array} \right. \end{aligned}$$
(6)
The triangle fuzzy number \(\tilde{z}=\langle \ell ,z^*,r \rangle \) specified by left and right limit \(\ell ,r\) and peak position \(z^*= z_\ell ^*= z_r^*\) is a special case of the trapezoidal fuzzy set. Figure 2 depicts the propagation of a trapezoidal fuzzy set.
For arbitrary fuzzy sets Zadeh’s extension principle is the way to propagate uncertainty through a mapping.
Theorem 1
(Zadeh’s extension principle [26])
Consider a function \(f:Z \rightarrow V\) with a non-empty set V. Let \(\tilde{z}\in \mathcal {F}(Z)\) with membership function \(\mu _{\tilde{z}}\). Define
$$\begin{aligned} \tilde{f} := f(\tilde{z}):=\{ (f(z), \mu _{\tilde{f}}(f(z))\in V\times [0,1], z\in Z\} \end{aligned}$$
(7)
with membership function \(\mu _{\tilde{f}}\) defined for all \( v\in V \) as
$$\begin{aligned} \mu _{\tilde{f}}(v):= {\left\{ \begin{array}{ll} \sup \limits _{z\in f^{-1}(v)} \mu _{\tilde{z}}(z) &{}\text {if } f^{-1}(v)\ne \emptyset , \\ \quad 0 &{}\text {if } f^{-1}(v)=\emptyset . \end{array}\right. } \end{aligned}$$
(8)
Then \(\tilde{f}\in \mathcal {F}(V)\) with membership function \(\mu _{\tilde{f}}\).
If more underlying structure is given, the extension principle can be formulated equivalently in terms of constrained optimization. Zadeh’s principle can be reformulated into the so-called \(\alpha \)-cut propagation for fuzzy vectors and fuzzy domains to reduce the computational costs.
Theorem 2
(\(\alpha \)-cut propagation [30])
Let \(f:Z\rightarrow V\) be continuous between metric spaces \((Z,d_1)\) and \((V,d_2)\) and let \(\tilde{z}\in \mathcal {F}(Z)\) with \(C_0[\tilde{z}]\subset K\subset Z\) for a compact set K with convex Z. Furthermore, let the membership function \(\mu _{\tilde{z}}\) be quasi-concave and upper semi-continuous. Then \(\mu _{\tilde{f}}\) can be characterised via \(\alpha \)-cuts as
$$\begin{aligned} C_\alpha [{\tilde{f}}] = f(C_\alpha [\tilde{z}]). \end{aligned}$$
(9)
Moreover, if \(V=\mathbb {R}\) then
$$\begin{aligned} C_\alpha [{\tilde{f}}] = \left[ \min \limits _{z\in C_\alpha [\tilde{z}]}f(z), \max \limits _{z\in C_\alpha [\tilde{z}]}f(z)\right] . \end{aligned}$$
(10)
Note that (9) follows from the assumption that \(C_\alpha [\tilde{z}]\) is closed and consequently compact as it is a closed subset of a compact set K in a metric space. Since f is assumed to be continuous, compact sets are mapped to compact sets.
Comparing Zadeh’s principle with the \(\alpha \)-cut propagation, we immediately see that the former requires the inverse image and an optimization step for each point v whereas the later only depends on two optimization steps of f for the number of \(\alpha \)-cuts. Since it is infeasible to perform Zadeh’s principle for all points \(v\in V\), it is combined or replaced with a sampling approach. We distinguish two variants to sample in V:
  • Semi sampling approach: directly solve the constrained optimization problem with a global optimiser. For a given sequence \((v_k)_k\subset V\), compute the supremum over \(Z_k:=\{z\in Z: f(z) = v_k\}\), see the red line in Fig. 2.
  • Full sampling approach: choose a sequence \((z_k)_k\subset Z\), compute \((f_k)_k = [f(z_k)]_k\) and \((\mu _k)_k = [\mu _{\tilde{z}}(z_k)]_k\). Use the data sample pairs \((v_k, \mu _k)\) and reconstruct \(\mu _{\tilde{f}}\), e.g. by convex hull or an envelope approach, see the orange/purple graphics in Fig. 2.
If f has high evaluation costs, the propagation inevitably becomes difficult and costly. This for example is the case if f represents a finite element solution with a large number of elements. Thus, to still render the propagation feasible the number of evaluations of f needs to be computationally feasible. In case of applicability of \(\alpha \)-cut propagation, the number of evaluations depends on the (global) optimization method. This number is possibly lower than the sampling counterparts. However, if the requirements of Theorem 2 are not met one may fall back to the sampling methods. By introducing a surrogate model \(f_h \approx f\), the computational costs can be further reduced for both strategies. In essence, this approach can be interpreted as a reduction of evaluations of f, determined by the number of evaluations needed to obtain \(f_h\), assuming that the evaluation costs of \(f_h\) are negligible.

2.2 Upscaling of heterogeneous linear elastic material

In the following we develop a numerical homogenisation method which yields a macroscopic material based on microscopic properties. The goal of the approach is to dispose of the computationally involved microstructure and construct an upscaled material surrogate with similar homogenised behaviour on a larger scale. Equivalent terms for “homogenised behaviour” are macroscopic, effective or upscaled behaviour [31], where the examined “behaviour” is subject to some quantity of interest (e.g. average displacement or stress) based on the system response.
In classical (asymptotic) homogenisation, an effective property is calculated based on the assumption of an infinite periodic domain [32]. The local microscopic structure is defined in terms of a representative volume element, which in our setting would consist of a single inclusion that has identical shape in the entire domain. For problems in non-periodic media, the methods of numerical homogenisation or numerical upscaling as e.g. described in [33] are used, where local boundary value problems are solved to calculate effective characteristics in each local domain [34], also see [9, 12, 35] for analytical stochastic homogenisation. Periodic boundary conditions (PBCs) are commonly used for numerical upscaling methods of matrix composite material [36, 37]. In this work, we focus on numerical homogenisation with PBCs for linear elasticity. Let \(D=[-1,1]^2\) be a unit cell domain and let the heterogeneous material law be encoded in a tensor \(\varvec{C}=\varvec{C}(x)\in \mathbb {R}^{2,2,2,2}, x\in D\). For given macroscopic strain \(\varvec{E}\) write the displacement \(\varvec{u}\) as \(\varvec{u}(x)=\varvec{E}\cdot x + \varvec{v}\) where the D-periodic fluctuation \(\varvec{v}\) solves
$$\begin{aligned} \left\{ \begin{array}{ll} {\text {div}}\varvec{\sigma } = 0 &{}\text {in } D, \\ \varvec{\sigma }= \varvec{C} :\varvec{\epsilon } &{} \text {in } D,\\ \varvec{\epsilon } = \varvec{E} + [\nabla \varvec{v}+\nabla ^T\varvec{v}]/2 &{} \text {in } D, \end{array} \right. \end{aligned}$$
(11)
such that \(\varvec{\sigma }\cdot \textbf{n}\) is antiperiodic on D with \(\varvec{n}\) denoting the outer normal with respect to \(\partial D\). Let \(\langle \varvec{\epsilon }\rangle \) and \(\langle \varvec{\sigma }\rangle \) denote the average strain and stress of the computed strain \(\varvec{\epsilon }\) and stress \(\varvec{\sigma }\). Then, there exists a tensor \(\mathcal {H}\in \mathbb {R}^{2,2,2,2}\) satisfying
$$\begin{aligned} \langle \varvec{\sigma }\rangle = \mathcal {H}\ :\langle \varvec{\epsilon }\rangle . \end{aligned}$$
(12)
The tensor \(\mathcal {H}\) is called the effective or upscaled (macroscopic) tensor, representing the elastic moduli of the homogenised medium. By construction it holds that \(\varvec{E} = \langle \varvec{\epsilon }\rangle \). Consequently, \(\mathcal {H}\) can be obtained from (12) by choosing macro strains \(\varvec{E}\) corresponding to different elementary load cases and subsequently computing \(\langle \varvec{\sigma }\rangle \) by solving (11). In what follows, the homogenisation technique will be applied for every shape parametrisation encoded in \(\varvec{C}\).

2.3 Measuring the distance between constitutive tensors

Constitutive tensors \(\varvec{C}=(C_{ijkl})\in \mathbb {R}^{d,d,d,d}\) in the planar case of \(d=2\) determine the behaviour of the linear elastic material by relating stress and strain. Such a tensor is said to be in \(\mathbb {E}\textrm{la}(d)\) if the symmetry property
$$\begin{aligned} C_{ijkl} = C_{jikl} = C_{ijlk} = C_{klij} \end{aligned}$$
holds. A tensor \(\varvec{C}\in \mathbb {E}\textrm{la}(d)\) may exhibit further symmetry properties. If there exist Lamé constants \(\lambda , \mu \in \mathbb {R}\) such that
$$\begin{aligned}&\varvec{C}_{\textrm{iso}} = \lambda \textbf{1}\otimes \textbf{1} + 2 \mu \varvec{I}, \quad {\varvec{I}}_{ijk\ell } = \frac{1}{2}\left( \delta _{ik}\delta _{j\ell } + \delta _{i\ell }\delta _{jk} \right) ,\\&\varvec{1}_{ij} = \delta _{ij}, \end{aligned}$$
the tensor is called isotropic. For orthotropic tensors, we define a rotation matrix
$$\begin{aligned} g(\theta ) = \begin{pmatrix} \cos (\theta ) &{} - \sin (\theta ) \\ \sin (\theta ) &{} \cos (\theta ) \end{pmatrix}, \quad \theta \in [0,2\pi ] \end{aligned}$$
and the full orthogonal group operation o(g) via
$$\begin{aligned} (o(g)\varvec{C})_{ijk\ell } = g_{ip}g_{jq}g_{kr}g_{\ell s} {C}_{pqrs}, \quad \varvec{C}\in \mathbb {E}\textrm{la}(d). \end{aligned}$$
We then say that \(\varvec{C}_{\text {ortho}}\in \mathbb {E}\textrm{la}(d)\) is orthotropic if it can be represented as
$$\begin{aligned} \varvec{C}_{\text {ortho}} = \varvec{C}_{\text {iso}} + {\varvec{T}}, \quad {\varvec{T}} = \rho (g(\theta )) \left( \Lambda _1 {\varvec{O}^1} + \Lambda _2 {\varvec{O}^2} \right) , \end{aligned}$$
(13)
for some isotropic \(\varvec{C}_{\text {iso}} \in \mathbb {E}\textrm{la}(d)\), \(\Lambda _1,\Lambda _2\in \mathbb {R}\), \(\theta \in [0,2\pi ]\) and forth order tensors \(\varvec{O}^1\) and \(\varvec{O}^2\) with a Kelvin-Mandel matrix representation [38] given as
$$\begin{aligned} \frac{1}{\sqrt{2}} \left[ \begin{array}{rrr} 1/2 &{} -1/2&{} 0 \\ -1/2 &{} 1/2 &{} 0 \\ 0 &{} 0 &{} -1 \end{array}\right] \qquad \text {and} \qquad \frac{1}{\sqrt{2}} \left[ \begin{array}{rrr} 1 &{} 0 &{} 0 \\ 0 &{} -1 &{} 0 \\ 0 &{} 0 &{} 0 \end{array}\right] . \end{aligned}$$
Note that any isotropic \(\varvec{C}_{\text {iso}}\in \mathbb {E}\textrm{la}(d)\) is invariant under action of \(\rho (g)\), in particular \( \rho (g)\varvec{C}_{\text {iso}} = \varvec{C}_{\text {iso}}\). With that representation, any orthotropic tensor has 5 degrees of freedom \((\lambda , \mu , \Lambda _1, \Lambda _2, \theta )\). The orthotropic tensors represented in normal form have only 4 degrees of freedom since \(\theta \) is chosen as zero. Let \(\mathbb {E}\textrm{la}(d,\text {iso}), \mathbb {E}\textrm{la}(d,\text {ortho})\subset \mathbb {E}\textrm{la}(d)\) be the set of all isotropic and orthotropic tensors, respectively. Given any anisotropic tensor \(\varvec{C}\), we define the distance to the symmetry class of isotropic and orthotropic tensors by
$$\begin{aligned} d_{\text {iso}}(\varvec{C})= & {} \min \limits _{\varvec{X}\in \mathbb {E}\textrm{la}(d,\text {iso})}\Vert \varvec{C}-\varvec{X}\Vert ,\nonumber \\ d_{\text {ortho}}(\varvec{C})= & {} \min \limits _{\varvec{X}\in \mathbb {E}\textrm{la}(d,\text {ortho})}\Vert \varvec{C}-\varvec{X}\Vert \end{aligned}$$
(14)
with the Frobenius norm \(\Vert \cdot \Vert \). The distances to the symmetry classes can be characterised as follows.
Proposition 3
(Distance to isotropy class [39])
Let \(\varvec{C}\in \mathbb {E}\textrm{la}(d)\) for \(d=2\). Define \(\varvec{C}_{\text {iso}} = 2\mu \varvec{I}+\lambda \varvec{1}\otimes \varvec{1}\) with
$$\begin{aligned} \lambda&:= \frac{1}{8}(C_{1111} + 6C_{1122} - 4C_{1212} + C_{2222})\\ \mu&:= \frac{1}{8}(C_{1111} - 2C_{1122} + 4C_{1212} + C_{2222}) \end{aligned}$$
as the orthogonal projection of \(\varvec{C}\) onto the isotropic material class. Then,
$$\begin{aligned} d_{\text {iso}}(\varvec{C}) = \Vert \varvec{C} - \varvec{C}_{\text {iso}}\Vert . \end{aligned}$$
(15)
Proposition 4
(Distance to orthotropy class [39])
Let \(\varvec{C}\in \mathbb {E}\textrm{la}(d)\) for \(d=2\) and \((\theta _k)_{k=1}^K\), \(K\in \mathbb {N}\) be the finite roots of
$$\begin{aligned} a\cos 8\theta + b \sin 8\theta + c\cos 4\theta + d\sin 4\theta = 0,\quad \theta \in [0,2\pi ] \end{aligned}$$
with \(a = 4X_1Y_1\), \(b=2(Y_1^2-X_1^2)\), \(c=2X_2Y_2\) and \(d=Y_2^2-X_2^2\) and
$$\begin{aligned} X_1&= \frac{1}{2\sqrt{2}}(C_{1111} - 2C_{1122}-4C_{1212} + C_{2222}), \\ Y_1&= \sqrt{2}(C_{1112}-C_{2212}),\\ X_2&= \frac{1}{\sqrt{2}}(C_{1111}-C_{2222}),\\ Y_2&= \sqrt{2}(C_{1112} + C_{2212}). \end{aligned}$$
For \(\theta \in [0,2\pi ]\), define
$$\begin{aligned}&\Lambda _1(\theta ) := X_1\cos 4\theta + Y_1\sin 4\theta ,\\&\Lambda _2(\theta ) := X_2\cos 2\theta + Y_2\sin 2\theta \end{aligned}$$
and
$$\begin{aligned} \varvec{X}(\theta _k):= & {} \rho (g(\theta _k))(\Lambda _1(\theta _k) \varvec{E}^1 + \Lambda _2(\theta _k) \varvec{E}^2) \\{} & {} \quad \text {for}\quad k=1,\ldots , K. \end{aligned}$$
Then,
$$\begin{aligned} d_{\text {ortho}}(\varvec{C}) = \min \limits _{k=1,\ldots ,K} \Vert \varvec{C} - \varvec{X}(\theta _k)\Vert . \end{aligned}$$
(16)
With the introduced material class concepts of constitutive tensors we are able to measure the distance of our effective material to the isotropic respectively orthotropic material class for each shape parametrisation.

3 Fuzzy edge detection

Given a blurred image as depicted in Fig. 3, the most common approach to reconstruct the original image is to use some method from the wide class of blind deconvolution methods [40]. These methods assume that an image y is the convolution of an original image x and some kernel k, distorted with additive noise n, i.e.,
$$\begin{aligned} y=x\star k + n. \end{aligned}$$
(17)
Finding a pair (xk) satisfying equation (17) is equivalent to de-blurring the image, leading to a original image x. This however is an ill-posed problem, since an infinite number of such pairs can be found [41], making some form of regularisation inevitable. The classical approaches assume zero noise and employ regularised least squared methods [4244]. Each regularisation is based on assumptions of the kernel and the original image. More recently, natural image statistics and the Bayesian framework were used to formulate such assumptions more precisely and thus improved the performance of blind deconvolution methods, see [4547] for a first overview. Simply put, prior knowledge and the ability to incorporate this knowledge improve the deconvolution result. However, we are in a different situation since we miss this prior knowledge and we are shall not be inclined to make unverifiable assumptions. Furthermore, we are not primarily interested in the most probable reconstructed image x. Instead, we want a set of possible boundaries given a blurred image and a small set of assumptions. Therefore, we introduce the fuzzy edge detection which is based on three simple assumptions
(a)
Each inclusion is a star-shaped, i.e. there exist a midpoint from which each boundary point is reached by a straight line.
 
(b)
The gradient of the blurred image yields a connected domain and the boundary of the inclusion lies in this connected domain.
 
(c)
Each boundary curve has a limited prescribed variation.
 
These three assumptions implicitly define a possible set of boundary curves provided a blurred image. Any of these boundary curves is possible but nothing can be said about the probability of each curve. Since we assume a star-shaped inclusion, a radial function \(r(\theta ):[0,2\pi )\mapsto \mathbb {R}_{+}\) is able to describe the boundary of this inclusion. The uncertainty of the boundary is then captured in a fuzzy function \(\tilde{r}\).
In order to construct a fuzzy function \(\tilde{r}\) from a blurred image with domain \(D=[-1,1]^2\), we use trigonometric interpolation on a vector of \(N\in \mathbb {N}\) fuzzy numbers. Each component of the fuzzy vector denoted by
$$\begin{aligned} \tilde{R}_N:= \tilde{R}_N[\Theta ] := \left( \tilde{r}_{\theta _1}\ \cdots \ \tilde{r}_{\theta _{N}} \right) \end{aligned}$$
(18)
is constructed from a radial cut (i.e. a line segment) that starts in the center and ends at the outer boundary for a prescribed set of angles \(\Theta = \left\{ \theta _i = \frac{2\pi i}{N+1}\mid i =1,\ldots ,N\right\} \). The continuous function \(\mathcal {I}_i: [0,R_i^\textrm{max}] \rightarrow \mathbb {R}_+\) with \(R_i^\textrm{max}= 1/\max \{\vert \cos \theta _i\vert ,\vert \sin \theta _i\vert \}\) determines the intensity of the gradient along this radial cut, see Fig. 4 for an illustration. The information encoded in \(\mathcal {I}_i\) is used to construct a trapezoidal fuzzy interval. For this, consider lower and upper percentage threshold values \(0<\underline{\delta }< \overline{\delta } <1\) and determine \(M :=\max _{r\in [0,R_i^\textrm{max}] }\mathcal {I}_i(r)\). If the intensity in one point is larger than \(\overline{\delta } M\), it is assigned a possibility of one. If it is smaller than \( \underline{\delta } M \), it gets assigned a possibility of zero. For simplicity, in between these thresholds we assign a possibility by linear interpolation. This algorithm yields a membership function \(\mu _{{\tilde{r}}_{\theta _i}}\) of trapezoidal form, which is illustrated in Fig. 4.
Gathering these fuzzy numbers into a vector yields the non-interactive fuzzy vector \(\tilde{R}_N\), encoded in the membership function
$$\begin{aligned} \mu _{\tilde{R}_N} = \min \{ \mu _{\tilde{r}_{\theta _1}}(r_1), \ldots , \mu _{\tilde{r}_{\theta _N}}(r_N) \}, \end{aligned}$$
where non-interactivity means that the value of one sample does not influence the membership function of another value. Let \(R=(r_1\ \cdots \ r_N ) \in C_{0}[\tilde{R}_N]\) and denote the trigonometric interpolation by
$$\begin{aligned} {\mathcal {T}(\cdot ,R) :[0,2\pi ]\rightarrow \mathbb {R}_+.} \end{aligned}$$
(19)
It represents the mapping from interpolation points onto the trigonometric polynomials. This mapping is bijective if the degree of freedom coincides with the number of interpolation points. The coefficients are efficiently computed via a discrete Fourier transformation and the interpolation scheme yields convergence rates as follows.
Proposition 5
(Trigonometric interpolation) Let \(f:(0,2\pi )\rightarrow \mathbb {R}\) be a k-times differentiable periodic function that describes the boundary of a star shaped inclusion in \(\mathbb {R}^2\), then there exists \(c = c(k,f,f^{(k)})>0\) such that
$$\begin{aligned} \left\| f - \mathcal {T}(\cdot ;[f(\theta _1),\ldots , f(\theta _N)]) \right\| _{L^2(0,2\pi )} \le c N^{-k}. \end{aligned}$$
Given all possible interpolation points \(R\in C_0[ \tilde{R}_N]\), we define the set of radial boundary functions describing the interface as
$$\begin{aligned} B_N = \left\{ \mathcal {T}(\cdot ;R) \mid R \in C_0[\tilde{R}_N] \right\} . \end{aligned}$$
Note that \(\mathcal {T}(\cdot ;\tilde{R}_N)\) defines a fuzzy function [29]. The number of cuts and thus the number of interpolation points N is setted according to the accuracy required by the considered problem. With sufficiently many interpolation points, the set \(B_N\) can represent highly oscillatory boundaries. For the sake of efficiency, the interpolation should be carried out with as few points as possible, which demands an adequate knowledge about the properties (in particular smoothness/roughness) of the physical system at hand. We choose the total variation of the radial function as a measure for roughness, which is defined for \(R\in C_0[\tilde{R}_N]\) as
$$\begin{aligned} \textrm{TV}(R) = \int _{0}^{2\pi } \left|\mathcal {T}'(\theta ;R) \right|\ \textrm{d}\theta . \end{aligned}$$
Given some bound \(0\le b \le \infty \), the TV restricted set of radial boundary function based on N grid points is defined by
$$\begin{aligned}{} & {} B_{N,b} := \left\{ \mathcal {T}(\cdot ;R):[0,2\pi ]\rightarrow \mathbb {R}_+ \mid R \in C_0[ \tilde{R}_N]\ \right. \nonumber \\{} & {} \qquad \qquad \left. \text {and}\ TV(R) \le b \right\} . \end{aligned}$$
(20)
With this construction, we may define the interactive fuzzy set
$$\begin{aligned} \tilde{R}_{N,b}&:= \left\{ (R, \mu _{\tilde{R}_{N,b} }) \mid \, R\in C_0[\tilde{R}_{N}]\text { and } \textrm{TV}(R)\le b\right\} , \end{aligned}$$
(21)
$$\begin{aligned} \mu _{\tilde{R}_{N,b}}&:= \mu _{\tilde{R}_N}\textbf{1}_{\{R \mid TV(R)\le b\}}, \end{aligned}$$
(22)
where \(\textbf{1}_X\) denotes the characteristic function on a set X. In Sect. 3.1 we make clear that the fuzzy set in (21) in fact defines a fuzzy vector. This in turn motivates the \(\alpha \)-cut propagation of fuzzy uncertainty, introduced in Sect. 2.1. In particular, consider a continuous real valued function \(\mathcal {Q}\) defined on \(C_0[\tilde{R}_N]\) which is the 0-cut of the non-interactive fuzzy vector \(\tilde{R}_N\) from (18). Note that \(C_0[\tilde{R}_N]\) is compact since the image domain \(D=[-1,1]^2\) is bounded. The propagated uncertainty of the interactive fuzzy set \(\tilde{R}_{N,b}\) through \(\mathcal {Q}\) is then captured on each \(\alpha \)-level
$$\begin{aligned} C_\alpha [\tilde{\mathcal {Q}}_b] := \left[ \min _{R\in C_\alpha [\tilde{R}_{N,b}] } \mathcal {Q}(R), \max _{R\in C_\alpha [\tilde{R}_{N,b}] } \mathcal {Q}(R) \right] \end{aligned}$$
(23)
of the fuzzy set \(\tilde{\mathcal {Q}}_b := \mathcal {Q}(\tilde{R}_{N,b})\). We emphasise that \(C_0[\tilde{R}_N]\) defines a tensor domain on which \(\mathcal {Q}\) is well defined, i.e.
https://static-content.springer.com/image/art%3A10.1007%2Fs00466-022-02261-z/MediaObjects/466_2022_2261_Equ24_HTML.png
(24)
a fact that becomes useful when applying low-rank tensor formats in the upcoming Sect. 4.2 as surrogate models for \(\mathcal {Q}\). However, the computation of \(\tilde{Q}_b\) only requires evaluation of \(\mathcal {Q}\) on \(C_0[\tilde{R}_{N,b}]\subset C_0[\tilde{R}_N] \), which in general is not a tensor domain.

3.1 Properties of the TV bounded fuzzy set \(\tilde{R}_{N,b}\)

The total variation determines the shape of the inclusion. With a very small total variation, the shapes become more and more circular. Any circular shape, independent of the radius, has a total variation of zero. We want to point out, that we do not measure the total variation of the trajectory, where the total variation of a circle would be larger than zero. Instead we measure the total variation of the radial function.
In Fig. 5 shapes with different total variations are depicted. The blue shape is generated by taking alternating radii, the white shape is the result of optimization. The maximal total variation is 5.2 for \(N=12\), whereas a randomly generated shape has a total variation of around 2. With increasing N the maximal total variation would also increase.
Note that the resulting shapes may violate the boundaries in between the sampling points. Especially, the maximum total variation solution. This can be resolved with more sampling points, a sufficiently strict TV bound or by replacing the trigonometrical interpolation with a corresponding spline interpolation.
Bounding the total variation leads to interaction of the fuzzy set \(\tilde{R}_{N,b}\). Figure 6 illustrates the interaction for the case \(N=3\). It shows that fixing one point constraints the remaining points to the light gray area. Hence the possibility outside this area is zero. In a non-interactive setting the possibility could be strictly larger than zero. Consequently, the total variation bound restricts the set of possible curves. The Figure indicates that the set of valid radial points is convex, which is shown in the following proposition.
Proposition 6
(Convexity of TV constrained domain) Let \(N\in \mathbb {N}\). Then for \(b\ge 0\) the 0-cut \(C_0[\tilde{R}_{N,b}]\) defines a convex set.
Proof
Let \(R_1, R_2 \in {\text {supp}}\tilde{R}_N\), \(t\in [0,1]\) and define \(R_t = tR_1 + (1-t)R_2\). The radial function to describe the boundary then takes the form
$$\begin{aligned} \mathcal {T}(\phi ,R_t ) = \sum \limits _{k=0}^{N-1} a_k[R_t] e^{ik\phi } \end{aligned}$$
with imaginary number i and coefficients \(a_i[R_t]\). Since the Fourier interpolation defines a linear operator, it holds \(a_i[R_t] = ta_i[R_1] + (1-t)a_i[R_2] \). Thus, by triangle inequality it follows that
$$\begin{aligned} TV(R_t) \le t TV(R_1) + (1-t) TV(R_2) \le b. \end{aligned}$$
\(\square \)
Consequently the \(\alpha \)-cuts of \(\tilde{R}_{N,b}\) are nested and convex.
Proposition 7
(Characterisation of \(\tilde{R}_{N,b}\)) Let \(b\ge 0\) and \(\tilde{R}_N\) be a fuzzy (domain) vector with \(C_1[\tilde{R}_N] \subset C_0[\tilde{R}_{N,b}]\). Then \(\tilde{R}_{N,b}\) defines a fuzzy (domain) vector.
Note that for given N, there always exists \(b=b(N)\) such that the conditions of Proposition 7 hold true. In particular, for b large enough it holds \(\tilde{R}_N = \tilde{R}_{N,b}\). Consequently, the \(\alpha \)-cut propagation (23) can be applied.
Remark 1
The convexity property from Proposition 6 is essential to apply Theorem 2. It is a consequence of the linear interpolation and the triangle inequality property of the TV based restriction. In the case, one of the latter properties does not hold, the fuzzy set may not define a fuzzy domain. Then, the propagation of the fuzzy set still can be realised by sampling approaches as discussed in the end of Sect. 2.1.

4 Accelerated emulation of fuzzy effective material

The fuzzy edge detection described above yields a fuzzy vector \(\tilde{R}_N\), where N is the number of radial cuts. Trigonometric interpolation for each realization \(R\in \tilde{R}_N\) then results in a boundary representation of the inclusion. In the following, we define a composite material based on this representation.
Consider the square \([-1,1]^2\) on which we define a two phase composite material \(\varvec{C}(x)= \varvec{C}(x,R)\) with \(\varvec{C}_\text {incl},\varvec{C}_\text {matrix} \in \mathbb {E}\textrm{la}(d,\text {iso})\). This description represents a piecewise isotropic material with star-shaped inclusion defined as
$$\begin{aligned} \varvec{C}(x,R) := {\left\{ \begin{array}{ll} \varvec{C}_\text {incl}, &{} r \le \mathcal {T}(\theta , R), \\ \varvec{C}_\text {matrix}, &{} \text {else}, \end{array}\right. }\quad x = r [\cos (\theta ), \sin (\theta )], \end{aligned}$$
(25)
such that \(0< R_{\text {min}}< \mathcal {T}(\theta , R)< R_\text {max} < 1\) uniform in \(\theta \) and R.
Given a fixed R, the homogenisation from Sect. 2.2 yields the constitutive tensor \(\varvec{C}(\cdot ,R) \in \mathbb {E}\textrm{la}(d)\) for the upscaled macroscopic material. In Voigt notation, this mapping is denoted as
$$\begin{aligned} R \mapsto \mathcal {H}(R) = \left[ \begin{array}{ccc} H_{11} &{} H_{12} &{} H_{13} \\ &{} H_{22} &{} H_{23} \\ \textit{sym} &{} &{} H_{33} \end{array} \right] (R) \in \mathbb {R}^{3,3}, \end{aligned}$$
(26)
with symmetric positive definite matrix \(\mathcal {H}(R)\in \mathbb {R}^{3,3}\). It in general describes an anisotropic material since the involved geometry of the inclusion may lack any type of symmetry.
We would like to point out that the parametric dependency \(R\rightarrow \varvec{C}(\cdot , R)\) does not define a continuous function with images in \(L^\infty (D)^{d,d,d,d}\) since marginal changes of the shape of the inclusion immediately yield an \(L^\infty \) error equal to the contrast \(\Vert \varvec{C}_\text {incl}-\varvec{C}_\text {matrix}\Vert _F\) with Frobenius norm \(\Vert \cdot \Vert _F\). Despite this discontinuity and its effect on the regularity of the parametric solution \(R\rightarrow \varvec{u}(R)\), the parametric homogenised tensor (26) defines a continuous map.
Recall that the evaluation of \(\mathcal {H}\) involves multiple simulations of periodic linear elastic boundary value problems of the form (11). To accelerate the upscaling process, in the following we replace the simulator with an emulator. The emulator relies on a mesh discretisation of the domain D and a group of tensor train surrogates to approximate \(\mathcal {H}\), which is discussed in Sect. 4.2.
If D is discretised by an automatic mesh generator under the constrained of equal amount of vertices on the inclusion’s boundary and on the boundary of D for any R, the resulting mesh map \(R\mapsto \mathcal {M}(R)\) in general is discontinuous. As a consequence, a mesh dependent finite element computation may inherit the lack of continuity even through the whole geometry has smooth dependence on R, see Fig. 7 for an example. This in turn aggravates optimization based on gradient information. We solve this issue in Sect. 4.1 by constructing a family of transformed meshes with smooth dependence on R.

4.1 Constructive smooth transformation of meshes

For \(n\in \mathbb {N}\) interpolation points, consider the composite interface discretisation
$$\begin{aligned}{} & {} \mathcal {T}_n(R) := \{ x = \mathcal {T}(\theta _i, R)[\cos (\theta _i)\ \sin (\theta _i)]^\top , \\{} & {} \qquad \theta _i = 2\pi i / n, i = 0,\ldots ,n-1 \}. \end{aligned}$$
We now construct a smooth diffeomorphism that creates meshes based on a single reference mesh \(\hat{\mathcal {M}}\) denoted as \(\Psi :\mathbb {R}^d\times \mathbb {R}^N\rightarrow \mathbb {R}^d\) such that \(\Psi (\hat{\mathcal {M}}, R)=\mathcal {M}(R)\). Let \(R_\text {ref} = [0.5,\ldots ,0.5]\in \mathbb {R}^N\) and \(\mathcal {T}(\cdot ,R_\text {ref})\) define a circular inclusion with radius 0.5. Consider a fixed reference mesh discretisation \(\mathcal {M}_\text {ref}\) of \([-1,1]^2\) such that
  • there is a fixed number of nodes on the boundary \(\partial [-1,1]^2\)
  • any node in \(\mathcal {F}_n(R_\text {ref})\) is part of the mesh.
Compute the sets \(\mathcal {F}_{2n}(R_\text {ref})\) and \(\mathcal {F}_{2n}(R)\) corresponding to higher discrete resolution of the reference or target interface. The transformation \(\Psi \) then consists of two parts as follows. First, let \(\{\phi _i\}\) form a smooth partition of unity of \([0,2\pi ]\) with \(\phi _i(i\pi / n) = 1\) and \({\text {supp}}\phi _i = [(i-1)\pi /n, (i+1)\pi /n]\). This function set deals with the angular part of the transformation. Second, let \(\chi _i\in \mathcal {C}^k[0,\sqrt{2}]\) be a set of splines with \(k\ge 2\) for the radial part of the transformation with the following properties:
1.
The reference radius is mapped onto the transformed radius, i.e.
$$\begin{aligned} \chi _i(\mathcal {T}(\theta _i, R_{\text {ref}})) = \mathcal {T}(\theta _i,R). \end{aligned}$$
 
2.
The spline is strong monotonically increasing on \((R_{\text {min}}, R_{\text {max}})\) and otherwise equals the identity map.
 
Altogether, the transformation reads
$$\begin{aligned} \Psi (\hat{x}, R) := \sum \limits _{i=1}^{2n} \chi _i(\hat{r})\phi _i(\hat{\theta }),\quad \hat{x} = \hat{r}[\cos \hat{\theta }, \sin \hat{\theta }]^\top . \end{aligned}$$
(27)
Note that the number of radial cuts N determines the range of shapes, whereas the number of interpolation points n has to be sufficient large to trace the shape and to account for the resolution of the finite element mesh. We refer to Fig. 8 for an illustration of the capacity of the constructed map \(\Psi \), which is characterised next.
Proposition 8
(\(\mathcal {C}^k\) diffeomorphism) The map \(\Psi : [-1,1]^2 \rightarrow [-1,1]^2\) is bijective and k-times continuously differentiable.
Proof
The bijectivity follows from strong monotonicity in radial direction and the smooth partition of unity in angular direction. The smoothness of \(\phi _i\) and of the polar coordinate mapping away from zero is given since \(\chi _i\) is k-times continuous differentiable. For \(i=1,\ldots ,2n\) the desired property follows immediately. \(\square \)

4.2 Tensor trains based emulation

Assume some function \(f:{C_0[{\tilde{R}}]}\subset \mathbb {R}^N\rightarrow \mathbb {R}\) with https://static-content.springer.com/image/art%3A10.1007%2Fs00466-022-02261-z/MediaObjects/466_2022_2261_IEq299_HTML.gif . For \(f = H_{ij}\), \(1\le i \le j\le 3\) we consider surrogates of the form
https://static-content.springer.com/image/art%3A10.1007%2Fs00466-022-02261-z/MediaObjects/466_2022_2261_Equ28_HTML.png
(28)
based on a polynomial feature class \(\Pi :=\{\Pi _{\beta }, {\beta }\in \mathbb {N}^N\}\) where each polynomial \(P_{\beta } \in \Pi \) is of tensorised form \(\Pi _{\beta } = \bigotimes _{i=1}^N q_{{\beta }_i}^i\) with one dimensional polynomials \(\{q_{{\beta }_i}^i:C_0[\tilde{R}_i] \rightarrow \mathbb {R}, {\beta }_i \in \Lambda _i\}\) for \(i=1,\ldots , N\) and unknown coefficient tensor \(\varvec{U}:\Lambda \rightarrow \mathbb {R}\). This model suffers from the curse of dimensionality since the cardinality \(\vert \Lambda \vert \) grows exponential with respect to the dimension N. A possibility to circumvent this challenge lies in a compressed representation of the coefficient tensor, here based on the tensor train format (TT format) [48]. Let \(\rho =(\rho _1,\ldots , \rho _{N-1})\in \mathbb {N}^{N-1}\) be the tensor train rank and let \(\rho _0=\rho _N=1\). We then choose \(\varvec{U}\) given in tensor train decomposition by
$$\begin{aligned} \varvec{U}[{\beta }] := U_1[{\beta }_1] U_2[{\beta }_2] \cdots U_n[{\beta }_N], \quad {\beta }\in \Lambda , \end{aligned}$$
(29)
with component tensors of order 3, such that \(U_i[:]\in \mathbb {R}^{\rho _{i-1}, \vert \Lambda _i\vert , \rho _i}\) for \(i=1,\ldots , N\). The number of degrees of freedom for this design is given as \(\sum _{i=1}^N \rho _{i-1} \vert \Lambda _i\vert \rho _i\), which does not grow exponentially but only linearly in dimension N. Given data \((R^{(k)}, y^{(k)})_{k=1}^K\), \(K\in \mathbb {N}\) with \(y^{(k)}=f(R^{(k)})\), we obtain a surrogate as in (28) by carrying out a regularised empirical regression, namely by solving the optimisation problem
$$\begin{aligned} \min \limits _{{\varvec{U}} \text { as in } (29)} \sum \limits _{k=1}^K \Vert f_\Lambda (R^{(k)}) - y^{(k)}\Vert + \lambda \Vert {\varvec{U}} \Vert _{{\textrm{Fro}}} \end{aligned}$$
(30)
with regularisation parameter \(\lambda >0\) and the Frobenius norm \(\Vert \cdot \Vert _{{\textrm{Fro}}}\). The optimisation problem (30) can be solved by regression with an alternating linear scheme (ALS) [49]. As a modification of the basic ALS, we introduce a scheme for rank adaptivity. This concept offers several advantages to obtain an adjusted tensor train rank which is in general unknown a priori. From a practical point of view, it reduces the computational cost during optimisation while obtaining a prescribed accuracy in the approximation class. Furthermore, starting with a small rank, the iterative process empirically enables the ALS to converge to a solution based on successive computations of initial guesses for models with higher ranks based on the given restricted number of samples. The proposed approach is summarised in Algorithm 1 and the rank adaptivity is presented in Algorithm 2.
A principle of the rank adaptivity is to keep one (control) singular value to monitor the importance of the related rank coupling value during the optimisation process. This additional singular value per rank coupling remains until the end of the regression scheme to prevent oscillation between rank growth and reduction. It ensures an upper bound of the related rank value throughout the entire process. After successful termination of Algorithm 2, the existing (possibly small) control singular value is removed by a final rounding [48] of the resulting tensor train. We refer to [49] for more technical details on the basic ALS, e.g. with regard to setting and moving the non-orthogonal component (the core) via left and right matricisation and orthogonalisation. Eventually, the overall emulator is obtained by evaluation of the 6 scalar valued tensor train surrogate maps, replacing \(R\mapsto [\mathcal {H}(R)]_{ij}\) and \(1\le i\le j\le 3\) corresponding to the upper triangular part of the symmetric matrix \(\mathcal {H}(R)\).
Note that non-zero values of \(H_{13}\) and \(H_{23}\) solely appear when the effective material behaviour is not isotropic or rotational-free orthotropic, i.e., this is described by (13) with \(\theta =0\).

5 Numerical experiments

This section is concerned with the assessment of the numerical performance of the approach presented above, for which two experiments are examined. In the first experiment, we measure the distance of the upscaled material to the isotropy and orthotropy class. Additionally, we identify configurations that maximise and minimise the respective distances. In the second experiment, we test if the upscaled material is suitable for a worst case analysis, i.e., we test if it is possible to find bounds that envelop the most extreme behaviour of the quantity of interest under consideration. For this, a material with \(8\times 8\) arbitrary inclusions placed on a checkerboard is compared to an upscaled material with the same geometrical dimensions. Both experiments demonstrate the efficacy of the fuzzy approach to model uncertainty of the inclusion to identify extremal behaviour and its source.
Three types of computational tasks are performed in the experiments. The foundation is laid by Finite Element (FE) simulations to generate the realisations of constitutive tensors by solving (11). Furthermore, an alternating linear scheme is used to train the surrogate model and for optimization computations to carry out the uncertainty propagation based on \(\alpha \)-cuts.
All FE computations are done with the FEniCS‘ package [50]. The mesh generation is realised using Gmsh‘ [51] and the Bubbles‘ package [52]. Moreover, the python package TensorTrain‘ [53] is utilised for the rank adaptive tensor train regression and ALEA‘ [54] for the underlying polynomial features. The optimization tasks are performed with restarted trust region optimization implemented in Scipy‘ [55].
The computations are performed with \(N=6\) and \(N=12\) radial sampling points. Here, the \(N=6\) case imposes significantly less computational burden on the surrogate and the optimization than the \(N=12\) case. In particular, our optimization scheme took up to \(10^{6}\) evaluations of the \(\mathcal {H}(R)\) for \(N=12\) and up to \(1\times 10^{5}\) evaluations for \(N=6\) per propagated quantity of interest. We underline that all optimization tasks involved are non-convex with non-linear cost functions and constraints. Consequently, we take arbitrary points \(R \in [0.3, 0.7]^6\) for \(N=6\) and \(R \in [0.4, 0.6]^{12}\)—a smaller domain—for \(N=12\). This enables sufficiently non-trivial shapes without rendering the surrogate model and the optimization infeasible. Each realisation R determines a boundary stellar inclusion in terms of a trigonometric interpolation. Inside and outside of the inclusion we assume isotropic material behaviour. Concretely, we let the Young’s modulus \(E_\text {incl} = 3230\) and Poisson ratio of \(\nu _\text {incl}=0.3\) in the inside. Outside of the inclusion we choose \(E_\text {matrix}= E_\text {incl}/4\) and \(\nu _\text {matrix}=0.2\). These values are transformed into Lamé constants by
$$\begin{aligned} \lambda = \frac{E\nu }{(1+\nu )(1-2\nu )},\quad \mu = \frac{E}{2(1+\nu )}. \end{aligned}$$
By adaptation to the plane, adapted Lamé constants are obtained, namely
$$\begin{aligned} \lambda ^*= \frac{2\lambda \mu }{\lambda + 2\mu },\quad \mu ^*= \mu . \end{aligned}$$
For a given \(R\in C_0[\tilde{R}_N]=[0.3,0.7]^N\) the homogenisation problem (11) is solved with FE of uniform polynomial degree \(p=2\). The reference mesh is based on 80 nodes on the related reference composite interface \(\mathcal {T}(\cdot ;R_\textrm{ref})\) is shown in Fig. 8. The computational mesh with composite interface \(\mathcal {T}(\cdot , R)\) is then obtained by applying the transformation \(\Psi (\cdot , R)\).

5.1 Surrogate validation

As a preparative step, we build an accelerated emulator in the tensor train format for \(N=6\) and \(N=12\) dimensional fuzzy input vectors \(\tilde{R}_N\), yielding a compressed surrogate of the map \(R\rightarrow \mathcal {H}(R)\).
Given \(K = 15635\) \((N=6)\) and \(K = 12000 + 4096\) \((N=12)\) samples that are normalised with respect to the sample mean and variance, we iteratively apply Algorithm 1 and Algorithm 2 with tensorised Chebyshev polynomials up to degree 5 in each coordinate. More precisely, we train the initial tensor train surrogate with a degree of two. Next, a new tensor train for polynomial degree \(\textrm{deg}+1\) up to \(\textrm{deg}=4\) is set to the last tensor approximation as initial guess instead of starting with a random rank-1 tensor train. All tensor train component entries associated to the higher polynomial degree are initially set to zero. This training approach turn out to have two advantages. First, in the iterative procedure the training and validation sets are split randomly, resulting in a pseudo stochastic solver. A strategy that is successful in the context of machine learning. Second, there is a huge speed-up in the training procedure for finding a local minimum. In fact, with a naive tensor train training for \(N=12\) with initial tensor polynomial degree 5, we rarely observed convergence to meaningful surrogates given random initial values.
We use \(\textrm{tol}_{\textrm{MSE}}=1\times 10^{-8}\), \(\textrm{iter}_{\textrm{max}}=200\), \(L_{\textrm{HIST}}=10\), \(\textrm{tol}_{\textrm{DECAY}}=1\times 10^{-5}\), \(r_\textrm{max}\le 5,7,9,20\) for \(\textrm{deg}=2,\ldots ,5\) , \(r_\textrm{add}\in \{1,2\}\), and \(\delta = 1\times 10^{-8}\). The results of the surrogate learning approach are depicted in Tables 1 and 2, which complement Figs. 10 and 11, showing mean square error or pointwise relative and absolute errors. The highlighted entries in both tables show a pointwise lack of accuracy of the tensor train surrogate for the \(H_{13}\) and \(H_{23}\) components, which represent the pure rotational or anisotropic contribution in the overall upscaled tensor \(\mathcal {H}\) on the used test sets. A further inspection of the high relative errors showed that these strictly belong to pointwise approximations of values close to zero. Nevertheless, strikingly small mean squared errors can be observed as a result of the proposed optimization algorithm. Figures 10 and 11 display the value ranges of the subcomponents of \(\mathcal {H}\) and the overall absolute and relative errors of the subcomponents and the full tensor \(\mathcal {H}\). We observe that the reduced (pointwise) accuracy of the surrogates for \(H_{13}\) and \(H_{23}\) does not influence the overall error for the full parametric tensor \(\mathcal {H}\). The final obtained ranks obtained with the rank adaptive algorithm are depicted in Fig. 9 for \(N=6\) and \(N=12\).
Table 1
Surrogate validation for \(N=6\) on \([0.3,0.7]^N\) based on 15635 data points randomly split into 14, 635 training and 1000 test samples
 
Validation set
Training set
 
MSE
Max error
Max rel. error
MSE
\(H_{11}\)
\(4.6\times 10^{-9}\)
\(2.9\times 10^{-1}\)
\(2.1\times 10^{-4}\)
\( 4.5\times 10^{-9}\)
\(H_{22}\)
\(10.2\times 10^{-9}\)
\(6.3\times 10^{-1}\)
\(6.0\times 10^{-4}\)
\( 9.4\times 10^{-9}\)
\(H_{33}\)
\(9.3\times 10^{-9}\)
\(1.3\times 10^{-1}\)
\(3.2\times 10^{-4}\)
\( 8.6\times 10^{-9}\)
\(H_{12}\)
\(9.1\times 10^{-9}\)
\(7.6\times 10^{-2}\)
\(3.1\times 10^{-4}\)
\( 8.6\times 10^{-9}\)
\(H_{13}\)
\(88.8\times 10^{-9}\)
\({\textbf {1.1}}\times {\textbf {10}}^{-{\textbf {2}}}\)
\({\textbf {1.4}}\times {\textbf {10}}^{+{\textbf {3}}}\)
\( 43.7\times 10^{-9}\)
\(H_{23}\)
\(97.9\times 10^{-9}\)
\({\textbf {9.7}}\times {\textbf {10}}^{-{\textbf {3}}}\)
\({\textbf {1.2}}\times {\textbf {10}}^{+{\textbf {3}}}\)
\( 45.5\times 10^{-9}\)
The bold values of maximum relative errors of \(H_{13}\) and \(H_{23}\) are the result of the surrogate model approximating a value \(1\times 10^{-7}\) with \(1\times 10^{-4}\). These effects do not spoil the accuracy of the total surrogate \(\mathcal {H}\) in subsequent experiments
Table 2
Surrogate validation for \(N=12\) on \([0.4,0.6]^N\) trained on normalised data
 
Validation set
Training set
 
MSE
Max error
Max rel. error
MSE
\(H_{11}\)
\(1.1\times 10^{-7}\)
\(3.5\times 10^{0}\)
\(3.1\times 10^{-3}\)
\( 5.9\times 10^{-8}\)
\(H_{22}\)
\(1.0\times 10^{-7}\)
\(3.8\times 10^{0}\)
\(3.3\times 10^{-3}\)
\( 7.4\times 10^{-8}\)
\(H_{33}\)
\(6.8\times 10^{-7}\)
\(5.1\times 10^{0}\)
\(1.2\times 10^{-2}\)
\( 5.2\times 10^{-7}\)
\(H_{12}\)
\(5.1\times 10^{-7}\)
\(1.8\times 10^{0}\)
\(8.0\times 10^{-3}\)
\( 3.0\times 10^{-7}\)
\(H_{13}\)
\(6.9\times 10^{-5}\)
\({\textbf {0.8}}\times {\textbf {10}}^{{\textbf {0}}}\)
\({\textbf {1.4}}\times {\textbf {10}}^{+{\textbf {4}}}\)
\( 4.4\times 10^{-5}\)
\(H_{23}\)
\(1.0\times 10^{-4}\)
\({\textbf {0.6}}\times {\textbf {10}}^{{\textbf {0}}}\)
\({\textbf {3.5}}\times {\textbf {10}}^{+{\textbf {3}}}\)
\( 7.5\times 10^{-5}\)
The mean squared error (MSE) is computed w.r.t. to the normalised value range of the data. The maximum pointwise relative and absolute errors are measured w.r.t to the surrogate values, which are shifted and scaled back to the original range. The bold values of maximum relative errors of \(H_{13}\) and \(H_{23}\) are the result of the surrogate model approximating a value \(1\times 10^{-7}\) with \(1\times 10^{-4}\)

5.2 Distances to the isotropy and orthotropy material class

In this section we apply the tensor train emulator to obtain an approximation of the parametric constitutive tensor \(\mathcal {H}(R)\) of the effective material defined in (26), given an parametrisation of the inclusion through the vector \(R\in C_0[\tilde{R}_{N,b}]\). For each constitutive tensor we compute the distance to the isotropy class, denoted by \(d_{\textrm{iso}}\), and orthotropy class, denoted by \(d_{\textrm{ortho}}\), according to (14). The fuzzy uncertainty of the boundary is propagated in terms of \(\alpha \)-cuts. On each \(\alpha \)-level, two optimization problems with constraints defined in (23) have to be solved, where \(\mathcal {Q}\) is either \(d_{\textrm{iso}}\) or \(d_{\textrm{ortho}}\). As optimiser we use a restarted trust region scheme to minimise the non-convex and non-linear target function with non-linear constraints.
Figure 12 shows the experimental results for \(\mathcal {Q}= d_{\textrm{iso}}\) with \(N=6\), zero \(\alpha \)-cut \(C_0[\tilde{R}_N] = [0.3,0.7]^N\) and the total variation bounds \(b=0.5, 1, \infty \). We point out that since \(D=[-1,1]^2\) is not of special hexagonal shape but of square shape, a circular inclusion does not yield isotropic effective material properties. Nevertheless, the circular shape—while exhibiting vanishing total variation—provides effective properties closest to the isotropy class. Since in our experiments we choose a stiffer material as inclusion, the circular inclusion minimises its size proportional to the possible maximal radius encoded in the \(\alpha \)-level. In this sense, the size of a circle is a measure of perturbation of the isotropic matrix material. Moreover, if the total variation and radii bounds allow it, the inclusion converges to a peanut shape. The green peanut shaped inclusion in Fig. 12 marks the inclusion with maximum distance to the isotropic material class for the given parameterisation family \(B_{N,\infty }\) from (20). Note that by rotational invariance also the peanut rotation of \(\pi /2\) yields the same maximising shape.
Figure 14 pictures the result for \(\mathcal {Q}= d_{\textrm{iso}}\) with \(N=12\), zero \(\alpha \)-cut \(C_0[\tilde{R}_N] = [0.4,0.6]^N\) and total variation bounds \(b=0.5, 1, \infty \). The shrinked domain has an immediate impact on the maximal possible total variation bound of all interface realisations. Consequently, maximum distances are caused by curves with similar shape up to rotation. Already for a total variation of 0.5, a symmetric shape emerges that becomes more prominent with increasing total variation.
Figure 13 depicts the experimental results for \(\mathcal {Q}= d_{\textrm{ortho}}\) with \(N=6\), zero \(\alpha \)-cut \(C_0[\tilde{R}_N] = [0.3,0.7]^N\) and total variation bounds \(b=0.5, 1, \infty \). A minimum of 0 is attained on each \(\alpha \) cut of the corresponding membership function, since the orthotropy class \(\mathbb {E}\textrm{la}(2,\textrm{ortho})\) contains the square symmetric class. Many boundaries in \(B_{N,b}\) lead to a shape that yields a square symmetric effective material.
As typical maximiser of the distance to the orthotropy class, a bean shape is found. We observe that the maximal possible total variation value of \(B_{N,b}\) is not exhausted when attaining the maximal distance. Note that due to rotational invariance, the bean shape can be rotated by multiples of \(\pi /2\) while still being a (in fact the same) maximiser.
Figure 15 shows the experimental results for \(\mathcal {Q}= d_{\textrm{ortho}}\) with \(N=12\), zero \(\alpha \)-cut \(C_0[\tilde{R}_N] = [0.4,0.6]^N\) and total variation bounds \(b=0.5, 1, \infty \). Again the shrinked domain has an immediate impact on the maximal possible total variation bound of all interface realisations. Consequently, maximum distances are attained by curves with similar shapes up to rotation. We observe that symmetric dented shape maximises the distance to the orthotropy class for the \(\alpha =1\) cut or a total variation smaller or equal to 0.5. However, as the total variation bound gets larger and we allow for larger fluctuations due to a wider parameter range the inclusion converges to a non-axis aligned structure with non-symmetric dented shape.

5.3 Best/worst case estimates for non-homogenised matrix composites

Let the domain \(D=[-N_s,N_s]\times [-N_s,N_s]\) with \(N_s=8\) and consider a checkerboard partitioning {\( D_{ij},i,j=1,\ldots ,N_s\}\) of D with square subdomains \(D_{ij}\) defined by
$$\begin{aligned}{} & {} D_{ij} := [-N_s + (i-1), -N_s + i]\\{} & {} \qquad \times [-N_s + (j-1), -N_s+ j]\,\text {for } i,j=1,\ldots ,N_s. \end{aligned}$$
On each subdomain \(D_{ij}\) we consider a single two phase composite as defined by (25) using a local polar coordinate system w.r.t. the midpoint of \(D_{ij}\), encoded in a material tensor \(\varvec{C}_\text {inhomo}\) with piecewise values \(\varvec{C}_\text {incl}\) and \(\varvec{C}_\text {matrix}\) specified in Sect. 5.1.
For this experiment we choose the non-interactive fuzzy set \(\tilde{R}_{N,\infty }\) with \(N=6\) trapezoidal fuzzy components characterised by
$$\begin{aligned} C_0[\tilde{R}_{N,\infty }] =\ {}&[0.32, 0.63] \times [0.35, 0.66] \times [0.36, 0.70] \\ \qquad&\times [0.32, 0.62] \times [0.38, 0.67] \times [0.31, 0.66] \\ C_1[\tilde{R}_{N,\infty }] =\ {}&[0.48, 0.55] \times [0.47, 0.56] \times [0.49,0.55] \\ \qquad&\times [0.47, 0.56]\times [0.45, 0.59] \times [0.48, 0.50]. \end{aligned}$$
In between these prescribed points, the \(\alpha \)-levels are obtained by linear interpolation. The TV bound is ignored in this experiment.
Furthermore, we consider a homogenised material tensor \(\varvec{C}_\text {homo} = \mathcal {H}(R)\) for \(R\in C_0[\tilde{R}_{N,b}]\) obtained from the homogenisation process of Sects. 2.2 and 5.1. In each domain, we model individual star shaped material inclusions as in (25) with boundaries of uniformly bounded total variation smaller or equal to b.
Given a material law \(\varvec{C}\in \{\varvec{C}_\text {inhomo},\varvec{C}_\text {homo}\}\), we then solve the linear elasticity problem
https://static-content.springer.com/image/art%3A10.1007%2Fs00466-022-02261-z/MediaObjects/466_2022_2261_Equ31_HTML.png
(31)
representing a linear elastic tensile test. Here, \(\Gamma _{\varvec{0}}\) = \(\{-N_s\}\times [-N_s,N_s]\) and
$$\begin{aligned} \varvec{g}(x) = \left\{ \begin{array}{ll} \varvec{0}, &{} x\in [-N_s,N_s] \times \{-N_s,N_s\}, \\ 100\varvec{e}_1,&{} x\in \{N_s\}\times [-N_s,N_s]. \end{array} \right. \end{aligned}$$
(32)
We are interested in the fuzzy propagation of the homogenised material through the average displacement \(\mathcal {Q}=\overline{u}\) defined by
$$\begin{aligned} \overline{u} := \frac{1}{{\text {vol}}(D)} \int \limits _D \Vert \varvec{u}(x)\Vert ^2\,\textrm{d}x, \end{aligned}$$
with the solution \(\varvec{u}\) of (31). Figure 16 shows the resulting membership function based on the fuzzy homogenised material law. The black dots mark resulting average displacements obtained for various non-periodic inclusions. Moreover, some configurations of possible realisations of the \(8\times 8\) inclusions are displayed. The corresponding average displacement is marked with colours, accordingly. Given any non-periodic configuration of prescribed matrix composites, we observe that the membership function at each \(\alpha \)-level bounds the various average displacements. In particular, for \(\alpha = 0., 0.5, 1\) we illustrated a corresponding single composite shape that (assumed in a periodic structure) results in the minimum and maximum average displacements. As the interior composite is slightly stiffer than the surrounding matrix material, it is to be expected that the shapes of the minimiser and maximiser attempt to exhaust or avoid the maximum capacity of the parameter range encoded in \(C_\alpha [\tilde{R}_{N,\infty }]\). It is noteworthy that the homogenised material is able to serve as worst/best case estimator for this particular quantity of interest (i.e. the average displacement). As long as the repetition of the same cell leads to extremal values of the quantity of interest, the fuzzy homogenised material functions as a worst/best case estimator.

6 Conclusion

We considered the possibilistic shape uncertainty of one inclusion induced by a blurred image. In this setting, computational stochastic homogenisation was carried out to propagate the uncertainty through the linear elasticity model, resulting in a fuzzy effective material. Eventually, the effective material was examined with respect to its distance to the orthotropic/isotropic material class. Moreover, it was used as a worst/best case estimator with respect to the global average displacement for a non-homogenised material.
To achieve this result, two major obstacles had to be overcome. The first was caused by the discontinuous mapping from boundary to effective material, which was a consequence of the automatic re-meshing for each new boundary instance. This was successfully resolved with an arbitrarily smooth transformation of a single reference mesh onto the respective mesh for a domain with a star-shaped inclusion. The second challenge was given by the computational effort needed to perform the (possibilistic) uncertainty propagation. Depending on the optimiser and the quantity of interest at hand, each \(\alpha \)-cut required up to \(5\times 10^4\) computational homogenisation simulations. This problem was successfully resolved with a highly accurate tensor train emulator.
We provide an (numerical) analysis for a problem where only little knowledge is given and only simple assumptions can be made. This may be distinguished from other works, where more prior knowledge is given and more elaborated assumptions are suitable. In the latter scenario, one usually can utilise the Bayesian framework to model the uncertainty, see [56, 57], where in our scenario stochastic modelling may lead to a false confidence [23]. While usually the uncertainty of material properties is modelled with precise probabilities, leading to statements about expectation values and probabilities of failure, cf. [58, 59], we model geometrical uncertainties with imprecise probability, leading to statements about possible intervals and worst/best case configurations. Even if an imprecise probability approach for geometrical uncertainty is chosen, the geometrical shape is often restricted to simple parametrisations like a circle of varying diameter [60]. This in fact is a very useful simplification, for instance if the observed quantities only depend on the volume ratio of the two materials. Nonetheless, our experiments demonstrate that the shape determines if the effective material is isotropic, orthotropic or anisotropic. A circle always results in an orthotropic effective material in the considered case with unit square cells. Furthermore, our tensor train surrogate shows a remarkable accuracy of a relative error of order \(\mathcal {O}(10^{-4})\). This is partly the result of a suitable choice of features, the novel rank-adaptive training strategy and the continuity induced by using a mesh transformation instead of automatic re-meshing. Compared to other works such as [61, 62], where standard generic methods like artificial neural networks and polynomial chaos are used in a similar context, we gain several orders of accuracy.
To focus on the influence of the shape uncertainty, other parts in our setting and analysis were kept simple. These parts are ideal starting points for future research. The homogenisation, for example, was performed with two phases of linear elastic materials. This can easily be substituted with more sophisticated materials—one may think of anisotropy, damage and higher contrast—or different homogenisation settings. The assumption of a star-shaped inclusion can be discarded with an adjusted edge detection and a different parametrisation of the boundary. Furthermore, the roughness measurement via total variation can be exchanged or extended with other restrictions of the boundary curves. In addition, it is possible to model the material constants appropriately with precise probabilities. By doing so, we add a stochastic dimension which together with the possibilistic uncertainty results in a fuzzy-stochastic model.
Although we only investigated 2D materials, our approach can be extended to 3D. The numerical upscaling and the surrogate model are applicable in higher dimensions. For the fuzzy edge detection the radial cuts should be replaced by cuts through the sphere. Under the assumption of star-shaped inclusions in 3D, the inclusion surface can be described in spherical coordinates with suitable interpolation for the radial component, which in turn can be constrained by a corresponding 2D TV measure. Moreover, in the 3D case there are more material classes. Depending on the symmetry class, the associated distances are—to our knowledge—not yet analytically given and thus have to be computed numerically [39, 63]. Thus, the theoretical adaption is feasible, but we expect the computational challenges of the 3D case to be significantly larger.
Another interesting research direction for our framework are more involved material models like hyperelastic materials or materials with damage modeling. In this case the numerical upscaling leads to an non-linear effective material law. In order to represent this non-linearity in our framework the surrogate model design needs to be adapted, e.g. mapping to interpolants of effective potentials based on [64].
Finally, we hope that our presented contributions, namely the treatment of discontinuity from re-meshing, the fuzzy edge detection, and the highly accurate tensor train surrogate, can be used and extended beneficially in other research to simulate composite materials in the presence of shape uncertainty.

Acknowledgements

The authors gratefully acknowledge the financial support of the German Research Foundation (DFG) within the Subproject 4 and 10 of the Priority Programme “Polymorphic uncertainty modelling for the numerical design of structures—SPP 1886”.
Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://​creativecommons.​org/​licenses/​by/​4.​0/​.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Literatur
1.
Zurück zum Zitat Jones RM (2018) Mechanics of composite materials, 2nd edn. CRC Press, Boca RatonCrossRef Jones RM (2018) Mechanics of composite materials, 2nd edn. CRC Press, Boca RatonCrossRef
2.
Zurück zum Zitat Casati R, Vedani M (2014) Metal matrix composites reinforced by nano-particles—a review. Metals 4(1):65–83CrossRef Casati R, Vedani M (2014) Metal matrix composites reinforced by nano-particles—a review. Metals 4(1):65–83CrossRef
3.
Zurück zum Zitat Seydibeyoglu MO, Mohanty AK, Misra M (2017) Fiber technology for fiber-reinforced composites. Woodhead Publishing, Swaston, Cambridge Seydibeyoglu MO, Mohanty AK, Misra M (2017) Fiber technology for fiber-reinforced composites. Woodhead Publishing, Swaston, Cambridge
4.
Zurück zum Zitat Oksman Niska K, Sain M (2008) Wood-polymer composites. Woodhead Publishing Materials, Swaston, CambridgeCrossRef Oksman Niska K, Sain M (2008) Wood-polymer composites. Woodhead Publishing Materials, Swaston, CambridgeCrossRef
5.
Zurück zum Zitat Vasiliev VV, Morozov EV (2018) Advanced mechanics of composite materials and structures. Elsevier, Netherlands Vasiliev VV, Morozov EV (2018) Advanced mechanics of composite materials and structures. Elsevier, Netherlands
6.
Zurück zum Zitat Babuška I (1976) Homogenization approach in engineering. In: Computing methods in applied sciences and engineering. Springer, Germany, pp 137–153 Babuška I (1976) Homogenization approach in engineering. In: Computing methods in applied sciences and engineering. Springer, Germany, pp 137–153
7.
Zurück zum Zitat Braides A et al (2002) Gamma-convergence for beginners, vol 22. Clarendon Press, Oxford University PressMATHCrossRef Braides A et al (2002) Gamma-convergence for beginners, vol 22. Clarendon Press, Oxford University PressMATHCrossRef
9.
Zurück zum Zitat Anantharaman A, Le Bris C (2012) Elements of mathematical foundations for numerical approaches for weakly random homogenization problems. Commun Comput Phys 11(4):1103–1143MathSciNetMATHCrossRef Anantharaman A, Le Bris C (2012) Elements of mathematical foundations for numerical approaches for weakly random homogenization problems. Commun Comput Phys 11(4):1103–1143MathSciNetMATHCrossRef
10.
11.
Zurück zum Zitat Bourgeat A, Piatnitski A (2004) Approximations of effective coefficients in stochastic homogenization. In: Annales de l’IHP Probabilités et Statistiques, vol 40, pp 153–165 Bourgeat A, Piatnitski A (2004) Approximations of effective coefficients in stochastic homogenization. In: Annales de l’IHP Probabilités et Statistiques, vol 40, pp 153–165
12.
Zurück zum Zitat Gloria A, Neukamm S, Otto F (2014) An optimal quantitative two-scale expansion in stochastic homogenization of discrete elliptic equations. ESAIM Math Model Numer Anal 48(2):325–346MathSciNetMATHCrossRef Gloria A, Neukamm S, Otto F (2014) An optimal quantitative two-scale expansion in stochastic homogenization of discrete elliptic equations. ESAIM Math Model Numer Anal 48(2):325–346MathSciNetMATHCrossRef
13.
Zurück zum Zitat Engquist B, Runborg O (2002) Wavelet-based numerical homogenization with applications. In: Multiscale and multiresolution methods. Springer, Germany, pp 97–148 Engquist B, Runborg O (2002) Wavelet-based numerical homogenization with applications. In: Multiscale and multiresolution methods. Springer, Germany, pp 97–148
14.
Zurück zum Zitat Ostoja-Starzewski M (2006) Material spatial randomness: from statistical to representative volume element. Probab Eng Mech 21(2):112–132CrossRef Ostoja-Starzewski M (2006) Material spatial randomness: from statistical to representative volume element. Probab Eng Mech 21(2):112–132CrossRef
15.
Zurück zum Zitat Ladeveze P, Nouy A (2003) On a multiscale computational strategy with time and space homogenization for structural mechanics. Comput Methods Appl Mech Eng 192(28–30):3061–3087MathSciNetMATHCrossRef Ladeveze P, Nouy A (2003) On a multiscale computational strategy with time and space homogenization for structural mechanics. Comput Methods Appl Mech Eng 192(28–30):3061–3087MathSciNetMATHCrossRef
16.
Zurück zum Zitat Tranquart B, Ladevèze P, Baranger E, Mouret A (2012) A computational approach for handling complex composite microstructures. Compos Struct 94(6):2097–2109CrossRef Tranquart B, Ladevèze P, Baranger E, Mouret A (2012) A computational approach for handling complex composite microstructures. Compos Struct 94(6):2097–2109CrossRef
17.
Zurück zum Zitat Babuška I, Andersson B, Smith PJ, Levin K (1999) Damage analysis of fiber composites part i: statistical analysis on fiber scale. Comput Methods Appl Mech Eng 172(1–4):27–77MATHCrossRef Babuška I, Andersson B, Smith PJ, Levin K (1999) Damage analysis of fiber composites part i: statistical analysis on fiber scale. Comput Methods Appl Mech Eng 172(1–4):27–77MATHCrossRef
18.
Zurück zum Zitat Babuška I, Motamed M (2016) A fuzzy-stochastic multiscale model for fiber composites: a one-dimensional study. Comput Methods Appl Mech Eng 302:109–130MathSciNetMATHCrossRef Babuška I, Motamed M (2016) A fuzzy-stochastic multiscale model for fiber composites: a one-dimensional study. Comput Methods Appl Mech Eng 302:109–130MathSciNetMATHCrossRef
19.
Zurück zum Zitat Shafer G (1976) A mathematical theory of evidence, vol 42. Princeton University Press, PrincetonMATHCrossRef Shafer G (1976) A mathematical theory of evidence, vol 42. Princeton University Press, PrincetonMATHCrossRef
20.
Zurück zum Zitat Matheron G (1974) Random sets and integral geometry. Wiley, New York, p 261 Matheron G (1974) Random sets and integral geometry. Wiley, New York, p 261
22.
23.
Zurück zum Zitat Balch MS, Martin R, Ferson S (2019) Satellite conjunction analysis and the false confidence theorem. Proc R Soc A 475(2227):20180565MathSciNetMATHCrossRef Balch MS, Martin R, Ferson S (2019) Satellite conjunction analysis and the false confidence theorem. Proc R Soc A 475(2227):20180565MathSciNetMATHCrossRef
24.
Zurück zum Zitat Drieschner M, Matthies HG, Hoang T-V, Rosić BV, Ricken T, Henning C, Ostermeyer G-P, Müller M, Brumme S, Srisupattarawanit T et al (2019) Analysis of polymorphic data uncertainties in engineering applications. GAMM-Mitteilungen 42(2):201900010MathSciNetCrossRef Drieschner M, Matthies HG, Hoang T-V, Rosić BV, Ricken T, Henning C, Ostermeyer G-P, Müller M, Brumme S, Srisupattarawanit T et al (2019) Analysis of polymorphic data uncertainties in engineering applications. GAMM-Mitteilungen 42(2):201900010MathSciNetCrossRef
25.
Zurück zum Zitat Drieschner M, Petryna Y (2019) Acquisition of polymorphic uncertain data based on computer tomographic scans and integration in numerical models of adhesive bonds. Preprint-Reihe des Fachgebiets Statik und Dynamik, Technische Universität Berlin Drieschner M, Petryna Y (2019) Acquisition of polymorphic uncertain data based on computer tomographic scans and integration in numerical models of adhesive bonds. Preprint-Reihe des Fachgebiets Statik und Dynamik, Technische Universität Berlin
26.
27.
Zurück zum Zitat Moens D, Hanss M (2011) Non-probabilistic finite element analysis for parametric uncertainty treatment in applied mechanics: recent advances. Finite Elem Anal Des 47(1):4–16CrossRef Moens D, Hanss M (2011) Non-probabilistic finite element analysis for parametric uncertainty treatment in applied mechanics: recent advances. Finite Elem Anal Des 47(1):4–16CrossRef
28.
Zurück zum Zitat Möller B, Beer M (2008) Engineering computation under uncertainty-capabilities of non-traditional models. Comput Struct 86(10):1024–1041CrossRef Möller B, Beer M (2008) Engineering computation under uncertainty-capabilities of non-traditional models. Comput Struct 86(10):1024–1041CrossRef
29.
Zurück zum Zitat Möller B, Graf W, Beer M, Sickert J-U (2002) Fuzzy randomness-towards a new modeling of uncertainty. In: Fifth world congress on computational mechanics, Vienna, Austria Möller B, Graf W, Beer M, Sickert J-U (2002) Fuzzy randomness-towards a new modeling of uncertainty. In: Fifth world congress on computational mechanics, Vienna, Austria
31.
Zurück zum Zitat Boutin C (2019) Homogenization methods and generalized continua in linear elasticity. In: Altenbach H, Oechsner A (eds) Encyclopedia of continuum mechanics. Springer, Berlin Boutin C (2019) Homogenization methods and generalized continua in linear elasticity. In: Altenbach H, Oechsner A (eds) Encyclopedia of continuum mechanics. Springer, Berlin
32.
Zurück zum Zitat Sánchez-Palencia E (1980) Non-homogeneous media and vibration theory, vol 127. Lecture notes in physics. Springer, BerlinMATH Sánchez-Palencia E (1980) Non-homogeneous media and vibration theory, vol 127. Lecture notes in physics. Springer, BerlinMATH
33.
Zurück zum Zitat Moravec F, Roman S (2009) Numerical computing of elastic homogenized coefficients for periodic fibrous tissue. Appl Comput Mech 3:141–152 Moravec F, Roman S (2009) Numerical computing of elastic homogenized coefficients for periodic fibrous tissue. Appl Comput Mech 3:141–152
34.
Zurück zum Zitat Ostoja-Starzewski M, Schulte J (1996) Bounding of effective thermal conductivities of multiscale materials by essential and natural boundary conditions. Phys Rev B 54(1):278CrossRef Ostoja-Starzewski M, Schulte J (1996) Bounding of effective thermal conductivities of multiscale materials by essential and natural boundary conditions. Phys Rev B 54(1):278CrossRef
35.
Zurück zum Zitat Le Bris C, Legoll F, Minvielle W (2016) Special quasirandom structures: a selection approach for stochastic homogenization. Monte Carlo Methods Appl 22(1):25–54MathSciNetMATH Le Bris C, Legoll F, Minvielle W (2016) Special quasirandom structures: a selection approach for stochastic homogenization. Monte Carlo Methods Appl 22(1):25–54MathSciNetMATH
36.
Zurück zum Zitat Christensen RM (2012) Mechanics of composite materials. Courier Corporation, North Chelmsford Christensen RM (2012) Mechanics of composite materials. Courier Corporation, North Chelmsford
37.
Zurück zum Zitat Pecullan S, Gibiansky L, Torquato S (1999) Scale effects on the elastic behavior of periodic and hierarchical two-dimensional composites. J Mech Phys Solids 47(7):1509–1542MathSciNetMATHCrossRef Pecullan S, Gibiansky L, Torquato S (1999) Scale effects on the elastic behavior of periodic and hierarchical two-dimensional composites. J Mech Phys Solids 47(7):1509–1542MathSciNetMATHCrossRef
38.
Zurück zum Zitat Mandel J (1965) Généralisation de la théorie de plasticité de wt koiter. Int J Solids Struct 1(3):273–295CrossRef Mandel J (1965) Généralisation de la théorie de plasticité de wt koiter. Int J Solids Struct 1(3):273–295CrossRef
39.
Zurück zum Zitat Antonelli A, Desmorat B, Kolev B, Desmorat R (2022) Distance to plane elasticity orthotropy by Euler–Lagrange method. C R Mécanique 350(G2):413–430CrossRef Antonelli A, Desmorat B, Kolev B, Desmorat R (2022) Distance to plane elasticity orthotropy by Euler–Lagrange method. C R Mécanique 350(G2):413–430CrossRef
40.
Zurück zum Zitat Kundur D, Hatzinakos D (1996) Blind image deconvolution. IEEE Signal Process Mag 13(3):43–64CrossRef Kundur D, Hatzinakos D (1996) Blind image deconvolution. IEEE Signal Process Mag 13(3):43–64CrossRef
41.
Zurück zum Zitat Sondhi MM (1972) Image restoration: the removal of spatially invariant degradations. Proc IEEE 60(7):842–853CrossRef Sondhi MM (1972) Image restoration: the removal of spatially invariant degradations. Proc IEEE 60(7):842–853CrossRef
42.
Zurück zum Zitat Cannon M (1976) Blind deconvolution of spatially invariant image blurs with phase. IEEE Trans Acoust Speech Signal Process 24(1):58–63CrossRef Cannon M (1976) Blind deconvolution of spatially invariant image blurs with phase. IEEE Trans Acoust Speech Signal Process 24(1):58–63CrossRef
43.
Zurück zum Zitat Hunt BR (1973) The application of constrained least squares estimation to image restoration by digital computer. IEEE Trans Comput 100(9):805–812CrossRef Hunt BR (1973) The application of constrained least squares estimation to image restoration by digital computer. IEEE Trans Comput 100(9):805–812CrossRef
44.
45.
Zurück zum Zitat Lam EY, Goodman JW (2000) Iterative statistical approach to blind image deconvolution. JOSA A 17(7):1177–1184CrossRef Lam EY, Goodman JW (2000) Iterative statistical approach to blind image deconvolution. JOSA A 17(7):1177–1184CrossRef
46.
Zurück zum Zitat Levin A, Weiss Y, Durand F, Freeman WT (2011) Understanding blind deconvolution algorithms. IEEE Trans Pattern Anal Mach Intell 33(12):2354–2367CrossRef Levin A, Weiss Y, Durand F, Freeman WT (2011) Understanding blind deconvolution algorithms. IEEE Trans Pattern Anal Mach Intell 33(12):2354–2367CrossRef
47.
Zurück zum Zitat Ruiz P, Zhou X, Mateos J, Molina R, Katsaggelos AK (2015) Variational bayesian blind image deconvolution: A review. Digit Signal Process 47:116–127MathSciNetCrossRef Ruiz P, Zhou X, Mateos J, Molina R, Katsaggelos AK (2015) Variational bayesian blind image deconvolution: A review. Digit Signal Process 47:116–127MathSciNetCrossRef
49.
Zurück zum Zitat Holtz S, Rohwedder T, Schneider R (2012) The alternating linear scheme for tensor optimization in the tensor train format. SIAM J Sci Comput 34(2):683–713MathSciNetMATHCrossRef Holtz S, Rohwedder T, Schneider R (2012) The alternating linear scheme for tensor optimization in the tensor train format. SIAM J Sci Comput 34(2):683–713MathSciNetMATHCrossRef
51.
Zurück zum Zitat Geuzaine C, Remacle J-F (2009) Gmsh: a 3-d finite element mesh generator with built-in pre-and post-processing facilities. Int J Numer Methods Eng 79(11):1309–1331MathSciNetMATHCrossRef Geuzaine C, Remacle J-F (2009) Gmsh: a 3-d finite element mesh generator with built-in pre-and post-processing facilities. Int J Numer Methods Eng 79(11):1309–1331MathSciNetMATHCrossRef
55.
Zurück zum Zitat Virtanen P, Gommers R, Oliphant TE, Haberland M, Reddy T, Cournapeau D, Burovski E, Peterson P, Weckesser W, Bright J, van der Walt SJ, Brett M, Wilson J, Millman KJ, Mayorov N, Nelson ARJ, Jones E, Kern R, Larson E, Carey CJ, Polat İ, Feng Y, Moore EW, VanderPlas J, Laxalde D, Perktold J, Cimrman R, Henriksen I, Quintero EA, Harris CR, Archibald AM, Ribeiro AH, Pedregosa F, van Mulbregt P (2020) SciPy 1.0 contributors: SciPy 1.0: fundamental algorithms for scientific computing in python. Nat Methods 17:261–272. https://doi.org/10.1038/s41592-019-0686-2 Virtanen P, Gommers R, Oliphant TE, Haberland M, Reddy T, Cournapeau D, Burovski E, Peterson P, Weckesser W, Bright J, van der Walt SJ, Brett M, Wilson J, Millman KJ, Mayorov N, Nelson ARJ, Jones E, Kern R, Larson E, Carey CJ, Polat İ, Feng Y, Moore EW, VanderPlas J, Laxalde D, Perktold J, Cimrman R, Henriksen I, Quintero EA, Harris CR, Archibald AM, Ribeiro AH, Pedregosa F, van Mulbregt P (2020) SciPy 1.0 contributors: SciPy 1.0: fundamental algorithms for scientific computing in python. Nat Methods 17:261–272. https://​doi.​org/​10.​1038/​s41592-019-0686-2
56.
Zurück zum Zitat Peng T, Saxena A, Goebel K, Xiang Y, Sankararaman S, Liu Y (2013) A novel Bayesian imaging method for probabilistic delamination detection of composite materials. Smart Mater Struct 22(12):125019CrossRef Peng T, Saxena A, Goebel K, Xiang Y, Sankararaman S, Liu Y (2013) A novel Bayesian imaging method for probabilistic delamination detection of composite materials. Smart Mater Struct 22(12):125019CrossRef
57.
Zurück zum Zitat Yan G, Sun H, Büyüköztürk O (2017) Impact load identification for composite structures using Bayesian regularization and unscented Kalman filter. Struct Control Health Monit 24(5):1910CrossRef Yan G, Sun H, Büyüköztürk O (2017) Impact load identification for composite structures using Bayesian regularization and unscented Kalman filter. Struct Control Health Monit 24(5):1910CrossRef
58.
Zurück zum Zitat Dimitrov N, Friis-Hansen P, Berggreen C (2013) Reliability analysis of a composite wind turbine blade section using the model correction factor method: numerical study and validation. Appl Compos Mater 20(1):17–39CrossRef Dimitrov N, Friis-Hansen P, Berggreen C (2013) Reliability analysis of a composite wind turbine blade section using the model correction factor method: numerical study and validation. Appl Compos Mater 20(1):17–39CrossRef
59.
Zurück zum Zitat Zhou X, Gosling P, Ullah Z, Pearce C et al (2016) Exploiting the benefits of multi-scale analysis in reliability analysis for composite structures. Compos Struct 155:197–212CrossRef Zhou X, Gosling P, Ullah Z, Pearce C et al (2016) Exploiting the benefits of multi-scale analysis in reliability analysis for composite structures. Compos Struct 155:197–212CrossRef
60.
Zurück zum Zitat Pivovarov D, Steinmann P (2016) Modified SFEM for computational homogenization of heterogeneous materials with microstructural geometric uncertainties. Comput Mech 57(1):123–147MathSciNetMATHCrossRef Pivovarov D, Steinmann P (2016) Modified SFEM for computational homogenization of heterogeneous materials with microstructural geometric uncertainties. Comput Mech 57(1):123–147MathSciNetMATHCrossRef
61.
Zurück zum Zitat Pilania G, Wang C, Jiang X, Rajasekaran S, Ramprasad R (2013) Accelerating materials property predictions using machine learning. Sci Rep 3(1):1–6CrossRef Pilania G, Wang C, Jiang X, Rajasekaran S, Ramprasad R (2013) Accelerating materials property predictions using machine learning. Sci Rep 3(1):1–6CrossRef
62.
Zurück zum Zitat Vasilyeva M, Tyrylgin A (2021) Machine learning for accelerating macroscopic parameters prediction for poroelasticity problem in stochastic media. Comput Math Appl 84:185–202MathSciNetMATHCrossRef Vasilyeva M, Tyrylgin A (2021) Machine learning for accelerating macroscopic parameters prediction for poroelasticity problem in stochastic media. Comput Math Appl 84:185–202MathSciNetMATHCrossRef
63.
64.
Zurück zum Zitat Yvonnet J, Monteiro E, He Q-C (2013) Computational homogenization method and reduced database model for hyperelastic heterogeneous structures. Int J Multiscale Comput Eng 11(3):201–225 Yvonnet J, Monteiro E, He Q-C (2013) Computational homogenization method and reduced database model for hyperelastic heterogeneous structures. Int J Multiscale Comput Eng 11(3):201–225
Metadaten
Titel
Numerical upscaling of parametric microstructures in a possibilistic uncertainty framework with tensor trains
verfasst von
Martin Eigel
Robert Gruhlke
Dieter Moser
Lars Grasedyck
Publikationsdatum
27.12.2022
Verlag
Springer Berlin Heidelberg
Erschienen in
Computational Mechanics / Ausgabe 4/2023
Print ISSN: 0178-7675
Elektronische ISSN: 1432-0924
DOI
https://doi.org/10.1007/s00466-022-02261-z

Weitere Artikel der Ausgabe 4/2023

Computational Mechanics 4/2023 Zur Ausgabe

Neuer Inhalt