Computation of the Response Surface in the Tensor Train data format
Sergey Dolgov, Boris N. Khoromskij, Alexander Litvinenko, and Hermann G. Matthies
Contact the author: Please use for correspondence this email.
Submission date: 12. Jun. 2014
MSC-Numbers: 15A69, 65F10, 60H15, 60H35, 65C30
Keywords and phrases: uncertainty quantification, polynomial chaos expansion, karhunen-loeve expansion, stochastic galerkin, tensor product methods, tensor train format, Adaptive Cross Approximation
We apply the Tensor Train (TT) approximation to construct the Polynomial Chaos Expansion (PCE) of a random field, and solve the stochastic elliptic diffusion PDE with the stochastic Galerkin discretization. We compare two strategies of the polynomial chaos expansion: sparse and full polynomial (multi-index) sets. In the full set, the polynomial orders are chosen independently in each variable, which provides higher flexibility and accuracy. However, the total amount of degrees of freedom grows exponentially with the number of stochastic coordinates. To cope with this curse of dimensionality, the data is kept compressed in the TT decomposition, a recurrent low-rank factorization. PCE computations on sparse grids sets are extensively studied, but the TT representation for PCE is a novel approach that is investigated in this paper. We outline how to deduce the PCE from the covariance matrix, assemble the Galerkin operator, and evaluate some post-processing (mean, variance, Sobol indices), staying within the low-rank framework. The most demanding are two stages. First, we interpolate PCE coefficients in the TT format using a few number of samples, which is performed via the block cross approximation method. Second, we solve the discretized equation (large linear system) via the alternating minimal energy algorithm. In the numerical experiments we demonstrate that the full expansion set encapsulated in the TT format is indeed preferable in cases when high accuracy and high polynomial orders are required.
The manuscript is located here.