TRANSFORMER MEETS BOUNDARY VALUE INVERSE PROBLEMS

Abstract

A Transformer-based deep direct sampling method is proposed for electrical impedance tomography, a well-known severely ill-posed nonlinear boundary value inverse problem. A real-time reconstruction is achieved by evaluating the learned inverse operator between carefully designed data and the reconstructed images. An effort is made to give a specific example to a fundamental question: whether and how one can benefit from the theoretical structure of a mathematical problem to develop task-oriented and structure-conforming deep neural networks? Specifically, inspired by direct sampling methods for inverse problems, the 1D boundary data in different frequencies are preprocessed by a partial differential equation-based feature map to yield 2D harmonic extensions as different input channels. Then, by introducing learnable non-local kernels, the direct sampling is recast to a modified attention mechanism. The new method achieves superior accuracy over its predecessors and contemporary operator learners and shows robustness to noises in benchmarks. This research shall strengthen the insights that, despite being invented for natural language processing tasks, the attention mechanism offers great flexibility to be modified in conformity with the a priori mathematical knowledge, which ultimately leads to the design of more physics-compatible neural architectures.

1. INTRODUCTION

Boundary value inverse problems aim to recover the internal structure or distribution of multiple media inside an object (2D reconstruction) based on only the data available on the boundary (1D signal input), which arise from many imaging techniques, e.g., electrical impedance tomography (EIT) (Holder, 2004) , diffuse optical tomography (DOT) (Culver et al., 2003) , magnetic induction tomography (MIT) (Griffiths et al., 1999) . Not needing any internal data renders these techniques generally non-invasive, safe, cheap, and thus quite suitable for monitoring applications. In this work, we shall take EIT as an example to illustrate how a more structure-conforming neural network architecture leads to better results in certain physics-based tasks. Given a 2D bounded domain Ω and an inclusion D, the forward model is the following partial differential equation (PDE) ∇ • (σ∇u) = 0 in Ω, where σ = σ 1 in D, and σ = σ 0 in Ω\D, where σ is a piecewise constant function defined on Ω with known function values σ 0 and σ 1 , but the shape of the inclusion D buried in Ω is unknown. The goal is to recover the shape of D using only the boundary data on ∂Ω (Figure 1 ). Specifically, by exerting a current g on the boundary, one solves (1) with the Neumann boundary condition σ∇u • n| ∂Ω = g, where n is the outwards unit normal direction of ∂Ω, to get a unique u on the whole domain Ω. In practice, only the Dirichlet boundary value representing the voltages f = u| ∂Ω on the boundary can be measured. This procedure is called Neumann-to-Dirichlet (NtD) mapping: Λ σ : H -1/2 (∂Ω) → H 1/2 (∂Ω), with g = σ∇u • n| ∂Ω → f = u| ∂Ω . (2)

