Abstract.

In recent years, deep learning techniques have shown great success in various tasks related to inverse problems, where a target quantity of interest can only be observed through indirect measurements by a forward operator. Common approaches apply deep neural networks in a postprocessing step to the reconstructions obtained by classical reconstruction methods. However, the latter methods can be computationally expensive and introduce artifacts that are not present in the measured data and, in turn, can deteriorate the performance on the given task. To overcome these limitations, we propose a class of equivariant neural networks that can be directly applied to the measurements to solve the desired task. To this end, we build appropriate network structures by developing layers that are equivariant with respect to data transformations induced by well-known symmetries in the domain of the forward operator. We rigorously analyze the relation between the measurement operator and the resulting group representations and prove a representer theorem that characterizes the class of linear operators that translate between a given pair of group actions. Based on this theory, we extend the existing concepts of Lie group equivariant deep learning to inverse problems and introduce new representations that result from the involved measurement operations. This allows us to efficiently solve classification, regression, or even reconstruction tasks based on indirect measurements also for very sparse data problems, where a classical reconstruction-based approach may be hard or even impossible. We illustrate the effectiveness of our approach in numerical experiments and compare with existing methods.

Keywords

  1. equivariant deep learning
  2. operator equivariance
  3. inverse problems

MSC codes

  1. 47B38
  2. 65J22
  3. 94A05

Get full access to this article

View all available purchase options and get full access to this article.

References

1.
S. Arridge, P. Maass, O. Öktem, and C.-B. Schönlieb, Solving inverse problems using data-driven models, Acta Numer., 28 (2019), pp. 1–174, https://doi.org/10.1017/S0962492919000059.
2.
E. Celledoni, M. J. Ehrhardt, C. Etmann, B. Owren, C.-B. Schönlieb, and F. Sherry, Equivariant neural networks for inverse problems, Inverse Problems, 37 (2021), 085006, https://doi.org/10.1088/1361-6420/ac104f.
3.
D. Chen, M. E. Davies, M. J. Ehrhardt, C.-B. Schönlieb, F. Sherry, and J. Tachella, Imaging with equivariant deep learning: From unrolled network design to fully unsupervised learning, IEEE Signal Process. Mag., 40 (2023), pp. 134–147, https://doi.org/10.1109/MSP.2022.3205430.
4.
D. Chen, J. Tachella, and M. E. Davies, Equivariant imaging: Learning beyond the range space, in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2021, pp. 4359–4368, https://doi.org/10.1109/ICCV48922.2021.00434.
5.
N. Chenouard and M. Unser, 3D steerable wavelets in practice, IEEE Trans. Image Process., 21 (2012), pp. 4522–4533, https://doi.org/10.1109/TIP.2012.2206044.
6.
T. S. Cohen, M. Geiger, and M. Weiler, A general theory of equivariant CNNs on homogeneous spaces, in Proceedings of Advances in Neural Information Processing Systems 32 (NeurIPS 2019), 2019.
7.
T. S. Cohen and M. Welling, Group equivariant convolutional networks, in Proceedings of the 33th International Conference on Machine Learning, 2016, pp. 2990–2999.
8.
T. S. Cohen and M. Welling, Steerable CNNs, in Proceedings of the International Conference on Learning Representations, 2017.
9.
M. Dax, S. R. Green, J. Gair, M. Deistler, B. Schölkopf, and J. H. Macke, Group equivariant neural posterior estimation, in Proceedings of the 10th International Conference on Learning Representations (ICLR), 2022, pp. 1–24.
10.
L. Deng, The MNIST database of handwritten digit images for machine learning research, IEEE Signal Process. Mag., 29 (2012), pp. 141–142.
11.
S. Dittmer, D. Erzmann, H. Harms, and P. Maass, SELTO: Sample-Efficient Learned Topology Optimization, 2022, https://arxiv.org/abs/2209.05098.
12.
K. Doherty, C. Simpson, S. Becker, and A. Doostan, QuadConv: Quadrature-Based Convolutions with Applications to Non-uniform PDE Data Compression, https://arxiv.org/abs/2211.05151, 2022.
13.
H. W. Engl, M. Hanke, and A. Neubauer, Regularization of Inverse Problems, Math. Appl. 375, Kluwer, Dordrecht, 2000.
14.
M. Finzi, S. Stanton, P. Izmailov, and A. G. Wilson, Generalizing convolutional neural networks for equivariance to Lie groups on arbitrary continuous data, in Proceedings of the 37th International Conference on Machine Learning, 2020, pp. 3165–3176.
15.
K. Fukushima, Cognitron: A self-organizing multilayered neural network, Biol. Cybernet., 20 (1975), pp. 121–136, https://doi.org/10.1007/BF00342633.
16.
I. M. Gel’fand and G. E. Shilov, Generalized Functions. Volume I: Properties and Operations, Academic Press, New York, 1964.
17.
K. He, X. Zhang, S. Ren, and J. Sun, Deep residual learning for image recognition, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015, pp. 770–778.
18.
L. Hörmander, The Analysis of Linear Partial Differential Operators I: Distribution Theory and Fourier Analysis, 2nd ed., Grundlehren Math. Wiss. 256, Springer, Berlin, 1990.
19.
S. Ioffe and C. Szegedy, Batch normalization: Accelerating deep network training by reducing internal covariate shift, in Proceedings of the 32nd International Conference on Machine Learning, 2015, pp. 448–456.
20.
J. P. Kaipio and E. Somersalo, Statistical and Computational Inverse Problems, Appl. Math. Sci. 160, Springer, New York, 2005.
21.
D. P. Kingma and J. Ba, Adam: A Method for Stochastic Optimization, https://arxiv.org/abs/1412.6980, 2014.
22.
R. Kondor and S. Trivedi, On the generalization of equivariance and convolution in neural networks to the action of compact groups, in Proceedings of the 35th International Conference on Machine Learning, 2018, pp. 2747–2755.
23.
H. Larochelle, D. Erhan, A. Courville, J. Bergstra, and Y. Bengio, An empirical evaluation of deep architectures on problems with many factors of variation, in Proceedings of the 24th International Conference on Machine Learning, 2007, pp. 473–480, https://doi.org/10.1145/1273496.1273556.
24.
Y. LeCun, L. D. Jackel, L. Bottou, C. Cortes, J. S. Denker, H. Drucker, I. Guyon, U. A. Müller, E. Säckinger, P. Simard, and V. Vapnik, Learning algorithms for classification: A comparison on handwritten digit recognition, in Neural Networks: The Statistical Mechanics Perspective, World Scientific, Englewood Cliffs, NJ, 1995, pp. 261–276.
25.
F. Natterer and F. Wübbeling, Mathematical Methods in Image Reconstruction, Math. Model. Comput. 5, SIAM, Philadelphia, 2001.
26.
D. Otero Baguer, J. Leuschner, and M. Schmidt, Computed tomography reconstruction using deep image prior and learned reconstruction methods, Inverse Problems, 36 (2020), 094004, https://doi.org/10.1088/1361-6420/aba415.
27.
A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga, A. Desmaison, A. Kopf, E. Yang, Z. DeVito, M. Raison, A. Tejani, S. Chilamkurthy, B. Steiner, L. Fang, J. Bai, and S. Chintala, PyTorch: An imperative style, high-performance deep learning library, in Proceedings of Advances in Neural Information Processing Systems 32 (NeurIPS 2019), 2019.
28.
P. Ramachandran, B. Zoph, and Q. V. Le, Searching for Activation Functions, https://arxiv.org/abs/1710.05941, 2017.
29.
J. Tang, Equivariance Regularization for Image Reconstruction, https://arxiv.org/abs/2202.05062, 2022.
30.
F. Treves, Differentiable functions with values in topological vector spaces. Tensor product of distributions, in Topological Vector Spaces, Distributions and Kernels, Pure Appl. Math. 25, Elsevier, Amsterdam, 1967, pp. 411–419, https://doi.org/10.1016/S0079-8169(08)60289-5.
31.
W. van Aarle, W. J. Palenstijn, J. Cant, E. Janssens, F. Bleichrodt, A. Dabravolski, J. De Beenhouwer, K. J. Batenburg, and J. Sijbers, Fast and flexible X-ray tomography using the ASTRA toolbox, Opt. Express, 24 (2016), pp. 25129–25147, https://doi.org/10.1364/OE.24.025129.
32.
W. van Aarle, W. J. Palenstijn, J. De Beenhouwer, T. Altantzis, S. Bals, K. J. Batenburg, and J. Sijbers, The ASTRA toolbox: A platform for advanced algorithm development in electron tomography, Ultramicroscopy, 157 (2015), pp. 35–47, https://doi.org/10.1016/j.ultramic.2015.05.002.
33.
M. Winkels and T. S. Cohen, Pulmonary nodule detection in CT scans with equivariant CNNs, Med. Image Anal., 55 (2019), pp. 15–26, https://doi.org/10.1016/j.media.2019.03.010.

Information & Authors

Information

Published In

cover image SIAM Journal on Mathematics of Data Science
SIAM Journal on Mathematics of Data Science
Pages: 579 - 601
ISSN (online): 2577-0187

History

Submitted: 29 June 2023
Accepted: 13 March 2024
Published online: 3 July 2024

Keywords

  1. equivariant deep learning
  2. operator equivariance
  3. inverse problems

MSC codes

  1. 47B38
  2. 65J22
  3. 94A05

Authors

Affiliations

Matthias Beckmann Contact the author
Center for Industrial Mathematics, University of Bremen, Bremen, Germany, and Department of Electrical and Electronic Engineering, Imperial College London, London, UK.
Nick Heilenkötter
Center for Industrial Mathematics, University of Bremen, Bremen, Germany.

Funding Information

Funding: The second author received funding from the Deutsche Forschungsgemeinschaft (DFG), project 281474342.

Metrics & Citations

Metrics

Citations

If you have the appropriate software installed, you can download article citation data to the citation manager of your choice. Simply select your manager software from the list below and click Download.

Cited By

View Options

View options

PDF

View PDF

Full Text

View Full Text

Figures

Tables

Media

Share

Share

Copy the content Link

Share with email

Email a colleague

Share on social media