deep inversion, autoencoders for learned regularization of ... · c3: non-linearity; c4:...
TRANSCRIPT
/ 43
Deep Inversion, Autoencoders for LearnedRegularization of Inverse Problems
Yoeri Boink*†, Srirang Manohar†, Leonie Zeune*‡, Leon Terstappen‡, Stephan van Gils*, Christoph Brune*Biomedical Photonic Imaging†
Medical Cell Biophysics‡Applied Mathematics*
08-06-2018 Regularisation methods for PAT 1
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
OUTLINE
2
1. Robustness of learned primal-dual (L-PD) reconstruction in photoacoustic tomography (PAT)
2. Functional learning with an unrolled gradient descent (GD) scheme
guaranteed convergence and stability!
3. Learn latent representation of data space and image space via variational autoencoder (VAE)
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
Classical Inverse ProblemsParameter Estimation
VARIATIONAL METHODS AND DEEP LEARNING
3
Partial Differential Equations
Deep VariationalNetworks
MeasurementsHidden Parameters
ComplexData Science
Generalization?Regularization Theory?
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
VARIATIONAL METHODS AND DEEP LEARNING
4
Chen, Pock - Trainable Nonlinear Reaction Diffusion, 2016
Normsnonconvex
Differential operators
Scale-space,Harmonic analysis
Regularization theory
Inverse problems
Vector fields,multimodality
Time dependent modeling
Activation functions
ReLU, sigmoid
Convolutionsfunctions per
layer
Scattering networks
Generalization properties
GANsVAEs?
Multiple populations?
Residual?Skip
connections?Deep networks
Variational methods
Deep Residual Neural Networksare connected to
Partial Differential Equations
Chen et al - Neural Ordinary Differential Equations, 2018
Ciccone et al - Stable Deep Networks from Non-Autonomous DEs, 2018Mallat - Understanding deep convolutional networks, 2016
Haber, Ruthotto - Stable architectures for deep neural networks, 2018
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
bi-level,
non-convex
large-scale, real-time
training data,
generalization
accurate,
stable,
reproducible
uncertainty,
parameters
Optimization
Regularization
Architecture
MATHEMATICS OF DEEP LEARNING
5
DL
Vidal et al. Mathematics of Deep Learning, 2018
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
DEEPER INSIGHTS INTO DEEP INVERSION
6
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
4TU PROPOSALPRECISION MEDICINE 2018
Challenges: C1: sparse/big data; C2: multi-dimensionality, heterogeneityC3: non-linearity; C4: super-resolution
Deep learning for model-driven imaging
“classical” physics-based reconstruction enriched by deep learning → latent operator parameters, robustness
“black-box” deep learning enriched by physical constraints → more natural, latent data structures, robustness
naturalrobust
performancesimple
7
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
FINDING NEEDLES IN A HAYSTACK
8
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
FINDING NEEDLES IN A HAYSTACK
9
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
Denoising and Scale Analysis by Local Diffusion
• given noisy input• denoise image by minimizing the Total Variation (TV) flow
→Time-discrete case: solving in every step the ROF [Rudin,92] problem:
10
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
Denoising and Scale Analysis by Local Diffusion
Idea: Solution of nonlinear eigenvalue problem
transformed to a peak in the spectral domain
→ time-point t where the eigenfunctions are completely removed depends on size and height of the disc, thus scale indicator
[Gilboa 2013,2014],[Horesh, Gilboa 2015],[Burger et al., 2015,2016][Aujol, Gilboa, Papadakis, 2015, 2017]
11
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
Contrast Invariance of L1-TV Denoising
12
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
Zeune et al - Multiscale Segmentation via Bregman Distances and Nonlinear Spectral Analysis, 2017
13
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
CNN CLASSIFICATION FOR CANCER-ID
14
Convolution network
with
max-pooling layers Fully-
connected
layers
Input:
image with
three
fluorescent
channels
32
64
128CTC probability
No CTC probability
0.9722 0.0278
0.010 0.990
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
• assume we have a trained AE with tied weights, i.e.• x and f(x) live in the same vector space, i.e.
is a vector field pointing from x to reconstructed f(x)
• under some (fulfilled) assumptions, G(x) is the gradient field of an energy E(x)
with
AUTOENCODERS AND GRADIENT FLOWS
15
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
Convolutional AETrained on no noise data, 6000 training sets, 1000 test sets3 Convolution (32 filters in total) + Pooling blocks for Encoder + Decoder → 4963 parameters
GT
Std.dev 0.5
Std.dev 0.2
Std.dev 0.1
Std.dev 0.05
No noise
16
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
Convolutional AETrained on noisy data (std dev 0.2), 6000 training sets, 1000 test sets3 Convolution (32 filters in total) + Pooling blocks for Encoder + Decoder → 4963 parameters
GT
Std.dev 0.5
Std.dev 0.2
Std.dev 0.1
Std.dev 0.05
No noise
17
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
Convolutional AETrained on no noise data, 6000 training sets, 1000 test sets3 Convolution (32 filters in total) + Pooling blocks for Encoder + Decoder → 4963 parameters
GT
No noise
Different Radius
Different Radius,Trained on correct
18
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
DEEP LEARNING OF CIRCULATING TUMOR CELLS
Leonie Zeune
19
Zeune et al - Deep learning for tumor cell classification (2019)
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
TUMOUR
BREAST
ULTRASOUND
DETECTORS
PHOTOACOUSTIC BREAST IMAGING(ILLUSTRATIONS MADE BY SJOUKJE SCHOUSTRA – BMPI GROUP, UNI TWENTE)
20
LASER
Folkman, 1996
angiogenesis
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
TUMOUR
BREAST
PULSEo LIGHT ABSORPTION
o TEMPERATURE RISE
o EXPANSION
o PRESSURE RISE
o ULTRASOUND WAVE
o SIGNALS
o RECONSTRUCTION
ULTRASOUND
DETECTORS
PHOTOACOUSTIC EFFECT
21
LASER
Folkman, 1996
angiogenesis
PHOTOACOUSTIC BREAST IMAGING(ILLUSTRATIONS MADE BY SJOUKJE SCHOUSTRA – BMPI GROUP, UNI TWENTE)
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
We make use of a projection model with calibration2:
2 Wang, Xing, Zeng, Chen - Photoacoustic imaging with deconvolution (2004)
1 Van Es, Vlieg, Biswas, Hondebrink, Van Hespen, Moens, Steenbergen, Manohar - Coregistered photoacoustic and ultrasound tomography of healthy and inflamed human interphalangeal joints (2015)
PHOTOACOUSTIC TOMOGRAPHY
22
Figure: 2D slice-based imaging with rotating fibres and sensor array.1
PAT-operator
Folkman, 1996
angiogenesis
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
INVERSE RECONSTRUCTION METHODS
23
FBP
operator
data reconstruction▪ Direct reconstruction: filtered backprojection (FBP)
▪ Iterative reconstruction: total variation (TV)
(solved with PDHG)
▪ Learned post-processing: U-Net 3
PDHG
operator
data reconstruction
reconstructionFBP
operator
dataU-Net
3 Jin, McCann, Froustey, Unser - Deep Convolutional Neural Network for Inverse Problems in Imaging (2017)
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
MODEL-BASED REGULARISED RECONSTRUCTION -CHOOSING FUNCTION SPACES
Kingsbury - The dual-tree complex wavelet transform a new efficient tool for image restauration and enhancement (1998)
Rudin, Osher, Fatemi - Nonlinear total variation based noise removal algorithms (1992)
Bredies, Kunisch, Pock - Total Generalised Variation (2010)
Boink, Lagerwerf, Steenbergen, van Gils, Manohar, Brune - A framework for directional and higher-order reconstruction in photoacoustic tomography (2018)
?
24
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
FROM MODEL-DRIVEN TO DATA-DRIVEN
Hauptmann, Lucka, Betcke, Huynh, Cox, Beard, Ourselin, Arridge - Model based learning for accelerated, limited-view 3D photoacoustic tomography (2017)
▪ No regularisation parameter;
▪ Better robustness to noise;
▪ Faster reconstruction.
However,
▪ No proven stability or convergence.
Meinhardt, Möller, Hazirbas, Cremers - Learning Proximal Operators: Using Denoising Networks for Regularizing Inverse Imaging Problems (2017)
Adler, Öktem – Learned Primal-Dual Reconstruction (2017)
25
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
=
=
+
CNN
+
+
26
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
MODEL CHOICES PAT SIMULATION AND RECONSTRUCTION NETWORK
Training
Resolution 1.5625 mm
Number of pixels 192x192
Number of sensors 32
• 768 training images• 192 test images• scaled between 0 and 1
DRIVE dataset: Staal, Abramoff, Niemeijer, Viergever, Ginneken -Ridge based vessel segmentation in color images of the retina (2004)
‘large’ network
smallnetwork
# primal-dual iterations (N) 10 5
# primal/dual channels (k) 5 2
# hidden layers 2 2
# channels in hidden layers 32 32
activation functions ReLU ReLU
filter size convolutions 3x3 3x3
27
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
ROBUSTNESS TO IMAGE CHANGES
28
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
ROBUSTNESS TO IMAGE CHANGES
▪ Strong noise removal and background identification;
▪ L-PD is robust against many changes in image;
▪ Training with more variety in data has positive effect on quality.
29
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
Initial pressureSegmentation
INCLUDING HIGHER-LEVEL TASK: SEGMENTATION
30
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
INCLUDING HIGHER-LEVEL TASK: SEGMENTATION
reconstruction
segmentation
▪ Learned post-processing: U-Net
▪ joint reconstruction and segmentation with L-PD
segmentationreconstructionFBP
operator
dataU-Net
reconstruction
L-PD
operator
datasegmentation
31
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
INCLUDING HIGHER-LEVEL TASK: SEGMENTATION
Figure: Reconstructions and segmentations using a 32 detector setting.
32
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
EXPERIMENTAL RESULTS: ONE-SIDED SAMPLING
33
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion 34
EXPERIMENTAL RESULTS: UNIFORM SAMPLING
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
ROBUSTNESS TO SYSTEM CHANGES
Figure: Reconstruction from data of 64 detectors.
How to achieve generalisability towards imaging operator uncertainty?Latent structures?!
35
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
CONVERGENCE OF PARTIALLY LEARNED METHODS
5 Banert, Ringh, Adler, Karlsson, Öktem – Data-driven nonsmooth optimization (2018)
Several recent papers give convergence proofs for:
▪ methods that use explicit (Tikhonov-like) regularisation in the form of a neural network.4
▪ methods with a proximal structure.5
▪ methods where the learned part only has influence on the null-space.6
6 Schwab, Antholzer, Haltmeier – Deep Null Space Learning for Inverse Problems: Convergence Analysis and Rates (2018)
4 Li, Schwab, Antholzer, Haltmeier - NETT Solving Inverse Problems with Deep Neural Networks (2018)
Lunz, Öktem, Schönlieb – Adversarial Regularizers in Inverse Problems (2018)
36
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
LEARNED (UNROLLED) GRADIENT DESCENT
Goal: learn a nonlinear function (functional) such that its minimiser is our desired reconstruction
where ,
and Hi represents a convolutional layer.
37
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
▪
▪
▪
▪ Train network with , f or .
LEARNED (UNROLLED) GRADIENT DESCENT
Learning of well-known gradient flows possible?Diffusion, convection-diffusion, thin-film, etc.
38
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
LEARNED (UNROLLED) GRADIENT DESCENT
NN
NN NN NN
39
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
TRAINING DETAILS: COIL DATASETColumbia Object Image Library
Nene, Nayar, Murase – Columbia Object Image Library (COIL-100) (1996)
Ground truth Noisy input
40
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
▪ #GD-steps = N = 20
▪ #layers = 3
▪ #channels = 8
▪ batch size = 9
▪ #epochs = 50
▪ Adam optimiser
rate 2∙10-2 →10-3
▪ kernel size = 3x3
TRAINING PARAMETERS AND RESULT
NN NN NN
Ground truth Noisy input L-GD output
41
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
LEARNED (UNROLLED) GRADIENT DESCENT
iterations 1-5
iterations 6-10
iterations 11-15
iterations 16-20
42
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
LEARNED (UNROLLED) GRADIENT DESCENT: Q2
iterations 1-5
iterations 6-10
iterations 11-15
iterations 16-20
LEARNED“DATA FIDELITY”OPERATOR
43
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
LEARNED (UNROLLED) GRADIENT DESCENT: Q1
iterations 1-5
iterations 6-10
iterations 11-15
iterations 16-20
LEARNED“REGULARIZATION”OPERATOR
44
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
LEARNED (UNROLLED) GRADIENT FLOW
Ground truth Noisy input
L-GD output(20 iterations)
L-GD output(40 iterations)
LearningLoss
Functional
45
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
LEARNING GRADIENT FLOWS, VARIATIONAL NETWORKS (PHYSICS-INFORMED)
46
Chen, Pock – Trainable Nonlinear Reaction Diffusion (2015)
Kobler et al – Variational Networks: Connecting Variational Methods and Deep Learning (2017)
Lusch et al - Deep learning for universal linear embeddings of nonlinear dynamics (2017)
Yang et al - Physics-informed deep generative models (2018)
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
DEEP DISCRIMINATIVE VS GENERATIVE MODELS
Genevay, Peyre, Cuturi – GAN and VAE from an Optimal Transport Point of View (2017)
Mescheder, Nowozin, Geiger – Adversarial Variational Bayes: Unifying VAEs and GANs (2017)
47
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
(BLIND) DECONVOLUTION PROBLEM IN PHOTOACOUSTICSLEARNING UNCERTAINTY -CALIBRATION IN FORWARD MODEL
48
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
LEARNEDDATA-SPACE WITH VARIATIONAL AUTOENCODER
Dai, Wipf - Diagnosing and Enhancing VAE Models (2019)
49
Kingma, Welling – Auto-Encoding Variational Bayes (2013)
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
LEARNEDDATA-SPACE WITH VARIATIONAL AUTOENCODER
r
50
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
LEARNED DATA-SPACE WITH VARIATIONAL AUTOENCODER
ground truth
reconstruction
51
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
LIMITATION: GEOMETRIC VAE LATENT SPACE REGULARIZATION
Regularization of latent variables z
52
Imaging and ML, IHP Paris, Apr 3, 2019 C. Brune - Deep Inversion
Thanks for your attention
CONCLUSIONS
• Nonlinear eigenvalue problems offer spectral decompositions similar to Autoencoders
• Robustness. Deep learning improves PAT reconstruction quality particularly in uncertain cases.
• Functional learning. Mathematical theory on stability and convergence not available. Construct learned methods that learn the functional to be minimised via a gradient flow.
• Generative networks. Generative networks can be successfully used for learning latent variables in inverse problems. VAEs show geometric limitations for latent space interpolation.
Boink, van Gils, Manohar, Brune - Sensitivity of a partially learned model‐based reconstruction algorithm (2018)
Boink, Manohar, Brune - A partially learned algorithm for joint photoacoustic reconstruction and segmentation (2019)
53
Zeune et al - Multiscale Segmentation via Bregman Distances and Nonlinear Spectral Analysis (2017)
Zeune et al - Deep learning for tumor cell classification (2019)