Structured Autoencoders for Operator-theoretic decomposition and - - PowerPoint PPT Presentation

structured autoencoders for operator theoretic
SMART_READER_LITE
LIVE PREVIEW

Structured Autoencoders for Operator-theoretic decomposition and - - PowerPoint PPT Presentation

Structured Autoencoders for Operator-theoretic decomposition and Model reduction Karthik Duraisamy Thanks to Motivation Decomposition & Reduced Order Modeling of Complex Multiscale Problems [K] [W/m3] Large scale simulations O(10 6 )-


slide-1
SLIDE 1

Karthik Duraisamy

Structured Autoencoders for Operator-theoretic decomposition and Model reduction

slide-2
SLIDE 2

Thanks to…

slide-3
SLIDE 3

Motivation

Decomposition & Reduced Order Modeling of Complex Multiscale Problems

[K] [W/m3]

Large scale simulations O(106)- O(108) CPU hours / run Complex physics : Flow, turbulence, combustion, heat transfer, etc

slide-4
SLIDE 4

The Autoencoder

4

Structure: encoder (compression) + decoder (decompression) – Encoder 𝚾 𝐲; 𝛊𝚾 – Decoder 𝛀 . ; 𝛊𝛀 – POD: 𝚾 → 𝐕) · , 𝛀 → 𝐕 · – Trained as one single network, 𝛊𝚾 and 𝛊𝛀 are optimized jointly – Automatically separated into encoder and decoder by cutting at the “bottleneck”

“Applied Deep Learning” https://towardsdatascience.com/

𝚾 𝛀

˜ x = Ψ(·, θΨ) Φ(x, θΦ)

x

slide-5
SLIDE 5

Embedding (in the right coordinates)

slide-6
SLIDE 6

Part 1

Operator-theoretic Learning & Decomposition

slide-7
SLIDE 7

Koopman operator and linear embedding

slide-8
SLIDE 8

Connections of Koopman to other operators

L := f · rx ∂u ∂t = Lu ; u(·, 0) = h Kth = h φt = etLh

Liouville operator Liouville PDE Generator

hh, Ptρi = hKth, ρi

ρ(·, t) = Pt ρ

Perron-Frobenius operator Duality

Liouville generates Koopman Perron-Frobenius is adjoint of Koopman

slide-9
SLIDE 9

Spectral expansion of Koopman operators

slide-10
SLIDE 10

Koopman operators & “Deep” Learning

Several works since 2018

slide-11
SLIDE 11

Extracting a Koopman-invariant subspace

Arbabi & Mezic 2019

Goal: Extracting the Koopman operator defined on Observation functionals

We are also interested in retrieving the state x

slide-12
SLIDE 12

Enforcing structure for Learning : “Physics information”

slide-13
SLIDE 13

Enforcing structure for Learning : Tractable optimization

slide-14
SLIDE 14

“Data-free”, “Physics-informed” Trajectory data, ”Unknown physics”

slide-15
SLIDE 15

Enforcing structure for Learning : “DMD ResNet”

slide-16
SLIDE 16

Enforcing structure for Learning : Stability

slide-17
SLIDE 17

x

f(x)

Naïve “Autoencoders”

slide-18
SLIDE 18

x

f(x)

  • Pan. S. & Duraisamy, K., Physics-Informed

Probabilistic Learning of Linear Embeddings of Non-linear Dynamics With Guaranteed Stability, SIAM J. of Applied Dynamical Systems, 2020.

Putting it all together (deterministic form)

slide-19
SLIDE 19

Bayesian Neural Networks & Variational Inference

slide-20
SLIDE 20

Variational Inference

slide-21
SLIDE 21

Verification on Model dynamical systems

Duffing oscillator: Eigenfunctions (with uncertainty) Prediction and sensitivity to data 100 data points. 1000 data points 10000 data points

slide-22
SLIDE 22

Flow over cylinder: Prediction with uncertainties

  • Gaussian white noise added
slide-23
SLIDE 23

Flow over cylinder: Prediction with uncertainties

Velocity magnitude (mean) Velocity magnitude (standard deviation)

Physics-Informed Probabilistic Learning of Linear Embeddings of Non-linear Dynamics With Guaranteed Stability, Pan, S., and Duraisamy, K., SIADS, 2020

slide-24
SLIDE 24

Proposed framework

Steps: I a-priori cross validation to choose an appropriate hyperparameter I mode-by-mode error analysis I choose a trade-off between reconstruction error and linear evolving error I sparse reconstruction of system with multi-task learning

4 / 9

Multi-task learning framework to extract sparse Koopman-invariant subspaces

Sparsity-promoting algorithms for the discovery of informative Koopman invariant subspaces, Pan, S., N. A-M and Duraisamy, K., arXiv:2002.10637

slide-25
SLIDE 25

I transient behavior is accurately reconstructed I stable modes are successfully extracted from strongly nonlinear transient data I left mode: due to side edge

  • f superstructure. right

mode: due to funnel

6 / 9

Turbulent Ship Airwake

Sparsity-promoting algorithms for the discovery of informative Koopman invariant subspaces, Pan, S., N. A-M and Duraisamy, K., arXiv:2002.10637

slide-26
SLIDE 26

26

Summary

Many opportunities to enforce structure in Autoencoders è flexible and powerful tools

slide-27
SLIDE 27

Part 2

Learning Reduced Order Models of Parametric Spatio-temporal dynamics

slide-28
SLIDE 28
  • B. Kramer, K. E. Willcox, AIAA Journal ,2019
  • M. Guo, J. S. Hesthaven, CMAME, 2018.
  • A. Mohan, D. Daniel, M. Chertkov, D. Livescu, arXiv, 2019
  • S. Lee, D. You, arXiv, 2019.
  • Q. Wang, J. S. Hesthaven, D. Ray, JCP, 2019.

Non-intrusive data-driven ROMs

qn+1

l

= f(qn+1

l

, qn

l , ....qn−l l

, B(un+1), µ)

Some recent works:

slide-29
SLIDE 29

Basic Component: Convolutional Layer

29

  • Convolutional layers preserve complex

spatio-temporal “information”

  • Convolutional operation on a local window w

– 𝑦 ∗ 𝑥 ./ = ∑ ∑ 𝑦.23,/25𝑥3,5

26 576 28 378

  • Ideal for “localized” feature identification
  • Rotation and translation invariant, if

properly constructed

http://cs231n.github.io/convolutional-networks/ : “Applied Deep Learning” https://towardsdatascience.com/

slide-30
SLIDE 30

Temporal Convolutional

  • Performs dilated 1D convolutional operation in temporal/sequential direction

– 𝑦 ∗9 𝑥 . = ∑ 𝑦.293𝑥3

:2; 37<

  • Exponential increase in reception field è an increasingly popular alternative to

RNN/LSTM

Input sequence Conv-1: d=1 reception field: 2 Conv-2: d=2 reception field: 4 Conv-3: d=4 reception field: 8 Output/Conv-4: d=8 reception field: 16 Example: k = 2, d = 2i Source of image: github.com/philipperemy/keras-tcn

slide-31
SLIDE 31

Training Multi-level convolutional AE networks

slide-32
SLIDE 32

Example CAE architecture

32

slide-33
SLIDE 33

Prediction using Multilevel AE networks

slide-34
SLIDE 34

Example TCAE architecture

slide-35
SLIDE 35

Prediction using Multilevel AE networks

slide-36
SLIDE 36

Time stepping

slide-37
SLIDE 37

37

Example TCN architecture

*: Convolution direction

1 1 1 1 1 1

Input

𝐑>

.,? Jumping dilated 1D convolution Dense

𝐫>

.A; Jumping dilated 1D convolution Jumping dilated 1D convolution 𝑜? steps 𝑜> channels *

Many-to-one

Input

𝐑>

.,? Non-strided dilated 1D convolution 1D convolution

𝐫>

.A; Non-strided dilated 1D convolution 𝑜? steps

Many-to-many

𝑜> input channels *

1 1 1 1 1 1 1 1 1 1

*: Convolution direction

Example TCN architecture

slide-38
SLIDE 38

38 Component CAE reconstruction CAE + TCN (final step) Training Testing Pressure 0.04% 0.1% 0.12% Density 0.01% 0.04% 0.14% Velocity 0.04% 0.08% 0.13%

Numerical Tests: Discontinuous compressible flow

slide-39
SLIDE 39

39

Discontinuous compressible flow : Impact of data

slide-40
SLIDE 40

40

Numerical Tests : 3D Ship Airwake

– Incompressible Navier-Stokes – 576k DOF, 400 time snapshots – Global parameter: sliding angle 𝛽 – Training: 𝛽 = 5° :5° :20° – Prediction: 𝛽 = 12.5

𝛽 = 5° 𝛽 = 20°

slide-41
SLIDE 41

41

Numerical Test: 3D Ship Airwake

Component CAE reconstruction MLP + TCAE + CAE Training Testing U 0.12% 0.30% 0.51% V 0.09% 0.38% 0.89% W 0.08% 0.29% 0.62%

Relative absolute error Truth Prediction Prediction vs Truth Latent variable

Manuscript “Multi-level Convolutional Autoencoder Networks for Parametric Prediction of Spatio-temporal Dynamics,” Submitted CMAME

slide-42
SLIDE 42
  • Fully Data-driven framework
  • Multi-level neural network architecture
  • Convolutions in space & time
  • Non-linear manifolds
  • Fast training, faster prediction
  • Up to 6 orders of reduction in DoF
  • Total training time: 3.6 hours on one NVIDIA Tesla P100 GPU for 3D ship air wake
  • Prediction time: Seconds for a new parameter or hundreds of future steps

Caveats

  • Require large amounts of data
  • No indicator for choice of latent dimensions è use singular values to find an

upper bound

42 Manuscript “Multi-level Convolutional Autoencoder Networks for Parametric Prediction of Spatio-temporal Dynamics” to be submitted to ArXiv in a week

Summary

slide-43
SLIDE 43
  • DARPA Physics of AI program (Technical Monitor: Dr. Ted Senator )
  • Air Force Center of Excellence grant (Program Managers: Dr. Mitat Birkan and Dr.

Fariba Fahroo)

  • Office of Naval Research (Program manager: Dr. Brian Holm-Hansen)
  • Computational infrastructure : NSF-MRI (Program manager: Dr. Stefan Robila)

43

Acknowledgments

slide-44
SLIDE 44

44

𝛽 = 5°

Numerical Tests