Ensemble Kalman Filters for wildfire simulation Jan Mandel Center - - PowerPoint PPT Presentation

ensemble kalman filters
SMART_READER_LITE
LIVE PREVIEW

Ensemble Kalman Filters for wildfire simulation Jan Mandel Center - - PowerPoint PPT Presentation

Ensemble Kalman Filters for wildfire simulation Jan Mandel Center for Computational Mathematics University of Colorado at Denver and Health Sciences Center, Denver, CO Mesoscale and Microscale Meteorology Division National Center for


slide-1
SLIDE 1

Ensemble Kalman Filters

for wildfire simulation

Jan Mandel

Center for Computational Mathematics University of Colorado at Denver and Health Sciences Center, Denver, CO Mesoscale and Microscale Meteorology Division National Center for Atmospheric Research, Boulder, CO Supported by NSF grant CNS-0325314 Industrial Affiliates Meeting Center for Subsurface Modeling University of Texas at Austin October 16, 2006

slide-2
SLIDE 2

The Wildfires Project Team

University of Colorado at Denver Department of Mathematical Sciences Jan Mandel (PI, Lead PI) Lynn Bennethum (Co-PI) Leo Franca (Co-PI) Craig Johns (Co-PI) Tolya Puhalskii (Co-PI Mingeong Kim (graduate student) Vaibhav Kulkarni (graduate student) Jonathan Beezley (graduate student) National Center for Atmospheric Research Janice Coen (PI) Texas A&M University

  • Dept. of Computer Science

Guan Qin (PI) Wei Zhao (PI) Jianjia Wu (graduate student) Rochester Institute of Technology Center for Imaging Science Anthony Vodacek (PI) Robert Kremens (Co-PI) Ambrose Onoye (postdoc) Ying Li (graduate student) Zhen Wang (graduate student) Matthew Weinstock (undergrad. student) University of Kentucky

  • Dept. of Computer Science

Craig Douglas (PI) Deng Li (postdoc) Wei Li (graduate student) Adam Zornes (graduate student) Other Collaborators: USDA Forest Service Missoula Tech. Development Center – UAVs, SAFE

  • Univ. of Montana (Natl. Cntr. Landscape Fire Analysis)

Univ of Utah - SCIRun enhancements

slide-3
SLIDE 3

The Objective

A Dynamic Data Driven Application System (DDDAS) for short-range forecasts of wildfire behavior with models steered by real-time weather data, fire- mapping images, and sensor streams.

slide-4
SLIDE 4

Goals

  • The model

– faster than real time – calibrated from measurements

  • Data assimilation

– sparse data (weather stations, field deployed sensors) – large image datasets (aerial photographs) – data acquisition steering – data arriving delayed and out of order – capable of adjusting a highly nonlinear model

  • Real-time visualization over the internet in the field
slide-5
SLIDE 5

Wildfire DDDAS Structure

Synthetic data Map sources (GIS) Fuel Data Sensors, telemetry Forecast Weather Fire Model Observation function Aerial imaging Adjust Compare Data Assimilation Initial conditions Weather data Data Acquisition Real data pool Real time data Interpret

slide-6
SLIDE 6

Modular Software Structure: Major components are interchangeable

Model 1. Weather model: Clark-Hall, WRF 2. Coupled with fire model

1. Convection-reaction-diffusion PDE 2. Fireline propagation (≈ reaction sheet asymptotics)

Data Assimilation

  • Ensemble Kalman Filter, improved efficiency
  • Regularized, Predictor-Corrector & Morphing

nonlinear filters (new)

slide-7
SLIDE 7

Stochastic Approach

“There are no guarantees in life,

  • nly probabilities”

(Jack Ryan in “Executive Orders” by Tom Clancy)

slide-8
SLIDE 8

Data Assimilation

  • Used in navigation, industrial control,

weather and ocean modeling for a long time

  • Sequential statistical estimation: best

estimate of the system state from all data available up to now

  • Ensemble data assimilation allows the

application to be used without any modifications, just encapsulated to advance a simulation state

slide-9
SLIDE 9

What is the model state?

  • An approximate probability distribution of the

system state

  • Represented numerically by

– The mean and the covariance matrix of gaussian distribution → Kalman filter – A sample from gaussian distribution → ensemble Kalman filters – Weigthed sample from general distribution → particle filters, sequential Monte Carlo,… – Other expansions and approximations, such as Gaussian mixtures, polynomial chaos, Karhunen – Loeve expansion,… (in progress)

slide-10
SLIDE 10

What is data?

Meaningful data items must

  • Have measurement values (duh… the data)
  • Have uncertainty estimate (“error bounds”)
  • Be related to the model state in a known way

(observation function: from a hypothetical model state, produce synthetic data to be compared to the measurements) Data item is a probability density of the measurements, conditional on a system state.

slide-11
SLIDE 11

Time Sequence of Fire Propagation Aerial Images from a Prescribed Burn

(Anthony Vodacek)

slide-12
SLIDE 12

Image Processing Algorithms

(AVIRIS Image from Vodacek et al. and Latham 2002, Int. J. Remote Sensing)

589 nm 770 nm/779 nm

Original image content

  • Pixel location
  • Spectral data
  • Algorithms to register to model grid
  • auto extraction of tie points
  • affine transform

Reduced image content

  • Normalized Thermal Index

(MWIR-LWIR)/(MWIR+LWIR)

  • Fire location
  • Derived temperatures
  • Direction fire is spreading
  • Derived fuels (NDVI)

(Anthony Vodacek)

slide-13
SLIDE 13

Autonomous Environmental Detectors (Primarily for local weather… but some burnovers)

We have developed a versatile electronic acquisition package ideally suited to field data collection

Major Features

Reconfigure to rapidly deploy? GPS - Position Aware Versatile Data Inputs Voice or Data Radio telemetry Inexpensive

Kremens, et al. 2003. Int. J. Wildland Fire

Data logger and thermocouples

100 200 300 400 500 600 700 800 11250 11750 12250 12750 13250 seconds after ignition temperature, C

Time (sec. after ignition)

T (oC)

slide-14
SLIDE 14

Sequential Statistical Estimation: Analysis Cycle

Forecast (prior) Synthetic data Measured data Analysis (posterior) Adjust Compare Observation function Advance in time Forecast (prior)

slide-15
SLIDE 15

Example: Kalman filter in GPS

  • GPS unit assimilates satelite measurements into a

forecast of its location (the last location, or, if smarter, extrapolation of movement)

  • With more measurements accuracy increases (esp.

if the unit stays still)

  • In 1D:

1 1 1 analysis variance forecast variance data variance = +

slide-16
SLIDE 16

Why ensemble filters?

Kalman filter must maintain covariance matrix of the state. This is

  • Complicated – much more programming than just

advancing the model in time

  • Expensive – covariance is a dense matrix, size equal

number of dofs in the model (easily millions!)

  • Limited to gaussian distribution

Solution: advance an ensemble of simulations, replace covariance by covariance of the ensemble. Also gain some robustness for the non-gaussian case this way.

slide-17
SLIDE 17

How? Ensemble Kalman Filter as Least Squares

  • Idealized formula: minimize in the span of the ensemble the

sum of

– Difference from forecast mean, weighted by the inverse of the forecast covariance – Difference of the output of the observation function from the data, weighted by the inverse of the covariance of data error distribution

  • To get the analysis ensemble:

– Apply the idealized formula to forecast ensemble members – replace the unknown forecast covariance by sample covariance – add random perturbation to each member’s data separately

  • There are other variants. But: in all variants, the analysis

ensemble is always a linear combination of the members of the forecast ensemble.

  • Dominant operations:

– advance ensemble members in time, embarrassingly parallel – dense linear algebra (parallel, e.g., Scalapack)

slide-18
SLIDE 18

Simple explanation of (Ensemble) Kalman Filter update step

  • Change the simulation state to balance two

contradictory objectives:

– The state should not change – The state should match the data

  • The more uncertainty (bigger covariance)
  • ne of the conditions has, the more it can be

violated (i.e., not be taken seriously)

slide-19
SLIDE 19

But Ensemble Kalman Filter fails for the wildfire problem (and does poorly for others, like hurricanes)

  • Ensemble Kalman Filters form the analysis ensemble by

solving a least squares problem, trying to match the data

  • The analysis ensemble is made of linear combinations
  • f the forecast ensemble. Consequently, if the forecast

ensemble is not rich enough, we are simply out of luck 1. In a desperate attempt to match the data, nonphysical states result. Observation: when bad things happen, gradients get really large 2. Probability distributions are strongly non-gaussian (burning/not burning) 3. State discrepancies are in fireline position as well as in values

slide-20
SLIDE 20

What are we doing about it: New developments in EnKF

  • Prevent nonphysical states:

Regularized Ensemble Kalman Filters

  • Nongaussian distribution:

Predictor-corrector filters

  • Position errors:

Morphing filters

slide-21
SLIDE 21

Dealing with nonphysical states: Regularized EnKF

slide-22
SLIDE 22

Simple PDE Wildfire Model

1 2 3

( ) ( ) (heat balance) ( ) (fuel balance)

a

T S k T c T c T T c t t S Sr T t ∂ ∂ = −∇ ∇ + ⋅∇ − − + ∂ ∂ ∂ = − ∂

is the temperature is the fuel supply ( ) is the rate of burning is the ignition temperature is the ambient temperature is white noise

i a

T S r T T T σ

Home

A simple model that however exhibits the correct qualitative behavior. Not captured: unburned spots caused by atmosphere interaction. Assumption that reaction intesity depends only on temperature may be too simplistic.

slide-23
SLIDE 23

Regularized EnKF

Filter developed to control spatial gradient of the solution. Example: 1D Fire Model Add penalty to the least squares: small constant times squared norm of the difference between gradient

  • f solution and gradient of

the forecast ensemble mean. Same as adding independent

  • bservation with large

variance gradient of the solution did not change.

Johns and Mandel – Envir. &

  • Ecol. Statistics. (in print)
slide-24
SLIDE 24

2D Simulated Fire Data Assimilation

  • Toy PDE model, 30x30 grid
  • Esemble 88 members
  • Observation function: 100 sample points
  • 2 display frames/assimilation cycle
  • Initial ensemble generated by random smooth spatial shift

and random smooth random additive perturbation of initial condition

  • The initial condition is different than “truth” to verify that

the ensemble can find and track the truth

  • Visualization:
  • Vertical axis is temperature, artificial color for perception only
  • demo-alpha: ensemble as overlay of transparent members
  • demo-mesh: ensemble mean, pointwise standard deviation
  • Reference solution is artificial “truth”
  • Comparison solution has the same initial conditions as the ensemble
slide-25
SLIDE 25

2D Simulated Fire Data Assimilation

The Reference solution represents the truth. Data assimilation by a standard ENKF algorithm results in an unstable solution because

  • f the nonlinear behavior of
  • wildfire. Stabilization gives

the regularized solution ENKF+reg. Without data assimilation, the solution would develop as in the Comparison; the data assimilation shifts the model towards the truth. The model state is a probability distribution, visualized in the two ENKF figures as the superposition of transparent temperature profiles of ensemble members.

slide-26
SLIDE 26
slide-27
SLIDE 27
slide-28
SLIDE 28
slide-29
SLIDE 29
slide-30
SLIDE 30
slide-31
SLIDE 31
slide-32
SLIDE 32
slide-33
SLIDE 33
slide-34
SLIDE 34
slide-35
SLIDE 35
slide-36
SLIDE 36
slide-37
SLIDE 37
slide-38
SLIDE 38
slide-39
SLIDE 39
slide-40
SLIDE 40
slide-41
SLIDE 41
slide-42
SLIDE 42
slide-43
SLIDE 43
slide-44
SLIDE 44
slide-45
SLIDE 45

Dealing with nongaussian distributions: Predictor-Corrector Ensemble Filters

slide-46
SLIDE 46

Non-gaussian probability distributions

  • Probability distributions (also of the solution) are too far from Gaussian
  • The problem is too nonlinear
  • Probability concentrated around burning and not burning states

Probability density Burns: 70% probability Does not burn: 30% probability Least squares solution: does not burn Temperature Ignition temperature

slide-47
SLIDE 47

Sequential Importance Sampling (SIS)

  • Model state is a weighted ensemble {wkuk}, sum of weights Σwk=1
  • Analysis ensemble is obtained by assigning so-called importance weights

from the Bayes theorem to a proposal ensemble, which is a sample from some proposal pdf:

( ) ~ ( | ) ( )

forecast proposal analysis proposal k k k proposal proposal k

p u w p d u p u

  • The formula requires the ratio of the forecast and the proposal densities

at the proposal ensemble members. In the SIS filter and variations, the same sample is taken for the forecast and the analysis, and the ratio equals to the forecast weights wk , giving

~ ( | )

analysis proposal proposal k k k

w p d u w

BIG problem: if the forecast is too far from the data, the data likelihoods p(d|u) and thus the importance weights are all numerically zero. Even if they are not zero, the weights become concentrated, and the ensemble degenerates.

slide-48
SLIDE 48

Comparison of EnKF and SIS

  • EnKF (Ensemble Kalman Filter)

– Requires the assumption that all distributions are normal, but used in practice in the general case anyway – Powerful algebra handles large differences between forecast and analysis OK – Weights all equal all the time, ensemble changes by the update – Relatively small ensembles (100s)

  • SIS (Sequential Importance Sampling)

– No normality assumption required – Ensemble members do not change in the update, only the weights updated – But the weights become concentrated on members in the region where the data likelihood and the forecast density are large, resulting in degenerate ensemble

  • r method failure (no such members). Consequently, SIS requires
  • small differences between forecast and analysis
  • frequent data input, before the model deviates from the reality too much
  • very large ensembles (easily 10000) to have the chance at some members

with large weights

  • Resampling to get rid of members with zero weights and to recover the

ensemble size

slide-49
SLIDE 49

Predictor-Corrector Filters (new)

  • Predictor (EnKF formulas, …) produces the

proposal ensemble. Needs to work on weighted

  • ensembles. Can make big adjustments.
  • Corrector assigns importance weights from

estimated ratio of the forecast and the proposal density at proposal ensemble members Need density estimation on spaces of smooth random functions (in infinite dimension)

slide-50
SLIDE 50

Predictor-Corrector EnKF+SIS

−1.5 −1 −0.5 0.5 1 1.5 0.2 0.4 0.6 0.8 1 1.2 1.4 EnKF −1.5 −1 −0.5 0.5 1 1.5 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 Prior distribution −1.5 −1 −0.5 0.5 1 1.5 1 2 3 4 5 6 7 8 EnKF SIS −1.5 −1 −0.5 0.5 1 1.5 1 2 3 4 5 6 Data likelihood

EnKF Predictor Data Forecast SIS Proposal Analysis Corrector

Similarly, nongaussian forecast + gaussian data → nongaussian analysis

slide-51
SLIDE 51

Dimension independent probability density estimation

  • Nonparametric “naïve kernel” estimate of probability density

ratio: p with respect to a measure µ: given a sample of size n and radius r>0 of a ball Br(x) with center x,

number of samples in the ball ( ) ( ) ( ( ))

r r

B x p x n B x µ ≈

  • Needs only the distance to define the ball, and the measure µ

(µ can be the Lebesgue measure in finite dimension,

  • therwise a Gaussian measure)
  • Convergence theorems known in Rk and for diffusion

stochastic processes, as n →∞, r=rn→0, nµ(Br(x)) →∞

slide-52
SLIDE 52

Dimension independent estimate of the density ratio

{ }

( ) number of forecast ensemble members ( ) ( ) number of proposal ensemble members ( ) is such that about proposal members ( ) |

forecast r proposal r r

p x B x p x B x r n B x y y x r ∈ ≈ ∈ ∈ = − <

How to choose the norm? The probability for a function in the state space should be positive to fall into a ball with r>0. In infinite dimension, this is far from automatic, and restricts the choice of the norm in the density estimate! Solution: choose the norm associated with the probability distribution the initial ensemble was sampled from (this is the norm of the so-called Cameron-Martin space associated with the probability distribution)

slide-53
SLIDE 53

Sobolev norms, gaussian measures, and smooth random fields

1 2 2 2 1 1

1D example: given 0, generate random function by ( ) sin , (0, ) (faster oscillating components are smaller) associated norm: for ( ) sin (fun

k k k k k k k k k k k

u x c kx c N c u u x c kx λ λ λ λ

∞ = ∞ ∞ = =

> → = = =

∑ ∑ ∑

1

ction with finite norm has fast oscillating components small) e.g. for 1/ we get the norm of the Sobolev space (0, ).

k

k H λ π =

slide-54
SLIDE 54
slide-55
SLIDE 55
slide-56
SLIDE 56
slide-57
SLIDE 57
slide-58
SLIDE 58
slide-59
SLIDE 59

Dealing with position errors: Morphing Ensemble Filters

slide-60
SLIDE 60

Image registration and morphing

(Picture Gao and Sederberg, 1998)

1 2 2 2

interpolate between two maps: ( ) ( ) given and , how to find ? solve minimization problem for registration distance ( , ) min ( ) for some suitable norms (e.g. involving der

T

f x f x Tx f f g f T d f g f I T g T

λ

λ = + = = = + − +

  • ivatives of )

can be done by multilevel optimization, reasonably fast T

If a good initial value of T is known, solving the registration problem by minimization of the registration distance is much easier.

slide-61
SLIDE 61

Morphing Ensemble Filters

  • Represent ensemble members as morphs of one fixed

member plus residual:

  • Instead of making linear combinations of the ensemble

members ui, make linear combinations of the morph mappings Ti and the residuals ri

  • After members are advanced in time, use previous

morph mapping as a good initial guess.

  • Now we can move the fireline easily!

( )

i i i

u u I T r = + +

slide-62
SLIDE 62

Software

  • Calls simulation as a black box
  • Massively parallel
  • Simulation must be able to

– Dump state, including information what the arrays in the state are, also nodal coordinates – Restart from modified state – User provides link between simulation state and data (observation function, problem dependent)

slide-63
SLIDE 63

Papers

  • www-math.cudenver.edu/ccm/reports