Generalized Approximate Survey Propagation for Hig igh-dimensional - - PowerPoint PPT Presentation
Generalized Approximate Survey Propagation for Hig igh-dimensional - - PowerPoint PPT Presentation
Generalized Approximate Survey Propagation for Hig igh-dimensional Estimation Luca Saglietti Yue Lu , Harvard University Carlo Lucibello , Bocconi University Outline Generalized Linear Models (GLM) Real-valued phase retrieval
Outline
- Generalized Linear Models (GLM)
- Real-valued phase retrieval
- Inference model
- Approximate message-passing
- Effective landscapes and competition
- Breaking the replica symmetry
- Changing the effective landscape
- Conclusions
Generalized Lin inear Models
3 ingredients : TRUE SIGNAL OBSERVATION MATRIX OBSERVED SIGNAL High-dimensional limit: with
- f
: : :
Generalized Lin inear Models
3 ingredients : TRUE SIGNAL OBSERVATION MATRIX OBSERVED SIGNAL High-dimensional limit: with
- f
: : :
Generalized Lin inear Models
3 ingredients : TRUE SIGNAL OBSERVATION MATRIX OBSERVED SIGNAL High-dimensional limit: with
- f
: : :
Generalized Lin inear Models
3 ingredients : TRUE SIGNAL OBSERVATION MATRIX OBSERVED SIGNAL High-dimensional limit: with
- f
: : :
An example: : Real-valued Phase Retrieval
Fun facts about phase retrieval:
- Physically meaningful!
- symmetry in the signal space.
- should provide enough information for a perfect reconstruction.
- Gradient descent struggles to reconstruct the signal until .
- Rigorous result about convexification in a regime.
( + noise )
An example: : Real-valued Phase Retrieval
Fun facts about phase retrieval:
- Physically meaningful!
- symmetry in the signal space.
- should provide enough information for a perfect reconstruction.
- Gradient descent struggles to reconstruct the signal until .
- Rigorous result about convexification in a regime.
( + noise )
An example: : Real-valued Phase Retrieval
Fun facts about phase retrieval:
- Physically meaningful!
- symmetry in the signal space.
- should provide enough information for a perfect reconstruction.
- Gradient descent struggles to reconstruct the signal until .
- Rigorous result about convexification in a regime.
( + noise )
An example: : Real-valued Phase Retrieval
Fun facts about phase retrieval:
- Physically meaningful!
- symmetry in the signal space.
- should provide enough information for a perfect reconstruction.
- Gradient descent struggles to reconstruct the signal until .
- Rigorous result about convexification in a regime.
( + noise )
An example: : Real-valued Phase Retrieval
Fun facts about phase retrieval:
- Physically meaningful!
- symmetry in the signal space.
- should provide enough information for a perfect reconstruction.
- Gradient descent struggles to reconstruct the signal until .
- Rigorous result about convexification in a regime.
( + noise )
An example: : Real-valued Phase Retrieval
Fun facts about phase retrieval:
- Physically meaningful!
- symmetry in the signal space.
- should provide enough information for a perfect reconstruction.
- Gradient descent struggles to reconstruct the signal until .
- Rigorous result about convexification in a regime.
( + noise )
In Inference Model
Sensible choice: GRAPHICAL MODEL MATCHED / MISMATCHED
Estimator :
Bayesian optimal: Maximum a posteriori:
In Inference Model
Sensible choice: GRAPHICAL MODEL MATCHED / MISMATCHED
Estimator :
Bayesian optimal: Maximum a posteriori:
In Inference Model
Sensible choice: GRAPHICAL MODEL MATCHED / MISMATCHED
Estimator :
Bayesian optimal: Maximum a posteriori:
Approximate Message-passing
How do we obtain ? Easy (if everything is i.i.d.)
BP Gaussian ansatz rBP Close on single-site quantities AMP (TAP)
DEFINE 2 SCALAR INFERENCE CHANNELS:
Encoding prior dependence Encoding data dependence
When do we get a good estimator?
Approximate Message-passing
How do we obtain ? Easy (if everything is i.i.d.)
BP Gaussian ansatz rBP Close on single-site quantities AMP (TAP)
DEFINE 2 SCALAR INFERENCE CHANNELS:
Encoding prior dependence Encoding data dependence
When do we get a good estimator?
Approximate Message-passing
How do we obtain ? Easy (if everything is i.i.d.)
BP Gaussian ansatz rBP Close on single-site quantities AMP (TAP)
DEFINE 2 SCALAR INFERENCE CHANNELS:
Encoding prior dependence Encoding data dependence
When do we get a good estimator?
Approximate Message-passing
How do we obtain ? Easy (if everything is i.i.d.)
BP Gaussian ansatz rBP Close on single-site quantities AMP (TAP)
DEFINE 2 SCALAR INFERENCE CHANNELS:
Encoding prior dependence Encoding data dependence
When do we get a good estimator?
Approximate Message-passing
How do we obtain ? Easy (if everything is i.i.d.)
BP Gaussian ansatz rBP Close on single-site quantities AMP (TAP)
DEFINE 2 SCALAR INFERENCE CHANNELS:
Encoding prior dependence Encoding data dependence
When do we get a good estimator?
Approximate Message-passing
How do we obtain ? Easy (if everything is i.i.d.)
BP Gaussian ansatz rBP Close on single-site quantities AMP (TAP)
DEFINE 2 SCALAR INFERENCE CHANNELS:
Encoding prior dependence Encoding data dependence
When do we get a good estimator?
Effective Landscapes and Competition
ρ energy
Low overlap High overlap
IMPOSSIBLE (SNR) (possible scenario)
1:
- verlap:
GD in this effective landscape Stationary points Fixed points
energy IMPOSSIBLE (SNR)
2:
Effective Landscapes and Competition
(possible scenario)
- verlap:
GD in this effective landscape Stationary points Fixed points ρ
Low overlap High overlap
energy HARD IMPOSSIBLE (SNR)
3:
Effective Landscapes and Competition
(possible scenario)
- verlap:
GD in this effective landscape Stationary points Fixed points ρ
Low overlap High overlap
energy EASY HARD IMPOSSIBLE (SNR)
4:
Effective Landscapes and Competition
(possible scenario)
- verlap:
GD in this effective landscape Stationary points Fixed points ρ
Low overlap High overlap
Breaking the symmetry ry
GAMP GASP(s)
vs Replica symmetry assumption 1RSB assumption Input scalar channel: Input scalar channel:
- Same computational complexity
- (Potentially) more expensive element-wise operations
- How to set the symmetry breaking parameter s ?
SYMMETRY BREAKING PARAMETER
Breaking the symmetry ry
GAMP GASP(s)
vs Replica symmetry assumption 1RSB assumption Input scalar channel: Input scalar channel:
- Same computational complexity
- (Potentially) more expensive element-wise operations
- How to set the symmetry breaking parameter s ?
SYMMETRY BREAKING PARAMETER
Breaking the symmetry ry
GAMP GASP(s)
vs Replica symmetry assumption 1RSB assumption Input scalar channel: Input scalar channel:
- Same computational complexity
- (Potentially) more expensive element-wise operations
- How to set the symmetry breaking parameter s ?
SYMMETRY BREAKING PARAMETER
Message-passing equations
GAMP GASP(s)
Changing the Effective Landscape
Phase retrieval, noiseless case No regularizer
RS RS
Energy
Perfect recovery!
: explore minima at different energy levels GROUND STATE
RS : Hard below 1RSB : Hard below
Changing the Effective Landscape
Phase retrieval, noiseless case No regularizer
RS 1RSB RS 1RSB
Energy
Perfect recovery!
: explore minima at different energy/complexity levels GROUND STATE
RS : Hard below 1RSB : Hard below
Changing the Effective Landscape
Phase retrieval, noiseless case No regularizer
RS 1RSB RS 1RSB
Energy
Perfect recovery!
: explore minima at different energy/complexity levels GROUND STATE
RS : Hard below 1RSB : Hard below
Conclusions
- In mismatched inference settings the RS assumption can be wrong.
- GASP can improve over GAMP. Same O(N^2) complexity.
- Simple continuation strategy can push GASP down to the BO algorithmic threshold.
- For more details please check my poster this evening!