Synthesis and Analysis Sparse Representation Models for Image Restoration Shuhang Gu 顾舒航
- Dept. of Computing
Models for Image Restoration Shuhang Gu Dept. of Computing The - - PowerPoint PPT Presentation
Synthesis and Analysis Sparse Representation Models for Image Restoration Shuhang Gu Dept. of Computing The Hong Kong Polytechnic University Outline Sparse representation models for image modeling Synthesis based representation
2
Sparse representation models for image modeling
Weighted nuclear norm and its applications in low level vision
Convolutional sparse coding for single image super-resolution
Synthesis and analysis sparse representation models
3
Synthesis based sparse representation model
Synthesis based sparse representation model assumes that a signal 𝑦 can be represented as a linear combination of a small number of atoms chosen out of a dictionary 𝐸: 𝑦 = 𝐸𝛽, s.t. 𝛽 0<𝜁
4
? ? . . . ? ?
A dense solution A sparse solution
Elad, M., Milanfar, P., Rubinstein, R. Analysis versus synthesis in signal priors. Inverse problems 2007.
simple multiplication operation, and assumes the coefficients are sparse: 𝑄𝑦 0<𝜁
5
? ? . . . ? ?
Elad, M., Milanfar, P., Rubinstein, R. Analysis versus synthesis in signal priors. Inverse problems 2007.
S&A representation models for image modeling
6
Synthesis model 𝑦 = 𝐸𝛽, where 𝛽 is sparse Analysis model 𝛾 = 𝑄𝑦, where 𝛾 is sparse
synthesis model emphasis the non-zero values in the sparse coefficient vector 𝛽, because these non-zero values select vectors in the dictionary to span the space of input signal
A hyperplane
Analysis model emphasis the zero values in the sparse coefficient vector 𝑄𝑦, because these zero values select vectors in the projection matrix to span the complementary space
Elad, M., Milanfar, P., Rubinstein, R. Analysis versus synthesis in signal priors. Inverse problems 2007.
S&A representation models for image modeling
Image restoration/enhancement problems 𝑧 = 𝑦 + 𝑜 𝑧 = 𝐸(𝑙⨂𝑦) + 𝑜 𝑧 = 𝑙⨂𝑦 + 𝑜 𝑧 = 𝑁 ⊙ 𝑦 + 𝑜
7
Image Inpainting Image Denoising Image Deconvolution
Image Super-resolution
S&A representation models for image modeling
8
– Sparsity priors – Non-local similarity priors – Color line priors …
Buades A, Coll B, Morel JM. A non-local algorithm for image denoising. In CVPR 2005.
S&A representation models for image modeling
𝑏𝑠𝑛𝑏𝑦𝑦𝑞 𝑦 𝑧 = 𝑞 𝑧 𝑦 𝑞(𝑦) Minimize the –log(𝑞 𝑦 𝑧 ): 𝑦 = 𝑏𝑠𝑛𝑗𝑜𝑌 1 2 𝑦 − 𝑧 𝐺 − log(𝑞(𝑦))
9
𝐪(𝐲)
Transformation domain Original signal domain Decomposition domain Long-tail dist. leads to sparse solution
discriminative enough
Gaussian likelihood
Data prior modeling
Prior modeling
Long-tail dist. leads to sparse solution Analysis model 𝜚(𝑄𝑦) Synthesis model 𝜔(𝛽)
S&A representation models for image modeling
Synthesis model
𝑛𝑗𝑜𝛽 1 2 𝑧 − 𝐸𝛽 𝐺 + 𝜔 𝛽 𝑦 = 𝐸𝛽
KSVD, BM3D, LSSC, NCSR, et. al.
10
Analysis model
𝑛𝑗𝑜𝑦 1 2 𝑧 − 𝑦 𝐺 + 𝜚(𝑄𝑦)
TV, wavelet methods, FRAME, FOE, CSF, TRD et. al.
S&A representation models for image modeling
Synthesis model
Methods: KSVD, BM3D, LSSC, NCSR, etc.
11
Analysis model
methods: TV, wavelet methods, FOE, CSF, TRD etc.
Patch based Filter based
S&A representation models for image modeling
Synthesis model
Methods: KSVD, BM3D, LSSC, NCSR, etc.
?
12
Analysis model
Methods: Analysis-KSVD et al. methods: TV, wavelet methods, FOE, CSF, TRD etc.
Patch based Filter based
S&A representation models for image modeling
Notes:
Aggregation: Overlap aggregation method may smooth image or generate ringing artifacts Non-local prior: Non-local prior helps to generate visual plausible results on highly noisy situation
13
Analysis model
Patch based Filter based Synthesis model
S&A representation models for image modeling
Notes:
Aggregation: Overlap aggregation method may smooth image or generate ringing artifacts Non-local prior: Non-local prior helps to generate visual plausible results on highly noisy situation
14
Analysis model
Patch based Filter based Synthesis model
Different applications may be better solved via different models!
S&A representation models for image modeling
15
– A analysis model with patch based implementation
– A synthesis model with filter based implementation
Denoising SR
Weighted nuclear norm minimization and its applications in low level vision
16
Low rank models
Loss functions are determined by different noise models: Gaussian noise model: PCA, Probabilistic PCA Sparse noise model: Robust PCAs Partial observations: Matrix completion Complex noise model: MoG etc.
17
Low rank models
A common used regularization term is the nuclear norm of matrix X
𝑌 ∗ = 𝜏𝑗(𝑌) 1 Pros:
Character:
18
Candès, Emmanuel J., et al. "Robust principal component analysis?." Journal of the ACM, 2011. Candès, Emmanuel J., and Benjamin Recht. "Exact matrix completion via convex optimization." Foundations of Computational mathematics 2009. Cai, J. F., Candès, E. J., & Shen, Z. A singular value thresholding algorithm for matrix completion. SIAM Journal on Optimization, 2010.
Low rank models
Analysis sparse model
𝑛𝑗𝑜𝑦 1 2 𝑧 − 𝑦 𝐺 + 𝜚(𝑄𝑦)
Nuclear norm regularization model
𝑛𝑗𝑜𝑌 1 2 𝑍 − 𝑌 𝐺 + 𝑉𝑈𝑌𝑊 1
Nuclear norm regularization model can be interpreted as a 2D analysis sparse model!
19
Weighted nuclear norm minimization
𝑛𝑗𝑜𝑌 1 2 𝑍 − 𝑌 𝐺 + 𝜇 𝑌 ∗ 𝑌∗ = 𝑉𝑇𝜇 𝜏𝑍 𝑊𝑈
– Tightest convex envelope of rank minimization. – Closed form solution.
– Treat equally all the singular values. Ignore the different significances of matrix singular values.
20
Cai, J. F., Candès, E. J., & Shen, Z. A singular value thresholding algorithm for matrix completion. SIAM Journal on Optimization, 2010.
Weighted nuclear norm minimization Weighted nuclear norm
𝑌 𝑥,∗ = 𝑥𝑗𝜏𝑗(𝑌) 1
Weighted nuclear norm proximal (WNNP)
𝑌 = 𝑏𝑠𝑛𝑗𝑜𝑌 𝑌 − 𝑍 𝐺 + 𝑌 𝑥,∗
– The WNNM is not convex for general weight vectors – The sub-gradient method cannot be used to analyze its
21
Weighted nuclear norm minimization
Theorem 1. ∀𝑍 ∈ 𝑆𝑛×𝑜, let 𝑍 = 𝑉Σ𝑊𝑈be its SVD. The optimal solution of the WNNP problem: 𝑌 = 𝑏𝑠𝑛𝑗𝑜𝑌 𝑌 − 𝑍 𝐺 + 𝑌 𝑥,∗ is 𝑌 = 𝑉𝐸𝑊𝑈 where D is a diagonal matrix with diagonal entries d=[d1, d2, … , dr] (r=min(m,n)) and d is determined by: 𝑛𝑗𝑜𝑒1, 𝑒2…𝑒𝑜 𝑗=1
𝑠
(𝑒𝑗−σ𝑗)2 + 𝑥𝑗 𝑒𝑗 𝑡. 𝑢. 𝑒1 ≥ 𝑒2≥ 𝑒𝑠 ≥0.
22
Weighted nuclear norm minimization Corollary 1. If the weights satisfy 0 ≤ 𝑥1 ≤ 𝑥2≤ 𝑥𝑜, the non-convex WNNP problem has a closed form
𝑌 = 𝑉𝑇𝑥(Σ)𝑊𝑈
where 𝑍 = 𝑉Σ𝑊𝑈 is the SVD of 𝑍, and
𝑇𝑥(Σ)𝑗𝑗 = max Σ𝑗𝑗 − 𝑥𝑗, 0 .
23
WNNM for image denosing
24
the image for its nonlocal similar patches to form matrix Y.
estimate the clean patches X from Y.
image.
several times to obtain the denoised image.
WNNM
𝑌 = 𝑏𝑠𝑛𝑗𝑜𝑌 𝑌 − 𝑍 𝐺 + 𝑌 𝑥,∗
WNNM for image denosing
25
Reweighting strategy to promote sparsity
𝑥𝑗 = 𝐷 𝜏𝑗 𝑌 + 𝜁
computation burden
WNNM for image denosing
26
WNNM-RPCA
27
𝑌, 𝑍 ∈ 𝑆𝑛×𝑛, 𝑆𝑏𝑜𝑙 𝑌 = 𝑄
𝑠 × 𝑛, 𝐹 0 = 𝑄 𝑓 × 𝑛2
28
WNNM-MC
29
𝑌, 𝑍 ∈ 𝑆𝑛×𝑛, 𝑆𝑏𝑜𝑙 𝑌 = 𝑄
𝑠 × 𝑛, 𝐹 0 = 𝑄 𝑓 × 𝑛2
WNNM-MC
30
WNNM summary
31
problem.
and achieved state-of-the-art performance.
WNNM-MC. WNNM achieved superior performance than NNM
Convolutional sparse coding for single image super-resolution
32
Convolutional sparse coding
Consistency constraint
33
Convolutional sparse coding
Aggregation method in patch based algorithms EPLL 𝑛𝑗𝑜𝑌 𝒀 − 𝒁 2 + 𝑆𝑗𝒀 − 𝑎𝑗
2 + 𝑄(𝑎𝑗)
34
Noisy Non Overlapping Center Pixel Overlapping
Zoran D, Weiss Y. From learning models of natural image patches to whole image restoration. In: ICCV 2011.
Convolutional sparse coding
Sparse coding
𝑛𝑗𝑜𝛽||𝑧 − 𝐸𝛽||𝐺
2+𝜚(𝛽)
Convolutional sparse coding
𝑛𝑗𝑜𝒂 𝒁 − 𝒈𝑗 ⊗ 𝒜𝑗
𝐺 2+ 𝜒(𝒜𝑗)
35
…
… … …
Mat atrix ix Form
Convolutional sparse coding for image SR
36
N LR feature maps M HR feature maps
M HR filters
Learning Joint HR Filter and Mapping Function Learning LR Filter Learning
The Testing Phase The Training Phase
N LR filters
CSC LR Filter Learning HR Filter Learning HR Feature Map Estimation Convolution
Convolutional sparse coding for image SR
37
N LR feature maps M HR feature maps
Learning Joint HR Filter and Mapping Function Learning LR Filter Learning
The Training Phase
CSC LR Filter Learning HR Filter Learning
– Pre-processing
Convolutional sparse coding for image SR
38
N LR feature maps M HR feature maps
Learning Joint HR Filter and Mapping Function Learning LR Filter Learning
The Training Phase
CSC LR Filter Learning HR Filter Learning
– LR filter training
Convolutional sparse coding for image SR
39
N LR feature maps M HR feature maps
Learning Joint HR Filter and Mapping Function Learning LR Filter Learning
The Training Phase
CSC LR Filter Learning HR Filter Learning
– Joint HR filter and mapping function learning
Convolutional sparse coding for image SR
40
M HR filters
The Testing Phase The Training Phase
N LR filters
HR Feature Map Estimation Convolution
Convolutional sparse coding for image SR
41
Optimization: SA-ADMM The original problem can be write as:
Convolutional sparse coding for image SR
42
Optimization: SA-ADMM SA-ADMM
Convolutional sparse coding for image SR
43
Convolutional sparse coding for image SR
44
Convolutional sparse coding for image SR
45
Convolutional sparse coding for image SR
46
CSC-SR: summary and future work
47
– To avoid patch aggregation in super-resolution, we utilize convolutional sparse coding to deal with SR problem. – SA-ADMM algorithm is used to train CSC-SR model for large scale training data. – State-of-the-art SR results with high PSNR and visual quality.
– End to end training strategy may be better. – Is there any optimization algorithm which is more suitable for CSC training.
Related Publication
Image Denoising,” In CVPR 2014.
Weighted Nuclear Norm Minimization” Technical Report.
Applications to Low Level Vision. Submitted to IJCV (Minor revision).
Super-resolution," In ICCV 2015.
References
2007.
for sparse representation. TSP 2006.
domain collaborative filtering. TIP, 2007.
48
References
In ICCV 2011.
Physica D: Nonlinear Phenomena 1992.
unified theory for texture modeling. IJCV 1998.
algorithm for the analysis sparse model." TSP 2013.
Michael E., and Christopher M. Bishop. "Probabilistic principal component analysis." Journal of the Royal Statistical Society: Series B (Statistical Methodology) 61.3 (1999): 611-622.
49
References
alternative convex programming. In: CVPR 2005.
Emmanuel J., and Benjamin Recht. "Exact matrix completion via convex
In: ICCV 2011.
2013. Note: References of comparison methods in the tables are omitted, all of these references can be found in my corresponding publications.
50
51