9.54, fall semester 2014
9.54 class 4
Shimon Ullman + Tomaso Poggio
Danny Harari + Daneil Zysman + Darren Seibert
9.54 class 4 Supervised learning Shimon Ullman + Tomaso Poggio - - PowerPoint PPT Presentation
9.54 class 4 Supervised learning Shimon Ullman + Tomaso Poggio Danny Harari + Daneil Zysman + Darren Seibert 9.54, fall semester 2014 Intro 9.54, fall semester 2014 An old and simple model of supervised learning associate b to a
9.54, fall semester 2014
Shimon Ullman + Tomaso Poggio
Danny Harari + Daneil Zysman + Darren Seibert
9.54, fall semester 2014
Intro
9.54, fall semester 2014
An old and simple model of supervised learning
associate b to a and store: retrieve output b from input a — if
9.54, fall semester 2014
An old and simple model of supervised learning
retrieve output b from input a — if
It is a special case…
9.54, fall semester 2014
Linear
9.54, fall semester 2014
“Linear” learning
Find linear operator (eg a matrix) such that Define
9.54, fall semester 2014
“Linear” learning
If X−1 exists, then
If X−1 does not exists, then
where the pseudo inverse is the solution of and if X is full column rank
||A||F = p ( X
i,j
|ai,j|2) with
X† = (XT X)−1XT
9.54, fall semester 2014
“Linear” learning is linear regression
e.g. the output y is scalar, then
with M = XY −1
9.54, fall semester 2014
Nonlinear
10
Nonlinear learning
Find operator N such that Define
In general impossible but…assume N is in the class of polynomial mappings of degree k in the vector space V (over the real field)…eg N has a convergent Taylor series expansion Weierstrass theorem ensures approximation of any continuous function
11
Nonlinear learning
f(x) is a polynomial with all monomials as in this 2D example
9.54, fall semester 2014
Classification and Regression
13
y = sign(L1x + L2(x, x)) = sign(a1u1 + a2u2 + bu1u2) = sign(u1u2)
XOR function is in fact enough. This corresponds to a universal, one-hidden layer network all monomials input variables
MLPs RBFs
9.54, fall semester 2014
Radial Basis Functions
Nonlinear learning
xk−x||2 = ∞
X
n=0
||ˆ xk − x||2n n! Later we will see that RBF expansions are a good approximation
9.54, fall semester 2014
Memory-based computation
2σ2
The training set is
: then it is a
2σ2
memory, a lookup table
9.54, fall semester 2014
Memory-based computation Of course learning is much more than memory but in this model the difference is between a Gaussian and a delta function
Poggio, Edelman Nature, 1990.
2σ2
Logothetis, Pauls, and Poggio, 1995
9.54, fall semester 2014
Garfield
Image Analysis
⇒ Bear (45° view)
34
9.54, fall semester 2014
Hyperbf
36
9.54, fall semester 2014
Cartooon male
9.54, fall semester 2014
Radial Basis Functions and MLPs
Sigmoidal units are radial basis functions (for normalized inputs)
and thus is a radial function Consider the MLP units
Since
Sigmoidal units are radial basis functions (for normalized inputs)
Sigmoidal units are radial basis functions (for normalized inputs)