CSC421 Intro to Artificial Intelligence UNIT 28: Learning from - - PowerPoint PPT Presentation

csc421 intro to artificial intelligence
SMART_READER_LITE
LIVE PREVIEW

CSC421 Intro to Artificial Intelligence UNIT 28: Learning from - - PowerPoint PPT Presentation

CSC421 Intro to Artificial Intelligence UNIT 28: Learning from observations (+ leftovers from Prob. Reasoning over Time) Filtering Example Smoothing Divide evidence e 1:t to e 1:k , e k+1,t P(X k | e 1:t ) = P(X k | e 1:k , e k+1,t )


slide-1
SLIDE 1

CSC421 Intro to Artificial Intelligence

UNIT 28: Learning from observations (+ leftovers from Prob. Reasoning

  • ver Time)
slide-2
SLIDE 2

Filtering Example

slide-3
SLIDE 3

Smoothing

  • Divide evidence e1:t to e1:k, ek+1,t

– P(Xk | e1:t) = P(Xk | e1:k, ek+1,t)

= a P(Xk | e1:k) P(ek+1,t| Xk,ek+1,t) = a P(Xk | e1:k) P(ek+1,t| Xk) = af1:kbk+1:t

  • Backward message computed by a

backwards recursion

slide-4
SLIDE 4

Hidden Markov Models

  • Dominant method for automatic speech

recognition

  • Temporal probabilistic model in which the

state of the process is described by a single discrete random variable

  • Simple and elegant matrix implementation
  • f all the basic algorithms
slide-5
SLIDE 5

HMM segmentation

Hidden

p( | )

Observed Model 1 2

P( | )

3 4 5 t t-1 Aucouturier & Sandler, AES 01

slide-6
SLIDE 6

Kalman Filter

Modelling systems described by a set of continuous variables

Tracking a bird flying Xt = X,Y,Z, dX, dY, dZ Airplanes, robots, ecosystems...

Gaussian prior, linear Gaussian transition model and sensor model

i.e next state is a linear function of current state plus some Gaussian noise Key property: linear Gaussian family remains closed under standard Baysian network

  • perations
slide-7
SLIDE 7

Kalman Filter

Position, velocity and position measurement Basically we forward messages (means + covariance matrix) to produce new message (means + covariance matrix)

slide-8
SLIDE 8

Kalman filtering

slide-9
SLIDE 9

Kalman Smoothing

slide-10
SLIDE 10

Outline

Learning agents Inductive Learning Decision tree learning Measuring learning performance

slide-11
SLIDE 11

Learning

Learning is essential for unknown environs

Designer lacks omniscience

Learning is useful as a system construction method

Expose the agent to reality rather than trying to write it down

Learning modifies agents decision making mechanisms to improve performance

slide-12
SLIDE 12

Learning Agents

slide-13
SLIDE 13

Learning Element

Design is dictated by:

Type of performance element Functional component to be learned How functional component is represented Type of feedback

slide-14
SLIDE 14

Example scenarios

Supervised learning: correct answers for each training instance Unsupervised learning: no feedback Reinforcement learning: occasional rewards Other terms: Classification, Regression, Clustering, Online, Offline

slide-15
SLIDE 15

Inductive Learning

Simplest form: “learn” a function from examples (training set)

An example is a pair x, f(x) Find hypothesis h(x) such that h(x) ≈ f(x) given a training set of example

Highly simplified. Why ?

slide-16
SLIDE 16

Inductive Learning

Learning a function is simplified:

Ignores prior knowledge Assumes deterministic observable environment Assumes examples are given Assumes that the agent wants to learn f – why ?

slide-17
SLIDE 17

Inductive Learning Method

Construct/adjust h to agree with f on training set e.g curve fitting

slide-18
SLIDE 18

Inductive Learning Method

Construct/adjust h to agree with f on training set e.g curve fitting

slide-19
SLIDE 19

Inductive Learning Method

Construct/adjust h to agree with f on training set e.g curve fitting

CONSISTENT HYPOTHESIS

slide-20
SLIDE 20

Inductive Learning Method

Construct/adjust h to agree with f on training set e.g curve fitting

CONSISTENT HYPOTHESIS

slide-21
SLIDE 21

Inductive Learning Method

Construct/adjust h to agree with f on training set e.g curve fitting

CONSISTENT HYPOTHESIS OCCAM'S RAZOR Maximize a combination

  • f consistency

and simplicity