Overview of Robot Perception Prof. Yuke Zhu Fall 2020 CS391R: - - PowerPoint PPT Presentation

overview of robot perception
SMART_READER_LITE
LIVE PREVIEW

Overview of Robot Perception Prof. Yuke Zhu Fall 2020 CS391R: - - PowerPoint PPT Presentation

Overview of Robot Perception Prof. Yuke Zhu Fall 2020 CS391R: Robot Learning (Fall 2020) 1 Logistics Office Hours Instructor: 4-5pm Wednesdays (Zoom) or by appointment TA: 10:15-11:15am Mondays (Zoom) or by appointment Presentation Sign-Up:


slide-1
SLIDE 1

CS391R: Robot Learning (Fall 2020)

Overview of Robot Perception

1

  • Prof. Yuke Zhu

Fall 2020

slide-2
SLIDE 2

CS391R: Robot Learning (Fall 2020) 2

Logistics

Office Hours

Instructor: 4-5pm Wednesdays (Zoom) or by appointment TA: 10:15-11:15am Mondays (Zoom) or by appointment Presentation Sign-Up: Deadline Today (EOD) First review due: Wednesday 9:59pm (one review: Mask-RCNN or YOLO) Student Self-Introduction

slide-3
SLIDE 3

CS391R: Robot Learning (Fall 2020) 3

Today’s Agenda

  • What is Robot Perception?
  • Robot Vision vs. Computer Vision
  • Landscape of Robot Perception

○ neural network architectures ○ representation learning algorithms ○ state estimation tasks ○ embodiment and active perception

  • Quick Review of Deep Learning (if time permits)
slide-4
SLIDE 4

CS391R: Robot Learning (Fall 2020) 4

[Levine et al. JMLR 2016] [Bohg et al. ICRA 2018] [Sa et al. IROS 2014] Perceive Act Perceive Act Act Perceive

A key challenge in Ro Robo bot Learning is to close the perception-action loop.

slide-5
SLIDE 5

CS391R: Robot Learning (Fall 2020) 5

What is Robot Perception?

Making sense of the unstructured real world…

  • Incomplete knowledge of objects and scene
  • Environment dynamics and other agents
  • Imperfect actions may lead to failure
slide-6
SLIDE 6

CS391R: Robot Learning (Fall 2020) 6

Robotic Sensors

Making contact of the physical world through multimodal senses

slide-7
SLIDE 7

CS391R: Robot Learning (Fall 2020) 7

Robotic Sensors

Making contact of the physical world through multimodal senses

[Source: HKU Advanced Robotics Laboratory]

slide-8
SLIDE 8

CS391R: Robot Learning (Fall 2020) 8

Robot Vision vs. Computer Vision

[Detectron - Facebook AI Research] [Zeng et al., IROS 2018]

Robot vision is embodied, active, and environmentally situated.

slide-9
SLIDE 9

CS391R: Robot Learning (Fall 2020) 9

Robot Vision vs. Computer Vision

Robot vision is embodied, active, and environmentally situated.

  • Embodied: Robots have physical bodies and experience the world directly. Their

actions are part of a dynamic with the world and have immediate feedback on their

  • wn sensation.
  • Active: Robots are active perceivers. It knows why it wishes to sense, and chooses

what to perceive, and determines how, when and where to achieve that perception.

  • Situated: Robots are situated in the world. They do not deal with abstract

descriptions, but with the here and now of the world directly influencing the behavior

  • f the system.

[Brooks 1991; Bajcsy 2018]

slide-10
SLIDE 10

CS391R: Robot Learning (Fall 2020) 10

Robot Perception: Landsc scape

What you will learn in the chapter of Robotics and Perception

  • 1. Modalities: neural network architectures designed for different sensory modalities
  • 2. Representations: representation learning algorithms without strong supervision
  • 3. Tasks: state estimation tasks for robot navigation and manipulation
  • 4. Embodiment: active perception for embodied visual intelligence
slide-11
SLIDE 11

CS391R: Robot Learning (Fall 2020) 11

Robot Perception: Mo Modalities

Pixels (from RGB cameras) Point cloud (from structure sensors)

(x1, y1, z1) (x2, y2, z2)

[Source: PointNet++; Qi et al. 2016]

Time series (from F/T sensors) Tactile data (from the GelSights sensors)

[Source: Calandra et al. 2018] [Source: Lee*, Zhu*, et al. 2018]

slide-12
SLIDE 12

CS391R: Robot Learning (Fall 2020) 12

Robot Perception: Mo Modalities

Week 2: Object Detection (Pixels) Week 3: 3D Point Cloud

More sensory modalities in later weeks…

How can we design the neural network architectures that can effectively process raw sensory data in vastly different forms?

slide-13
SLIDE 13

CS391R: Robot Learning (Fall 2020) 13

Robot Perception: Represe sentations

A fundamental problem in robot perception is to learn the proper representations

  • f the unstructured world.

[Source: Stanford CS331b]

slide-14
SLIDE 14

CS391R: Robot Learning (Fall 2020) 14

Robot Perception: Represe sentations

“Solving a problem simply means representing it so as to make the solution transparent.” Herbert A. Simon, Sciences of the Artificial Our secret weapon? Learning

slide-15
SLIDE 15

CS391R: Robot Learning (Fall 2020) 15

[6.S094, MIT]

slide-16
SLIDE 16

CS391R: Robot Learning (Fall 2020) 16

Robot Perception: Represe sentations

How can we learn representations of the world with limited supervision?

Structural priors (inductive biases) Interaction and movement (embodiment)

“N “Nature” “N “Nurture”

+

Week k 3 (Thu) Week k 4 (Tue)

babies learning by playing

slide-17
SLIDE 17

CS391R: Robot Learning (Fall 2020) 17

Robot Perception: Represe sentations

How can we learn representations that fuse multiple sensory modalities together?

[The McGurk Effect, BBC]

Is seeing believing?

https://www.youtube.com/watch?v=2k8fHR9jKVM

slide-18
SLIDE 18

CS391R: Robot Learning (Fall 2020) 18

Robot Perception: Represe sentations

How can we learn representations that fuse multiple sensory modalities together?

[Lee*, Zhu*, et al. 2018]

1 2 3 4 5 6

1 2 Reaching 3 4 Alignment 5 6 Insertion

combining vision and force for manipulation Week k 4 Thu: Multimodal Sensor Fusion

slide-19
SLIDE 19

CS391R: Robot Learning (Fall 2020) 19

Robot Perception: Tasks sks

State Representation

Perception & Computer Vision Robot Control & Decision Making

Noisy Sensory Data

slide-20
SLIDE 20

CS391R: Robot Learning (Fall 2020) 20

Robot Perception: Tasks sks

State Representation

Perception & Computer Vision Robot Control & Decision Making

Noisy Sensory Data

Localization (Week 5 Tue) Pose Estimation (Week 5 Thu) Visual Tracking (Week 6 Tue)

slide-21
SLIDE 21

CS391R: Robot Learning (Fall 2020) 21

Robot Perception: Tasks sks

State Representation

Robot Control & Decision Making

Noisy Sensory Data

Perception & Computer Vision

http://www.probabilistic-robotics.org/

slide-22
SLIDE 22

CS391R: Robot Learning (Fall 2020) 22

Robot Perception: Tasks sks

: state : observation : action : transition model (motion model) : measurement model (observation model)

xt

<latexit sha1_base64="TheT5UxEhslRmdBZl0X0uTguJY=">AB6nicbVBNS8NAEJ3Ur1q/qh69LBbBU0lEqMeiF48V7Qe0oWy2m3bpZhN2J2IJ/QlePCji1V/kzX/jts1BWx8MPN6bYWZekEh0HW/ncLa+sbmVnG7tLO7t39QPjxqmTjVjDdZLGPdCajhUijeRIGSdxLNaRI3g7GNzO/ci1EbF6wEnC/YgOlQgFo2il+6c+9sVt+rOQVaJl5MK5Gj0y1+9QczSiCtkhrT9dwE/YxqFEzyamXGp5QNqZD3rVU0YgbP5ufOiVnVhmQMNa2FJK5+nsio5ExkyiwnRHFkVn2ZuJ/XjfF8MrPhEpS5IotFoWpJBiT2d9kIDRnKCeWUKaFvZWwEdWUoU2nZEPwl9eJa2LqndZrd1dVurXeRxFOIFTOAcPalCHW2hAExgM4Rle4c2Rzovz7nwsWgtOPnMf+B8/gB0eI3t</latexit>

zt

<latexit sha1_base64="/JoDb3w1Lbpxa+WZO8WaYqU2K8=">AB6nicbVBNS8NAEJ3Ur1q/qh69LBbBU0lEqMeiF48V7Qe0oWy2m3bpZhN2J0IN/QlePCji1V/kzX/jts1BWx8MPN6bYWZekEh0HW/ncLa+sbmVnG7tLO7t39QPjxqmTjVjDdZLGPdCajhUijeRIGSdxLNaRI3g7GNzO/ci1EbF6wEnC/YgOlQgFo2il+6c+9sVt+rOQVaJl5MK5Gj0y1+9QczSiCtkhrT9dwE/YxqFEzyamXGp5QNqZD3rVU0YgbP5ufOiVnVhmQMNa2FJK5+nsio5ExkyiwnRHFkVn2ZuJ/XjfF8MrPhEpS5IotFoWpJBiT2d9kIDRnKCeWUKaFvZWwEdWUoU2nZEPwl9eJa2LqndZrd1dVurXeRxFOIFTOAcPalCHW2hAExgM4Rle4c2Rzovz7nwsWgtOPnMf+B8/gB3hI3v</latexit>

ut

<latexit sha1_base64="/eqKUhd47miGuwmnBPvPmbKGWM=">AB6nicbVBNS8NAEJ3Ur1q/qh69LBbBU0mkUI9FLx4r2g9oQ9lsN+3SzSbsToQS+hO8eFDEq7/Im/GbZuDtj4YeLw3w8y8IJHCoOt+O4WNza3tneJuaW/4PCofHzSNnGqGW+xWMa6G1DpVC8hQIl7ya0yiQvBNMbud+54lrI2L1iNOE+xEdKREKRtFKD+kAB+WKW3UXIOvEy0kFcjQH5a/+MGZpxBUySY3peW6CfkY1Cib5rNRPDU8om9AR71mqaMSNny1OnZELqwxJGtbCslC/T2R0ciYaRTYzoji2Kx6c/E/r5dieO1nQiUpcsWi8JUEozJ/G8yFJozlFNLKNPC3krYmGrK0KZTsiF4qy+vk/ZV1atV6/e1SuMmj6MIZ3AOl+BHRpwB01oAYMRPMrvDnSeXHenY9la8HJZ07hD5zPH2/mjeo=</latexit>

p(xt|ut, xt−1)

<latexit sha1_base64="rcCQoP/aEiLnvtOiarySz7BnQE=">AB+3icbVDLSsNAFJ3UV62vWJduBotQUsihbosunFZwT6gDWEynbRDJw9mbqQl5lfcuFDErT/izr9x2mah1QMXDufcy73eLHgCizryisrW9sbhW3Szu7e/sH5mG5o6JEUtamkYhkzyOKCR6yNnAQrBdLRgJPsK43uZn73QcmFY/Ce5jFzAnIKOQ+pwS05JrluDp14TFx4XzqpnBhZ2euWbFq1gL4L7FzUkE5Wq75ORhGNAlYCFQpfq2FYOTEgmcCpaVBoliMaETMmJ9TUMSMOWki9szfKqVIfYjqSsEvFB/TqQkUGoWeLozIDBWq95c/M/rJ+BfOSkP4wRYSJeL/ERgiPA8CDzklEQM0IlVzfiumYSEJBx1XSIdirL/8lncuaXa817uqV5nUeRxEdoxNURTZqoCa6RS3URhRN0RN6Qa9GZjwb8b7srVg5DNH6BeMj29hwJQG</latexit>

p(zt|xt)

<latexit sha1_base64="1EdS5PnvuyIMSClqyAfSQ+0cy5M=">AB8XicbVBNS8NAEJ34WetX1aOXYBHqpSRSqMeiF48V7Ae2IWy2m3bpZhN2J2Kt/RdePCji1X/jzX/jts1BWx8MPN6bYWZekAiu0XG+rZXVtfWNzdxWfntnd2+/cHDY1HGqKGvQWMSqHRDNBJesgRwFayeKkSgQrBUMr6Z+654pzWN5i6OEeRHpSx5yStBId0np0cenBx/P/ELRKTsz2MvEzUgRMtT9wle3F9M0YhKpIFp3XCdBb0wUcirYJN9NUsIHZI+6xgqScS0N5dPLFPjdKzw1iZkmjP1N8TYxJpPYoC0xkRHOhFbyr+53VSDC+8MZdJikzS+aIwFTbG9vR9u8cVoyhGhCquLnVpgOiCEUTUt6E4C6+vEya52W3Uq7eVIq1yOHBzDCZTAhSrU4Brq0AKEp7hFd4sb1Y79bHvHXFymaO4A+szx9grJC9</latexit>

bel(xt)

<latexit sha1_base64="7dqwAjIevYMGVrenjYTqQEQdk/c=">AB73icbVDLSgNBEOz1GeMr6tHLYBDiJexKIB6DXjxGMA9IljA7mU2GzM6uM71iCPkJLx4U8ervePNvnCR70MSChqKqm+6uIJHCoOt+O2vrG5tb27md/O7e/sFh4ei4aeJUM95gsYx1O6CGS6F4AwVK3k40p1EgeSsY3cz81iPXRsTqHscJ9yM6UCIUjKV2gGXpaceXvQKRbfszkFWiZeRImSo9wpf3X7M0ogrZJIa0/HcBP0J1SiY5N8NzU8oWxEB7xjqaIRN/5kfu+UnFulT8JY21JI5urviQmNjBlHge2MKA7NsjcT/M6KYZX/kSoJEWu2GJRmEqCMZk9T/pCc4ZybAlWthbCRtSTRnaiPI2BG/5VXSvCx7lXL1rlKsXWdx5OAUzqAEHlShBrdQhwYwkPAMr/DmPDgvzrvzsWhdc7KZE/gD5/MHgmSPow=</latexit>

: belief

State estimation methods: Bayes Filtering

slide-23
SLIDE 23

CS391R: Robot Learning (Fall 2020) 23

Robot Perception: Tasks sks

State estimation methods: Bayes Filtering

: state : observation : action : transition model (motion model) : measurement model (observation model)

xt

<latexit sha1_base64="TheT5UxEhslRmdBZl0X0uTguJY=">AB6nicbVBNS8NAEJ3Ur1q/qh69LBbBU0lEqMeiF48V7Qe0oWy2m3bpZhN2J2IJ/QlePCji1V/kzX/jts1BWx8MPN6bYWZekEh0HW/ncLa+sbmVnG7tLO7t39QPjxqmTjVjDdZLGPdCajhUijeRIGSdxLNaRI3g7GNzO/ci1EbF6wEnC/YgOlQgFo2il+6c+9sVt+rOQVaJl5MK5Gj0y1+9QczSiCtkhrT9dwE/YxqFEzyamXGp5QNqZD3rVU0YgbP5ufOiVnVhmQMNa2FJK5+nsio5ExkyiwnRHFkVn2ZuJ/XjfF8MrPhEpS5IotFoWpJBiT2d9kIDRnKCeWUKaFvZWwEdWUoU2nZEPwl9eJa2LqndZrd1dVurXeRxFOIFTOAcPalCHW2hAExgM4Rle4c2Rzovz7nwsWgtOPnMf+B8/gB0eI3t</latexit>

zt

<latexit sha1_base64="/JoDb3w1Lbpxa+WZO8WaYqU2K8=">AB6nicbVBNS8NAEJ3Ur1q/qh69LBbBU0lEqMeiF48V7Qe0oWy2m3bpZhN2J0IN/QlePCji1V/kzX/jts1BWx8MPN6bYWZekEh0HW/ncLa+sbmVnG7tLO7t39QPjxqmTjVjDdZLGPdCajhUijeRIGSdxLNaRI3g7GNzO/ci1EbF6wEnC/YgOlQgFo2il+6c+9sVt+rOQVaJl5MK5Gj0y1+9QczSiCtkhrT9dwE/YxqFEzyamXGp5QNqZD3rVU0YgbP5ufOiVnVhmQMNa2FJK5+nsio5ExkyiwnRHFkVn2ZuJ/XjfF8MrPhEpS5IotFoWpJBiT2d9kIDRnKCeWUKaFvZWwEdWUoU2nZEPwl9eJa2LqndZrd1dVurXeRxFOIFTOAcPalCHW2hAExgM4Rle4c2Rzovz7nwsWgtOPnMf+B8/gB3hI3v</latexit>

ut

<latexit sha1_base64="/eqKUhd47miGuwmnBPvPmbKGWM=">AB6nicbVBNS8NAEJ3Ur1q/qh69LBbBU0mkUI9FLx4r2g9oQ9lsN+3SzSbsToQS+hO8eFDEq7/Im/GbZuDtj4YeLw3w8y8IJHCoOt+O4WNza3tneJuaW/4PCofHzSNnGqGW+xWMa6G1DpVC8hQIl7ya0yiQvBNMbud+54lrI2L1iNOE+xEdKREKRtFKD+kAB+WKW3UXIOvEy0kFcjQH5a/+MGZpxBUySY3peW6CfkY1Cib5rNRPDU8om9AR71mqaMSNny1OnZELqwxJGtbCslC/T2R0ciYaRTYzoji2Kx6c/E/r5dieO1nQiUpcsWi8JUEozJ/G8yFJozlFNLKNPC3krYmGrK0KZTsiF4qy+vk/ZV1atV6/e1SuMmj6MIZ3AOl+BHRpwB01oAYMRPMrvDnSeXHenY9la8HJZ07hD5zPH2/mjeo=</latexit>

p(xt|ut, xt−1)

<latexit sha1_base64="rcCQoP/aEiLnvtOiarySz7BnQE=">AB+3icbVDLSsNAFJ3UV62vWJduBotQUsihbosunFZwT6gDWEynbRDJw9mbqQl5lfcuFDErT/izr9x2mah1QMXDufcy73eLHgCizryisrW9sbhW3Szu7e/sH5mG5o6JEUtamkYhkzyOKCR6yNnAQrBdLRgJPsK43uZn73QcmFY/Ce5jFzAnIKOQ+pwS05JrluDp14TFx4XzqpnBhZ2euWbFq1gL4L7FzUkE5Wq75ORhGNAlYCFQpfq2FYOTEgmcCpaVBoliMaETMmJ9TUMSMOWki9szfKqVIfYjqSsEvFB/TqQkUGoWeLozIDBWq95c/M/rJ+BfOSkP4wRYSJeL/ERgiPA8CDzklEQM0IlVzfiumYSEJBx1XSIdirL/8lncuaXa817uqV5nUeRxEdoxNURTZqoCa6RS3URhRN0RN6Qa9GZjwb8b7srVg5DNH6BeMj29hwJQG</latexit>

p(zt|xt)

<latexit sha1_base64="1EdS5PnvuyIMSClqyAfSQ+0cy5M=">AB8XicbVBNS8NAEJ34WetX1aOXYBHqpSRSqMeiF48V7Ae2IWy2m3bpZhN2J2Kt/RdePCji1X/jzX/jts1BWx8MPN6bYWZekAiu0XG+rZXVtfWNzdxWfntnd2+/cHDY1HGqKGvQWMSqHRDNBJesgRwFayeKkSgQrBUMr6Z+654pzWN5i6OEeRHpSx5yStBId0np0cenBx/P/ELRKTsz2MvEzUgRMtT9wle3F9M0YhKpIFp3XCdBb0wUcirYJN9NUsIHZI+6xgqScS0N5dPLFPjdKzw1iZkmjP1N8TYxJpPYoC0xkRHOhFbyr+53VSDC+8MZdJikzS+aIwFTbG9vR9u8cVoyhGhCquLnVpgOiCEUTUt6E4C6+vEya52W3Uq7eVIq1yOHBzDCZTAhSrU4Brq0AKEp7hFd4sb1Y79bHvHXFymaO4A+szx9grJC9</latexit>

What if models are hard to specify? Learning

bel(xt)

<latexit sha1_base64="7dqwAjIevYMGVrenjYTqQEQdk/c=">AB73icbVDLSgNBEOz1GeMr6tHLYBDiJexKIB6DXjxGMA9IljA7mU2GzM6uM71iCPkJLx4U8ervePNvnCR70MSChqKqm+6uIJHCoOt+O2vrG5tb27md/O7e/sFh4ei4aeJUM95gsYx1O6CGS6F4AwVK3k40p1EgeSsY3cz81iPXRsTqHscJ9yM6UCIUjKV2gGXpaceXvQKRbfszkFWiZeRImSo9wpf3X7M0ogrZJIa0/HcBP0J1SiY5N8NzU8oWxEB7xjqaIRN/5kfu+UnFulT8JY21JI5urviQmNjBlHge2MKA7NsjcT/M6KYZX/kSoJEWu2GJRmEqCMZk9T/pCc4ZybAlWthbCRtSTRnaiPI2BG/5VXSvCx7lXL1rlKsXWdx5OAUzqAEHlShBrdQhwYwkPAMr/DmPDgvzrvzsWhdc7KZE/gD5/MHgmSPow=</latexit>

: belief Exa xample: Particle Filter Localization

slide-24
SLIDE 24

CS391R: Robot Learning (Fall 2020) 24

Robot Perception: Em Embo bodi diment

Input-Output Picture (Susan Hurley, 1998) Conve ventional Vi View of

  • f Pe

Percept ption

[Action in Perception, Alva Noë 2004]

  • Perception is the process of building an internal

representation of the environment

  • Perception is input from world to mind, and action

is output from mind to world, thought is the mediating process.

slide-25
SLIDE 25

CS391R: Robot Learning (Fall 2020) 25

Robot Perception: Em Embo bodi diment

Kitten Carousel (Held and Hein, 1963) Em Embo bodi died Vi View of

  • f Pe

Percept ption

  • As the active cat (A) walks, the other cat (P) moves

and perceives the environment passively.

  • Only the active cat develops normal perception

through self-actuated movement.

  • The passive cat suffers from perception problems,

such as 1) not blinking when objects approach, and 2) hitting the walls.

slide-26
SLIDE 26

CS391R: Robot Learning (Fall 2020) 26

Robot Perception: Em Embo bodi diment

Pebbles (James J. Gibson 1966) Em Embo bodi died Vi View of

  • f Pe

Percept ption

  • Subjects asked to find a reference object among a

set of irregularly-shaped objects

  • Three groups

a. Passive observers of one static image (49%) b. Observers of moving shapes (72%) c. Interactive observers (99%)

  • The ability to condition input signals with actions is

crucial to perception.

slide-27
SLIDE 27

CS391R: Robot Learning (Fall 2020) 27

Robot Perception: Em Embo bodi diment

Take ke-home messa ssages

  • Perceptual experiences do not present the sense in the way that a photograph does.
  • Perception is developed by an embodied agent through actively exploring in the

physical world.

  • “We see in order to move; we move in order to see.” – William Gibson
slide-28
SLIDE 28

CS391R: Robot Learning (Fall 2020) 28

Robot Perception: Em Embo bodi diment

Week k 6 (Thu) – Active ve Perception: How can embodied agents (robots) improve perception based on visual experiences through active exploration?

View Selection Physical Interaction

[Ramakrishnan et al. 2019] [Pinto et al. 2016]

slide-29
SLIDE 29

CS391R: Robot Learning (Fall 2020) 29

Research Frontier: Closi sing the Perception-Ac Action Loop Perception Action Robots

How robots’ intelligent behaviors are guided by their interactive perception How robots develop better perception from embodied sensorimotor experiences

slide-30
SLIDE 30

CS391R: Robot Learning (Fall 2020) 30

Visual Processing Methods

Staged Visual Recognition Pipeline End-to-end Deep Learning What is new since 1980s?

slide-31
SLIDE 31

CS391R: Robot Learning (Fall 2020) 31

Quick Review of Deep Learning: Ar Artificial Neurons

Bi Biological Ne Neuron Computational building block for the brain Ar Artificial Neuron Computational building block for the neural network

No Note: Many differences exist – be careful with the brain analogies!

[Dendritic Computation, Michael London and Michael Hausser 2015]

slide-32
SLIDE 32

CS391R: Robot Learning (Fall 2020) 32

Quick Review of Deep Learning: Convo volutional Networks ks

slide-33
SLIDE 33

CS391R: Robot Learning (Fall 2020) 33

Quick Review of Deep Learning: Ful Fully-Connected Laye yers

[Source: Stanford CS231N]

What is the dimension of W ?

slide-34
SLIDE 34

CS391R: Robot Learning (Fall 2020) 34

Quick Review of Deep Learning: Convo volutional Laye yers

[Source: Stanford CS231N]

slide-35
SLIDE 35

CS391R: Robot Learning (Fall 2020) 35

Quick Review of Deep Learning: Convo volutional Laye yers

[Source: Stanford CS231N]

slide-36
SLIDE 36

CS391R: Robot Learning (Fall 2020) 36

Quick Review of Deep Learning: Convo volutional Laye yers

[Source: Stanford CS231N]

slide-37
SLIDE 37

CS391R: Robot Learning (Fall 2020) 37

Quick Review of Deep Learning: Convo volutional Laye yers

[Source: Stanford CS231N]

slide-38
SLIDE 38

CS391R: Robot Learning (Fall 2020) 38

Quick Review of Deep Learning: Convo volutional Laye yers

[Source: Stanford CS231N]

slide-39
SLIDE 39

CS391R: Robot Learning (Fall 2020) 39

Quick Review of Deep Learning: Convo volutional Laye yers

[Source: Stanford CS231N]

slide-40
SLIDE 40

CS391R: Robot Learning (Fall 2020) 40

Quick Review of Deep Learning: Convo volutional Laye yers

[Source: Stanford CS231N]

slide-41
SLIDE 41

CS391R: Robot Learning (Fall 2020) 41

Quick Review of Deep Learning: Po Pooling Ope Operati tions

slide-42
SLIDE 42

CS391R: Robot Learning (Fall 2020) 42

Quick Review of Deep Learning: Activa vation Functions

slide-43
SLIDE 43

CS391R: Robot Learning (Fall 2020) 43

Quick Review of Deep Learning: CNN CNN Archit itectures

AlexNet VGG-16 ResNet LeNet

slide-44
SLIDE 44

CS391R: Robot Learning (Fall 2020) 44

Quick Review of Deep Learning: Optimiza zation

Stochastic Gradient Descent (SGD)

θ = θ ηrθJ(θ; x(i); y(i))

<latexit sha1_base64="BKFxODiGs0UWCdpl4EQg4DJQSQo=">ACKHicbZDLSgMxFIYzXmu9V26CRahLiwzIiIWHQjrhTsBTq1nElTG8xkhuSMWIY+jhtfxY2Im59EtN2Fmo9EPLx/+eQnD+IpTDoup/O1PTM7Nx8biG/uLS8slpYW6+ZKNGMV1kI90IwHApFK+iQMkbseYQBpLXg7uzoV+/59qISF1jP+atEG6V6AoGaKV24cTHkegxzSDXeoPL19BIKGdiRelMRw93KQlsTM4ov0x7LQLRbfsjopOgpdBkWR12S68+p2IJSFXyCQY0/TcGFspaBRM8kHeTwyPgd3BLW9aVBy0pHiw7otlU6tBtpexTSkfpzIoXQmH4Y2M4QsGf+ekPxP6+ZYPewlQoVJ8gVGz/UTSTFiA5Tox2hOUPZtwBMC/tXynqgaHNm9D8P6uPAm1vbK3Xz642i9WTrM4cmSTbJES8cgBqZBzckmqhJFH8kzeyLvz5Lw4H87nuHXKyWY2yK9yvr4B7UCkrg=</latexit>

input label learning rate weights

Backpropagation

slide-45
SLIDE 45

CS391R: Robot Learning (Fall 2020) 45

Quick Review of Deep Learning: Feat Featur ures es

[Source: Stanford CS231N]

slide-46
SLIDE 46

CS391R: Robot Learning (Fall 2020) 46

Quick Review of Deep Learning: Im Impl plementa tati tion

Tutorial coming in late September / early October

slide-47
SLIDE 47

CS391R: Robot Learning (Fall 2020) 47

Quick Review of Deep Learning: Reso sources

Online Courses

  • CS231N: Convolutional Neural Networks for Visual Recognition

http://cs231n.stanford.edu/

  • MIT 6.S191: Introduction to Deep Learning

http://introtodeeplearning.com/

Textbooks:

  • Deep Learning. Ian Goodfellow, Yoshua Bengio, Aaron Courville

http://www.deeplearningbook.org/

slide-48
SLIDE 48

CS391R: Robot Learning (Fall 2020) 48

Resources

Related courses at UTCS

  • CS342: Neural Networks
  • CS 376: Computer Vision
  • CS 378 Autonomous Driving
  • CS 393R: Autonomous Robots
  • CS394R: Reinforcement Learning: Theory and Practice

Extended readings:

  • Action-based Theories of Perception, Stanford Encyclopedia of Philosophy
  • Action in Perception, Alva Noë