Biologically inspired Vision on a Modular Reconfigurable System - - PowerPoint PPT Presentation

biologically inspired vision on a modular reconfigurable
SMART_READER_LITE
LIVE PREVIEW

Biologically inspired Vision on a Modular Reconfigurable System - - PowerPoint PPT Presentation

Biologically inspired Vision on a Modular Reconfigurable System (BMV) Jon Binney RESL Lior Elazary ILAB Nadeesha Ranasinghe PRL Overview of Presentation BMV Combination of research ideas from our labs Search and Rescue


slide-1
SLIDE 1

Biologically inspired Vision on a Modular Reconfigurable System (BMV)

Jon Binney RESL Lior Elazary ILAB Nadeesha Ranasinghe PRL

slide-2
SLIDE 2

Overview of Presentation

BMV

Combination of research ideas from our labs

Search and Rescue (S&R) Scenario

Locate injured people Identify hazards for the safety of the rescue crews Obstacle avoidance and exploration

Modular Reconfigurable Robot Stereo Vision & Structure From Motion Saliency Based Identification & Tracking

slide-3
SLIDE 3

Robots & Vision

Why should we use robots for S&R?

Dangers present in a disaster area

Collapsing structures Elemental hazards (fire, water, electricity, gas) Minimize risking lives

Cheaper Faster Tolerant to Elemental hazards to some extent

Why should we use Vision?

Very powerful sensor for high-level task-based sensing More noise tolerant than most other sensors

slide-4
SLIDE 4

Modular Reconfigurable Robot (SuperBot)

  • Standard Module’s Capabilities
  • 3 DOF (x - yaw, y - pitch, z - yaw)
  • IR Communication & Proximity Sensing
  • 3D Accelerometer
  • One-way Radio Communications
  • 6 Docks for Reconfiguration
  • Additional WiFi Communications & Wireless Camera Module
slide-5
SLIDE 5

Go Anywhere with Shapes & Gaits

  • Reconfigurable Shape
  • Changing the global shape to maneuver through various obstacles
  • Reposition cameras
  • Track, Snake, Spiral, Biped, Quadruped, Hexapod etc.
  • Gaits
  • Ways of moving
  • Restrictions imposed by Shape
  • Rolling, Sidewinder, Caterpillar, Walkers, Climbers etc.
slide-6
SLIDE 6

What will the robot see?

slide-7
SLIDE 7

Stereo Vision & Structure From Motion

Calibration Dense Stereo Structure from Motion Fitting a Model to Data

slide-8
SLIDE 8

Calibration

Each camera creates

images which are warped in slightly different ways

Calibration uses a known

target to calculate these parameters

Zhang, Z. (2000), 'A Flexible New Technique for Camera Calibration', IEEE Transactions on Pattern Analysis and Machine Intelligence 22(11), 1330-1334.

slide-9
SLIDE 9

Dense Stereo

Goal: Find the 3D depth of

each pixel

Input: A left and a right

image

slide-10
SLIDE 10

Steps in Dense Stereo

  • 1. For each pixel in the left image, find the

corresponding pixel in the right image (disparity)

  • 2. Use knowledge of the relative positions of

the two cameras to triangulate the 3D position

  • f the point
slide-11
SLIDE 11

Use of Dense Stereo for Robotics

Provides a TON of information about the depth of 3D points in

front of the robot, which can be used for obstacle avoidance

Can be combined with a SLAM technique to provide estimates

  • f robot position over time

1) Hirschmuller, H.; Innocent, P.R. & Garibaldi, J. (2002), 'Real-Time Correlation-Based Stereo Vision with Reduced Border Errors', International Journal of Computer Vision 47, 229-246. 2) D. Scharstein and R. Szeliski. A Taxonomy and Evaluation of Dense Two-Frame Stereo Correspondence Algorithms. IJCV 47(1/2/3):7-42, April- June 2002. (for the test images)

slide-12
SLIDE 12

Structure from Motion

Goal: Use multiple images to build a 3D model of the

environment (and possibly calibrate the camera at the same time)

How is this different from stereo reconstruction?

slide-13
SLIDE 13

SfM for Robotics

SfM works naturally with robotics because the

robot gets a sequence of images as it moves through its environment

Many SfM techniques have the advantage of

not needing calibrated cameras

Solves for the positions of the cameras (and

robot) as it solves for the structure of the environment

Pollefeys, M.; Gool, L.V.; Vergauwen, M.; Verbiest, F.; Cornelis, K.; Tops, J. & Koch, R. (2004), 'Visual Modeling with a Hand-Held Camera', International Journal of Computer Vision 59(3), 207-232.

slide-14
SLIDE 14

Fitting a Model to Data

Dense Stereo and Structure from Motion (SfM) result

in a set of points in 3D. How do we turn this into a more useful model of the environment?

slide-15
SLIDE 15

Fitting a Model to Data

Assume some basic structure for the environment Assume some error distribution for points Find the most 'Likely' model using a technique like

EM x

Liu, Y.; Emery, R.; Chakrabarti, D.; Burgard, W. & Thrun, S. (2001),'Using EM to Learn 3D Models of Indoor Environments with Mobile Robots''Proceedings of the International Conference on Machine Learning (ICML)'.

slide-16
SLIDE 16

Saliency Based Identification & Tracking

Bottom-Up Saliency Top-Down Saliency Navigation

slide-17
SLIDE 17

Saliency Based Identification & Tracking

slide-18
SLIDE 18

Saliency Based Identification & Tracking

slide-19
SLIDE 19

Visual Search

  • Free examination
  • Estimate material circumstances
  • f family
  • Give ages of the people
  • Surmise what family has been doing

before arrival of “unexpected visitor”

  • Remember clothes worn by the

people

  • Remember position of people and
  • bjects
  • Estimate how long the “unexpected

visitor” has been away from family

slide-20
SLIDE 20

Attention

Given an input image, predict which location in

the image will automatically attract your attention.

Vision is expensive and ambiguous.

Requires a large amount of processing to compute

high-order information, i.e object recognition.

Too much information at one time can hinder the

  • system. Categorizing two objects at one time as appose

to one at a time.

Don’t want to search the whole image.

Saliency gives clues of where the object of interest

would be.

Need to find likely interest positions based on

simple features.

slide-21
SLIDE 21
slide-22
SLIDE 22

Natural scenes

slide-23
SLIDE 23

Many applications

Including… Video compression Automatic target detection Driver alerting & monitoring Surveillance Robotics Animation of virtual agents Analysis of satellite imagery Star Wars binoculars … many more.

slide-24
SLIDE 24
slide-25
SLIDE 25

Example - Beobot

slide-26
SLIDE 26

Top Down Attention Where is Waldo?

Knowing the target in a

visual space leads to faster search (Vickery et. al 2005, Wolfe 1994)

What features from the

  • bject do we learn for

biasing.

How can biases be

applied in the most efficient manner.

slide-27
SLIDE 27

Selecting Features Using Saliency

  • Salient location within an
  • bject would remain the same

under various transformations.

  • Get the raw

center-surround features from the most salient location.

  • Choose the most salient

location within each submap.

  • Using Bayesian decision

theory to decide object classification (Richard et al. 2001)

  • Bias image using learned

likelihood function.

slide-28
SLIDE 28

Results: Search Task for houses

slide-29
SLIDE 29

Results: Search Task for houses

slide-30
SLIDE 30

Results: Search Task for houses and roads

slide-31
SLIDE 31

Landmark Navigation Using Biased Attention

Toy jeep fitted with a

wireless (1.2GHz) camera and a standard RC remote control.

The camera receiver was

connected to a capture card.

The RC remote control was

connected to a device (sc8000) which allowed the computer to control the robot.

slide-32
SLIDE 32

Landmark Navigation Using Biased Attention

  • Left Image
  • Yellow box and dot represent the landmark found using SIFT.
  • Blue dot represents the current tracking location using biased attention.
  • Right Image is the resulting saliency map.
  • Saliency map is reversed.
  • Left bottom text (from left to right).
  • Image capture rate frame per second.
  • Biased saliency map frame per second (tracking).
  • Current landmark id the robot is navigating toward.
  • Can be thought of as the current leg in the path.
slide-33
SLIDE 33

Landmark Navigation Using Biased Attention

slide-34
SLIDE 34

Conclusion

Modular reconfigurable robots in S&R Saliency to identify possible targets Reconstruct the structure around the target