SlideShare ist ein Scribd-Unternehmen logo
1 von 47
Downloaden Sie, um offline zu lesen
estimation of terrain gradient conditions &
obstacle detection using a monocular
vision-based system
Final Demonstration
Connor Luke Goddard (clg11)
Department of Computer Science, Aberystwyth University
Background
1
background
Figure 1: Tracked features from MER [2].
The ability for an autonomous robot to
navigate from one location to another
in a manner that is both safe, yet
objective is vital to the survivability
of the machine, and the success
of the mission it is undertaking.
Vision based obstacle detection has enjoyed a high level of research
focus over recent years. Less work has concentrated on providing a
means of reliably detecting changes in terrain slope through the
exploitation of observed changes in motion.
2
related work
[1] A Robust Visual Odometry and Precipice Detection System Using
Consumer-grade Monocular Vision, J. Campbell et al., IEEE, 2005
[2] Obstacle Detection using Optical Flow, T. Low and G. Wyeth, School
of Information Technology and Electrical Engineering, University of
Queensland, 2011
[3] Appearance-based Obstacle Detection with Monocular Color
Vision, I. Ulrich and I. Nourbakhsh, AAAI, 2000
Figure 2: Left: Precipice detection [1], Right: Obstacle colour detection [3]
3
project overview
Focus
Investigation into a system capable of utilising a single,
forward-facing colour camera to provide an estimation into current
terrain gradient conditions, obstacle location & characteristics and
robot ego-motion.
Potential Metrics For Observation
• Changes in terrain gradient (i.e. slopes).
• Location & characteristics of positive and negative obstacles (i.e.
rock or pit respectively).
• Current robot speed of travel.
• Changes in robot orientation.
4
Working Hypothesis
5
working hypothesis
“From images captured using a non-calibrated monocular
camera system, analysis of optical flow vectors extracted using
localised appearance-based dense optical flow techniques can
provide certain estimates into the current condition of terrain
gradient, the location and characteristics of obstacles, and robot
ego-motion.”
6
motion parallax
Approach focussed on the effects of motion parallax.
i.e. Objects/features that are at a greater distance from the camera
appear to move less from frame-to-frame than those that are closer.
Figure 3: Typical example of motion parallax [4].
7
inference of terrain gradient & obstacle detection
Key Aspects
1. The exclusive use of appearance-based template matching
techniques to provide a localised variant of dense optical flow
analysis over the use of sparse optical flow techniques relying
upon a high number of available image features.
2. The use of a formalised model to represent detected changes in
vertical displacement as part of efforts to estimate changes in
terrain gradient in addition to the location characteristics of
potential obstacles.
8
vertical displacement model
Provides a mechanism by which to evaluate the prediction:
0Vertical
Displacement(px) 0
Image Row (Height)
(px)
Expected general trend between image row
and vertical displacement demonstrated.
Figure 4: Example of “perfect” vertical
displacement model.
“Features captured
towards the bottom of an image
will show greater displacement
between subsequent frames,
than those features captured
towards the top of the image.”
9
inference of terrain slope & potential obstacles
From observing discrepancies between the current model and
“baseline” model, it should be possible to infer the presence of
potential obstacles:
0
Vertical
Displacement(px)
0
Image Row (Height)
(px)
160 400
Recorded vertical
displacement for row #160
matching model vertical
displacement for row #230
230
Model vertical displacement
Last-recorded vertical displacement
Figure 5: Vertical displacement model indicating the potential presence of a positive obstacle.
10
inference of terrain slope & potential obstacles
Differences in model discrepancy behaviour should provide enough
information to establish between a potential obstacle and a change
in terrain slope:
Figure 6: Example of observed differences between detection of obstacles and change in terrain
slope.
11
robot rotation
Two types of observed motion:
• Translation (Vertical component of optical flow vectors)
• Rotation (Horizontal component of optical flow vectors)
Horizontal component of optical flow vector
(Rotation)
Vertical component of optical flow vector
(Translation)
Figure 7: Diagram demonstrating the expected difference in behaviour between the separated
vertical and horizontal components of observed optical flow vectors calculated from
forward-turning motion.
12
robot speed
By recording multiple “calibration” models at different speeds, it
should be possible to later infer the current speed using the
standard speed-distance-time equation.
Optimum point to infer speed
as lowest risk of interference
from moving objects during
recording and comparing
current displacement with
model.
0
Vertical
Displacement(px)
0
Image Row (Height)
(px)
Vertical displacement model recorded at
constant 5mph
Vertical displacement model recorded at
constant 10mph
Vertical displacement model recorded at
constant 15mph
Figure 8: Diagram indicating examples of alternative vertical displacement models calculated
while travelling at different constant speeds.
13
investigation aims
Primary Aims
1. Establish which out the following appearance-based template
matching similarity measures:
• Euclidean Distance
• Correlation Coefficient
• Histogram Correlation
• Histogram Chi-Square
best supports the predicted positive correlation between the
vertical position of features within an image, and the level of
vertical displacement demonstrated.
2. Use the generated model of vertical displacement to identify
potential obstacles and changes in terrain gradient based upon
the predicted ‘comparative trend’ behaviour.
14
investigation aims
Secondary Aims
1. Further extend the capabilities of the vertical displacement model
to provide estimates into the current speed of travel
demonstrated by the robot.
2. Add support for providing an estimation into the change in
orientation.
15
Experiment Methods
16
gathering datasets
No existing imagery datasets were found to be suitable, therefore
new sets had to be collected.
Figure 9: Camera-rig setup used to capture the experiment datasets from front and back profiles.
17
gathering datasets
Images taken at a number of locations, but four were eventually
selected based upon their variety in terms of terrain texture and
lighting conditions.
Figure 10: Examples of the first frame captured from each of the four location datasets.
18
experiment methods
Over time, investigation focus changed from exploring a series of
reasonably “generalised” research aims, to instead concentrating
almost exclusively on one small, but very important aspect;
Ensuring that accurate results for appearance-based motion
displacement could continue to be achieved, across terrain lacking
any significant features and/or demonstrating highly repetitive
patterns.
Three Experiments
1. Template Matching - Multiple Small Patches
2. Template Matching - Full-width Patches (Non-Scaled)
3. Template Matching - Full-width Patches (Geometrically Scaled)
19
experiment one - multiple small patches
Stage One
Import two consecutive images, and convert from RGB (BGR in
OpenCV) colour space to HSV.
The ’V’ channel is then removed in order to improve robustness to
lighting changes between frames.
20
experiment one - multiple small patches
Stage Two
Extract a percentage-width region of interest (ROI) from centre of
first image.
21
region-of-interest
Why do we need to extract a ROI?
Focus-of-expansion: Objects in images do not actually move in
1-dimension (i.e. straight down the image).
This effect is minimised towards the centre of the image.
Figure 11: Diagram indicating the perceived effects caused by Focus of Expansion. Courtesy: [5]
22
experiment one - multiple small patches
Stage Three
Extract patches of a fixed size around each pixel within the extracted
ROI.
Figure 12: Simplified example of patch extraction within ROI.
23
experiment one - multiple small patches
Stage Four
For each patch extracted from image one, move down through a
localised search window (column) in image two searching for the
best match against the template patch.
Figure 13: Example of “best match” search within local column.
24
experiment one - multiple small patches
Stage Five
Identify the patch within the localised search window that provides
the “best match” via correlation-based matching (e.g. Euclidean
Distance, SSD or Correlation coefficient).
25
experiment one - multiple small patches
Stage Six
Average all measured displacements for each pixel along a given row.
Outliers are removed by ignoring any displacements that lie outside
of (2 x Standard Deviation) of the mean.
26
experiment one - multiple small patches
Repeat Stages 1-6
Repeat stages 1-6 for an entire collection of “calibration” images
taken of flat, unobstructed terrain.
27
experiment one - multiple small patches
Stage Seven
Plot the average displacement for each ROI row, calculated from the
displacements recorded over all calibration images.
28
experiment two - full-width patches (non-scaled)
Focus
Establishing if adopting larger patches in fewer numbers provided
better appearance-based matching accuracy than using many more,
but critically much smaller overlapping patches.
While the underlying method remained the same as in the first
experiment, some key changes were made:
• Extracting a central region of interest relative to the ground plane,
as opposed to extracting a fixed-width column from the image
plane.
• Moving away from multiple overlapping small patches, in favour of
adopting a single, full-width patch to represent a single row in the
image.
29
exhaustive vs. non-exhaustive search
Issue discovered regarding the use of non-exhaustive searching
causing erroneous results. All tests in experiments two and three
were conducted using both exhaustive and non-exhaustive
approaches.
0
EuclideanDistance
0
Point at which non-exhaustive search
potentially stops (local minima)
Location of true "best match" score for
Euclidean Distance (global minima)
Figure 14: Example of potential early termination of non-exhaustive search due to entering local
minima caused by noise, rather than global minima which represents to true best match.
30
perspective distortion calibration tool
To extract a region-of-interest along the ground plane, the system
would need to take account of perspective distortion. Therefore, a
simple calibration tool was implemented, enabling users to define
the required region of interest within an image.
Figure 15: Results of the verification test for the approach towards geometric scaling of template
patch pixel coordinates.
31
experiment three - full-width patches (scaled)
Focus
Add geometric scaling of template patch in order to account for
objects appearing larger as they approach the camera.
Original Template Patch Re-scaled Template Patch (Geometric Transformation of Pixel Coordinates) Re-scaled Template Patch (Linear Interpolation)
Figure 16: Comparison between various approaches of performing scaling of the original template
patch image. In the case of linear interpolation, additional information has been added to image
as a result of “estimating” the colour that lies between two original pixels.
32
experiment three - full-width patches (scaled)
6 Stage Process
1. Obtain the width and height of the current template patch.
2. Obtain the calibrated width of the search window with respect to the
current image row.
3. Calculate the independent scale factor for the width and height between
the template patch and the search window.
4. Treating each pixel as a 2D coordinate in geometric space, scale the
position of each pixel coordinate within the template patch relative to the
position of centre coordinate.
5. For each scaled pixel coordinate, extract the value of the pixel at the
corresponding position with the search window and add to a temporary
“image” data structure.
6. Perform template matching between the original template patch, and the
new “extracted” temporary image.
33
experiment three - full-width patches (scaled)
Testing of Scaling Approach
A simple test was devised to confirm geometric scaling worked as
expected under “perfect” conditions.
Figure 17: Results of the verification test for the approach towards geometric scaling of template
patch pixel coordinates.
34
additional experimental work
Some additional experimental work was conducted as part of the larger
project investigation, but these did not actively contribute to the final
investigation results.
• Investigation of Campbell et al. [1] approach to feature tracking to provide
optical flow field (based heavily upon C# implementation provided by Dr.
Rainer Hessmer (Source))
• Investigation of edge-based template matching discussed by Perveen et
al. [3]
Figure 18: Left: Example of feature tracking to establish optical flow field. Right: Edge-based
template matching using Canny edge detector.
35
Investigation Results
View Results (Courtesy: nbviewer)
36
results
Results can be described as generally inconclusive.
However from the results obtained, we have learned:
1. Under certain terrain conditions, it is possible to potentially
establish a relationship between row height in an image, and
average downwards pixel displacement.
2. The performance of each appearance-based similarity measure
can vary widely, based upon factors including:
• Terrain texture (e.g. highly patterned, isotropic or homogenous)
• Patch size (Balance between reducing “noise” and representing
true displacement behaviour)
• Exhaustive vs. Non-exhaustive search
• Scaling vs. Non-scaling
37
results: approach comparison - ‘living room rug’ dataset
Figure 19: Comparison of results for the ‘Living room rug’ dataset using an exhaustive search and
100px patch size. Small patches (Graph 1), full-width patches (Graph 2) and geometrically scaled
full-width patches (Graph 3).
38
results: approach comparison - ‘brick-paved road’ dataset
Figure 20: Comparison of results for the ‘Brick-paved road’ dataset using an exhaustive search
and 100px patch size. Small patches (Graph 1), full-width patches (Graph 2) and geometrically
scaled full-width patches (Graph 3).
39
results: exhaustive vs. non-exhaustive search
Non-Exhaustive Exhaustive
Figure 21: Comparison between non-exhaustive and exhaustive search performance using
geometrically scaled template patches and the ‘Living room rug’ dataset.
40
results: patch size comparison - ‘slate footpath’ dataset
Figure 22: Comparison of results for the ‘Living room rug’ dataset using geometrically-scaled
patches of 50px, 100px and 200px heights under an exhaustive search approach.
41
Conclusion
42
future work
Great potential for future work, most likely focussing on areas
including:
• Further experimentation to establish most appropriate
combination of similarity-measure, search approach and patch
size for a given terrain type.
• Development of a system to identify “significant” changes in
vertical displacement model that are potentially indicative of a
potential obstacle or change in terrain gradient.
• Development of system to estimate change in robot heading.
• Development of system to estimate current robot speed.
43
conclusion summary
Project has laid the groundwork upon which further development
can subsequently be conducted.
While not all of the original aims were accomplished, research
efforts had to be focussed on confirming the underlying hypothesis,
which as a general concept, has been proven.
Not one single approach or solution will suffice for all terrain
conditions.
44
Questions?
Project Blog
Slide Design: Matthias Vogelgesang - (Github)
45
bibliography
[1] Jason Campbell, Rahul Sukthankar, Illah Nourbakhsh, and Aroon Pahwa. A robust
visual odometry and precipice detection system using consumer-grade monocular
vision. In Robotics and Automation, 2005. ICRA 2005. Proceedings of the 2005 IEEE
International Conference on, pages 3421–3427. IEEE, 2005.
[2] Mark Maimone, Yang Cheng, and Larry Matthies. Two years of visual odometry on
the mars exploration rovers. Journal of Field Robotics, 24(3):169–186, 2007.
[3] Nazil Perveen, Darshan Kumar, and Ishan Bhardwaj. An Overview on Template
Matching Methodologies and its Applications. IJRCCT, 2(10):988–995, 2013.
[4] Jonathan Pillow. Motion Perception 2.
http://homepage.psy.utexas.edu/homepage/faculty/pillow/
courses/perception09/slides/Lec13A_Motion_part2.pdf, October
2009.
[5] Akshay Roongta. Depth Cue Theory.
http://akshayroongta.in/notes/depth-cue-theory/, October 2013.
46

Weitere ähnliche Inhalte

Was ist angesagt?

Image reconstruction in nuclear medicine
Image reconstruction in nuclear medicineImage reconstruction in nuclear medicine
Image reconstruction in nuclear medicineshokoofeh mousavi
 
Detection of Bridges using Different Types of High Resolution Satellite Images
Detection of Bridges using Different Types of High Resolution Satellite ImagesDetection of Bridges using Different Types of High Resolution Satellite Images
Detection of Bridges using Different Types of High Resolution Satellite Imagesidescitation
 
AUTOMATIC IDENTIFICATION OF CLOUD COVER REGIONS USING SURF
AUTOMATIC IDENTIFICATION OF CLOUD COVER REGIONS USING SURF AUTOMATIC IDENTIFICATION OF CLOUD COVER REGIONS USING SURF
AUTOMATIC IDENTIFICATION OF CLOUD COVER REGIONS USING SURF ijcseit
 
Radar reflectance model for the extraction of height from shape from shading ...
Radar reflectance model for the extraction of height from shape from shading ...Radar reflectance model for the extraction of height from shape from shading ...
Radar reflectance model for the extraction of height from shape from shading ...eSAT Journals
 
An automatic algorithm for object recognition and detection based on asift ke...
An automatic algorithm for object recognition and detection based on asift ke...An automatic algorithm for object recognition and detection based on asift ke...
An automatic algorithm for object recognition and detection based on asift ke...Kunal Kishor Nirala
 
FR3T10-3-IGARSS2011_Geolocation_20110720.pptx
FR3T10-3-IGARSS2011_Geolocation_20110720.pptxFR3T10-3-IGARSS2011_Geolocation_20110720.pptx
FR3T10-3-IGARSS2011_Geolocation_20110720.pptxgrssieee
 
Urban 3D Semantic Modelling Using Stereo Vision, ICRA 2013
Urban 3D Semantic Modelling Using Stereo Vision, ICRA 2013Urban 3D Semantic Modelling Using Stereo Vision, ICRA 2013
Urban 3D Semantic Modelling Using Stereo Vision, ICRA 2013Sunando Sengupta
 
Satellite Imaging System
Satellite Imaging SystemSatellite Imaging System
Satellite Imaging SystemCSCJournals
 
Review on Various Algorithm for Cloud Detection and Removal for Images
Review on Various Algorithm for Cloud Detection and Removal for ImagesReview on Various Algorithm for Cloud Detection and Removal for Images
Review on Various Algorithm for Cloud Detection and Removal for ImagesIJERA Editor
 
A new approach of edge detection in sar images using region based active cont...
A new approach of edge detection in sar images using region based active cont...A new approach of edge detection in sar images using region based active cont...
A new approach of edge detection in sar images using region based active cont...eSAT Journals
 
Passive stereo vision with deep learning
Passive stereo vision with deep learningPassive stereo vision with deep learning
Passive stereo vision with deep learningYu Huang
 
Matching algorithm performance analysis for autocalibration method of stereo ...
Matching algorithm performance analysis for autocalibration method of stereo ...Matching algorithm performance analysis for autocalibration method of stereo ...
Matching algorithm performance analysis for autocalibration method of stereo ...TELKOMNIKA JOURNAL
 
A new approach of edge detection in sar images using
A new approach of edge detection in sar images usingA new approach of edge detection in sar images using
A new approach of edge detection in sar images usingeSAT Publishing House
 
Uncalibrated View Synthesis Using Planar Segmentation of Images
Uncalibrated View Synthesis Using Planar Segmentation of Images  Uncalibrated View Synthesis Using Planar Segmentation of Images
Uncalibrated View Synthesis Using Planar Segmentation of Images ijcga
 
An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...
An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...
An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...CSCJournals
 
The flow of baseline estimation using a single omnidirectional camera
The flow of baseline estimation using a single omnidirectional cameraThe flow of baseline estimation using a single omnidirectional camera
The flow of baseline estimation using a single omnidirectional cameraTELKOMNIKA JOURNAL
 

Was ist angesagt? (20)

Image reconstruction in nuclear medicine
Image reconstruction in nuclear medicineImage reconstruction in nuclear medicine
Image reconstruction in nuclear medicine
 
Detection of Bridges using Different Types of High Resolution Satellite Images
Detection of Bridges using Different Types of High Resolution Satellite ImagesDetection of Bridges using Different Types of High Resolution Satellite Images
Detection of Bridges using Different Types of High Resolution Satellite Images
 
AUTOMATIC IDENTIFICATION OF CLOUD COVER REGIONS USING SURF
AUTOMATIC IDENTIFICATION OF CLOUD COVER REGIONS USING SURF AUTOMATIC IDENTIFICATION OF CLOUD COVER REGIONS USING SURF
AUTOMATIC IDENTIFICATION OF CLOUD COVER REGIONS USING SURF
 
06466595
0646659506466595
06466595
 
Radar reflectance model for the extraction of height from shape from shading ...
Radar reflectance model for the extraction of height from shape from shading ...Radar reflectance model for the extraction of height from shape from shading ...
Radar reflectance model for the extraction of height from shape from shading ...
 
An automatic algorithm for object recognition and detection based on asift ke...
An automatic algorithm for object recognition and detection based on asift ke...An automatic algorithm for object recognition and detection based on asift ke...
An automatic algorithm for object recognition and detection based on asift ke...
 
FR3T10-3-IGARSS2011_Geolocation_20110720.pptx
FR3T10-3-IGARSS2011_Geolocation_20110720.pptxFR3T10-3-IGARSS2011_Geolocation_20110720.pptx
FR3T10-3-IGARSS2011_Geolocation_20110720.pptx
 
Urban 3D Semantic Modelling Using Stereo Vision, ICRA 2013
Urban 3D Semantic Modelling Using Stereo Vision, ICRA 2013Urban 3D Semantic Modelling Using Stereo Vision, ICRA 2013
Urban 3D Semantic Modelling Using Stereo Vision, ICRA 2013
 
Satellite Imaging System
Satellite Imaging SystemSatellite Imaging System
Satellite Imaging System
 
Review on Various Algorithm for Cloud Detection and Removal for Images
Review on Various Algorithm for Cloud Detection and Removal for ImagesReview on Various Algorithm for Cloud Detection and Removal for Images
Review on Various Algorithm for Cloud Detection and Removal for Images
 
A new approach of edge detection in sar images using region based active cont...
A new approach of edge detection in sar images using region based active cont...A new approach of edge detection in sar images using region based active cont...
A new approach of edge detection in sar images using region based active cont...
 
Passive stereo vision with deep learning
Passive stereo vision with deep learningPassive stereo vision with deep learning
Passive stereo vision with deep learning
 
Matching algorithm performance analysis for autocalibration method of stereo ...
Matching algorithm performance analysis for autocalibration method of stereo ...Matching algorithm performance analysis for autocalibration method of stereo ...
Matching algorithm performance analysis for autocalibration method of stereo ...
 
Lm342080283
Lm342080283Lm342080283
Lm342080283
 
Graphics
GraphicsGraphics
Graphics
 
A new approach of edge detection in sar images using
A new approach of edge detection in sar images usingA new approach of edge detection in sar images using
A new approach of edge detection in sar images using
 
Principles of photogrammetry
Principles of photogrammetryPrinciples of photogrammetry
Principles of photogrammetry
 
Uncalibrated View Synthesis Using Planar Segmentation of Images
Uncalibrated View Synthesis Using Planar Segmentation of Images  Uncalibrated View Synthesis Using Planar Segmentation of Images
Uncalibrated View Synthesis Using Planar Segmentation of Images
 
An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...
An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...
An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...
 
The flow of baseline estimation using a single omnidirectional camera
The flow of baseline estimation using a single omnidirectional cameraThe flow of baseline estimation using a single omnidirectional camera
The flow of baseline estimation using a single omnidirectional camera
 

Andere mochten auch

Application of Fourier Transform in Agriculture (Robotic Cultivators Technology)
Application of Fourier Transform in Agriculture (Robotic Cultivators Technology)Application of Fourier Transform in Agriculture (Robotic Cultivators Technology)
Application of Fourier Transform in Agriculture (Robotic Cultivators Technology)Eisha Razia
 
Signal Processing Course : Fourier
Signal Processing Course : FourierSignal Processing Course : Fourier
Signal Processing Course : FourierGabriel Peyré
 
Signal Processing Introduction using Fourier Transforms
Signal Processing Introduction using Fourier TransformsSignal Processing Introduction using Fourier Transforms
Signal Processing Introduction using Fourier TransformsArvind Devaraj
 
Eece 301 note set 14 fourier transform
Eece 301 note set 14 fourier transformEece 301 note set 14 fourier transform
Eece 301 note set 14 fourier transformSandilya Sridhara
 
Optics Fourier Transform Ii
Optics Fourier Transform IiOptics Fourier Transform Ii
Optics Fourier Transform Iidiarmseven
 
Basics of edge detection and forier transform
Basics of edge detection and forier transformBasics of edge detection and forier transform
Basics of edge detection and forier transformSimranjit Singh
 
Introduction to Fourier transform and signal analysis
Introduction to Fourier transform and signal analysisIntroduction to Fourier transform and signal analysis
Introduction to Fourier transform and signal analysis宗翰 謝
 
Introduction to Wavelet Transform with Applications to DSP
Introduction to Wavelet Transform with Applications to DSPIntroduction to Wavelet Transform with Applications to DSP
Introduction to Wavelet Transform with Applications to DSPHicham Berkouk
 

Andere mochten auch (11)

Fourier Transforms
Fourier TransformsFourier Transforms
Fourier Transforms
 
Application of Fourier Transform in Agriculture (Robotic Cultivators Technology)
Application of Fourier Transform in Agriculture (Robotic Cultivators Technology)Application of Fourier Transform in Agriculture (Robotic Cultivators Technology)
Application of Fourier Transform in Agriculture (Robotic Cultivators Technology)
 
Signal Processing Course : Fourier
Signal Processing Course : FourierSignal Processing Course : Fourier
Signal Processing Course : Fourier
 
Signal Processing Introduction using Fourier Transforms
Signal Processing Introduction using Fourier TransformsSignal Processing Introduction using Fourier Transforms
Signal Processing Introduction using Fourier Transforms
 
Lecture 9
Lecture 9Lecture 9
Lecture 9
 
Eece 301 note set 14 fourier transform
Eece 301 note set 14 fourier transformEece 301 note set 14 fourier transform
Eece 301 note set 14 fourier transform
 
Optics Fourier Transform Ii
Optics Fourier Transform IiOptics Fourier Transform Ii
Optics Fourier Transform Ii
 
SDEE: Lectures 3 and 4
SDEE: Lectures 3 and 4SDEE: Lectures 3 and 4
SDEE: Lectures 3 and 4
 
Basics of edge detection and forier transform
Basics of edge detection and forier transformBasics of edge detection and forier transform
Basics of edge detection and forier transform
 
Introduction to Fourier transform and signal analysis
Introduction to Fourier transform and signal analysisIntroduction to Fourier transform and signal analysis
Introduction to Fourier transform and signal analysis
 
Introduction to Wavelet Transform with Applications to DSP
Introduction to Wavelet Transform with Applications to DSPIntroduction to Wavelet Transform with Applications to DSP
Introduction to Wavelet Transform with Applications to DSP
 

Ähnlich wie Estimation of Terrain Gradient Conditions & Obstacle Detection Using a Monocular Vision-based System

6 - Conception of an Autonomous UAV using Stereo Vision (presented in an Indo...
6 - Conception of an Autonomous UAV using Stereo Vision (presented in an Indo...6 - Conception of an Autonomous UAV using Stereo Vision (presented in an Indo...
6 - Conception of an Autonomous UAV using Stereo Vision (presented in an Indo...Youness Lahdili
 
Real-time Moving Object Detection using SURF
Real-time Moving Object Detection using SURFReal-time Moving Object Detection using SURF
Real-time Moving Object Detection using SURFiosrjce
 
International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)IJERD Editor
 
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...IJERD Editor
 
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...IJERD Editor
 
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...IJERD Editor
 
Research Paper v2.0
Research Paper v2.0Research Paper v2.0
Research Paper v2.0Kapil Tiwari
 
Algorithmic Analysis to Video Object Tracking and Background Segmentation and...
Algorithmic Analysis to Video Object Tracking and Background Segmentation and...Algorithmic Analysis to Video Object Tracking and Background Segmentation and...
Algorithmic Analysis to Video Object Tracking and Background Segmentation and...Editor IJCATR
 
Efficient 3D stereo vision stabilization for multi-camera viewpoints
Efficient 3D stereo vision stabilization for multi-camera viewpointsEfficient 3D stereo vision stabilization for multi-camera viewpoints
Efficient 3D stereo vision stabilization for multi-camera viewpointsjournalBEEI
 
An Assessment of Image Matching Algorithms in Depth Estimation
An Assessment of Image Matching Algorithms in Depth EstimationAn Assessment of Image Matching Algorithms in Depth Estimation
An Assessment of Image Matching Algorithms in Depth EstimationCSCJournals
 
Gait Based Person Recognition Using Partial Least Squares Selection Scheme
Gait Based Person Recognition Using Partial Least Squares Selection Scheme Gait Based Person Recognition Using Partial Least Squares Selection Scheme
Gait Based Person Recognition Using Partial Least Squares Selection Scheme ijcisjournal
 
Three-dimensional structure from motion recovery of a moving object with nois...
Three-dimensional structure from motion recovery of a moving object with nois...Three-dimensional structure from motion recovery of a moving object with nois...
Three-dimensional structure from motion recovery of a moving object with nois...IJECEIAES
 
Removal of Transformation Errors by Quarterion In Multi View Image Registration
Removal of Transformation Errors by Quarterion In Multi View Image RegistrationRemoval of Transformation Errors by Quarterion In Multi View Image Registration
Removal of Transformation Errors by Quarterion In Multi View Image RegistrationIDES Editor
 
Visual tracking using particle swarm optimization
Visual tracking using particle swarm optimizationVisual tracking using particle swarm optimization
Visual tracking using particle swarm optimizationcsandit
 
VISUAL TRACKING USING PARTICLE SWARM OPTIMIZATION
VISUAL TRACKING USING PARTICLE SWARM OPTIMIZATIONVISUAL TRACKING USING PARTICLE SWARM OPTIMIZATION
VISUAL TRACKING USING PARTICLE SWARM OPTIMIZATIONcsandit
 
An efficient image segmentation approach through enhanced watershed algorithm
An efficient image segmentation approach through enhanced watershed algorithmAn efficient image segmentation approach through enhanced watershed algorithm
An efficient image segmentation approach through enhanced watershed algorithmAlexander Decker
 
Object tracking with SURF: ARM-Based platform Implementation
Object tracking with SURF: ARM-Based platform ImplementationObject tracking with SURF: ARM-Based platform Implementation
Object tracking with SURF: ARM-Based platform ImplementationEditor IJCATR
 

Ähnlich wie Estimation of Terrain Gradient Conditions & Obstacle Detection Using a Monocular Vision-based System (20)

6 - Conception of an Autonomous UAV using Stereo Vision (presented in an Indo...
6 - Conception of an Autonomous UAV using Stereo Vision (presented in an Indo...6 - Conception of an Autonomous UAV using Stereo Vision (presented in an Indo...
6 - Conception of an Autonomous UAV using Stereo Vision (presented in an Indo...
 
D018112429
D018112429D018112429
D018112429
 
J017377578
J017377578J017377578
J017377578
 
Real-time Moving Object Detection using SURF
Real-time Moving Object Detection using SURFReal-time Moving Object Detection using SURF
Real-time Moving Object Detection using SURF
 
International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)
 
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
 
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
 
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
 
I0343065072
I0343065072I0343065072
I0343065072
 
Research Paper v2.0
Research Paper v2.0Research Paper v2.0
Research Paper v2.0
 
Algorithmic Analysis to Video Object Tracking and Background Segmentation and...
Algorithmic Analysis to Video Object Tracking and Background Segmentation and...Algorithmic Analysis to Video Object Tracking and Background Segmentation and...
Algorithmic Analysis to Video Object Tracking and Background Segmentation and...
 
Efficient 3D stereo vision stabilization for multi-camera viewpoints
Efficient 3D stereo vision stabilization for multi-camera viewpointsEfficient 3D stereo vision stabilization for multi-camera viewpoints
Efficient 3D stereo vision stabilization for multi-camera viewpoints
 
An Assessment of Image Matching Algorithms in Depth Estimation
An Assessment of Image Matching Algorithms in Depth EstimationAn Assessment of Image Matching Algorithms in Depth Estimation
An Assessment of Image Matching Algorithms in Depth Estimation
 
Gait Based Person Recognition Using Partial Least Squares Selection Scheme
Gait Based Person Recognition Using Partial Least Squares Selection Scheme Gait Based Person Recognition Using Partial Least Squares Selection Scheme
Gait Based Person Recognition Using Partial Least Squares Selection Scheme
 
Three-dimensional structure from motion recovery of a moving object with nois...
Three-dimensional structure from motion recovery of a moving object with nois...Three-dimensional structure from motion recovery of a moving object with nois...
Three-dimensional structure from motion recovery of a moving object with nois...
 
Removal of Transformation Errors by Quarterion In Multi View Image Registration
Removal of Transformation Errors by Quarterion In Multi View Image RegistrationRemoval of Transformation Errors by Quarterion In Multi View Image Registration
Removal of Transformation Errors by Quarterion In Multi View Image Registration
 
Visual tracking using particle swarm optimization
Visual tracking using particle swarm optimizationVisual tracking using particle swarm optimization
Visual tracking using particle swarm optimization
 
VISUAL TRACKING USING PARTICLE SWARM OPTIMIZATION
VISUAL TRACKING USING PARTICLE SWARM OPTIMIZATIONVISUAL TRACKING USING PARTICLE SWARM OPTIMIZATION
VISUAL TRACKING USING PARTICLE SWARM OPTIMIZATION
 
An efficient image segmentation approach through enhanced watershed algorithm
An efficient image segmentation approach through enhanced watershed algorithmAn efficient image segmentation approach through enhanced watershed algorithm
An efficient image segmentation approach through enhanced watershed algorithm
 
Object tracking with SURF: ARM-Based platform Implementation
Object tracking with SURF: ARM-Based platform ImplementationObject tracking with SURF: ARM-Based platform Implementation
Object tracking with SURF: ARM-Based platform Implementation
 

Kürzlich hochgeladen

CRM Contender Series: HubSpot vs. Salesforce
CRM Contender Series: HubSpot vs. SalesforceCRM Contender Series: HubSpot vs. Salesforce
CRM Contender Series: HubSpot vs. SalesforceBrainSell Technologies
 
Innovate and Collaborate- Harnessing the Power of Open Source Software.pdf
Innovate and Collaborate- Harnessing the Power of Open Source Software.pdfInnovate and Collaborate- Harnessing the Power of Open Source Software.pdf
Innovate and Collaborate- Harnessing the Power of Open Source Software.pdfYashikaSharma391629
 
20240415 [Container Plumbing Days] Usernetes Gen2 - Kubernetes in Rootless Do...
20240415 [Container Plumbing Days] Usernetes Gen2 - Kubernetes in Rootless Do...20240415 [Container Plumbing Days] Usernetes Gen2 - Kubernetes in Rootless Do...
20240415 [Container Plumbing Days] Usernetes Gen2 - Kubernetes in Rootless Do...Akihiro Suda
 
Odoo 14 - eLearning Module In Odoo 14 Enterprise
Odoo 14 - eLearning Module In Odoo 14 EnterpriseOdoo 14 - eLearning Module In Odoo 14 Enterprise
Odoo 14 - eLearning Module In Odoo 14 Enterprisepreethippts
 
Catch the Wave: SAP Event-Driven and Data Streaming for the Intelligence Ente...
Catch the Wave: SAP Event-Driven and Data Streaming for the Intelligence Ente...Catch the Wave: SAP Event-Driven and Data Streaming for the Intelligence Ente...
Catch the Wave: SAP Event-Driven and Data Streaming for the Intelligence Ente...confluent
 
Large Language Models for Test Case Evolution and Repair
Large Language Models for Test Case Evolution and RepairLarge Language Models for Test Case Evolution and Repair
Large Language Models for Test Case Evolution and RepairLionel Briand
 
SensoDat: Simulation-based Sensor Dataset of Self-driving Cars
SensoDat: Simulation-based Sensor Dataset of Self-driving CarsSensoDat: Simulation-based Sensor Dataset of Self-driving Cars
SensoDat: Simulation-based Sensor Dataset of Self-driving CarsChristian Birchler
 
Dealing with Cultural Dispersion — Stefano Lambiase — ICSE-SEIS 2024
Dealing with Cultural Dispersion — Stefano Lambiase — ICSE-SEIS 2024Dealing with Cultural Dispersion — Stefano Lambiase — ICSE-SEIS 2024
Dealing with Cultural Dispersion — Stefano Lambiase — ICSE-SEIS 2024StefanoLambiase
 
How to submit a standout Adobe Champion Application
How to submit a standout Adobe Champion ApplicationHow to submit a standout Adobe Champion Application
How to submit a standout Adobe Champion ApplicationBradBedford3
 
Powering Real-Time Decisions with Continuous Data Streams
Powering Real-Time Decisions with Continuous Data StreamsPowering Real-Time Decisions with Continuous Data Streams
Powering Real-Time Decisions with Continuous Data StreamsSafe Software
 
Understanding Flamingo - DeepMind's VLM Architecture
Understanding Flamingo - DeepMind's VLM ArchitectureUnderstanding Flamingo - DeepMind's VLM Architecture
Understanding Flamingo - DeepMind's VLM Architecturerahul_net
 
VK Business Profile - provides IT solutions and Web Development
VK Business Profile - provides IT solutions and Web DevelopmentVK Business Profile - provides IT solutions and Web Development
VK Business Profile - provides IT solutions and Web Developmentvyaparkranti
 
UI5ers live - Custom Controls wrapping 3rd-party libs.pptx
UI5ers live - Custom Controls wrapping 3rd-party libs.pptxUI5ers live - Custom Controls wrapping 3rd-party libs.pptx
UI5ers live - Custom Controls wrapping 3rd-party libs.pptxAndreas Kunz
 
Cyber security and its impact on E commerce
Cyber security and its impact on E commerceCyber security and its impact on E commerce
Cyber security and its impact on E commercemanigoyal112
 
Simplifying Microservices & Apps - The art of effortless development - Meetup...
Simplifying Microservices & Apps - The art of effortless development - Meetup...Simplifying Microservices & Apps - The art of effortless development - Meetup...
Simplifying Microservices & Apps - The art of effortless development - Meetup...Rob Geurden
 
Taming Distributed Systems: Key Insights from Wix's Large-Scale Experience - ...
Taming Distributed Systems: Key Insights from Wix's Large-Scale Experience - ...Taming Distributed Systems: Key Insights from Wix's Large-Scale Experience - ...
Taming Distributed Systems: Key Insights from Wix's Large-Scale Experience - ...Natan Silnitsky
 
Balasore Best It Company|| Top 10 IT Company || Balasore Software company Odisha
Balasore Best It Company|| Top 10 IT Company || Balasore Software company OdishaBalasore Best It Company|| Top 10 IT Company || Balasore Software company Odisha
Balasore Best It Company|| Top 10 IT Company || Balasore Software company Odishasmiwainfosol
 
Global Identity Enrolment and Verification Pro Solution - Cizo Technology Ser...
Global Identity Enrolment and Verification Pro Solution - Cizo Technology Ser...Global Identity Enrolment and Verification Pro Solution - Cizo Technology Ser...
Global Identity Enrolment and Verification Pro Solution - Cizo Technology Ser...Cizo Technology Services
 

Kürzlich hochgeladen (20)

CRM Contender Series: HubSpot vs. Salesforce
CRM Contender Series: HubSpot vs. SalesforceCRM Contender Series: HubSpot vs. Salesforce
CRM Contender Series: HubSpot vs. Salesforce
 
Innovate and Collaborate- Harnessing the Power of Open Source Software.pdf
Innovate and Collaborate- Harnessing the Power of Open Source Software.pdfInnovate and Collaborate- Harnessing the Power of Open Source Software.pdf
Innovate and Collaborate- Harnessing the Power of Open Source Software.pdf
 
20240415 [Container Plumbing Days] Usernetes Gen2 - Kubernetes in Rootless Do...
20240415 [Container Plumbing Days] Usernetes Gen2 - Kubernetes in Rootless Do...20240415 [Container Plumbing Days] Usernetes Gen2 - Kubernetes in Rootless Do...
20240415 [Container Plumbing Days] Usernetes Gen2 - Kubernetes in Rootless Do...
 
Odoo 14 - eLearning Module In Odoo 14 Enterprise
Odoo 14 - eLearning Module In Odoo 14 EnterpriseOdoo 14 - eLearning Module In Odoo 14 Enterprise
Odoo 14 - eLearning Module In Odoo 14 Enterprise
 
Catch the Wave: SAP Event-Driven and Data Streaming for the Intelligence Ente...
Catch the Wave: SAP Event-Driven and Data Streaming for the Intelligence Ente...Catch the Wave: SAP Event-Driven and Data Streaming for the Intelligence Ente...
Catch the Wave: SAP Event-Driven and Data Streaming for the Intelligence Ente...
 
Large Language Models for Test Case Evolution and Repair
Large Language Models for Test Case Evolution and RepairLarge Language Models for Test Case Evolution and Repair
Large Language Models for Test Case Evolution and Repair
 
SensoDat: Simulation-based Sensor Dataset of Self-driving Cars
SensoDat: Simulation-based Sensor Dataset of Self-driving CarsSensoDat: Simulation-based Sensor Dataset of Self-driving Cars
SensoDat: Simulation-based Sensor Dataset of Self-driving Cars
 
Dealing with Cultural Dispersion — Stefano Lambiase — ICSE-SEIS 2024
Dealing with Cultural Dispersion — Stefano Lambiase — ICSE-SEIS 2024Dealing with Cultural Dispersion — Stefano Lambiase — ICSE-SEIS 2024
Dealing with Cultural Dispersion — Stefano Lambiase — ICSE-SEIS 2024
 
How to submit a standout Adobe Champion Application
How to submit a standout Adobe Champion ApplicationHow to submit a standout Adobe Champion Application
How to submit a standout Adobe Champion Application
 
Powering Real-Time Decisions with Continuous Data Streams
Powering Real-Time Decisions with Continuous Data StreamsPowering Real-Time Decisions with Continuous Data Streams
Powering Real-Time Decisions with Continuous Data Streams
 
Understanding Flamingo - DeepMind's VLM Architecture
Understanding Flamingo - DeepMind's VLM ArchitectureUnderstanding Flamingo - DeepMind's VLM Architecture
Understanding Flamingo - DeepMind's VLM Architecture
 
Advantages of Odoo ERP 17 for Your Business
Advantages of Odoo ERP 17 for Your BusinessAdvantages of Odoo ERP 17 for Your Business
Advantages of Odoo ERP 17 for Your Business
 
VK Business Profile - provides IT solutions and Web Development
VK Business Profile - provides IT solutions and Web DevelopmentVK Business Profile - provides IT solutions and Web Development
VK Business Profile - provides IT solutions and Web Development
 
UI5ers live - Custom Controls wrapping 3rd-party libs.pptx
UI5ers live - Custom Controls wrapping 3rd-party libs.pptxUI5ers live - Custom Controls wrapping 3rd-party libs.pptx
UI5ers live - Custom Controls wrapping 3rd-party libs.pptx
 
Cyber security and its impact on E commerce
Cyber security and its impact on E commerceCyber security and its impact on E commerce
Cyber security and its impact on E commerce
 
Simplifying Microservices & Apps - The art of effortless development - Meetup...
Simplifying Microservices & Apps - The art of effortless development - Meetup...Simplifying Microservices & Apps - The art of effortless development - Meetup...
Simplifying Microservices & Apps - The art of effortless development - Meetup...
 
2.pdf Ejercicios de programación competitiva
2.pdf Ejercicios de programación competitiva2.pdf Ejercicios de programación competitiva
2.pdf Ejercicios de programación competitiva
 
Taming Distributed Systems: Key Insights from Wix's Large-Scale Experience - ...
Taming Distributed Systems: Key Insights from Wix's Large-Scale Experience - ...Taming Distributed Systems: Key Insights from Wix's Large-Scale Experience - ...
Taming Distributed Systems: Key Insights from Wix's Large-Scale Experience - ...
 
Balasore Best It Company|| Top 10 IT Company || Balasore Software company Odisha
Balasore Best It Company|| Top 10 IT Company || Balasore Software company OdishaBalasore Best It Company|| Top 10 IT Company || Balasore Software company Odisha
Balasore Best It Company|| Top 10 IT Company || Balasore Software company Odisha
 
Global Identity Enrolment and Verification Pro Solution - Cizo Technology Ser...
Global Identity Enrolment and Verification Pro Solution - Cizo Technology Ser...Global Identity Enrolment and Verification Pro Solution - Cizo Technology Ser...
Global Identity Enrolment and Verification Pro Solution - Cizo Technology Ser...
 

Estimation of Terrain Gradient Conditions & Obstacle Detection Using a Monocular Vision-based System

  • 1. estimation of terrain gradient conditions & obstacle detection using a monocular vision-based system Final Demonstration Connor Luke Goddard (clg11) Department of Computer Science, Aberystwyth University
  • 3. background Figure 1: Tracked features from MER [2]. The ability for an autonomous robot to navigate from one location to another in a manner that is both safe, yet objective is vital to the survivability of the machine, and the success of the mission it is undertaking. Vision based obstacle detection has enjoyed a high level of research focus over recent years. Less work has concentrated on providing a means of reliably detecting changes in terrain slope through the exploitation of observed changes in motion. 2
  • 4. related work [1] A Robust Visual Odometry and Precipice Detection System Using Consumer-grade Monocular Vision, J. Campbell et al., IEEE, 2005 [2] Obstacle Detection using Optical Flow, T. Low and G. Wyeth, School of Information Technology and Electrical Engineering, University of Queensland, 2011 [3] Appearance-based Obstacle Detection with Monocular Color Vision, I. Ulrich and I. Nourbakhsh, AAAI, 2000 Figure 2: Left: Precipice detection [1], Right: Obstacle colour detection [3] 3
  • 5. project overview Focus Investigation into a system capable of utilising a single, forward-facing colour camera to provide an estimation into current terrain gradient conditions, obstacle location & characteristics and robot ego-motion. Potential Metrics For Observation • Changes in terrain gradient (i.e. slopes). • Location & characteristics of positive and negative obstacles (i.e. rock or pit respectively). • Current robot speed of travel. • Changes in robot orientation. 4
  • 7. working hypothesis “From images captured using a non-calibrated monocular camera system, analysis of optical flow vectors extracted using localised appearance-based dense optical flow techniques can provide certain estimates into the current condition of terrain gradient, the location and characteristics of obstacles, and robot ego-motion.” 6
  • 8. motion parallax Approach focussed on the effects of motion parallax. i.e. Objects/features that are at a greater distance from the camera appear to move less from frame-to-frame than those that are closer. Figure 3: Typical example of motion parallax [4]. 7
  • 9. inference of terrain gradient & obstacle detection Key Aspects 1. The exclusive use of appearance-based template matching techniques to provide a localised variant of dense optical flow analysis over the use of sparse optical flow techniques relying upon a high number of available image features. 2. The use of a formalised model to represent detected changes in vertical displacement as part of efforts to estimate changes in terrain gradient in addition to the location characteristics of potential obstacles. 8
  • 10. vertical displacement model Provides a mechanism by which to evaluate the prediction: 0Vertical Displacement(px) 0 Image Row (Height) (px) Expected general trend between image row and vertical displacement demonstrated. Figure 4: Example of “perfect” vertical displacement model. “Features captured towards the bottom of an image will show greater displacement between subsequent frames, than those features captured towards the top of the image.” 9
  • 11. inference of terrain slope & potential obstacles From observing discrepancies between the current model and “baseline” model, it should be possible to infer the presence of potential obstacles: 0 Vertical Displacement(px) 0 Image Row (Height) (px) 160 400 Recorded vertical displacement for row #160 matching model vertical displacement for row #230 230 Model vertical displacement Last-recorded vertical displacement Figure 5: Vertical displacement model indicating the potential presence of a positive obstacle. 10
  • 12. inference of terrain slope & potential obstacles Differences in model discrepancy behaviour should provide enough information to establish between a potential obstacle and a change in terrain slope: Figure 6: Example of observed differences between detection of obstacles and change in terrain slope. 11
  • 13. robot rotation Two types of observed motion: • Translation (Vertical component of optical flow vectors) • Rotation (Horizontal component of optical flow vectors) Horizontal component of optical flow vector (Rotation) Vertical component of optical flow vector (Translation) Figure 7: Diagram demonstrating the expected difference in behaviour between the separated vertical and horizontal components of observed optical flow vectors calculated from forward-turning motion. 12
  • 14. robot speed By recording multiple “calibration” models at different speeds, it should be possible to later infer the current speed using the standard speed-distance-time equation. Optimum point to infer speed as lowest risk of interference from moving objects during recording and comparing current displacement with model. 0 Vertical Displacement(px) 0 Image Row (Height) (px) Vertical displacement model recorded at constant 5mph Vertical displacement model recorded at constant 10mph Vertical displacement model recorded at constant 15mph Figure 8: Diagram indicating examples of alternative vertical displacement models calculated while travelling at different constant speeds. 13
  • 15. investigation aims Primary Aims 1. Establish which out the following appearance-based template matching similarity measures: • Euclidean Distance • Correlation Coefficient • Histogram Correlation • Histogram Chi-Square best supports the predicted positive correlation between the vertical position of features within an image, and the level of vertical displacement demonstrated. 2. Use the generated model of vertical displacement to identify potential obstacles and changes in terrain gradient based upon the predicted ‘comparative trend’ behaviour. 14
  • 16. investigation aims Secondary Aims 1. Further extend the capabilities of the vertical displacement model to provide estimates into the current speed of travel demonstrated by the robot. 2. Add support for providing an estimation into the change in orientation. 15
  • 18. gathering datasets No existing imagery datasets were found to be suitable, therefore new sets had to be collected. Figure 9: Camera-rig setup used to capture the experiment datasets from front and back profiles. 17
  • 19. gathering datasets Images taken at a number of locations, but four were eventually selected based upon their variety in terms of terrain texture and lighting conditions. Figure 10: Examples of the first frame captured from each of the four location datasets. 18
  • 20. experiment methods Over time, investigation focus changed from exploring a series of reasonably “generalised” research aims, to instead concentrating almost exclusively on one small, but very important aspect; Ensuring that accurate results for appearance-based motion displacement could continue to be achieved, across terrain lacking any significant features and/or demonstrating highly repetitive patterns. Three Experiments 1. Template Matching - Multiple Small Patches 2. Template Matching - Full-width Patches (Non-Scaled) 3. Template Matching - Full-width Patches (Geometrically Scaled) 19
  • 21. experiment one - multiple small patches Stage One Import two consecutive images, and convert from RGB (BGR in OpenCV) colour space to HSV. The ’V’ channel is then removed in order to improve robustness to lighting changes between frames. 20
  • 22. experiment one - multiple small patches Stage Two Extract a percentage-width region of interest (ROI) from centre of first image. 21
  • 23. region-of-interest Why do we need to extract a ROI? Focus-of-expansion: Objects in images do not actually move in 1-dimension (i.e. straight down the image). This effect is minimised towards the centre of the image. Figure 11: Diagram indicating the perceived effects caused by Focus of Expansion. Courtesy: [5] 22
  • 24. experiment one - multiple small patches Stage Three Extract patches of a fixed size around each pixel within the extracted ROI. Figure 12: Simplified example of patch extraction within ROI. 23
  • 25. experiment one - multiple small patches Stage Four For each patch extracted from image one, move down through a localised search window (column) in image two searching for the best match against the template patch. Figure 13: Example of “best match” search within local column. 24
  • 26. experiment one - multiple small patches Stage Five Identify the patch within the localised search window that provides the “best match” via correlation-based matching (e.g. Euclidean Distance, SSD or Correlation coefficient). 25
  • 27. experiment one - multiple small patches Stage Six Average all measured displacements for each pixel along a given row. Outliers are removed by ignoring any displacements that lie outside of (2 x Standard Deviation) of the mean. 26
  • 28. experiment one - multiple small patches Repeat Stages 1-6 Repeat stages 1-6 for an entire collection of “calibration” images taken of flat, unobstructed terrain. 27
  • 29. experiment one - multiple small patches Stage Seven Plot the average displacement for each ROI row, calculated from the displacements recorded over all calibration images. 28
  • 30. experiment two - full-width patches (non-scaled) Focus Establishing if adopting larger patches in fewer numbers provided better appearance-based matching accuracy than using many more, but critically much smaller overlapping patches. While the underlying method remained the same as in the first experiment, some key changes were made: • Extracting a central region of interest relative to the ground plane, as opposed to extracting a fixed-width column from the image plane. • Moving away from multiple overlapping small patches, in favour of adopting a single, full-width patch to represent a single row in the image. 29
  • 31. exhaustive vs. non-exhaustive search Issue discovered regarding the use of non-exhaustive searching causing erroneous results. All tests in experiments two and three were conducted using both exhaustive and non-exhaustive approaches. 0 EuclideanDistance 0 Point at which non-exhaustive search potentially stops (local minima) Location of true "best match" score for Euclidean Distance (global minima) Figure 14: Example of potential early termination of non-exhaustive search due to entering local minima caused by noise, rather than global minima which represents to true best match. 30
  • 32. perspective distortion calibration tool To extract a region-of-interest along the ground plane, the system would need to take account of perspective distortion. Therefore, a simple calibration tool was implemented, enabling users to define the required region of interest within an image. Figure 15: Results of the verification test for the approach towards geometric scaling of template patch pixel coordinates. 31
  • 33. experiment three - full-width patches (scaled) Focus Add geometric scaling of template patch in order to account for objects appearing larger as they approach the camera. Original Template Patch Re-scaled Template Patch (Geometric Transformation of Pixel Coordinates) Re-scaled Template Patch (Linear Interpolation) Figure 16: Comparison between various approaches of performing scaling of the original template patch image. In the case of linear interpolation, additional information has been added to image as a result of “estimating” the colour that lies between two original pixels. 32
  • 34. experiment three - full-width patches (scaled) 6 Stage Process 1. Obtain the width and height of the current template patch. 2. Obtain the calibrated width of the search window with respect to the current image row. 3. Calculate the independent scale factor for the width and height between the template patch and the search window. 4. Treating each pixel as a 2D coordinate in geometric space, scale the position of each pixel coordinate within the template patch relative to the position of centre coordinate. 5. For each scaled pixel coordinate, extract the value of the pixel at the corresponding position with the search window and add to a temporary “image” data structure. 6. Perform template matching between the original template patch, and the new “extracted” temporary image. 33
  • 35. experiment three - full-width patches (scaled) Testing of Scaling Approach A simple test was devised to confirm geometric scaling worked as expected under “perfect” conditions. Figure 17: Results of the verification test for the approach towards geometric scaling of template patch pixel coordinates. 34
  • 36. additional experimental work Some additional experimental work was conducted as part of the larger project investigation, but these did not actively contribute to the final investigation results. • Investigation of Campbell et al. [1] approach to feature tracking to provide optical flow field (based heavily upon C# implementation provided by Dr. Rainer Hessmer (Source)) • Investigation of edge-based template matching discussed by Perveen et al. [3] Figure 18: Left: Example of feature tracking to establish optical flow field. Right: Edge-based template matching using Canny edge detector. 35
  • 37. Investigation Results View Results (Courtesy: nbviewer) 36
  • 38. results Results can be described as generally inconclusive. However from the results obtained, we have learned: 1. Under certain terrain conditions, it is possible to potentially establish a relationship between row height in an image, and average downwards pixel displacement. 2. The performance of each appearance-based similarity measure can vary widely, based upon factors including: • Terrain texture (e.g. highly patterned, isotropic or homogenous) • Patch size (Balance between reducing “noise” and representing true displacement behaviour) • Exhaustive vs. Non-exhaustive search • Scaling vs. Non-scaling 37
  • 39. results: approach comparison - ‘living room rug’ dataset Figure 19: Comparison of results for the ‘Living room rug’ dataset using an exhaustive search and 100px patch size. Small patches (Graph 1), full-width patches (Graph 2) and geometrically scaled full-width patches (Graph 3). 38
  • 40. results: approach comparison - ‘brick-paved road’ dataset Figure 20: Comparison of results for the ‘Brick-paved road’ dataset using an exhaustive search and 100px patch size. Small patches (Graph 1), full-width patches (Graph 2) and geometrically scaled full-width patches (Graph 3). 39
  • 41. results: exhaustive vs. non-exhaustive search Non-Exhaustive Exhaustive Figure 21: Comparison between non-exhaustive and exhaustive search performance using geometrically scaled template patches and the ‘Living room rug’ dataset. 40
  • 42. results: patch size comparison - ‘slate footpath’ dataset Figure 22: Comparison of results for the ‘Living room rug’ dataset using geometrically-scaled patches of 50px, 100px and 200px heights under an exhaustive search approach. 41
  • 44. future work Great potential for future work, most likely focussing on areas including: • Further experimentation to establish most appropriate combination of similarity-measure, search approach and patch size for a given terrain type. • Development of a system to identify “significant” changes in vertical displacement model that are potentially indicative of a potential obstacle or change in terrain gradient. • Development of system to estimate change in robot heading. • Development of system to estimate current robot speed. 43
  • 45. conclusion summary Project has laid the groundwork upon which further development can subsequently be conducted. While not all of the original aims were accomplished, research efforts had to be focussed on confirming the underlying hypothesis, which as a general concept, has been proven. Not one single approach or solution will suffice for all terrain conditions. 44
  • 46. Questions? Project Blog Slide Design: Matthias Vogelgesang - (Github) 45
  • 47. bibliography [1] Jason Campbell, Rahul Sukthankar, Illah Nourbakhsh, and Aroon Pahwa. A robust visual odometry and precipice detection system using consumer-grade monocular vision. In Robotics and Automation, 2005. ICRA 2005. Proceedings of the 2005 IEEE International Conference on, pages 3421–3427. IEEE, 2005. [2] Mark Maimone, Yang Cheng, and Larry Matthies. Two years of visual odometry on the mars exploration rovers. Journal of Field Robotics, 24(3):169–186, 2007. [3] Nazil Perveen, Darshan Kumar, and Ishan Bhardwaj. An Overview on Template Matching Methodologies and its Applications. IJRCCT, 2(10):988–995, 2013. [4] Jonathan Pillow. Motion Perception 2. http://homepage.psy.utexas.edu/homepage/faculty/pillow/ courses/perception09/slides/Lec13A_Motion_part2.pdf, October 2009. [5] Akshay Roongta. Depth Cue Theory. http://akshayroongta.in/notes/depth-cue-theory/, October 2013. 46