SlideShare ist ein Scribd-Unternehmen logo
1 von 1
Downloaden Sie, um offline zu lesen
 4	
  	
  	
  	
  Method	
  
Camera	
  calibra1on	
  
Each	
  camera	
  has	
  its	
  own	
  unique	
  intrinsic	
  parameters,	
  which	
  is	
  expressed	
  by	
  a	
  3x3	
  
camera	
  calibra9on	
  matrix,	
  K.	
  We	
  obtain	
  this	
  matrix	
  by	
  using	
  a	
  video	
  calibra9on	
  
technique	
  implemented	
  with	
  OpenCV	
  using	
  mul9ple	
  checkerboard	
  images	
  as	
  
shown	
  below.	
  (3	
  of	
  25	
  images	
  are	
  displayed	
  here)	
  
Raw	
  image	
   Image	
  with	
  lens	
  distor9on	
  removed	
  
Mo1va1on	
  
•  Explora9on	
  using	
  robots	
  can	
  be	
  useful	
  for	
  examining	
  interior	
  
structures	
  
•  Panoramas	
   create	
   a	
   wider	
   field	
   of	
   view	
   to	
   increase	
   the	
  	
  
teleoperator’s	
  3D	
  awareness	
  
•  To	
   render	
   a	
   real-­‐9me	
   visualiza9on	
   to	
   the	
   teleoperator,	
   the	
  	
  
implemented	
   algorithm	
   should	
   s9tch	
   video	
   frames	
   together	
  
as	
  efficiently	
  as	
  possible	
  
	
  1	
  	
  	
  	
  Introduc1on	
  
	
  
	
  
	
  
	
  
Assump1ons	
  
•  Lens	
  distor9on	
  is	
  removed	
  from	
  video	
  
•  Intrinsic	
  camera	
  parameters	
  are	
  known	
  
	
  
Problem	
  statement	
  
From	
  a	
  video,	
  detect	
  features	
  in	
  frame	
  In	
  and	
  match	
  these	
  features	
  in	
  consecu9ve	
  
frames	
  in	
  order	
  to	
  detect	
  the	
  key	
  frame	
  In+1	
  from	
  the	
  calculated	
  homography	
  Hn,n+1	
  
Use	
  the	
  saved	
  key	
  frames	
  and	
  homographies	
  to	
  create	
  a	
  s9tched	
  panorama	
  of	
  the	
  
environment.	
  
	
  
2	
  	
  	
  	
  Problem	
  Statement	
  
	
  5	
  	
  Experimental	
  Results	
  
	
  6	
  	
  References	
  
[1]	
  Brown,	
  MaYhew,	
  and	
  David	
  G.	
  Lowe.	
  "Automa9c	
  Panoramic	
  Image	
  S9tching	
  
Using	
  Invariant	
  Features."	
  Int	
  J	
  Comput	
  Vision	
  Interna1onal	
  Journal	
  of	
  Computer	
  
Vision	
  74.1	
  (2006):	
  59-­‐73.	
  
[2]	
  Shi,	
  Jianbo,	
  and	
  Tomasi.	
  "Good	
  Features	
  to	
  Track."	
  Proceedings	
  of	
  IEEE	
  
Conference	
  on	
  Computer	
  Vision	
  and	
  Pa9ern	
  Recogni1on	
  CVPR-­‐94	
  (1994).	
  
	
  7	
  	
  Acknowledgments	
  
This	
  research	
  was	
  supported	
  by	
  the	
  Na9onal	
  Science	
  Founda9on	
  
Research	
  Experiences	
  for	
  Undergraduates	
  (REU)	
  program	
  on	
  
Interdisciplinary	
  Research	
  in	
  Mechatronics,	
  Robo9cs,	
  and	
  
Automated	
  System	
  Design	
  (NSF	
  grant	
  no.	
  1263293).	
  Any	
  
opinions,	
  findings,	
  and	
  conclusions	
  or	
  recommenda9ons	
  
expressed	
  in	
  this	
  material	
  are	
  those	
  of	
  the	
  author(s)	
  and	
  do	
  not	
  
necessarily	
  reflect	
  the	
  views	
  of	
  the	
  Na9onal	
  Science	
  Founda9on.	
  
	
  3	
  	
  	
  	
  System	
  Overview	
  
25 50 75 100
0
0.02
0.04
0.06
0.08
0.1
0.12
0.14
0.16
0.18
Number of Features
Time(s)
Time Comparison Between
Shi−Tomasi/Optical Flow and SIFT/FLANN
Shi−Tomasi Corner/Optical Flow
SIFT/FLANN
Sample	
  images	
  for	
  calibra9on	
  
Feature	
  Detec1on	
  and	
  Matching	
  
There	
  are	
  many	
  different	
  methods	
  used	
  to	
  detect	
  features.	
  Scale	
  Invariant	
  Feature	
  
Detec9on,	
  SIFT,	
  is	
  a	
  popular	
  method	
  that	
  is	
  known	
  to	
  be	
  robust	
  as	
  it	
  is	
  invariant	
  to	
  
changes	
  in	
  image	
  scale	
  [1].	
  Shi-­‐Tomasi’s	
  method	
  is	
  a	
  technique	
  which	
  only	
  tacks	
  
corner	
  features	
  in	
  an	
  image	
  [2],	
  which	
  is	
  ohen	
  faster.	
  SIFT	
  features	
  are	
  matched	
  
between	
  frames	
  using	
  the	
  Fast	
  Library	
  for	
  Approximate	
  Nearest	
  Neighbors,	
  FLANN,	
  	
  
method.	
  The	
  Shi-­‐Tomasi	
  technique	
  tracks	
  the	
  corners	
  with	
  op9cal	
  flow	
  which	
  uses	
  
the	
  gradient	
  in	
  the	
  neighborhood	
  of	
  a	
  feature	
  to	
  track	
  its	
  mo9on	
  among	
  frames.	
  
	
  
	
  
	
  
	
  
	
  
	
  
	
  
	
  
	
  
	
  
	
  
Feature	
  detec9on	
  using	
  SIFT	
  and	
  matching	
  using	
  FLANN	
  
	
  
	
  
	
  
	
  
	
  
	
  
	
  
	
  
	
  
Feature	
  detec9on	
  using	
  Shi-­‐Tomasi	
  Corner	
  and	
  matching	
  using	
  Op9cal	
  Flow	
  
	
  
Key	
  Frame	
  Selec1on	
  
In	
  order	
  to	
  get	
  a	
  well	
  s9tched	
  panoramic	
  image,	
  it	
  is	
  important	
  to	
  choose	
  a	
  frame	
  I’	
  
that	
   will	
   overlap	
   well	
   with	
   its	
   previous	
   frame	
   I.	
   A	
   homography	
   H	
   is	
   a	
   projec9ve	
  
transforma9on	
   that	
   relates	
   image	
   points	
   by	
   x’	
   =	
   Hx,	
   where	
   x	
   and	
   x’	
   are	
   image	
  
points	
   in	
   I	
   and	
   I’,	
   respec9vely.	
   Image	
   s9tching	
   is	
   accomplished	
   by	
   calcula9ng	
   H	
  
between	
   the	
   two	
   images.	
   Op9cal	
   flow	
   tracks	
   features	
   among	
   frames.	
   When	
  
tracking	
  n	
  features,	
  once	
  the	
  successfully	
  tracked	
  features	
  n’	
  in	
  the	
  current	
  frame	
  is	
  
less	
   than	
   n/2,	
   the	
   camera	
   has	
   moved	
   a	
   sufficient	
   amount	
   to	
   provide	
   a	
   good	
  
homography.	
  
	
  
Image	
  S1tching	
  
The	
  key	
  frames	
  are	
  s9tched	
  to	
  a	
  panoramic	
  texture	
  that	
  is	
  mapped	
  onto	
  a	
  cylinder.	
  
From	
  H	
  and	
  K,	
  the	
  camera	
  rota9on	
  R	
  is	
  es9mated.	
  Based	
  on	
  this	
  pose,	
  an	
  image	
  
point	
  x	
  in	
  the	
  world	
  space	
  of	
  the	
  cylinder	
  is	
  determined	
  by	
  X	
  =	
  K-­‐1R-­‐1x	
  where	
  X	
  =	
  (x,	
  
y,	
  z)	
  .	
  These	
  world	
  coordinates	
  are	
  then	
  mapped	
  to	
  UV	
  texture	
  coordinates	
  where	
  u	
  
=	
  arctan(x/z)	
  and	
  v	
  =	
  y/√(x2	
  +	
  z2).	
  The	
  final	
  step	
  is	
  to	
  texture	
  map	
  the	
  cylinder	
  by	
  x	
  
=	
  r*cos(u),	
  y	
  =	
  v,	
  z	
  =	
  r*sin(u)	
  where	
  r	
  is	
  the	
  radius	
  of	
  the	
  cylinder.	
  
Shi-­‐Tomasi	
  and	
  Op9cal	
  Flow	
   SIFT	
  and	
  FLANN	
  
Shi-­‐Tomasi	
  and	
  Op9cal	
  Flow	
   SIFT	
  and	
  FLANN	
  
S9tched	
  images	
  with	
  cylindrical	
  projec9on	
  Input	
  Images	
  

Weitere ähnliche Inhalte

Was ist angesagt?

An automatic algorithm for object recognition and detection based on asift ke...
An automatic algorithm for object recognition and detection based on asift ke...An automatic algorithm for object recognition and detection based on asift ke...
An automatic algorithm for object recognition and detection based on asift ke...
Kunal Kishor Nirala
 
Open GL T0074 56 sm3
Open GL T0074 56 sm3Open GL T0074 56 sm3
Open GL T0074 56 sm3
Roziq Bahtiar
 
Temporal Frequency Probing for 5D Transient Analysis of Global Light Transport
Temporal Frequency Probing for 5D Transient Analysis of Global Light TransportTemporal Frequency Probing for 5D Transient Analysis of Global Light Transport
Temporal Frequency Probing for 5D Transient Analysis of Global Light Transport
Matthew O'Toole
 

Was ist angesagt? (20)

Passive stereo vision with deep learning
Passive stereo vision with deep learningPassive stereo vision with deep learning
Passive stereo vision with deep learning
 
An automatic algorithm for object recognition and detection based on asift ke...
An automatic algorithm for object recognition and detection based on asift ke...An automatic algorithm for object recognition and detection based on asift ke...
An automatic algorithm for object recognition and detection based on asift ke...
 
Neural Scene Representation & Rendering: Introduction to Novel View Synthesis
Neural Scene Representation & Rendering: Introduction to Novel View SynthesisNeural Scene Representation & Rendering: Introduction to Novel View Synthesis
Neural Scene Representation & Rendering: Introduction to Novel View Synthesis
 
Introductory Level of SLAM Seminar
Introductory Level of SLAM SeminarIntroductory Level of SLAM Seminar
Introductory Level of SLAM Seminar
 
An accurate retrieval through R-MAC+ descriptors for landmark recognition
An accurate retrieval through R-MAC+ descriptors for landmark recognitionAn accurate retrieval through R-MAC+ descriptors for landmark recognition
An accurate retrieval through R-MAC+ descriptors for landmark recognition
 
06466595
0646659506466595
06466595
 
zernike moments for image classification
zernike moments for image classificationzernike moments for image classification
zernike moments for image classification
 
Dense Image Matching - Challenges and Potentials (Keynote 3D-ARCH 2015)
Dense Image Matching - Challenges and Potentials (Keynote 3D-ARCH 2015) Dense Image Matching - Challenges and Potentials (Keynote 3D-ARCH 2015)
Dense Image Matching - Challenges and Potentials (Keynote 3D-ARCH 2015)
 
Automatic Dense Semantic Mapping From Visual Street-level Imagery
Automatic Dense Semantic Mapping From Visual Street-level ImageryAutomatic Dense Semantic Mapping From Visual Street-level Imagery
Automatic Dense Semantic Mapping From Visual Street-level Imagery
 
Open GL T0074 56 sm3
Open GL T0074 56 sm3Open GL T0074 56 sm3
Open GL T0074 56 sm3
 
Primal-Dual Coding to Probe Light Transport
Primal-Dual Coding to Probe Light TransportPrimal-Dual Coding to Probe Light Transport
Primal-Dual Coding to Probe Light Transport
 
Lane Detection
Lane DetectionLane Detection
Lane Detection
 
3D Shape and Indirect Appearance by Structured Light Transport
3D Shape and Indirect Appearance by Structured Light Transport3D Shape and Indirect Appearance by Structured Light Transport
3D Shape and Indirect Appearance by Structured Light Transport
 
Ray casting algorithm by mhm
Ray casting algorithm by mhmRay casting algorithm by mhm
Ray casting algorithm by mhm
 
Lane detection by use of canny edge
Lane detection by use of canny edgeLane detection by use of canny edge
Lane detection by use of canny edge
 
Pengantar Structure from Motion Photogrammetry
Pengantar Structure from Motion PhotogrammetryPengantar Structure from Motion Photogrammetry
Pengantar Structure from Motion Photogrammetry
 
mid_presentation
mid_presentationmid_presentation
mid_presentation
 
DimEye Corp Presents Revolutionary VLS (Video Laser Scan) at SS IMMR 2013
DimEye Corp Presents Revolutionary VLS (Video Laser Scan) at SS IMMR 2013DimEye Corp Presents Revolutionary VLS (Video Laser Scan) at SS IMMR 2013
DimEye Corp Presents Revolutionary VLS (Video Laser Scan) at SS IMMR 2013
 
3-d interpretation from single 2-d image for autonomous driving
3-d interpretation from single 2-d image for autonomous driving3-d interpretation from single 2-d image for autonomous driving
3-d interpretation from single 2-d image for autonomous driving
 
Temporal Frequency Probing for 5D Transient Analysis of Global Light Transport
Temporal Frequency Probing for 5D Transient Analysis of Global Light TransportTemporal Frequency Probing for 5D Transient Analysis of Global Light Transport
Temporal Frequency Probing for 5D Transient Analysis of Global Light Transport
 

Ähnlich wie Poster_Final

IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
ijceronline
 
194Martin LeungUnerd Poster
194Martin LeungUnerd Poster194Martin LeungUnerd Poster
194Martin LeungUnerd Poster
Martin Leung
 
Super Resolution of Image
Super Resolution of ImageSuper Resolution of Image
Super Resolution of Image
Satheesh K
 

Ähnlich wie Poster_Final (20)

Fisheye Omnidirectional View in Autonomous Driving
Fisheye Omnidirectional View in Autonomous DrivingFisheye Omnidirectional View in Autonomous Driving
Fisheye Omnidirectional View in Autonomous Driving
 
Video Stitching using Improved RANSAC and SIFT
Video Stitching using Improved RANSAC and SIFTVideo Stitching using Improved RANSAC and SIFT
Video Stitching using Improved RANSAC and SIFT
 
Final Paper
Final PaperFinal Paper
Final Paper
 
SkyStitch: a Cooperative Multi-UAV-based Real-time Video Surveillance System ...
SkyStitch: a Cooperative Multi-UAV-based Real-time Video Surveillance System ...SkyStitch: a Cooperative Multi-UAV-based Real-time Video Surveillance System ...
SkyStitch: a Cooperative Multi-UAV-based Real-time Video Surveillance System ...
 
Leader Follower Formation Control of Ground Vehicles Using Dynamic Pixel Coun...
Leader Follower Formation Control of Ground Vehicles Using Dynamic Pixel Coun...Leader Follower Formation Control of Ground Vehicles Using Dynamic Pixel Coun...
Leader Follower Formation Control of Ground Vehicles Using Dynamic Pixel Coun...
 
Isvc08
Isvc08Isvc08
Isvc08
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 
Visual Environment by Semantic Segmentation Using Deep Learning: A Prototype ...
Visual Environment by Semantic Segmentation Using Deep Learning: A Prototype ...Visual Environment by Semantic Segmentation Using Deep Learning: A Prototype ...
Visual Environment by Semantic Segmentation Using Deep Learning: A Prototype ...
 
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
 
Calibration Issues in FRC: Camera, Projector, Kinematics based Hybrid Approac...
Calibration Issues in FRC: Camera, Projector, Kinematics based Hybrid Approac...Calibration Issues in FRC: Camera, Projector, Kinematics based Hybrid Approac...
Calibration Issues in FRC: Camera, Projector, Kinematics based Hybrid Approac...
 
A ROS IMPLEMENTATION OF THE MONO-SLAM ALGORITHM
A ROS IMPLEMENTATION OF THE MONO-SLAM ALGORITHMA ROS IMPLEMENTATION OF THE MONO-SLAM ALGORITHM
A ROS IMPLEMENTATION OF THE MONO-SLAM ALGORITHM
 
Recognition and tracking moving objects using moving camera in complex scenes
Recognition and tracking moving objects using moving camera in complex scenesRecognition and tracking moving objects using moving camera in complex scenes
Recognition and tracking moving objects using moving camera in complex scenes
 
194Martin LeungUnerd Poster
194Martin LeungUnerd Poster194Martin LeungUnerd Poster
194Martin LeungUnerd Poster
 
Tracking Chessboard Corners Using Projective Transformation for Augmented Rea...
Tracking Chessboard Corners Using Projective Transformation for Augmented Rea...Tracking Chessboard Corners Using Projective Transformation for Augmented Rea...
Tracking Chessboard Corners Using Projective Transformation for Augmented Rea...
 
Super Resolution of Image
Super Resolution of ImageSuper Resolution of Image
Super Resolution of Image
 
DICTA 2017 poster
DICTA 2017 posterDICTA 2017 poster
DICTA 2017 poster
 
Computer Vision panoramas
Computer Vision  panoramasComputer Vision  panoramas
Computer Vision panoramas
 
AR/SLAM for end-users
AR/SLAM for end-usersAR/SLAM for end-users
AR/SLAM for end-users
 
Deep Learning for Structure-from-Motion (SfM)
Deep Learning for Structure-from-Motion (SfM)Deep Learning for Structure-from-Motion (SfM)
Deep Learning for Structure-from-Motion (SfM)
 
SIGGRAPH 2014 Course on Computational Cameras and Displays (part 4)
SIGGRAPH 2014 Course on Computational Cameras and Displays (part 4)SIGGRAPH 2014 Course on Computational Cameras and Displays (part 4)
SIGGRAPH 2014 Course on Computational Cameras and Displays (part 4)
 

Poster_Final

  • 1.  4        Method   Camera  calibra1on   Each  camera  has  its  own  unique  intrinsic  parameters,  which  is  expressed  by  a  3x3   camera  calibra9on  matrix,  K.  We  obtain  this  matrix  by  using  a  video  calibra9on   technique  implemented  with  OpenCV  using  mul9ple  checkerboard  images  as   shown  below.  (3  of  25  images  are  displayed  here)   Raw  image   Image  with  lens  distor9on  removed   Mo1va1on   •  Explora9on  using  robots  can  be  useful  for  examining  interior   structures   •  Panoramas   create   a   wider   field   of   view   to   increase   the     teleoperator’s  3D  awareness   •  To   render   a   real-­‐9me   visualiza9on   to   the   teleoperator,   the     implemented   algorithm   should   s9tch   video   frames   together   as  efficiently  as  possible    1        Introduc1on           Assump1ons   •  Lens  distor9on  is  removed  from  video   •  Intrinsic  camera  parameters  are  known     Problem  statement   From  a  video,  detect  features  in  frame  In  and  match  these  features  in  consecu9ve   frames  in  order  to  detect  the  key  frame  In+1  from  the  calculated  homography  Hn,n+1   Use  the  saved  key  frames  and  homographies  to  create  a  s9tched  panorama  of  the   environment.     2        Problem  Statement    5    Experimental  Results    6    References   [1]  Brown,  MaYhew,  and  David  G.  Lowe.  "Automa9c  Panoramic  Image  S9tching   Using  Invariant  Features."  Int  J  Comput  Vision  Interna1onal  Journal  of  Computer   Vision  74.1  (2006):  59-­‐73.   [2]  Shi,  Jianbo,  and  Tomasi.  "Good  Features  to  Track."  Proceedings  of  IEEE   Conference  on  Computer  Vision  and  Pa9ern  Recogni1on  CVPR-­‐94  (1994).    7    Acknowledgments   This  research  was  supported  by  the  Na9onal  Science  Founda9on   Research  Experiences  for  Undergraduates  (REU)  program  on   Interdisciplinary  Research  in  Mechatronics,  Robo9cs,  and   Automated  System  Design  (NSF  grant  no.  1263293).  Any   opinions,  findings,  and  conclusions  or  recommenda9ons   expressed  in  this  material  are  those  of  the  author(s)  and  do  not   necessarily  reflect  the  views  of  the  Na9onal  Science  Founda9on.    3        System  Overview   25 50 75 100 0 0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16 0.18 Number of Features Time(s) Time Comparison Between Shi−Tomasi/Optical Flow and SIFT/FLANN Shi−Tomasi Corner/Optical Flow SIFT/FLANN Sample  images  for  calibra9on   Feature  Detec1on  and  Matching   There  are  many  different  methods  used  to  detect  features.  Scale  Invariant  Feature   Detec9on,  SIFT,  is  a  popular  method  that  is  known  to  be  robust  as  it  is  invariant  to   changes  in  image  scale  [1].  Shi-­‐Tomasi’s  method  is  a  technique  which  only  tacks   corner  features  in  an  image  [2],  which  is  ohen  faster.  SIFT  features  are  matched   between  frames  using  the  Fast  Library  for  Approximate  Nearest  Neighbors,  FLANN,     method.  The  Shi-­‐Tomasi  technique  tracks  the  corners  with  op9cal  flow  which  uses   the  gradient  in  the  neighborhood  of  a  feature  to  track  its  mo9on  among  frames.                         Feature  detec9on  using  SIFT  and  matching  using  FLANN                     Feature  detec9on  using  Shi-­‐Tomasi  Corner  and  matching  using  Op9cal  Flow     Key  Frame  Selec1on   In  order  to  get  a  well  s9tched  panoramic  image,  it  is  important  to  choose  a  frame  I’   that   will   overlap   well   with   its   previous   frame   I.   A   homography   H   is   a   projec9ve   transforma9on   that   relates   image   points   by   x’   =   Hx,   where   x   and   x’   are   image   points   in   I   and   I’,   respec9vely.   Image   s9tching   is   accomplished   by   calcula9ng   H   between   the   two   images.   Op9cal   flow   tracks   features   among   frames.   When   tracking  n  features,  once  the  successfully  tracked  features  n’  in  the  current  frame  is   less   than   n/2,   the   camera   has   moved   a   sufficient   amount   to   provide   a   good   homography.     Image  S1tching   The  key  frames  are  s9tched  to  a  panoramic  texture  that  is  mapped  onto  a  cylinder.   From  H  and  K,  the  camera  rota9on  R  is  es9mated.  Based  on  this  pose,  an  image   point  x  in  the  world  space  of  the  cylinder  is  determined  by  X  =  K-­‐1R-­‐1x  where  X  =  (x,   y,  z)  .  These  world  coordinates  are  then  mapped  to  UV  texture  coordinates  where  u   =  arctan(x/z)  and  v  =  y/√(x2  +  z2).  The  final  step  is  to  texture  map  the  cylinder  by  x   =  r*cos(u),  y  =  v,  z  =  r*sin(u)  where  r  is  the  radius  of  the  cylinder.   Shi-­‐Tomasi  and  Op9cal  Flow   SIFT  and  FLANN   Shi-­‐Tomasi  and  Op9cal  Flow   SIFT  and  FLANN   S9tched  images  with  cylindrical  projec9on  Input  Images