SlideShare ist ein Scribd-Unternehmen logo
1 von 4
Downloaden Sie, um offline zu lesen
Visual Servoing in Presence of Non-Rigid Motion

                                                        D Santosh Kumar, C V Jawahar
                                                   Center for Visual Information Technology
                                               International Institute of Information Technology
                                                           Hyderabad 500032, India
                                                    {santosh@students.,jawahar@}iiit.ac.in


                                       Abstract                               are non-rigid [2] in nature. For example, the articulated hu-
                                                                              man motion; the deforming nature of the beating heart, fly-
                 Most robotic vision algorithms have been proposed by
                                                                              ing birds and moving aquatic animals; the near-rigid nature
              envisaging robots operating in industrial environments,
                                                                              of the human face etc. If robots are to be used in pres-
              where the world is assumed to be static and rigid. These
                                                                              ence of such objects, we need servoing algorithms that are
              algorithms cannot be used in environments where the as-
                                                                              adaptable to non-rigidity. The work done by [7] and [10]
              sumption of a rigid world does not hold. In this paper, we
                                                                              deals with only moving rigid object visual servoing. Re-
              study the problem of visual servoing in presence of non-
                                                                              cently in [4], Cavusoglu et al. proposed a method to handle
              rigid objects and analyze the design of servoing strategies
                                                                              non-rigidness by canceling the relative motion between the
              needed to perform optimally even in unconventional envi-
                                                                              non-rigid object (heart) and the camera by using a model
              ronments. We also propose a servoing algorithm that is ro-
                                                                              predictive controller so as to facilitate easy manipulation of
              bust to non-rigidity. The algorithm extracts invariant fea-
                                                                              the object. In this paper, we address the issue of position-
              tures of the non-rigid object and uses these features in the
                                                                              ing of the end-effector of a robot with respect to a non-rigid
              servoing process. We validate the technique with experi-
                                                                              object. Often, the positioning task is a pre-requisite for the
              ments and demonstrate the applicability.
                                                                              manipulation task.
              1. Introduction                                                     As a non-rigid object undergoes a persistent change in
                                                                              its pose, it is preferable to extract pose-invariant features
                 Visual servoing [8] describes a broad class of navigation    of the object and use them in the servoing process. As a
              problems in which a robot is positioned with respect to a       first attempt towards analyzing this problem, we investigate
              target using computer vision as the primary feedback sen-       the various issues that need to be considered while solving it
              sor. It is achieved by processing the visual feedback (2D       and also propose one simple yet an effective visual servoing
              features or 3D pose or both) and minimizing an appropriate      algorithm to handle deformations. The algorithm is based
              error function.                                                 on a not so strict assumption that any non-rigid motion can
                 Several algorithms [1, 3, 7, 10] have been proposed to       be approximated to a generalized quasi-periodic motion.
              perform the task of visual servoing; however, much of the       And for a quasi-periodic motion, gross measures such as
              research until now focuses only on rigid objects. As far as     the moving average are sufficient to reveal its general sta-
              we are aware, the problem of visual servoing in presence of     tionary characteristics, provided the change in pose can be
              non-rigid objects has not been addressed in literature. The     compensated. By using the recent results of multiple-view
              challenge with non-rigidity is that the motion planned based    geometry, we efficiently extract such static features of the
              on the image-clues extracted at current time instant might      non-rigid object, which enable us to use the conventional
              not be valid in the next time instant as the object under-      visual servoing techniques even in this case.
              goes a change in its pose. This problem is different from
              that of servoing with respect to a moving object, where all
                                                                              2. Background
              the features that are being tracked undergo the same or pre-
              dictable transformations in the image space. Whereas, for a
              deforming object, each feature point transforms differently        In image-based visual servo control, the error signal e
              and thus generates different control signals at the same time   is defined directly in terms of image feature parameters as
              instant, thereby destabilizing the motion planning process.     e(S) = S − S ∗ , where S and S ∗ are the current and the
                 Many of the living forms that we come across everyday        desired feature parameters respectively. By differentiating




The 18th International Conference on Pattern Recognition (ICPR'06)
0-7695-2521-0/06 $20.00 © 2006
this error function with respect to time, we get:                     Any servoing algorithm that computes an optimal path
                                                                                 for the robot to traverse from the initial position to the de-
                             de   dS   ∂S dr                                     sired position, can be divided into two basic steps. After de-
                                =    =( )    = LS V,                      (1)
                             dt   dt   ∂r dt                                     ciding on an appropriate feature to represent the non-rigid
                                                                                 object, the following steps need to be repeated until the de-
              where V = (v T , ω T )T is the camera screw, r is the cam-
                                                                                 sired position is reached.
              era pose and LS is the interaction matrix [8]. It relates the
              motion of the features in the image space to the camera              • Compute the error estimate between the current and
              motion in the Cartesian space. The main objective of the               the desired feature vector i.e., S − S ∗
              visual servoing process is to minimize the error function            • Using a proportional control law (see Section 2), move
              e(S). For exponential convergence, we use a proportional               in the direction that minimizes the above error
              control law i.e., de(S) = −λe(S). By substituting this in
                                  dt                                                 It is crucial to have a good feature vector S to represent
              (1), the required velocity of the camera can be computed as        the object state and thereby to perform the servoing task.
              V = −λL+ e(S), where L+ is the pseudo-inverse of the in-
                          S               S                                      However for non-rigid objects as the object pose contin-
              teraction matrix and λ is a scale factor. In case of a moving      uously changes, we cannot obtain a feature vector from a
              object, the expression for V will get modified as (see [8])         single image that characterizes its state as it might lead to
                                                                                 an unstable end-effector trajectory. For instance, a surgical
                                                        ∂e
                               V = −λL+ e(S) − L+
                                      S         S          ,              (2)    robot positioning itself only using the current image of the
                                                        ∂t                       heart, will oscillate corresponding to the systole and dias-
                                                                                 tole states of the heart during its motion. Moreover, if only
              where ∂e represents the object motion model. The image-
                      ∂t                                                         a single image is used to describe the desired position S ∗ ,
              based visual servoing algorithm has gained prominence in           it will lead to oscillations of the manipulator even after the
              literature due to its robustness to depth estimation errors.       goal position is reached. One possible solution to circum-
                  Note that the above analysis assumes the objects to be         vent this problem would be to use the sequence of images
              rigid. Though non-rigid objects have not been studied in the       representing all states of the object to perform the servo-
              visual servoing framework, they have been well analyzed in         ing task. This will generate better control signals as it takes
              computer vision in the context of non-rigid tracking and 3D        into account the object deformations. However, a more effi-
              reconstruction problems. Most of these algorithms employ           cient solution could be to extract gross features from those
              models like the linear subspace model, appearance mani-            sequence of images by exploiting the stationary signal prop-
              fold [11], kinematic chains [5, 9] etc in order to maintain        erty of the quasi-periodic object deformations.
              a representation of the non-rigid object and also to infer its         At each pose, a sequence of images (I1 , I2 , ..., In ) are
              internal motion model. Such models allow prediction of the         acquired by the camera, capturing the deformations of the
              future states of the object, given its current state. Prediction   non-rigid object (where n depends on the periodicity of the
              is necessary in case of manipulation tasks where the end-          deformations). Then, m point features are extracted from
              effector needs to interact with the deforming object. For the      each image Ii to obtain a feature vector Fit = [f1 f2 ...fm ]T ,
              positioning task, only an approximate pose of the object is        where each feature fk corresponds to the projection of a 3D
              sufficient to align the camera with respect to the object. Any      point on the surface of the non-rigid object at time instant
              error resulting due to this approximation can be corrected         t. The invariant feature Gt corresponding to the image se-
              by using the visual feedback.                                      quence is computed as a function of all the feature vectors.
                                                                                                    Gt = f (F1 , F2 , ...., Fn )
                                                                                                             t    t          t
                                                                                                                                             (3)
              3. Visual Servoing in Presence of Deformations
                                                                                 However, the camera can use its previous knowledge and
                  Our problem of visual servoing is to position the end-         predict the next gross appearance of the deforming object.
              effector of the robot with respect to a deforming object.          Thus at each position, only a single image is captured by
              Several strategies could be conceived to solve this problem.       the camera. The rest of the deformations of the object are
              One possible solution is to use clues from object appearance       obtained by projecting the past images into the new pose.
              to aid the positioning task. Alternatively, one could gain a       Hence, the gross feature at time t can be predictively calcu-
              stable representation of the object from its motion charac-        lated as a function of the features extracted from the current
              teristics, which can be used to calibrate and move the end-        image and the previous images as
              effector. Another solution is to extract some pose-invariant                 Gt = f (g(F t−n , ...., F t−2 , F t−1 ), F t ),   (4)
              features of the non-rigid object and use these features in the
              servoing process. Here we adopt a relatively simple strategy       where F t−i denotes the feature vector obtained from
              to solve this problem.                                             the image captured at camera pose rt−i , F t denotes the




The 18th International Conference on Pattern Recognition (ICPR'06)
0-7695-2521-0/06 $20.00 © 2006
feature vector extracted from the current image and g(·)            We demonstrate our results using two kinds of features
              is a function that projects the past image features into the    namely the mean and the eigen feature. Such features have
              current view as explained below.                                been used in computer vision to capture statistical regulari-
                                                                              ties and redundancies shared by a set of images. In the first
              Projection of Images Given two camera matrices M1               method, all the feature vectors Fi , i = 1, 2, ..., n obtained
              and M2 and an image I1 taken from the pose corresponding        from the n images are averaged to get the mean feature vec-
              to M1 , the task is to obtain the image I2 as seen from the     tor M . In this context, the mean operator corresponds to
              pose corresponding to M2 using a partial 3D model of the        the function f (·) and M represents the invariant feature Gt .
              world that is assumed to be available in the image-based        This feature is used as the input S to the servoing algorithm.
              visual servoing process [8]. Let M1 = K1 [R1 t1 ] and           As explained in Section 3, the new mean is predictively cal-
              M2 = K2 [R2 t2 ] where Ki denotes the intrinsic para-           culated using the past image features.
              meter matrix and [Ri ti ] denotes the pose (rotation and            In the second method, the eigen feature of the sequence
              translation). Let p1 represent a point in I1 and Z1 be an       of image features is computed and used as the input feature
              estimate of the depth of the corresponding 3D point P1 in       to the servoing process. This feature indicates the direction
              the camera frame given by the 3D model. The coordinates         in which the point features undergo maximum variation in
                                                 −1
              of point P1 are given by P1 = K1 Z1 p1 (see [6]).               the image space. In estimating the camera screw from (2),
                 Now to obtain the 3D coordinate of this point in the other   the line Jacobian is used to compute V since the eigen fea-
              camera frame, we have                                           ture can be conceived of as a line segment in the image.
                                                                                  In comparison to classical image-based visual servoing,
                               P2 = [R2 t2 ][R1 t1 ]−1 P1 .            (5)    the use of above features generated a stable camera trajec-
                                                                              tory. Fig. 1a shows the velocity screw obtained in the case
              Finally, the image coordinates of the point with respect to     of the classical servoing algorithm. As only a single image
                                                        1
              the second camera are given by p2 = Z2 K2 P2 . Thus by          is used, the generated control instructions resulted in an os-
              projecting the past image features into the new view, all       cillatory camera trajectory (see the jagged behavior of the
              the deformations of the object as viewed from the current       velocity screw in Fig. 1a as compared to that of Fig. 1b),
              camera pose can be obtained. These features are used to         affecting the stability of the robot. Also the system set-
              compute the representative feature of the non-rigid object.     tled in a critically stable condition as the oscillations of the
                                                                              end-effector continued even after the desired position was
              4. Results and Discussions                                      reached.

                 In our experiments, we consider a set of four 3D points                                            Classical Image Based Visual Servoing                                                       Servoing using Eigen Feature
                                                                                                               1                                                                                            2



                                                                                                                                                                                       Trans. Vel (m/sec)
                                                                                        Trans. Vel. (m/sec)




                                                                                                                                                              vx                                                                                  vx
                                                                                                                                                                                                                                                  vy
              on the surface of a non-rigid object being observed by a                                         0
                                                                                                                                                              vy
                                                                                                                                                              vz
                                                                                                                                                                                                            0
                                                                                                                                                                                                                                                  vz


              perspective camera. The projection of the points onto the                                       −1


                                                                                                                                                                                                     −2
              image are considered as features. The deformations of the                                       −2
                                                                                                                0        2     4      6      8       10
                                                                                                                                          Time (in seconds)
                                                                                                                                                                   12                                  0             2        4          6
                                                                                                                                                                                                                                  Time (In Seconds)
                                                                                                                                                                                                                                                       8
                                                                                                                                                                                                            1
                                                                                                                                                                            Rot. Vel (rad/sec)




                                                                                                              0.6
              non-rigid object are simulated by using a sinusoidal motion
                                                                               Rot. Vel. (rad/sec)




                                                                                                                                                          wx                                                                                     wx
                                                                                                              0.4                                         wy                                                                                     wy
                                                                                                                                                          wz                                                                                     wz
              perturbed by a white gaussian noise. This makes the object                                      0.2                                                                                           0
                                                                                                               0
              undergo quasi-periodic deformations. The task of position-                             −0.2                                                                                            −1
                                                                                                         0               2     4      6      8       10            12                                  0             2        4          6             8
              ing a camera with respect to the non-rigid object is studied.                                                               Time (in seconds)             a                                                         Time (In Seconds)        b
                 The basic implementation of the proposed algorithm can
              be summarized into the following steps.                                                         Figure 1. Velocity Screw (a) Oscillatory be-
                                                                                                              havior (features only from single image are
                1. In an offline step, acquire images from the desired fi-                                      used) (b) Smooth convergence
                   nal position. Extract feature S ∗ from these images us-
                   ing (3)                                                       We also tested the use of image sequence as an input
                2. Acquire a sequence of images from the current camera       feature to the servoing algorithm. In this case, the feature
                   pose and calculate S, again using (3)                      S was the entire sequence of features extracted from each
                3. Compute the velocity screw V using (2)                     image in the sequence i.e., S = [F1 , F2 , ..., Fn ]T . As a
                4. Using V , move the end-effector to the new pose and        large set of features are considered, minimizing the error
                   capture a new image                                        difference between the current and the desired feature
                5. Project the features in the old images into the new pose   vector consumed more time and led to slower convergence
                   as discussed in the latter half of Section 3               of the error function (compare the error norm and the time
                6. Using the features extracted from the old images and       taken in Fig. 2a to that of Fig. 2b).
                   the current image, extract S using (4)
                7. Repeat steps 3 to 6 until e(S) < θ                         Stability and Robustness We empirically verify the




The 18th International Conference on Pattern Recognition (ICPR'06)
0-7695-2521-0/06 $20.00 © 2006
Convergence of Error Norm                                                                           Error Convergence using features
                                      200
                                                                                                                                     1400
                                                                                                                                                           from Image Sequence                                5. Conclusion
                                                                         Eigen feature
                                                                         Mean feature                                                1200
                                      150
                                                                                                                                                                                                                 Existing servoing schemes cannot adapt to the non-




                                                                                                             Error(in pixels)
                                                                                                                                     1000
               Error(in pixels)




                                      100
                                                                                                                                           800
                                                                                                                                                                                                              rigidity present in domestic environments. In this paper, we
                                                                                                                                           600
                                                                                                                                                                                                              studied the issues involved in designing servoing strategies
                                       50                                                                                                  400

                                                                                                                                           200
                                                                                                                                                                                                              with optimal performance in presence of non-rigid objects.
                                        0
                                         0           2          4           6            8                                                      0
                                                                                                                                                                                                              We also proposed a novel image-based visual servoing al-
                                                         Time(in seconds)                    a                                                   0       2      4         6
                                                                                                                                                              Time(in seconds)
                                                                                                                                                                                    8                10
                                                                                                                                                                                                          b   gorithm, which efficiently extracted gross features from se-
                                                                                                                                                                                                              quence of images and used them in the servoing process.
                                        Figure 2. Error norm in case of (a) Eigen and                                                                                                                         This algorithm generated a stable camera trajectory unlike
                                        mean feature (b) Image sequence                                                                                                                                       the classical servoing algorithms. In future, we plan to im-
              robustness of our approach to calibration errors. Often                                                                                                                                         prove the algorithm by considering other invariant features
              due to system calibration errors, the obtained image                                                                                                                                            which better characterize the non-rigid object motion.
              measurements are corrupted. The current image will be
              corrupted by the error in the current camera pose whereas                                                                                                                                       References
              the projected images will be corrupted by the error in the
              previous camera poses. This leads to the incorrect compu-                                                                                                                                        [1] J. Adachi and J. Sato. Uncalibrated visual servoing from
              tation of the error function as e (S) = e(S) + Δe where                                                                                                                                              projective reconstruction of control values. International
              e(S) is as defined in Section 2 and Δe is the measurement                                                                                                                                             Conference on Pattern Recognition, 4:297–300, August
              error obtained due to the corresponding error in pose. As                                                                                                                                            2004.
                                                                                                                                                                                                               [2] J. K. Aggarwal, Q. Cai, W. Liao, and B. Sabata. Non-
              a result, the velocity screw changes to V = V + ΔV                                                                                                                                                   rigid motion analysis: articulated and elastic motion. Com-
              where V is as given in (2) and ΔV is the velocity command                                                                                                                                            puter Vision and Image Understanding, 70(2):142–156,
              generated to compensate the corresponding measurement                                                                                                                                                May 1998.
              error Δe. This ΔV accommodates the uncertainty in pose                                                                                                                                           [3] J. P. Barreto, J. Batista, and H. Araujo. Model predictive
              and thus ensures the convergence of error to zero. The                                                                                                                                               control to improve visual control of motion: Applications in
              stability of the algorithm was tested by introducing noise                                                                                                                                           active tracking of moving targets. International Conference
              into the camera extrinsic parameters. The noise level was                                                                                                                                            on Pattern Recognition, 4:4732–4735, September 2000.
                                                                                                                                                                                                               [4] M. C. Cavusoglu, J. Rotella, W. S. Newman, S. Choi,
              varied and the performance of the algorithm was studied.
                                                                                                                                                                                                                   J. Ustin, and S. S. Sastry. Control algorithms for active rel-
              In Fig. 3a, it is observed that in spite of the calibration
                                                                                                                                                                                                                   ative motion cancelling for robotic assisted off-pump coro-
              errors (a 10% gaussain noise was added), the convergence                                                                                                                                             nary artery bypass graft surgery. International Conference
              of velocity screw is achieved. This supports our analysis                                                                                                                                            on Advanced Robotics, pages 431–436, July 2005.
              that uncertainty in pose is corrected by the visual feedback                                                                                                                                     [5] F. Chaumette, E. Marchand, and A. Comport. Object-
              rather than getting propagated due to the prediction step.                                                                                                                                           based visual 3d tracking of articulated objects via kinematic
                                                                                                                                                                                                                   sets. IEEE Workshop on Articulated and Non-Rigid Motion,
                                       2
                                            Velocity Screw with calibration errors                                                         Velocity Screw with depth estimation errors
                                                                                                                                            2
                                                                                                                                                                                                                   CVPRW, 1:2–9, June 2004.
                                                                                                                      Trans. Vel (m/sec)
                 Trans. Vel (m/sec)




                                                                                    vx
                                                                                    vy
                                                                                    vz
                                                                                                                                                                                            vx
                                                                                                                                                                                            vy                 [6] R. Hartley and A. Zisserman. Multiple view geometry in
                                                                                                                                                                                            vz
                                       0                                                                                                    0                                                                      computer vision. Cambridge University Press, 2003.
                                      −2
                                                                                                                                                                                                               [7] K. Hashimoto, K. Nagahama, T. Noritsugu, and M. Taka-
                                                                                                                                           −2
                                        0           2          4          6              8                                                   0           2          4          6                 8
                                                                    Time (In Seconds)                                                                                   Time (in Seconds)                          iawa. Visual servoing based on object motion estimation.
                                       1
                Rot. Vel (rad/sec)




                                                                                                                                            1
                                                                                                 Rot. Vel (rad/sec)




                                      0.5
                                                                                   wx
                                                                                   wy
                                                                                                                                                                                        wx
                                                                                                                                                                                        wy                         IEEE/RSJ International Conference on Intelligent Robots
                                                                                   wz                                              0.5                                                  wz

                                       0                                                                                                    0
                                                                                                                                                                                                                   and Systems, 1:245–250, October 2000.
                             −0.5
                                                                                                                                                                                                               [8] S. A. Hutchinson, G. D. Hager, and P. I. Corke. A tutorial
                                                                                                                       −0.5
                                 0                  2          4          6              8                                 0                             2          4          6                 8
                                                                    Time (In Seconds)        a                                                                           Time (in Seconds)            b            on visual servo control. IEEE Transactions on Robotics and
                                                                                                                                                                                                                   Automation, 12(5):651–670, October 1996.
                                        Figure 3. Convergence of velocity screw ob-                                                                                                                            [9] T. Kanade, D. D. Morris, and J. Rehg. Ambiguities in vi-
                                        served even in presence of errors (noise)                                                                                                                                  sual tracking of articulated objects using two- and three-
                                                                                                                                                                                                                   dimensional models. International Journal of Robotics Re-
                 One of the crucial steps in the algorithm is the predic-                                                                                                                                          search, 22(6):393–418, June 2003.
              tion step that assumes the depth of the feature points. The                                                                                                                                     [10] E. Malis and S. Benhimane. A unified approach to visual
              robustness of the algorithm to depth estimation errors was                                                                                                                                           tracking and servoing. Robotics and Autonomous Systems,
              tested by introducing noise into the depth values. Fig. 3b                                                                                                                                           52(1):39–52, July 2005.
                                                                                                                                                                                                              [11] A. Shashua, A. Levin, and S. Avidan. Manifold pursuit:
              shows the velocity screw obtained when the estimated depth
                                                                                                                                                                                                                   A new approach to appearance based recognition. Interna-
              was corrupted by 15% gaussian noise. This experiment                                                                                                                                                 tional Conference of Pattern Recognition, 3:590–594, Au-
              demonstrated that minor error in the depth values can be                                                                                                                                             gust 2002.
              accommodated by our algorithm.




The 18th International Conference on Pattern Recognition (ICPR'06)
0-7695-2521-0/06 $20.00 © 2006

Weitere ähnliche Inhalte

Was ist angesagt?

Fcv bio cv_simoncelli
Fcv bio cv_simoncelliFcv bio cv_simoncelli
Fcv bio cv_simoncellizukun
 
Fingerprint recognition using correlation
Fingerprint recognition using correlationFingerprint recognition using correlation
Fingerprint recognition using correlationWABCO
 
Sparse feature analysis for detection of clustered microcalcifications in mam...
Sparse feature analysis for detection of clustered microcalcifications in mam...Sparse feature analysis for detection of clustered microcalcifications in mam...
Sparse feature analysis for detection of clustered microcalcifications in mam...Wesley De Neve
 
Optimally Learnt, Neural Network Based Autonomous Mobile Robot Navigation System
Optimally Learnt, Neural Network Based Autonomous Mobile Robot Navigation SystemOptimally Learnt, Neural Network Based Autonomous Mobile Robot Navigation System
Optimally Learnt, Neural Network Based Autonomous Mobile Robot Navigation SystemIDES Editor
 
Critical Features for Recognition of Biological Motion
Critical Features for Recognition of Biological MotionCritical Features for Recognition of Biological Motion
Critical Features for Recognition of Biological MotionDimitris Agorastos
 
Biometric simulator for visually impaired (1)
Biometric simulator for visually impaired (1)Biometric simulator for visually impaired (1)
Biometric simulator for visually impaired (1)Rahul Bhagat
 
Uniform and non uniform single image deblurring based on sparse representatio...
Uniform and non uniform single image deblurring based on sparse representatio...Uniform and non uniform single image deblurring based on sparse representatio...
Uniform and non uniform single image deblurring based on sparse representatio...ijma
 
Kccsi 2012 a real-time robust object tracking-v2
Kccsi 2012   a real-time robust object tracking-v2Kccsi 2012   a real-time robust object tracking-v2
Kccsi 2012 a real-time robust object tracking-v2Prarinya Siritanawan
 
An Object Detection, Tracking And Parametric Classification– A Review
An Object Detection, Tracking And Parametric Classification– A ReviewAn Object Detection, Tracking And Parametric Classification– A Review
An Object Detection, Tracking And Parametric Classification– A ReviewIRJET Journal
 
Komogortsev Qualitative And Quantitative Scoring And Evaluation Of The Eye Mo...
Komogortsev Qualitative And Quantitative Scoring And Evaluation Of The Eye Mo...Komogortsev Qualitative And Quantitative Scoring And Evaluation Of The Eye Mo...
Komogortsev Qualitative And Quantitative Scoring And Evaluation Of The Eye Mo...Kalle
 
201109CVIM/PRMU Inverse Composite Alignment of a sphere under orthogonal proj...
201109CVIM/PRMU Inverse Composite Alignment of a sphere under orthogonal proj...201109CVIM/PRMU Inverse Composite Alignment of a sphere under orthogonal proj...
201109CVIM/PRMU Inverse Composite Alignment of a sphere under orthogonal proj...Toru Tamaki
 
Abaqus users (commercial finite element code
Abaqus users (commercial finite element codeAbaqus users (commercial finite element code
Abaqus users (commercial finite element codebasma2006
 
Ostu Thresholding
Ostu ThresholdingOstu Thresholding
Ostu Thresholdingasd_gtgh
 

Was ist angesagt? (17)

Fcv bio cv_simoncelli
Fcv bio cv_simoncelliFcv bio cv_simoncelli
Fcv bio cv_simoncelli
 
Fingerprint recognition using correlation
Fingerprint recognition using correlationFingerprint recognition using correlation
Fingerprint recognition using correlation
 
Sparse feature analysis for detection of clustered microcalcifications in mam...
Sparse feature analysis for detection of clustered microcalcifications in mam...Sparse feature analysis for detection of clustered microcalcifications in mam...
Sparse feature analysis for detection of clustered microcalcifications in mam...
 
Two
TwoTwo
Two
 
Optimally Learnt, Neural Network Based Autonomous Mobile Robot Navigation System
Optimally Learnt, Neural Network Based Autonomous Mobile Robot Navigation SystemOptimally Learnt, Neural Network Based Autonomous Mobile Robot Navigation System
Optimally Learnt, Neural Network Based Autonomous Mobile Robot Navigation System
 
Critical Features for Recognition of Biological Motion
Critical Features for Recognition of Biological MotionCritical Features for Recognition of Biological Motion
Critical Features for Recognition of Biological Motion
 
Biometric simulator for visually impaired (1)
Biometric simulator for visually impaired (1)Biometric simulator for visually impaired (1)
Biometric simulator for visually impaired (1)
 
Uniform and non uniform single image deblurring based on sparse representatio...
Uniform and non uniform single image deblurring based on sparse representatio...Uniform and non uniform single image deblurring based on sparse representatio...
Uniform and non uniform single image deblurring based on sparse representatio...
 
F0164348
F0164348F0164348
F0164348
 
Kccsi 2012 a real-time robust object tracking-v2
Kccsi 2012   a real-time robust object tracking-v2Kccsi 2012   a real-time robust object tracking-v2
Kccsi 2012 a real-time robust object tracking-v2
 
An Object Detection, Tracking And Parametric Classification– A Review
An Object Detection, Tracking And Parametric Classification– A ReviewAn Object Detection, Tracking And Parametric Classification– A Review
An Object Detection, Tracking And Parametric Classification– A Review
 
L010427275
L010427275L010427275
L010427275
 
Komogortsev Qualitative And Quantitative Scoring And Evaluation Of The Eye Mo...
Komogortsev Qualitative And Quantitative Scoring And Evaluation Of The Eye Mo...Komogortsev Qualitative And Quantitative Scoring And Evaluation Of The Eye Mo...
Komogortsev Qualitative And Quantitative Scoring And Evaluation Of The Eye Mo...
 
K0966468
K0966468K0966468
K0966468
 
201109CVIM/PRMU Inverse Composite Alignment of a sphere under orthogonal proj...
201109CVIM/PRMU Inverse Composite Alignment of a sphere under orthogonal proj...201109CVIM/PRMU Inverse Composite Alignment of a sphere under orthogonal proj...
201109CVIM/PRMU Inverse Composite Alignment of a sphere under orthogonal proj...
 
Abaqus users (commercial finite element code
Abaqus users (commercial finite element codeAbaqus users (commercial finite element code
Abaqus users (commercial finite element code
 
Ostu Thresholding
Ostu ThresholdingOstu Thresholding
Ostu Thresholding
 

Andere mochten auch

An Emerging Market Opportunity For Global Manufacturers & Suppliers Ofapparel...
An Emerging Market Opportunity For Global Manufacturers & Suppliers Ofapparel...An Emerging Market Opportunity For Global Manufacturers & Suppliers Ofapparel...
An Emerging Market Opportunity For Global Manufacturers & Suppliers Ofapparel...FNian
 
15기무적공감선발용_공감소개자료
15기무적공감선발용_공감소개자료15기무적공감선발용_공감소개자료
15기무적공감선발용_공감소개자료Seunghyuk Bang
 
Mafalda
MafaldaMafalda
MafaldaNeo22
 

Andere mochten auch (6)

An Emerging Market Opportunity For Global Manufacturers & Suppliers Ofapparel...
An Emerging Market Opportunity For Global Manufacturers & Suppliers Ofapparel...An Emerging Market Opportunity For Global Manufacturers & Suppliers Ofapparel...
An Emerging Market Opportunity For Global Manufacturers & Suppliers Ofapparel...
 
Shopeeg
ShopeegShopeeg
Shopeeg
 
Cmasummit08sms
Cmasummit08smsCmasummit08sms
Cmasummit08sms
 
15기무적공감선발용_공감소개자료
15기무적공감선발용_공감소개자료15기무적공감선발용_공감소개자료
15기무적공감선발용_공감소개자료
 
啞夫妻
啞夫妻啞夫妻
啞夫妻
 
Mafalda
MafaldaMafalda
Mafalda
 

Ähnlich wie 2007 112

Real-time 3D Object Pose Estimation and Tracking for Natural Landmark Based V...
Real-time 3D Object Pose Estimation and Tracking for Natural Landmark Based V...Real-time 3D Object Pose Estimation and Tracking for Natural Landmark Based V...
Real-time 3D Object Pose Estimation and Tracking for Natural Landmark Based V...c.choi
 
Humanoid Dual Arm Robot with Neural Learning of Skill Transferring Control
Humanoid Dual Arm Robot with Neural Learning of Skill Transferring ControlHumanoid Dual Arm Robot with Neural Learning of Skill Transferring Control
Humanoid Dual Arm Robot with Neural Learning of Skill Transferring ControlIRJET Journal
 
Motion planning and controlling algorithm for grasping and manipulating movin...
Motion planning and controlling algorithm for grasping and manipulating movin...Motion planning and controlling algorithm for grasping and manipulating movin...
Motion planning and controlling algorithm for grasping and manipulating movin...ijscai
 
Motion Planning and Controlling Algorithm for Grasping and Manipulating Movin...
Motion Planning and Controlling Algorithm for Grasping and Manipulating Movin...Motion Planning and Controlling Algorithm for Grasping and Manipulating Movin...
Motion Planning and Controlling Algorithm for Grasping and Manipulating Movin...ijscai
 
Wireless Vision based Real time Object Tracking System Using Template Matching
Wireless Vision based Real time Object Tracking System Using Template MatchingWireless Vision based Real time Object Tracking System Using Template Matching
Wireless Vision based Real time Object Tracking System Using Template MatchingIDES Editor
 
Human Action Recognition Based on Spacio-temporal features-Poster
Human Action Recognition Based on Spacio-temporal features-PosterHuman Action Recognition Based on Spacio-temporal features-Poster
Human Action Recognition Based on Spacio-temporal features-Posternikhilus85
 
Face recognition across non uniform motion blur, illumination, and pose
Face recognition across non uniform motion blur, illumination, and poseFace recognition across non uniform motion blur, illumination, and pose
Face recognition across non uniform motion blur, illumination, and posePvrtechnologies Nellore
 
In tech vision-based_obstacle_detection_module_for_a_wheeled_mobile_robot
In tech vision-based_obstacle_detection_module_for_a_wheeled_mobile_robotIn tech vision-based_obstacle_detection_module_for_a_wheeled_mobile_robot
In tech vision-based_obstacle_detection_module_for_a_wheeled_mobile_robotSudhakar Spartan
 
Mapping virtual and physical reality
Mapping virtual and physical realityMapping virtual and physical reality
Mapping virtual and physical realityChheang Vuthea
 
HUMAN ACTION RECOGNITION IN VIDEOS USING STABLE FEATURES
HUMAN ACTION RECOGNITION IN VIDEOS USING STABLE FEATURES HUMAN ACTION RECOGNITION IN VIDEOS USING STABLE FEATURES
HUMAN ACTION RECOGNITION IN VIDEOS USING STABLE FEATURES sipij
 
IRJET- Behavior Analysis from Videos using Motion based Feature Extraction
IRJET-  	  Behavior Analysis from Videos using Motion based Feature ExtractionIRJET-  	  Behavior Analysis from Videos using Motion based Feature Extraction
IRJET- Behavior Analysis from Videos using Motion based Feature ExtractionIRJET Journal
 
Feature Tracking of Objects in Underwater Video Sequences
Feature Tracking of Objects in Underwater Video SequencesFeature Tracking of Objects in Underwater Video Sequences
Feature Tracking of Objects in Underwater Video SequencesIDES Editor
 
Deep Learning for Structure-from-Motion (SfM)
Deep Learning for Structure-from-Motion (SfM)Deep Learning for Structure-from-Motion (SfM)
Deep Learning for Structure-from-Motion (SfM)PetteriTeikariPhD
 
Intelligent vision based snake robot
Intelligent vision based snake robotIntelligent vision based snake robot
Intelligent vision based snake roboteSAT Journals
 
Intelligent vision based snake robot
Intelligent vision based snake robotIntelligent vision based snake robot
Intelligent vision based snake roboteSAT Publishing House
 
Object Detection and Tracking AI Robot
Object Detection and Tracking AI RobotObject Detection and Tracking AI Robot
Object Detection and Tracking AI RobotIRJET Journal
 
MIT 6.870 - Template Matching and Histograms (Nicolas Pinto, MIT)
MIT 6.870 - Template Matching and Histograms (Nicolas Pinto, MIT)MIT 6.870 - Template Matching and Histograms (Nicolas Pinto, MIT)
MIT 6.870 - Template Matching and Histograms (Nicolas Pinto, MIT)npinto
 
IEEE CASE 2016 On Avoiding Moving Objects for Indoor Autonomous Quadrotors
IEEE CASE 2016 On Avoiding Moving Objects for Indoor Autonomous QuadrotorsIEEE CASE 2016 On Avoiding Moving Objects for Indoor Autonomous Quadrotors
IEEE CASE 2016 On Avoiding Moving Objects for Indoor Autonomous QuadrotorsPeter SHIN
 

Ähnlich wie 2007 112 (20)

Real-time 3D Object Pose Estimation and Tracking for Natural Landmark Based V...
Real-time 3D Object Pose Estimation and Tracking for Natural Landmark Based V...Real-time 3D Object Pose Estimation and Tracking for Natural Landmark Based V...
Real-time 3D Object Pose Estimation and Tracking for Natural Landmark Based V...
 
Humanoid Dual Arm Robot with Neural Learning of Skill Transferring Control
Humanoid Dual Arm Robot with Neural Learning of Skill Transferring ControlHumanoid Dual Arm Robot with Neural Learning of Skill Transferring Control
Humanoid Dual Arm Robot with Neural Learning of Skill Transferring Control
 
Motion planning and controlling algorithm for grasping and manipulating movin...
Motion planning and controlling algorithm for grasping and manipulating movin...Motion planning and controlling algorithm for grasping and manipulating movin...
Motion planning and controlling algorithm for grasping and manipulating movin...
 
Motion Planning and Controlling Algorithm for Grasping and Manipulating Movin...
Motion Planning and Controlling Algorithm for Grasping and Manipulating Movin...Motion Planning and Controlling Algorithm for Grasping and Manipulating Movin...
Motion Planning and Controlling Algorithm for Grasping and Manipulating Movin...
 
Wireless Vision based Real time Object Tracking System Using Template Matching
Wireless Vision based Real time Object Tracking System Using Template MatchingWireless Vision based Real time Object Tracking System Using Template Matching
Wireless Vision based Real time Object Tracking System Using Template Matching
 
Hq2513761382
Hq2513761382Hq2513761382
Hq2513761382
 
Hq2513761382
Hq2513761382Hq2513761382
Hq2513761382
 
Human Action Recognition Based on Spacio-temporal features-Poster
Human Action Recognition Based on Spacio-temporal features-PosterHuman Action Recognition Based on Spacio-temporal features-Poster
Human Action Recognition Based on Spacio-temporal features-Poster
 
Face recognition across non uniform motion blur, illumination, and pose
Face recognition across non uniform motion blur, illumination, and poseFace recognition across non uniform motion blur, illumination, and pose
Face recognition across non uniform motion blur, illumination, and pose
 
In tech vision-based_obstacle_detection_module_for_a_wheeled_mobile_robot
In tech vision-based_obstacle_detection_module_for_a_wheeled_mobile_robotIn tech vision-based_obstacle_detection_module_for_a_wheeled_mobile_robot
In tech vision-based_obstacle_detection_module_for_a_wheeled_mobile_robot
 
Mapping virtual and physical reality
Mapping virtual and physical realityMapping virtual and physical reality
Mapping virtual and physical reality
 
HUMAN ACTION RECOGNITION IN VIDEOS USING STABLE FEATURES
HUMAN ACTION RECOGNITION IN VIDEOS USING STABLE FEATURES HUMAN ACTION RECOGNITION IN VIDEOS USING STABLE FEATURES
HUMAN ACTION RECOGNITION IN VIDEOS USING STABLE FEATURES
 
IRJET- Behavior Analysis from Videos using Motion based Feature Extraction
IRJET-  	  Behavior Analysis from Videos using Motion based Feature ExtractionIRJET-  	  Behavior Analysis from Videos using Motion based Feature Extraction
IRJET- Behavior Analysis from Videos using Motion based Feature Extraction
 
Feature Tracking of Objects in Underwater Video Sequences
Feature Tracking of Objects in Underwater Video SequencesFeature Tracking of Objects in Underwater Video Sequences
Feature Tracking of Objects in Underwater Video Sequences
 
Deep Learning for Structure-from-Motion (SfM)
Deep Learning for Structure-from-Motion (SfM)Deep Learning for Structure-from-Motion (SfM)
Deep Learning for Structure-from-Motion (SfM)
 
Intelligent vision based snake robot
Intelligent vision based snake robotIntelligent vision based snake robot
Intelligent vision based snake robot
 
Intelligent vision based snake robot
Intelligent vision based snake robotIntelligent vision based snake robot
Intelligent vision based snake robot
 
Object Detection and Tracking AI Robot
Object Detection and Tracking AI RobotObject Detection and Tracking AI Robot
Object Detection and Tracking AI Robot
 
MIT 6.870 - Template Matching and Histograms (Nicolas Pinto, MIT)
MIT 6.870 - Template Matching and Histograms (Nicolas Pinto, MIT)MIT 6.870 - Template Matching and Histograms (Nicolas Pinto, MIT)
MIT 6.870 - Template Matching and Histograms (Nicolas Pinto, MIT)
 
IEEE CASE 2016 On Avoiding Moving Objects for Indoor Autonomous Quadrotors
IEEE CASE 2016 On Avoiding Moving Objects for Indoor Autonomous QuadrotorsIEEE CASE 2016 On Avoiding Moving Objects for Indoor Autonomous Quadrotors
IEEE CASE 2016 On Avoiding Moving Objects for Indoor Autonomous Quadrotors
 

Kürzlich hochgeladen

AMERICAN LANGUAGE HUB_Level2_Student'sBook_Answerkey.pdf
AMERICAN LANGUAGE HUB_Level2_Student'sBook_Answerkey.pdfAMERICAN LANGUAGE HUB_Level2_Student'sBook_Answerkey.pdf
AMERICAN LANGUAGE HUB_Level2_Student'sBook_Answerkey.pdfphamnguyenenglishnb
 
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptx
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptxECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptx
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptxiammrhaywood
 
Judging the Relevance and worth of ideas part 2.pptx
Judging the Relevance  and worth of ideas part 2.pptxJudging the Relevance  and worth of ideas part 2.pptx
Judging the Relevance and worth of ideas part 2.pptxSherlyMaeNeri
 
Visit to a blind student's school🧑‍🦯🧑‍🦯(community medicine)
Visit to a blind student's school🧑‍🦯🧑‍🦯(community medicine)Visit to a blind student's school🧑‍🦯🧑‍🦯(community medicine)
Visit to a blind student's school🧑‍🦯🧑‍🦯(community medicine)lakshayb543
 
GRADE 4 - SUMMATIVE TEST QUARTER 4 ALL SUBJECTS
GRADE 4 - SUMMATIVE TEST QUARTER 4 ALL SUBJECTSGRADE 4 - SUMMATIVE TEST QUARTER 4 ALL SUBJECTS
GRADE 4 - SUMMATIVE TEST QUARTER 4 ALL SUBJECTSJoshuaGantuangco2
 
ACC 2024 Chronicles. Cardiology. Exam.pdf
ACC 2024 Chronicles. Cardiology. Exam.pdfACC 2024 Chronicles. Cardiology. Exam.pdf
ACC 2024 Chronicles. Cardiology. Exam.pdfSpandanaRallapalli
 
Student Profile Sample - We help schools to connect the data they have, with ...
Student Profile Sample - We help schools to connect the data they have, with ...Student Profile Sample - We help schools to connect the data they have, with ...
Student Profile Sample - We help schools to connect the data they have, with ...Seán Kennedy
 
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITYISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITYKayeClaireEstoconing
 
Influencing policy (training slides from Fast Track Impact)
Influencing policy (training slides from Fast Track Impact)Influencing policy (training slides from Fast Track Impact)
Influencing policy (training slides from Fast Track Impact)Mark Reed
 
ANG SEKTOR NG agrikultura.pptx QUARTER 4
ANG SEKTOR NG agrikultura.pptx QUARTER 4ANG SEKTOR NG agrikultura.pptx QUARTER 4
ANG SEKTOR NG agrikultura.pptx QUARTER 4MiaBumagat1
 
How to Add Barcode on PDF Report in Odoo 17
How to Add Barcode on PDF Report in Odoo 17How to Add Barcode on PDF Report in Odoo 17
How to Add Barcode on PDF Report in Odoo 17Celine George
 
How to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERPHow to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERPCeline George
 
Concurrency Control in Database Management system
Concurrency Control in Database Management systemConcurrency Control in Database Management system
Concurrency Control in Database Management systemChristalin Nelson
 
4.18.24 Movement Legacies, Reflection, and Review.pptx
4.18.24 Movement Legacies, Reflection, and Review.pptx4.18.24 Movement Legacies, Reflection, and Review.pptx
4.18.24 Movement Legacies, Reflection, and Review.pptxmary850239
 
Difference Between Search & Browse Methods in Odoo 17
Difference Between Search & Browse Methods in Odoo 17Difference Between Search & Browse Methods in Odoo 17
Difference Between Search & Browse Methods in Odoo 17Celine George
 
Transaction Management in Database Management System
Transaction Management in Database Management SystemTransaction Management in Database Management System
Transaction Management in Database Management SystemChristalin Nelson
 
Field Attribute Index Feature in Odoo 17
Field Attribute Index Feature in Odoo 17Field Attribute Index Feature in Odoo 17
Field Attribute Index Feature in Odoo 17Celine George
 

Kürzlich hochgeladen (20)

LEFT_ON_C'N_ PRELIMS_EL_DORADO_2024.pptx
LEFT_ON_C'N_ PRELIMS_EL_DORADO_2024.pptxLEFT_ON_C'N_ PRELIMS_EL_DORADO_2024.pptx
LEFT_ON_C'N_ PRELIMS_EL_DORADO_2024.pptx
 
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
 
AMERICAN LANGUAGE HUB_Level2_Student'sBook_Answerkey.pdf
AMERICAN LANGUAGE HUB_Level2_Student'sBook_Answerkey.pdfAMERICAN LANGUAGE HUB_Level2_Student'sBook_Answerkey.pdf
AMERICAN LANGUAGE HUB_Level2_Student'sBook_Answerkey.pdf
 
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptx
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptxECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptx
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptx
 
Judging the Relevance and worth of ideas part 2.pptx
Judging the Relevance  and worth of ideas part 2.pptxJudging the Relevance  and worth of ideas part 2.pptx
Judging the Relevance and worth of ideas part 2.pptx
 
Visit to a blind student's school🧑‍🦯🧑‍🦯(community medicine)
Visit to a blind student's school🧑‍🦯🧑‍🦯(community medicine)Visit to a blind student's school🧑‍🦯🧑‍🦯(community medicine)
Visit to a blind student's school🧑‍🦯🧑‍🦯(community medicine)
 
GRADE 4 - SUMMATIVE TEST QUARTER 4 ALL SUBJECTS
GRADE 4 - SUMMATIVE TEST QUARTER 4 ALL SUBJECTSGRADE 4 - SUMMATIVE TEST QUARTER 4 ALL SUBJECTS
GRADE 4 - SUMMATIVE TEST QUARTER 4 ALL SUBJECTS
 
ACC 2024 Chronicles. Cardiology. Exam.pdf
ACC 2024 Chronicles. Cardiology. Exam.pdfACC 2024 Chronicles. Cardiology. Exam.pdf
ACC 2024 Chronicles. Cardiology. Exam.pdf
 
Student Profile Sample - We help schools to connect the data they have, with ...
Student Profile Sample - We help schools to connect the data they have, with ...Student Profile Sample - We help schools to connect the data they have, with ...
Student Profile Sample - We help schools to connect the data they have, with ...
 
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITYISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
 
Influencing policy (training slides from Fast Track Impact)
Influencing policy (training slides from Fast Track Impact)Influencing policy (training slides from Fast Track Impact)
Influencing policy (training slides from Fast Track Impact)
 
ANG SEKTOR NG agrikultura.pptx QUARTER 4
ANG SEKTOR NG agrikultura.pptx QUARTER 4ANG SEKTOR NG agrikultura.pptx QUARTER 4
ANG SEKTOR NG agrikultura.pptx QUARTER 4
 
How to Add Barcode on PDF Report in Odoo 17
How to Add Barcode on PDF Report in Odoo 17How to Add Barcode on PDF Report in Odoo 17
How to Add Barcode on PDF Report in Odoo 17
 
How to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERPHow to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERP
 
YOUVE GOT EMAIL_FINALS_EL_DORADO_2024.pptx
YOUVE GOT EMAIL_FINALS_EL_DORADO_2024.pptxYOUVE GOT EMAIL_FINALS_EL_DORADO_2024.pptx
YOUVE GOT EMAIL_FINALS_EL_DORADO_2024.pptx
 
Concurrency Control in Database Management system
Concurrency Control in Database Management systemConcurrency Control in Database Management system
Concurrency Control in Database Management system
 
4.18.24 Movement Legacies, Reflection, and Review.pptx
4.18.24 Movement Legacies, Reflection, and Review.pptx4.18.24 Movement Legacies, Reflection, and Review.pptx
4.18.24 Movement Legacies, Reflection, and Review.pptx
 
Difference Between Search & Browse Methods in Odoo 17
Difference Between Search & Browse Methods in Odoo 17Difference Between Search & Browse Methods in Odoo 17
Difference Between Search & Browse Methods in Odoo 17
 
Transaction Management in Database Management System
Transaction Management in Database Management SystemTransaction Management in Database Management System
Transaction Management in Database Management System
 
Field Attribute Index Feature in Odoo 17
Field Attribute Index Feature in Odoo 17Field Attribute Index Feature in Odoo 17
Field Attribute Index Feature in Odoo 17
 

2007 112

  • 1. Visual Servoing in Presence of Non-Rigid Motion D Santosh Kumar, C V Jawahar Center for Visual Information Technology International Institute of Information Technology Hyderabad 500032, India {santosh@students.,jawahar@}iiit.ac.in Abstract are non-rigid [2] in nature. For example, the articulated hu- man motion; the deforming nature of the beating heart, fly- Most robotic vision algorithms have been proposed by ing birds and moving aquatic animals; the near-rigid nature envisaging robots operating in industrial environments, of the human face etc. If robots are to be used in pres- where the world is assumed to be static and rigid. These ence of such objects, we need servoing algorithms that are algorithms cannot be used in environments where the as- adaptable to non-rigidity. The work done by [7] and [10] sumption of a rigid world does not hold. In this paper, we deals with only moving rigid object visual servoing. Re- study the problem of visual servoing in presence of non- cently in [4], Cavusoglu et al. proposed a method to handle rigid objects and analyze the design of servoing strategies non-rigidness by canceling the relative motion between the needed to perform optimally even in unconventional envi- non-rigid object (heart) and the camera by using a model ronments. We also propose a servoing algorithm that is ro- predictive controller so as to facilitate easy manipulation of bust to non-rigidity. The algorithm extracts invariant fea- the object. In this paper, we address the issue of position- tures of the non-rigid object and uses these features in the ing of the end-effector of a robot with respect to a non-rigid servoing process. We validate the technique with experi- object. Often, the positioning task is a pre-requisite for the ments and demonstrate the applicability. manipulation task. 1. Introduction As a non-rigid object undergoes a persistent change in its pose, it is preferable to extract pose-invariant features Visual servoing [8] describes a broad class of navigation of the object and use them in the servoing process. As a problems in which a robot is positioned with respect to a first attempt towards analyzing this problem, we investigate target using computer vision as the primary feedback sen- the various issues that need to be considered while solving it sor. It is achieved by processing the visual feedback (2D and also propose one simple yet an effective visual servoing features or 3D pose or both) and minimizing an appropriate algorithm to handle deformations. The algorithm is based error function. on a not so strict assumption that any non-rigid motion can Several algorithms [1, 3, 7, 10] have been proposed to be approximated to a generalized quasi-periodic motion. perform the task of visual servoing; however, much of the And for a quasi-periodic motion, gross measures such as research until now focuses only on rigid objects. As far as the moving average are sufficient to reveal its general sta- we are aware, the problem of visual servoing in presence of tionary characteristics, provided the change in pose can be non-rigid objects has not been addressed in literature. The compensated. By using the recent results of multiple-view challenge with non-rigidity is that the motion planned based geometry, we efficiently extract such static features of the on the image-clues extracted at current time instant might non-rigid object, which enable us to use the conventional not be valid in the next time instant as the object under- visual servoing techniques even in this case. goes a change in its pose. This problem is different from that of servoing with respect to a moving object, where all 2. Background the features that are being tracked undergo the same or pre- dictable transformations in the image space. Whereas, for a deforming object, each feature point transforms differently In image-based visual servo control, the error signal e and thus generates different control signals at the same time is defined directly in terms of image feature parameters as instant, thereby destabilizing the motion planning process. e(S) = S − S ∗ , where S and S ∗ are the current and the Many of the living forms that we come across everyday desired feature parameters respectively. By differentiating The 18th International Conference on Pattern Recognition (ICPR'06) 0-7695-2521-0/06 $20.00 © 2006
  • 2. this error function with respect to time, we get: Any servoing algorithm that computes an optimal path for the robot to traverse from the initial position to the de- de dS ∂S dr sired position, can be divided into two basic steps. After de- = =( ) = LS V, (1) dt dt ∂r dt ciding on an appropriate feature to represent the non-rigid object, the following steps need to be repeated until the de- where V = (v T , ω T )T is the camera screw, r is the cam- sired position is reached. era pose and LS is the interaction matrix [8]. It relates the motion of the features in the image space to the camera • Compute the error estimate between the current and motion in the Cartesian space. The main objective of the the desired feature vector i.e., S − S ∗ visual servoing process is to minimize the error function • Using a proportional control law (see Section 2), move e(S). For exponential convergence, we use a proportional in the direction that minimizes the above error control law i.e., de(S) = −λe(S). By substituting this in dt It is crucial to have a good feature vector S to represent (1), the required velocity of the camera can be computed as the object state and thereby to perform the servoing task. V = −λL+ e(S), where L+ is the pseudo-inverse of the in- S S However for non-rigid objects as the object pose contin- teraction matrix and λ is a scale factor. In case of a moving uously changes, we cannot obtain a feature vector from a object, the expression for V will get modified as (see [8]) single image that characterizes its state as it might lead to an unstable end-effector trajectory. For instance, a surgical ∂e V = −λL+ e(S) − L+ S S , (2) robot positioning itself only using the current image of the ∂t heart, will oscillate corresponding to the systole and dias- tole states of the heart during its motion. Moreover, if only where ∂e represents the object motion model. The image- ∂t a single image is used to describe the desired position S ∗ , based visual servoing algorithm has gained prominence in it will lead to oscillations of the manipulator even after the literature due to its robustness to depth estimation errors. goal position is reached. One possible solution to circum- Note that the above analysis assumes the objects to be vent this problem would be to use the sequence of images rigid. Though non-rigid objects have not been studied in the representing all states of the object to perform the servo- visual servoing framework, they have been well analyzed in ing task. This will generate better control signals as it takes computer vision in the context of non-rigid tracking and 3D into account the object deformations. However, a more effi- reconstruction problems. Most of these algorithms employ cient solution could be to extract gross features from those models like the linear subspace model, appearance mani- sequence of images by exploiting the stationary signal prop- fold [11], kinematic chains [5, 9] etc in order to maintain erty of the quasi-periodic object deformations. a representation of the non-rigid object and also to infer its At each pose, a sequence of images (I1 , I2 , ..., In ) are internal motion model. Such models allow prediction of the acquired by the camera, capturing the deformations of the future states of the object, given its current state. Prediction non-rigid object (where n depends on the periodicity of the is necessary in case of manipulation tasks where the end- deformations). Then, m point features are extracted from effector needs to interact with the deforming object. For the each image Ii to obtain a feature vector Fit = [f1 f2 ...fm ]T , positioning task, only an approximate pose of the object is where each feature fk corresponds to the projection of a 3D sufficient to align the camera with respect to the object. Any point on the surface of the non-rigid object at time instant error resulting due to this approximation can be corrected t. The invariant feature Gt corresponding to the image se- by using the visual feedback. quence is computed as a function of all the feature vectors. Gt = f (F1 , F2 , ...., Fn ) t t t (3) 3. Visual Servoing in Presence of Deformations However, the camera can use its previous knowledge and Our problem of visual servoing is to position the end- predict the next gross appearance of the deforming object. effector of the robot with respect to a deforming object. Thus at each position, only a single image is captured by Several strategies could be conceived to solve this problem. the camera. The rest of the deformations of the object are One possible solution is to use clues from object appearance obtained by projecting the past images into the new pose. to aid the positioning task. Alternatively, one could gain a Hence, the gross feature at time t can be predictively calcu- stable representation of the object from its motion charac- lated as a function of the features extracted from the current teristics, which can be used to calibrate and move the end- image and the previous images as effector. Another solution is to extract some pose-invariant Gt = f (g(F t−n , ...., F t−2 , F t−1 ), F t ), (4) features of the non-rigid object and use these features in the servoing process. Here we adopt a relatively simple strategy where F t−i denotes the feature vector obtained from to solve this problem. the image captured at camera pose rt−i , F t denotes the The 18th International Conference on Pattern Recognition (ICPR'06) 0-7695-2521-0/06 $20.00 © 2006
  • 3. feature vector extracted from the current image and g(·) We demonstrate our results using two kinds of features is a function that projects the past image features into the namely the mean and the eigen feature. Such features have current view as explained below. been used in computer vision to capture statistical regulari- ties and redundancies shared by a set of images. In the first Projection of Images Given two camera matrices M1 method, all the feature vectors Fi , i = 1, 2, ..., n obtained and M2 and an image I1 taken from the pose corresponding from the n images are averaged to get the mean feature vec- to M1 , the task is to obtain the image I2 as seen from the tor M . In this context, the mean operator corresponds to pose corresponding to M2 using a partial 3D model of the the function f (·) and M represents the invariant feature Gt . world that is assumed to be available in the image-based This feature is used as the input S to the servoing algorithm. visual servoing process [8]. Let M1 = K1 [R1 t1 ] and As explained in Section 3, the new mean is predictively cal- M2 = K2 [R2 t2 ] where Ki denotes the intrinsic para- culated using the past image features. meter matrix and [Ri ti ] denotes the pose (rotation and In the second method, the eigen feature of the sequence translation). Let p1 represent a point in I1 and Z1 be an of image features is computed and used as the input feature estimate of the depth of the corresponding 3D point P1 in to the servoing process. This feature indicates the direction the camera frame given by the 3D model. The coordinates in which the point features undergo maximum variation in −1 of point P1 are given by P1 = K1 Z1 p1 (see [6]). the image space. In estimating the camera screw from (2), Now to obtain the 3D coordinate of this point in the other the line Jacobian is used to compute V since the eigen fea- camera frame, we have ture can be conceived of as a line segment in the image. In comparison to classical image-based visual servoing, P2 = [R2 t2 ][R1 t1 ]−1 P1 . (5) the use of above features generated a stable camera trajec- tory. Fig. 1a shows the velocity screw obtained in the case Finally, the image coordinates of the point with respect to of the classical servoing algorithm. As only a single image 1 the second camera are given by p2 = Z2 K2 P2 . Thus by is used, the generated control instructions resulted in an os- projecting the past image features into the new view, all cillatory camera trajectory (see the jagged behavior of the the deformations of the object as viewed from the current velocity screw in Fig. 1a as compared to that of Fig. 1b), camera pose can be obtained. These features are used to affecting the stability of the robot. Also the system set- compute the representative feature of the non-rigid object. tled in a critically stable condition as the oscillations of the end-effector continued even after the desired position was 4. Results and Discussions reached. In our experiments, we consider a set of four 3D points Classical Image Based Visual Servoing Servoing using Eigen Feature 1 2 Trans. Vel (m/sec) Trans. Vel. (m/sec) vx vx vy on the surface of a non-rigid object being observed by a 0 vy vz 0 vz perspective camera. The projection of the points onto the −1 −2 image are considered as features. The deformations of the −2 0 2 4 6 8 10 Time (in seconds) 12 0 2 4 6 Time (In Seconds) 8 1 Rot. Vel (rad/sec) 0.6 non-rigid object are simulated by using a sinusoidal motion Rot. Vel. (rad/sec) wx wx 0.4 wy wy wz wz perturbed by a white gaussian noise. This makes the object 0.2 0 0 undergo quasi-periodic deformations. The task of position- −0.2 −1 0 2 4 6 8 10 12 0 2 4 6 8 ing a camera with respect to the non-rigid object is studied. Time (in seconds) a Time (In Seconds) b The basic implementation of the proposed algorithm can be summarized into the following steps. Figure 1. Velocity Screw (a) Oscillatory be- havior (features only from single image are 1. In an offline step, acquire images from the desired fi- used) (b) Smooth convergence nal position. Extract feature S ∗ from these images us- ing (3) We also tested the use of image sequence as an input 2. Acquire a sequence of images from the current camera feature to the servoing algorithm. In this case, the feature pose and calculate S, again using (3) S was the entire sequence of features extracted from each 3. Compute the velocity screw V using (2) image in the sequence i.e., S = [F1 , F2 , ..., Fn ]T . As a 4. Using V , move the end-effector to the new pose and large set of features are considered, minimizing the error capture a new image difference between the current and the desired feature 5. Project the features in the old images into the new pose vector consumed more time and led to slower convergence as discussed in the latter half of Section 3 of the error function (compare the error norm and the time 6. Using the features extracted from the old images and taken in Fig. 2a to that of Fig. 2b). the current image, extract S using (4) 7. Repeat steps 3 to 6 until e(S) < θ Stability and Robustness We empirically verify the The 18th International Conference on Pattern Recognition (ICPR'06) 0-7695-2521-0/06 $20.00 © 2006
  • 4. Convergence of Error Norm Error Convergence using features 200 1400 from Image Sequence 5. Conclusion Eigen feature Mean feature 1200 150 Existing servoing schemes cannot adapt to the non- Error(in pixels) 1000 Error(in pixels) 100 800 rigidity present in domestic environments. In this paper, we 600 studied the issues involved in designing servoing strategies 50 400 200 with optimal performance in presence of non-rigid objects. 0 0 2 4 6 8 0 We also proposed a novel image-based visual servoing al- Time(in seconds) a 0 2 4 6 Time(in seconds) 8 10 b gorithm, which efficiently extracted gross features from se- quence of images and used them in the servoing process. Figure 2. Error norm in case of (a) Eigen and This algorithm generated a stable camera trajectory unlike mean feature (b) Image sequence the classical servoing algorithms. In future, we plan to im- robustness of our approach to calibration errors. Often prove the algorithm by considering other invariant features due to system calibration errors, the obtained image which better characterize the non-rigid object motion. measurements are corrupted. The current image will be corrupted by the error in the current camera pose whereas References the projected images will be corrupted by the error in the previous camera poses. This leads to the incorrect compu- [1] J. Adachi and J. Sato. Uncalibrated visual servoing from tation of the error function as e (S) = e(S) + Δe where projective reconstruction of control values. International e(S) is as defined in Section 2 and Δe is the measurement Conference on Pattern Recognition, 4:297–300, August error obtained due to the corresponding error in pose. As 2004. [2] J. K. Aggarwal, Q. Cai, W. Liao, and B. Sabata. Non- a result, the velocity screw changes to V = V + ΔV rigid motion analysis: articulated and elastic motion. Com- where V is as given in (2) and ΔV is the velocity command puter Vision and Image Understanding, 70(2):142–156, generated to compensate the corresponding measurement May 1998. error Δe. This ΔV accommodates the uncertainty in pose [3] J. P. Barreto, J. Batista, and H. Araujo. Model predictive and thus ensures the convergence of error to zero. The control to improve visual control of motion: Applications in stability of the algorithm was tested by introducing noise active tracking of moving targets. International Conference into the camera extrinsic parameters. The noise level was on Pattern Recognition, 4:4732–4735, September 2000. [4] M. C. Cavusoglu, J. Rotella, W. S. Newman, S. Choi, varied and the performance of the algorithm was studied. J. Ustin, and S. S. Sastry. Control algorithms for active rel- In Fig. 3a, it is observed that in spite of the calibration ative motion cancelling for robotic assisted off-pump coro- errors (a 10% gaussain noise was added), the convergence nary artery bypass graft surgery. International Conference of velocity screw is achieved. This supports our analysis on Advanced Robotics, pages 431–436, July 2005. that uncertainty in pose is corrected by the visual feedback [5] F. Chaumette, E. Marchand, and A. Comport. Object- rather than getting propagated due to the prediction step. based visual 3d tracking of articulated objects via kinematic sets. IEEE Workshop on Articulated and Non-Rigid Motion, 2 Velocity Screw with calibration errors Velocity Screw with depth estimation errors 2 CVPRW, 1:2–9, June 2004. Trans. Vel (m/sec) Trans. Vel (m/sec) vx vy vz vx vy [6] R. Hartley and A. Zisserman. Multiple view geometry in vz 0 0 computer vision. Cambridge University Press, 2003. −2 [7] K. Hashimoto, K. Nagahama, T. Noritsugu, and M. Taka- −2 0 2 4 6 8 0 2 4 6 8 Time (In Seconds) Time (in Seconds) iawa. Visual servoing based on object motion estimation. 1 Rot. Vel (rad/sec) 1 Rot. Vel (rad/sec) 0.5 wx wy wx wy IEEE/RSJ International Conference on Intelligent Robots wz 0.5 wz 0 0 and Systems, 1:245–250, October 2000. −0.5 [8] S. A. Hutchinson, G. D. Hager, and P. I. Corke. A tutorial −0.5 0 2 4 6 8 0 2 4 6 8 Time (In Seconds) a Time (in Seconds) b on visual servo control. IEEE Transactions on Robotics and Automation, 12(5):651–670, October 1996. Figure 3. Convergence of velocity screw ob- [9] T. Kanade, D. D. Morris, and J. Rehg. Ambiguities in vi- served even in presence of errors (noise) sual tracking of articulated objects using two- and three- dimensional models. International Journal of Robotics Re- One of the crucial steps in the algorithm is the predic- search, 22(6):393–418, June 2003. tion step that assumes the depth of the feature points. The [10] E. Malis and S. Benhimane. A unified approach to visual robustness of the algorithm to depth estimation errors was tracking and servoing. Robotics and Autonomous Systems, tested by introducing noise into the depth values. Fig. 3b 52(1):39–52, July 2005. [11] A. Shashua, A. Levin, and S. Avidan. Manifold pursuit: shows the velocity screw obtained when the estimated depth A new approach to appearance based recognition. Interna- was corrupted by 15% gaussian noise. This experiment tional Conference of Pattern Recognition, 3:590–594, Au- demonstrated that minor error in the depth values can be gust 2002. accommodated by our algorithm. The 18th International Conference on Pattern Recognition (ICPR'06) 0-7695-2521-0/06 $20.00 © 2006