SlideShare ist ein Scribd-Unternehmen logo
1 von 10
Downloaden Sie, um offline zu lesen
ISSN: 2278 – 1323
            International Journal of Advanced Research in Computer Engineering & Technology
                                                                Volume 1, Issue 4, June 2012


      Intelligent System for detecting, Modeling,
     Classification of human behavior using image
       processing, machine vision and OpenCV
                       Niraj .B. Gadhe,              Prof: Dr. B .K . Lande,           Prof: Dr. B.B.Meshram

Abstract: Surveillance Cameras has proven to be a key                detect malicious activities. It would be so easy if
factor in enhancing the public security in many                      system on its own recognizes some events as
countries around the world . In spite of advancements                malicious and extract only those frames which are
in image processing and machine vision techniques very               providing suspicious activity. The endless hours of
less is applied to actual implementation of such
surveillance systems. Traditional methods wherein
                                                                     watching the video can be avoided. Moreover
stand alone cameras provide just live feed to the                    assistance provided by system during live feed of
observer are still deployed at most of the places but the            surveillance may prove beneficial as it can alert the
need of today’s world is to add some intelligence to this            observer of a possible danger and drag his attention
systems. In this paper we will be focusing on important              to a particular activity which he might have neglected
features to build an intelligent surveillance system and             in absence of intelligent system.
its architecture using computer vision libraries of                        In the light of such weak security infrastructure
opencv. Background Subtraction , segmentation,                       we propose a method to develop an intelligent system
tracking and human behavior analysis are step by step                which not only will capture and feedback the video at
implemented to build such a system. A detailed analysis
of background subtraction methods and their
                                                                     real time but also model human behavior.[2] We
comparison is provided. Lastly we discuss the future                 think of detecting , modeling and classifying human
work which can lead to human behavior analysis like                  behavior using image processing to detect malicious
crowd counting, intrusion detection, loitering , crowd               behavior. Further we may maintain a database of
movement , vehicle tracking and all.                                 suspects found over time and use face detection [ 24]
                                                                     and recognition to identify the probable matches for
Index Terms:         surveillance systems, opencv,                   the suspects.
background subtraction, segmentation, machine                          There is lot of work done in Image processing as
vision, computer vision.                                             well as machine vision techniques. A much of it is
                                                                     local and developed for particular domain. The
I.         INTRODUCTION                                              object recognition frame works consist of
                                                                     algorithms like Pattern matching, feature
        In traditional systems[ 1] a network of passive              matching , boundary detection, knowledge based
cameras able of only recording the moving images                     repository systems.[18 ] The behavioral model
and converting it to video files are deployed at the                 uses heuristics and classification models like
site of prime importance. In case of accidents,                      Motion based classification , Shape based
terrorist attacks , theft, malicious activities such video           classification, color based classification.[19][18 ]
data comes to be very important for analysis. But if                 [17][14] Once the object is classified and detected
traditional system is used we require to watch the                   it can be tracked using algorithms like mean shift
video data over and over for hours together because                  algorithm.[10 ][22] Human behavior is studied
collected data is very large and we don’t know the                   under Bayesian Model and Markov Models.[12 ]
exact time when the malicious activity took place.                   [8] Face recognition and detection done via Eigen
Such manual method is prone to human failure as                      faces, SIFT based, Edge based, Binary pattern face
human may miss small details like intrusion[2] or                    recognitions, viola jones face detection are famous.[
may neglect some activity. So its necessary to build                 21][24]. Much of the work today is focused on
an intelligent system that would not only provide                    detecting occluded object.[4]
data as per traditional system but also assist us to                    The rest of the paper is organized as follows:
                                                                     Section 2 discusses our architecture of system and
Niraj Babasaheb Gadhe, Dept of computers- NIMS, VJTI, (e-mail:       features we decided to incorporate in it. Section 3
nirajgadhe@gmail.com). Mumbai.India,
Prof: Dr. B.K.Lande, Computer engg dept, Shah and anchor             presents a detailed analysis of famous background
college of engg, University of Mumbai, Mumbai, India ,               subtraction methods. . A detailed review of present
Prof: Dr. B.B.Meshram, Computer engg dept, VJTI, Mumbai,             image processing algorithms and their advantages
India ,                                                              are discussed . Section 4 we study segmentation of




                                                                                                                       590
                                                All Rights Reserved © 2012 IJARCET
ISSN: 2278 – 1323
          International Journal of Advanced Research in Computer Engineering & Technology
                                                              Volume 1, Issue 4, June 2012

object and motion tracking of object here. Section 5           some human behaviors of malicious intent. We will
discusses the behavior which can be detected using             take a simple real world problem like Trespassing or
the system. Lastly we conclude with the future work.           intrusion, loitering, bag being left by a human and
                                                               flee, crowd analysis. [2]
II.      BUILDING AN INTELLIGENT                               D) Capturing frames containing suspicious
         SURVEILLANCE SYSTEM                                        activities module:
                                                               Capturing frames containing probable suspicious
We are to develop a system based on image and                  behavior, classifying them and maintaining it for
video processing algorithms that will help us detect           further use and alarming the system . Face detection
and classify human behavior. As detecting a human              and storing face data can be done.[24][21] Further
behavior from a scene we need to follow some                   such data can be used during investigation of crimes.
sequential steps in our processing of the video. These         According to our architecture(FIG 1 ) the working of
sequential steps can be generalized and explained in           our system can be summarized as follows
brief as follows . [1]
                                                               1.   We first acquire frames from a camera or video
A) Frame detection module.                                          file.
                                                               2.   We first acquire frames from a camera or video
This routine will detect each frame at real time from               file.
a video sequence and present its status in form of             3.   Now we pass on these frames to background
matrix, its frame no, pixels, color content and sample              subtraction module. For convenience we are
rate, etc.[23] Now we can treat each captured frame                 using the frame differencing for background
as a image data type and apply image processing on                  subtraction.
it.                                                            4.   We first acquire frames from a camera or video
                                                                    file.
B) Background Removal and segmentation                         5.   Now we pass on these frames to background
     module:                                                        subtraction module. For convenience we are
This will detect and remove background features to                  using the frame differencing for background
detect the object in the scene . Algorithms like frame              subtraction.
differencing, averaging background method and
GMM based will be a tool for such removal.[22] We
are interested in analyzing human behavior.[2] .
Different types of background subtracting
algorithms[ 5] like Gaussian mixture model[6][11],
adaptive Gaussian mixture model[4][20] and many
more are available. Then we apply segmentation
algorithms to separate moving objects from
background. Segmentation can be carried out by
Mean shift method [10], Bayesian method [8],
wrapper based approach [7], etc.

C) Detecting human and their behavior module:

Once we get a human object in frame we can track
the motion of such object in subsequent frames. We
achieved this by drawing contour around the
foreground object using connected component
technique [13 ] of opencv.[22 ][23] Then we draw a
bounding rectangle around moving object. Then we
can use this rectangle parameters to detect the                FIG: 1 Architecture of intelligent surveillance system.
position of object in the scene. Ot her techniques like
event detection using rejection based approach[3] ,            6.   We first acquire frames from a camera or video
kalman filters[9], optical flow[22], object recognition             file.
methods[12][13][14], detection of salient object[15],          7.   Now we pass on this frames to background
shape filters[16][17][19], repository based object                  subtraction module. For convenience we are
recognition[18] can be used.                                        using the frame differencing for background
We provide some parameters to system to detect                      subtraction.



                                                                                                                         591
                                          All Rights Reserved © 2012 IJARCET
ISSN: 2278 – 1323
            International Journal of Advanced Research in Computer Engineering & Technology
                                                                Volume 1, Issue 4, June 2012

8.  Background subtraction separates background                i.   Grab current frame. Save it in an image and
    and gives us the frame with foreground object.                  name it “current frame” .
9. After background subtraction we detect object               ii. Take absolute difference between current and
    and draw a counter using connected point                        background frame and save it in a “foreground
    concept and opencv modules.                                     frame”. Apply threshold and erosion if needed.
10. After this we draw a bounding rectangle around             iii. Display the foreground frame. Move to next
    the counter and save its center .                               frame.
11. Now we track motion of moving object using the             3. Stop when all frames are over and release all the
    coordinates of the centre.                                      data types to avoid stack overflow.
12. Once we track the motion we can analyse the
    behavior of the person in the scene and study his          Experimental results:
    interaction with the environment.
                                                               A. Indoor scene with minimum fluctuations:
III.       BACKGROUND             SUBTRACTION
           ANALYSIS


This section focuses on how to isolate objects or parts
of objects from the rest of the image.[22] We want to
isolate those events and to be able to ignore the
endless hours when nothing is changing. We start               .FIG 2 Without threshold gray level image
with separating foreground objects from learned
background scenes. We will be dealing with three
types of background subtraction methods for our
analysis. We will implement frame differencing,
averaging background method and mixture of
Gaussian( MOG) model and study the results for
different cases.
                                                               FIG 3 After applying threshold : binary image TH=80
Absolute differencing        method      or    Frame
differencing:

Basic idea here is to subtract one frame from another
subsequent frames and label any difference that is big
enough ( threshold) as Foreground .[ 1] This method
mainly catches motion in the frames. For simplicity,
let’s say we have three single-channel images:
frame1, frame2, and Foreground. We could then use              FIG 4 After applying threshold              and erosion
the following code to detect the magnitude (absolute           technique
value) of foreground differences in Foreground:
Opencv’s cvAbsDiff( ) function can be used for such
differencing.[22 ]. To reduce noise, we can apply
threshold concept. This removes unwanted pixels
which add to noise in our data. Thresholds are
manually adjusted to suit our needs. cvThreshold( ) is
a tool for applying thresholding. We will clean up
small noise areas with cvErode( ) Function .

Algorithm:

1.     Grab a random initial frame and save it in an           FIG 5 Outputs for different thresholds The thresholds for
                                                               images from Left to right and top to bottom are
       image datatype and name it as “background               Th=10,Th=30,Th=60,Th=80,Th=110.
       frame”.
2.     Now for each other frame do the following




                                                                                                                     592
                                          All Rights Reserved © 2012 IJARCET
ISSN: 2278 – 1323
           International Journal of Advanced Research in Computer Engineering & Technology
                                                               Volume 1, Issue 4, June 2012

Experimental Conclusions: [Fig 2,3,4,5]                        all up using cvAcc() and then divide by the total
                                                               number of images to obtain the mean. We can also
1.   Simple gray level absolute differencing causes            accumulate squared images, which will allow us to
     foreground detection with shadow effects and              compute quickly the variance of individual pixels.
     noise.                                                    For finding the covariance[22 ] we can also see how
2.   Thresholding reduces noise and gives binary               images vary over time by selecting a specific lag and
     image which is more predictable.                          then multiplying the current image by the image from
3.   Selection of thresholds depends upon lighting             the past that corresponds to the given lag. Th e
     conditions.                                               function cvMultiplyAcc() will perform a pixelwise
4.   Erosion technique further helps reducing noise.           multiplication of the two images and then add the
                                                               result to the “running total” in acc:[1 ][ 22] [ 5]
Our sample video shows following results:[Fig 5]
                                                                Algorithm:
1.   Th <30 gives foreground detection with lot of
     noise.                                                    i. First, we create pointers for the various scratch
2.   Th>30 and Th<60 gives satisfactory results. Still                and statistics-keeping images we will need along
     has noise contents.                                              the way.
3.   Th >60 and Th<80 gives good results without               ii.    we create a single call to allocate all the
     noise. But shows little loss of foreground data.                 necessary intermediate images. For convenience
4.   Th> 110 Gives only foreground data with great                    we pass in a single image (from our video) that
     loss but without noise.                                          can be used as a reference for sizing the
5.   This means that 30<Th<60 is a better window                      intermediate images. This is done in allocate
     for indoor scenes to achieve better foreground                   images routine.
     detection.                                                iii.   Next we learn the accumulated background
                                                                      image and the accumulated absolute value of
Advantages:                                                           frame-to-frame image differences. This is done
                                                                      in accumulatebackground routine.
Simplest method for background subtraction. Less               iv.    We first use cvCvtScale() to turn the raw
memory requirement as only initial frame is                           background          8-bit-per-channel,       three-
considered as background. Can work well for indoor                    colorchannel image into a floating-point three-
lighting conditions well.                                             channel image. We then accumulate the raw
                                                                      floating-point images into IavgF. Next, we
Disadvantages: Lighting fluctuations are not handled                  calculate the frame-to-frame absolute difference
properly.Threshold    depends      upon      lighting                 image using cvAbsDiff() and accumulate that
conditions.Cannot work well for outdoor scenes                        into image IdiffF. Each time we accumulate
where background may change considerably.                             these images, we increment the image count
                                                                      Icount, a global, to use for averaging later
2. Averaging background method                                 v.     Once we have accumulated enough frames, we
                                                                      convert them into a statistical model of the
                                                                      background. That is, we compute the means and
The averaging method makes use of four OpenCV
                                                                      deviation measures (the average absolute diff
routines: [22 ] [23]
                                                                      erences) of each pixel: using createmodel from
1. cvAcc(), to accumulate images over time;
                                                                      stats routine.
2. cvAbsDiff(), to accumulate frame-to-frame
     image diff erences over time;                             vi.    setHighThreshold() and setLowThreshold() are
3. cvInRange(), to segment the image (once a                          utility functions that set a threshold based on the
     background model has been learned) into                          frame-to-frame average absolute diff erences.
     foreground and background regions;                               The call setHighThreshold(7.0) fi xes a threshold
4. cvOr(), to compile segmentations from different                    such that any value that is 7 times the average
     color channels into a single mask image.                         frame-to-frame absolute diff erence above the
                                                                      average value for that pixel is considered
                                                                      foreground; likewise, setLowThreshold(6.0) sets
Finding Mean , variance and covariance:
                                                                      a threshold bound that is 6 times the average
                                                                      frame-to-frame absolute difference below the
To compute a mean[22 ] value for each pixel across a                  average value for that pixel.
large set of images, the easiest method is to add them




                                                                                                                    593
                                          All Rights Reserved © 2012 IJARCET
ISSN: 2278 – 1323
          International Journal of Advanced Research in Computer Engineering & Technology
                                                              Volume 1, Issue 4, June 2012

vii. Once we have our background model, complete             2.   Needs more memory and time to execute than
     with high and low thresholds, we use it to                   previous method.
     segment the image into foreground (things not
     “explained” by the background image) and the            3.   Advanced background subtraction -Mixture
     background (anything that fits within the high               of Gaussian:
     and low thresholds of our background model).
     Segmentation is done by calling backgrounddiff          Gaussian Function:
     routine.
viii. This function first converts the input image I         In mathematics, a Gaussian function (named after
     (the image to be segmented) into a floating-            Carl Friedrich Gauss) is a function of the form:
     point image by calling cvCvtScale(). We then
     convert the three-channel image into separate
     one-channel image planes using cvSplit(). These
     color channel planes are then checked to see if
     they are within the high and low range of the           for some real constants a, b, c > 0, and e ≈
     average background pixel via the cvInRange()            2.718281828 (Euler's number).The graph of a
     function, which sets the grayscale 8-bit depth          Gaussian is a characteristic symmetric "bell curve"
     image Imaskt to max (255) when it’s in range            shape that quickly falls off towards plus/minus
     and to 0 otherwise. For each color channel we           infinity. (FIG 7 )The parameter a is the height of the
     logically OR the segmentation results into a            curve's peak, b is the position of the centre of the
     mask image Imask, since strong differences in           peak, and c controls the width of the "bell". Gaussian
     any color channel are considered evidence of a          functions are widely used in statistics where they
     foreground pixel here. Finally, we invert Imask         describe the normal distributions, in signal processing
     using cvSubRS(), because foreground should be           where they serve to define Gaussian filters, in image
     the values out of range, not in range. The mask         processing where two-dimensional Gaussians are
     image is the output result.                             used for Gaussian blurs, and in mathematics where
ix. Lastly we de-allocate all the variables and              they are used to solve heat equations and diffusion
     pointers.                                               equations and to define the Weierstrass transform.

                                                             Running Gaussian average was first proposed by
                                                             Wren , Azarbayejani, Darell, pentland in 1997 which
                                                             fits one Gaussian distribution (μ, σ) over the
                                                             histogram.[1]Generalized Gaussian model was
                                                             proposed by H.Kim, R. Sakamoto et all[1 ][ 4] to
                                                             cope up with the background changes and shadows
                                                             using this model. In Mixture of Gaussian ( MOG)
                                                             model background is termed as parametric frame of
                                                             values where each pixel is represented with number
                                                             of Gaussian Functions as Probability Distribution
 FIG: 6 Averaging background method (Background is           function.[6 ]
average of first 10 frames)

Advantages:

1.   Builds a unique model of background depending
     upon earlier “n” frames
2.   Predicts the foreground based on lighting
     condition changes in consecutive frames.
3.   Works well if more frames are taken for
     averaging the background.

Disadvantages:

1.   More complex calculations.                              FIG: 7 Normalized Gaussian curves with expected




                                                                                                               594
                                        All Rights Reserved © 2012 IJARCET
ISSN: 2278 – 1323
           International Journal of Advanced Research in Computer Engineering & Technology
                                                               Volume 1, Issue 4, June 2012

    value μ and variance σ2. The corresponding                   Algorithm:
    parameters are a = 1/(σ√(2π)), b = μ, c = σ[25 ]
                                                                 i.   For             background            subtraction
Mixture of Gaussian basics:                                           CvGaussBGStatModelParams is used where it
                                                                      takes some parameter like bg_threshold and
Sometimes different objects are likely to appear at                   std_threshold. This two parameters are
the same pixel (i,j) location. For example, in an                     responsible to make difference between
outdoor scene with trees partially covering a                         background and foreground.
building: a same (i,j) pixel location will show values           ii. There       is     another     function     called
from tree leaves, tree branches, and the building                     cvCreateGaussianBGModel which takes video
itself. Other examples can be easily drawn from                       frame to create standard background model
snowing, raining, or watching sea waves from a                        according to parameters.
beach. In these cases, a single valued background is             iii. After       that,     extracting      foreground
not an adequate model. Stauffer and Grimson [1 ]                      cvUpdateBGStatModel function is used where it
[22] [26] raised the case for a multi-valued                          takes video frame and background model to
background model able to cope with multiple                           compute differences frame by frame.
background objects. Stauffer and Grimson describe                iv. As an output bgModel->background represent
the probability of observing a certain pixel value, x,                background of the video and bgModel-
at time t by means of a mixture of Gaussians. All                     >foreground shows foreground of the video
weights wi are updated (or normalised) at every new
frame. In practical cases, K is set to be between 3 and          Our chosen Gaussian parameters:
5. Gaussians are multi-variable to describe red, green
and blue values. If these values are assumed                     params->win_size=2; - Learning rate;
independent, the co-variance matrix, ∑i,j simplifies             alpha = 1/CV_GBG_WINDOW_SIZE .
to diagonal.[26 ]
                                                                 params->n_gauss=5; = K = number of
                                                                 Gaussians in mixture .
   At each t frame time, two problems must be                    params->bg_threshold=0.7; threshold sum of
simultaneously solved a) assigning the new observed              weights for background test .
value, x,, to the best matching distribution and b)              params->std_threshold=3;      lambda=2.5 is
estimating the updated model parameters. These                   99%     params->minArea=15;
concurrent problems can he solved by AN
                                                                 params->weight_init=0.25;
EXPECTATION MAXIMIZATION (EM) algorithm.
                                                                 params->variance_init=30;
AN EXPECTATION MAXIMIZATION works on                             CV_BGFG_MOG_SIGMA_INIT*CV_BGFG_MOG_
the buffer of the last n frames. However, as this                SIGMA_INIT
would prove extremely costly, the matching is
approximated          in         these        terms
                                                                 Results:



The first in ranking order is accepted as a match for
      . furthermore the parameters                    are
updated only for this matching distribution and by
using simple online cumulative averages . At every
new frame, some of the Gaussians “match” the
                                                                      Fig:8 mixture of Gaussians background subtraction
current value (those at a distance < 2.5 σi ): for them,              method
μi ,σi are updated by the running average. If no match
is found the last ranked distribution is replaced by a
                                                                 We kept all parameters constant and only changed
new one centered in        , with low weight and high            one parameter at a time to see the probable changes
variance. The mixture of Gaussians actually models               in the output and best values for the parameters. We
both the foreground and the background All                       concluded following results from our observations.
distributions are ranked according to their ωi /σi and
the first ones chosen as “background” .[26 ]




                                                                                                                  595
                                            All Rights Reserved © 2012 IJARCET
ISSN: 2278 – 1323
                International Journal of Advanced Research in Computer Engineering & Technology
                                                                    Volume 1, Issue 4, June 2012

     Mixture of gaussian parameter analysis:                       Sample
                                                                   output
 Paramter                            Observed best value
                                     range
 threshold sum of weights for        0.1 to 0.7                               Better           Good
 background test (bg_threshold)                                                                                 Best
 Standard threshold(                 2 to 5
 std_threshold)
 Number of Gaussians(n_gauss)        3 to 6 ( 3 is better)
 Alpha-learning rate                 0.3 to 0.7( 0 to values)      IV.       SEGMENTATION                      USING
                                                                             CONTOURS   AND                   MOTION
     Advantages: [Fig 8]                                                     TRACKING.
     1.   Best background prediction.
     2.   Model can be adapted to specific scenes by               A contour is a list of points that represent, a curve in
          changing the parameters.                                 an image. [22][23]A contour is represented in
     3.   Least amount of noise incurred when parameters           OpenCV by a CvSeq sequence that is, a sequence of
          are adjusted accordingly.                                points. The function cvFindContours() [22]
     4.   Works well in all lighting conditions.                   computes contours from binary images. The return
                                                                   value of cvFindContours() is the total number of
                                                                   contours found. The first argument is the input
     Disadvantages:                                                image. The next argument, storage, indicates a place
                                                                   where cvFindContours() can find memory in which
1.   Takes largest time and memory among all the                   to record the contours. In a binary image white
     methods implemented by us.                                    regions are labeled as contour and black lines as
2.   Parameter adjustment is a trial and error method .            holes . One of our most basic tasks is drawing a
3.   Needs high processing speeds for real time                    contour on the screen.           For this we have
     environments.                                                 cvDrawContours( ).

     Summary of our analysis:                                      Motion Tracking:[fig 9]

                Frame          Averaging        Mixture            Algorithm:
                differencing   background       of
                               method           Gaussians
                                                method             A) Find a rectangle around counter:

                                                                   1.    BoundingRect() Calculates up-right bounding
                                                                         rectangle of point set. Here we use it to draw a
     Compl      Very less      Medium           More
                                                                         bounding rectangle around the counter we just
     exity
                                                                         got in above step.
     Memor      Very less      Medium           More               2.    BoundingRect() takes two parameters:
     y                                                                   1. Counter type and name= here we have
     require                                                                  taken counter type of cvseq type.
     ments                                                               2. Update flag= update flag is zero in case of
                                                                              cvseq type.
     Executi    Fast           Medium           Very slow
     on time                                                       B) Finding the centre of rectangle:
     Param      Threshold is   Few Needed .     Many
     eter       only needed.                    parameters         1.    Boundingrect function will return us with an
     selectio                                   are needed               object say BNDRECT.
     n                                                             2.    BNDRECT has following information
                                                                         X coordinate of left upper vertex
                                                                         Y coordinate of left upper vertex
                                                                         Width
                                                                         Height




                                                                                                                       596
                                              All Rights Reserved © 2012 IJARCET
ISSN: 2278 – 1323
           International Journal of Advanced Research in Computer Engineering & Technology
                                                               Volume 1, Issue 4, June 2012

3.   Calculate Point 1= (X, Y)                                   Loitering is defined as the presence of an individual
4.   Calculate point2=(X+width, Y+height)                        in an area for a period of time longer than a given
5.   Calculate Centre on X axis(here for our                     time threshold. [ 2][Fig 11]For example loitering can
     convenience)= avgX=( point1.x+point2.x)/2                   be possible at museums, hotels, shops, at car parking,
6.   This avgX can be used as a parameter to detect              hospitals, monuments and business places. Such
     the position of moving object we have detected.             suspicious events where a person stands at an area for
7.   Lastly for visibility draw a rectangle using point          a longer duration can be caught by providing such
     1 and point 2 on moving counter.                            thresholds and defining the area of importance within
                                                                 the scene.
Results:




FIG:9 First frame shows raw video, second shows
foreground detected, third shows contour and fourth shows
a rectangle drawn around the person in the running video.

V.         BEHAVIOR            DETECTION    –
           DETECTING               SUSPICIOUS
           ACTIVITY.

Now as we have done with detection and tracking of               FIG:11 first frame shows person enters an area( marked by
moving object we can predict the motion and                      green rectangle). This area is area of strategic importance.
behavior of moving object.                                       Second frame shows person still inside the area. The third
                                                                 frame shows a “loitering” alert generated when the person
                                                                 stays within the area beyond a certain threshold.
1. Trip Wire detection:
We will take a simple scenario where one is not
supposed to enter a prohibited area. [ 22][Fig 10]We
will take this problem as Crossing the green line and            As soon as we get a suspicious activity we will term
or being present on prohibited area will be termed as            the frame as suspicious and display it in a window.
intrusion. We will be checking avgX value and                    We will extract the intruder image and save it using
monitoring it. If the value is greater or equal to any           bounding rectangle coordinate. Along with
coordinates of trip wire we will term the action as              suspicious frames for further use we can also save the
intrusion:                                                       timing, date, and location data for further ease during
                                                                 investigations.
Result:
                                                                 3.   Abandoned baggage or boxes:

                                                                 Here the system continuously checks for motion in
                                                                 the frames, as well as checks for objects that are not
                                                                 present before. Any stationary object not present
                                                                 before can act as a threat for the public. So such
                                                                 baggage or box can be termed as harmful and can be
                                                                 highlighted.[Fig 12]

                                                                 Face Detection:
Fig: 10The person enters a prohibited area at a
particular time and so an alarm is generated.                    If face is detectable [24 ] we will detect and save it
                                                                 too. These suspicious frames, intruders and face data
2.   Loitering activity:                                         can be used for future use in case of emergency.




                                                                                                                         597
                                            All Rights Reserved © 2012 IJARCET
ISSN: 2278 – 1323
           International Journal of Advanced Research in Computer Engineering & Technology
                                                               Volume 1, Issue 4, June 2012




                                                                   FIG: 13 Faces are detected and rectangle are drawn
                                                                   bounding the faces.

                                                                                          CONCLUSION
     FIG: 12 The first frame shows a person carrying a bag,
     second shows the person keeping the bag on the floor,            Thus we have shown that use of present day
     the third frame shows bag being detected as                   machine vision algorithms along with opencv can
     abandoned object by the system.                               lead us to develop an intelligent surveillance
                                                                   infrastructure. Yet a lot of work needs to be done in
The face data is detectable using HAAR classifiers [               refinement of this behavior detection process. More
22] and viola jones algorithms[ 22][24] provided by                and more behavior needs to be considered and
opencv library. [Fig 13]The haar classifier used is                identified.
haarcascade_frontalface_alt2.xml file which can
detect frontal faces in an image. Whenever a face is                                 ACKNOWLEDGMENT
detected we draw a rectangle around it and save it.
Future work may comprise of supplying such face                    I would like to thank all those people whose support
data to a face matching system which will find out                 and cooperation has been an invaluable asset during
correct matching faces of the suspicious person.                   the course of this paper. I would also like to thank my
Future work also focuses mainly on creating concrete               Guide Prof. Dr. B.K.Lande and our dept head Dr:
classifiers for detecting human and non human                      B.B. MEshram for guiding me throughout this paper
objects in a scene.                                                and giving it the present shape. It would have been
                                                                   impossible to complete the paper without their
                                                                   support,      valuable       suggestions,      criticism,
Algorithm:                                                         encouragement and guidance.

1.  Create a cascade of haar classifiers[22 ] using                I am also grateful for all other teaching and non-
    haarcascade_frontalface_alt2.xml file and                      teaching staff members of the Computer Technology
CvHaarClassifierCascade*           cascade        =                for directly or indirectly helping us for the
(CvHaarClassifierCascade*)cvLoad(                                  completion of this project and the resources provided.
"haarcascade_frontalface_alt2.xml" );
2. Detect objects using cvHaarDetectObjects( );                    REFERENCES
3. This will return number of objects and the
    objects will be stored in memory.                              [1] Niraj Gadhe, Dr:B.K . Lande, VJTI, IOSR Journal of
4. For each object detected( each face detected)                   Engineering (IOSRJEN),Vol. 2 Issue 2, Feb.2012
    draw a rectangle bounding the face.
                                                                   [2] Joshua Candamo, Matthew Shreve, Member, IEEE, Dmitry B.
                                                                   Goldgof, Fellow, IEEE, Deborah B. Sapper, and Rangachar
                                                                   Kasturi, Fellow, IEEE “Understanding Transit Scenes: A Survey
                                                                   on Human Behavior-Recognition Algorithms” in IEEE
     Result:                                                       TRANSACTIONS ON INTELLIGENT TRANSPORTATION
                                                                   SYSTEMS, VOL. 11, NO. 1, MARCH 2010

                                                                   [3] Margarita Osadchy and Daniel Keren “A Rejection-Based
                                                                   Method for Event Detection in Video” in IEEE TRANSACTIONS




                                                                                                                           598
                                              All Rights Reserved © 2012 IJARCET
ISSN: 2278 – 1323
            International Journal of Advanced Research in Computer Engineering & Technology
                                                                Volume 1, Issue 4, June 2012

ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY,                           Conference on Neural Information Processing (ICONIP'O2), Vol.
VOL. 14, NO. 4, APRIL 2004                                              1
[4] H. Kim, R. Sakamoto, I. Kitahara, T. Toriyama and K. Kogure
“Background subtraction using generalized Gaussian family               [17] Cosmin Grigorescu, Student Member, IEEE, and Nicolai
model” in ELECTRONICS LETTERS 31st January 2008 Vol. 44                 Petkov “Distance Sets for Shape Filters
No. 3                                                                   and Shape Recognition” in IEEE TRANSACTIONS ON IMAGE
                                                                        PROCESSING, VOL. 12, NO. 10, OCTOBER 2003
[5] Shahrizat Shaik Mohamed, Nooritawati Md Tahir, Ramli
Adnan “Background Modelling and Background Subtraction                  [18] Chensheng Wang, Fei Wang “A Knowledge-based Strategy
Performance for Object Detection” in 2010 6th International             for Object Recognition and Reconstruction” in 2009 International
Colloquium on Signal Processing & Its Applications (CSPA)               Conference on Information Technology and Computer Science

[6] Dar-Shyang Lee, Member, IEEE “Effective Gaussian Mixture            [19] Junqiu Wang, Member, IEEE, and Yasushi Yagi, Member,
Learning for Video Background Subtraction “ in IEEE                     IEEE “Integrating Color and Shape-Texture Features for Adaptive
TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE                            Real-Time Object Tracking” in IEEE TRANSACTIONS ON
INTELLIGENCE, VOL. 27, NO. 5, MAY 2005                                  IMAGE PROCESSING, VOL. 17, NO. 2, FEBRUARY 2008

[7] Michael E. Farmer, Member, IEEE, and Anil K. Jain, Fellow,          [20] Zoran Zivkovic “Improved Adaptive Gaussian Mixture Model
IEEE “A Wrapper-Based Approach to Image Segmentation and                for Background Subtraction” in Proc. ICPR, 2004
Classification” in IEEE TRANSACTIONS ON IMAGE
PROCESSING, VOL. 14, NO. 12, DECEMBER 2005
                                                                        [21]    Implementing the Viola-Jones Face             Detection
[8] A. Colmenarejo, M. Escudero-Vin˜olo and J. Besco´s “Class-          Algorithm Ole Helvig Jensen Kongens Lyngby            Technical
driven Bayesian background modelling for video object                   University of Denmark, 2008,
segmentation” in ELECTRONICS LETTERS 1st September 2011
Vol. 47 No. 18                                                          [22] Book on “ Learning Open Cv – Computer vision with Open
                                                                        CV library” by Gary Bradski and Adrian Kaebler – o’reilly
[9] Jesse Scott et all “Kalman Filter Based Video Background
Estimation” in 2007IEEE
                                                                        [23]Book on “OpenCV 2 Computer Vision Application
[10] Madhurima , et all “Object tracking in a video sequence using      Programming Cookbook” by Robert Laganière-PACKT
Mean-Shift Based Approach” in IJCEM International Journal of            PUBLISHING.
Computational Engineering & Management, Vol. 11, January 2011

[11] Sen-Ching S. Cheung and Chandrika Kamath “Robust                   [24] Image Processing Project on “Face Recognition Techniques
techniques for background subtraction in urban traffic video”           “by Jorge Orts , Univerity of Wisconsin, Madison.

[12] Padmini Jaikumar, Abhishek Singh and Suman K Mitra
“Background Subtraction in Videos using                                  [25]www.wikipedia.org
Bayesian Learning with Motion Information” in 2008.

[13] Scott Helmer and David G. Lowe “Object Class Recognition           [26] “MIXTURE OF GAUSSIANS USING OPENCV” AS IN
                                                                        HTTP://MMLAB.DISI.UNITN.IT/WIKI/INDEX.PHP/MIXTURE_OF_GAUS
with Many Local Features” in Proceedings of the 2004 IEEE
                                                                        SIANS_USING_OPENCV.
Computer Society Conference on Computer Vision and Pattern
Recognition Workshops.
                                                                        [27]    IMAGE AND VIDEO DATABASES OF CVONLINE: IMAGE
[14] Lun Zhang, Stan Z. Li, Xiaotong Yuan and Shiming Xiang             DATABASES
“Real-time Object Classification in Video Surveillance Based on         HTTP://HOMEPAGES.INF.ED.AC.UK/RBF/CVONLINE/IMAGEDBASE.H
Appearance Learning” in 2007 IEEE                                       TM#RETRIEVE


[15] Tie Liu, Zejian Yuan, Jian Sun, Jingdong Wang, Nanning             [28] Surveillance databases – VISOR- video urviellance online
Zheng, Fellow, IEEE,et all “Learning to Detect a Salient Object”        repository-
in IEEE TRANSACTIONS ON PATTERN ANALYSIS AND                            http://imagelab.ing.unimore.it/visor/video_videosInCategory.asp?i
MACHINE INTELLIGENCE, VOL. 33, NO. 2, FEBRUARY                          dcategory=14
2011

[16] Zhiyong Wang, Zheru Chi and David Feng “STRUCTURAL
REPRESENTATION AND BPTS LEARNING FOR SHAPE
CLASSIFICATION” in Proceedings of the 9th International




                                                                                                                                   599
                                                   All Rights Reserved © 2012 IJARCET

Weitere ähnliche Inhalte

Was ist angesagt?

IRJET - Deep Learning Applications and Frameworks – A Review
IRJET -  	  Deep Learning Applications and Frameworks – A ReviewIRJET -  	  Deep Learning Applications and Frameworks – A Review
IRJET - Deep Learning Applications and Frameworks – A ReviewIRJET Journal
 
User identity verification via mouse dynamics
User identity verification via mouse dynamicsUser identity verification via mouse dynamics
User identity verification via mouse dynamicsBalwant Gorad
 
IRJET- Smart Ship Detection using Transfer Learning with ResNet
IRJET- Smart Ship Detection using Transfer Learning with ResNetIRJET- Smart Ship Detection using Transfer Learning with ResNet
IRJET- Smart Ship Detection using Transfer Learning with ResNetIRJET Journal
 
A real time aggressive human behaviour detection system in cage environment a...
A real time aggressive human behaviour detection system in cage environment a...A real time aggressive human behaviour detection system in cage environment a...
A real time aggressive human behaviour detection system in cage environment a...Journal Papers
 
IRJET - Direct Me-Nevigation for Blind People
IRJET -  	  Direct Me-Nevigation for Blind PeopleIRJET -  	  Direct Me-Nevigation for Blind People
IRJET - Direct Me-Nevigation for Blind PeopleIRJET Journal
 
Graphical Password by Watermarking for security
Graphical Password by Watermarking for securityGraphical Password by Watermarking for security
Graphical Password by Watermarking for securityIJERA Editor
 
IRJET- Lossless Encryption Technique for Finger Biometric Images
IRJET-  	  Lossless Encryption Technique for Finger Biometric ImagesIRJET-  	  Lossless Encryption Technique for Finger Biometric Images
IRJET- Lossless Encryption Technique for Finger Biometric ImagesIRJET Journal
 
Graphical Based Password for Android Phones using Keystroke Dynamics - A Survey
Graphical Based Password for Android Phones using Keystroke Dynamics - A SurveyGraphical Based Password for Android Phones using Keystroke Dynamics - A Survey
Graphical Based Password for Android Phones using Keystroke Dynamics - A SurveyIJSRD
 
Usable security- It isn't secure if people can't use it. O-ISC conference 14m...
Usable security- It isn't secure if people can't use it. O-ISC conference 14m...Usable security- It isn't secure if people can't use it. O-ISC conference 14m...
Usable security- It isn't secure if people can't use it. O-ISC conference 14m...Darren Kall
 
Video Analysis with Convolutional Neural Networks (Master Computer Vision Bar...
Video Analysis with Convolutional Neural Networks (Master Computer Vision Bar...Video Analysis with Convolutional Neural Networks (Master Computer Vision Bar...
Video Analysis with Convolutional Neural Networks (Master Computer Vision Bar...Universitat Politècnica de Catalunya
 
IRJET-Design and Fabrication of Automatic Plastic Cup Thermoforming Machine
IRJET-Design and Fabrication of Automatic Plastic Cup Thermoforming MachineIRJET-Design and Fabrication of Automatic Plastic Cup Thermoforming Machine
IRJET-Design and Fabrication of Automatic Plastic Cup Thermoforming MachineIRJET Journal
 
A deep awareness framework for pervasiv video cloud
A deep awareness framework for pervasiv video cloudA deep awareness framework for pervasiv video cloud
A deep awareness framework for pervasiv video cloudredpel dot com
 
Usable security it isn't secure if people can't use it mwux 2 jun2012
Usable security   it isn't secure if people can't use it mwux 2 jun2012Usable security   it isn't secure if people can't use it mwux 2 jun2012
Usable security it isn't secure if people can't use it mwux 2 jun2012Darren Kall
 
V Code And V Data Illustrating A New Framework For Supporting The Video Annot...
V Code And V Data Illustrating A New Framework For Supporting The Video Annot...V Code And V Data Illustrating A New Framework For Supporting The Video Annot...
V Code And V Data Illustrating A New Framework For Supporting The Video Annot...GoogleTecTalks
 
International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)IJERD Editor
 
Image Retrieval for Visual Understanding in Dynamic and Sensor Rich Environments
Image Retrieval for Visual Understanding in Dynamic and Sensor Rich EnvironmentsImage Retrieval for Visual Understanding in Dynamic and Sensor Rich Environments
Image Retrieval for Visual Understanding in Dynamic and Sensor Rich EnvironmentsMichael Ippolito
 

Was ist angesagt? (18)

IRJET - Deep Learning Applications and Frameworks – A Review
IRJET -  	  Deep Learning Applications and Frameworks – A ReviewIRJET -  	  Deep Learning Applications and Frameworks – A Review
IRJET - Deep Learning Applications and Frameworks – A Review
 
User identity verification via mouse dynamics
User identity verification via mouse dynamicsUser identity verification via mouse dynamics
User identity verification via mouse dynamics
 
IRJET- Smart Ship Detection using Transfer Learning with ResNet
IRJET- Smart Ship Detection using Transfer Learning with ResNetIRJET- Smart Ship Detection using Transfer Learning with ResNet
IRJET- Smart Ship Detection using Transfer Learning with ResNet
 
A real time aggressive human behaviour detection system in cage environment a...
A real time aggressive human behaviour detection system in cage environment a...A real time aggressive human behaviour detection system in cage environment a...
A real time aggressive human behaviour detection system in cage environment a...
 
IRJET - Direct Me-Nevigation for Blind People
IRJET -  	  Direct Me-Nevigation for Blind PeopleIRJET -  	  Direct Me-Nevigation for Blind People
IRJET - Direct Me-Nevigation for Blind People
 
Graphical Password by Watermarking for security
Graphical Password by Watermarking for securityGraphical Password by Watermarking for security
Graphical Password by Watermarking for security
 
IRJET- Lossless Encryption Technique for Finger Biometric Images
IRJET-  	  Lossless Encryption Technique for Finger Biometric ImagesIRJET-  	  Lossless Encryption Technique for Finger Biometric Images
IRJET- Lossless Encryption Technique for Finger Biometric Images
 
Graphical Based Password for Android Phones using Keystroke Dynamics - A Survey
Graphical Based Password for Android Phones using Keystroke Dynamics - A SurveyGraphical Based Password for Android Phones using Keystroke Dynamics - A Survey
Graphical Based Password for Android Phones using Keystroke Dynamics - A Survey
 
Usable security- It isn't secure if people can't use it. O-ISC conference 14m...
Usable security- It isn't secure if people can't use it. O-ISC conference 14m...Usable security- It isn't secure if people can't use it. O-ISC conference 14m...
Usable security- It isn't secure if people can't use it. O-ISC conference 14m...
 
Video Analysis with Convolutional Neural Networks (Master Computer Vision Bar...
Video Analysis with Convolutional Neural Networks (Master Computer Vision Bar...Video Analysis with Convolutional Neural Networks (Master Computer Vision Bar...
Video Analysis with Convolutional Neural Networks (Master Computer Vision Bar...
 
IRJET-Design and Fabrication of Automatic Plastic Cup Thermoforming Machine
IRJET-Design and Fabrication of Automatic Plastic Cup Thermoforming MachineIRJET-Design and Fabrication of Automatic Plastic Cup Thermoforming Machine
IRJET-Design and Fabrication of Automatic Plastic Cup Thermoforming Machine
 
1709 1715
1709 17151709 1715
1709 1715
 
A deep awareness framework for pervasiv video cloud
A deep awareness framework for pervasiv video cloudA deep awareness framework for pervasiv video cloud
A deep awareness framework for pervasiv video cloud
 
Usable security it isn't secure if people can't use it mwux 2 jun2012
Usable security   it isn't secure if people can't use it mwux 2 jun2012Usable security   it isn't secure if people can't use it mwux 2 jun2012
Usable security it isn't secure if people can't use it mwux 2 jun2012
 
Dz31840846
Dz31840846Dz31840846
Dz31840846
 
V Code And V Data Illustrating A New Framework For Supporting The Video Annot...
V Code And V Data Illustrating A New Framework For Supporting The Video Annot...V Code And V Data Illustrating A New Framework For Supporting The Video Annot...
V Code And V Data Illustrating A New Framework For Supporting The Video Annot...
 
International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)
 
Image Retrieval for Visual Understanding in Dynamic and Sensor Rich Environments
Image Retrieval for Visual Understanding in Dynamic and Sensor Rich EnvironmentsImage Retrieval for Visual Understanding in Dynamic and Sensor Rich Environments
Image Retrieval for Visual Understanding in Dynamic and Sensor Rich Environments
 

Andere mochten auch (9)

102 105
102 105102 105
102 105
 
24 26
24 2624 26
24 26
 
666 670
666 670666 670
666 670
 
75 78
75 7875 78
75 78
 
42 46
42 4642 46
42 46
 
27 32
27 3227 32
27 32
 
75 78
75 7875 78
75 78
 
315 319
315 319315 319
315 319
 
1 7
1 71 7
1 7
 

Ähnlich wie 590 599

IRJET- Real-Time Object Detection System using Caffe Model
IRJET- Real-Time Object Detection System using Caffe ModelIRJET- Real-Time Object Detection System using Caffe Model
IRJET- Real-Time Object Detection System using Caffe ModelIRJET Journal
 
YOLOv4: A Face Mask Detection System
YOLOv4: A Face Mask Detection SystemYOLOv4: A Face Mask Detection System
YOLOv4: A Face Mask Detection SystemIRJET Journal
 
Human Activity Recognition
Human Activity RecognitionHuman Activity Recognition
Human Activity RecognitionIRJET Journal
 
Object and Currency Detection for the Visually Impaired
Object and Currency Detection for the Visually ImpairedObject and Currency Detection for the Visually Impaired
Object and Currency Detection for the Visually ImpairedIRJET Journal
 
IRJET- Object Detection in an Image using Deep Learning
IRJET- Object Detection in an Image using Deep LearningIRJET- Object Detection in an Image using Deep Learning
IRJET- Object Detection in an Image using Deep LearningIRJET Journal
 
NoxEye.pptx
NoxEye.pptxNoxEye.pptx
NoxEye.pptxMathanE5
 
Review of Pose Recognition Systems
Review of Pose Recognition SystemsReview of Pose Recognition Systems
Review of Pose Recognition Systemsvivatechijri
 
Java Implementation based Heterogeneous Video Sequence Automated Surveillance...
Java Implementation based Heterogeneous Video Sequence Automated Surveillance...Java Implementation based Heterogeneous Video Sequence Automated Surveillance...
Java Implementation based Heterogeneous Video Sequence Automated Surveillance...CSCJournals
 
Object Detection and Tracking AI Robot
Object Detection and Tracking AI RobotObject Detection and Tracking AI Robot
Object Detection and Tracking AI RobotIRJET Journal
 
SMART SURVEILLANCE SYSTEM USING LBPH ALGORITHM
SMART SURVEILLANCE SYSTEM USING LBPH ALGORITHMSMART SURVEILLANCE SYSTEM USING LBPH ALGORITHM
SMART SURVEILLANCE SYSTEM USING LBPH ALGORITHMIRJET Journal
 
IRJET- Review on Human Action Detection in Stored Videos using Support Vector...
IRJET- Review on Human Action Detection in Stored Videos using Support Vector...IRJET- Review on Human Action Detection in Stored Videos using Support Vector...
IRJET- Review on Human Action Detection in Stored Videos using Support Vector...IRJET Journal
 
Image Recognition Expert System based on deep learning
Image Recognition Expert System based on deep learningImage Recognition Expert System based on deep learning
Image Recognition Expert System based on deep learningPRATHAMESH REGE
 
Analysis of Human Behavior Based On Centroid and Treading Track
Analysis of Human Behavior Based On Centroid and Treading  TrackAnalysis of Human Behavior Based On Centroid and Treading  Track
Analysis of Human Behavior Based On Centroid and Treading TrackIJMER
 
Classification Rule Discovery Using Ant-Miner Algorithm: An Application Of N...
Classification Rule Discovery Using Ant-Miner Algorithm: An  Application Of N...Classification Rule Discovery Using Ant-Miner Algorithm: An  Application Of N...
Classification Rule Discovery Using Ant-Miner Algorithm: An Application Of N...IJMER
 
SENSITIVITY OF A VIDEO SURVEILLANCE SYSTEM BASED ON MOTION DETECTION
SENSITIVITY OF A VIDEO SURVEILLANCE SYSTEM BASED ON MOTION DETECTIONSENSITIVITY OF A VIDEO SURVEILLANCE SYSTEM BASED ON MOTION DETECTION
SENSITIVITY OF A VIDEO SURVEILLANCE SYSTEM BASED ON MOTION DETECTIONsipij
 
Event-Handling Based Smart Video Surveillance System
Event-Handling Based Smart Video Surveillance SystemEvent-Handling Based Smart Video Surveillance System
Event-Handling Based Smart Video Surveillance SystemCSCJournals
 
Inspection of Suspicious Human Activity in the Crowd Sourced Areas Captured i...
Inspection of Suspicious Human Activity in the Crowd Sourced Areas Captured i...Inspection of Suspicious Human Activity in the Crowd Sourced Areas Captured i...
Inspection of Suspicious Human Activity in the Crowd Sourced Areas Captured i...IRJET Journal
 
IRJET - Real-Time Analysis of Video Surveillance using Machine Learning a...
IRJET -  	  Real-Time Analysis of Video Surveillance using Machine Learning a...IRJET -  	  Real-Time Analysis of Video Surveillance using Machine Learning a...
IRJET - Real-Time Analysis of Video Surveillance using Machine Learning a...IRJET Journal
 
Human Motion Detection in Video Surveillance using Computer Vision Technique
Human Motion Detection in Video Surveillance using Computer Vision TechniqueHuman Motion Detection in Video Surveillance using Computer Vision Technique
Human Motion Detection in Video Surveillance using Computer Vision TechniqueIRJET Journal
 
AUTOMATIC THEFT SECURITY SYSTEM (SMART SURVEILLANCE CAMERA)
AUTOMATIC THEFT SECURITY SYSTEM (SMART SURVEILLANCE CAMERA)AUTOMATIC THEFT SECURITY SYSTEM (SMART SURVEILLANCE CAMERA)
AUTOMATIC THEFT SECURITY SYSTEM (SMART SURVEILLANCE CAMERA)cscpconf
 

Ähnlich wie 590 599 (20)

IRJET- Real-Time Object Detection System using Caffe Model
IRJET- Real-Time Object Detection System using Caffe ModelIRJET- Real-Time Object Detection System using Caffe Model
IRJET- Real-Time Object Detection System using Caffe Model
 
YOLOv4: A Face Mask Detection System
YOLOv4: A Face Mask Detection SystemYOLOv4: A Face Mask Detection System
YOLOv4: A Face Mask Detection System
 
Human Activity Recognition
Human Activity RecognitionHuman Activity Recognition
Human Activity Recognition
 
Object and Currency Detection for the Visually Impaired
Object and Currency Detection for the Visually ImpairedObject and Currency Detection for the Visually Impaired
Object and Currency Detection for the Visually Impaired
 
IRJET- Object Detection in an Image using Deep Learning
IRJET- Object Detection in an Image using Deep LearningIRJET- Object Detection in an Image using Deep Learning
IRJET- Object Detection in an Image using Deep Learning
 
NoxEye.pptx
NoxEye.pptxNoxEye.pptx
NoxEye.pptx
 
Review of Pose Recognition Systems
Review of Pose Recognition SystemsReview of Pose Recognition Systems
Review of Pose Recognition Systems
 
Java Implementation based Heterogeneous Video Sequence Automated Surveillance...
Java Implementation based Heterogeneous Video Sequence Automated Surveillance...Java Implementation based Heterogeneous Video Sequence Automated Surveillance...
Java Implementation based Heterogeneous Video Sequence Automated Surveillance...
 
Object Detection and Tracking AI Robot
Object Detection and Tracking AI RobotObject Detection and Tracking AI Robot
Object Detection and Tracking AI Robot
 
SMART SURVEILLANCE SYSTEM USING LBPH ALGORITHM
SMART SURVEILLANCE SYSTEM USING LBPH ALGORITHMSMART SURVEILLANCE SYSTEM USING LBPH ALGORITHM
SMART SURVEILLANCE SYSTEM USING LBPH ALGORITHM
 
IRJET- Review on Human Action Detection in Stored Videos using Support Vector...
IRJET- Review on Human Action Detection in Stored Videos using Support Vector...IRJET- Review on Human Action Detection in Stored Videos using Support Vector...
IRJET- Review on Human Action Detection in Stored Videos using Support Vector...
 
Image Recognition Expert System based on deep learning
Image Recognition Expert System based on deep learningImage Recognition Expert System based on deep learning
Image Recognition Expert System based on deep learning
 
Analysis of Human Behavior Based On Centroid and Treading Track
Analysis of Human Behavior Based On Centroid and Treading  TrackAnalysis of Human Behavior Based On Centroid and Treading  Track
Analysis of Human Behavior Based On Centroid and Treading Track
 
Classification Rule Discovery Using Ant-Miner Algorithm: An Application Of N...
Classification Rule Discovery Using Ant-Miner Algorithm: An  Application Of N...Classification Rule Discovery Using Ant-Miner Algorithm: An  Application Of N...
Classification Rule Discovery Using Ant-Miner Algorithm: An Application Of N...
 
SENSITIVITY OF A VIDEO SURVEILLANCE SYSTEM BASED ON MOTION DETECTION
SENSITIVITY OF A VIDEO SURVEILLANCE SYSTEM BASED ON MOTION DETECTIONSENSITIVITY OF A VIDEO SURVEILLANCE SYSTEM BASED ON MOTION DETECTION
SENSITIVITY OF A VIDEO SURVEILLANCE SYSTEM BASED ON MOTION DETECTION
 
Event-Handling Based Smart Video Surveillance System
Event-Handling Based Smart Video Surveillance SystemEvent-Handling Based Smart Video Surveillance System
Event-Handling Based Smart Video Surveillance System
 
Inspection of Suspicious Human Activity in the Crowd Sourced Areas Captured i...
Inspection of Suspicious Human Activity in the Crowd Sourced Areas Captured i...Inspection of Suspicious Human Activity in the Crowd Sourced Areas Captured i...
Inspection of Suspicious Human Activity in the Crowd Sourced Areas Captured i...
 
IRJET - Real-Time Analysis of Video Surveillance using Machine Learning a...
IRJET -  	  Real-Time Analysis of Video Surveillance using Machine Learning a...IRJET -  	  Real-Time Analysis of Video Surveillance using Machine Learning a...
IRJET - Real-Time Analysis of Video Surveillance using Machine Learning a...
 
Human Motion Detection in Video Surveillance using Computer Vision Technique
Human Motion Detection in Video Surveillance using Computer Vision TechniqueHuman Motion Detection in Video Surveillance using Computer Vision Technique
Human Motion Detection in Video Surveillance using Computer Vision Technique
 
AUTOMATIC THEFT SECURITY SYSTEM (SMART SURVEILLANCE CAMERA)
AUTOMATIC THEFT SECURITY SYSTEM (SMART SURVEILLANCE CAMERA)AUTOMATIC THEFT SECURITY SYSTEM (SMART SURVEILLANCE CAMERA)
AUTOMATIC THEFT SECURITY SYSTEM (SMART SURVEILLANCE CAMERA)
 

Mehr von Editor IJARCET

Electrically small antennas: The art of miniaturization
Electrically small antennas: The art of miniaturizationElectrically small antennas: The art of miniaturization
Electrically small antennas: The art of miniaturizationEditor IJARCET
 
Volume 2-issue-6-2205-2207
Volume 2-issue-6-2205-2207Volume 2-issue-6-2205-2207
Volume 2-issue-6-2205-2207Editor IJARCET
 
Volume 2-issue-6-2195-2199
Volume 2-issue-6-2195-2199Volume 2-issue-6-2195-2199
Volume 2-issue-6-2195-2199Editor IJARCET
 
Volume 2-issue-6-2200-2204
Volume 2-issue-6-2200-2204Volume 2-issue-6-2200-2204
Volume 2-issue-6-2200-2204Editor IJARCET
 
Volume 2-issue-6-2190-2194
Volume 2-issue-6-2190-2194Volume 2-issue-6-2190-2194
Volume 2-issue-6-2190-2194Editor IJARCET
 
Volume 2-issue-6-2186-2189
Volume 2-issue-6-2186-2189Volume 2-issue-6-2186-2189
Volume 2-issue-6-2186-2189Editor IJARCET
 
Volume 2-issue-6-2177-2185
Volume 2-issue-6-2177-2185Volume 2-issue-6-2177-2185
Volume 2-issue-6-2177-2185Editor IJARCET
 
Volume 2-issue-6-2173-2176
Volume 2-issue-6-2173-2176Volume 2-issue-6-2173-2176
Volume 2-issue-6-2173-2176Editor IJARCET
 
Volume 2-issue-6-2165-2172
Volume 2-issue-6-2165-2172Volume 2-issue-6-2165-2172
Volume 2-issue-6-2165-2172Editor IJARCET
 
Volume 2-issue-6-2159-2164
Volume 2-issue-6-2159-2164Volume 2-issue-6-2159-2164
Volume 2-issue-6-2159-2164Editor IJARCET
 
Volume 2-issue-6-2155-2158
Volume 2-issue-6-2155-2158Volume 2-issue-6-2155-2158
Volume 2-issue-6-2155-2158Editor IJARCET
 
Volume 2-issue-6-2148-2154
Volume 2-issue-6-2148-2154Volume 2-issue-6-2148-2154
Volume 2-issue-6-2148-2154Editor IJARCET
 
Volume 2-issue-6-2143-2147
Volume 2-issue-6-2143-2147Volume 2-issue-6-2143-2147
Volume 2-issue-6-2143-2147Editor IJARCET
 
Volume 2-issue-6-2119-2124
Volume 2-issue-6-2119-2124Volume 2-issue-6-2119-2124
Volume 2-issue-6-2119-2124Editor IJARCET
 
Volume 2-issue-6-2139-2142
Volume 2-issue-6-2139-2142Volume 2-issue-6-2139-2142
Volume 2-issue-6-2139-2142Editor IJARCET
 
Volume 2-issue-6-2130-2138
Volume 2-issue-6-2130-2138Volume 2-issue-6-2130-2138
Volume 2-issue-6-2130-2138Editor IJARCET
 
Volume 2-issue-6-2125-2129
Volume 2-issue-6-2125-2129Volume 2-issue-6-2125-2129
Volume 2-issue-6-2125-2129Editor IJARCET
 
Volume 2-issue-6-2114-2118
Volume 2-issue-6-2114-2118Volume 2-issue-6-2114-2118
Volume 2-issue-6-2114-2118Editor IJARCET
 
Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113Editor IJARCET
 
Volume 2-issue-6-2102-2107
Volume 2-issue-6-2102-2107Volume 2-issue-6-2102-2107
Volume 2-issue-6-2102-2107Editor IJARCET
 

Mehr von Editor IJARCET (20)

Electrically small antennas: The art of miniaturization
Electrically small antennas: The art of miniaturizationElectrically small antennas: The art of miniaturization
Electrically small antennas: The art of miniaturization
 
Volume 2-issue-6-2205-2207
Volume 2-issue-6-2205-2207Volume 2-issue-6-2205-2207
Volume 2-issue-6-2205-2207
 
Volume 2-issue-6-2195-2199
Volume 2-issue-6-2195-2199Volume 2-issue-6-2195-2199
Volume 2-issue-6-2195-2199
 
Volume 2-issue-6-2200-2204
Volume 2-issue-6-2200-2204Volume 2-issue-6-2200-2204
Volume 2-issue-6-2200-2204
 
Volume 2-issue-6-2190-2194
Volume 2-issue-6-2190-2194Volume 2-issue-6-2190-2194
Volume 2-issue-6-2190-2194
 
Volume 2-issue-6-2186-2189
Volume 2-issue-6-2186-2189Volume 2-issue-6-2186-2189
Volume 2-issue-6-2186-2189
 
Volume 2-issue-6-2177-2185
Volume 2-issue-6-2177-2185Volume 2-issue-6-2177-2185
Volume 2-issue-6-2177-2185
 
Volume 2-issue-6-2173-2176
Volume 2-issue-6-2173-2176Volume 2-issue-6-2173-2176
Volume 2-issue-6-2173-2176
 
Volume 2-issue-6-2165-2172
Volume 2-issue-6-2165-2172Volume 2-issue-6-2165-2172
Volume 2-issue-6-2165-2172
 
Volume 2-issue-6-2159-2164
Volume 2-issue-6-2159-2164Volume 2-issue-6-2159-2164
Volume 2-issue-6-2159-2164
 
Volume 2-issue-6-2155-2158
Volume 2-issue-6-2155-2158Volume 2-issue-6-2155-2158
Volume 2-issue-6-2155-2158
 
Volume 2-issue-6-2148-2154
Volume 2-issue-6-2148-2154Volume 2-issue-6-2148-2154
Volume 2-issue-6-2148-2154
 
Volume 2-issue-6-2143-2147
Volume 2-issue-6-2143-2147Volume 2-issue-6-2143-2147
Volume 2-issue-6-2143-2147
 
Volume 2-issue-6-2119-2124
Volume 2-issue-6-2119-2124Volume 2-issue-6-2119-2124
Volume 2-issue-6-2119-2124
 
Volume 2-issue-6-2139-2142
Volume 2-issue-6-2139-2142Volume 2-issue-6-2139-2142
Volume 2-issue-6-2139-2142
 
Volume 2-issue-6-2130-2138
Volume 2-issue-6-2130-2138Volume 2-issue-6-2130-2138
Volume 2-issue-6-2130-2138
 
Volume 2-issue-6-2125-2129
Volume 2-issue-6-2125-2129Volume 2-issue-6-2125-2129
Volume 2-issue-6-2125-2129
 
Volume 2-issue-6-2114-2118
Volume 2-issue-6-2114-2118Volume 2-issue-6-2114-2118
Volume 2-issue-6-2114-2118
 
Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113Volume 2-issue-6-2108-2113
Volume 2-issue-6-2108-2113
 
Volume 2-issue-6-2102-2107
Volume 2-issue-6-2102-2107Volume 2-issue-6-2102-2107
Volume 2-issue-6-2102-2107
 

Kürzlich hochgeladen

[Webinar] SpiraTest - Setting New Standards in Quality Assurance
[Webinar] SpiraTest - Setting New Standards in Quality Assurance[Webinar] SpiraTest - Setting New Standards in Quality Assurance
[Webinar] SpiraTest - Setting New Standards in Quality AssuranceInflectra
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsNathaniel Shimoni
 
UiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPathCommunity
 
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...Alkin Tezuysal
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentPim van der Noll
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .Alan Dix
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsPixlogix Infotech
 
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxThe Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxLoriGlavin3
 
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...Wes McKinney
 
So einfach geht modernes Roaming fuer Notes und Nomad.pdf
So einfach geht modernes Roaming fuer Notes und Nomad.pdfSo einfach geht modernes Roaming fuer Notes und Nomad.pdf
So einfach geht modernes Roaming fuer Notes und Nomad.pdfpanagenda
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.Curtis Poe
 
Connecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfConnecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfNeo4j
 
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyesHow to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyesThousandEyes
 
Manual 508 Accessibility Compliance Audit
Manual 508 Accessibility Compliance AuditManual 508 Accessibility Compliance Audit
Manual 508 Accessibility Compliance AuditSkynet Technologies
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteDianaGray10
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsSergiu Bodiu
 
A Framework for Development in the AI Age
A Framework for Development in the AI AgeA Framework for Development in the AI Age
A Framework for Development in the AI AgeCprime
 
How to write a Business Continuity Plan
How to write a Business Continuity PlanHow to write a Business Continuity Plan
How to write a Business Continuity PlanDatabarracks
 
2024 April Patch Tuesday
2024 April Patch Tuesday2024 April Patch Tuesday
2024 April Patch TuesdayIvanti
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfLoriGlavin3
 

Kürzlich hochgeladen (20)

[Webinar] SpiraTest - Setting New Standards in Quality Assurance
[Webinar] SpiraTest - Setting New Standards in Quality Assurance[Webinar] SpiraTest - Setting New Standards in Quality Assurance
[Webinar] SpiraTest - Setting New Standards in Quality Assurance
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directions
 
UiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to Hero
 
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and Cons
 
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxThe Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
 
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
 
So einfach geht modernes Roaming fuer Notes und Nomad.pdf
So einfach geht modernes Roaming fuer Notes und Nomad.pdfSo einfach geht modernes Roaming fuer Notes und Nomad.pdf
So einfach geht modernes Roaming fuer Notes und Nomad.pdf
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.
 
Connecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfConnecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdf
 
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyesHow to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
 
Manual 508 Accessibility Compliance Audit
Manual 508 Accessibility Compliance AuditManual 508 Accessibility Compliance Audit
Manual 508 Accessibility Compliance Audit
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test Suite
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platforms
 
A Framework for Development in the AI Age
A Framework for Development in the AI AgeA Framework for Development in the AI Age
A Framework for Development in the AI Age
 
How to write a Business Continuity Plan
How to write a Business Continuity PlanHow to write a Business Continuity Plan
How to write a Business Continuity Plan
 
2024 April Patch Tuesday
2024 April Patch Tuesday2024 April Patch Tuesday
2024 April Patch Tuesday
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdf
 

590 599

  • 1. ISSN: 2278 – 1323 International Journal of Advanced Research in Computer Engineering & Technology Volume 1, Issue 4, June 2012 Intelligent System for detecting, Modeling, Classification of human behavior using image processing, machine vision and OpenCV Niraj .B. Gadhe, Prof: Dr. B .K . Lande, Prof: Dr. B.B.Meshram Abstract: Surveillance Cameras has proven to be a key detect malicious activities. It would be so easy if factor in enhancing the public security in many system on its own recognizes some events as countries around the world . In spite of advancements malicious and extract only those frames which are in image processing and machine vision techniques very providing suspicious activity. The endless hours of less is applied to actual implementation of such surveillance systems. Traditional methods wherein watching the video can be avoided. Moreover stand alone cameras provide just live feed to the assistance provided by system during live feed of observer are still deployed at most of the places but the surveillance may prove beneficial as it can alert the need of today’s world is to add some intelligence to this observer of a possible danger and drag his attention systems. In this paper we will be focusing on important to a particular activity which he might have neglected features to build an intelligent surveillance system and in absence of intelligent system. its architecture using computer vision libraries of In the light of such weak security infrastructure opencv. Background Subtraction , segmentation, we propose a method to develop an intelligent system tracking and human behavior analysis are step by step which not only will capture and feedback the video at implemented to build such a system. A detailed analysis of background subtraction methods and their real time but also model human behavior.[2] We comparison is provided. Lastly we discuss the future think of detecting , modeling and classifying human work which can lead to human behavior analysis like behavior using image processing to detect malicious crowd counting, intrusion detection, loitering , crowd behavior. Further we may maintain a database of movement , vehicle tracking and all. suspects found over time and use face detection [ 24] and recognition to identify the probable matches for Index Terms: surveillance systems, opencv, the suspects. background subtraction, segmentation, machine There is lot of work done in Image processing as vision, computer vision. well as machine vision techniques. A much of it is local and developed for particular domain. The I. INTRODUCTION object recognition frame works consist of algorithms like Pattern matching, feature In traditional systems[ 1] a network of passive matching , boundary detection, knowledge based cameras able of only recording the moving images repository systems.[18 ] The behavioral model and converting it to video files are deployed at the uses heuristics and classification models like site of prime importance. In case of accidents, Motion based classification , Shape based terrorist attacks , theft, malicious activities such video classification, color based classification.[19][18 ] data comes to be very important for analysis. But if [17][14] Once the object is classified and detected traditional system is used we require to watch the it can be tracked using algorithms like mean shift video data over and over for hours together because algorithm.[10 ][22] Human behavior is studied collected data is very large and we don’t know the under Bayesian Model and Markov Models.[12 ] exact time when the malicious activity took place. [8] Face recognition and detection done via Eigen Such manual method is prone to human failure as faces, SIFT based, Edge based, Binary pattern face human may miss small details like intrusion[2] or recognitions, viola jones face detection are famous.[ may neglect some activity. So its necessary to build 21][24]. Much of the work today is focused on an intelligent system that would not only provide detecting occluded object.[4] data as per traditional system but also assist us to The rest of the paper is organized as follows: Section 2 discusses our architecture of system and Niraj Babasaheb Gadhe, Dept of computers- NIMS, VJTI, (e-mail: features we decided to incorporate in it. Section 3 nirajgadhe@gmail.com). Mumbai.India, Prof: Dr. B.K.Lande, Computer engg dept, Shah and anchor presents a detailed analysis of famous background college of engg, University of Mumbai, Mumbai, India , subtraction methods. . A detailed review of present Prof: Dr. B.B.Meshram, Computer engg dept, VJTI, Mumbai, image processing algorithms and their advantages India , are discussed . Section 4 we study segmentation of 590 All Rights Reserved © 2012 IJARCET
  • 2. ISSN: 2278 – 1323 International Journal of Advanced Research in Computer Engineering & Technology Volume 1, Issue 4, June 2012 object and motion tracking of object here. Section 5 some human behaviors of malicious intent. We will discusses the behavior which can be detected using take a simple real world problem like Trespassing or the system. Lastly we conclude with the future work. intrusion, loitering, bag being left by a human and flee, crowd analysis. [2] II. BUILDING AN INTELLIGENT D) Capturing frames containing suspicious SURVEILLANCE SYSTEM activities module: Capturing frames containing probable suspicious We are to develop a system based on image and behavior, classifying them and maintaining it for video processing algorithms that will help us detect further use and alarming the system . Face detection and classify human behavior. As detecting a human and storing face data can be done.[24][21] Further behavior from a scene we need to follow some such data can be used during investigation of crimes. sequential steps in our processing of the video. These According to our architecture(FIG 1 ) the working of sequential steps can be generalized and explained in our system can be summarized as follows brief as follows . [1] 1. We first acquire frames from a camera or video A) Frame detection module. file. 2. We first acquire frames from a camera or video This routine will detect each frame at real time from file. a video sequence and present its status in form of 3. Now we pass on these frames to background matrix, its frame no, pixels, color content and sample subtraction module. For convenience we are rate, etc.[23] Now we can treat each captured frame using the frame differencing for background as a image data type and apply image processing on subtraction. it. 4. We first acquire frames from a camera or video file. B) Background Removal and segmentation 5. Now we pass on these frames to background module: subtraction module. For convenience we are This will detect and remove background features to using the frame differencing for background detect the object in the scene . Algorithms like frame subtraction. differencing, averaging background method and GMM based will be a tool for such removal.[22] We are interested in analyzing human behavior.[2] . Different types of background subtracting algorithms[ 5] like Gaussian mixture model[6][11], adaptive Gaussian mixture model[4][20] and many more are available. Then we apply segmentation algorithms to separate moving objects from background. Segmentation can be carried out by Mean shift method [10], Bayesian method [8], wrapper based approach [7], etc. C) Detecting human and their behavior module: Once we get a human object in frame we can track the motion of such object in subsequent frames. We achieved this by drawing contour around the foreground object using connected component technique [13 ] of opencv.[22 ][23] Then we draw a bounding rectangle around moving object. Then we can use this rectangle parameters to detect the FIG: 1 Architecture of intelligent surveillance system. position of object in the scene. Ot her techniques like event detection using rejection based approach[3] , 6. We first acquire frames from a camera or video kalman filters[9], optical flow[22], object recognition file. methods[12][13][14], detection of salient object[15], 7. Now we pass on this frames to background shape filters[16][17][19], repository based object subtraction module. For convenience we are recognition[18] can be used. using the frame differencing for background We provide some parameters to system to detect subtraction. 591 All Rights Reserved © 2012 IJARCET
  • 3. ISSN: 2278 – 1323 International Journal of Advanced Research in Computer Engineering & Technology Volume 1, Issue 4, June 2012 8. Background subtraction separates background i. Grab current frame. Save it in an image and and gives us the frame with foreground object. name it “current frame” . 9. After background subtraction we detect object ii. Take absolute difference between current and and draw a counter using connected point background frame and save it in a “foreground concept and opencv modules. frame”. Apply threshold and erosion if needed. 10. After this we draw a bounding rectangle around iii. Display the foreground frame. Move to next the counter and save its center . frame. 11. Now we track motion of moving object using the 3. Stop when all frames are over and release all the coordinates of the centre. data types to avoid stack overflow. 12. Once we track the motion we can analyse the behavior of the person in the scene and study his Experimental results: interaction with the environment. A. Indoor scene with minimum fluctuations: III. BACKGROUND SUBTRACTION ANALYSIS This section focuses on how to isolate objects or parts of objects from the rest of the image.[22] We want to isolate those events and to be able to ignore the endless hours when nothing is changing. We start .FIG 2 Without threshold gray level image with separating foreground objects from learned background scenes. We will be dealing with three types of background subtraction methods for our analysis. We will implement frame differencing, averaging background method and mixture of Gaussian( MOG) model and study the results for different cases. FIG 3 After applying threshold : binary image TH=80 Absolute differencing method or Frame differencing: Basic idea here is to subtract one frame from another subsequent frames and label any difference that is big enough ( threshold) as Foreground .[ 1] This method mainly catches motion in the frames. For simplicity, let’s say we have three single-channel images: frame1, frame2, and Foreground. We could then use FIG 4 After applying threshold and erosion the following code to detect the magnitude (absolute technique value) of foreground differences in Foreground: Opencv’s cvAbsDiff( ) function can be used for such differencing.[22 ]. To reduce noise, we can apply threshold concept. This removes unwanted pixels which add to noise in our data. Thresholds are manually adjusted to suit our needs. cvThreshold( ) is a tool for applying thresholding. We will clean up small noise areas with cvErode( ) Function . Algorithm: 1. Grab a random initial frame and save it in an FIG 5 Outputs for different thresholds The thresholds for images from Left to right and top to bottom are image datatype and name it as “background Th=10,Th=30,Th=60,Th=80,Th=110. frame”. 2. Now for each other frame do the following 592 All Rights Reserved © 2012 IJARCET
  • 4. ISSN: 2278 – 1323 International Journal of Advanced Research in Computer Engineering & Technology Volume 1, Issue 4, June 2012 Experimental Conclusions: [Fig 2,3,4,5] all up using cvAcc() and then divide by the total number of images to obtain the mean. We can also 1. Simple gray level absolute differencing causes accumulate squared images, which will allow us to foreground detection with shadow effects and compute quickly the variance of individual pixels. noise. For finding the covariance[22 ] we can also see how 2. Thresholding reduces noise and gives binary images vary over time by selecting a specific lag and image which is more predictable. then multiplying the current image by the image from 3. Selection of thresholds depends upon lighting the past that corresponds to the given lag. Th e conditions. function cvMultiplyAcc() will perform a pixelwise 4. Erosion technique further helps reducing noise. multiplication of the two images and then add the result to the “running total” in acc:[1 ][ 22] [ 5] Our sample video shows following results:[Fig 5] Algorithm: 1. Th <30 gives foreground detection with lot of noise. i. First, we create pointers for the various scratch 2. Th>30 and Th<60 gives satisfactory results. Still and statistics-keeping images we will need along has noise contents. the way. 3. Th >60 and Th<80 gives good results without ii. we create a single call to allocate all the noise. But shows little loss of foreground data. necessary intermediate images. For convenience 4. Th> 110 Gives only foreground data with great we pass in a single image (from our video) that loss but without noise. can be used as a reference for sizing the 5. This means that 30<Th<60 is a better window intermediate images. This is done in allocate for indoor scenes to achieve better foreground images routine. detection. iii. Next we learn the accumulated background image and the accumulated absolute value of Advantages: frame-to-frame image differences. This is done in accumulatebackground routine. Simplest method for background subtraction. Less iv. We first use cvCvtScale() to turn the raw memory requirement as only initial frame is background 8-bit-per-channel, three- considered as background. Can work well for indoor colorchannel image into a floating-point three- lighting conditions well. channel image. We then accumulate the raw floating-point images into IavgF. Next, we Disadvantages: Lighting fluctuations are not handled calculate the frame-to-frame absolute difference properly.Threshold depends upon lighting image using cvAbsDiff() and accumulate that conditions.Cannot work well for outdoor scenes into image IdiffF. Each time we accumulate where background may change considerably. these images, we increment the image count Icount, a global, to use for averaging later 2. Averaging background method v. Once we have accumulated enough frames, we convert them into a statistical model of the background. That is, we compute the means and The averaging method makes use of four OpenCV deviation measures (the average absolute diff routines: [22 ] [23] erences) of each pixel: using createmodel from 1. cvAcc(), to accumulate images over time; stats routine. 2. cvAbsDiff(), to accumulate frame-to-frame image diff erences over time; vi. setHighThreshold() and setLowThreshold() are 3. cvInRange(), to segment the image (once a utility functions that set a threshold based on the background model has been learned) into frame-to-frame average absolute diff erences. foreground and background regions; The call setHighThreshold(7.0) fi xes a threshold 4. cvOr(), to compile segmentations from different such that any value that is 7 times the average color channels into a single mask image. frame-to-frame absolute diff erence above the average value for that pixel is considered foreground; likewise, setLowThreshold(6.0) sets Finding Mean , variance and covariance: a threshold bound that is 6 times the average frame-to-frame absolute difference below the To compute a mean[22 ] value for each pixel across a average value for that pixel. large set of images, the easiest method is to add them 593 All Rights Reserved © 2012 IJARCET
  • 5. ISSN: 2278 – 1323 International Journal of Advanced Research in Computer Engineering & Technology Volume 1, Issue 4, June 2012 vii. Once we have our background model, complete 2. Needs more memory and time to execute than with high and low thresholds, we use it to previous method. segment the image into foreground (things not “explained” by the background image) and the 3. Advanced background subtraction -Mixture background (anything that fits within the high of Gaussian: and low thresholds of our background model). Segmentation is done by calling backgrounddiff Gaussian Function: routine. viii. This function first converts the input image I In mathematics, a Gaussian function (named after (the image to be segmented) into a floating- Carl Friedrich Gauss) is a function of the form: point image by calling cvCvtScale(). We then convert the three-channel image into separate one-channel image planes using cvSplit(). These color channel planes are then checked to see if they are within the high and low range of the for some real constants a, b, c > 0, and e ≈ average background pixel via the cvInRange() 2.718281828 (Euler's number).The graph of a function, which sets the grayscale 8-bit depth Gaussian is a characteristic symmetric "bell curve" image Imaskt to max (255) when it’s in range shape that quickly falls off towards plus/minus and to 0 otherwise. For each color channel we infinity. (FIG 7 )The parameter a is the height of the logically OR the segmentation results into a curve's peak, b is the position of the centre of the mask image Imask, since strong differences in peak, and c controls the width of the "bell". Gaussian any color channel are considered evidence of a functions are widely used in statistics where they foreground pixel here. Finally, we invert Imask describe the normal distributions, in signal processing using cvSubRS(), because foreground should be where they serve to define Gaussian filters, in image the values out of range, not in range. The mask processing where two-dimensional Gaussians are image is the output result. used for Gaussian blurs, and in mathematics where ix. Lastly we de-allocate all the variables and they are used to solve heat equations and diffusion pointers. equations and to define the Weierstrass transform. Running Gaussian average was first proposed by Wren , Azarbayejani, Darell, pentland in 1997 which fits one Gaussian distribution (μ, σ) over the histogram.[1]Generalized Gaussian model was proposed by H.Kim, R. Sakamoto et all[1 ][ 4] to cope up with the background changes and shadows using this model. In Mixture of Gaussian ( MOG) model background is termed as parametric frame of values where each pixel is represented with number of Gaussian Functions as Probability Distribution FIG: 6 Averaging background method (Background is function.[6 ] average of first 10 frames) Advantages: 1. Builds a unique model of background depending upon earlier “n” frames 2. Predicts the foreground based on lighting condition changes in consecutive frames. 3. Works well if more frames are taken for averaging the background. Disadvantages: 1. More complex calculations. FIG: 7 Normalized Gaussian curves with expected 594 All Rights Reserved © 2012 IJARCET
  • 6. ISSN: 2278 – 1323 International Journal of Advanced Research in Computer Engineering & Technology Volume 1, Issue 4, June 2012 value μ and variance σ2. The corresponding Algorithm: parameters are a = 1/(σ√(2π)), b = μ, c = σ[25 ] i. For background subtraction Mixture of Gaussian basics: CvGaussBGStatModelParams is used where it takes some parameter like bg_threshold and Sometimes different objects are likely to appear at std_threshold. This two parameters are the same pixel (i,j) location. For example, in an responsible to make difference between outdoor scene with trees partially covering a background and foreground. building: a same (i,j) pixel location will show values ii. There is another function called from tree leaves, tree branches, and the building cvCreateGaussianBGModel which takes video itself. Other examples can be easily drawn from frame to create standard background model snowing, raining, or watching sea waves from a according to parameters. beach. In these cases, a single valued background is iii. After that, extracting foreground not an adequate model. Stauffer and Grimson [1 ] cvUpdateBGStatModel function is used where it [22] [26] raised the case for a multi-valued takes video frame and background model to background model able to cope with multiple compute differences frame by frame. background objects. Stauffer and Grimson describe iv. As an output bgModel->background represent the probability of observing a certain pixel value, x, background of the video and bgModel- at time t by means of a mixture of Gaussians. All >foreground shows foreground of the video weights wi are updated (or normalised) at every new frame. In practical cases, K is set to be between 3 and Our chosen Gaussian parameters: 5. Gaussians are multi-variable to describe red, green and blue values. If these values are assumed params->win_size=2; - Learning rate; independent, the co-variance matrix, ∑i,j simplifies alpha = 1/CV_GBG_WINDOW_SIZE . to diagonal.[26 ] params->n_gauss=5; = K = number of Gaussians in mixture . At each t frame time, two problems must be params->bg_threshold=0.7; threshold sum of simultaneously solved a) assigning the new observed weights for background test . value, x,, to the best matching distribution and b) params->std_threshold=3; lambda=2.5 is estimating the updated model parameters. These 99% params->minArea=15; concurrent problems can he solved by AN params->weight_init=0.25; EXPECTATION MAXIMIZATION (EM) algorithm. params->variance_init=30; AN EXPECTATION MAXIMIZATION works on CV_BGFG_MOG_SIGMA_INIT*CV_BGFG_MOG_ the buffer of the last n frames. However, as this SIGMA_INIT would prove extremely costly, the matching is approximated in these terms Results: The first in ranking order is accepted as a match for . furthermore the parameters are updated only for this matching distribution and by using simple online cumulative averages . At every new frame, some of the Gaussians “match” the Fig:8 mixture of Gaussians background subtraction current value (those at a distance < 2.5 σi ): for them, method μi ,σi are updated by the running average. If no match is found the last ranked distribution is replaced by a We kept all parameters constant and only changed new one centered in , with low weight and high one parameter at a time to see the probable changes variance. The mixture of Gaussians actually models in the output and best values for the parameters. We both the foreground and the background All concluded following results from our observations. distributions are ranked according to their ωi /σi and the first ones chosen as “background” .[26 ] 595 All Rights Reserved © 2012 IJARCET
  • 7. ISSN: 2278 – 1323 International Journal of Advanced Research in Computer Engineering & Technology Volume 1, Issue 4, June 2012 Mixture of gaussian parameter analysis: Sample output Paramter Observed best value range threshold sum of weights for 0.1 to 0.7 Better Good background test (bg_threshold) Best Standard threshold( 2 to 5 std_threshold) Number of Gaussians(n_gauss) 3 to 6 ( 3 is better) Alpha-learning rate 0.3 to 0.7( 0 to values) IV. SEGMENTATION USING CONTOURS AND MOTION Advantages: [Fig 8] TRACKING. 1. Best background prediction. 2. Model can be adapted to specific scenes by A contour is a list of points that represent, a curve in changing the parameters. an image. [22][23]A contour is represented in 3. Least amount of noise incurred when parameters OpenCV by a CvSeq sequence that is, a sequence of are adjusted accordingly. points. The function cvFindContours() [22] 4. Works well in all lighting conditions. computes contours from binary images. The return value of cvFindContours() is the total number of contours found. The first argument is the input Disadvantages: image. The next argument, storage, indicates a place where cvFindContours() can find memory in which 1. Takes largest time and memory among all the to record the contours. In a binary image white methods implemented by us. regions are labeled as contour and black lines as 2. Parameter adjustment is a trial and error method . holes . One of our most basic tasks is drawing a 3. Needs high processing speeds for real time contour on the screen. For this we have environments. cvDrawContours( ). Summary of our analysis: Motion Tracking:[fig 9] Frame Averaging Mixture Algorithm: differencing background of method Gaussians method A) Find a rectangle around counter: 1. BoundingRect() Calculates up-right bounding rectangle of point set. Here we use it to draw a Compl Very less Medium More bounding rectangle around the counter we just exity got in above step. Memor Very less Medium More 2. BoundingRect() takes two parameters: y 1. Counter type and name= here we have require taken counter type of cvseq type. ments 2. Update flag= update flag is zero in case of cvseq type. Executi Fast Medium Very slow on time B) Finding the centre of rectangle: Param Threshold is Few Needed . Many eter only needed. parameters 1. Boundingrect function will return us with an selectio are needed object say BNDRECT. n 2. BNDRECT has following information X coordinate of left upper vertex Y coordinate of left upper vertex Width Height 596 All Rights Reserved © 2012 IJARCET
  • 8. ISSN: 2278 – 1323 International Journal of Advanced Research in Computer Engineering & Technology Volume 1, Issue 4, June 2012 3. Calculate Point 1= (X, Y) Loitering is defined as the presence of an individual 4. Calculate point2=(X+width, Y+height) in an area for a period of time longer than a given 5. Calculate Centre on X axis(here for our time threshold. [ 2][Fig 11]For example loitering can convenience)= avgX=( point1.x+point2.x)/2 be possible at museums, hotels, shops, at car parking, 6. This avgX can be used as a parameter to detect hospitals, monuments and business places. Such the position of moving object we have detected. suspicious events where a person stands at an area for 7. Lastly for visibility draw a rectangle using point a longer duration can be caught by providing such 1 and point 2 on moving counter. thresholds and defining the area of importance within the scene. Results: FIG:9 First frame shows raw video, second shows foreground detected, third shows contour and fourth shows a rectangle drawn around the person in the running video. V. BEHAVIOR DETECTION – DETECTING SUSPICIOUS ACTIVITY. Now as we have done with detection and tracking of FIG:11 first frame shows person enters an area( marked by moving object we can predict the motion and green rectangle). This area is area of strategic importance. behavior of moving object. Second frame shows person still inside the area. The third frame shows a “loitering” alert generated when the person stays within the area beyond a certain threshold. 1. Trip Wire detection: We will take a simple scenario where one is not supposed to enter a prohibited area. [ 22][Fig 10]We will take this problem as Crossing the green line and As soon as we get a suspicious activity we will term or being present on prohibited area will be termed as the frame as suspicious and display it in a window. intrusion. We will be checking avgX value and We will extract the intruder image and save it using monitoring it. If the value is greater or equal to any bounding rectangle coordinate. Along with coordinates of trip wire we will term the action as suspicious frames for further use we can also save the intrusion: timing, date, and location data for further ease during investigations. Result: 3. Abandoned baggage or boxes: Here the system continuously checks for motion in the frames, as well as checks for objects that are not present before. Any stationary object not present before can act as a threat for the public. So such baggage or box can be termed as harmful and can be highlighted.[Fig 12] Face Detection: Fig: 10The person enters a prohibited area at a particular time and so an alarm is generated. If face is detectable [24 ] we will detect and save it too. These suspicious frames, intruders and face data 2. Loitering activity: can be used for future use in case of emergency. 597 All Rights Reserved © 2012 IJARCET
  • 9. ISSN: 2278 – 1323 International Journal of Advanced Research in Computer Engineering & Technology Volume 1, Issue 4, June 2012 FIG: 13 Faces are detected and rectangle are drawn bounding the faces. CONCLUSION FIG: 12 The first frame shows a person carrying a bag, second shows the person keeping the bag on the floor, Thus we have shown that use of present day the third frame shows bag being detected as machine vision algorithms along with opencv can abandoned object by the system. lead us to develop an intelligent surveillance infrastructure. Yet a lot of work needs to be done in The face data is detectable using HAAR classifiers [ refinement of this behavior detection process. More 22] and viola jones algorithms[ 22][24] provided by and more behavior needs to be considered and opencv library. [Fig 13]The haar classifier used is identified. haarcascade_frontalface_alt2.xml file which can detect frontal faces in an image. Whenever a face is ACKNOWLEDGMENT detected we draw a rectangle around it and save it. Future work may comprise of supplying such face I would like to thank all those people whose support data to a face matching system which will find out and cooperation has been an invaluable asset during correct matching faces of the suspicious person. the course of this paper. I would also like to thank my Future work also focuses mainly on creating concrete Guide Prof. Dr. B.K.Lande and our dept head Dr: classifiers for detecting human and non human B.B. MEshram for guiding me throughout this paper objects in a scene. and giving it the present shape. It would have been impossible to complete the paper without their support, valuable suggestions, criticism, Algorithm: encouragement and guidance. 1. Create a cascade of haar classifiers[22 ] using I am also grateful for all other teaching and non- haarcascade_frontalface_alt2.xml file and teaching staff members of the Computer Technology CvHaarClassifierCascade* cascade = for directly or indirectly helping us for the (CvHaarClassifierCascade*)cvLoad( completion of this project and the resources provided. "haarcascade_frontalface_alt2.xml" ); 2. Detect objects using cvHaarDetectObjects( ); REFERENCES 3. This will return number of objects and the objects will be stored in memory. [1] Niraj Gadhe, Dr:B.K . Lande, VJTI, IOSR Journal of 4. For each object detected( each face detected) Engineering (IOSRJEN),Vol. 2 Issue 2, Feb.2012 draw a rectangle bounding the face. [2] Joshua Candamo, Matthew Shreve, Member, IEEE, Dmitry B. Goldgof, Fellow, IEEE, Deborah B. Sapper, and Rangachar Kasturi, Fellow, IEEE “Understanding Transit Scenes: A Survey on Human Behavior-Recognition Algorithms” in IEEE Result: TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 11, NO. 1, MARCH 2010 [3] Margarita Osadchy and Daniel Keren “A Rejection-Based Method for Event Detection in Video” in IEEE TRANSACTIONS 598 All Rights Reserved © 2012 IJARCET
  • 10. ISSN: 2278 – 1323 International Journal of Advanced Research in Computer Engineering & Technology Volume 1, Issue 4, June 2012 ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, Conference on Neural Information Processing (ICONIP'O2), Vol. VOL. 14, NO. 4, APRIL 2004 1 [4] H. Kim, R. Sakamoto, I. Kitahara, T. Toriyama and K. Kogure “Background subtraction using generalized Gaussian family [17] Cosmin Grigorescu, Student Member, IEEE, and Nicolai model” in ELECTRONICS LETTERS 31st January 2008 Vol. 44 Petkov “Distance Sets for Shape Filters No. 3 and Shape Recognition” in IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 12, NO. 10, OCTOBER 2003 [5] Shahrizat Shaik Mohamed, Nooritawati Md Tahir, Ramli Adnan “Background Modelling and Background Subtraction [18] Chensheng Wang, Fei Wang “A Knowledge-based Strategy Performance for Object Detection” in 2010 6th International for Object Recognition and Reconstruction” in 2009 International Colloquium on Signal Processing & Its Applications (CSPA) Conference on Information Technology and Computer Science [6] Dar-Shyang Lee, Member, IEEE “Effective Gaussian Mixture [19] Junqiu Wang, Member, IEEE, and Yasushi Yagi, Member, Learning for Video Background Subtraction “ in IEEE IEEE “Integrating Color and Shape-Texture Features for Adaptive TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE Real-Time Object Tracking” in IEEE TRANSACTIONS ON INTELLIGENCE, VOL. 27, NO. 5, MAY 2005 IMAGE PROCESSING, VOL. 17, NO. 2, FEBRUARY 2008 [7] Michael E. Farmer, Member, IEEE, and Anil K. Jain, Fellow, [20] Zoran Zivkovic “Improved Adaptive Gaussian Mixture Model IEEE “A Wrapper-Based Approach to Image Segmentation and for Background Subtraction” in Proc. ICPR, 2004 Classification” in IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 14, NO. 12, DECEMBER 2005 [21] Implementing the Viola-Jones Face Detection [8] A. Colmenarejo, M. Escudero-Vin˜olo and J. Besco´s “Class- Algorithm Ole Helvig Jensen Kongens Lyngby Technical driven Bayesian background modelling for video object University of Denmark, 2008, segmentation” in ELECTRONICS LETTERS 1st September 2011 Vol. 47 No. 18 [22] Book on “ Learning Open Cv – Computer vision with Open CV library” by Gary Bradski and Adrian Kaebler – o’reilly [9] Jesse Scott et all “Kalman Filter Based Video Background Estimation” in 2007IEEE [23]Book on “OpenCV 2 Computer Vision Application [10] Madhurima , et all “Object tracking in a video sequence using Programming Cookbook” by Robert Laganière-PACKT Mean-Shift Based Approach” in IJCEM International Journal of PUBLISHING. Computational Engineering & Management, Vol. 11, January 2011 [11] Sen-Ching S. Cheung and Chandrika Kamath “Robust [24] Image Processing Project on “Face Recognition Techniques techniques for background subtraction in urban traffic video” “by Jorge Orts , Univerity of Wisconsin, Madison. [12] Padmini Jaikumar, Abhishek Singh and Suman K Mitra “Background Subtraction in Videos using [25]www.wikipedia.org Bayesian Learning with Motion Information” in 2008. [13] Scott Helmer and David G. Lowe “Object Class Recognition [26] “MIXTURE OF GAUSSIANS USING OPENCV” AS IN HTTP://MMLAB.DISI.UNITN.IT/WIKI/INDEX.PHP/MIXTURE_OF_GAUS with Many Local Features” in Proceedings of the 2004 IEEE SIANS_USING_OPENCV. Computer Society Conference on Computer Vision and Pattern Recognition Workshops. [27] IMAGE AND VIDEO DATABASES OF CVONLINE: IMAGE [14] Lun Zhang, Stan Z. Li, Xiaotong Yuan and Shiming Xiang DATABASES “Real-time Object Classification in Video Surveillance Based on HTTP://HOMEPAGES.INF.ED.AC.UK/RBF/CVONLINE/IMAGEDBASE.H Appearance Learning” in 2007 IEEE TM#RETRIEVE [15] Tie Liu, Zejian Yuan, Jian Sun, Jingdong Wang, Nanning [28] Surveillance databases – VISOR- video urviellance online Zheng, Fellow, IEEE,et all “Learning to Detect a Salient Object” repository- in IEEE TRANSACTIONS ON PATTERN ANALYSIS AND http://imagelab.ing.unimore.it/visor/video_videosInCategory.asp?i MACHINE INTELLIGENCE, VOL. 33, NO. 2, FEBRUARY dcategory=14 2011 [16] Zhiyong Wang, Zheru Chi and David Feng “STRUCTURAL REPRESENTATION AND BPTS LEARNING FOR SHAPE CLASSIFICATION” in Proceedings of the 9th International 599 All Rights Reserved © 2012 IJARCET