SlideShare a Scribd company logo
1 of 4
Download to read offline
A PHYSICAL APPROACH TO MOVING CAST SHADOW DETECTION

                                                  Jia-Bin Huang and Chu-Song Chen

                                        Academia Sinica, Institute of Information Science
                                                    Taipei, Taiwan, R.O.C.


                             ABSTRACT                                          light scattered by the sky, and other colored light from nearby
This paper presents a physics-based approach capable of detecting              surfaces (i.e. color bleeding). These lights can have significantly
cast shadows in video sequence effectively. We develop a new phys-             different spectral power distributions (SPD). Therefore, the RGB
ical model of cast shadows without making prior assumption of the              value of a shaded pixel may not attenuate linearly along the line
spectral power distribution (SPD) of the light sources and ambi-               between the value of the corresponding background and the origin.
ent illumination in the scene. The background appearance varia-                     Little attention has been directed toward this problem before.
tion caused by cast shadows is characterized as the interaction of             Nadami et al. [7] addressed this non-linearity by dichromatic reflec-
the blocked light sources and the background surface reflectance.               tion model for both the sun and the sky illuminations in an outdoor
We then take advantage of the statistical prevalence of cast shadows           scene. However, the method is designed only for an outdoor envi-
to learn and update the shadow model parameters using the Gaus-                ronment. Recently, a more general shadow model was presented in
sian mixture model (GMM) over time. The proposed algorithm is                  [8]. This model introduced an ambient illumination term which de-
completely unsupervised and can adapt to specific environment with              termines a direction in the RGB space where the background values
complex illumination condition as well as changing shadow condi-               under cast shadows can be found. Since the ambient term may have
tions. Experimental results on three challenging sequences demon-              spectrum which differs from the incident light sources, the values of
strate the effectiveness of the proposed method.                               the shaded pixels may not be proportional to the values under direct
                                                                               light. Nonparametric density estimation is used to model the surface
    Index Terms— Visual surveillance, Object detection, Moving                 variation under cast shadows in an unsupervised way. By provid-
Cast Shadow Detection                                                          ing a better description of cast shadows on background surfaces, the
                                                                               shadow model presented in [8] considerably improved the previous
                        1. INTRODUCTION                                        approaches based on linear model. However, in [8] the SPDs of all
                                                                               light sources are assumed having the same profile but with different
Detection of moving objects is an important low-level vision task              power factor.This assumption may not be hold for many of the real-
for many vision-based applications. One of the most challenging                world cases. In addition, the ambient illuminations are assumed to
problems in these applications is identifying cast shadows in video.           be constant, but they may be slightly changed due to circulation of
Cast shadows are often misclassified as parts of foreground regions,            foreground objects.
resulting in distortion of the true shape and color properties of the               In this paper, we extend the shadow model presented in [8] by
target objects. An effective shadow detection method is thus neces-            releasing these assumptions. The proposed shadow model is more
sary for accurate foreground segmentation.                                     general and can be applied to real-world cases more appropriately.
     Moving cast shadows are caused by the occlusion of light                  However, a more general model may suffer from a longer period of
sources. Shadows reduce the total energy incident at the background            time for learning the model parameters. To address this problem,
surfaces where the light sources are partially or totally blocked by           we introduce the confidence-rated Gaussian mixture learning which
the foreground objects. Hence, points under shadows have lower                 accelerates the parameters convergent rate. Moreover, we observe
luminance values but similar chromaticity values compared to their             that shadows on the light/shadow borders show different behavior
reference background points.                                                   from shadows inside the shaded region. The shadows detection rate
     Many works have been proposed for cast shadow detection.                  can be further improved by addressing this issue.
Most of them are based on the assumption that the RGB values of                     The remainder of this paper is organized as follows. We first
shadow pixels will fall on the line between the illuminated values             propose in Section 2 a new physical model for describing appearance
and the origin in the RGB color space (linear model). This linear              variation of a background value under cast shadows. Based on the
model has been employed in different color spaces [1, 2, 3]. One               proposed model, we present an unsupervised learning approach to
main limitation of these methods is that they require explicit tuning          learn the behavior of cast shadows on surfaces by taking advantage
in a large set of parameters for each new scene and are difficult to            of the statistical prevalence of shadows in Section 3. In Section 4,
work for on-line applications. To adapt to environment changes,                we derive the posterior probability of cast shadows and foreground
statistical learning methods have been employed to describe the cast           given an observed sample. Experimental results are presented in
shadows [4, 5, 6]. However, the linear model assumption may not                Section 5. Section 6 concludes the paper.
be fulfilled in a real-world environment. For instance, in an outdoor
scene the illumination may consist of direct sunlight, the diffused
                                                                                      2. PHYSICAL MODEL FOR CAST SHADOWS
     This research was supported by the National Science Council of Taiwan
under Grant No. NSC 95-2221-E-001-028-MY3 and 97-EC-17-A-02-S1-                According to the Lambertian model, the camera sensor response
032.                                                                           gk (p), k ∈ {R, G, B} at pixel p depends on the intensity func-




978-1-4244-2354-5/09/$25.00 ©2009 IEEE                                   769                                                        ICASSP 2009
from the contribution of the blocked incident and ambient illumina-
                                                                             tions. We then develop a 3D color feature of cast shadows xs,t (p) =
                                                                             [αt (p), θt (p), φt (p)]T to encode this information, where αt (p) rep-
                                                                             resents the illumination attenuation and θt (p) and φt (p) indicate the
                                                                             direction of the vector vt (p) in spherical coordinate. Similar to Hor-
                                                                             prasert et al. [1], we define the illumination attenuation αt (p) by
                                                                                                                  ||vt (p)||
                                                                                                     αt (p) =                ,                   (3)
                                                                                                                ||BGt (p)||
                        (a)                    (b)
                                                                             where BGt (p) is the background value at the pixel p. The illumi-
Fig. 1. Physical cast shadow model. (a) Linear attenuation model             nation attenuation, which is smaller than one, describes the strength
and the proposed model. The linear model will fail when the SPD              of brightness of the blocked lights with respect to that of the back-
of ambient illumination does not consistent with that of the light           ground value. The direction of the vector vt (p), the chromaticity
sources. (b) Background appearance variation due to cast shadows.            line direction, is determined by θt (p) and φt (p):
                                                                                                       vG (p)
                                                                                                        t                         vB (p)
                                                                                    θt (p) = arctan(          ), φt (p) = arccos( t         ),   (4)
tion of the light sources E(λ, p), the reflectance of an object surface                                 vR (p)
                                                                                                        t                        ||vt (p)||
ρ(λ, p), and sensor spectral sensitivity Sk (λ):                             where the superscript R, G, B indicate the components in the RGB
                                                                             space. We can use this 3-D color feature vector vt (p) to describe the
                gk (p) =      E(λ, p)ρ(λ, p)Sk (λ)dλ.               (1)      appearance variation induced by the blocked light sources on shaded
                                                                             regions. Unlike [8], our model takes the variation of chromaticity di-
We describe the illumination incident to a surface point p by N light        rection into consideration and thus better describes the cast shadows
sources and lights reflected from M ambient objects (the position p           on background surface.
is ignore):                                                                       Since different foreground objects often block the light sources
                  N                        M
                                                                             in a similar way, cast shadows on background surfaces are thus sim-
        E(λ) =         Eincident,n (λ) +         Eambient,m (λ).    (2)      ilar and uncorrelated with foreground objects. We present in the fol-
                 n=1                       m=1
                                                                             lowing section how to learn the cast shadow model in a completely
                                                                             unsupervised way.
For a color camera with a linear response, the response to a specific
illumination follows a line in the RGB space. We show our model
                                                                                 3. LEARNING AND DETECTING CAST SHADOWS
in Fig. 1(a), where BGA is the ambient term from the contribution
of all ambient lightings and BG is the background surface value.             Shadow detection methods with fixed parameters can not adapt to
The zigzag lines starting from BGA and ending at BG represent                time-varying illumination. In this paper, we use Gaussian mixture
the response from the incident light sources. Based on the additive          model (GMM) [9] to model cast shadows by using scene activities.
color model, their net effect forms the vector from BGA to BG.               We first apply weak shadow detector to moving pixels detected by
Note that in this model, the ambient illumination can have different         background model to select possible shadow samples. The color fea-
SPDs from that of the total contribution of incident light sources (i.e.     tures of these possible shadow pixels are then learned by GMM over
the direction of BGA may not be consistent with the direction from           time. Then, we can detect shadows by using the learned shadow
BGA to BG). In addition, the SPD of individual light source may              model.Note that our approach is pixel-based, model the SPDs of
differ from that of other light sources.                                     ambient lighting and light sources locally, and thus can deal with
     In comparison, in [8] the SPDs of the incident light sources are        environment with complex illumination conditions.
assumed sharing the same profile. That is, the first term in (2) de-
generates into Eincident (λ) N γn , where γn is the scaling fac-
                                  n=1
tor of the nth light sources. Moreover, the ambient illumination is          3.1. Weak Shadow Detector
assumed to be constant. Under these two assumptions, all shadow              To model the cast shadows, impossible shadow samples that belong
values are expected to be found on the line between background BG            to background and foreground should be excluded. Therefore, we
and the ambient value BGA . This approach may fail to deal with              apply a weak shadow detector that evaluates every moving pixels de-
environment with complex illumination conditions.                            tected by the background model to filter out some impossible shadow
     Given a surface point, the background surface value variation           samples. Since cast shadows on surface reduce luminance values and
under cast shadows is resulted from the occlusion of lights by fore-         change the saturation, we define the potential shadow values fall into
ground objects. The blocked lights may contain partial or total direct       the conic volume around the corresponding background color [4].
incident light sources and some reflected lights from nearby ambient               The weak shadow detector is applied to moving regions detected
objects. Here, some of the incident lights will disappear since they         by the background model to obtain possible shadow points. We call
are blocked by foreground objects. The first term in (2) is then re-          these possible shadow points as “shadow candidates”, which form
duced to N Eincident,n (λ) − n∈blocked Eincident,n (λ). The
             n=1                                                             the set Q. At the pixel level, we model xs,t (p), p ∈ Q using the
ambient lights are also changed slightly due to the circulation of           GMM to learn the color feature of shadows.
foreground objects. Therefore, according to (1), the vector from the
shadow value (i.e. SD in Fig. 1) to its corresponding background             3.2. Spatial Information
value can be interpreted as the interaction between the blocked lights
and surface reflectance (see Fig. 1(b)).                                      With the proposed physical shadow model, we can accurately de-
     For a pixel p at time t, the vector from shadow to background           scribe the appearance variation of background surface under shad-
value is denoted as vt (p). The vector vt (p) is the net response            ows at the pixel level. However, two shortcomings remain. One



                                                                       770
than ambient illumination, blocking them induces a sharp shadow.
                                                                             In this case, shadows at light/shadow border show different behavior
                                                                             from shadows inside the shaded region. This is caused by the in-
           (a)                                                               sufficient resolution of CCD camera. Pixels at light/shadow border
                                                                             may receive part of energy from light sources. We demonstrate this
                                                                             observation in Fig. 2. We first collected shadow samples throughout
                                                                             the sequence “Highway I”. Shadow border pixels can be detected
                                                                             from background posterior image where large gradient intensities
           (b)                                                               are present. We classify these shadow samples into two groups:
                                                                             “inside” and “at border”. Fig. 2(a) presents that the illumination
                                                                             attenuation have significantly different distribution for shadows at
                                                                             border and inside. On the other hand, Fig. 2(b)(c) show that the
           (c)                                                               distribution on angle information are very similar. The reason is
                                                                             that neighboring pixels often receive incident energies with similar
Fig. 2. Color feature value distribution of shadows inside (in red)          SPDs.Consequently, we propose to address shadows at border by
and at the light/shadow border (in blue). Samples are taken from             evaluating only their angle information.
the sequence “Highway I”. (a) Illumination attenuation αt (p) (b)(c)
angle information: θt (p), and φt (p), respectively.
                                                                                            4. POSTERIOR PROBABILITIES

                                                                             In this section, we present how to derive the posterior probabilities of
is that the proposed shadow model may suffer from slow learning              cast shadows and foreground given the observed sample xt (p) using
of model parameters if motion is rare on a pixel. With the con-              the proposed shadow model and spatial information.
ventional Gaussian mixture learning [9], training the pixel-based
shadow model requires a long time while the illumination conditions
                                                                             4.1. Cast Shadow Posterior
should remain stable. The other problem arises from foreground ob-
jects having similar colors as background, which may be misclassi-           The shadow posterior is first computed by decomposing P (SD|xt (p))
fied as cast shadows. Therefore, we propose to incorporate spatial            over (BG, FS) domain, where F S indicates moving pixels (real
information for improving the shadow discriminative ability and the          foreground and cast shadows). Since P (SD|xt (p), BG) = 0, the
convergent rate of the model parameters.                                     decomposition is
     In addition to the illumination attenuation, when a surface point
is shaded, the gradient intensity around the point will also be smaller              P (SD|xt (p)) = P (SD|xt (p), F S)P (F S|xt (p)),           (6)
than that of the illuminated surface. Consequently, points having            where P (F S|xt (p)) = 1 − P (BG|xt (p)) can be directly com-
larger gradient intensities than background are more likely to be            puted from the background model. Second, we remove pixels that
considered as foreground objects. We thus penalize these samples             are definitely foreground and consider only shadow candidate (SC).
by lessening the learning rate in this situation. We define a con-            Also, we penalize pixels whose gradient intensities are amplified
fidence value of shadow using spatial gradient intensities of back-           instead of reduced by multiplying ωt (p): P (SD|xt (p), F S) =
ground |∇(Bt )| and current image |∇(It )|:                                  P (SD|xt (p), F S, SC) ∗ ωt (p). Here, the color feature vector
                             ε + |∇(Bt (p))|                                 xc,t (p) is the sufficient statistics for the observed sample xt (p).
           ωt (p) =                                     ,           (5)      Therefore, we have P (SD|xt (p), F S, SC) = P (SD|xc,t (p), F S, SC).
                      ε + max{|∇(It (p))|, |∇(Bt (p))|}
                                                                             As mentioned in section 3, we learn the cast shadows using the
where ε is a smooth term to alleviate the effect of noise in textureless     GMM [9]. The posterior probability can be expressed in terms of
regions. Thus, we can discriminate foreground objects with similar           the mixture components in GMM:
color as background from shadow with the aid of gradient informa-
tion. To improve the convergent rate, we use ωt (p) as the confi-               P (SD|xc,t (p), F S, SC) =          P (SD|Gk )P (Gk |xc,t (p)), (7)
                                                                                                               k
dence value to update the GMM at pixel p through confidence-rated
learning. In [10], Lee has shown that fast updating can be achieved          where Gk is the kth Gaussian state in the mixture. Given the ob-
by combining recursive filter type of learning with incremental EM            servation that color change due to objects has higher variance com-
in background modeling. We propose to weight these two learning              pared to the color change due to cast shadows, Gaussian states with
scheme using the confidence value ωt (p) extracted from spatial do-           higher prior probabilities and smaller variances would be considered
main [11]. If ωt (p) is large, the learning rate depends more on the         as shadows. Therefore, we approximate P (SD|Gk ) using logistic
formulation of incremental EM algorithm, which converges quickly             regression similar to that in [10] for background subtraction.
with few observations. On the other hand, if ωt (p) is small, the
learning rate follows the recursive type of learning to maintain model       4.2. Foreground Posterior
stability.
                                                                             Computing foreground posterior probability is much easier. Given
3.3. Shadows at Light/Shadow Border                                          a pixel p at time t, we first compute the background posterior
                                                                             P (BG|xt (p)) from the background model. Then, we can ob-
Once the shadow model for a pixel p is built, we can estimate                tain shadow posterior probability P (SD|xt (p)) using the learned
the likelihood of the observation belong to shadows by checking              shadow model. Based on the probability theory, the foreground
whether the 3-D color feature vector xs,t (p) is associated with one         posterior can be obtained as:
of the working Gaussian distribution in the mixture. However, if the
light sources are clustered in the scene and are relatively brighter               P (F G|xt (p)) = 1 − P (BG|xt (p)) − P (SD|xt (p)).           (8)




                                                                       771
(a)                         (b)                        (c)

                                                                                Fig. 4. Effect of shadows at shadow/light border (a) Original frame
                                                                                of sequence ”Highway I”. (b)(c) Foreground posterior without/with
                                                                                considering shadows at shadow/light border.


        (a)                (b)                (c)                (d)                                       6. CONCLUSION

Fig. 3. Visual results of the proposed algorithm. (a) Original                  In this paper, we have introduced a novel physical model for cast
frame. (b) Background posterior P (BG|xt (p)). (c) Shadow pos-                  shadow detection in video sequences. Qualitative and quantitative
terior P (SD|xt (p)). (d) Foreground posterior P (F G|xt (p)).                  evaluation of the physical shadow model validate that our approach
                                                                                is more accurate in describing background surface variation under
                                                                                cast shadows. Spatial information is used to help discriminat-
     Table 1.   Quantitative results on surveillance sequences                  ing foreground with similar chroma from cast shadows as well
 Sequence         Highway I          Highway II          Hallway                as improving the convergent rate of the shadow model through
   Method        η%        ξ%       η%       ξ%        η%      ξ%               confidence-rated learning. We also address the problem of shadows
  Proposed      72.34 84.98 72.70 79.89 71.69 88.25                             at light/shadow border where these shadows have different behavior
 Kernel [8]     70.50 84.40 68.40 71.20 72.40 86.70                             from other regions. Note that the proposed approach is pixel-based;
   LGf [6]      72.10 79.70           -        -        -       -               a contextual model that incorporate spatial and temporal coherence
 GMSM [5]       63.30 71.30 58.51 44.40 60.50 87.00                             can further enhance the performance.

                                                                                                            7. REFERENCES
                 5. EXPERIMENTAL RESULTS                                         [1] T. Horprasert, D. Harwood, and L. S. Davis, “A statistical approach
                                                                                     for real-time robust background subtraction and shadow detection,” in
We validate the proposed method on three challenging video se-                       ICCV Frame-rate Workshop, 1999.
quences. Both visual results and quantitative measurement are                    [2] R. Cucchiara, C. Grana, M. Piccardi, and A. Prati, “Detecting moving
presented. We compare our algorithm quantitatively with previous                     objects, ghosts, and shadows in video streams,” IEEE Trans. PAMI,
shadow detection methods when results are available. The effect of                   vol. 25, no. 10, pp. 1337–1342, 2003.
considering shadows at light/shadow border is also present.                      [3] E. Salvador, A. Cavallaro, and T. Ebrahimi, “Cast shadow segmentation
                                                                                     using invariant color features,” CVIU, vol. 95, no. 2, pp. 238–259, 2004.
5.1. Assessing Shadow Detection Accuracy                                         [4] F. Porikli and J. Thornton, “Shadow flow: a recursive method to learn
                                                                                     moving cast shadows,” in ICCV, Oct. 2005, pp. 891–898 Vol. 1.
We show in Fig. 3 the visual results of three video sequences from
the proposed algorithm. Fig. 3(a) shows one frame from the se-                   [5] N. Martel-Brisson and A. Zaccarin, “Learning and removing cast shad-
                                                                                     ows through a multidistribution approach,” IEEE Trans. PAMI, vol. 29,
quence. The background, shadow, and foreground posterior are                         no. 7, pp. 1133–1146, 2007.
given in Fig. 3 (b)(c)(d), respectively. We can see that in all three
test sequences the cast shadows are correctly detected.                          [6] Z. Liu, K. Huang, T. Tan, and L. Wang, “Cast shadow removal com-
                                                                                     bining local and global features,” in CVPR, 2007, pp. 1–8.
     We evaluate our method through the quantitative metrics (shadow
detection η% and discrimination rate ξ%) proposed in [12]. For ex-               [7] S. Nadimi and B. Bhanu, “Physical models for moving shadow and
                                                                                     object detection in video,” IEEE Trans. PAMI, vol. 26, no. 8, pp. 1079–
act equations, please refer to [12]. We report our results in Table
                                                                                     1087, 2004.
1 and compare them with recent works using statistical method to
detect shadows. Our method can achieve higher shadow detection                   [8] N. Martel-Brisson and A. Zaccarin, “Kernel-based learning of cast
                                                                                     shadows from a physical model of light sources and surfaces for low-
and discrimination rate than the state-of-the-art approaches, where
                                                                                     level segmentation,” in CVPR, June. 2008.
the quantitative results of previous works are directly taken from [8]
and [6].                                                                         [9] C. Stauffer and W. E. L. Grimson, “Learning patterns of activity using
                                                                                     real-time tracking,” IEEE Trans. PAMI, vol. 22, no. 8, pp. 747–757,
                                                                                     2000.
5.2. Effect of Shadows at Shadow/Light border                                   [10] D. S. Lee, “Effective gaussian mixture learning for video background
                                                                                     subtraction,” IEEE Trans. PAMI, vol. 27, no. 5, pp. 827–832, 2005.
We show the visual results with and without considering shadows at
shadow/light border. Fig. 4(a) is a traffic scene where sharp shad-              [11] J. B. Huang and C. S. Chen, “Learning moving cast shadows for fore-
ows are induced when blocking strong light sources (e.g. sun). Fig.                  ground detection,” in The Eighth International Workshop on Visual
                                                                                     Surveillance (VS), in conjunction with ECCV, 2008.
4(b) shows the foreground posterior without taking shadows at bor-
der into consideration. Shadows at border are misclassified as fore-             [12] A. Prati, I. Mikic, M.M. Trivedi, and R. Cucchiara, “Detecting moving
ground due to the distinct distribution of color feature values. On the              shadows: algorithms and evaluation,” IEEE Trans. PAMI, vol. 25, no.
                                                                                     7, pp. 918–923, 2003.
other hand, in Fig. 4(c), shadows at border can be correctly detected
when evaluating their angle information only.




                                                                          772

More Related Content

What's hot

SIGGRAPH 2014論文紹介 - Sound & Light + Fabrication Session
SIGGRAPH 2014論文紹介 - Sound & Light + Fabrication SessionSIGGRAPH 2014論文紹介 - Sound & Light + Fabrication Session
SIGGRAPH 2014論文紹介 - Sound & Light + Fabrication Sessionyamo_o
 
Deferred Pixel Shading on the PLAYSTATION®3
Deferred Pixel Shading on the PLAYSTATION®3Deferred Pixel Shading on the PLAYSTATION®3
Deferred Pixel Shading on the PLAYSTATION®3Slide_N
 
A novel method for mounting gunn diode in active slot ring
A novel method for mounting gunn diode in active slot ringA novel method for mounting gunn diode in active slot ring
A novel method for mounting gunn diode in active slot ringAlexander Decker
 
B140715
B140715B140715
B140715irjes
 
Implementing kohonen's som with missing data in OTB
Implementing kohonen's som with missing data in OTBImplementing kohonen's som with missing data in OTB
Implementing kohonen's som with missing data in OTBmelaneum
 
SAL3D presentation - AQSENSE's 3D machine vision library
SAL3D presentation - AQSENSE's 3D machine vision librarySAL3D presentation - AQSENSE's 3D machine vision library
SAL3D presentation - AQSENSE's 3D machine vision libraryAQSENSE S.L.
 
Analysis
AnalysisAnalysis
AnalysisGorybox
 
Introduction to compressive sensing
Introduction to compressive sensingIntroduction to compressive sensing
Introduction to compressive sensingAhmed Nasser Agag
 
Literature study article
Literature study articleLiterature study article
Literature study articleakaspers
 
Link and Energy Adaptive Design of Sustainable IR-UWB Communications and Sensing
Link and Energy Adaptive Design of Sustainable IR-UWB Communications and SensingLink and Energy Adaptive Design of Sustainable IR-UWB Communications and Sensing
Link and Energy Adaptive Design of Sustainable IR-UWB Communications and SensingDong Zhao
 
Sneutrino Cold Dark Matter, Oxford
Sneutrino Cold Dark Matter, OxfordSneutrino Cold Dark Matter, Oxford
Sneutrino Cold Dark Matter, OxfordSpinor
 
Advanced lithographic technologies
Advanced lithographic technologiesAdvanced lithographic technologies
Advanced lithographic technologiesLen Mei
 
Unit i mm_chap4_color in image and video
Unit i mm_chap4_color in image and  videoUnit i mm_chap4_color in image and  video
Unit i mm_chap4_color in image and videoEellekwameowusu
 
Projected Barzilai-Borwein Methods Applied to Distributed Compressive Spectru...
Projected Barzilai-Borwein Methods Applied to Distributed Compressive Spectru...Projected Barzilai-Borwein Methods Applied to Distributed Compressive Spectru...
Projected Barzilai-Borwein Methods Applied to Distributed Compressive Spectru...Polytechnique Montreal
 
Accelerometers to Augmented Reality
Accelerometers to Augmented RealityAccelerometers to Augmented Reality
Accelerometers to Augmented Realityjblocksom
 

What's hot (20)

SIGGRAPH 2014論文紹介 - Sound & Light + Fabrication Session
SIGGRAPH 2014論文紹介 - Sound & Light + Fabrication SessionSIGGRAPH 2014論文紹介 - Sound & Light + Fabrication Session
SIGGRAPH 2014論文紹介 - Sound & Light + Fabrication Session
 
Compressed Sensing In Spectral Imaging
Compressed Sensing In Spectral Imaging  Compressed Sensing In Spectral Imaging
Compressed Sensing In Spectral Imaging
 
Projet Ma2
Projet Ma2Projet Ma2
Projet Ma2
 
Deferred Pixel Shading on the PLAYSTATION®3
Deferred Pixel Shading on the PLAYSTATION®3Deferred Pixel Shading on the PLAYSTATION®3
Deferred Pixel Shading on the PLAYSTATION®3
 
Slides Hair Rendering
Slides Hair RenderingSlides Hair Rendering
Slides Hair Rendering
 
A novel method for mounting gunn diode in active slot ring
A novel method for mounting gunn diode in active slot ringA novel method for mounting gunn diode in active slot ring
A novel method for mounting gunn diode in active slot ring
 
B140715
B140715B140715
B140715
 
Implementing kohonen's som with missing data in OTB
Implementing kohonen's som with missing data in OTBImplementing kohonen's som with missing data in OTB
Implementing kohonen's som with missing data in OTB
 
Space time
Space timeSpace time
Space time
 
SAL3D presentation - AQSENSE's 3D machine vision library
SAL3D presentation - AQSENSE's 3D machine vision librarySAL3D presentation - AQSENSE's 3D machine vision library
SAL3D presentation - AQSENSE's 3D machine vision library
 
Analysis
AnalysisAnalysis
Analysis
 
Introduction to compressive sensing
Introduction to compressive sensingIntroduction to compressive sensing
Introduction to compressive sensing
 
ICPRAM 2012
ICPRAM 2012ICPRAM 2012
ICPRAM 2012
 
Literature study article
Literature study articleLiterature study article
Literature study article
 
Link and Energy Adaptive Design of Sustainable IR-UWB Communications and Sensing
Link and Energy Adaptive Design of Sustainable IR-UWB Communications and SensingLink and Energy Adaptive Design of Sustainable IR-UWB Communications and Sensing
Link and Energy Adaptive Design of Sustainable IR-UWB Communications and Sensing
 
Sneutrino Cold Dark Matter, Oxford
Sneutrino Cold Dark Matter, OxfordSneutrino Cold Dark Matter, Oxford
Sneutrino Cold Dark Matter, Oxford
 
Advanced lithographic technologies
Advanced lithographic technologiesAdvanced lithographic technologies
Advanced lithographic technologies
 
Unit i mm_chap4_color in image and video
Unit i mm_chap4_color in image and  videoUnit i mm_chap4_color in image and  video
Unit i mm_chap4_color in image and video
 
Projected Barzilai-Borwein Methods Applied to Distributed Compressive Spectru...
Projected Barzilai-Borwein Methods Applied to Distributed Compressive Spectru...Projected Barzilai-Borwein Methods Applied to Distributed Compressive Spectru...
Projected Barzilai-Borwein Methods Applied to Distributed Compressive Spectru...
 
Accelerometers to Augmented Reality
Accelerometers to Augmented RealityAccelerometers to Augmented Reality
Accelerometers to Augmented Reality
 

Viewers also liked

Final Portfolio and Self Evaluation Denise Clarke-Hang
Final Portfolio and Self Evaluation   Denise Clarke-HangFinal Portfolio and Self Evaluation   Denise Clarke-Hang
Final Portfolio and Self Evaluation Denise Clarke-Hangguest733717f
 
Caber potensi besar
Caber potensi besarCaber potensi besar
Caber potensi besarbogasi
 
Sport a közösségi médiában
Sport a közösségi médiábanSport a közösségi médiában
Sport a közösségi médiábanMihály Zsiros
 
Virtual Government Network at PS Engage 2011
Virtual Government Network at PS Engage 2011Virtual Government Network at PS Engage 2011
Virtual Government Network at PS Engage 2011Thom Kearney
 
The Crowded Boardroom WSSF 2013
The Crowded Boardroom WSSF 2013The Crowded Boardroom WSSF 2013
The Crowded Boardroom WSSF 2013Thom Kearney
 
Artworks
ArtworksArtworks
Artworkspepe
 
Mind Behind Movement
Mind Behind MovementMind Behind Movement
Mind Behind Movementguest2a787e1
 
At Boekblad aan Zee
At Boekblad aan ZeeAt Boekblad aan Zee
At Boekblad aan Zee24Symbols
 
Collaboration keys and a twitter example march 23 2011 v2
Collaboration keys and a twitter example march 23 2011 v2Collaboration keys and a twitter example march 23 2011 v2
Collaboration keys and a twitter example march 23 2011 v2Thom Kearney
 
Mobile phones can cook your brain
Mobile phones can cook your brainMobile phones can cook your brain
Mobile phones can cook your brainChirag Chaudasi
 
The future is a cloud hosted retail store
The future is a cloud hosted retail storeThe future is a cloud hosted retail store
The future is a cloud hosted retail store24Symbols
 
How To Choose An Installer
How To Choose An InstallerHow To Choose An Installer
How To Choose An InstallerMarkElson
 
Victor molev metamorphosis
Victor molev metamorphosisVictor molev metamorphosis
Victor molev metamorphosisemve
 
A magyar atlétikai klubok működési feltételei és esélyei tartalom
A magyar atlétikai klubok működési feltételei és esélyei tartalomA magyar atlétikai klubok működési feltételei és esélyei tartalom
A magyar atlétikai klubok működési feltételei és esélyei tartalomMihály Zsiros
 
Spanishpower point
Spanishpower pointSpanishpower point
Spanishpower pointguest3b9684d
 
Espn wsummit blogpdf
Espn wsummit blogpdfEspn wsummit blogpdf
Espn wsummit blogpdfEmma Cookson
 
Change, communication and collaboration
Change, communication and collaborationChange, communication and collaboration
Change, communication and collaborationThom Kearney
 
Ikt ak gure ikastetxeetan
Ikt ak gure ikastetxeetanIkt ak gure ikastetxeetan
Ikt ak gure ikastetxeetanMaitane
 

Viewers also liked (20)

Final Portfolio and Self Evaluation Denise Clarke-Hang
Final Portfolio and Self Evaluation   Denise Clarke-HangFinal Portfolio and Self Evaluation   Denise Clarke-Hang
Final Portfolio and Self Evaluation Denise Clarke-Hang
 
Confirmation lesson 4 8 b
Confirmation lesson 4 8 bConfirmation lesson 4 8 b
Confirmation lesson 4 8 b
 
Caber potensi besar
Caber potensi besarCaber potensi besar
Caber potensi besar
 
Sport a közösségi médiában
Sport a közösségi médiábanSport a közösségi médiában
Sport a közösségi médiában
 
Virtual Government Network at PS Engage 2011
Virtual Government Network at PS Engage 2011Virtual Government Network at PS Engage 2011
Virtual Government Network at PS Engage 2011
 
The Crowded Boardroom WSSF 2013
The Crowded Boardroom WSSF 2013The Crowded Boardroom WSSF 2013
The Crowded Boardroom WSSF 2013
 
Artworks
ArtworksArtworks
Artworks
 
Mind Behind Movement
Mind Behind MovementMind Behind Movement
Mind Behind Movement
 
At Boekblad aan Zee
At Boekblad aan ZeeAt Boekblad aan Zee
At Boekblad aan Zee
 
Collaboration keys and a twitter example march 23 2011 v2
Collaboration keys and a twitter example march 23 2011 v2Collaboration keys and a twitter example march 23 2011 v2
Collaboration keys and a twitter example march 23 2011 v2
 
Mobile phones can cook your brain
Mobile phones can cook your brainMobile phones can cook your brain
Mobile phones can cook your brain
 
The future is a cloud hosted retail store
The future is a cloud hosted retail storeThe future is a cloud hosted retail store
The future is a cloud hosted retail store
 
How To Choose An Installer
How To Choose An InstallerHow To Choose An Installer
How To Choose An Installer
 
Victor molev metamorphosis
Victor molev metamorphosisVictor molev metamorphosis
Victor molev metamorphosis
 
A magyar atlétikai klubok működési feltételei és esélyei tartalom
A magyar atlétikai klubok működési feltételei és esélyei tartalomA magyar atlétikai klubok működési feltételei és esélyei tartalom
A magyar atlétikai klubok működési feltételei és esélyei tartalom
 
10 02 16 Veb Strategic Planning Meeting
10 02 16 Veb Strategic Planning Meeting10 02 16 Veb Strategic Planning Meeting
10 02 16 Veb Strategic Planning Meeting
 
Spanishpower point
Spanishpower pointSpanishpower point
Spanishpower point
 
Espn wsummit blogpdf
Espn wsummit blogpdfEspn wsummit blogpdf
Espn wsummit blogpdf
 
Change, communication and collaboration
Change, communication and collaborationChange, communication and collaboration
Change, communication and collaboration
 
Ikt ak gure ikastetxeetan
Ikt ak gure ikastetxeetanIkt ak gure ikastetxeetan
Ikt ak gure ikastetxeetan
 

Similar to A Physical Approach to Moving Cast Shadow Detection (ICASSP 2009)

A Physical Approach to Moving Cast Shadow Detection (ICASSP 2009)
A Physical Approach to Moving Cast Shadow Detection (ICASSP 2009)A Physical Approach to Moving Cast Shadow Detection (ICASSP 2009)
A Physical Approach to Moving Cast Shadow Detection (ICASSP 2009)Jia-Bin Huang
 
Image Restoration and Denoising By Using Nonlocally Centralized Sparse Repres...
Image Restoration and Denoising By Using Nonlocally Centralized Sparse Repres...Image Restoration and Denoising By Using Nonlocally Centralized Sparse Repres...
Image Restoration and Denoising By Using Nonlocally Centralized Sparse Repres...IJERA Editor
 
APPEARANCE-BASED REPRESENTATION AND RENDERING OF CAST SHADOWS
APPEARANCE-BASED REPRESENTATION AND RENDERING OF CAST SHADOWSAPPEARANCE-BASED REPRESENTATION AND RENDERING OF CAST SHADOWS
APPEARANCE-BASED REPRESENTATION AND RENDERING OF CAST SHADOWSijcga
 
A Survey on Single Image Dehazing Approaches
A Survey on Single Image Dehazing ApproachesA Survey on Single Image Dehazing Approaches
A Survey on Single Image Dehazing ApproachesIRJET Journal
 
Interferogram Filtering Using Gaussians Scale Mixtures in Steerable Wavelet D...
Interferogram Filtering Using Gaussians Scale Mixtures in Steerable Wavelet D...Interferogram Filtering Using Gaussians Scale Mixtures in Steerable Wavelet D...
Interferogram Filtering Using Gaussians Scale Mixtures in Steerable Wavelet D...CSCJournals
 
Learning Moving Cast Shadows for Foreground Detection (VS 2008)
Learning Moving Cast Shadows for Foreground Detection (VS 2008)Learning Moving Cast Shadows for Foreground Detection (VS 2008)
Learning Moving Cast Shadows for Foreground Detection (VS 2008)Jia-Bin Huang
 
VISUAL MODEL BASED SINGLE IMAGE DEHAZING USING ARTIFICIAL BEE COLONY OPTIMIZA...
VISUAL MODEL BASED SINGLE IMAGE DEHAZING USING ARTIFICIAL BEE COLONY OPTIMIZA...VISUAL MODEL BASED SINGLE IMAGE DEHAZING USING ARTIFICIAL BEE COLONY OPTIMIZA...
VISUAL MODEL BASED SINGLE IMAGE DEHAZING USING ARTIFICIAL BEE COLONY OPTIMIZA...ijistjournal
 
VISUAL MODEL BASED SINGLE IMAGE DEHAZING USING ARTIFICIAL BEE COLONY OPTIMIZA...
VISUAL MODEL BASED SINGLE IMAGE DEHAZING USING ARTIFICIAL BEE COLONY OPTIMIZA...VISUAL MODEL BASED SINGLE IMAGE DEHAZING USING ARTIFICIAL BEE COLONY OPTIMIZA...
VISUAL MODEL BASED SINGLE IMAGE DEHAZING USING ARTIFICIAL BEE COLONY OPTIMIZA...ijistjournal
 
Segmentation Based Multilevel Wide Band Compression for SAR Images Using Coif...
Segmentation Based Multilevel Wide Band Compression for SAR Images Using Coif...Segmentation Based Multilevel Wide Band Compression for SAR Images Using Coif...
Segmentation Based Multilevel Wide Band Compression for SAR Images Using Coif...CSCJournals
 
Flat panel ---nihms864608
Flat panel ---nihms864608Flat panel ---nihms864608
Flat panel ---nihms864608Maria Vergakh
 
SHARP OR BLUR: A FAST NO-REFERENCE QUALITY METRIC FOR REALISTIC PHOTOS
SHARP OR BLUR: A FAST NO-REFERENCE QUALITY METRIC FOR REALISTIC PHOTOSSHARP OR BLUR: A FAST NO-REFERENCE QUALITY METRIC FOR REALISTIC PHOTOS
SHARP OR BLUR: A FAST NO-REFERENCE QUALITY METRIC FOR REALISTIC PHOTOScsandit
 
Image Restoration UsingNonlocally Centralized Sparse Representation and histo...
Image Restoration UsingNonlocally Centralized Sparse Representation and histo...Image Restoration UsingNonlocally Centralized Sparse Representation and histo...
Image Restoration UsingNonlocally Centralized Sparse Representation and histo...IJERA Editor
 
Shadow Detection and Removal using Tricolor Attenuation Model Based on Featur...
Shadow Detection and Removal using Tricolor Attenuation Model Based on Featur...Shadow Detection and Removal using Tricolor Attenuation Model Based on Featur...
Shadow Detection and Removal using Tricolor Attenuation Model Based on Featur...ijtsrd
 
A Novel Approach for Qualitative Imaging of Buried PEC Scatterers
A Novel Approach for Qualitative Imaging of Buried PEC ScatterersA Novel Approach for Qualitative Imaging of Buried PEC Scatterers
A Novel Approach for Qualitative Imaging of Buried PEC ScatterersTELKOMNIKA JOURNAL
 
120_SEM_Special_Topics.ppt
120_SEM_Special_Topics.ppt120_SEM_Special_Topics.ppt
120_SEM_Special_Topics.pptzaki194502
 
videoMotionTrackingPCA
videoMotionTrackingPCAvideoMotionTrackingPCA
videoMotionTrackingPCAKellen Betts
 
Image Denoising Based On Sparse Representation In A Probabilistic Framework
Image Denoising Based On Sparse Representation In A Probabilistic FrameworkImage Denoising Based On Sparse Representation In A Probabilistic Framework
Image Denoising Based On Sparse Representation In A Probabilistic FrameworkCSCJournals
 

Similar to A Physical Approach to Moving Cast Shadow Detection (ICASSP 2009) (20)

A Physical Approach to Moving Cast Shadow Detection (ICASSP 2009)
A Physical Approach to Moving Cast Shadow Detection (ICASSP 2009)A Physical Approach to Moving Cast Shadow Detection (ICASSP 2009)
A Physical Approach to Moving Cast Shadow Detection (ICASSP 2009)
 
Image Restoration and Denoising By Using Nonlocally Centralized Sparse Repres...
Image Restoration and Denoising By Using Nonlocally Centralized Sparse Repres...Image Restoration and Denoising By Using Nonlocally Centralized Sparse Repres...
Image Restoration and Denoising By Using Nonlocally Centralized Sparse Repres...
 
APPEARANCE-BASED REPRESENTATION AND RENDERING OF CAST SHADOWS
APPEARANCE-BASED REPRESENTATION AND RENDERING OF CAST SHADOWSAPPEARANCE-BASED REPRESENTATION AND RENDERING OF CAST SHADOWS
APPEARANCE-BASED REPRESENTATION AND RENDERING OF CAST SHADOWS
 
A Survey on Single Image Dehazing Approaches
A Survey on Single Image Dehazing ApproachesA Survey on Single Image Dehazing Approaches
A Survey on Single Image Dehazing Approaches
 
Interferogram Filtering Using Gaussians Scale Mixtures in Steerable Wavelet D...
Interferogram Filtering Using Gaussians Scale Mixtures in Steerable Wavelet D...Interferogram Filtering Using Gaussians Scale Mixtures in Steerable Wavelet D...
Interferogram Filtering Using Gaussians Scale Mixtures in Steerable Wavelet D...
 
Learning Moving Cast Shadows for Foreground Detection (VS 2008)
Learning Moving Cast Shadows for Foreground Detection (VS 2008)Learning Moving Cast Shadows for Foreground Detection (VS 2008)
Learning Moving Cast Shadows for Foreground Detection (VS 2008)
 
VISUAL MODEL BASED SINGLE IMAGE DEHAZING USING ARTIFICIAL BEE COLONY OPTIMIZA...
VISUAL MODEL BASED SINGLE IMAGE DEHAZING USING ARTIFICIAL BEE COLONY OPTIMIZA...VISUAL MODEL BASED SINGLE IMAGE DEHAZING USING ARTIFICIAL BEE COLONY OPTIMIZA...
VISUAL MODEL BASED SINGLE IMAGE DEHAZING USING ARTIFICIAL BEE COLONY OPTIMIZA...
 
VISUAL MODEL BASED SINGLE IMAGE DEHAZING USING ARTIFICIAL BEE COLONY OPTIMIZA...
VISUAL MODEL BASED SINGLE IMAGE DEHAZING USING ARTIFICIAL BEE COLONY OPTIMIZA...VISUAL MODEL BASED SINGLE IMAGE DEHAZING USING ARTIFICIAL BEE COLONY OPTIMIZA...
VISUAL MODEL BASED SINGLE IMAGE DEHAZING USING ARTIFICIAL BEE COLONY OPTIMIZA...
 
Segmentation Based Multilevel Wide Band Compression for SAR Images Using Coif...
Segmentation Based Multilevel Wide Band Compression for SAR Images Using Coif...Segmentation Based Multilevel Wide Band Compression for SAR Images Using Coif...
Segmentation Based Multilevel Wide Band Compression for SAR Images Using Coif...
 
Flat panel ---nihms864608
Flat panel ---nihms864608Flat panel ---nihms864608
Flat panel ---nihms864608
 
SHARP OR BLUR: A FAST NO-REFERENCE QUALITY METRIC FOR REALISTIC PHOTOS
SHARP OR BLUR: A FAST NO-REFERENCE QUALITY METRIC FOR REALISTIC PHOTOSSHARP OR BLUR: A FAST NO-REFERENCE QUALITY METRIC FOR REALISTIC PHOTOS
SHARP OR BLUR: A FAST NO-REFERENCE QUALITY METRIC FOR REALISTIC PHOTOS
 
Structlight
StructlightStructlight
Structlight
 
Image Restoration UsingNonlocally Centralized Sparse Representation and histo...
Image Restoration UsingNonlocally Centralized Sparse Representation and histo...Image Restoration UsingNonlocally Centralized Sparse Representation and histo...
Image Restoration UsingNonlocally Centralized Sparse Representation and histo...
 
Shadow Detection and Removal using Tricolor Attenuation Model Based on Featur...
Shadow Detection and Removal using Tricolor Attenuation Model Based on Featur...Shadow Detection and Removal using Tricolor Attenuation Model Based on Featur...
Shadow Detection and Removal using Tricolor Attenuation Model Based on Featur...
 
Lm342080283
Lm342080283Lm342080283
Lm342080283
 
A Novel Approach for Qualitative Imaging of Buried PEC Scatterers
A Novel Approach for Qualitative Imaging of Buried PEC ScatterersA Novel Approach for Qualitative Imaging of Buried PEC Scatterers
A Novel Approach for Qualitative Imaging of Buried PEC Scatterers
 
120_SEM_Special_Topics.ppt
120_SEM_Special_Topics.ppt120_SEM_Special_Topics.ppt
120_SEM_Special_Topics.ppt
 
Determining Optical Flow
Determining Optical FlowDetermining Optical Flow
Determining Optical Flow
 
videoMotionTrackingPCA
videoMotionTrackingPCAvideoMotionTrackingPCA
videoMotionTrackingPCA
 
Image Denoising Based On Sparse Representation In A Probabilistic Framework
Image Denoising Based On Sparse Representation In A Probabilistic FrameworkImage Denoising Based On Sparse Representation In A Probabilistic Framework
Image Denoising Based On Sparse Representation In A Probabilistic Framework
 

More from Jia-Bin Huang

How to write a clear paper
How to write a clear paperHow to write a clear paper
How to write a clear paperJia-Bin Huang
 
Research 101 - Paper Writing with LaTeX
Research 101 - Paper Writing with LaTeXResearch 101 - Paper Writing with LaTeX
Research 101 - Paper Writing with LaTeXJia-Bin Huang
 
Computer Vision Crash Course
Computer Vision Crash CourseComputer Vision Crash Course
Computer Vision Crash CourseJia-Bin Huang
 
Applying for Graduate School in S.T.E.M.
Applying for Graduate School in S.T.E.M.Applying for Graduate School in S.T.E.M.
Applying for Graduate School in S.T.E.M.Jia-Bin Huang
 
Linear Algebra and Matlab tutorial
Linear Algebra and Matlab tutorialLinear Algebra and Matlab tutorial
Linear Algebra and Matlab tutorialJia-Bin Huang
 
Single Image Super-Resolution from Transformed Self-Exemplars (CVPR 2015)
Single Image Super-Resolution from Transformed Self-Exemplars (CVPR 2015)Single Image Super-Resolution from Transformed Self-Exemplars (CVPR 2015)
Single Image Super-Resolution from Transformed Self-Exemplars (CVPR 2015)Jia-Bin Huang
 
Lecture 29 Convolutional Neural Networks - Computer Vision Spring2015
Lecture 29 Convolutional Neural Networks -  Computer Vision Spring2015Lecture 29 Convolutional Neural Networks -  Computer Vision Spring2015
Lecture 29 Convolutional Neural Networks - Computer Vision Spring2015Jia-Bin Huang
 
Lecture 21 - Image Categorization - Computer Vision Spring2015
Lecture 21 - Image Categorization -  Computer Vision Spring2015Lecture 21 - Image Categorization -  Computer Vision Spring2015
Lecture 21 - Image Categorization - Computer Vision Spring2015Jia-Bin Huang
 
Writing Fast MATLAB Code
Writing Fast MATLAB CodeWriting Fast MATLAB Code
Writing Fast MATLAB CodeJia-Bin Huang
 
Image Completion using Planar Structure Guidance (SIGGRAPH 2014)
Image Completion using Planar Structure Guidance (SIGGRAPH 2014)Image Completion using Planar Structure Guidance (SIGGRAPH 2014)
Image Completion using Planar Structure Guidance (SIGGRAPH 2014)Jia-Bin Huang
 
Toward Accurate and Robust Cross-Ratio based Gaze Trackers Through Learning F...
Toward Accurate and Robust Cross-Ratio based Gaze Trackers Through Learning F...Toward Accurate and Robust Cross-Ratio based Gaze Trackers Through Learning F...
Toward Accurate and Robust Cross-Ratio based Gaze Trackers Through Learning F...Jia-Bin Huang
 
Real-time Face Detection and Recognition
Real-time Face Detection and RecognitionReal-time Face Detection and Recognition
Real-time Face Detection and RecognitionJia-Bin Huang
 
Transformation Guided Image Completion ICCP 2013
Transformation Guided Image Completion ICCP 2013Transformation Guided Image Completion ICCP 2013
Transformation Guided Image Completion ICCP 2013Jia-Bin Huang
 
Jia-Bin Huang's Curriculum Vitae
Jia-Bin Huang's Curriculum VitaeJia-Bin Huang's Curriculum Vitae
Jia-Bin Huang's Curriculum VitaeJia-Bin Huang
 
Pose aware online visual tracking
Pose aware online visual trackingPose aware online visual tracking
Pose aware online visual trackingJia-Bin Huang
 
Face Expression Enhancement
Face Expression EnhancementFace Expression Enhancement
Face Expression EnhancementJia-Bin Huang
 
Image Smoothing for Structure Extraction
Image Smoothing for Structure ExtractionImage Smoothing for Structure Extraction
Image Smoothing for Structure ExtractionJia-Bin Huang
 
Three Reasons to Join FVE at uiuc
Three Reasons to Join FVE at uiucThree Reasons to Join FVE at uiuc
Three Reasons to Join FVE at uiucJia-Bin Huang
 
Static and Dynamic Hand Gesture Recognition
Static and Dynamic Hand Gesture RecognitionStatic and Dynamic Hand Gesture Recognition
Static and Dynamic Hand Gesture RecognitionJia-Bin Huang
 
Real-Time Face Detection, Tracking, and Attributes Recognition
Real-Time Face Detection, Tracking, and Attributes RecognitionReal-Time Face Detection, Tracking, and Attributes Recognition
Real-Time Face Detection, Tracking, and Attributes RecognitionJia-Bin Huang
 

More from Jia-Bin Huang (20)

How to write a clear paper
How to write a clear paperHow to write a clear paper
How to write a clear paper
 
Research 101 - Paper Writing with LaTeX
Research 101 - Paper Writing with LaTeXResearch 101 - Paper Writing with LaTeX
Research 101 - Paper Writing with LaTeX
 
Computer Vision Crash Course
Computer Vision Crash CourseComputer Vision Crash Course
Computer Vision Crash Course
 
Applying for Graduate School in S.T.E.M.
Applying for Graduate School in S.T.E.M.Applying for Graduate School in S.T.E.M.
Applying for Graduate School in S.T.E.M.
 
Linear Algebra and Matlab tutorial
Linear Algebra and Matlab tutorialLinear Algebra and Matlab tutorial
Linear Algebra and Matlab tutorial
 
Single Image Super-Resolution from Transformed Self-Exemplars (CVPR 2015)
Single Image Super-Resolution from Transformed Self-Exemplars (CVPR 2015)Single Image Super-Resolution from Transformed Self-Exemplars (CVPR 2015)
Single Image Super-Resolution from Transformed Self-Exemplars (CVPR 2015)
 
Lecture 29 Convolutional Neural Networks - Computer Vision Spring2015
Lecture 29 Convolutional Neural Networks -  Computer Vision Spring2015Lecture 29 Convolutional Neural Networks -  Computer Vision Spring2015
Lecture 29 Convolutional Neural Networks - Computer Vision Spring2015
 
Lecture 21 - Image Categorization - Computer Vision Spring2015
Lecture 21 - Image Categorization -  Computer Vision Spring2015Lecture 21 - Image Categorization -  Computer Vision Spring2015
Lecture 21 - Image Categorization - Computer Vision Spring2015
 
Writing Fast MATLAB Code
Writing Fast MATLAB CodeWriting Fast MATLAB Code
Writing Fast MATLAB Code
 
Image Completion using Planar Structure Guidance (SIGGRAPH 2014)
Image Completion using Planar Structure Guidance (SIGGRAPH 2014)Image Completion using Planar Structure Guidance (SIGGRAPH 2014)
Image Completion using Planar Structure Guidance (SIGGRAPH 2014)
 
Toward Accurate and Robust Cross-Ratio based Gaze Trackers Through Learning F...
Toward Accurate and Robust Cross-Ratio based Gaze Trackers Through Learning F...Toward Accurate and Robust Cross-Ratio based Gaze Trackers Through Learning F...
Toward Accurate and Robust Cross-Ratio based Gaze Trackers Through Learning F...
 
Real-time Face Detection and Recognition
Real-time Face Detection and RecognitionReal-time Face Detection and Recognition
Real-time Face Detection and Recognition
 
Transformation Guided Image Completion ICCP 2013
Transformation Guided Image Completion ICCP 2013Transformation Guided Image Completion ICCP 2013
Transformation Guided Image Completion ICCP 2013
 
Jia-Bin Huang's Curriculum Vitae
Jia-Bin Huang's Curriculum VitaeJia-Bin Huang's Curriculum Vitae
Jia-Bin Huang's Curriculum Vitae
 
Pose aware online visual tracking
Pose aware online visual trackingPose aware online visual tracking
Pose aware online visual tracking
 
Face Expression Enhancement
Face Expression EnhancementFace Expression Enhancement
Face Expression Enhancement
 
Image Smoothing for Structure Extraction
Image Smoothing for Structure ExtractionImage Smoothing for Structure Extraction
Image Smoothing for Structure Extraction
 
Three Reasons to Join FVE at uiuc
Three Reasons to Join FVE at uiucThree Reasons to Join FVE at uiuc
Three Reasons to Join FVE at uiuc
 
Static and Dynamic Hand Gesture Recognition
Static and Dynamic Hand Gesture RecognitionStatic and Dynamic Hand Gesture Recognition
Static and Dynamic Hand Gesture Recognition
 
Real-Time Face Detection, Tracking, and Attributes Recognition
Real-Time Face Detection, Tracking, and Attributes RecognitionReal-Time Face Detection, Tracking, and Attributes Recognition
Real-Time Face Detection, Tracking, and Attributes Recognition
 

A Physical Approach to Moving Cast Shadow Detection (ICASSP 2009)

  • 1. A PHYSICAL APPROACH TO MOVING CAST SHADOW DETECTION Jia-Bin Huang and Chu-Song Chen Academia Sinica, Institute of Information Science Taipei, Taiwan, R.O.C. ABSTRACT light scattered by the sky, and other colored light from nearby This paper presents a physics-based approach capable of detecting surfaces (i.e. color bleeding). These lights can have significantly cast shadows in video sequence effectively. We develop a new phys- different spectral power distributions (SPD). Therefore, the RGB ical model of cast shadows without making prior assumption of the value of a shaded pixel may not attenuate linearly along the line spectral power distribution (SPD) of the light sources and ambi- between the value of the corresponding background and the origin. ent illumination in the scene. The background appearance varia- Little attention has been directed toward this problem before. tion caused by cast shadows is characterized as the interaction of Nadami et al. [7] addressed this non-linearity by dichromatic reflec- the blocked light sources and the background surface reflectance. tion model for both the sun and the sky illuminations in an outdoor We then take advantage of the statistical prevalence of cast shadows scene. However, the method is designed only for an outdoor envi- to learn and update the shadow model parameters using the Gaus- ronment. Recently, a more general shadow model was presented in sian mixture model (GMM) over time. The proposed algorithm is [8]. This model introduced an ambient illumination term which de- completely unsupervised and can adapt to specific environment with termines a direction in the RGB space where the background values complex illumination condition as well as changing shadow condi- under cast shadows can be found. Since the ambient term may have tions. Experimental results on three challenging sequences demon- spectrum which differs from the incident light sources, the values of strate the effectiveness of the proposed method. the shaded pixels may not be proportional to the values under direct light. Nonparametric density estimation is used to model the surface Index Terms— Visual surveillance, Object detection, Moving variation under cast shadows in an unsupervised way. By provid- Cast Shadow Detection ing a better description of cast shadows on background surfaces, the shadow model presented in [8] considerably improved the previous 1. INTRODUCTION approaches based on linear model. However, in [8] the SPDs of all light sources are assumed having the same profile but with different Detection of moving objects is an important low-level vision task power factor.This assumption may not be hold for many of the real- for many vision-based applications. One of the most challenging world cases. In addition, the ambient illuminations are assumed to problems in these applications is identifying cast shadows in video. be constant, but they may be slightly changed due to circulation of Cast shadows are often misclassified as parts of foreground regions, foreground objects. resulting in distortion of the true shape and color properties of the In this paper, we extend the shadow model presented in [8] by target objects. An effective shadow detection method is thus neces- releasing these assumptions. The proposed shadow model is more sary for accurate foreground segmentation. general and can be applied to real-world cases more appropriately. Moving cast shadows are caused by the occlusion of light However, a more general model may suffer from a longer period of sources. Shadows reduce the total energy incident at the background time for learning the model parameters. To address this problem, surfaces where the light sources are partially or totally blocked by we introduce the confidence-rated Gaussian mixture learning which the foreground objects. Hence, points under shadows have lower accelerates the parameters convergent rate. Moreover, we observe luminance values but similar chromaticity values compared to their that shadows on the light/shadow borders show different behavior reference background points. from shadows inside the shaded region. The shadows detection rate Many works have been proposed for cast shadow detection. can be further improved by addressing this issue. Most of them are based on the assumption that the RGB values of The remainder of this paper is organized as follows. We first shadow pixels will fall on the line between the illuminated values propose in Section 2 a new physical model for describing appearance and the origin in the RGB color space (linear model). This linear variation of a background value under cast shadows. Based on the model has been employed in different color spaces [1, 2, 3]. One proposed model, we present an unsupervised learning approach to main limitation of these methods is that they require explicit tuning learn the behavior of cast shadows on surfaces by taking advantage in a large set of parameters for each new scene and are difficult to of the statistical prevalence of shadows in Section 3. In Section 4, work for on-line applications. To adapt to environment changes, we derive the posterior probability of cast shadows and foreground statistical learning methods have been employed to describe the cast given an observed sample. Experimental results are presented in shadows [4, 5, 6]. However, the linear model assumption may not Section 5. Section 6 concludes the paper. be fulfilled in a real-world environment. For instance, in an outdoor scene the illumination may consist of direct sunlight, the diffused 2. PHYSICAL MODEL FOR CAST SHADOWS This research was supported by the National Science Council of Taiwan under Grant No. NSC 95-2221-E-001-028-MY3 and 97-EC-17-A-02-S1- According to the Lambertian model, the camera sensor response 032. gk (p), k ∈ {R, G, B} at pixel p depends on the intensity func- 978-1-4244-2354-5/09/$25.00 ©2009 IEEE 769 ICASSP 2009
  • 2. from the contribution of the blocked incident and ambient illumina- tions. We then develop a 3D color feature of cast shadows xs,t (p) = [αt (p), θt (p), φt (p)]T to encode this information, where αt (p) rep- resents the illumination attenuation and θt (p) and φt (p) indicate the direction of the vector vt (p) in spherical coordinate. Similar to Hor- prasert et al. [1], we define the illumination attenuation αt (p) by ||vt (p)|| αt (p) = , (3) ||BGt (p)|| (a) (b) where BGt (p) is the background value at the pixel p. The illumi- Fig. 1. Physical cast shadow model. (a) Linear attenuation model nation attenuation, which is smaller than one, describes the strength and the proposed model. The linear model will fail when the SPD of brightness of the blocked lights with respect to that of the back- of ambient illumination does not consistent with that of the light ground value. The direction of the vector vt (p), the chromaticity sources. (b) Background appearance variation due to cast shadows. line direction, is determined by θt (p) and φt (p): vG (p) t vB (p) θt (p) = arctan( ), φt (p) = arccos( t ), (4) tion of the light sources E(λ, p), the reflectance of an object surface vR (p) t ||vt (p)|| ρ(λ, p), and sensor spectral sensitivity Sk (λ): where the superscript R, G, B indicate the components in the RGB space. We can use this 3-D color feature vector vt (p) to describe the gk (p) = E(λ, p)ρ(λ, p)Sk (λ)dλ. (1) appearance variation induced by the blocked light sources on shaded regions. Unlike [8], our model takes the variation of chromaticity di- We describe the illumination incident to a surface point p by N light rection into consideration and thus better describes the cast shadows sources and lights reflected from M ambient objects (the position p on background surface. is ignore): Since different foreground objects often block the light sources N M in a similar way, cast shadows on background surfaces are thus sim- E(λ) = Eincident,n (λ) + Eambient,m (λ). (2) ilar and uncorrelated with foreground objects. We present in the fol- n=1 m=1 lowing section how to learn the cast shadow model in a completely unsupervised way. For a color camera with a linear response, the response to a specific illumination follows a line in the RGB space. We show our model 3. LEARNING AND DETECTING CAST SHADOWS in Fig. 1(a), where BGA is the ambient term from the contribution of all ambient lightings and BG is the background surface value. Shadow detection methods with fixed parameters can not adapt to The zigzag lines starting from BGA and ending at BG represent time-varying illumination. In this paper, we use Gaussian mixture the response from the incident light sources. Based on the additive model (GMM) [9] to model cast shadows by using scene activities. color model, their net effect forms the vector from BGA to BG. We first apply weak shadow detector to moving pixels detected by Note that in this model, the ambient illumination can have different background model to select possible shadow samples. The color fea- SPDs from that of the total contribution of incident light sources (i.e. tures of these possible shadow pixels are then learned by GMM over the direction of BGA may not be consistent with the direction from time. Then, we can detect shadows by using the learned shadow BGA to BG). In addition, the SPD of individual light source may model.Note that our approach is pixel-based, model the SPDs of differ from that of other light sources. ambient lighting and light sources locally, and thus can deal with In comparison, in [8] the SPDs of the incident light sources are environment with complex illumination conditions. assumed sharing the same profile. That is, the first term in (2) de- generates into Eincident (λ) N γn , where γn is the scaling fac- n=1 tor of the nth light sources. Moreover, the ambient illumination is 3.1. Weak Shadow Detector assumed to be constant. Under these two assumptions, all shadow To model the cast shadows, impossible shadow samples that belong values are expected to be found on the line between background BG to background and foreground should be excluded. Therefore, we and the ambient value BGA . This approach may fail to deal with apply a weak shadow detector that evaluates every moving pixels de- environment with complex illumination conditions. tected by the background model to filter out some impossible shadow Given a surface point, the background surface value variation samples. Since cast shadows on surface reduce luminance values and under cast shadows is resulted from the occlusion of lights by fore- change the saturation, we define the potential shadow values fall into ground objects. The blocked lights may contain partial or total direct the conic volume around the corresponding background color [4]. incident light sources and some reflected lights from nearby ambient The weak shadow detector is applied to moving regions detected objects. Here, some of the incident lights will disappear since they by the background model to obtain possible shadow points. We call are blocked by foreground objects. The first term in (2) is then re- these possible shadow points as “shadow candidates”, which form duced to N Eincident,n (λ) − n∈blocked Eincident,n (λ). The n=1 the set Q. At the pixel level, we model xs,t (p), p ∈ Q using the ambient lights are also changed slightly due to the circulation of GMM to learn the color feature of shadows. foreground objects. Therefore, according to (1), the vector from the shadow value (i.e. SD in Fig. 1) to its corresponding background 3.2. Spatial Information value can be interpreted as the interaction between the blocked lights and surface reflectance (see Fig. 1(b)). With the proposed physical shadow model, we can accurately de- For a pixel p at time t, the vector from shadow to background scribe the appearance variation of background surface under shad- value is denoted as vt (p). The vector vt (p) is the net response ows at the pixel level. However, two shortcomings remain. One 770
  • 3. than ambient illumination, blocking them induces a sharp shadow. In this case, shadows at light/shadow border show different behavior from shadows inside the shaded region. This is caused by the in- (a) sufficient resolution of CCD camera. Pixels at light/shadow border may receive part of energy from light sources. We demonstrate this observation in Fig. 2. We first collected shadow samples throughout the sequence “Highway I”. Shadow border pixels can be detected from background posterior image where large gradient intensities (b) are present. We classify these shadow samples into two groups: “inside” and “at border”. Fig. 2(a) presents that the illumination attenuation have significantly different distribution for shadows at border and inside. On the other hand, Fig. 2(b)(c) show that the (c) distribution on angle information are very similar. The reason is that neighboring pixels often receive incident energies with similar Fig. 2. Color feature value distribution of shadows inside (in red) SPDs.Consequently, we propose to address shadows at border by and at the light/shadow border (in blue). Samples are taken from evaluating only their angle information. the sequence “Highway I”. (a) Illumination attenuation αt (p) (b)(c) angle information: θt (p), and φt (p), respectively. 4. POSTERIOR PROBABILITIES In this section, we present how to derive the posterior probabilities of is that the proposed shadow model may suffer from slow learning cast shadows and foreground given the observed sample xt (p) using of model parameters if motion is rare on a pixel. With the con- the proposed shadow model and spatial information. ventional Gaussian mixture learning [9], training the pixel-based shadow model requires a long time while the illumination conditions 4.1. Cast Shadow Posterior should remain stable. The other problem arises from foreground ob- jects having similar colors as background, which may be misclassi- The shadow posterior is first computed by decomposing P (SD|xt (p)) fied as cast shadows. Therefore, we propose to incorporate spatial over (BG, FS) domain, where F S indicates moving pixels (real information for improving the shadow discriminative ability and the foreground and cast shadows). Since P (SD|xt (p), BG) = 0, the convergent rate of the model parameters. decomposition is In addition to the illumination attenuation, when a surface point is shaded, the gradient intensity around the point will also be smaller P (SD|xt (p)) = P (SD|xt (p), F S)P (F S|xt (p)), (6) than that of the illuminated surface. Consequently, points having where P (F S|xt (p)) = 1 − P (BG|xt (p)) can be directly com- larger gradient intensities than background are more likely to be puted from the background model. Second, we remove pixels that considered as foreground objects. We thus penalize these samples are definitely foreground and consider only shadow candidate (SC). by lessening the learning rate in this situation. We define a con- Also, we penalize pixels whose gradient intensities are amplified fidence value of shadow using spatial gradient intensities of back- instead of reduced by multiplying ωt (p): P (SD|xt (p), F S) = ground |∇(Bt )| and current image |∇(It )|: P (SD|xt (p), F S, SC) ∗ ωt (p). Here, the color feature vector ε + |∇(Bt (p))| xc,t (p) is the sufficient statistics for the observed sample xt (p). ωt (p) = , (5) Therefore, we have P (SD|xt (p), F S, SC) = P (SD|xc,t (p), F S, SC). ε + max{|∇(It (p))|, |∇(Bt (p))|} As mentioned in section 3, we learn the cast shadows using the where ε is a smooth term to alleviate the effect of noise in textureless GMM [9]. The posterior probability can be expressed in terms of regions. Thus, we can discriminate foreground objects with similar the mixture components in GMM: color as background from shadow with the aid of gradient informa- tion. To improve the convergent rate, we use ωt (p) as the confi- P (SD|xc,t (p), F S, SC) = P (SD|Gk )P (Gk |xc,t (p)), (7) k dence value to update the GMM at pixel p through confidence-rated learning. In [10], Lee has shown that fast updating can be achieved where Gk is the kth Gaussian state in the mixture. Given the ob- by combining recursive filter type of learning with incremental EM servation that color change due to objects has higher variance com- in background modeling. We propose to weight these two learning pared to the color change due to cast shadows, Gaussian states with scheme using the confidence value ωt (p) extracted from spatial do- higher prior probabilities and smaller variances would be considered main [11]. If ωt (p) is large, the learning rate depends more on the as shadows. Therefore, we approximate P (SD|Gk ) using logistic formulation of incremental EM algorithm, which converges quickly regression similar to that in [10] for background subtraction. with few observations. On the other hand, if ωt (p) is small, the learning rate follows the recursive type of learning to maintain model 4.2. Foreground Posterior stability. Computing foreground posterior probability is much easier. Given 3.3. Shadows at Light/Shadow Border a pixel p at time t, we first compute the background posterior P (BG|xt (p)) from the background model. Then, we can ob- Once the shadow model for a pixel p is built, we can estimate tain shadow posterior probability P (SD|xt (p)) using the learned the likelihood of the observation belong to shadows by checking shadow model. Based on the probability theory, the foreground whether the 3-D color feature vector xs,t (p) is associated with one posterior can be obtained as: of the working Gaussian distribution in the mixture. However, if the light sources are clustered in the scene and are relatively brighter P (F G|xt (p)) = 1 − P (BG|xt (p)) − P (SD|xt (p)). (8) 771
  • 4. (a) (b) (c) Fig. 4. Effect of shadows at shadow/light border (a) Original frame of sequence ”Highway I”. (b)(c) Foreground posterior without/with considering shadows at shadow/light border. (a) (b) (c) (d) 6. CONCLUSION Fig. 3. Visual results of the proposed algorithm. (a) Original In this paper, we have introduced a novel physical model for cast frame. (b) Background posterior P (BG|xt (p)). (c) Shadow pos- shadow detection in video sequences. Qualitative and quantitative terior P (SD|xt (p)). (d) Foreground posterior P (F G|xt (p)). evaluation of the physical shadow model validate that our approach is more accurate in describing background surface variation under cast shadows. Spatial information is used to help discriminat- Table 1. Quantitative results on surveillance sequences ing foreground with similar chroma from cast shadows as well Sequence Highway I Highway II Hallway as improving the convergent rate of the shadow model through Method η% ξ% η% ξ% η% ξ% confidence-rated learning. We also address the problem of shadows Proposed 72.34 84.98 72.70 79.89 71.69 88.25 at light/shadow border where these shadows have different behavior Kernel [8] 70.50 84.40 68.40 71.20 72.40 86.70 from other regions. Note that the proposed approach is pixel-based; LGf [6] 72.10 79.70 - - - - a contextual model that incorporate spatial and temporal coherence GMSM [5] 63.30 71.30 58.51 44.40 60.50 87.00 can further enhance the performance. 7. REFERENCES 5. EXPERIMENTAL RESULTS [1] T. Horprasert, D. Harwood, and L. S. Davis, “A statistical approach for real-time robust background subtraction and shadow detection,” in We validate the proposed method on three challenging video se- ICCV Frame-rate Workshop, 1999. quences. Both visual results and quantitative measurement are [2] R. Cucchiara, C. Grana, M. Piccardi, and A. Prati, “Detecting moving presented. We compare our algorithm quantitatively with previous objects, ghosts, and shadows in video streams,” IEEE Trans. PAMI, shadow detection methods when results are available. The effect of vol. 25, no. 10, pp. 1337–1342, 2003. considering shadows at light/shadow border is also present. [3] E. Salvador, A. Cavallaro, and T. Ebrahimi, “Cast shadow segmentation using invariant color features,” CVIU, vol. 95, no. 2, pp. 238–259, 2004. 5.1. Assessing Shadow Detection Accuracy [4] F. Porikli and J. Thornton, “Shadow flow: a recursive method to learn moving cast shadows,” in ICCV, Oct. 2005, pp. 891–898 Vol. 1. We show in Fig. 3 the visual results of three video sequences from the proposed algorithm. Fig. 3(a) shows one frame from the se- [5] N. Martel-Brisson and A. Zaccarin, “Learning and removing cast shad- ows through a multidistribution approach,” IEEE Trans. PAMI, vol. 29, quence. The background, shadow, and foreground posterior are no. 7, pp. 1133–1146, 2007. given in Fig. 3 (b)(c)(d), respectively. We can see that in all three test sequences the cast shadows are correctly detected. [6] Z. Liu, K. Huang, T. Tan, and L. Wang, “Cast shadow removal com- bining local and global features,” in CVPR, 2007, pp. 1–8. We evaluate our method through the quantitative metrics (shadow detection η% and discrimination rate ξ%) proposed in [12]. For ex- [7] S. Nadimi and B. Bhanu, “Physical models for moving shadow and object detection in video,” IEEE Trans. PAMI, vol. 26, no. 8, pp. 1079– act equations, please refer to [12]. We report our results in Table 1087, 2004. 1 and compare them with recent works using statistical method to detect shadows. Our method can achieve higher shadow detection [8] N. Martel-Brisson and A. Zaccarin, “Kernel-based learning of cast shadows from a physical model of light sources and surfaces for low- and discrimination rate than the state-of-the-art approaches, where level segmentation,” in CVPR, June. 2008. the quantitative results of previous works are directly taken from [8] and [6]. [9] C. Stauffer and W. E. L. Grimson, “Learning patterns of activity using real-time tracking,” IEEE Trans. PAMI, vol. 22, no. 8, pp. 747–757, 2000. 5.2. Effect of Shadows at Shadow/Light border [10] D. S. Lee, “Effective gaussian mixture learning for video background subtraction,” IEEE Trans. PAMI, vol. 27, no. 5, pp. 827–832, 2005. We show the visual results with and without considering shadows at shadow/light border. Fig. 4(a) is a traffic scene where sharp shad- [11] J. B. Huang and C. S. Chen, “Learning moving cast shadows for fore- ows are induced when blocking strong light sources (e.g. sun). Fig. ground detection,” in The Eighth International Workshop on Visual Surveillance (VS), in conjunction with ECCV, 2008. 4(b) shows the foreground posterior without taking shadows at bor- der into consideration. Shadows at border are misclassified as fore- [12] A. Prati, I. Mikic, M.M. Trivedi, and R. Cucchiara, “Detecting moving ground due to the distinct distribution of color feature values. On the shadows: algorithms and evaluation,” IEEE Trans. PAMI, vol. 25, no. 7, pp. 918–923, 2003. other hand, in Fig. 4(c), shadows at border can be correctly detected when evaluating their angle information only. 772