Dark scene compensation in a background-foreground module of...

Image analysis – Image segmentation

Reexamination Certificate

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

C375S240080

Reexamination Certificate

active

08064695

ABSTRACT:
Embodiments of the present invention provide a method and a module for identifying a background of a scene depicted in an acquired stream of video frames that may be used by a video-analysis system. For each pixel or block of pixels in an acquired video frame a comparison measure is determined. The comparison measure depends on difference of color values exhibited in the acquired video frame and in a background image respectively by the pixel or block of pixels and a corresponding pixel and block of pixels in the background image. To determine the comparison measure, the resulting difference is considered in relation to a range of possible color values. If the comparison measure is above a dynamically adjusted threshold, the pixel or the block of pixels is classified as a part of the background of the scene.

REFERENCES:
patent: 4998162 (1991-03-01), Kondo et al.
patent: 5677733 (1997-10-01), Yoshimura et al.
patent: 7136097 (2006-11-01), Toyoda et al.
patent: 7236190 (2007-06-01), Yanof et al.
patent: 7349574 (2008-03-01), Sodini et al.
patent: 7535511 (2009-05-01), Wu et al.
patent: 7639741 (2009-12-01), Holt et al.
patent: 7710466 (2010-05-01), Drader et al.
patent: 2004/0032906 (2004-02-01), Lillig
patent: 2008/0193010 (2008-08-01), Eaton et al.
patent: 2008/0266413 (2008-10-01), Cohen et al.
patent: 2009/0087024 (2009-04-01), Eaton et al.
patent: 2009/0087085 (2009-04-01), Eaton et al.
patent: 2009/0087086 (2009-04-01), Eaton et al.
patent: 2009/0087093 (2009-04-01), Eaton et al.
patent: 2009/0087096 (2009-04-01), Eaton et al.
S. Apewokin, et al. “Multimodal Mean Adaptive Backgrounding for Embedded Real-Time Video Surveillance,” (2007) IEEE 6 pages.
Chris Stauffer, et al. “Adaptive Background Mixture Models for Real-Time Tracking,” The Artificial Intelligence Laboratory, Massachussetts Institute of Technology 7 pages, 2007.
Ahmed Elgammal, et al. “Non-parametric Model for Background Substraction,” Computer Vision Laboratory, University of Maryland 17 pages, 2007.
Ismail Haritaoglu, et al. “W4: Real-Time Surveillance of People and Their Activities,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 22, No. 8 (Aug. 2000), pp. 809-830.
Yuri Ivanov, et al. “Video Surveillance of Interactions,” MIT Media Laboratory, Cambridge, MA 8 pages, 2000.
Chris Stauffer, et al. “Learning Patterns of Activity Using Real-Time Tracking,” Artificial Intelligence Laboratory, Massachussetts Institute of Technology, Cambridge, MA 17 pages, 1993
Pentti Kanerva, “Sparse Distributed Memory and Related Models,” M.H. Hassoun, ed., Associative Neural Memories:Theory and Implementation, pp. 50-76. New York: Oxford University Press (1993).
Andrew Senior, et al. “Appearance Models for Occlusion Handling,” IBM T.J. Watson Research Center, Yorktown, Heights, NY, 8 pages, 2000.
U.S. Appl. No. 12/129,521, filed May 29, 2008, entitled, “Background-Foreground Module for Video Analysis System,”.
U.S. Appl. No. 12/129,551, filed May 29, 2008, entitled, “Identifying Stale Background Pixels in a Video Analysis System,”.

LandOfFree

Say what you really think

Search LandOfFree.com for the USA inventors and patents. Rate them and share your experience with other people.

Rating

Dark scene compensation in a background-foreground module of... does not yet have a rating. At this time, there are no reviews or comments for this patent.

If you have personal experience with Dark scene compensation in a background-foreground module of..., we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Dark scene compensation in a background-foreground module of... will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFUS-PAI-O-4292599

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.