BACKGROUND MODEL FOR COMPLEX AND DYNAMIC SCENES
First Claim
Patent Images
1. A computer-implemented method for generating a background model of a scene depicted in a sequence of video frames captured by a video camera, the method comprising:
- receiving a video frame, wherein the video frame includes one or more appearance values for each of a plurality of pixels;
for one or more of the pixels;
passing the appearance values for the pixel to an input layer of an adaptive resonance theory (ART) network corresponding to the pixel;
mapping, by the ART network, the appearance values to one of one or more clusters of the ART network;
classifying the pixel as depicting one of scene background and scene foreground, based on the mapping of the appearance values to the cluster of the ART network.
6 Assignments
0 Petitions
Accused Products
Abstract
Techniques are disclosed for learning and modeling a background for a complex and/or dynamic scene over a period of observations without supervision. A background/foreground component of a computer vision engine may be configured to model a scene using an array of ART networks. The ART networks learn the regularity and periodicity of the scene by observing the scene over a period of time. Thus, the ART networks allow the computer vision engine to model complex and dynamic scene backgrounds in video.
-
Citations
25 Claims
-
1. A computer-implemented method for generating a background model of a scene depicted in a sequence of video frames captured by a video camera, the method comprising:
-
receiving a video frame, wherein the video frame includes one or more appearance values for each of a plurality of pixels; for one or more of the pixels; passing the appearance values for the pixel to an input layer of an adaptive resonance theory (ART) network corresponding to the pixel; mapping, by the ART network, the appearance values to one of one or more clusters of the ART network; classifying the pixel as depicting one of scene background and scene foreground, based on the mapping of the appearance values to the cluster of the ART network. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9)
-
-
10. A computer-readable storage medium containing a program, which when executed on a processor, performs an operation for generating a background model of a scene depicted in a sequence of video frames captured by a video camera, the operation comprising:
-
receiving a video frame, wherein the video frame includes one or more appearance values for each of a plurality of pixels; for one or more of the pixels; passing the appearance values for the pixel to an input layer of an adaptive resonance theory (ART) network corresponding to the pixel; mapping, by the ART network, the appearance values to one of one or more clusters of the ART network; classifying the pixel as depicting one of scene background and scene foreground, based on the mapping of the appearance values to the cluster of the ART network. - View Dependent Claims (11, 12, 13, 14, 15, 16, 17)
-
-
18. A system, comprising:
-
a video input source configured to provide a sequence of video frames, each depicting a scene; a processor; and a memory containing a program, which, when executed on the processor is configured to perform an operation for generating a background model of a scene depicted in a sequence of video frames captured by a video camera, the operation comprising; receiving a video frame, wherein the video frame includes one or more appearance values for each of a plurality of pixels; for one or more of the pixels; passing the appearance values for the pixel to an input layer of an adaptive resonance theory (ART) network corresponding to the pixel; mapping, by the ART network, the appearance values to one of one or more clusters of the ART network; classifying the pixel as depicting one of scene background and scene foreground, based on the mapping of the appearance values to the cluster of the ART network. - View Dependent Claims (19, 20, 21, 22, 23, 24, 25)
-
Specification