Pixel-level based micro-feature extraction
First Claim
1. A processor-implemented method for tracking an object in a video scene, the method comprising:
- determining, via at least one processor, at least one pixel-level characteristic for a foreground patch in a first video frame from a plurality of video frames of a video scene, the foreground patch including a set of pixels containing a portion of scene foreground in the first video frame of the video scene;
determining, via the at least one processor, kinematics of an object in the video scene based at least in part on the at least one pixel-level characteristic;
generating, via the at least one processor, a micro-feature vector based at least in part on the kinematics of the object, the micro-feature vector corresponding to at least one micro-feature value;
updating, via the at least one processor, a neural network based at least in part on the micro-feature vector, the neural network including a plurality of clusters, each cluster in the plurality of clusters representing an object type;
classifying, via the at least one processor, the object in the first video frame of the video scene as a first object type based on the updated neural network, the first object type being represented by a first cluster in the plurality of clusters; and
tracking, via the at least one processor, the object in the video scene based on the micro-feature vector and the first object type.
69 Assignments
0 Petitions
Accused Products
Abstract
Techniques are disclosed for extracting micro-features at a pixel-level based on characteristics of one or more images. Importantly, the extraction is unsupervised, i.e., performed independent of any training data that defines particular objects, allowing a behavior-recognition system to forgo a training phase and for object classification to proceed without being constrained by specific object definitions. A micro-feature extractor that does not require training data is adaptive and self-trains while performing the extraction. The extracted micro-features are represented as a micro-feature vector that may be input to a micro-classifier which groups objects into object type clusters based on the micro-feature vectors.
-
Citations
20 Claims
-
1. A processor-implemented method for tracking an object in a video scene, the method comprising:
-
determining, via at least one processor, at least one pixel-level characteristic for a foreground patch in a first video frame from a plurality of video frames of a video scene, the foreground patch including a set of pixels containing a portion of scene foreground in the first video frame of the video scene; determining, via the at least one processor, kinematics of an object in the video scene based at least in part on the at least one pixel-level characteristic; generating, via the at least one processor, a micro-feature vector based at least in part on the kinematics of the object, the micro-feature vector corresponding to at least one micro-feature value; updating, via the at least one processor, a neural network based at least in part on the micro-feature vector, the neural network including a plurality of clusters, each cluster in the plurality of clusters representing an object type; classifying, via the at least one processor, the object in the first video frame of the video scene as a first object type based on the updated neural network, the first object type being represented by a first cluster in the plurality of clusters; and tracking, via the at least one processor, the object in the video scene based on the micro-feature vector and the first object type. - View Dependent Claims (2, 3, 4, 5, 6, 7)
-
-
8. A system, comprising:
-
a video input receiver configured to receive a plurality of video frames of a video scene; at least one processor; and a memory having processor-executable instructions and in communication with the at least one processor, the instructions in the memory comprising instructions for the at least one processor to; determine at least one pixel-level characteristic for a foreground patch in a first video frame from the plurality of video frames of a video scene received by the video input receiver, the foreground patch including a set of pixels containing a portion of scene foreground in the first video frame of the video scene, determine kinematics of an object in the video scene based at least in part on the at least one pixel-level characteristic, generate a micro-feature vector based at least in part on at least on the kinematics of the object, the micro-feature vector corresponding to at least one micro-feature value; update a neural network based at least in part on the micro-feature vector, the neural network including a plurality of clusters, each cluster in the plurality of clusters representing an object type, classify the object in the first video frame of the video scene as a first object type based on the updated neural network, the first object type being represented by a first cluster in the plurality of clusters, and track the object in the video scene based on the micro-feature vector and the first object type. - View Dependent Claims (9, 10, 11, 12, 13)
-
-
14. A non-transitory computer-readable storage medium storing computer-executable instructions for tracking an object in a video scene, the computer-executable instructions comprising instructions to:
-
determine at least one pixel-level characteristic for a foreground patch in a first video frame from a plurality of video frames of a video scene, the foreground patch including a set of pixels containing a portion of scene foreground in the first video frame of the video scene; determine kinematics of an object in the video scene based at least in part on that at least one pixel-level characteristic; generate a micro-feature vector based at least in part on the kinematics of the object, the micro-feature vector corresponding to at least one micro-feature value; update a neural network based at least in part on the micro-feature vector, the neural network including a plurality of clusters, each cluster in the plurality of clusters representative of an object type; classify the object in the first video frame of the video scene as a first object type based on the updated neural network, the first object type being represented by a first cluster in the plurality of clusters; and track the object in the video scene based on the micro-feature vector and the first object type. - View Dependent Claims (15, 16, 17, 18, 19, 20)
-
Specification