Methods and systems for representation and matching of video content
First Claim
Patent Images
1. A method of determining spatio-temporal correspondence between different sets of video data, the method comprising:
- inputting the sets of video data;
wherein the video data is a collection of video sequences comprising query video data and corpus video data, or subsets of a single video sequence or modified subsets of a video sequence from the corpus video data;
representing the video data as ordered sequences of visual nucleotides;
determining temporally corresponding subsets of video data by aligning the sequences of visual nucleotides;
computing a spatial correspondence between the temporally corresponding subsets of video data (spatio-temporal correspondence); and
outputting the spatio-temporal correspondence between subsets of the video data.
4 Assignments
0 Petitions
Accused Products
Abstract
The described methods and systems provide for the representation and matching of video content, including spatio-temporal matching of different video sequences. A particular method of determining temporal correspondence between different sets of video data inputs the sets of video data and represents the video data as ordered sequences of visual nucleotides. Temporally corresponding subsets of video data are determined by aligning the sequences of visual nucleotides.
20 Citations
51 Claims
-
1. A method of determining spatio-temporal correspondence between different sets of video data, the method comprising:
-
inputting the sets of video data; wherein the video data is a collection of video sequences comprising query video data and corpus video data, or subsets of a single video sequence or modified subsets of a video sequence from the corpus video data; representing the video data as ordered sequences of visual nucleotides; determining temporally corresponding subsets of video data by aligning the sequences of visual nucleotides; computing a spatial correspondence between the temporally corresponding subsets of video data (spatio-temporal correspondence); and outputting the spatio-temporal correspondence between subsets of the video data. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50)
-
-
51. An apparatus comprising:
-
a source of video data, said video data comprising a collection of video sequences comprising query video data and corpus video data, or subsets of a single video sequence or modified subsets of a video sequence from the corpus video data; a video segmenter coupled to the source of video data and configured to segment video data into temporal intervals; a video processor coupled to the source of video data and configured to detect feature locations within the video data, generate feature descriptors associated with the feature locations, and prune the detected feature locations to generate a subset of feature locations; and a video aggregator coupled to the video segmenter and the video processor, the video aggregator configured to generate a video DNA associated with the video data, wherein the video DNA includes video data ordered as sequences of visual nucleotides.
-
Specification