Methods and systems of spatiotemporal pattern recognition for video content development
First Claim
1. A method for providing enhanced video content, comprising:
- processing at least one video feed through at least one spatiotemporal pattern recognition algorithm that uses a machine learning system to determine at least one event type for each of a plurality of events within the at least one video feed;
extracting a plurality of video cuts from the at least one video feed using a combination of the determination of the at least one event type and a determination of an event type of another input feed selected from the group consisting of a broadcast video feed, an audio feed and a closed caption feed, wherein the determination of an event type of the another input feed is responsive to applying the machine learning system to a portion of content of a broadcast commentary present in the another input feed;
indexing the extracted plurality of video cuts based on a similarity of the at least one event type determined by the machine learning system with the event type determined by applying the machine learning system to the portion of content of broadcast commentary present in the another input feed;
identifying at least one pattern in the extracted plurality of video cuts, the pattern spanning a plurality of frames of at least one of the extracted plurality of video cuts; and
indexing at least a portion of the plurality of the extracted video cuts that comprise the at least one pattern with an attribute indicative of the at least one pattern;
wherein the at least one spatiotemporal pattern recognition algorithm is based on at least one pattern recognized by adjusting an input feature of a plurality of input features and a weight thereof within the machine learning system, wherein the machine learning system is constructed to process the plurality of input features of the at least one video feed, the plurality of input features comprising;
relative direction of motion of at least two visible features,duration of relative motion of visible features with respect to each other,rate of motion of at least two visible features with respect to each other,relative acceleration of motion of at least two visible features, andrelative projected point of intersection of at least two visible features.
3 Assignments
0 Petitions
Accused Products
Abstract
Providing enhanced video content includes processing at least one video feed through at least one spatiotemporal pattern recognition algorithm that uses machine learning to develop an understanding of a plurality of events and to determine at least one event type for each of the plurality of events. The event type includes an entry in a relationship library detailing a relationship between two visible features. Extracting and indexing a plurality of video cuts from the video feed is performed based on the at least one event type determined by the understanding that corresponds to an event in the plurality of events detectable in the video cuts. Lastly, automatically and under computer control, an enhanced video content data structure is generated using the extracted plurality of video cuts based on the indexing of the extracted plurality of video cuts.
82 Citations
22 Claims
-
1. A method for providing enhanced video content, comprising:
-
processing at least one video feed through at least one spatiotemporal pattern recognition algorithm that uses a machine learning system to determine at least one event type for each of a plurality of events within the at least one video feed; extracting a plurality of video cuts from the at least one video feed using a combination of the determination of the at least one event type and a determination of an event type of another input feed selected from the group consisting of a broadcast video feed, an audio feed and a closed caption feed, wherein the determination of an event type of the another input feed is responsive to applying the machine learning system to a portion of content of a broadcast commentary present in the another input feed; indexing the extracted plurality of video cuts based on a similarity of the at least one event type determined by the machine learning system with the event type determined by applying the machine learning system to the portion of content of broadcast commentary present in the another input feed; identifying at least one pattern in the extracted plurality of video cuts, the pattern spanning a plurality of frames of at least one of the extracted plurality of video cuts; and indexing at least a portion of the plurality of the extracted video cuts that comprise the at least one pattern with an attribute indicative of the at least one pattern; wherein the at least one spatiotemporal pattern recognition algorithm is based on at least one pattern recognized by adjusting an input feature of a plurality of input features and a weight thereof within the machine learning system, wherein the machine learning system is constructed to process the plurality of input features of the at least one video feed, the plurality of input features comprising; relative direction of motion of at least two visible features, duration of relative motion of visible features with respect to each other, rate of motion of at least two visible features with respect to each other, relative acceleration of motion of at least two visible features, and relative projected point of intersection of at least two visible features. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22)
-
Specification