Methods and apparatus for identifying segments of content in a presentation stream using signature data
First Claim
Patent Images
1. A method for processing an audio/video stream, the method comprising:
- receiving location information including a text string associated with a first video location within the audio/video stream, the text string referencing closed captioning data associated with the audio/video stream to identify the first video location within the audio/video stream, the location information being received separately from the audio/video stream;
receiving a signature of a portion of the audio/video stream, wherein the signature refers to waveform characteristics of the portion of the audio/video stream;
receiving at least one search boundary offset specified relative to the first video location;
processing the location information and the closed captioning data to locate an instance of the text string in the closed captioning data;
identifying an intermediate video location within the audio/video stream, based on the instance of the text string located in the closed captioning data;
applying the at least one search boundary offset to the identified intermediate video location to identify search boundaries within the audio/video stream;
processing content of the audio/video stream within the search boundaries to locate boundaries of a segment of the audio/video stream, based on the received signature and at least one segment boundary offset associated with the received signature;
receiving user input requesting presentation of the segment of the audio/video stream; and
outputting the segment for presentation by a display device, wherein the outputting uses the located boundaries of the segment.
3 Assignments
0 Petitions
Accused Products
Abstract
Described herein are methods and apparatus for the identification of locations in a presentation stream based on metadata associated with the presentation stream. Locations within a presentation stream are identified using signature data associated with the presentation stream. The identified locations within a presentation stream may be utilized to identify boundaries of segments within the presentation stream, such as logical chapters, scenes or other sections of the content. The identified segments of the presentation stream may then be selectably viewed by a user.
-
Citations
14 Claims
-
1. A method for processing an audio/video stream, the method comprising:
-
receiving location information including a text string associated with a first video location within the audio/video stream, the text string referencing closed captioning data associated with the audio/video stream to identify the first video location within the audio/video stream, the location information being received separately from the audio/video stream; receiving a signature of a portion of the audio/video stream, wherein the signature refers to waveform characteristics of the portion of the audio/video stream; receiving at least one search boundary offset specified relative to the first video location; processing the location information and the closed captioning data to locate an instance of the text string in the closed captioning data; identifying an intermediate video location within the audio/video stream, based on the instance of the text string located in the closed captioning data; applying the at least one search boundary offset to the identified intermediate video location to identify search boundaries within the audio/video stream; processing content of the audio/video stream within the search boundaries to locate boundaries of a segment of the audio/video stream, based on the received signature and at least one segment boundary offset associated with the received signature; receiving user input requesting presentation of the segment of the audio/video stream; and outputting the segment for presentation by a display device, wherein the outputting uses the located boundaries of the segment. - View Dependent Claims (2, 3, 4)
-
-
5. A method for processing an audio/video stream, the method comprising:
-
receiving location information identifying a first location within an audio/video stream, the location information including a text string corresponding to closed captioning data for the audio/video stream, and the location information including at least one search boundary offset relative to the first location; receiving a signature of a portion of the audio/video stream, the signature identifying a transition in the video data from a first luminance value for a first frame of the video data to a second luminance value for a second frame of the video data; processing the closed captioning data to locate an instance of the text string in the closed captioning data and to identify an intermediate location within the audio/video stream based on the instance of the text string; applying the at least one search boundary offset to the identified intermediate location to identify search boundaries within the audio/video stream; computing average luminance values for a plurality of frames of the video data of the audio/video stream, wherein the plurality of frames are within the search boundaries; processing the average luminance values to identify the transition from the first luminance value to the second luminance value based on the signature, the transition corresponding with a signature-based location within the audio/video stream; identifying boundaries of a segment of the audio/video stream based on the signature-based location; receiving user input requesting presentation of the segment of the audio/video stream; and outputting the segment for presentation by a display device, wherein the outputting uses the identified boundaries of the segment. - View Dependent Claims (6, 7, 8)
-
-
9. A digital video recorder comprising:
-
a communication interface that receives an audio/video stream including a plurality of segments and that further receives location information identifying a first location within the audio/video stream, a signature of a portion of the audio/video stream, and at least one search boundary offset specified relative to the first location, wherein the location information, the signature and the at least one search boundary offset are received separately from the audio/video stream, wherein the signature identifies a transition in the video data from a first luminance value for a first frame of the video data to a second luminance value for a second frame of the video data, and wherein the location information includes a text string corresponding to closed captioning data for the audio/video stream; a storage medium; and control logic communicatively coupled to the communication interface and the storage medium that is configured to; process the closed captioning data to locate an instance of the text string in the closed captioning data and to identify an intermediate location within the audio/video stream based on the instance of the text string; apply the at least one search boundary offset to the identified intermediate location to identify search boundaries within the audio/video stream; compute average luminance values for a plurality of frames of the video data of the audio/video stream, wherein the plurality of frames are within the search boundaries; process the average luminance values to identify the transition from the first luminance value to the second luminance value based on the signature, the transition corresponding with a signature-based location within the audio/video stream; identify the boundaries of one of the plurality of segments of the audio/video stream based on the signature-based location and at least one segment boundary offset; receive user input requesting presentation of the one of the plurality of segments of the audio/video stream; and output the one of the plurality of segments for presentation by a display device, wherein the outputting uses the identified boundaries. - View Dependent Claims (10)
-
-
11. A digital video recorder comprising:
-
a communication interface that receives an audio/video stream including a plurality of segments and associated closed captioning data, the communication interface further receiving location information including a text string referencing the closed captioning data to identify a first video location within the audio/video stream, and the communication interface further receiving a signature of a portion of the audio/video stream, and at least one search boundary offset specified relative to the first video location, wherein the signature refers to waveform characteristics of the portion of the audio/video stream; a storage medium; control logic communicatively coupled to the communication interface and the storage medium configured to; process the location information and the closed captioning data to locate an instance of the text string in the closed captioning data; identify an intermediate video location within the audio/video stream, based on the instance of the text string located in the closed captioning data; apply the at least one search boundary offset to the identified intermediate video location to identify search boundaries within the audio/video stream; process content of the audio/video stream within the search boundaries to locate boundaries of a segment of the audio/video stream, based on the received signature and at least one segment boundary offset associated with the received signature; receive user input requesting presentation of the segment of the audio/video stream; and an audio/video output interface that outputs the segment for presentation by a display device, wherein the outputting uses the located boundaries of the segment. - View Dependent Claims (12, 13, 14)
-
Specification