Methods and apparatus for identifying portions of a video stream based on characteristics of the video stream
First Claim
1. A method for processing an audio/video stream, the method comprising:
- providing a first audio/video stream including at least one segment of a show, at least one interstitial of the show and closed captioning data that is shifted in time relative to the at least one segment of the show;
receiving autonomous location information referencing the closed captioning data to identify a first identified audio/video location within the first audio/video stream, wherein the first identified audio/video location is shifted in time relative to a corresponding referenced audio/video location within the first audio/video stream;
receiving a first off-set and a second off-set;
applying the first off-set to the first identified audio/video location to identify a second identified audio/video location as a possible segment boundary of the first audio/video stream;
determining whether the second identified audio/video location corresponds with a first boundary of a segment of the show based on characteristics of a video frame at the second identified audio/video location;
identifying a third identified audio/video location that corresponds with the first boundary of the segment responsive to determining that the second identified audio/video location does not correspond with the first boundary;
determining a shift offset of the closed captioning data, the shift offset corresponding to a difference between the second identified audio/video location and the third identified audio/video location;
identifying a second boundary of the segment of the show, using the first identified audio/video location, the second off-set, and the shift offset of the closed captioning data;
filtering the interstitial from the first audio/video stream based on the first boundary of the segment of the show and the second boundary of the segment of the show to generate a second audio/video stream including the segment of the show; and
outputting the second audio/video stream for presentation by a display device.
3 Assignments
0 Petitions
Accused Products
Abstract
The various embodiments described herein generally provide for the identification of locations in a presentation stream based on metadata associated with the presentation stream. Locations within an audio/video stream are identified by processing associated text data using autonomous location information referencing the text data. The identified locations within a presentation stream may be utilized to identify boundaries of segments within the presentation stream, such as segments of a show and interstitials of the show. Processing is then performed to determine whether the identified boundaries possess specific characteristics. If the identified boundaries do not possess the identified characteristics, then additional processing is performed to identify other locations, temporally near the identified boundaries, which correspond with boundaries of the portion of the audio/video stream identified by the autonomous location information.
-
Citations
23 Claims
-
1. A method for processing an audio/video stream, the method comprising:
-
providing a first audio/video stream including at least one segment of a show, at least one interstitial of the show and closed captioning data that is shifted in time relative to the at least one segment of the show; receiving autonomous location information referencing the closed captioning data to identify a first identified audio/video location within the first audio/video stream, wherein the first identified audio/video location is shifted in time relative to a corresponding referenced audio/video location within the first audio/video stream; receiving a first off-set and a second off-set; applying the first off-set to the first identified audio/video location to identify a second identified audio/video location as a possible segment boundary of the first audio/video stream; determining whether the second identified audio/video location corresponds with a first boundary of a segment of the show based on characteristics of a video frame at the second identified audio/video location; identifying a third identified audio/video location that corresponds with the first boundary of the segment responsive to determining that the second identified audio/video location does not correspond with the first boundary; determining a shift offset of the closed captioning data, the shift offset corresponding to a difference between the second identified audio/video location and the third identified audio/video location; identifying a second boundary of the segment of the show, using the first identified audio/video location, the second off-set, and the shift offset of the closed captioning data; filtering the interstitial from the first audio/video stream based on the first boundary of the segment of the show and the second boundary of the segment of the show to generate a second audio/video stream including the segment of the show; and outputting the second audio/video stream for presentation by a display device. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8)
-
-
9. A digital video recorder comprising:
-
a communication interface that receives a first audio/video stream including at least one segment of a show, at least one interstitial of the show and closed captioning data that is shifted in time relative to the at least one segment of the show; a storage medium; control logic communicatively coupled to the communication interface and the storage medium that; stores the first audio/video stream on the storage medium for subsequent presentation to a user; receives autonomous location information referencing the closed captioning data to identify a first identified audio/video location within the first audio/video stream, wherein the first identified audio/video location is shifted in time relative to a corresponding referenced audio/video location within the first audio/video stream; receives a first off-set and a second off-set; applies the first off-set to the first identified audio/video location to identify a second identified audio/video location as a possible segment boundary of the first audio/video stream; determines whether the second identified audio/video location corresponds with a first boundary of a segment of the show based on characteristics of a video frame at the second identified audio/video location; identifies a third identified audio/video location that corresponds with the first boundary of the segment responsive to determining that the second identified audio/video location does not correspond with the first boundary; determines a shift offset of the closed captioning data, the shift offset corresponding to a difference between the second identified audio/video location and the third identified audio/video location; identifies a second boundary of the segment of the show, using the first identified audio/video location, the second off-set, and the shift offset of the closed captioning data; filters the interstitial from the first audio/video stream based on the first boundary of the segment of the show and the second boundary of the segment of the show to generate a second audio/video stream including the segment of the show; and outputs the second audio/video stream for presentation by a display device. - View Dependent Claims (10, 11, 12, 13, 14, 15, 16, 17, 18)
-
-
19. A method for processing an audio/video stream, the method comprising:
-
recording a first audio/video stream including a plurality of segments of a show, interstitials interspersed between the plurality of segments, and closed captioning data that is shifted in time relative to the plurality of segments; receiving autonomous location information referencing the closed captioning data, the autonomous location information comprising a plurality of text strings associated with the closed captioning data for one or more video frames of the first audio/video stream, wherein each of the plurality of text strings uniquely identifies a respective audio/video location within the first audio/video stream corresponding to a respective one of the plurality of segments; receiving a first offset and a second offset for each respective audio/video location uniquely identified by the plurality of text strings; for a first segment of the plurality of segments; reviewing the closed captioning data to locate one of the plurality of text strings in the closed captioning data and to identify a first identified audio/video location within the first audio/video stream, wherein the first identified audio/video location is shifted in time relative to a corresponding referenced audio/video location within the first audio/video stream; applying the first offset of the first segment to the first identified audio/video location to identify a second identified audio/video location as a possible segment boundary of the first audio/video stream; determining whether the second identified audio/video location corresponds with a first boundary of the first segment based on characteristics of a video frame at the second identified audio/video location; and identifying a third identified audio/video location that corresponds with the first boundary of the first segment in response to determining that the second identified audio/video location does not correspond with the first boundary of the first segment; determining a shift offset of the closed captioning data, the shift offset corresponding to a difference between the second identified audio/video location and the third identified audio/video location; using the shift offset of the closed captioning data to identify boundaries of other segments of the show; filtering the interstitials from the first audio/video stream based on the identified boundaries to generate a second audio/video stream including the segments of the show; and outputting the second audio/video stream for presentation by a display device. - View Dependent Claims (20, 21, 22, 23)
-
Specification