Identifying scene boundaries using group sparsity analysis
First Claim
1. A method for determining scene boundaries within a video sequence including a time sequence of video frames, each video frame including an array of image pixels having pixel values, comprising:
- a) selecting a set of video frames from the video sequence, the set of video frames comprising a starting point and an ending point;
b) extracting a feature vector for each video frame in the set of video frames;
c) applying a group sparsity algorithm to represent the feature vector for a particular video frame as a group sparse combination of the feature vectors for the other video frames in the set of video frames, each feature vector for the other video frames in the group sparse combination having an associated weighting coefficient, wherein the weighting coefficients for feature vectors corresponding to other video frames that are most similar to the particular video frame are non-zero, and the weighting coefficients for feature vectors corresponding to other video frames that are most dissimilar from the particular video frame are zero;
d) analyzing the weighting coefficients to determine a video frame cluster of temporally-contiguous, similar video frames that includes the particular video frame;
e) repeating steps c)-d) for a plurality of particular video frames to provide a plurality of video frame clusters;
f) identifying one or more scene boundaries corresponding scenes in the video sequence based on the locations of boundaries between the determined video frame clusters; and
g) storing an indication of the identified scene boundaries in a processor-accessible memory;
wherein selecting a set of video frames from the video sequence comprises temporally sub-sampling the video sequence to select a subset of video frames.
5 Assignments
0 Petitions
Accused Products
Abstract
A method for identifying a set of key video frames from a video sequence comprising extracting feature vectors for each video frame and applying a group sparsity algorithm to represent the feature vector for a particular video frame as a group sparse combination of the feature vectors for the other video frames. Weighting coefficients associated with the group sparse combination are analyzed to determine video frame clusters of temporally-contiguous, similar video frames. The video sequence is segmented into scenes by identifying scene boundaries based on the determined video frame clusters.
-
Citations
10 Claims
-
1. A method for determining scene boundaries within a video sequence including a time sequence of video frames, each video frame including an array of image pixels having pixel values, comprising:
-
a) selecting a set of video frames from the video sequence, the set of video frames comprising a starting point and an ending point; b) extracting a feature vector for each video frame in the set of video frames; c) applying a group sparsity algorithm to represent the feature vector for a particular video frame as a group sparse combination of the feature vectors for the other video frames in the set of video frames, each feature vector for the other video frames in the group sparse combination having an associated weighting coefficient, wherein the weighting coefficients for feature vectors corresponding to other video frames that are most similar to the particular video frame are non-zero, and the weighting coefficients for feature vectors corresponding to other video frames that are most dissimilar from the particular video frame are zero; d) analyzing the weighting coefficients to determine a video frame cluster of temporally-contiguous, similar video frames that includes the particular video frame; e) repeating steps c)-d) for a plurality of particular video frames to provide a plurality of video frame clusters; f) identifying one or more scene boundaries corresponding scenes in the video sequence based on the locations of boundaries between the determined video frame clusters; and g) storing an indication of the identified scene boundaries in a processor-accessible memory; wherein selecting a set of video frames from the video sequence comprises temporally sub-sampling the video sequence to select a subset of video frames. - View Dependent Claims (2, 3, 4)
-
-
5. A method for determining scene boundaries within a video sequence including a time sequence of video frames, each video frame including an array of image pixels having pixel values, comprising:
-
a) selecting a set of video frames from the video sequence, the set of video frames comprising a starting point and an ending point; b) extracting a feature vector for each video frame in the set of video frames; c) applying a group sparsity algorithm to represent the feature vector for a particular video frame as a group sparse combination of the feature vectors for the other video frames in the set of video frames, each feature vector for the other video frames in the group sparse combination having an associated weighting coefficient, wherein the weighting coefficients for feature vectors corresponding to other video frames that are most similar to the particular video frame are non-zero, and the weighting coefficients for feature vectors corresponding to other video frames that are most dissimilar from the particular video frame are zero; d) analyzing the weighting coefficients to determine a video frame cluster of temporally-contiguous, similar video frames that includes the particular video frame; e) repeating steps c)-d) for a plurality of particular video frames to provide a plurality of video frame clusters; f) identifying one or more scene boundaries corresponding scenes in the video sequence based on the locations of boundaries between the determined video frame clusters; and g) storing an indication of the identified scene boundaries in a processor-accessible memory; wherein the video sequence comprises a plurality of independently encoded video frames and a plurality of video frames that are encoded using inter-frame coding, and wherein selecting a set of video frames from the video sequence comprises selecting at least a subset of the plurality of independently encoded video frames. - View Dependent Claims (6)
-
-
7. A method for determining scene boundaries within a video sequence including a time sequence of video frames, each video frame including an array of image pixels having pixel values, comprising:
-
a) selecting a set of video frames from the video sequence, the set of video frames comprising a starting point and an ending point; b) extracting a feature vector for each video frame in the set of video frames; c) applying a group sparsity algorithm to represent the feature vector for a particular video frame as a group sparse combination of the feature vectors for the other video frames in the set of video frames, each feature vector for the other video frames in the group sparse combination having an associated weighting coefficient, wherein the weighting coefficients for feature vectors corresponding to other video frames that are most similar to the particular video frame are non-zero, and the weighting coefficients for feature vectors corresponding to other video frames that are most dissimilar from the particular video frame are zero; d) analyzing the weighting coefficients to determine a video frame cluster of temporally-contiguous, similar video frames that includes the particular video frame; e) repeating steps c)-d) for a plurality of particular video frames to provide a plurality of video frame clusters; f) identifying one or more scene boundaries corresponding scenes in the video sequence based on the locations of boundaries between the determined video frame clusters; and g) storing an indication of the identified scene boundaries in a processor-accessible memory; wherein extracting a feature vector for each video frame comprises extracting a color channel of each video frame. - View Dependent Claims (8, 9, 10)
-
Specification