×

Training of adapted classifiers for video categorization

  • US 8,452,778 B1
  • Filed: 09/01/2010
  • Issued: 05/28/2013
  • Est. Priority Date: 11/19/2009
  • Status: Active Grant
First Claim
Patent Images

1. A computer implemented method of training video classifiers, the method comprising:

  • storing a taxonomy of hierarchically-arranged categories;

    storing a set of labeled videos, each of the labeled videos having associated textual metadata and being initially labeled as representing one or more of the categories;

    storing labels initially associated with a set of text documents distinct from the labeled videos, each stored label corresponding to one of the categories and indicating that the associated text document represents the category;

    identifying, for each of the categories, a positive training subset of the text documents that represent the category based on their stored labels, and a negative training subset of the text documents that do not represent the category based on their stored labels;

    training a set of text-based classifiers based on the positive training subsets and the negative training subsets, each text-based classifier associated with one of the categories and producing, when applied to text, a score providing a measure of how strongly the text represents the associated category;

    identifying, for each of the categories, a positive training subset of the labeled videos that represent the category based on their labels, and a negative training subset of the labeled videos that do not represent the category based on their labels;

    for each video of the positive training subsets of the labeled videos and of the negative training subsets of the labeled videos;

    applying the text-based classifiers to the associated textual metadata of the video, thereby producing a vector of scores for the video, the scores providing measures of how strongly the textual metadata of the video represents the categories associated with the text-based classifiers;

    extracting a content feature vector from video content of frames of the video;

    forming a hybrid feature vector comprising the vector of scores and the content feature vector for that video; and

    training a set of adapted classifiers based on the hybrid feature vectors of the videos in the positive training subsets of the labeled videos and on the hybrid feature vectors of the videos in the negative training subsets of the labeled videos, each adapted classifier associated with one of the categories and producing, when applied to an unlabeled video, a score providing a measure of how strongly the unlabeled video represents the associated category.

View all claims
  • 2 Assignments
Timeline View
Assignment View
    ×
    ×