Voice-based video tagging
First Claim
1. A method for identifying events of interest in a captured video, the method comprising:
- storing multiple stored speech patterns for multiple input types, the multiple stored speech patterns corresponding to a command for identifying the events of interest within the captured video, wherein the multiple stored speech patterns include a first stored speech pattern for a first input type, wherein storing the first stored speech pattern comprises;
receiving, from a user, an input configuring a camera into a training mode to learn the first stored speech pattern;
capturing the first stored speech pattern from the user; and
storing the first stored speech pattern, wherein the first stored speech pattern is stored in response to capturing the first stored speech pattern from the user a threshold number of times;
accessing a captured speech pattern, the captured speech pattern captured from the user during capture of the captured video;
determining that the captured speech pattern corresponds to the first stored speech pattern; and
in response to determining that the captured speech pattern corresponds to the first stored speech pattern, storing event of interest information in metadata associated with the captured video, the event of interest information identifying (i) the first input type for a first event of interest, and (ii) an event moment during the capture of the captured video at which the captured speech pattern was captured from the user.
3 Assignments
0 Petitions
Accused Products
Abstract
Video and corresponding metadata is accessed. Events of interest within the video are identified based on the corresponding metadata, and best scenes are identified based on the identified events of interest. A video summary can be generated including one or more of the identified best scenes. The video summary can be generated using a video summary template with slots corresponding to video clips selected from among sets of candidate video clips. Best scenes can also be identified by receiving an indication of an event of interest within video from a user during the capture of the video. Metadata patterns representing activities identified within video clips can be identified within other videos, which can subsequently be associated with the identified activities.
-
Citations
15 Claims
-
1. A method for identifying events of interest in a captured video, the method comprising:
-
storing multiple stored speech patterns for multiple input types, the multiple stored speech patterns corresponding to a command for identifying the events of interest within the captured video, wherein the multiple stored speech patterns include a first stored speech pattern for a first input type, wherein storing the first stored speech pattern comprises; receiving, from a user, an input configuring a camera into a training mode to learn the first stored speech pattern; capturing the first stored speech pattern from the user; and storing the first stored speech pattern, wherein the first stored speech pattern is stored in response to capturing the first stored speech pattern from the user a threshold number of times; accessing a captured speech pattern, the captured speech pattern captured from the user during capture of the captured video; determining that the captured speech pattern corresponds to the first stored speech pattern; and in response to determining that the captured speech pattern corresponds to the first stored speech pattern, storing event of interest information in metadata associated with the captured video, the event of interest information identifying (i) the first input type for a first event of interest, and (ii) an event moment during the capture of the captured video at which the captured speech pattern was captured from the user. - View Dependent Claims (2, 3, 4, 5)
-
-
6. A system for identifying events of interest in a captured video, the system comprising:
a processor configured by instructions to; store multiple stored speech patterns for multiple input types, the multiple stored speech patterns corresponding to a command for identifying the events of interest within the captured video, wherein the multiple stored speech patterns include a first stored speech pattern for a first input type, wherein storing the first stored speech pattern comprises; receiving, from a user, an input configuring a camera into a training mode to learn the first stored speech pattern; capturing the first stored speech pattern from the user; and storing the first stored speech pattern, wherein the first stored speech pattern is stored in response to capturing the first stored speech pattern from the user a threshold number of times; access a captured speech pattern, the captured speech pattern captured from the user during capture of the captured video; determine that the captured speech pattern corresponds to the first stored speech pattern; and in response to determining that the captured speech pattern corresponds to the first stored speech pattern, store event of interest information in metadata associated with the captured video, the event of interest information identifying (i) the first input type for a first event of interest, and (ii) an event moment during the capture of the captured video at which the captured speech pattern was captured from the user. - View Dependent Claims (7, 8, 9, 10)
-
11. A non-transitory computer-readable storage medium storing instructions for identifying events of interest in a captured video, the instructions, when executed, causing a processor to:
-
store multiple stored speech patterns for multiple input types, the multiple stored speech patterns corresponding to a command for identifying the events of interest within the captured video, wherein the multiple stored speech patterns include a first stored speech pattern for a first input type, wherein storing the first stored speech pattern comprises; receiving, from a user, an input configuring a camera into a training mode to learn the first stored speech pattern; capturing the first stored speech pattern from the user; and storing the first stored speech pattern, wherein the first stored speech pattern is stored in response to capturing the first stored speech pattern from the user a threshold number of times; access a captured speech pattern, the captured speech pattern captured from the user during capture of the captured video; determine that the captured speech pattern corresponds to the first stored speech pattern; and in response to determining that the captured speech pattern corresponds to the first stored speech pattern, store event of interest information in metadata associated with the captured video, the event of interest information identifying (i) the first input type for a first event of interest, and (ii) an event moment during the capture of the captured video at which the captured speech pattern was captured from the user. - View Dependent Claims (12, 13, 14, 15)
-
Specification