Using automated content analysis for audio/video content consumption
First Claim
Patent Images
1. An audio/video (A/V) processing system, comprising:
- an automatic content analyzer configured to receive A/V content and analyze the A/V content using speech recognition and natural language processing to generate speech metadata and natural language metadata corresponding to the A/V content; and
a player configured to display a plurality of different metadata displays, the metadata displays displaying information based on the speech and natural language metadata, and being indicative of where in the A/V content a speaker is speaking or a subject matter is mentioned.
2 Assignments
0 Petitions
Accused Products
Abstract
Audio/video (A/V) content is analyzed using speech and language analysis components. Metadata is automatically generated based upon the analysis. The metadata is used in generating user interface interaction components which allow a user to view subject matter in various segments of the A/V content and to interact with the A/V content based on the automatically generated metadata.
131 Citations
20 Claims
-
1. An audio/video (A/V) processing system, comprising:
-
an automatic content analyzer configured to receive A/V content and analyze the A/V content using speech recognition and natural language processing to generate speech metadata and natural language metadata corresponding to the A/V content; and a player configured to display a plurality of different metadata displays, the metadata displays displaying information based on the speech and natural language metadata, and being indicative of where in the A/V content a speaker is speaking or a subject matter is mentioned. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10)
-
-
11. An audio/video (A/V) processing system, comprising:
-
an automatic content analyzer configured to receive A/V content and analyze the A/V content using speech recognition and natural language processing to generate speech metadata and natural language metadata corresponding to the A/V content; and a player configured to play the A/V content and to generate a user interface that receives community metadata, corresponding to the A/V content, from one or more users and displays a plurality of different metadata displays, the metadata displays displaying information based on the speech metadata, the natural language metadata and the community metadata, wherein the community metadata is associated with a position in the A/V content such that the metadata display generated based on the community metadata changes as the player plays the A/V content. - View Dependent Claims (12, 13, 14, 15)
-
-
16. A computer implemented method of displaying audio/video (A/V) content, comprising:
-
receiving user A/V content preferences; analyzing a plurality of different portions of A/V content using automatic speech recognition and natural language processing to generate subject matter metadata and speaker related metadata corresponding to different points in the A/V content; selecting portions of interest to the user based on the user A/V content preferences; and concatenating the selected portions of interest together to generate a user-specific A/V program. - View Dependent Claims (17, 18, 19, 20)
-
Specification