Systems and methods for remote control of interactive video
First Claim
1. A system to control interaction with a first source of video data and a second source of associated metadata, the video data containing video images including a plurality of objects, the metadata associated with at least one of the plurality of objects, the system comprising:
- a video device configured to process the first source of video data and generate the video images, the video device further configured to use artificial image recognition methods to automatically identify object features in the video images and use these object features to identify metadata in the second source of metadata that is associated with the identified objects;
and a portable device configured to communicate with the video device and to display the video images, the portable device further configured to synchronize the display of video images with the video device, the portable device further configured to receive spatio-temporal video image coordinate input from a user that selects an object in the video image, and wherein the portable device communicates the selected object to the video device, receives metadata from said second source that is associated with the selected object, and displays the received metadata to the user.
3 Assignments
0 Petitions
Accused Products
Abstract
The present invention describes a system and method for controlling interactive video, including a remote control device allowing the user to interact with the video. The method allows bringing the “lean-forward” hypervideo experience to a “lay-back” environment such as the TV screen. The method also allows for the creation of a single system that combines both the social aspects of the “lay-back” and the “lean-forward experience” into a new type of rich media viewing experience.
20 Citations
20 Claims
-
1. A system to control interaction with a first source of video data and a second source of associated metadata, the video data containing video images including a plurality of objects, the metadata associated with at least one of the plurality of objects, the system comprising:
-
a video device configured to process the first source of video data and generate the video images, the video device further configured to use artificial image recognition methods to automatically identify object features in the video images and use these object features to identify metadata in the second source of metadata that is associated with the identified objects; and a portable device configured to communicate with the video device and to display the video images, the portable device further configured to synchronize the display of video images with the video device, the portable device further configured to receive spatio-temporal video image coordinate input from a user that selects an object in the video image, and wherein the portable device communicates the selected object to the video device, receives metadata from said second source that is associated with the selected object, and displays the received metadata to the user. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8)
-
-
9. A method of controlling interaction with video data received from a first source, and associated metadata received from a second source, the video data containing video images, the method comprising:
-
using artificial image recognition methods to process the video images on a video device to automatically identify a plurality of object features in the video images;
synchronizing the video images with a portable device having a video display;receiving user spatio-temporal video image coordinate input from the portable device, wherein the user input selects an object in the video image; using said object features to identify metadata in said second source of metadata that is associated with the selected object; and communicating at least a portion of the identified metadata to the portable device for display on the portable device video display. - View Dependent Claims (10, 11, 12, 13, 14, 15, 16, 17, 18)
-
-
19. A method comprising:
-
receiving identification of a selected object from a portable device; using artificial image recognition methods and user spatio-temporal video image coordinate input from said portable device to automatically analyze a first source of video data to identify a plurality of image features; wherein said artificial image recognition methods comprise one or more operations selected from the group consisting of Smallest Univalue Segment Assimilating Nucleus (SUSAN) invariant feature detection, scale-invariant feature transform (SIFT) descriptor detection, and Random Sample Consensus (RANSAC) operations; communicating the image features to a remote server; receiving a list of objects associated with the image features from the remote server; receiving a second source of metadata associated with the list of objects from the remote server; and communicating the list of objects and the video data to said portable device. - View Dependent Claims (20)
-
Specification