Providing virtual inserts using image tracking with camera and position sensors
First Claim
1. A processor-implemented method for processing image data, comprising the processor-implemented steps of:
- obtaining an image of a live event, the image includes a pixel representation of an object which is moving in the live event, the object have an appearance, and the image is captured at a point in time in the live event by at least one camera;
determining an orientation of the object in the live event for the point in time;
determining a position of the object in the live event for the point in time;
based on the position, determining a distance between the at least one camera and the object for the point in time;
based on the distance and a zoom setting of the at least one camera, estimating a size of the pixel representation of the object in the image;
based on the position and registration data of the at least one camera, estimating a position of the pixel representation of the object in the image;
providing a template comprising a model which mimics the appearance of the object, the model has an orientation which corresponds to the determined orientation of the object, and the template has a size which corresponds to the estimated size of the pixel representation of the object in the image; and
detecting the pixel representation of the object in the image by correlating the template with different candidate pixel representations of the object in the image to select an optimum candidate pixel representation of the object in the image, the candidate pixel representations are based on the estimated position of the pixel representation of the object in the image.
9 Assignments
0 Petitions
Accused Products
Abstract
Camera registration and/or sensor data is updated during a live event by determining a difference between an estimated position of an object in an image and an actual position of the object in the image. The estimated position of the object in the image can be based on an estimated position of the object in the live event, e.g., based on GPS or other location data. This position is transformed to the image space using current camera registration and/or sensor data. The actual position of the object in the image can be determined by template matching which accounts for an orientation of the object, a shape of the object, an estimated size of the representation of the object in the image, and the estimated position of the object in the image. The updated camera registration/sensor data can be used in detecting an object in a subsequent image.
96 Citations
25 Claims
-
1. A processor-implemented method for processing image data, comprising the processor-implemented steps of:
-
obtaining an image of a live event, the image includes a pixel representation of an object which is moving in the live event, the object have an appearance, and the image is captured at a point in time in the live event by at least one camera; determining an orientation of the object in the live event for the point in time; determining a position of the object in the live event for the point in time; based on the position, determining a distance between the at least one camera and the object for the point in time; based on the distance and a zoom setting of the at least one camera, estimating a size of the pixel representation of the object in the image; based on the position and registration data of the at least one camera, estimating a position of the pixel representation of the object in the image; providing a template comprising a model which mimics the appearance of the object, the model has an orientation which corresponds to the determined orientation of the object, and the template has a size which corresponds to the estimated size of the pixel representation of the object in the image; and detecting the pixel representation of the object in the image by correlating the template with different candidate pixel representations of the object in the image to select an optimum candidate pixel representation of the object in the image, the candidate pixel representations are based on the estimated position of the pixel representation of the object in the image. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16)
-
-
17. A system for processing image data, comprising:
-
at least one camera for capturing an image of a live event at a point in time, the live event including an object; and at least one processing facility which receives the image, the at least one processing facility;
a) transforms an estimated position of the object in the live event at the point in time to an estimated position in the image based on registration data of the at least one camera, and b) detects a pixel representation of the object in the image based on data indicating an orientation of the object in the live event for the point in time, a shape of the object, an estimated size of the pixel representation of the object in the image, and the estimated position of the object in the image. - View Dependent Claims (18, 19, 20, 21, 22)
-
-
23. A processor-implemented method for processing image data, comprising the processor-implemented steps of:
-
obtaining successive images of a live event which are captured by at least one camera, the live event including an object; for each image, transforming an estimated position of the object in the live event to an estimated position in the image based on registration data of the at least one camera; for each image, detecting a representation of the object in the image and detecting an offset in pixel space of a position of the representation of the object in the image from the estimated position of the object in the image; and for at least one of the images, filtering the associated offset based on at least one offset of a past and/or future image. - View Dependent Claims (24, 25)
-
Specification