METHOD FOR ESTIMATING A POSE OF AN ARTICULATED OBJECT MODEL
First Claim
1. A computer-implemented method for estimating a pose of an articulated object model (4), wherein the articulated object model (4) is a computer based 3D model (1) of a real world object (14) observed by one or more source cameras (9), and the articulated object model (4) represents a plurality of joints (2) and of links (3) that link the joints (2), and wherein the pose of the articulated object model (4) is defined by the spatial location of the joints (2), the method comprising the steps ofobtaining at least one source image (10) from a video stream comprising a view of the real world object (14) recorded by a source camera (9);
- processing the at least one source image (10) to extract a corresponding source image segment (13) comprising the view of the real world object (14) separated from the image background;
maintaining, in a database in computer readable form, a set of reference silhouettes, each reference silhouette being associated with an articulated object model (4) and with a particular reference pose of this articulated object model (4);
comparing the at least one source image segment (13) to the reference silhouettes and selecting a predetermined number of reference silhouettes by taking into account, for each reference silhouette,a matching error that indicates how closely the reference silhouette matches the source image segment (13) and/ora coherence error that indicates how much the reference pose is consistent with the pose of the same real world object (14) as estimated from at least one of preceding and following source images (10) of the video stream;
retrieving the reference poses of the articulated object models (4) associated with the selected of reference silhouettes; and
computing an estimate of the pose of the articulated object model (4) from the reference poses of the selected reference silhouettes.
2 Assignments
0 Petitions
Accused Products
Abstract
A computer-implemented method for estimating a pose of an articulated object model (4), wherein the articulated object model (4) is a computer based 3D model (1) of a real world object (14) observed by one or more source cameras (9), and wherein the pose of the articulated object model (4) is defined by the spatial location of joints (2) of the articulated object model (4), comprises the steps of
- obtaining a source image (10) from a video stream;
- processing the source image (10) to extract a source image segment (13);
- maintaining, in a database, a set of reference silhouettes, each being associated with an articulated object model (4) and a corresponding reference pose;
- comparing the source image segment (13) to the reference silhouettes and selecting reference silhouettes by taking into account, for each reference silhouette,
- a matching error that indicates how closely the reference silhouette matches the source image segment (13) and/or
- a coherence error that indicates how much the reference pose is consistent with the pose of the same real world object (14) as estimated from a preceding source image (10);
- retrieving the corresponding reference poses of the articulated object models (4); and
- computing an estimate of the pose of the articulated object model (4) from the reference poses of the selected reference silhouettes.
115 Citations
15 Claims
-
1. A computer-implemented method for estimating a pose of an articulated object model (4), wherein the articulated object model (4) is a computer based 3D model (1) of a real world object (14) observed by one or more source cameras (9), and the articulated object model (4) represents a plurality of joints (2) and of links (3) that link the joints (2), and wherein the pose of the articulated object model (4) is defined by the spatial location of the joints (2), the method comprising the steps of
obtaining at least one source image (10) from a video stream comprising a view of the real world object (14) recorded by a source camera (9); -
processing the at least one source image (10) to extract a corresponding source image segment (13) comprising the view of the real world object (14) separated from the image background; maintaining, in a database in computer readable form, a set of reference silhouettes, each reference silhouette being associated with an articulated object model (4) and with a particular reference pose of this articulated object model (4); comparing the at least one source image segment (13) to the reference silhouettes and selecting a predetermined number of reference silhouettes by taking into account, for each reference silhouette, a matching error that indicates how closely the reference silhouette matches the source image segment (13) and/or a coherence error that indicates how much the reference pose is consistent with the pose of the same real world object (14) as estimated from at least one of preceding and following source images (10) of the video stream; retrieving the reference poses of the articulated object models (4) associated with the selected of reference silhouettes; and computing an estimate of the pose of the articulated object model (4) from the reference poses of the selected reference silhouettes. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 15)
-
-
14. A computer-implemented method for rendering a virtual image (12) as seen from a virtual camera (11), given an articulated object model (4), wherein the articulated object model (4) is a computer based 3D model (1) of a real world object (14) observed by two or more source cameras (9), and the articulated object model (4) represents a plurality of joints (2) and of links (3) that link the joints (2), and wherein the pose of the articulated object model (4) is defined by the spatial location of the joints (2), the method comprising the steps of
determining an estimate of the 3D pose, that is, the 3D joint positions of the articulated object model (4); -
associating each link (3) with one or more projection surfaces (5), wherein the projection surfaces (5) are surfaces defined in the 3D model, and the position and orientation of each projection surface (5) is determined by the position and orientation of the associated link (3); wherein the projection surfaces (5), for each link (3), comprise a fan (7) of billboards (6), each billboard (6) being associated with a source camera (9), and each billboard being a planar surface spanned by its associated link (3) and a vector that is normal to both this link (3) and to a line connecting a point of the link (3) to the source camera (9); for each source camera (9), projecting segments of the associated source image (10) onto the associated billboard (6), creating billboard images; for each link (3), projecting the billboard images into the virtual image (12) and blending the billboard images to form a corresponding part of the virtual image (12).
-
Specification