Lip-password based speaker verification system
First Claim
1. A lip-based speaker verification system for identifying a speaker, comprising one or more computer processors for executing a process of verification of identity of the speaker using one modality of lip motions;
- wherein an identification key of the speaker comprising one or more passwords;
wherein the one or more passwords are embedded into lip motions of the speaker;
wherein the speaker is verified by underlying dynamic characteristics of the lip motions and extracted area-based features wherein the extracted area-based features further comprise teeth, tongue and oral cavity during the utterance; and
wherein the speaker is required to match the one or more passwords embedded in the lip motions with registered information in a database such that the matching between the dynamic characteristics of the speaker lip motions and the extracted area-based features with the one or more passwords is verified by using one or more multi-boosted hidden Markov models (HMMs);
wherein the process comprises the steps of;
(1) extracting visual features for each lip frame;
(2) performing lip motion segmentation of D to yield D={D1, D2, . . . , Dp} where D denotes the one or more passwords, and p is the number of password components;
(3) for each value of m=1, . . . , p, performing the steps of;
(3.1) getting a training set DmT={X1T, X2T, . . . , XNaT} of the speaker and DmI={X11, X21, . . . , XNbI} of an imposer, and forming a novel training set using a data sharing scheme (DSS);
(3.2) initializing wi,jT, wi,jI, r and ε
0 respectively with;
1 Assignment
0 Petitions
Accused Products
Abstract
A lip-based speaker verification system for identifying a speaker using a modality of lip motions; wherein an identification key of the speaker comprising one or more passwords; wherein the one or more passwords are embedded into lip motions of the speaker; wherein the speaker is verified by underlying dynamic characteristics of the lip motions; and wherein the speaker is required to match the one or more passwords embedded in the lip motions with registered information in a database. That is, in the case where the target speaker saying the wrong password or even in the case where an impostor knowing and saying the correct password, the nonconformities will be detected and the authentications/accesses will be denied.
16 Citations
11 Claims
-
1. A lip-based speaker verification system for identifying a speaker, comprising one or more computer processors for executing a process of verification of identity of the speaker using one modality of lip motions;
-
wherein an identification key of the speaker comprising one or more passwords; wherein the one or more passwords are embedded into lip motions of the speaker; wherein the speaker is verified by underlying dynamic characteristics of the lip motions and extracted area-based features wherein the extracted area-based features further comprise teeth, tongue and oral cavity during the utterance; and
wherein the speaker is required to match the one or more passwords embedded in the lip motions with registered information in a database such that the matching between the dynamic characteristics of the speaker lip motions and the extracted area-based features with the one or more passwords is verified by using one or more multi-boosted hidden Markov models (HMMs);wherein the process comprises the steps of; (1) extracting visual features for each lip frame; (2) performing lip motion segmentation of D to yield D={D1, D2, . . . , Dp} where D denotes the one or more passwords, and p is the number of password components; (3) for each value of m=1, . . . , p, performing the steps of; (3.1) getting a training set DmT={X1T, X2T, . . . , XN a T} of the speaker and DmI={X11, X21, . . . , XNb I} of an imposer, and forming a novel training set using a data sharing scheme (DSS);(3.2) initializing wi,jT, wi,jI, r and ε
0 respectively with; - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11)
-
Specification