Method and apparatus for recognizing identity of individuals employing synchronized biometrics
First Claim
1. A method for recognizing an individual based on attributes associated with the individual, comprising the steps of:
- pre-storing at least two distinctive attributes of the individual during at least one enrollment session;
contemporaneously extracting the at least two distinctive attributes from the individual during a common recognition session;
segmenting the pre-stored attributes and the extracted attributes according to a sequence of segmentation units;
indexing the segmented pre-stored and extracted attributes so that the segmented pre-stored and extracted attributes corresponding to an identical segmentation unit in the sequence of segmentation units are associated to an identical index; and
respectively comparing the segmented pre-stored and extracted attributes associated to the identical index to each other to recognize the individual wherein at least one of the at least two distinctive attributes is lip shape and the comparing step is performed by a lip reading system and a lip recognition system.
2 Assignments
0 Petitions
Accused Products
Abstract
A method for recognizing an individual based on attributes associated with the individual comprises the steps of: pre-storing at least two distinctive attributes of the individual during at least one enrollment session; contemporaneously extracting the at least two distinctive attributes from the individual during a common recognition session; segmenting the pre-stored attributes and the extracted attributes according to a sequence of segmentation units; indexing the segmented pre-stored and extracted attributes so that the segmented pre-stored and extracted attributes corresponding to an identical segmentation unit in the sequence of segmentation units are associated to an identical index; and respectively comparing the segmented pre-stored and extracted attributes associated to the identical index to each other to recognize the individual.
375 Citations
43 Claims
-
1. A method for recognizing an individual based on attributes associated with the individual, comprising the steps of:
-
pre-storing at least two distinctive attributes of the individual during at least one enrollment session;
contemporaneously extracting the at least two distinctive attributes from the individual during a common recognition session;
segmenting the pre-stored attributes and the extracted attributes according to a sequence of segmentation units;
indexing the segmented pre-stored and extracted attributes so that the segmented pre-stored and extracted attributes corresponding to an identical segmentation unit in the sequence of segmentation units are associated to an identical index; and
respectively comparing the segmented pre-stored and extracted attributes associated to the identical index to each other to recognize the individual wherein at least one of the at least two distinctive attributes is lip shape and the comparing step is performed by a lip reading system and a lip recognition system. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28)
(a) decoding synchronized utterances of audio and video data corresponding to the individual via application of a joint audio-video joint likelihood function to produce a decoded password phrase W, the joint likelihood function being a combination of likelihood functions for audio and video components;
(b) reducing an effect of the video component of the joint audio-video likelihood function;
(c) decoding the video data utilizing only the audio component of the joint likelihood function to produce a decoded password phrase V;
(d) adjusting the audio component of the joint likelihood function to produce a better match of the decoded password phrase V to the decoded password phrase W;
(e) decoding the synchronized utterances of audio and video data utilizing an iteratively updated joint audio-video likelihood function that includes the video component and the adjusted audio component to produce a new decoded password phrase W′
; and
(f) iteratively repeating steps (b) through (e).
-
-
16. The method of claim 15, wherein the adjustment step further comprises the step of comparing the decoded password phrase V obtained via the joint audio-video likelihood function to the audio and video components from a previous iteration to produce a better match to the decoded password phrase W.
-
17. The method of claim 16, wherein the adjustment step further comprises the step of modifying some parameters of the video component of the joint likelihood function.
-
18. The method of claim 15, wherein the repeating step repeats steps (b) through (e) until a match between the decoded password phrase W′
- and the correct password phrase is established.
-
19. The method of claim 18, wherein if the match is a full match, then the individual is considered recognized.
-
20. The method of claim 18, wherein if the match is one of varying and degrading with each iteration, then the individual is considered to be an imposter.
-
21. The method of claim 18, wherein if the match is one of varying and degrading with each iteration, then the data are not properly indexed.
-
22. The method of claim 15, wherein the repeating step repeats steps (b) through (e) until the decoded password phrase W′
- does not change.
-
23. The method of claim 15, wherein the audio and video data are produced over small time segments.
-
24. The method of claim 15, wherein the video component corresponds to likelihood functions for lip contour geometry and mouth region.
-
25. The method of claim 8, wherein two of the at least two attributes correspond to audio and video characteristics of the individual, the recognition of the individual depending on the individual correctly speaking a password phrase, the segmenting and indexing steps comprising the steps of:
-
(a) decoding synchronized utterances of audio and video data corresponding to the individual via application of a joint audio-video joint likelihood function to produce a decoded password phrase W, the joint likelihood function being a combination of likelihood functions for audio and video components;
(b) reducing an effect of the audio component of the joint audio-video likelihood function;
(c) decoding the audio data utilizing only the video component of the joint likelihood function to produce a decoded password phrase A;
(d) adjusting the video component of the joint likelihood function to produce a better match of the decoded password phrase A to the decoded password phrase W;
(e) decoding the synchronized utterances of audio and video data utilizing an iteratively updated joint audio-video likelihood function that includes the audio component and the adjusted video component to produce a new decoded password phrase W′
; and
(f) iteratively repeating steps (b) through (e).
-
-
26. The method of claim 8, wherein two of the at least two attributes correspond to audio and video characteristics of the individual, the recognition of the individual depending on the individual correctly speaking a password phrase, the segmenting and indexing steps comprising the steps of:
-
(a) decoding synchronized utterances of audio and video data corresponding to the individual via application of a joint audio-video joint likelihood function to produce a decoded password phrase W, the joint likelihood function being a combination of likelihood functions for audio and video components;
(b) reducing an effect of the video component of the joint audio-video likelihood function;
(c) decoding the video data utilizing only the audio component of the joint likelihood function to produce a decoded password phrase V;
(d) adjusting the audio component of the joint likelihood function to produce a better match of the decoded password phrase V to the decoded password phrase W;
(e) reducing an effect of the audio component of the joint audio-video likelihood function;
(f) decoding the audio data utilizing only the video component of the joint likelihood function to produce a decoded password phrase A;
(g) adjusting the video component of the joint likelihood function to produce a better match of the decoded password phrase A to the decoded password phrase W;
(h) decoding the synchronized utterances of audio and video data utilizing an iteratively updated joint audio-video likelihood function that includes the adjusted audio and video components to produce a new decoded password phrase W′
; and
(i) iteratively repeating steps (b) through (h).
-
-
27. The method of claim 26, wherein the repeating step repeats steps (b) through (h) until a match between the decoded password phrase W′
- and the correct password phrase is established.
-
28. The method of claim 26, wherein the repeating step repeats steps (b) through (h) until the decoded password phrase W′
- does not change.
-
29. A method for recognizing an individual based on attributes associated with the individual, comprising the steps of:
-
pre-storing at least two distinctive attributes of the individual during at least one enrollment session;
contemporaneously extracting the at least two distinctive attributes from the individual during a common recognition session;
segmenting the pre-stored attributes and the extracted attributes according to a sequence of segmentation units;
indexing the segmented pre-stored and extracted attributes so that the segmented pre-stored and extracted attributes corresponding to an identical segmentation unit in the sequence of segmentation units are associated to an identical index; and
respectively comparing the segmented pre-stored and extracted attributes associated to the identical index to each other to recognize the individual wherein three of the at least two distinctive attributes are pen pressure, pen velocity, and hand geometry.
-
-
30. An apparatus for recognizing an individual based on attributes associated with the individual, comprising:
-
a store for pre-storing at least two distinctive attributes of the individual during at least one enrollment session;
contemporaneous extraction means for contemporaneously extracting the at least two distinctive attributes from the individual during a common recognition session;
segmentation means for segmenting the pre-stored attributes and the extracted attributes according to a sequence of segmentation units;
indexing means for indexing the segmented pre-stored and extracted attributes so that the segmented pre-stored and extracted attributes corresponding to an identical segmentation unit in the sequence of segmentation units are associated to an identical index; and
comparing means for respectively comparing the segmented pre-stored and extracted attributes associated to the identical index to each other to recognize the individual, wherein at least one of the at least two distinctive attributes is lip shape and said comparing means comprises a lip reading system and a lip recognition system. - View Dependent Claims (31, 32, 33)
(a) an audio and video joint likelihood function decoder for decoding synchronized utterances of audio and video data corresponding to the individual via application of a joint audio-video joint likelihood function to produce a decoded password phrase W, the joint likelihood function being a combination of likelihood functions for audio and video components;
(b) an audio likelihood function decoder operatively coupled to the audio and video joint likelihood function decoder for decoding the video data utilizing only the audio component of the joint likelihood function to produce a decoded password phrase V;
(c) a video likelihood function decoder operatively coupled to the audio and video joint likelihood function decoder for decoding the audio data utilizing only the video component of the joint likelihood function to produce a decoded password phrase A; and
(d) an adjustment module operatively coupled to the audio and video joint likelihood function decoder, the audio likelihood function decoder and the video likelihood function decoder for adjusting the audio component of the joint likelihood function to produce a better match of the decoded password phrase V to the decoded password phrase W, and for adjusting the video component of the joint likelihood function to produce a better match of the decoded password phrase A to the decoded password phrase W.
-
-
34. An apparatus employing synchronized speaker recognition, lip reading and lip recognition for recognizing an individual, comprising:
-
a speaker recognition system for performing speaker recognition;
a lip reading system for performing lip reading;
a lip recognition system for performing lip recognition;
a camera and pointing system for generating images of the individual for use by the lip reading and lip recognition systems, the speaker recognition system, the lip reading system, the lip recognition system, and the camera and pointing system contemporaneously extracting biometric attributes corresponding to speaker recognition, lip reading and lip recognition;
a store operatively coupled to the speaker recognition system, the lip reading system, and the lip recognition system for pre-storing biometric attributes during at least one enrollment session, the pre-stored attributes respectively corresponding to speaker recognition, lip reading and lip recognition;
segmentation means for segmenting the pre-stored attributes and the extracted attributes according to a sequence of segmentation units;
indexing means for indexing the segmented pre-stored and extracted attributes so that the segmented pre-stored and extracted attributes corresponding to an identical segmentation unit in the sequence of segmentation units are associated to an identical index, the speaker recognition system, the lip reading system, and the lip recognition system respectively comparing the segmented pre-stored and extracted attributes associated to the identical index to each other to recognize the individual; and
a controller for processing the results of the comparisons such that the individual is considered recognized if the segmented pre-stored attributes associated to the identical index substantially match the segmented extracted attributes associated to the identical index. - View Dependent Claims (35, 36, 37, 38, 39, 40, 41, 42, 43)
-
Specification