Method and system for facial recognition for a videoconference
First Claim
Patent Images
1. A method comprising:
- identifying a participant from audio information, wherein identifying the participant from the audio information comprises;
performing a feature extraction and a speaker segmentation on the audio information to determine a voice model, andcomparing the determined voice model from the audio information with a plurality of voice models stored in a database to identify the participant;
identifying the participant in video information, wherein identifying the participant in the video information comprises;
identifying a plurality of facial images in the video information;
determining a one of the plurality of facial images in the video information as having the most movement as compared to others of the plurality of facial images; and
identifying the participant as the determined one of the plurality of facial images;
capturing, from the video information, a plurality of images of the participant identified in the video information wherein ones of the plurality of captured images of the participant include respective different expressions of a face of the participant and wherein other ones of the plurality of captured images of the participant include respective different illumination conditions of the face of the participant;
associating a unique identifier with the captured plurality of images, the unique identifier corresponding to the participant identified from the audio information; and
saving the captured plurality of images and the associated unique identifier in the database.
1 Assignment
0 Petitions
Accused Products
Abstract
Videoconferencing may be provided. A participant may be identified from audio information and in video information. From the video information, a plurality of images may be captured of the participant identified in the video information. A unique identifier may be associated with the captured plurality of images. The unique identifier may correspond to the participant identified from the audio information. The captured plurality of images and the associated unique identifier may be saved in a database.
-
Citations
23 Claims
-
1. A method comprising:
-
identifying a participant from audio information, wherein identifying the participant from the audio information comprises; performing a feature extraction and a speaker segmentation on the audio information to determine a voice model, and comparing the determined voice model from the audio information with a plurality of voice models stored in a database to identify the participant; identifying the participant in video information, wherein identifying the participant in the video information comprises; identifying a plurality of facial images in the video information; determining a one of the plurality of facial images in the video information as having the most movement as compared to others of the plurality of facial images; and identifying the participant as the determined one of the plurality of facial images; capturing, from the video information, a plurality of images of the participant identified in the video information wherein ones of the plurality of captured images of the participant include respective different expressions of a face of the participant and wherein other ones of the plurality of captured images of the participant include respective different illumination conditions of the face of the participant; associating a unique identifier with the captured plurality of images, the unique identifier corresponding to the participant identified from the audio information; and saving the captured plurality of images and the associated unique identifier in the database. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16)
-
-
17. An apparatus comprising:
-
a memory storage; and a processing unit coupled to the memory storage, wherein the processing unit is operative to; identify a participant from video information in a teleconference, wherein the processing unit being operative to identify the participant in the video information comprises the processing unit being operative to; identify a plurality of facial images in the video information, determine a one of the plurality of facial images in the video information as having the most movement as compared to others of the plurality of facial images, and identify the participant as the determined one of the plurality of facial images; capture, from the video information in the teleconference, a plurality of images of the participant identified in the video information wherein ones of the plurality of captured images of the participant include respective different expressions of a face of the participant and wherein other ones of the plurality of captured images of the participant include respective different illumination conditions of the face of the participant; associate a unique identifier with the captured plurality of images, the unique identifier corresponding to the participant identified from audio information in the teleconference, wherein the processing unit being operative to associate the unique identifier comprises the processing unit being operative to; perform a feature extraction and a speaker segmentation on the audio information to determine a voice model, and compare the determined voice model from the audio information with a plurality of voice models stored in a database to identify the participant; receive participant information corresponding to the unique identifier; and save the captured plurality of images and the associated participant information in a database. - View Dependent Claims (18, 19, 20)
-
-
21. A non-transitory computer-readable medium that stores a set of instructions which when executed perform a method comprising:
-
identifying a participant in video information, wherein identifying the participant in the video information comprises; identifying a plurality of facial images in the video information; determining a one of the plurality of facial images in the video information as having the most movement as compared to others of the plurality of facial images; and identifying the participant as the determined one of the plurality of facial images; capturing, from the video information, a plurality of images of the participant identified in the video information wherein ones of the plurality of captured images of the participant include respective different expressions of a face of the participant and wherein other ones of the plurality of captured images of the participant include respective different illumination conditions of the face of the participant; associating a unique identifier with the captured plurality of images, the unique identifier corresponding to the participant identified from audio information, wherein associating the unique identifier with the captured plurality of images comprises; performing a feature extraction and a speaker segmentation on the audio information to determine a voice model, and comparing the determined voice model from the audio information with a plurality of voice models stored in a database to identify the participant; and saving the captured plurality of images and the associated unique identifier in the database. - View Dependent Claims (22, 23)
-
Specification