COMPUTER GENERATED EMULATION OF A SUBJECT
First Claim
Patent Images
1. A system for emulating a subject, to allow a user to interact with a computer generated talking head with the subject'"'"'s face and voice;
- said system comprising a processor, a user interface and a personality storage section,the user interface being configured to emulate the subject, by displaying a talking head which comprises the subject'"'"'s face and output speech from the mouth of the face with the subject'"'"'s voice, the user interface further comprising a receiver for receiving a query from the user, the emulated subject being configured to respond to the query received from the user,the processor comprising a dialogue section and a talking head generation section,wherein said dialogue section is configured to generate a response to a query inputted by a user from the user interface and generate a response to be outputted by the talking head, the response being generated by retrieving information from said personality storage section, said personality storage section comprising content created by or about the subject,and said talking head generation section is configured to;
convert said response into a sequence of acoustic units, the talking head generation section further comprising a statistical model, said statistical model comprising a plurality of model parameters, said model parameters being derived from said personality storage section, the model parameters describing probability distributions which relate an acoustic unit to an image vector and speech vector, said image vector comprising a plurality of parameters which define the subject'"'"'s face and said speech vector comprising a plurality of parameters which define the subject'"'"'s voice, the talking head generation section being further configured to output a sequence of speech vectors and image vectors which are synchronised such that the head appears to talk.
1 Assignment
0 Petitions
Accused Products
Abstract
A system for emulating a subject, to allow a user to interact with a computer generated talking head with the subject'"'"'s face and voice;
- said system comprising a processor, a user interface and a personality storage section,
- the user interface being configured to emulate the subject, by displaying a talking head which comprises the subject'"'"'s face and output speech from the mouth of the face with the subject'"'"'s voice, the user interface further comprising a receiver for receiving a query from the user, the emulated subject being configured to respond to the query received from the user,
- the processor comprising a dialogue section and a talking head generation section,
- wherein said dialogue section is configured to generate a response to a query inputted by a user from the user interface and generate a response to be outputted by the talking head, the response being generated by retrieving information from said personality storage section, said personality storage section comprising content created by or about the subject,
- and said talking head generation section is configured to:
- convert said response into a sequence of acoustic units, the talking head generation section further comprising a statistical model, said statistical model comprising a plurality of model parameters, said model parameters being derived from said personality storage section, the model parameters describing probability distributions which relate an acoustic unit to an image vector and speech vector, said image vector comprising a plurality of parameters which define the subject'"'"'s face and said speech vector comprising a plurality of parameters which define the subject'"'"'s voice, the talking head generation section being further configured to output a sequence of speech vectors and image vectors which are synchronised such that the head appears to talk.
-
Citations
20 Claims
-
1. A system for emulating a subject, to allow a user to interact with a computer generated talking head with the subject'"'"'s face and voice;
-
said system comprising a processor, a user interface and a personality storage section, the user interface being configured to emulate the subject, by displaying a talking head which comprises the subject'"'"'s face and output speech from the mouth of the face with the subject'"'"'s voice, the user interface further comprising a receiver for receiving a query from the user, the emulated subject being configured to respond to the query received from the user, the processor comprising a dialogue section and a talking head generation section, wherein said dialogue section is configured to generate a response to a query inputted by a user from the user interface and generate a response to be outputted by the talking head, the response being generated by retrieving information from said personality storage section, said personality storage section comprising content created by or about the subject, and said talking head generation section is configured to; convert said response into a sequence of acoustic units, the talking head generation section further comprising a statistical model, said statistical model comprising a plurality of model parameters, said model parameters being derived from said personality storage section, the model parameters describing probability distributions which relate an acoustic unit to an image vector and speech vector, said image vector comprising a plurality of parameters which define the subject'"'"'s face and said speech vector comprising a plurality of parameters which define the subject'"'"'s voice, the talking head generation section being further configured to output a sequence of speech vectors and image vectors which are synchronised such that the head appears to talk. - View Dependent Claims (2, 3, 4, 5, 6, 11, 12, 13, 14, 15, 16, 17, 18, 19)
-
-
7. A system for creating a response to an inputted user query, said system comprising:
-
a personality file storage section, said personality file storage section comprising a plurality of documents stored in an unstructured form; a query conversion section configured to convert said query into a word vector; a first comparison section configured to compare said word vector generated from said query with word vectors generated from the documents in said personality file storage section and output identified documents; a second comparison section configured to compare said word vector selected from said query and passages from said identified documents and to rank said selected passages, said ranking being based on the number of matches between said selected passage and said query; and a concatenation section adapted to concatenate selected passages together using sentence connectors, wherein said sentence connectors are chosen from a plurality of sentence connectors, said sentence connectors being chosen on the basis of a statistical model. - View Dependent Claims (8, 9, 10)
-
-
20. A method for emulating a subject, to allow a user to interact with a computer generated talking head with the subject'"'"'s face and voice;
-
the method comprising; receiving a user inputted query; generating a response to a query inputted by a user from the user interface and generate a response to be outputted by the talking head, the response being generated by retrieving information from said personality storage section, said personality storage section comprising content created by or about the subject; and outputting said response by displaying a talking head which comprises the subject'"'"'s face and output speech from the mouth of the face with the subject'"'"'s voice, wherein said talking head outputs said response by; converting said response into a sequence of acoustic units using a statistical model, said statistical model comprising a plurality of model parameters, the model parameters describing probability distributions which relate an acoustic unit to an image vector and speech vector, said image vector comprising a plurality of parameters which define the subject'"'"'s face and said speech vector comprising a plurality of parameters which define the subject'"'"'s voice, the talking head appearing to talk by outputting a sequence of speech vectors and image vectors which are synchronised.
-
Specification