Method and system for simulated interactive conversation
First Claim
Patent Images
1. A computer-implemented method of simulating interactive communication between a user and a human subject, comprising:
- assigning at least one phrase to a stored content sequence, wherein the content sequence comprises a content clip of the subject, the subject being a human recorded on video, the content clip including a contemporaneously-recorded head and mouth of the subject and contemporaneously-recorded audio of the subject, wherein the content clip is free of any superimposed facial features;
parsing the at least one phrase to produce at least one phonetic clone;
associating the at least one phonetic clone with the stored content sequence;
creating a transition between the content clip and the second content sequence by frame-matching a frame of the stored content sequence, the content sequence including the human subject speaking, with a frame of a second content sequence, the frame-matching being performed with respect to the recorded video of the entire head and facial features of the human subject;
receiving an utterance from the user;
matching the utterance to the at least one phonetic clone; and
in response to matching the utterance, displaying the stored content sequence associated with the at least one phonetic clone in succession with the second content sequence.
1 Assignment
0 Petitions
Accused Products
Abstract
A method of simulating interactive communication between a user and a human subject. The method comprises: assigning at least one phrase to a stored content sequence, wherein the content sequence comprises a content clip of the subject; parsing the at least one phrase to produce at least one phonetic clone; associating the at least one phonetic clone with the stored content sequence; receiving an utterance from the user; matching the utterance to the at least one phonetic clone; and displaying the stored content sequence associated with the at least one phonetic clone.
109 Citations
33 Claims
-
1. A computer-implemented method of simulating interactive communication between a user and a human subject, comprising:
-
assigning at least one phrase to a stored content sequence, wherein the content sequence comprises a content clip of the subject, the subject being a human recorded on video, the content clip including a contemporaneously-recorded head and mouth of the subject and contemporaneously-recorded audio of the subject, wherein the content clip is free of any superimposed facial features; parsing the at least one phrase to produce at least one phonetic clone; associating the at least one phonetic clone with the stored content sequence; creating a transition between the content clip and the second content sequence by frame-matching a frame of the stored content sequence, the content sequence including the human subject speaking, with a frame of a second content sequence, the frame-matching being performed with respect to the recorded video of the entire head and facial features of the human subject; receiving an utterance from the user; matching the utterance to the at least one phonetic clone; and in response to matching the utterance, displaying the stored content sequence associated with the at least one phonetic clone in succession with the second content sequence. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19)
-
-
20. A system for simulating interactive communication between a user and a human subject, the system comprising:
-
a display for displaying the subject, the subject being a human recorded on video; a memory; and a processor, coupled to the memory and the display, the processor operable to; assign at least one phrase to a stored content sequence of the subject, wherein the content sequence comprises a content clip of the subject, the content sequence including contemporaneously-recorded audio of the human that is played simultaneously with the video of the human, wherein the video of the human includes a contemporaneously-recorded head and mouth of the human, wherein the content clip is free of any superimposed facial features; parse the at least one phrase to produce at least one phonetic clone of the at least one phrase; associate the at least one phonetic clone with the stored content sequence; create a transition between the content clip and the second content sequence by frame-matching a frame of the stored content sequence, the content sequence including the human subject speaking, with a frame of a second content sequence, the frame-matching being performed with respect to the recorded video of the entire head and facial features of the human subject; receive an utterance from the user; match the utterance to the at least one phonetic clone; and in response to the match, display the stored content sequence associated with the at least one phonetic clone in succession with the second content sequence.
-
-
21. A computer-implemented method of simulating interactive communication between a user and a human subject, comprising:
-
receiving a voice input from the user; matching the voice input to one of a plurality of a stored phonetic clones, the phonetic clones each corresponding to a target speech phrase associated with a stored content sequence file depicting the subject, the number of stored phonetic clones being greater than the number of stored content sequence files, the subject being a human recorded on video, the content sequence including contemporaneously-recorded audio of the human that is played simultaneously with the video of the human, wherein the video of the human includes a contemporaneously-recorded head and mouth of the human, wherein the content clip is free of any superimposed facial features; creating a transition between the content clip and the second content sequence by frame-matching a frame of the stored content sequence, the content sequence including the human subject speaking, with a frame of a second content sequence, the frame-matching being performed with respect to the recorded video of the entire head and facial features of the human subject; and in response to the matching, displaying the stored content sequence file matched to the phonetic clone and the second content sequence in succession. - View Dependent Claims (22, 23, 24, 25, 26, 27, 28, 29, 30)
-
-
31. A conversation system for simulating interactive communication between a user and a first human subject and a second human subject, comprising:
-
a display for displaying the first and second subjects, the first and second subjects being humans recorded on video; a memory; and a processor, coupled to the memory and the display, the processor operable to; receive a voice input from the user; match the voice input to one of a plurality of a stored phonetic clones, a first portion of the phonetic clones each corresponding to a target speech phrase associated with a stored content sequence file depicting the first subject and a second portion of the phonetic clones each corresponding to a target speech phrase associated with a stored content sequence file depicting the second subject, the number of stored phonetic clones being greater than the number of stored content sequence files, the content sequence including a contemporaneously-recorded head and mouth of the second subject and contemporaneously-recorded audio of the second subject, wherein the content clip is free of any superimposed facial features; in response to the match, display the stored content sequence file matched to the phonetic clone in succession with a second content sequence; and frame-match a frame of the stored content sequence with a frame of the second content sequence, the stored content sequence including a human subject speaking, the frame-matching being performed with respect to the recorded video of the entire head and facial features of the second human subject to create a transition for the stored content clip. - View Dependent Claims (32, 33)
-
Specification