Voice-based determination of physical and emotional characteristics of users
First Claim
Patent Images
1. A speaker device comprising:
- a microphone;
at least one memory that stores computer-executable instructions;
at least one processor configured to access the at least one memory and execute the computer-executable instructions to;
receive, using the microphone, first voice input from a user comprising a user utterance;
determine background noise in the first voice data;
determine that the user is in an ambient environment with multiple users;
generate a first tag indicative of a multiple user audience;
process the first voice data of the first voice input using a first signal processing algorithm;
determine that a physical status of the user is abnormal;
select a sore throat physical status for the user;
generate a second tag indicative of the sore throat physical status;
apply a second signal processing algorithm to the first voice data;
determine that an emotional status of the user indicates the user is excited;
select an excited emotional status for the user;
generate a third tag indicative of the excited emotional status;
send a content request comprising the first voice data, the first tag, the second tag, and the third tag to a server, wherein the server determines first audio content for presentation at the speaker device;
receive an indication of the first audio content; and
present the first audio content, wherein targeting criteria for the first audio content comprises the sore throat physical status, the excited emotional status, and the multiple user audience.
1 Assignment
0 Petitions
Accused Products
Abstract
Systems, methods, and computer-readable media are disclosed for voice-based determination of physical and emotional characteristics of users. Example methods may include determining first voice data, wherein the first voice data is generated by a user, determining a first real-time user status of the user using the first voice data, generating a first data tag indicative of the first real-time user status, determining first audio content for presentation at a speaker device using the first data tag and the first voice data, and causing presentation of the first audio content via a speaker of the speaker device.
168 Citations
19 Claims
-
1. A speaker device comprising:
-
a microphone; at least one memory that stores computer-executable instructions; at least one processor configured to access the at least one memory and execute the computer-executable instructions to; receive, using the microphone, first voice input from a user comprising a user utterance; determine background noise in the first voice data; determine that the user is in an ambient environment with multiple users; generate a first tag indicative of a multiple user audience; process the first voice data of the first voice input using a first signal processing algorithm; determine that a physical status of the user is abnormal; select a sore throat physical status for the user; generate a second tag indicative of the sore throat physical status; apply a second signal processing algorithm to the first voice data; determine that an emotional status of the user indicates the user is excited; select an excited emotional status for the user; generate a third tag indicative of the excited emotional status; send a content request comprising the first voice data, the first tag, the second tag, and the third tag to a server, wherein the server determines first audio content for presentation at the speaker device; receive an indication of the first audio content; and present the first audio content, wherein targeting criteria for the first audio content comprises the sore throat physical status, the excited emotional status, and the multiple user audience. - View Dependent Claims (2, 3)
-
-
4. A method comprising:
-
determining, by one or more computer processors coupled to at least one memory, first voice data, wherein the first voice data is generated by a user; determining a first real-time user status of the user using the first voice data; generating a first data tag indicative of the first real-time user status; determining candidate audio content for presentation using the first data tag, the candidate audio content comprising first audio content and second audio content; determining that a first score for the first audio content is greater than a second score for the second audio content, wherein the first score is determined using a first targeting criteria, and the second score is determined using a second targeting criteria; determining the first audio content for presentation at a speaker device; and causing presentation of the first audio content via a speaker of the speaker device. - View Dependent Claims (5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16)
-
-
17. A device comprising:
-
at least one memory that stores computer-executable instructions; and at least one processor configured to access the at least one memory and execute the computer-executable instructions to; determine first voice data, wherein the first voice data is generated by a user; determine a first real-time user status of the user using the first voice data; generate a first data tag indicative of the first real-time user status; determine candidate audio content for presentation using the first data tag, the candidate audio content comprising first audio content and second audio content; determine that a first score for the first audio content is greater than a second score for the second audio content, wherein the first score is determined using a first targeting criteria, and the second score is determined using a second targeting criteria; determine the first audio content for presentation at a speaker device; and present the first audio content via a speaker. - View Dependent Claims (18, 19)
-
Specification