Health monitoring system and appliance
First Claim
1. An electronic device configured to process audible expressions from users, comprising:
- a network interface;
at least one computing device; and
computer readable memory including instructions operable to be executed by the at least one computing device to perform a set of actions, configuring the at least one computing device to;
receive in real time, over a network via the network interface, a digitized human vocal expression of a first user and one or more digital images from a remote device;
process, remotely from the remote device, the received digitized human vocal expression using digital signal processing to convert the digitized audible expression from a time domain to a frequency domain;
use the processed digitized human vocal expression to determine characteristics of the human vocal expression, including;
determining, using a volume analysis module a volume of the human vocal expression,determining, using a rapidity analysis module that detects quiet time using a power spectrum of the human vocal expression, how rapidly the first user is speaking in the human vocal expression,determining, using a vocal tract analysis module, a magnitude spectrum of the human vocal expression, andidentifying, using a non-speech analysis module, pauses and the length of pauses in speech in the human vocal expression;
use a natural language module to;
identify phonemes in the human vocal expression and map the phonemes to words, to convert audible speech in the human vocal expression to text,divide the text into text elements including words, sentences, and paragraphs,understand audible speech in the human vocal expression using semantic analysis that assigns respective logical and grammatical roles to the text elements, anddetect violations of grammar rules in the text obtained from the human vocal expression to obtain detected grammar violations;
compare the determined characteristics of the human vocal expression with baseline, historical characteristics of human vocal expressions associated with the first user to identify changes in human vocal expression characteristics of the first user as identified vocal changes;
process the received one or more images to detect characteristics of the first user face, including determining the presence of;
a sagging lip, a crooked smile, uneven eyebrows, or facial droop;
compare the detected characteristics of the first user face with baseline, historical characteristics of the first user face accessed from a data store, and identify changes in characteristics of the first user face as identified facial changes;
weight, using a first weight, a first identified change, of the identified vocal changes, with respect to a first vocal expression characteristic of the first user;
weight, using a second weight, a second identified change, of the identified vocal changes, with respect to a second vocal expression characteristic of the first user;
weight, using a third weight, a third identified change, of the identified facial changes, with respect to a first characteristic of the first user face;
weight, using a fourth weight, a fourth identified change, of the identified facial changes, with respect to a second characteristic of the first user face;
weight, using a fifth weight, the detected grammar violations;
infer a change in health status of the first user using the weighted first identified change with respect to the first vocal expression characteristic of the first user, the weighted second identified change with respect to the second vocal expression characteristic of the first user, the weighted third identified change with respect to the first characteristic of the first user face, the weighted fourth identified change with respect to the second characteristic of the first user face and the weighted detected grammar violations;
based at least in part on the inferred change in health status of the first user determine if a vehicle is to be deployed to the first user; and
at least partly in response to a determination that a vehicle is to be deployed to the first user, enable a vehicle to be deployed to a location of the first user.
1 Assignment
0 Petitions
Accused Products
Abstract
Systems and methods are disclosed. A digitized human vocal expression of a user and digital images are received over a network from a remote device. The digitized human vocal expression is processed to determine characteristics of the human vocal expression, including: pitch, volume, rapidity, a magnitude spectrum identify, and/or pauses in speech. Digital images are received and processed to detect characteristics of the user face, including detecting if one or more of the following is present: a sagging lip, a crooked smile, uneven eyebrows, and/or facial droop. Based at least on part on the human vocal expression characteristics and face characteristics, a determination is made as to what action is to be taken. A cepstrum pitch may be determined using an inverse Fourier transform of a logarithm of a spectrum of a human vocal expression signal. The volume may be determined using peak heights in a power spectrum of the human vocal expression.
174 Citations
26 Claims
-
1. An electronic device configured to process audible expressions from users, comprising:
-
a network interface; at least one computing device; and computer readable memory including instructions operable to be executed by the at least one computing device to perform a set of actions, configuring the at least one computing device to; receive in real time, over a network via the network interface, a digitized human vocal expression of a first user and one or more digital images from a remote device; process, remotely from the remote device, the received digitized human vocal expression using digital signal processing to convert the digitized audible expression from a time domain to a frequency domain; use the processed digitized human vocal expression to determine characteristics of the human vocal expression, including; determining, using a volume analysis module a volume of the human vocal expression, determining, using a rapidity analysis module that detects quiet time using a power spectrum of the human vocal expression, how rapidly the first user is speaking in the human vocal expression, determining, using a vocal tract analysis module, a magnitude spectrum of the human vocal expression, and identifying, using a non-speech analysis module, pauses and the length of pauses in speech in the human vocal expression; use a natural language module to; identify phonemes in the human vocal expression and map the phonemes to words, to convert audible speech in the human vocal expression to text, divide the text into text elements including words, sentences, and paragraphs, understand audible speech in the human vocal expression using semantic analysis that assigns respective logical and grammatical roles to the text elements, and detect violations of grammar rules in the text obtained from the human vocal expression to obtain detected grammar violations; compare the determined characteristics of the human vocal expression with baseline, historical characteristics of human vocal expressions associated with the first user to identify changes in human vocal expression characteristics of the first user as identified vocal changes; process the received one or more images to detect characteristics of the first user face, including determining the presence of;
a sagging lip, a crooked smile, uneven eyebrows, or facial droop;compare the detected characteristics of the first user face with baseline, historical characteristics of the first user face accessed from a data store, and identify changes in characteristics of the first user face as identified facial changes; weight, using a first weight, a first identified change, of the identified vocal changes, with respect to a first vocal expression characteristic of the first user; weight, using a second weight, a second identified change, of the identified vocal changes, with respect to a second vocal expression characteristic of the first user; weight, using a third weight, a third identified change, of the identified facial changes, with respect to a first characteristic of the first user face; weight, using a fourth weight, a fourth identified change, of the identified facial changes, with respect to a second characteristic of the first user face; weight, using a fifth weight, the detected grammar violations; infer a change in health status of the first user using the weighted first identified change with respect to the first vocal expression characteristic of the first user, the weighted second identified change with respect to the second vocal expression characteristic of the first user, the weighted third identified change with respect to the first characteristic of the first user face, the weighted fourth identified change with respect to the second characteristic of the first user face and the weighted detected grammar violations; based at least in part on the inferred change in health status of the first user determine if a vehicle is to be deployed to the first user; and at least partly in response to a determination that a vehicle is to be deployed to the first user, enable a vehicle to be deployed to a location of the first user. - View Dependent Claims (2, 3, 4, 5, 6, 7)
-
-
8. An electronic device, comprising:
-
a network interface; at least one computing device; and computer readable memory including instructions operable to be executed by the at least one computing device to perform a set of actions, configuring the at least one computing device to; receive, over a network via the network interface, a digitized human vocal expression of a first user and one or more digital images of the first user from a first source; process, remotely from the first source, the received digitized human vocal expression using digital signal processing to convert the digitized audible expression from a time domain to a frequency domain; use the processed digitized human vocal expression to determine characteristics of the human vocal expression, including; determining a volume, magnitude, and a power spectrum of the human vocal expression, and detecting quiet time using the power spectrum of the human vocal expression to determine pauses and the length of pauses in speech in the human vocal expression, and to determine how rapidly the first user is speaking in the human vocal expression; use a natural language module to; identify phonemes in the human vocal expression and map the phonemes to words, to convert audible speech in the human vocal expression to text, divide the text into text elements including words, sentences, and/or paragraphs, understand audible speech in the human vocal expression using semantic analysis, and detect violations of grammar rules in the text obtained from the human vocal expression to obtain detected grammar violations; compare the determined characteristics of the human vocal expression with baseline, historical characteristics of human vocal expressions associated with the first user to identify changes in human vocal expression characteristics of the first user as identified vocal changes; process the received one or more images to detect characteristics of the first user face, including determining the presence of;
a sagging lip, a crooked smile, uneven eyebrows, or facial droop;compare the detected characteristics of the first user face with baseline, historical characteristics of the first user face accessed from a data store, and identify changes in characteristics of the first user face as identified facial changes; weight, using a first weight, a first identified change, of the identified vocal changes, with respect to a first vocal expression characteristic of the first user; weight, using a second weight, a second identified change, of the identified vocal changes, with respect to a second vocal expression characteristic of the first user; weight, using a third weight, a third identified change, of the identified facial changes, with respect to a first characteristic of the first user face; weight, using a fourth weight, a fourth identified change, of the identified facial changes, with respect to a second characteristic of the first user face; weight, using a fifth weight, the detected grammar violations; infer a change in health status of the first user based at least in part on the weighted first identified change with respect to the first vocal expression characteristic of the first user, the weighted second identified change with respect to the second vocal expression characteristic of the first user, the weighted third identified change with respect to the first characteristic of the first user face, the weighted fourth identified change with respect to the second characteristic of the first user face, and the weighted detected grammar violations; and based at least in part on the inferred change in health status of the first user, cause a first action is to be taken. - View Dependent Claims (9, 10, 11, 12, 13, 14, 15, 16)
-
-
17. A computer implemented method, comprising:
-
receiving, at a system configured to process digitized human vocal expressions using digital signal processing, a digitized human vocal expression of a first user and one or more digital images of the first user from a first device; processing remotely from the first source, using digital signal processing, the received digitized human vocal expression to convert the digitized audible expression from a time domain to a frequency domain; using, by the system, the processed digitized human vocal expression to determine characteristics of the human vocal expression, including; determining a volume, magnitude, and a power spectrum of the human vocal expression, detecting quiet time using the power spectrum of the human vocal expression to determine pauses and the length of pauses in speech in the human vocal expression, and to determine how rapidly the first user is speaking in the human vocal expression, using natural language processing to; identify phonemes in the human vocal expression and map the phonemes to words, to convert audible speech in the human vocal expression to text, divide the text into text elements including words, sentences, and/or paragraphs, understand audible speech in the human vocal expression using semantic analysis, and detect violations of grammar rules in the text obtained from the human vocal expression to obtain detected grammar violations; and comparing one or more of the determined characteristics of the human vocal expression with one or more baseline, historical characteristics of human vocal expressions associated with the first user as identified vocal changes; processing the received one or more images to detect characteristics of the first user face, including determining the presence of;
a sagging lip, a crooked smile, uneven eyebrows, or facial droop;comparing the detected characteristics of the first user face with baseline, historical characteristics of the first user face accessed from a data store, and identify changes in characteristics of the first user face as identified facial changes; weighting by the system, using a first weight, a first identified change, of the identified vocal changes, with respect to a first vocal expression characteristic of the first user; weighting by the system, using a second weight, a second identified change, of the identified vocal changes, with respect to a second vocal expression characteristic of the first user; weighting by the system, using a third weight, a third identified change, of the identified facial changes, with respect to a first characteristic of the first user face; weighting by the system, using a fourth weight, a fourth identified change, of the identified facial changes, with respect to a second characteristic of the first user face; weighting by the system, using a fifth weight, the detected grammar violations; inferring, by the system, a change in health status of the first user based at least in part on the weighted first identified change with respect to the first vocal expression characteristic of the first user, the weighted second identified change with respect to the second vocal expression characteristic of the first user, the weighted third identified change with respect to the first characteristic of the first user face, the weighted fourth identified change with respect to the second characteristic of the first user face, and weighted detected grammar violations; and based at least in part on the inferred change in health status of the first user, causing a first action is to be taken. - View Dependent Claims (18, 19, 20, 21, 22)
-
-
23. A computer implemented method, comprising:
-
receiving from a first source, at a computerized device configured to process digitized human vocal expressions using digital signal processing, a digitized human vocal expression of a first user and one or more digital images of the first user; processing remote from the first source, using digital signal processing, the received digitized human vocal expression to convert the digitized audible expression from a time domain to a frequency domain; using, by the system, the processed digitized human vocal expression to determine characteristics of the human vocal expression, including; determining a volume, magnitude, and a power spectrum of the human vocal expression, determining how rapidly the first user is speaking in the human vocal expression by detecting quiet time using a power spectrum of the human vocal expression, and determining pauses and the length of pauses in speech in the human vocal expression; using natural language processing to; identify phonemes in the human vocal expression and map the phonemes to words, to convert audible speech in the human vocal expression to text, divide the text into text elements including words, sentences, and/or paragraphs, understand audible speech in the human vocal expression using semantic analysis, and detect violations of grammar rules in the text obtained from the human vocal expression to obtain detected grammar violations; comparing, using the computerized device, one or more of the determined characteristics of the human vocal expression with one or more baseline, historical characteristics of human vocal expressions associated with the first user as identified vocal changes; processing the received one or more images to detect characteristics of the first user face, including determining the presence of;
a sagging lip, a crooked smile, uneven eyebrows, or facial droop;comparing the detected characteristics of the first user face with baseline, historical characteristics of the first user face accessed from a data store, and identify changes in characteristics of the first user face as identified facial changes; weighting by the system, using a first weight, a first identified change, of the identified vocal changes, with respect to a first vocal expression characteristic of the first user; weighting by the system, using a second weight, a second identified change, of the identified vocal changes, with respect to a second vocal expression characteristic of the first user; weighting by the system, using a third weight, a third identified change, of the identified facial changes, with respect to a first characteristic of the first user face; weighting by the system, using a fourth weight, a fourth identified change, of the identified facial changes, with respect to a second characteristic of the first user face; weighting, using a fifth weight, the detected grammar violations; inferring, by the system, a change in health status of the first user based at least in part on the weighted first identified change with respect to the first vocal expression characteristic of the first user, the weighted second identified change with respect to the second vocal expression characteristic of the first user, the weighted third identified change with respect to the first characteristic of the first user face, the weighted fourth identified change with respect to the second characteristic of the first user face, and the weighted detected grammar violations; and based at least in part on the inferred change in health status of the first user, enabling a first action is to be taken. - View Dependent Claims (24, 25, 26)
-
Specification