Multi-modal neural interfacing for prosthetic devices
First Claim
Patent Images
1. A neural interface system comprising:
- a prosthetic device;
a plurality of types of sensors configured to receive a plurality of types of physiological activity signals from a prosthetic device user;
a neural interface configured to decode a user movement intent from each of the plurality of signals types, and fuse the movement intents into a joint decision to control moveable elements of the prosthetic device, the neural interface comprising;
a plurality of classifier modules, each classifier module associated with a corresponding one of the signal types and configured to determine a user movement state from signals of the signal type;
a plurality of decoders, each associated with a corresponding one of the signal types and configured to decode a movement intent from signals of the signal type and from one or more of the user movement states; and
a fusion module to fuse movement intents from a plurality of the decoders into the joint movement decision;
wherein the plurality of signal types include a combination of two or more of,a local field potential (LFP) signal,a unit activity (spike) signal,an epidural electrocorticography grid (ECoG) signal,an electromyography (EMG) signal,an electroencephalography (EEG) signal, andan electronystagmography (ENG) signal.
1 Assignment
0 Petitions
Accused Products
Abstract
Methods and systems to interface between physiological devices and a prosthetic device, including to receive a plurality of types of physiological activity signals from a user, decode a user movement intent from each of the plurality of signals types, and fuse the movement intents into a joint decision to control moveable elements of the prosthetic device.
-
Citations
12 Claims
-
1. A neural interface system comprising:
-
a prosthetic device; a plurality of types of sensors configured to receive a plurality of types of physiological activity signals from a prosthetic device user; a neural interface configured to decode a user movement intent from each of the plurality of signals types, and fuse the movement intents into a joint decision to control moveable elements of the prosthetic device, the neural interface comprising; a plurality of classifier modules, each classifier module associated with a corresponding one of the signal types and configured to determine a user movement state from signals of the signal type; a plurality of decoders, each associated with a corresponding one of the signal types and configured to decode a movement intent from signals of the signal type and from one or more of the user movement states; and a fusion module to fuse movement intents from a plurality of the decoders into the joint movement decision; wherein the plurality of signal types include a combination of two or more of, a local field potential (LFP) signal, a unit activity (spike) signal, an epidural electrocorticography grid (ECoG) signal, an electromyography (EMG) signal, an electroencephalography (EEG) signal, and an electronystagmography (ENG) signal. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12)
-
Specification