Methods and apparatus for early sensory integration and robust acquisition of real world knowledge
First Claim
1. A method for forming at least one neural network representation of a robot with respect to an environment, the method comprising:
- (A) receiving, by a processor, a plurality of sets of data readings from at least one sensor, the plurality of sets of data readings representing independent measurements by the at least one sensor of the robot with respect to the environment;
(B) mapping the plurality of sets of data readings received in (A) to respective first sets of data cells, each first set of data cells forming a corresponding first neural representation of the robot with respect to the environment;
(C) applying respective weights to values in the respective first sets of data cells mapped in (B), the respective weights representing respective precisions of the independent measurements represented by the corresponding sets of data readings;
(D) combining the values in the first sets of data cells weighted in (C) to form a plurality of aggregated data readings; and
(E) mapping the plurality of aggregated data readings formed in (D) to a second set of data cells stored in the memory, the second set of data cells forming a second neural representation of the robot with respect to the environment that is more precise than the first neural representations of the robot with respect to the environment.
4 Assignments
0 Petitions
Accused Products
Abstract
The systems and methods disclosed herein include a path integration system that calculates optic flow, infers angular velocity from the flow field, and incorporates this velocity estimate into heading calculations. The resulting system fuses heading estimates from accelerometers, gyroscopes, engine torques, and optic flow to determine self-localization. The system also includes a motivational system that implements a reward drive, both positive and negative, into the system. In some implementations, the drives can include: a) a curiosity drive that encourages exploration of new areas, b) a resource drive that attracts the agent towards the recharging base when the battery is low, and c) a mineral reward drive that attracts the agent towards previously explored scientific targets.
68 Citations
31 Claims
-
1. A method for forming at least one neural network representation of a robot with respect to an environment, the method comprising:
-
(A) receiving, by a processor, a plurality of sets of data readings from at least one sensor, the plurality of sets of data readings representing independent measurements by the at least one sensor of the robot with respect to the environment; (B) mapping the plurality of sets of data readings received in (A) to respective first sets of data cells, each first set of data cells forming a corresponding first neural representation of the robot with respect to the environment; (C) applying respective weights to values in the respective first sets of data cells mapped in (B), the respective weights representing respective precisions of the independent measurements represented by the corresponding sets of data readings; (D) combining the values in the first sets of data cells weighted in (C) to form a plurality of aggregated data readings; and (E) mapping the plurality of aggregated data readings formed in (D) to a second set of data cells stored in the memory, the second set of data cells forming a second neural representation of the robot with respect to the environment that is more precise than the first neural representations of the robot with respect to the environment. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13)
-
-
14. A system for estimating at least one position of a robot moving through an environment, the system comprising:
-
an interface to receive first sensor data from a first sensor and second sensor data from a second sensor, the first sensor data and the second sensor data representing measurements of the environment with respect to the robot; a memory, operably coupled to the interface, to store the first sensor data in a first neural representation of possible readings from the first sensor and to store the second sensor data in a second neural representation of possible readings from the second sensor; and a path integration module, operably coupled to the memory, to; (i) apply a first weight to the first sensor data in the first neural representation, the first weight representing a first measurement precision of the first sensor; (ii) apply a second weight to the second sensor data in the second neural representation, the second weight representing a first measurement precision of the first sensor; (iii) combine the first sensor data in the first neural representation with the second sensor data in the second neural representation so as to produce aggregated sensor data in an aggregated neural representation of the environment; and (iv) estimate a position of the robot with respect to the environment based at least in part on the aggregated neural representation of the environment. - View Dependent Claims (15, 16, 17, 18, 19, 20, 21, 22, 23)
-
-
24. A method of providing an estimate of at least one of a position of a robot and an orientation of the robot, the method comprising:
-
(A) receiving a vestibular signal representative of the at least one of the orientation of the robot and the position of the robot; (B) receiving a motor outflow signal representative of commanded motion of the robot; (C) receiving an optical flow signal representative of the at least of the orientation of the robot and the position of the robot with respect to an environment of the robot; (D) weighting the vestibular signal received in (A), the motor outflow signal received in (B), and the optical flow signal received in (C) based on a type of the motion of the robot; (E) combining the vestibular signal, motor outflow signal, and optical flow signal weighted in (D) to produce a localization signal; and (F) generating, via at least one processor, the estimate of the at least one of the position of the robot and the orientation of the robot based on the localization signal produced in (E). - View Dependent Claims (25, 26, 27, 28, 29, 30, 31)
-
Specification