Optical flow estimation using a neural network and egomotion optimization
First Claim
1. A method comprising the steps of:
- providing a processor having a neural network, localization filter, and robotic model implemented therein, wherein the processor is operatively connected to a camera connected to a platform, wherein one or more of the camera and the platform is moving;
using the camera to capture a current image of a scene at a current time, wherein the scene includes one or more moving objects therein;
transmitting the current image and one or more previous images of the scene to the neural network, wherein the one or more previous images of the scene are taken at respective one or more previous times prior to the current time;
receiving, at the neural network, current position information and previous position information from the robot model, wherein the current position information is the position of the camera at the current time and the previous position information is the one or more positions of the camera at the respective one or more previous times; and
generating, using the neural network, an estimated optical flow image using the current image of the scene, the current position information, one or more previous images of the scene, and the previous position information;
wherein, prior to the providing step, training the neural network using a ground truth optical flow image, one or more training images of a scene having one or more moving objects therein, and position information corresponding to positions of the camera when the one or more training images are taken; and
wherein the ground truth optical flow image includes a flow of salient objects but does not include a flow of other pixels even when the other pixels visually flow due to movement of the camera.
1 Assignment
0 Petitions
Accused Products
Abstract
A camera is connected to a trained neural network. The camera takes an image of a scene and transmits the image to the neural network. A processor connected to the neural network has a localization filter and a robot model implemented therein. A global positioning system (GPS) receiver and inertial measurement unit (IMU) transmit GPS information and IMU information, respectively, to the processor. The localization filter filters the received GPS and IMU information and inputs the filtered information into the robot model. The robot model outputs current position information corresponding to the current image and previous position information corresponding to the respective one or more previous images. The neural network uses the current image and associated current position information and the one or more previous images and respective associated previous position information to generate an estimated optical flow image, which is transmitted to an object detection system.
9 Citations
18 Claims
-
1. A method comprising the steps of:
-
providing a processor having a neural network, localization filter, and robotic model implemented therein, wherein the processor is operatively connected to a camera connected to a platform, wherein one or more of the camera and the platform is moving; using the camera to capture a current image of a scene at a current time, wherein the scene includes one or more moving objects therein; transmitting the current image and one or more previous images of the scene to the neural network, wherein the one or more previous images of the scene are taken at respective one or more previous times prior to the current time; receiving, at the neural network, current position information and previous position information from the robot model, wherein the current position information is the position of the camera at the current time and the previous position information is the one or more positions of the camera at the respective one or more previous times; and generating, using the neural network, an estimated optical flow image using the current image of the scene, the current position information, one or more previous images of the scene, and the previous position information; wherein, prior to the providing step, training the neural network using a ground truth optical flow image, one or more training images of a scene having one or more moving objects therein, and position information corresponding to positions of the camera when the one or more training images are taken; and wherein the ground truth optical flow image includes a flow of salient objects but does not include a flow of other pixels even when the other pixels visually flow due to movement of the camera. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8)
-
-
9. A system comprising:
-
a neural network; a camera operatively connected to the neural network, the camera configured to take an image of a scene and transmit the image of the scene to the neural network; and a processor operatively connected to the neural network, the processor having a localization filter and a robot model implemented therein, wherein the processor is configured to output from the robot model current position information corresponding to the current image and previous position information corresponding to the respective one or more previous images, wherein the neural network is configured to use the current image and the associated current position information, and the one or more previous images and the respective associated previous position information to generate an estimated optical flow image; wherein training the neural network using a ground truth optical flow image, one or more training images of a scene having one or more moving objects therein, and position information corresponding to positions of the camera when the one or more training images are taken; and wherein the ground truth optical flow image includes a flow of salient objects but does not include a flow of other pixels even when the other pixels visually flow due to movement of the camera. - View Dependent Claims (10, 11, 12, 13, 14, 15, 16, 17)
-
-
18. A system comprising:
-
a neural network; a camera operatively connected to the neural network, the camera configured to take an image of a scene and transmit the image of the scene to the neural network; a processor operatively connected to the neural network, the processor having a localization filter and a robot model implemented therein; a global positioning system (GPS) receiver operatively connected to the processor; and an inertial measurement unit (IMU) operatively connected to the processor, wherein the GPS receiver and the IMU are configured to transmit GPS information and IMU information, respectively, to the processor, wherein the processor is configured to use the localization filter to filter the received GPS information and IMU information and input the filtered information into the robot model, wherein the processor is configured to output from the robot model current position information corresponding to the current image and previous position information corresponding to the respective one or more previous images, wherein the neural network is configured to use the current image and the associated current position information while ignoring a motion field due to ego-motion, and the one or more previous images and the respective associated previous position information to generate an estimated optical flow image; wherein training the neural network using a ground truth optical flow image, one or more training images of a scene having one or more moving objects therein, and position information corresponding to positions of the camera when the one or more training images are taken; and wherein the ground truth optical flow image includes a flow of salient objects but does not include a flow of other pixels even when the other pixels visually flow due to movement of the camera.
-
Specification