Output of a neural network method for deep odometry assisted by static scene optical flow
First Claim
1. A method of visual odometry for a non-transitory computer readable storage medium storing one or more programs, the one or more programs comprising instructions, which when executed by a computing device, causes the computing device to perform the following steps comprising:
- performing data alignment among sensors including a light detection and ranging (LiDAR) sensor, cameras, and an IMU-GPS module;
collecting image data and generating point clouds;
processing a pair of consecutive images in the image data to recognize pixels corresponding to a same point in the point clouds;
establishing an optical flow for visual odometry;
receiving a first image of a first pair of image frames, and extracting representative features from the first image of the first pair in a first convolution neural network (CNN);
receiving a second image of the first pair, and extracting representative features from the second image of the first pair in the first CNN;
merging, in a first merge module, outputs from the first CNN;
decreasing feature map size in a second CNN;
generating a first flow output for each layer in a first deconvolution neural network (DNN); and
merging, in a second merge module, outputs from the second CNN and the first DNN to generate a first motion estimate.
2 Assignments
0 Petitions
Accused Products
Abstract
A method of visual odometry for a non-transitory computer readable storage medium storing one or more programs is disclosed. The one or more programs includes instructions, which when executed by a computing device, causes the computing device to perform the following steps comprising: performing data alignment among sensors including a LiDAR, cameras and an IMU-GPS module; collecting image data and generating point clouds; processing, in the IMU-GPS module, a pair of consecutive images in the image data to recognize pixels corresponding to a same point in the point clouds; and establishing an optical flow for visual odometry.
190 Citations
16 Claims
-
1. A method of visual odometry for a non-transitory computer readable storage medium storing one or more programs, the one or more programs comprising instructions, which when executed by a computing device, causes the computing device to perform the following steps comprising:
-
performing data alignment among sensors including a light detection and ranging (LiDAR) sensor, cameras, and an IMU-GPS module; collecting image data and generating point clouds; processing a pair of consecutive images in the image data to recognize pixels corresponding to a same point in the point clouds; establishing an optical flow for visual odometry; receiving a first image of a first pair of image frames, and extracting representative features from the first image of the first pair in a first convolution neural network (CNN); receiving a second image of the first pair, and extracting representative features from the second image of the first pair in the first CNN; merging, in a first merge module, outputs from the first CNN; decreasing feature map size in a second CNN; generating a first flow output for each layer in a first deconvolution neural network (DNN); and merging, in a second merge module, outputs from the second CNN and the first DNN to generate a first motion estimate. - View Dependent Claims (2, 3, 4)
-
-
5. A method of visual odometry for a non-transitory computer readable storage medium storing one or more programs, the one or more programs comprising instructions, which when executed by a computing device, causes the computing device to perform the following steps comprising:
-
performing data alignment among sensors including a light detection and ranging (LiDAR) sensor, cameras and an IMU-GPS module; collecting image data and generating point clouds; processing a pair of consecutive images in the image data to recognize pixels corresponding to a same point in the point clouds; establishing an optical flow for visual odometry; receiving a first image of a second pair of image frames, and extracting representative features from the first image of the second pair in a first convolutional neural network (CNN); receiving a second image of the second pair, and extracting representative features from the second image of the second pair in the first CNN; merging, in a first merge module, outputs from the first CNN; decreasing feature map size in a second CNN; and generating a first flow output for each layer in a first deconvolutional neural network (DNN); and merging, in the second merge module, outputs from the second CNN and the first DNN to generate a second motion estimate. - View Dependent Claims (6, 7, 8)
-
-
9. A system for visual odometry, the system comprising:
-
an interne server, comprising; an I/O port, configured to transmit and receive electrical signals to and from a client device; a memory; one or more processing units; and one or more programs stored in the memory and configured for execution by the one or more processing units, the one or more programs including instructions for; performing data alignment among sensors including a light detection and ranging (LiDAR) sensor, cameras and an IMU-GPS module; collecting image data and generating point clouds; processing, in the IMU-GPS module, a pair of consecutive images in the image data to recognize pixels corresponding to a same point in the point clouds; establishing an optical flow for visual odometry; receiving a first image of a first pair of image frames, and extracting representative features from the first image of the first pair in a first convolution neural network (CNN); receiving a second image of the first pair and extracting representative features from the second image of the first pair in the first CNN; merging, in a first merge module, outputs from the first CNN; decreasing a feature map size in a second CNN; generating a first flow output for each layer in a first deconvolution neural network (DNN); and merging, in a second merge module, outputs from the second CNN and the first DNN to generate a first motion estimate. - View Dependent Claims (10, 11, 12)
-
-
13. A system for visual odometry, the system comprising:
-
an interne server, comprising; an I/O port, configured to transmit and receive electrical signals to and from a client device; a memory; one or more processing units; and one or more programs stored in the memory and configured for execution by the one or more processing units, the one or more programs including instructions for; performing data alignment among sensors including a light detection and ranging (LiDAR) sensor, cameras and an IMU-GPS module; collecting image data and generating point clouds; processing, in the IMU-GPS module, a pair of consecutive images in the image data to recognize pixels corresponding to a same point in the point clouds; establishing an optical flow for visual odometry; receiving a first image of a second pair of image frames, and extracting representative features from the first image of the second pair in a first convolution neural network (CNN); and receiving a second image of the second pair and extracting representative features from the second image of the second pair in the first CNN; merging, in a first merge module, outputs from the first CNN; decreasing feature map size in a second CNN; generating a first flow output for each layer in a first deconvolutional neural network (DNN); and merging, in the second merge module, outputs from the second CNN and the first DNN to generate a second motion estimate. - View Dependent Claims (14, 15, 16)
-
Specification