Systems and methods for compressing geotagged video
First Claim
1. A method of encoding a video sequence using a geotagged video database, comprising:
- receiving a captured video sequence including at least one video segment at an encoding server, where at least one geotag indicating geographic capture location information is associated with the captured video sequence;
selecting a segment from the at least one video segment in the captured video sequence using the encoding server;
determining geographic capture location information associated with the selected segment from the at least one geotag associated with the captured video sequence using the encoding server;
identifying a set of relevant video segments from a geotagged video database using the encoding server based on the geographic capture location information associated with the selected video segment, location information from geotags associated with video segments stored in the geotagged video database, and a velocity of a video capture device during capture of the selected segment;
comparing content of the selected segment to content of each video segment in the set of segments using the encoding server to determine a similarity of content between the selected segment and each video segment in the set of relevant video segments by;
performing feature matching with respect to at least one frame in the selected video segment and at least one frame from a video segment within the set of relevant video segments; and
comparing the photometric similarity of the at least one frame in the selected video segment and the at least one frame from the video segment within the set of relevant video segments;
determining a most relevant video segment from the set of relevant video segments that has content that is most similar to the content of the selected segment from the captured video sequence using the encoding server based upon the similarity of content in each video segment within the set of relevant video segments to content of the selected segment from the captured video sequence using the encoding server;
encoding the selected segment from the captured video sequence using the encoding server, where the selected segment is encoded using predictions that include references to the most relevant video segment from the geotagged video database, wherein the most relevant video segment from the geotagged video database was captured at a different time from the captured video sequence; and
storing the encoded video segment in the geotagged video database using the encoding server.
5 Assignments
0 Petitions
Accused Products
Abstract
Systems and methods for compressing and sharing geotagged video in accordance with embodiments of the invention are disclosed. One embodiment includes receiving a captured video sequence, where at least one geographic location is associated with the captured video sequence, selecting a segment of the captured video sequence, identifying a set of relevant video segments from a geotagged video database based on the at least one geotag associated with the captured video sequence, determining the video segment from the set of relevant video segments that is the best match by comparing the similarity of the content in the video segments to the content of the selected segment from the captured video sequence, encoding the selected segment, where the selected segment is encoded using predictions that include references to the video segment that is the best match, and storing the encoded video segment in the geotagged video database.
288 Citations
23 Claims
-
1. A method of encoding a video sequence using a geotagged video database, comprising:
-
receiving a captured video sequence including at least one video segment at an encoding server, where at least one geotag indicating geographic capture location information is associated with the captured video sequence; selecting a segment from the at least one video segment in the captured video sequence using the encoding server; determining geographic capture location information associated with the selected segment from the at least one geotag associated with the captured video sequence using the encoding server; identifying a set of relevant video segments from a geotagged video database using the encoding server based on the geographic capture location information associated with the selected video segment, location information from geotags associated with video segments stored in the geotagged video database, and a velocity of a video capture device during capture of the selected segment; comparing content of the selected segment to content of each video segment in the set of segments using the encoding server to determine a similarity of content between the selected segment and each video segment in the set of relevant video segments by; performing feature matching with respect to at least one frame in the selected video segment and at least one frame from a video segment within the set of relevant video segments; and comparing the photometric similarity of the at least one frame in the selected video segment and the at least one frame from the video segment within the set of relevant video segments; determining a most relevant video segment from the set of relevant video segments that has content that is most similar to the content of the selected segment from the captured video sequence using the encoding server based upon the similarity of content in each video segment within the set of relevant video segments to content of the selected segment from the captured video sequence using the encoding server; encoding the selected segment from the captured video sequence using the encoding server, where the selected segment is encoded using predictions that include references to the most relevant video segment from the geotagged video database, wherein the most relevant video segment from the geotagged video database was captured at a different time from the captured video sequence; and storing the encoded video segment in the geotagged video database using the encoding server. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21)
-
-
22. A video sharing server system, comprising:
-
an encoding server; and a geotagged video database including a plurality of video sequences tagged with geotags indicating geographic locations; wherein the encoding server is configured to; receive a captured video sequence including at least one video segment and associated with the captured video sequence at least one geotag, where the at least one geotag indicates geographic capture location information indicating at least one geographic location; select a segment from the at least one video segment in the captured video sequence; determine geographic capture location information associated with the selected segment from the at least one geotag associated with the captured video sequence; identify a set of relevant video segments from a geotagged video database using the geographic capture location information associated with the selected video segment, location information from geotags associated with video segments stored in the geotagged video database, and a velocity of a video capture device during capture of the selected segment; compare content of the selected segment to content of each video segment in the set of segments to determine a similarity of content between the selected segment and each video segment in the set of relevant video segments by; performing feature matching with respect to at least one frame in the selected video segment and at least one frame from a video segment within the set of relevant video segments; and comparing the photometric similarity of the at least one frame in the selected video segment and the at least one frame from the video segment within the set of relevant video segments; determine a most relevant video segment from the set of relevant video segments that has content that is most similar to the content of the selected segment from the captured video sequence based on the similarity of content between the selected segment and each video segment in the set of relevant video segments; encode the selected segment from the captured video sequence using predictions that include references to the most relevant video segment from the geotagged video database, wherein the most relevant video segment from the geotagged video database was captured at a different time from the captured video sequence; and store the encoded video segment in the geotagged video database.
-
-
23. A non-transitory machine readable medium containing processor instructions, where execution of the instructions by a processor causes the processor to perform a process that comprises:
-
receiving a captured video sequence including at least one video segment and is associated with the captured video sequence at least one geotag, where the at least one geotag indicates geographic capture location information indicating at least one geographic location; selecting a segment from the at least one video segment in the captured video sequence using the encoding server; obtaining a set of relevant video segments from a geotagged video database using the geographic capture location information associated with the selected video segment, location information from geotags associated with video segments stored in the geotagged video database, and a velocity of a video capture device during capture of the selected segment; comparing content of the selected segment to content of each video segment in the set of segments to determine a similarity of content between the selected segment and each video segment in the set of relevant video segments by; performing feature matching with respect to at least one frame in the selected video segment and at least one frame from a video segment within the set of relevant video segments; and comparing the photometric similarity of the at least one frame in the selected video segment and the at least one frame from the video segment within the set of relevant video segments; determining a most relevant video segment from the set of relevant video segments that has content that is most similar to the content of the captured video sequence based on the similarity of content between the selected segment and each video segment in the set of relevant video segments; and encoding the selected segment from the captured video sequence using predictions that include references to the video segment from the geotagged video database that is the best match, wherein the most relevant video segment from the geotagged video database was captured at a different time from the captured video sequence.
-
Specification