Visual search in real world using optical see-through head mounted display with augmented reality and user interaction tracking
First Claim
1. A method of conducting an online visual search through an augmented reality (AR) device having a display, said method comprising:
- capturing, via an image capture device of the AR device, a scene in a field of view of the display;
identifying, via at least one processor of the AR device, a portion of the scene based on a first user interaction with the display;
displaying AR content on the display in response to the first user interaction, the AR content comprising indicia associated with the identified portion of the scene;
receiving, after displaying the AR content on the display, an indication to initiate an online visual search of the identified portion of the scene based on a second user interaction with a search icon displayed on the display, the second user interaction occurring after the first user interaction, wherein the second user interaction comprises a non-eye gaze gesture; and
transmitting, by the AR device in response to the second user interaction, an image of the identified portion of the scene to a search engine, wherein the image includes the identified portion of the scene and does not include content in the field of view of the display outside of the identified portion of the scene.
1 Assignment
0 Petitions
Accused Products
Abstract
A method, an apparatus, and a computer program product conduct online visual searches through an augmented reality (AR) device having an optical see-through head mounted display (HMD). An apparatus identifies a portion of an object in a field of view of the HMD based on user interaction with the HMD. The portion includes searchable content, such as a barcode. The user interaction may be an eye gaze or a gesture. A user interaction point in relation to the HMD screen is tracked to locate a region of the object that includes the portion and the portion is detected within the region. The apparatus captures an image of the portion. The identified portion of the object does not encompass the entirety of the object. Accordingly, the size of the image is less than the size of the object in the field of view. The apparatus transmits the image to a visual search engine.
34 Citations
27 Claims
-
1. A method of conducting an online visual search through an augmented reality (AR) device having a display, said method comprising:
-
capturing, via an image capture device of the AR device, a scene in a field of view of the display; identifying, via at least one processor of the AR device, a portion of the scene based on a first user interaction with the display; displaying AR content on the display in response to the first user interaction, the AR content comprising indicia associated with the identified portion of the scene; receiving, after displaying the AR content on the display, an indication to initiate an online visual search of the identified portion of the scene based on a second user interaction with a search icon displayed on the display, the second user interaction occurring after the first user interaction, wherein the second user interaction comprises a non-eye gaze gesture; and transmitting, by the AR device in response to the second user interaction, an image of the identified portion of the scene to a search engine, wherein the image includes the identified portion of the scene and does not include content in the field of view of the display outside of the identified portion of the scene. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8)
-
-
9. An apparatus for conducting an online visual search through an augmented reality (AR) device having a display, said apparatus comprising:
-
means for capturing a scene in a field of view of the display; means for identifying a portion of the scene based on a first user interaction with the display; means for displaying AR content on the display in response to the first user interaction, the AR content comprising indicia associated with the identified portion of the scene; means for receiving, after displaying the AR content on the display, an indication to initiate an online visual search of the identified portion of the scene based on a second user interaction with a search icon displayed on the display, the second user interaction occurring after the first user interaction, wherein the second user interaction comprises a non-eye gaze gesture; and means for transmitting, in response to the second user interaction, an image of the identified portion of the scene to a search engine, wherein the image includes the identified portion of the scene and does not include content in the field of view of the display outside of the identified portion of the scene. - View Dependent Claims (10, 11, 12, 13, 14, 15)
-
-
16. An apparatus for conducting an online visual search through an augmented reality (AR) device having a display, said apparatus comprising:
-
a memory; an image capture device configured to capture a scene in a field of view of the display; a transceiver; and at least one processor coupled to the memory and transceiver, wherein the at least one processor is configured to; identify a portion of the scene based on a first user interaction with the display; display AR content on the display in response to the first user interaction, the AR content comprising indicia associated with the identified portion of the scene; receive, after the AR content is displayed on the display, an indication to initiate an online visual search of the identified portion of the scene based on a second user interaction with a search icon displayed on the display, the second user interaction occurring after the first user interaction, wherein the second user interaction comprises a non-eye gaze gesture; and cause, in response to the second user interaction, the transceiver to transmit an image of the identified portion of the scene to a search engine, wherein the image includes the identified portion of the scene and does not include content in the field of view of the display outside of the identified portion of the scene. - View Dependent Claims (17, 18, 19, 20, 21, 22, 23)
-
-
24. A non-transitory computer-readable medium having instructions stored thereon that, when executed, cause at least one processor of an augmented reality (AR) device having a display to:
-
cause an image capture device of the AR device to capture a scene in a field of view of the display; identify, via the at least one processor of the AR device, a portion of the scene based on a first user interaction with the display; display AR content on the display in response to the first user interaction, the AR content comprising indicia associated with the identified portion of the scene; receive, after the AR content is displayed on the display, an indication to initiate an online visual search of the identified portion of the scene based on a second user interaction with a search icon displayed on the display, the second user interaction occurring after the first user interaction, wherein the second user interaction comprises a non-eye gaze gesture; and cause, in response to the second user interaction, the AR device to transmit an image of the identified portion of the scene to a search engine, wherein the image includes the identified portion of the scene and does not include content in the field of view of the display outside of the identified portion of the scene. - View Dependent Claims (25, 26, 27)
-
Specification