Projected image item tracking system
First Claim
1. A projected image item tracking system comprising:
- a processor coupled toa sensor configured to generatean enter signal when a shopper reaches into or towards an item storage area in a store, wherein said item storage area contains items; and
,an exit signal when said shopper retracts from said item storage area; and
,a plurality of cameras oriented to view said item storage area;
wherein said processor is configured toobtain a plurality of before images captured by said plurality of cameras, each before image of said plurality of before images corresponding to a camera of said plurality of cameras, wherein said each before image is captured at a time before said enter signal;
obtain a plurality of after images captured by said plurality of cameras, each after image of said plurality of after images corresponding to said camera of said plurality of cameras, wherein said each after image is captured at a time after said exit signal;
project said plurality of before images onto two or more planes in said item storage area to generate a plurality of projected before images corresponding to each combination of a plane of said two or more planes and said camera of said plurality of cameras;
project said plurality of after images onto said two or more planes to generate a plurality of projected after images corresponding to said each combination of said plane of said two or more planes and said camera of said plurality of cameras;
analyzesaid plurality of projected before images, andsaid plurality of projected after images, toidentify an item of said items taken from or put into said item storage area between said enter signal and said exit signal; and
,associate said item with said shopper; and
,wherein said analyze said plurality of projected before images and said plurality of projected after images comprisescalculate a 3D volume difference betweencontents of said item storage area at said time before said enter signal andcontents of said item storage area at said time after said exit signal;
when said 3D volume difference indicates that said contents of said item storage area at said time after said exit signal is smaller than said contents of said item storage area at said time before said enter signal,input at least a portion of one or more of said plurality of projected before images that intersects said 3D volume difference into a classifier;
when said 3D volume difference indicates that said contents of said item storage area at said time after said exit signal is larger than said contents of said item storage area at said time before said enter signal,input at least a portion of one or more of said plurality of projected after images that intersects said 3D volume difference into said classifier; and
,identify said item of said items taken from or put into said item storage area as an output of said classifier.
1 Assignment
0 Petitions
Accused Products
Abstract
A projected image item tracking system that analyzes projected camera images to determine items taken from, placed on, or moved on a shelf or other area in an autonomous store. The items and actions performed on them may then be attributed to a shopper near the area. Projected images may be combined to generate a 3D volume difference between the state of the area before and after shopper interaction. The volume difference may be calculated using plane-sweep stereo, or using convolutional neural networks. Because these methods may be computationally intensive, the system may first localize a change volume where items appear to have been displaced, and then generate a volume difference only within that change volume. This optimization results in significant savings in power consumption and in more rapid identification of items. The 3D volume difference may also indicate the quantity of items displaced, for example from a vertical stack.
55 Citations
22 Claims
-
1. A projected image item tracking system comprising:
-
a processor coupled to a sensor configured to generate an enter signal when a shopper reaches into or towards an item storage area in a store, wherein said item storage area contains items; and
,an exit signal when said shopper retracts from said item storage area; and
,a plurality of cameras oriented to view said item storage area; wherein said processor is configured to obtain a plurality of before images captured by said plurality of cameras, each before image of said plurality of before images corresponding to a camera of said plurality of cameras, wherein said each before image is captured at a time before said enter signal; obtain a plurality of after images captured by said plurality of cameras, each after image of said plurality of after images corresponding to said camera of said plurality of cameras, wherein said each after image is captured at a time after said exit signal; project said plurality of before images onto two or more planes in said item storage area to generate a plurality of projected before images corresponding to each combination of a plane of said two or more planes and said camera of said plurality of cameras; project said plurality of after images onto said two or more planes to generate a plurality of projected after images corresponding to said each combination of said plane of said two or more planes and said camera of said plurality of cameras; analyze said plurality of projected before images, and said plurality of projected after images, to identify an item of said items taken from or put into said item storage area between said enter signal and said exit signal; and
,associate said item with said shopper; and
,wherein said analyze said plurality of projected before images and said plurality of projected after images comprises calculate a 3D volume difference between contents of said item storage area at said time before said enter signal and contents of said item storage area at said time after said exit signal; when said 3D volume difference indicates that said contents of said item storage area at said time after said exit signal is smaller than said contents of said item storage area at said time before said enter signal, input at least a portion of one or more of said plurality of projected before images that intersects said 3D volume difference into a classifier; when said 3D volume difference indicates that said contents of said item storage area at said time after said exit signal is larger than said contents of said item storage area at said time before said enter signal, input at least a portion of one or more of said plurality of projected after images that intersects said 3D volume difference into said classifier; and
,identify said item of said items taken from or put into said item storage area as an output of said classifier. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21)
-
-
22. A projected image item tracking system comprising:
-
a processor coupled to a sensor configured to generate an enter signal when a shopper reaches into or towards an item storage area in a store, wherein said item storage area contains items; and
,an exit signal when said shopper retracts from said item storage area; and
,a plurality of cameras oriented to view said item storage area; wherein said processor is configured to obtain a plurality of before images captured by said plurality of cameras, each before image of said plurality of before images corresponding to a camera of said plurality of cameras, wherein said each before image is captured at a time before said enter signal; obtain a plurality of after images captured by said plurality of cameras, each after image of said plurality of after images corresponding to said camera of said plurality of cameras, wherein said each after image is captured at a time after said exit signal; project said plurality of before images onto two or more planes in said item storage area to generate a plurality of projected before images corresponding to each combination of a plane of said two or more planes and said camera of said plurality of cameras; project said plurality of after images onto said two or more planes to generate a plurality of projected after images corresponding to said each combination of said plane of said two or more planes and said camera of said plurality of cameras; analyze said plurality of projected before images, and said plurality of projected after images, to identify an item of said items taken from or put into said item storage area between said enter signal and said exit signal; and
,associate said item with said shopper; and
,wherein said analyze said plurality of projected before images and said plurality of projected after images comprises input at least a portion of said plurality of projected before images and at least a portion of said plurality of projected after images into a neural network trained to output said item of said items taken from or put into said item storage area between said enter signal and said exit signal, wherein said neural network is further trained to output an action that indicates whether said item of said items is taken from or is put into said item storage area between said enter signal and said exit signal; and
,wherein said neural network comprises a feature extraction layer applied to each of said at least a portion of said plurality of projected before images and at least a portion of said plurality of projected after images, wherein said feature extraction layer outputs image features; a differencing layer applied to each pair of said image features associated with a portion of a projected before image from said camera and said image features associated with a portion of a projected after image from said camera, wherein said differencing layer outputs feature differences associated with said camera; one or more convolutional layers applied to said feature differences associated with each camera of said plurality of cameras; an item classifier layer applied to an output of said one or more convolutional layers; and
,an action classifier layer applied to said output of said one or more convolutional layers.
-
Specification