Online asynchronous reinforcement learning from concurrent customer histories
First Claim
1. An apparatus, comprising;
- one or more computing devices, each of the computing devices having one or more processors and memories configured to perform a method of asynchronous reinforcement learning (RL), including;
obtaining an indication of a Decision Request;
receiving, obtaining, accessing or constructing a user state pertaining to at least one user; and
in response to the Decision Request;
scoring a plurality of actions according to one or more value functions based, at least in part, upon the user state;
applying a policy to identify one of the scored actions as a decision; and
providing an indication of the decision or applying the decision to the at least one user;
obtaining an indication of an Update Request, the Update Request being activated independent of user activity;
receiving, obtaining, accessing or constructing a further user state pertaining to the at least one user; and
in response to the Update Request;
updating at least one of;
the one or functions and the policy based, at least in part, upon the further user state,wherein the Decision Request is activated in response to an event timer and the event timer operates to periodically generate Decision Requests, wherein a frequency with which the event timer generates the Decision Requests is based at least in part, upon a period of time from a last user event pertaining to the at least one user or from a last user action, the last user action including the providing of the indication of the decision or the applying of the decision to the at least one user.
3 Assignments
0 Petitions
Accused Products
Abstract
In one embodiment, an indication of a Decision Request or an Update Request may be received, where the Update Request is activated independent of user activity. A user state pertaining to at least one user may be received, obtained, accessed or constructed. For the Decision Request, one or more actions may be scored according to one or more value functions associated with a computing device, a policy associated with the computing device may be applied to identify one of the scored actions as a decision, and an indication of the decision may be provided or applied. For the Update Request, the one or more value functions and/or the policy may be updated. An indication of updates to the one or more value functions and/or an indication of updates to the policy may be provided.
41 Citations
20 Claims
-
1. An apparatus, comprising;
-
one or more computing devices, each of the computing devices having one or more processors and memories configured to perform a method of asynchronous reinforcement learning (RL), including; obtaining an indication of a Decision Request; receiving, obtaining, accessing or constructing a user state pertaining to at least one user; and in response to the Decision Request; scoring a plurality of actions according to one or more value functions based, at least in part, upon the user state; applying a policy to identify one of the scored actions as a decision; and providing an indication of the decision or applying the decision to the at least one user; obtaining an indication of an Update Request, the Update Request being activated independent of user activity; receiving, obtaining, accessing or constructing a further user state pertaining to the at least one user; and in response to the Update Request; updating at least one of;
the one or functions and the policy based, at least in part, upon the further user state,wherein the Decision Request is activated in response to an event timer and the event timer operates to periodically generate Decision Requests, wherein a frequency with which the event timer generates the Decision Requests is based at least in part, upon a period of time from a last user event pertaining to the at least one user or from a last user action, the last user action including the providing of the indication of the decision or the applying of the decision to the at least one user. - View Dependent Claims (2, 3, 4, 5, 6)
-
-
7. A computer-implemented method of performing asynchronous reinforcement learning (RL), comprising:
-
obtaining an indication of a Decision Request pertaining to at least one user; obtaining an indication of an Update Request pertaining to the at least one user; receiving, obtaining, accessing or constructing a user state pertaining to at least one user, the Update Request being activated independent of activity of the at least one user; in response to the indication of the Update Request, updating at least one of;
one or more value functions and a policy based, at least in part, upon the user state; andperforming an action with respect to the at least one user in response to the Decision Request, wherein updating the one or more value functions includes incorporating non-response data into the one or more value functions, wherein a response to the action taken with respect to the at least one user has not been received or detected. - View Dependent Claims (8, 9, 10, 11, 14, 15, 19, 20)
-
-
12. A computer-implemented method of performing asynchronous reinforcement learning (RL) comprising:
-
obtaining an indication of a Decision Request pertaining to at least one user; obtaining an indication of an Update Request pertaining to the at least one user; receiving, obtaining, accessing or constructing a user state pertaining to at least one user, the Update Request being activated independent of activity of the at least one user; in response to the indication of the Update Request, updating at least one of;
one or more value functions and a policy based, at least in part, upon the user state; andrecording or determining a tune since an action was taken with respect to the at least one user or a time since a last user event pertaining to the at least one user; wherein time(s) at which the updating is performed is determined based, at least in part, upon the time since the action was taken with respect to the at least one user and/or the time since the last user event pertaining to the at least one user.
-
-
13. A computer-implemented method of performing asynchronous reinforcement learning (RL) comprising:
-
obtaining an indication of a Decision Request pertaining to at least one user; obtaining an indication of an Update Request pertaining to the at least one user; receiving, obtaining, accessing or constructing a user state pertaining to at least one user, the Update Request being activated independent of activity of the at least one user; in response to the indication of the Update Request, updating at least one of;
one or more value functions and a policy based, at least in part, upon the user state; andperforming an action with respect to the at least one user in response to the Decision Request; and determining a time since the action was performed with respect to the at least one user; wherein updating includes updating the one or more value functions based, at least in part, upon the time since the action was taken with respect to the at least one user.
-
-
16. A computer implemented method of performing asynchronous reinforcement learning (RL) comprising;
-
obtaining an indication of a Decision Request pertaining to at least one user; obtaining an indication of an Update Request pertaining to the at least one user; receiving, obtaining, accessing or constructing a user state pertaining to at least one user, the Update Request being activated independent of activity of the at least one user; and in response to the indication of the Update Request, updating at least one of;
one or more value functions and a policy based, at least in part, upon the user state,wherein an event timer operates to periodically generate an Update Request, wherein a frequency with which the event timer generates an Update Request is based, at least in part, upon a period of time from a last user event pertaining to the at least one user or a last action performed with respect to the at least one user. - View Dependent Claims (17, 18)
-
Specification