Training encoder model and/or using trained encoder model to determine responsive action(s) for natural language input
First Claim
1. A method implemented by one or more processors, comprising:
- identifying a plurality of positive training instances that each include an input and a response, wherein for each of the positive training instances;
the input is based on content of a corresponding electronic communication, andthe response is based on a corresponding responsive electronic communication that is responsive to the corresponding electronic communication;
training an encoder model based on the positive training instances, wherein training the encoder model based on a given instance of the positive training instances comprises;
generating an input encoding based on processing the input using the encoder model;
generating a response encoding based on processing the response using the encoder model;
generating a final response encoding based on processing the response encoding using a reasoning model;
determining a value based on comparison of the input encoding and the final response encoding; and
updating both the reasoning model and the encoder model based on comparison of the value to a given value indicated by the given instance; and
after training the encoder model;
using the trained encoder model, independent of the reasoning model, to determine a similarity value of two textual segments, wherein the similarity value indicates semantic similarity of the two textual segments, and wherein using the trained encoder model to determine the similarity value of the two textual segments comprises;
receiving a query directed to an automated assistant;
generating a query encoding based on processing the query using the trained encoder model;
comparing the query encoding to a plurality of pre-determined query encodings each stored in association with one or more corresponding actions;
determining, based on the comparing, a given predetermined query encoding to which the query encoding is most similar; and
in response to the query and based on the given predetermined query encoding being most similar to the query encoding, causing the automated assistant to perform the one or more corresponding actions that are stored in association with the given predetermined query encoding.
1 Assignment
0 Petitions
Accused Products
Abstract
Systems, methods, and computer readable media related to: training an encoder model that can be utilized to determine semantic similarity of a natural language textual string to each of one or more additional natural language textual strings (directly and/or indirectly); and/or using a trained encoder model to determine one or more responsive actions to perform in response to a natural language query. The encoder model is a machine learning model, such as a neural network model. In some implementations of training the encoder model, the encoder model is trained as part of a larger network architecture trained based on one or more tasks that are distinct from a “semantic textual similarity” task for which the encoder model can be used.
11 Citations
15 Claims
-
1. A method implemented by one or more processors, comprising:
-
identifying a plurality of positive training instances that each include an input and a response, wherein for each of the positive training instances; the input is based on content of a corresponding electronic communication, and the response is based on a corresponding responsive electronic communication that is responsive to the corresponding electronic communication; training an encoder model based on the positive training instances, wherein training the encoder model based on a given instance of the positive training instances comprises; generating an input encoding based on processing the input using the encoder model; generating a response encoding based on processing the response using the encoder model; generating a final response encoding based on processing the response encoding using a reasoning model; determining a value based on comparison of the input encoding and the final response encoding; and updating both the reasoning model and the encoder model based on comparison of the value to a given value indicated by the given instance; and after training the encoder model; using the trained encoder model, independent of the reasoning model, to determine a similarity value of two textual segments, wherein the similarity value indicates semantic similarity of the two textual segments, and wherein using the trained encoder model to determine the similarity value of the two textual segments comprises; receiving a query directed to an automated assistant; generating a query encoding based on processing the query using the trained encoder model; comparing the query encoding to a plurality of pre-determined query encodings each stored in association with one or more corresponding actions; determining, based on the comparing, a given predetermined query encoding to which the query encoding is most similar; and in response to the query and based on the given predetermined query encoding being most similar to the query encoding, causing the automated assistant to perform the one or more corresponding actions that are stored in association with the given predetermined query encoding. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11)
-
-
12. A method implemented by one or more processors, comprising:
-
identifying a plurality of positive training instances that each include an input and a response, wherein for each of the positive training instances; the input is based on content of a corresponding electronic communication, and the response is based on a corresponding responsive electronic communication that is responsive to the corresponding electronic communication; training an encoder model based on the positive training instances, wherein training the encoder model based on a given instance of the positive training instances comprises; generating an input encoding based on processing the input using the encoder model; generating a response encoding based on processing the response using the encoder model; generating a final response encoding based on processing the response encoding using a reasoning model; determining a value based on comparison of the input encoding and the final response encoding; and updating both the reasoning model and the encoder model based on comparison of the value to a given value indicated by the given instance; and training the encoder model based on a plurality of distinct additional training instances, wherein the plurality of distinct additional training instances are for a task that is distinct from the task of the plurality of positive training instances, and wherein training the encoder model based on a given distinct instance of the distinct additional training instances comprises; generating a first encoding based on processing a first input of the given distinct instance using the encoder model; generating a second encoding based on processing a second input of the given distinct instance using the encoder model; generating a prediction based on processing of the first encoding and the second encoding using an additional model, wherein the additional model is not utilized in training the encoder model based on the positive training instances; and updating both the additional model and the encoder model based on comparison of the prediction to a labeled output of the given distinct instance; after training the encoder model; using the trained encoder model, independent of the reasoning model, to determine a similarity value of two textual segments, wherein the similarity value indicates semantic similarity of the two textual segments. - View Dependent Claims (13, 14, 15)
-
Specification