Data transformation caching in an artificial intelligence infrastructure
First Claim
1. A method of data transformation caching in an artificial intelligence infrastructure that includes one or more storage systems and one or more graphical processing unit (‘
- GPU’
) servers, the method comprising;
identifying, in dependence upon one or more machine learning models to be executed on the GPU servers, one or more transformations to apply to a dataset;
generating, in dependence upon the one or more transformations, a transformed dataset;
storing, within one or more of the storage systems, the transformed dataset;
receiving a plurality of requests to transmit the transformed dataset to one or more of the GPU servers; and
responsive to each request, transmitting, from the one or more storage systems to the one or more GPU servers without re-performing the one or more transformations on the dataset, the transformed dataset.
2 Assignments
0 Petitions
Accused Products
Abstract
Data transformation caching in an artificial intelligence infrastructure that includes one or more storage systems and one or more graphical processing unit (‘GPU’) servers, including: identifying, in dependence upon one or more machine learning models to be executed on the GPU servers, one or more transformations to apply to a dataset; generating, in dependence upon the one or more transformations, a transformed dataset; storing, within one or more of the storage systems, the transformed dataset; receiving a plurality of requests to transmit the transformed dataset to one or more of the GPU servers; and responsive to each request, transmitting, from the one or more storage systems to the one or more GPU servers without re-performing the one or more transformations on the dataset, the transformed dataset.
213 Citations
20 Claims
-
1. A method of data transformation caching in an artificial intelligence infrastructure that includes one or more storage systems and one or more graphical processing unit (‘
- GPU’
) servers, the method comprising;identifying, in dependence upon one or more machine learning models to be executed on the GPU servers, one or more transformations to apply to a dataset; generating, in dependence upon the one or more transformations, a transformed dataset; storing, within one or more of the storage systems, the transformed dataset; receiving a plurality of requests to transmit the transformed dataset to one or more of the GPU servers; and responsive to each request, transmitting, from the one or more storage systems to the one or more GPU servers without re-performing the one or more transformations on the dataset, the transformed dataset. - View Dependent Claims (2, 3, 4, 5, 6, 7)
- GPU’
-
8. An artificial intelligence infrastructure that includes one or more storage systems and one or more graphical processing unit (‘
- GPU’
) servers, the artificial intelligence infrastructure configured to carry out the steps of;identifying, in dependence upon one or more machine learning models to be executed on the GPU servers, one or more transformations to apply to a dataset; generating, in dependence upon the one or more transformations, a transformed dataset; storing, within one or more of the storage systems, the transformed dataset; receiving a plurality of requests to transmit the transformed dataset to one or more of the GPU servers; and responsive to each request, transmitting, from the one or more storage systems to the one or more GPU servers without re-performing the one or more transformations on the dataset, the transformed dataset. - View Dependent Claims (9, 10, 11, 12, 13, 14)
- GPU’
-
15. An apparatus for data transformation offloading in an artificial intelligence infrastructure that includes one or more storage systems and one or more graphical processing unit (‘
- GPU’
) servers, the apparatus comprising a computer processor, a computer memory operatively coupled to the computer processor, the computer memory having disposed within it computer program instructions that, when executed by the computer processor, cause the apparatus to carry out the steps of;identifying, in dependence upon one or more machine learning models to be executed on the GPU servers, one or more transformations to apply to a dataset; generating, in dependence upon the one or more transformations, a transformed dataset; storing, within one or more of the storage systems, the transformed dataset; receiving a plurality of requests to transmit the transformed dataset to one or more of the GPU servers; and responsive to each request, transmitting, from the one or more storage systems to the one or more GPU servers without re-performing the one or more transformations on the dataset, the transformed dataset. - View Dependent Claims (16, 17, 18, 19, 20)
- GPU’
Specification