DEEP NEURAL NETWORK PROCESSING ON HARDWARE ACCELERATORS WITH STACKED MEMORY
First Claim
1. A method for processing on an acceleration component a deep neural network, the acceleration component comprising an acceleration component die and a memory stack disposed in an integrated circuit package, the memory stack comprising a memory bandwidth greater than about 50 GB/sec and a power efficiency of greater than about 20 MB/sec/mW, the method comprising:
- configuring the acceleration component to perform forward propagation and backpropagation stages of the deep neural network.
1 Assignment
0 Petitions
Accused Products
Abstract
A method is provided for processing on an acceleration component a deep neural network. The method includes configuring the acceleration component to perform forward propagation and backpropagation stages of the deep neural network. The acceleration component includes an acceleration component die and a memory stack disposed in an integrated circuit package. The memory stack has a memory bandwidth greater than about 50 GB/sec and a power efficiency of greater than about 20 MB/sec/mW.
121 Citations
20 Claims
-
1. A method for processing on an acceleration component a deep neural network, the acceleration component comprising an acceleration component die and a memory stack disposed in an integrated circuit package, the memory stack comprising a memory bandwidth greater than about 50 GB/sec and a power efficiency of greater than about 20 MB/sec/mW, the method comprising:
configuring the acceleration component to perform forward propagation and backpropagation stages of the deep neural network. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9)
-
10. A system for processing a deep neural network, the system comprising:
-
an acceleration component comprising an acceleration component die and a memory stack disposed in an integrated circuit package, the memory stack comprising a memory bandwidth greater than about 50 GB/sec and a power efficiency of greater than about 20 MB/sec/mW; a plurality of neural engines configured on acceleration component die, wherein the neural engines comprise logic to implement forward propagation and backpropagation stages of the deep neural network. - View Dependent Claims (11, 12, 13, 14, 15, 16, 17, 20)
-
-
18. A system for processing a deep neural network, the system comprising:
-
an acceleration component comprising an acceleration component die and a memory stack disposed in an integrated circuit package, the memory stack comprising a memory bandwidth greater than about 50 GB/sec and a power efficiency of greater than about 20 MB/sec/mW; a plurality of neural engines configured on the acceleration component die; and a plurality of DRAM channels on the memory stack, each of the DRAM channels coupled to the neural engines, wherein the neural engines comprise logic to implement forward propagation and backpropagation stages of the deep neural network. - View Dependent Claims (19)
-
Specification