-
公开(公告)号:US11741345B2
公开(公告)日:2023-08-29
申请号:US17033573
申请日:2020-09-25
Applicant: Amazon Technologies, Inc.
Inventor: Randy Huang , Ron Diamant
CPC classification number: G06N3/045 , G06F3/061 , G06F3/065 , G06F3/0683 , G06F13/28 , G06F13/4068 , G06F15/80
Abstract: Provided are systems, methods, and integrated circuits for a neural network processing system. In various implementations, the system can include a first array of processing engines coupled to a first set of memory banks and a second array of processing engines coupled to a second set of memory banks. The first and second set of memory banks be storing all the weight values for a neural network, where the weight values are stored before any input data is received. Upon receiving input data, the system performs a task defined for the neural network. Performing the task can include computing an intermediate result using the first array of processing engines, copying the intermediate result to the second set of memory banks, and computing a final result using the second array of processing engines, where the final result corresponds to an outcome of performing the task.
-
公开(公告)号:US11250319B1
公开(公告)日:2022-02-15
申请号:US15714924
申请日:2017-09-25
Applicant: Amazon Technologies, Inc.
Inventor: Randy Huang , Ron Diamant
Abstract: Disclosed herein are techniques for classifying data with a data processing circuit. In one embodiment, the data processing circuit includes a probabilistic circuit configurable to generate a decision at a pre-determined probability, and an output generation circuit including an output node and configured to receive input data and a weight, and generate output data at the output node for approximating a product of the input data and the weight. The generation of the output data includes propagating the weight to the output node according a first decision of the probabilistic circuit. The probabilistic circuit is configured to generate the first decision at a probability determined based on the input data.
-
公开(公告)号:US20190294968A1
公开(公告)日:2019-09-26
申请号:US15933201
申请日:2018-03-22
Applicant: Amazon Technologies, Inc.
Inventor: Dana Michelle Vantrease , Ron Diamant , Thomas A. Volpe , Randy Huang
Abstract: Disclosed herein are techniques for performing multi-layer neural network processing for multiple contexts. In one embodiment, a computing engine is set in a first configuration to implement a second layer of a neural network and to process first data related to a first context to generate first context second layer output. The computing engine can be switched from the first configuration to a second configuration to implement a first layer of the neural network. The computing engine can be used to process second data related to a second context to generate second context first layer output. The computing engine can be set to a third configuration to implement a third layer of the neural network to process the first context second layer output and the second context first layer output to generate a first processing result of the first context and a second processing result of the second context.
-
公开(公告)号:US20230334294A1
公开(公告)日:2023-10-19
申请号:US18339954
申请日:2023-06-22
Applicant: Amazon Technologies, Inc.
Inventor: Randy Huang , Ron Diamant
CPC classification number: G06N3/045 , G06F15/80 , G06F13/28 , G06F3/0683 , G06F3/061 , G06F3/065 , G06F13/4068
Abstract: Provided are systems, methods, and integrated circuits for neural network processing. In various implementations, an integrated circuit for neural network processing can include a plurality of memory banks storing weight values for a neural network. The memory banks can be on the same chip as an array of processing engines. Upon receiving input data, the circuit can be configured to use the set of weight values to perform a task defined for the neural network. Performing the task can include reading weight values from the memory banks, inputting the weight values into the array of processing engines, and computing a result using the array of processing engines, where the result corresponds to an outcome of performing the task.
-
公开(公告)号:US11475306B2
公开(公告)日:2022-10-18
申请号:US15933201
申请日:2018-03-22
Applicant: Amazon Technologies, Inc.
Inventor: Dana Michelle Vantrease , Ron Diamant , Thomas A. Volpe , Randy Huang
Abstract: Disclosed herein are techniques for performing multi-layer neural network processing for multiple contexts. In one embodiment, a computing engine is set in a first configuration to implement a second layer of a neural network and to process first data related to a first context to generate first context second layer output. The computing engine can be switched from the first configuration to a second configuration to implement a first layer of the neural network. The computing engine can be used to process second data related to a second context to generate second context first layer output. The computing engine can be set to a third configuration to implement a third layer of the neural network to process the first context second layer output and the second context first layer output to generate a first processing result of the first context and a second processing result of the second context.
-
公开(公告)号:US10817260B1
公开(公告)日:2020-10-27
申请号:US16007749
申请日:2018-06-13
Applicant: Amazon Technologies, Inc.
Inventor: Randy Huang , Ron Diamant , Thomas Elmer , Sundeep Amirineni , Thomas A. Volpe
Abstract: Systems and methods are provided to skip multiplication operations with zeros in processing elements of the systolic array to reduce dynamic power consumption. A value of zero can be detected on an input data element entering each row of the array and respective zero indicators may be generated. These respective zero indicators may be passed to all the processing elements in the respective rows. The multiplication operation with the zero value can be skipped in each processing element based on the zero indicators, thus reducing dynamic power consumption.
-
公开(公告)号:US10445638B1
公开(公告)日:2019-10-15
申请号:US15908236
申请日:2018-02-28
Applicant: Amazon Technologies, Inc.
Inventor: Sundeep Amirineni , Ron Diamant , Randy Huang , Thomas A. Volpe
Abstract: Disclosed herein are techniques for performing neural network computations. In one embodiment, an apparatus may include an array of processing elements, the array having a configurable first effective dimension and a configurable second effective dimension. The apparatus may also include a controller configured to determine at least one of: a first number of input data sets to be provided to the array at the first time or a second number of output data sets to be generated by the array at the second time, and to configure, based on at least one of the first number or the second number, at least one of the first effective dimension or the second effective dimension of the array.
-
公开(公告)号:US12008466B1
公开(公告)日:2024-06-11
申请号:US15934469
申请日:2018-03-23
Applicant: Amazon Technologies, Inc.
Inventor: Randy Huang , Ron Diamant , Thomas A. Volpe
IPC: G06N3/08
CPC classification number: G06N3/08
Abstract: In various implementations, provided are systems and methods for operating a neural network that includes conditional structures. In some implementations, an integrated circuit can compute a result using a set of intermediate results, where the intermediate results are computed from the outputs of a hidden layer of the neural network. The integrated circuit can further test the result against a condition. The outcome of the test can determine a next layer that the integrated circuit is to execute, or can be used to determine that further execution of the neural network can be terminated.
-
公开(公告)号:US20230014783A1
公开(公告)日:2023-01-19
申请号:US17951084
申请日:2022-09-22
Applicant: Amazon Technologies, Inc.
Inventor: Dana Michelle Vantrease , Ron Diamant , Thomas A. Volpe , Randy Huang
Abstract: Disclosed herein are techniques for performing multi-layer neural network processing for multiple contexts. In one embodiment, a computing engine is set in a first configuration to implement a second layer of a neural network and to process first data related to a first context to generate first context second layer output. The computing engine can be switched from the first configuration to a second configuration to implement a first layer of the neural network. The computing engine can be used to process second data related to a second context to generate second context first layer output. The computing engine can be set to a third configuration to implement a third layer of the neural network to process the first context second layer output and the second context first layer output to generate a first processing result of the first context and a second processing result of the second context.
-
公开(公告)号:US10943167B1
公开(公告)日:2021-03-09
申请号:US16538698
申请日:2019-08-12
Applicant: Amazon Technologies, Inc.
Inventor: Sundeep Amirineni , Ron Diamant , Randy Huang , Thomas A. Volpe
Abstract: Disclosed herein are techniques for performing neural network computations. In one embodiment, an apparatus includes an array of processing elements, the array having configurable dimensions. The apparatus further includes a controller configured to set the dimensions of the array of processing elements based on at least one of: a first number of input data sets to be received by the array, or a second number of output data sets to be output by the array.
-
-
-
-
-
-
-
-
-