NEURAL PROCESSING UNIT (NPU) DIRECT MEMORY ACCESS (NDMA) MEMORY BANDWIDTH OPTIMIZATION

    公开(公告)号:US20220230058A1

    公开(公告)日:2022-07-21

    申请号:US17713176

    申请日:2022-04-04

    Abstract: A neural processing unit (NPU) is described. The NPU includes an NPU direct memory access (NDMA) core. The NDMA core includes a read engine having a read buffer. The NDMA core also includes a write engine having a write buffer. The NPU also includes a controller. The controller is configured to direct the NDMA core to perform hardware memory bandwidth optimization for reading/writing NDMA data in the read buffer and/or NDMA data in the write buffer. The NDMA core is also configured to transparently combine NDMA transaction requests for a data stripe to increase local access to available tensors in artificial neural networks.

    OPTIMIZING PERFORMANCE OF RECURRENT NEURAL NETWORKS

    公开(公告)号:US20190325289A1

    公开(公告)日:2019-10-24

    申请号:US15956674

    申请日:2018-04-18

    Abstract: An apparatus for optimizing a computational network is configure to receive an input at a first processing component. The first processing component may include at least a first programmable processing component and a second programmable processing component. The first programmable processing component is configured to compute a first nonlinear function and the second programmable processing component is configured to compute a second nonlinear function which is different than the second nonlinear function. The computational network which may be a recurrent neural network such as a long short-term memory may be operated to generate an inference based at least in part on outputs of the first programmable processing component and the second programmable processing component.

Patent Agency Ranking