-
公开(公告)号:US20230229921A1
公开(公告)日:2023-07-20
申请号:US17576101
申请日:2022-01-14
Applicant: Arm Limited
Inventor: Igor Fedorov , Paul Nicholas Whatmough
Abstract: Neural network systems and methods are provided. One method for processing a neural network includes, for at least one neural network layer that includes a plurality of weights, applying an offset function to each of a plurality of weight values in the plurality of weights to generate an offset weight value, and quantizing the offset weight values to form quantized offset weight values. The plurality of weights are pruned. One method for executing a neural network includes reading, from a memory, at least one neural network layer that includes quantized offset weight values and an offset value α, and performing a neural network layer operation on an input feature map, based on the quantized offset weight values and the offset value α, to generate an output feature map. The quantized offset weight values are signed integer numbers.
-
2.
公开(公告)号:US20240046065A1
公开(公告)日:2024-02-08
申请号:US17817142
申请日:2022-08-03
Applicant: Arm Limited
Inventor: Hokchhay Tann , Ramon Matas Navarro , Igor Fedorov , Chuteng Zhou , Paul Nicholas Whatmough , Matthew Mattina
IPC: G06N3/04
CPC classification number: G06N3/04
Abstract: Example methods, apparatuses, and/or articles of manufacture are disclosed that may be implemented, in whole or in part, using one or more computing devices to determine options for decisions in connection with design features of a computing device. In a particular implementation, design options for two or more design decisions of neural network processing device may be identified based, at least in part, on combination of a definition of available computing resources and one or more predefined performance constraints.
-
公开(公告)号:US20230042271A1
公开(公告)日:2023-02-09
申请号:US17394048
申请日:2021-08-04
Applicant: Arm Limited
Inventor: Igor Fedorov , Ramon Matas Navarro , Chuteng Zhou , Hokchhay Tann , Paul Nicholas Whatmough , Matthew Mattina
Abstract: Example methods, apparatuses, and/or articles of manufacture are disclosed that may be implemented, in whole or in part, using one or more computing devices to select options for decisions in connection with design features of a computing device. In a particular implementation, design options for two or more design decisions of neural network processing device may be selected based, at least in part, on combination of function values that are computed based, at least in part, on a tensor expressing sample neural network weights.
-
-