-
公开(公告)号:US11720789B2
公开(公告)日:2023-08-08
申请号:US16672352
申请日:2019-11-01
Applicant: Apple Inc.
Inventor: Hessam Bagherinezhad , Dmitry Belenko
CPC classification number: G06N3/08 , G06F16/90335 , G06F17/16 , G06F40/30 , G06N3/04
Abstract: In one embodiment, a method includes receiving an input vector corresponding to a query at a neural network model comprising a plurality of layers, wherein the plurality of layers comprise a last layer associated with a mapping matrix, generating a binary matrix based on the mapping matrix, an identity matrix, and one or more Gaussian vectors, generating an integer vector based on the binary matrix and a binary vector associated with the input vector, identifying a plurality of indices corresponding to a plurality of top values of the integer vector for the integer vector, generating an output vector based on the input vector and a plurality of rows of the mapping matrix, wherein the plurality of rows is associated with the plurality of identified indices, respectively, and determining the query is associated with one or more classes based on the output vector.
-
公开(公告)号:US11657124B2
公开(公告)日:2023-05-23
申请号:US16215540
申请日:2018-12-10
Applicant: Apple Inc.
Inventor: Peter Zatloukal , Matthew Weaver , Alexander Kirchhoff , Dmitry Belenko , Ali Farhadi , Mohammad Rastegari , Andrew Luke Chronister , Keith Patrick Wyss , Chenfan Sun
CPC classification number: G06F21/105 , G06F21/12 , G06N3/08 , G06N3/10 , H04L9/0891 , H04L9/30 , G06F2221/0755
Abstract: In one embodiment, a method includes receiving a user request from a client device associated with a user, accessing an instructional file comprising one or more binary inference engines and one or more encrypted model data corresponding to the one or more binary inference engines, respectively, selecting a binary inference engine from the one or more binary inference engines in the accessed instructional file based on the user request, sending a validation request for a permission to execute the binary inference engine to a licensing server, receiving the permission from the licensing server, decrypting the encrypted model data corresponding to the binary inference engine by a decryption key, executing the binary inference engine based on the user request and the decrypted model data, and sending one or more execution results responsive to the execution of the binary inference engine to the client device.
-