-
公开(公告)号:US20230153565A1
公开(公告)日:2023-05-18
申请号:US17430644
申请日:2021-07-09
Applicant: SAMSUNG ELECTRONICS CO., LTD
Inventor: Brijraj SINGH , Mayukh DAS , Yash Hemant JAIN , Sharan Kumar ALLUR , Venkappa MALA , Praveen Doreswamy NAIDU
Abstract: A method of deep neural network (DNN) modularization for optimal loading includes receiving, by an electronic device, a DNN model for execution, obtaining, by the electronic device, a plurality of parameters associated with the electronic device and a plurality of parameters associated with the DNN model, determining, by the electronic device, a number of sub-models of the DNN model and a splitting index, based on the obtained plurality of parameters associated with the electronic device and the obtained plurality of parameters associated with the DNN model, and splitting, by the electronic device, the received DNN model into a plurality of sub-models, based on the determined number of sub-models of the DNN model and the determined splitting index.
-
2.
公开(公告)号:US20200019854A1
公开(公告)日:2020-01-16
申请号:US16480545
申请日:2018-02-23
Applicant: Samsung Electronics Co., Ltd.
Inventor: Arun ABRAHAM , Suhas Parlathaya KUDRAL , Balaji Srinivas HOLUR , Sarbojit GANGULY , Venkappa MALA , Suneel Kumar SURIMANI , Sharan Kumar ALLUR
Abstract: The present invention describes a method of accelerating execution of one or more application tasks in a computing device using machine learning (ML) based model. According to one embodiment, a neural accelerating engine present in the computing device receives a ML input task for execution on the computing device from a user. The neural accelerating engine further retrieves a trained ML model and a corresponding optimal configuration file based on the received ML input task. Also, the current performance status of the computing device for executing the ML input task is obtained. Then, the neural accelerating engine dynamically schedules and dispatches parts of the ML input task to one or more processing units in the computing device for execution based on the retrieved optimal configuration file and the obtained current performance status of the computing device.
-
公开(公告)号:US20230127001A1
公开(公告)日:2023-04-27
申请号:US18082305
申请日:2022-12-15
Applicant: SAMSUNG ELECTRONICS CO., LTD.
Inventor: Mayukh DAS , Brijraj SINGH , Pradeep NELAHONNE SHIVAMURTHAPPA , Aakash KAPOOR , Rajath Elias SOANS , Soham Vijay DIXIT , Sharan Kumar ALLUR , Venkappa MALA
IPC: G06N3/082 , G06V10/776 , G06V10/82 , G06V40/16
Abstract: A method for generating an optimal neural network (NN) model may include determining intermediate outputs of the NN model by passing an input dataset through each intermediate exit gate of the plurality of intermediate exit gates, determining an accuracy score for each intermediate exit gate of the plurality of intermediate exit gates based on a comparison of the final output of the NN model with the intermediate output, identifying an earliest intermediate exit gate that produces the intermediate output closer to the final output based on the accuracy score, and generating the optimal NN model by removing remaining layers of the plurality of layers and remaining intermediate exit gates of the plurality of intermediate exit gates located after the determined earliest intermediate exit gate.
-
公开(公告)号:US20180260243A1
公开(公告)日:2018-09-13
申请号:US15571191
申请日:2016-06-01
Applicant: Samsung Electronics Co., Ltd.
Inventor: Aniruddha BANERJEE , Sharan Kumar ALLUR , Syam Prasad KUNCHA
Abstract: The embodiments herein provide a method for scheduling an entity in a multi-core processor system including a big-core processor, and a little-core processor. The method includes detecting, by a scheduler, that a load contribution of the entity exceeds a load threshold. Further, the method includes determining, by the scheduler, whether the entity is one of a background entity, an IO intensive entity, a non-background entity, and a non-IO intensive entity based on at least one parameter. Further, the method includes instructing, by the scheduler, one of to schedule the entity on a little-core processor when the entity is at least one of the background entity and the IO intensive entity; and to schedule the entity on the big-core processor when the entity is at least one of the non-background entity and the non-IO intensive entity.
-
-
-