METHOD OF EXECUTING TASK FOR LARGE LANGUAGE MODEL, DEVICE, AND STORAGE MEDIUM

    公开(公告)号:US20240378077A1

    公开(公告)日:2024-11-14

    申请号:US18782617

    申请日:2024-07-24

    Abstract: A method of executing a task for a large language model, a device, and a storage medium are provided, which relate to a field of artificial intelligence technology, and in particular to fields of deep learning, large language model, natural language processing and computer vision technologies. The method includes: determining, by using a determination unit, a target attention task from a plurality of attention tasks to be processed, based on a sparse representation corresponding to a feature to be processed, where the target attention task is a task corresponding to a non-fully masked region of the feature, the sparse representation represents a mask position of the feature, and the mask position represents mask endpoint positions in at least two non-intersecting intervals in a mask matrix corresponding to the feature; and executing the target attention task by using a computing unit, so as to obtain an attention feature.

    Method and Apparatus for Generating and Applying Deep Learning Model based on Deep Learning Framework

    公开(公告)号:US20230185702A1

    公开(公告)日:2023-06-15

    申请号:US17856091

    申请日:2022-07-01

    CPC classification number: G06F11/3688 G06N3/08

    Abstract: A method and apparatus is provided for generating and applying a deep learning model based on a deep learning framework, and relates to the field of computers. A specific implementation solution includes that a basic operating environment is established on a target device, where the basic operating environment is used for providing environment preparation for an overall generation process of a deep learning model; a basic function of the deep learning model is generated in the basic operating environment according to at least one of a service requirement and a hardware requirement, to obtain a first processing result; an extended function of the deep learning model is generated in the basic operating environment based on the first processing result, to obtain a second processing result; and a preset test script is used to perform function test on the second processing result, to output a test result.

    Neural Network Training Method and Apparatus, Electronic Device, Medium and Program Product

    公开(公告)号:US20220374704A1

    公开(公告)日:2022-11-24

    申请号:US17558355

    申请日:2021-12-21

    Abstract: The disclosure provides a neural network training method and apparatus, an electronic device, a medium and a program product, and relates to the field of artificial intelligence, in particular to the fields of deep learning and distributed learning. The method includes: acquiring a neural network for deep learning; constructing a deep reinforcement learning model for the neural network; and determining, through the deep reinforcement learning model, a processing unit selection for the plurality of the network layers based on a duration for training each of the network layers by each type of the plurality of types of the processing units, and a cost of each type of the plurality of types of the processing units, wherein the processing unit selection comprises the type of the processing unit to be used for each of the plurality of the network layers, and the processing unit selection is used for making a total cost of the processing units used by the neural network below a cost threshold, in response to a duration for pipelining parallel computing for training the neural network being shorter than a present duration.

Patent Agency Ranking