-
1.
公开(公告)号:US20250077780A1
公开(公告)日:2025-03-06
申请号:US18748642
申请日:2024-06-20
Inventor: Yongkang Xie , Guming Gao , Penghao Zhao , Xue Xiong , Qian Wang , Dongze Xu , En Shi , Yuxuan Li , Sheng Zhou , Shupeng Li , Yao Wang , Zhou Xin
Abstract: A method for invoking a plugin of a large language model includes: acquiring natural language content; performing semantic understanding on the natural language content and detecting whether the natural language content hits a plugin to obtain a first plugin pointed to by the plugin hit result; comparing the first plugin with a second plugin corresponding to the current session understanding task to determine a to-be-executed session understanding task and a third plugin corresponding to the to-be-executed session understanding task; acquiring the language understanding content of the to-be-executed session understanding task and sending the language understanding content to the large language model to obtain the input parameter of the third plugin; and calling the third plugin according to the input parameter of the third plugin to obtain the calling result of the to-be-executed session understanding task.
-
公开(公告)号:US20220391182A1
公开(公告)日:2022-12-08
申请号:US17820095
申请日:2022-08-16
Inventor: En Shi , Yongkang Xie , Zihao Pan , Shupeng Li , Xiaoyu Chen , Zhengyu Qian , Jingqiu Li
Abstract: A method for model production includes acquiring a related operation for model production from a user interface layer of a model production system, and determining a software platform of the model production system; acquiring a model service corresponding to the related operation by invoking an application programming interface (API) corresponding to the related operation, wherein the API is located between the user interface layer and other layer in the model production system; performing the model service by invoking local resources of the software platform with a tool of the software platform adapted to the model service, to generate a target model; and applying the target model in a target usage scene.
-
公开(公告)号:US20220374742A1
公开(公告)日:2022-11-24
申请号:US17817015
申请日:2022-08-03
Inventor: Zhengxiong Yuan , Zhengyu Qian , En Shi , Mingren Hu , Jinqi Li , Zhenfang Chu , Runqing Li , Yue Huang
Abstract: A method for running an inference service platform, includes: determining inference tasks to be allocated for the inference service platform, in which the inference service platform includes two or more inference service groups, versions of the inference service groups are different, and the inference service groups are configured to perform a same type of inference services; determining a flow weight of each of the inference service groups, in which the flow weight is configured to indicate a proportion of a number of inference tasks to which the corresponding inference service group need to be allocated in a total number of inference tasks; and allocating the corresponding number of inference tasks in the inference tasks to be allocated to each of the inference service groups based on the flow weight of each of the inference service groups; and performing the inference tasks by the inference service group.
-
公开(公告)号:US20230401484A1
公开(公告)日:2023-12-14
申请号:US18076711
申请日:2022-12-07
Inventor: Chao WANG , Xiangyue Lin , Yang Liang , En Shi , Shuangshuang QIAO
IPC: G06N20/00
CPC classification number: G06N20/00
Abstract: Provided are a data processing method and apparatus, an electronic device, and a storage medium. The data processing method includes acquiring a target directed acyclic graph (DAG) corresponding to the service processing logic of a model self-taught learning service, where the service processing logic includes execution logic for acquiring service data generated by an online released service model, execution logic for training a to-be-trained service model based on the service data, and execution logic for releasing the trained service model online; and performing self-taught learning on the to-be-trained service model according to the target DAG.
-
5.
公开(公告)号:US20240005182A1
公开(公告)日:2024-01-04
申请号:US17982128
申请日:2022-11-07
Inventor: Jinqi Li , En Shi , Mingren Hu , Zhengyu Qian , Zhengxiong Yuan , Zhenfang Chu , Yue Huang , Yang Luo , Guobin Wang
IPC: G06N5/04
CPC classification number: G06N5/04
Abstract: Provided are a streaming media processing method based on inference service, an electronic device, and a storage medium, which relates to the field of artificial intelligence, and in particular, to the field of inference service of artificial intelligence models. The method includes: detecting, in a process of processing a k-th channel of streaming media through an i-th inference service pod, the i-th inference service pod, to obtain a detection result of the i-th inference service pod, i and k being positive integers; determining a replacement object of the i-th inference service pod, in the case where it is determined that the i-th inference service pod is in an abnormal state based on the detection result of the i-th inference service pod; and processing the k-th channel of streaming media through the replacement object of the i-th inference service pod.
-
公开(公告)号:US12182546B2
公开(公告)日:2024-12-31
申请号:US17820095
申请日:2022-08-16
Inventor: En Shi , Yongkang Xie , Zihao Pan , Shupeng Li , Xiaoyu Chen , Zhengyu Qian , Jingqiu Li
Abstract: A method for model production includes acquiring a related operation for model production from a user interface layer of a model production system, and determining a software platform of the model production system; acquiring a model service corresponding to the related operation by invoking an application programming interface (API) corresponding to the related operation, wherein the API is located between the user interface layer and other layer in the model production system; performing the model service by invoking local resources of the software platform with a tool of the software platform adapted to the model service, to generate a target model; and applying the target model in a target usage scene.
-
-
-
-
-