Efficient serverless method and system of serving artificial intelligence models

    公开(公告)号:US12231491B1

    公开(公告)日:2025-02-18

    申请号:US18509247

    申请日:2023-11-14

    Abstract: A method for forecasting server demand includes collecting a historical number of scoring requests from a network using a serverless architecture. A scoring request capacity per server is determined using the historical number of scoring requests. A prediction model predicts a first future value of scoring requests for a first future time span. A current number of servers in a pool of servers handling the scoring requests. Using the prediction model, a determination of whether the current number of servers is capable of handling the first future value of scoring requests for the first future time span is made. Upon determining that the current number of servers is incapable of handling the first future value of scoring requests, one or more additional servers are warmed up. The warmed-up additional servers are added to the pool of servers prior to an arrival of the first future time span.

    POST-MODELING VISUALIZATION
    2.
    发明申请

    公开(公告)号:US20250053858A1

    公开(公告)日:2025-02-13

    申请号:US18446149

    申请日:2023-08-08

    Abstract: In an approach, a processor selects a top N features for a machine learning (ML) model; discretizes values of each continuous feature of the top N features; generates a set of combination values that each represent a unique combination of feature values in for a data record; predicts, using the ML model, a target value for each record generating predicted target values; groups the predicted target values based on the combination value for each respective record; fits a distribution for each grouping of the predicted target values associated with a respective combination value generating a set of distributions; clusters and refits the set of distributions using a clustering algorithm resulting in a set of clusters and a refitted distribution for each cluster of the set of clusters; and outputs a visualization of the refitted distribution for each cluster as a distribution curve on a graph along with the associated records.

    Compatible and secure software upgrades

    公开(公告)号:US11836483B1

    公开(公告)日:2023-12-05

    申请号:US17804322

    申请日:2022-05-27

    CPC classification number: G06F8/71 G06N20/00

    Abstract: Described are techniques for machine learning library management. The techniques include generating a table including a plurality of machine learning libraries and their current versions that are used in a deployed machine learning platform (MLP) instance, a first available version upgrade for a first machine learning library of the plurality of machine learning libraries, a security indication associated with the first available version upgrade relative to a current version implemented by the first machine learning library, and a compatibility indication between the first available version upgrade and the current version of the first machine learning library. The techniques further include generating a recommendation related to upgrading the first machine learning library based on the security indication and the compatibility indication.

    SYSTEM AND METHOD FOR EFFICIENT TRANSFORMATION PREDICTION IN A DATA ANALYTICS PREDICTION MODEL PIPELINE

    公开(公告)号:US20230359941A1

    公开(公告)日:2023-11-09

    申请号:US17739716

    申请日:2022-05-09

    CPC classification number: G06N20/20 G06Q20/4016

    Abstract: A computer-implemented system, platform, programing product, and/or method for improving transformation selection in an ensemble machine learning (ML) model that includes: providing all base ML models of the ensemble ML model; identifying all of a plurality of Derived Fields in all the base ML models; performing a Derived Field run prediction analysis for all the Derived Fields; computing the Derived Field Importance Weight for Field (DFIW4F) and the Derived Field Importance Weight for Model (DFIW4M) for all the Derived Fields; clustering all the Derived Fields into a plurality of Derived Field clusters, wherein each Derived Field cluster is based upon the DFIW4M and the DFIW4F for the Derived Field; sorting all the Derived Field clusters by best cluster based upon DFIW4M and DFIW4F; and running the base ML models based upon the Derived Fields in the best Derived Field cluster until sufficient base ML models have been run.

    INTELLIGENT EXPANSION OF REVIEWER FEEDBACK ON TRAINING DATA

    公开(公告)号:US20230214454A1

    公开(公告)日:2023-07-06

    申请号:US17568305

    申请日:2022-01-04

    CPC classification number: G06K9/6257 G06K9/6263 G06K9/6219 G06N20/20

    Abstract: An embodiment generates an initial set of training data from monitoring data. The initial set of training data is generated by combining outputs from a plurality of pretrained classifiers. The embodiment trains a new classification model using the initial set of training data to identify anomalies in monitoring data. The embodiment performs a multiple-level clustering of the data samples resulting in a plurality of clusters of sub-clusters of data samples, and generates a review list of data samples by selecting a representative data sample from each of the clusters. The embodiment receives an updated data sample from the expert review that includes a revised target classification for at least one of the data samples of the expert review list. The embodiment then trains another replacement classification model using a revised set of training data that includes the updated data sample and associated revised target classification.

    EFFICIENT MACHINE LEARNING MODEL INFERENCE

    公开(公告)号:US20230138987A1

    公开(公告)日:2023-05-04

    申请号:US17453565

    申请日:2021-11-04

    Abstract: One or more computer processors calculate a cache prediction for a received inference request within an inference cache structured as a self-learning tree, wherein the inference request comprises a set of input values. The one or more computer processors responsive to the retrieved cache prediction exceeding a cache prediction threshold, transmit the cache prediction. The one or more computer processors parallel compute a model prediction for the received inference request utilizing a trained model. The one or more computer processors responsive to the retrieved model prediction exceeding a model prediction threshold, convert the trained model into a tree structure. The one or more computer processors update the inference cache with the converted train model. The one or more computer processors transmit the model prediction.

Patent Agency Ranking