METHODS, APPARATUS, AND ARTICLES OF MANUFACTURE TO IMPROVE PERFORMANCE OF AN ARTIFICIAL INTELLIGENCE BASED MODEL ON DATASETS HAVING DIFFERENT DISTRIBUTIONS

    公开(公告)号:US20220335285A1

    公开(公告)日:2022-10-20

    申请号:US17853518

    申请日:2022-06-29

    Abstract: Methods, apparatus, systems, and articles of manufacture are disclosed to improve performance of an artificial intelligence based (AI-based) model on datasets having different distributions. An example apparatus includes interface circuitry to access data, computer readable instructions, and processor circuitry to at least one of instantiate or execute the computer readable instructions to implement adversarial evaluation circuitry, convolution circuitry, and output control circuitry. The example adversarial evaluation circuitry is to determine whether the data is to be processed as adversarial data. The example convolution circuitry is to, based on whether the data is to be processed as the adversarial data, determine a convolution of an input tensor and (1) a parameter tensor corresponding to a layer of the AI-based model or (2) a noisy parameter tensor generated based on the parameter tensor. The example output control circuitry is to output a classification of the data based on the convolution.

    DEEP NEURAL NETWORK OPTIMIZATION SYSTEM FOR MACHINE LEARNING MODEL SCALING

    公开(公告)号:US20220036194A1

    公开(公告)日:2022-02-03

    申请号:US17504282

    申请日:2021-10-18

    Abstract: The present disclosure is related to techniques for optimizing artificial intelligence (AI) and/or machine learning (ML) models to reduce resource consumption while maintaining or improving AI/ML model performance. A sparse distillation framework (SDF) is provided for producing a class of parameter and compute efficient AI/ML models suitable for resource constrained applications. The SDF simultaneously distills knowledge from a compute heavy teacher model while also pruning a student model in a single pass of training, thereby reducing training and tuning times considerably. A self-attention mechanism may also replace CNNs or convolutional layers of a CNN to have better translational equivariance. Other embodiments may be described and/or claimed.

Patent Agency Ranking