Joint Architecture And Hyper-Parameter Search For Machine Learning Models

    公开(公告)号:US20210383223A1

    公开(公告)日:2021-12-09

    申请号:US17337834

    申请日:2021-06-03

    Applicant: Google LLC

    Abstract: The present disclosure provides a differentiable joint hyper-parameter and architecture search approach, with some implementations including the idea of discretizing the continuous space into a linear combination of multiple categorical basis. One example element of the proposed approach is the use of weight sharing across all architecture- and hyper-parameters which enables it to search efficiently over the large joint search space. Experimental results on MobileNet/ResNet/EfficientNet/BERT show that the proposed systems significantly improve the accuracy by up to 2% on ImageNet and the F1 by up to 0.4 on SQuAD, with search cost comparable to training a single model. Compared to other AutoML methods, such as random search or Bayesian method, the proposed techniques can achieve better accuracy with 10× less compute cost.

    MACHINE LEARNING ALGORITHM SEARCH WITH SYMBOLIC PROGRAMMING

    公开(公告)号:US20230144138A1

    公开(公告)日:2023-05-11

    申请号:US17905196

    申请日:2021-06-04

    Applicant: Google LLC

    CPC classification number: G06N20/00 G06N5/01

    Abstract: A method for searching for an output machine learning (ML) algorithm to perform an ML task is described. The method comprising: receiving data specifying an input ML algorithm; receiving data specifying a search algorithm that searches for candidate ML algorithms and an evaluation function that evaluates the performance of candidate ML algorithms; generating data representing a symbolic tree from the input ML algorithm; generating data representing a hyper symbolic tree from the symbolic tree; searching an algorithm search space that defines a set of possible concrete symbolic trees from the hyper symbolic tree for candidate ML algorithms and training the candidate ML algorithms to determine a respective performance metric for each candidate ML algorithm; and selecting one or more trained candidate ML algorithms among the trained candidate ML algorithms based on the determined performance metrics.

    REINFORCEMENT LEARNING ALGORITHM SEARCH

    公开(公告)号:US20220391687A1

    公开(公告)日:2022-12-08

    申请号:US17338093

    申请日:2021-06-03

    Applicant: Google LLC

    Abstract: Methods, computer systems, and apparatus, including computer programs encoded on computer storage media, for generating and searching reinforcement learning algorithms. In some implementations, a computer-implemented system generates a sequence of candidate reinforcement learning algorithms. Each candidate reinforcement learning algorithm in the sequence is configured to receive an input environment state characterizing a state of an environment and to generate an output that specifies an action to be performed by an agent interacting with the environment. For each candidate reinforcement learning algorithm in the sequence, the system performs a performance evaluation for a set of a plurality of training environments. For each training environment, the system adjusts a set of environment-specific parameters of the candidate reinforcement learning algorithm by performing training of the candidate reinforcement learning algorithm to control a corresponding agent in the training environment. The system generates an environment-specific performance metric for the candidate reinforcement learning algorithm that measures a performance of the candidate reinforcement learning algorithm in controlling the corresponding agent in the training environment as a result of the training. After performing training in the set of training environments, the system generates a summary performance metric for the candidate reinforcement learning algorithm by combining the environment-specific performance metrics generated for the set of training environments. After evaluating each of the candidate reinforcement learning algorithms in the sequence, the system selects one or more output reinforcement learning algorithms from the sequence based on the summary performance metrics of the candidate reinforcement learning algorithms.

    Proxy Task Design Tools for Neural Architecture Search

    公开(公告)号:US20240289605A1

    公开(公告)日:2024-08-29

    申请号:US18173347

    申请日:2023-02-23

    Applicant: Google LLC

    CPC classification number: G06N3/08

    Abstract: Aspects of the disclosure are directed to proxy task design tools that automatically find proxy tasks, such as optimal proxy tasks, for neural architecture searches. The proxy task design tools can include one or more tools to search for an optimal proxy task having the lowest neural architecture search cost while meeting a minimum correlation requirement threshold after being provided with a proxy task search space definition. The proxy task design tools can further include one or more tools to select candidate models for computing correlation scores of proxy tasks as well as one or more tools to measure variance of a model. The proxy task design tools can minimize time and effort involved in designing the proxy task.

    NEURAL ARCHITECTURE AND HARDWARE ACCELERATOR SEARCH

    公开(公告)号:US20240005129A1

    公开(公告)日:2024-01-04

    申请号:US18029849

    申请日:2021-10-01

    Applicant: Google LLC

    CPC classification number: G06N3/045 G06N3/092 G06N3/0464 G06N3/044 G06N3/063

    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for jointly determining neural network architectures and hardware accelerator architectures. In one aspect, a method includes: generating, using a controller policy, a batch of one or more output sequences, each output sequence in the batch defining a respective architecture of a child neural network and a respective architecture of a hardware accelerator; for each output sequence in the batch: training a respective instance of the child neural network having the architecture defined by the output sequence; evaluating a network performance of the trained instance of the child neural; and evaluating an accelerator performance of a respective instance of the hardware accelerator having the architecture defined by the output sequence to determine an accelerator performance metric for the instance of the hardware accelerator; and using the network performance metrics and the accelerator performance metrics to adjust the controller policy.

    PREDICTING NEURAL NETWORK PERFORMANCE USING NEURAL NETWORK GAUSSIAN PROCESS

    公开(公告)号:US20220019856A1

    公开(公告)日:2022-01-20

    申请号:US17377142

    申请日:2021-07-15

    Applicant: Google LLC

    Abstract: A method for predicting performance of a neural network (NN) is described. The method includes receiving a training data set having a set of training samples; receiving a validation data set having a set of validation pairs; initializing (i) a validation-training kernel matrix representing similarities of the validation inputs in the validation data set and the training inputs in the training data set and (ii) a training-training kernel matrix representing similarities across the training inputs within the training data set; generating a final updated validation-training kernel matrix and a final updated training-training kernel matrix; performing the following operations at least once: generating predicted validation outputs for the validation inputs, and updating an accuracy score of the NN based on the predicted validation outputs and the validation outputs; and outputting the updated accuracy score as a final accuracy score representing performance of the NN.

Patent Agency Ranking