INCORPORATION OF DECISION TREES IN A NEURAL NETWORK

    公开(公告)号:US20240220867A1

    公开(公告)日:2024-07-04

    申请号:US18289173

    申请日:2021-05-10

    Applicant: Google LLC

    CPC classification number: G06N20/20

    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for scheduling operations represented on a computation graph. One of the methods comprises receiving data representing a neural network comprising a plurality of layers arranged in a sequence; selecting one or more groups of layers each comprising one or more layers adjacent to each other in the sequence; generating a new machine learning model, comprising: for each group of layers, a respective decision tree that replaces the group of layers, wherein the respective decision tree receives as input a quantized version of the inputs to a respective first layer in the group and generates as output a quantized version of the outputs of a respective last layer in the group, wherein a tree depth of the respective decision tree is based at least in part on a number of layers of the group.

    Automatic Selection of Quantization and Filter Pruning Optimization Under Energy Constraints

    公开(公告)号:US20230229895A1

    公开(公告)日:2023-07-20

    申请号:US18007871

    申请日:2021-06-02

    Applicant: Google LLC

    CPC classification number: G06N3/0495 G06N3/092

    Abstract: Systems and methods for producing a neural network architecture with improved energy consumption and performance tradeoffs are disclosed, such as would be deployed for use on mobile or other resource-constrained devices. In particular, the present disclosure provides systems and methods for searching a network search space for joint optimization of a size of a layer of a reference neural network model (e.g., the number of filters in a convolutional layer or the number of output units in a dense layer) and of the quantization of values within the layer. By defining the search space to correspond to the architecture of a reference neural network model, examples of the disclosed network architecture search can optimize models of arbitrary complexity. The resulting neural network models are able to be run using relatively fewer computing resources (e.g., less processing power, less memory usage, less power consumption, etc.), all while remaining competitive with or even exceeding the performance (e.g., accuracy) of current state-of-the-art, mobile-optimized models.

    GENERATING AND GLOBALLY TUNING APPLICATION-SPECIFIC MACHINE LEARNING ACCELERATORS

    公开(公告)号:US20240232594A1

    公开(公告)日:2024-07-11

    申请号:US18289292

    申请日:2021-05-03

    Applicant: Google LLC

    CPC classification number: G06N3/063 G06N3/0464 G06N3/082

    Abstract: Methods, systems, and apparatus, including computer-readable media, are described for globally tuning and generating ML hardware accelerators. A design system selects an architecture representing a baseline processor configuration. An ML cost model of the system generates performance data about the architecture at least by modeling how the architecture executes computations of a neural network that includes multiple layers. Based on the performance data, the architecture is dynamically tuned to satisfy a performance objective when the architecture implements the neural network and executes machine-learning computations for a target application. In response to dynamically tuning the architecture, the system generates a configuration of an ML accelerator that specifies customized hardware configurations for implementing each of the multiple layers of the neural network.

Patent Agency Ranking