GRANULAR NEURAL NETWORK ARCHITECTURE SEARCH OVER LOW-LEVEL PRIMITIVES

    公开(公告)号:WO2022251719A1

    公开(公告)日:2022-12-01

    申请号:PCT/US2022/031468

    申请日:2022-05-27

    Applicant: GOOGLE LLC

    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for determining neural network architectures. One of the methods includes receiving initial neural network architecture data; generating, from the initial neural network architecture data, search space data defining a plurality of sub-model architectures, each sub-model architecture comprising an ordered set of primitive neural network operations each associated with one or more operation parameters; and determining a final architecture of a neural network for performing a machine learning task comprising running an evolutionary architecture search algorithm over the search space data to identify a respective optimized value for each of the one or more operation parameters of the primitive neural network operations in at least one of the plurality of sub-model architectures.

    SYSTEMS AND METHODS FOR MACHINE-LEARNED MODELS HAVING CONVOLUTION AND ATTENTION

    公开(公告)号:WO2022251602A1

    公开(公告)日:2022-12-01

    申请号:PCT/US2022/031304

    申请日:2022-05-27

    Applicant: GOOGLE LLC

    Abstract: A computer-implemented method for performing computer vision with reduced computational cost and improved accuracy can include obtaining, by a computing system including one or more computing devices, input data comprising an input tensor having one or more dimensions, providing, by the computing system, the input data to a machine-learned convolutional attention network, the machine-learned convolutional attention network including two or more network stages, and, in response to providing the input data to the machine-learned convolutional attention network, receiving, by the computing system, a machine-learning prediction from the machine-learned convolutional attention network. The convolutional attention network can include at least one attention block, wherein the attention block includes a relative attention mechanism, the relative attention mechanism including the sum of a static convolution kernel with an adaptive attention matrix. This provides for improved generalization, capacity, and efficiency of the convolutional attention network relative to some existing models.

Patent Agency Ranking