Parameter-efficient multi-task and transfer learning

    公开(公告)号:US11676008B2

    公开(公告)日:2023-06-13

    申请号:US16577698

    申请日:2019-09-20

    Applicant: Google LLC

    CPC classification number: G06N3/08 G06N3/045 G10L15/16

    Abstract: The present disclosure provides systems and methods that enable parameter-efficient transfer learning, multi-task learning, and/or other forms of model re-purposing such as model personalization or domain adaptation. In particular, as one example, a computing system can obtain a machine-learned model that has been previously trained on a first training dataset to perform a first task. The machine-learned model can include a first set of learnable parameters. The computing system can modify the machine-learned model to include a model patch, where the model patch includes a second set of learnable parameters. The computing system can train the machine-learned model on a second training dataset to perform a second task that is different from the first task, which may include learning new values for the second set of learnable parameters included in the model patch while keeping at least some (e.g., all) of the first set of parameters fixed.

    Parameter-efficient multi-task and transfer learning

    公开(公告)号:US12169779B2

    公开(公告)日:2024-12-17

    申请号:US18310638

    申请日:2023-05-02

    Applicant: Google LLC

    Abstract: The present disclosure provides systems and methods that enable parameter-efficient transfer learning, multi-task learning, and/or other forms of model re-purposing such as model personalization or domain adaptation. In particular, as one example, a computing system can obtain a machine-learned model that has been previously trained on a first training dataset to perform a first task. The machine-learned model can include a first set of learnable parameters. The computing system can modify the machine-learned model to include a model patch, where the model patch includes a second set of learnable parameters. The computing system can train the machine-learned model on a second training dataset to perform a second task that is different from the first task, which may include learning new values for the second set of learnable parameters included in the model patch while keeping at least some (e.g., all) of the first set of parameters fixed.

    Highly efficient convolutional neural networks

    公开(公告)号:US11734545B2

    公开(公告)日:2023-08-22

    申请号:US15898566

    申请日:2018-02-17

    Applicant: Google LLC

    CPC classification number: G06N3/04 G06N3/045 G06N3/08 G06N3/048

    Abstract: The present disclosure provides directed to new, more efficient neural network architectures. As one example, in some implementations, the neural network architectures of the present disclosure can include a linear bottleneck layer positioned structurally prior to and/or after one or more convolutional layers, such as, for example, one or more depthwise separable convolutional layers. As another example, in some implementations, the neural network architectures of the present disclosure can include one or more inverted residual blocks where the input and output of the inverted residual block are thin bottleneck layers, while an intermediate layer is an expanded representation. For example, the expanded representation can include one or more convolutional layers, such as, for example, one or more depthwise separable convolutional layers. A residual shortcut connection can exist between the thin bottleneck layers that play a role of an input and output of the inverted residual block.

    Highly efficient convolutional neural networks

    公开(公告)号:US11823024B2

    公开(公告)日:2023-11-21

    申请号:US17382503

    申请日:2021-07-22

    Applicant: Google LLC

    CPC classification number: G06N3/04 G06N3/045 G06N3/08 G06N3/048

    Abstract: The present disclosure provides directed to new, more efficient neural network architectures. As one example, in some implementations, the neural network architectures of the present disclosure can include a linear bottleneck layer positioned structurally prior to and/or after one or more convolutional layers, such as, for example, one or more depthwise separable convolutional layers. As another example, in some implementations, the neural network architectures of the present disclosure can include one or more inverted residual blocks where the input and output of the inverted residual block are thin bottleneck layers, while an intermediate layer is an expanded representation. For example, the expanded representation can include one or more convolutional layers, such as, for example, one or more depthwise separable convolutional layers. A residual shortcut connection can exist between the thin bottleneck layers that play a role of an input and output of the inverted residual block.

    Parameter-Efficient Multi-Task and Transfer Learning

    公开(公告)号:US20230267330A1

    公开(公告)日:2023-08-24

    申请号:US18310638

    申请日:2023-05-02

    Applicant: Google LLC

    CPC classification number: G06N3/08 G06N3/045 G10L15/16

    Abstract: The present disclosure provides systems and methods that enable parameter-efficient transfer learning, multi-task learning, and/or other forms of model re-purposing such as model personalization or domain adaptation. In particular, as one example, a computing system can obtain a machine-learned model that has been previously trained on a first training dataset to perform a first task. The machine-learned model can include a first set of learnable parameters. The computing system can modify the machine-learned model to include a model patch, where the model patch includes a second set of learnable parameters. The computing system can train the machine-learned model on a second training dataset to perform a second task that is different from the first task, which may include learning new values for the second set of learnable parameters included in the model patch while keeping at least some (e.g., all) of the first set of parameters fixed.

    Convolutional Neural Network Compression
    8.
    发明申请

    公开(公告)号:US20190279092A1

    公开(公告)日:2019-09-12

    申请号:US16346313

    申请日:2017-09-29

    Applicant: Google LLC

    Abstract: Systems and methods of convolutional neural network compression are provided. For instance, a convolutional neural network can include an input convolutional layer having a plurality of associated input filters and an output convolutional layer having a plurality of associated output filters. The convolutional neural network implements a connection pattern defining connections between the plurality of input filters and the plurality of output filers. The connection pattern specifies that at least one output filter of the plurality of output filters is connected to only a subset of the plurality of input filters.

    Highly Efficient Convolutional Neural Networks

    公开(公告)号:US20190147318A1

    公开(公告)日:2019-05-16

    申请号:US15898566

    申请日:2018-02-17

    Applicant: Google LLC

    Abstract: The present disclosure provides directed to new, more efficient neural network architectures. As one example, in some implementations, the neural network architectures of the present disclosure can include a linear bottleneck layer positioned structurally prior to and/or after one or more convolutional layers, such as, for example, one or more depthwise separable convolutional layers. As another example, in some implementations, the neural network architectures of the present disclosure can include one or more inverted residual blocks where the input and output of the inverted residual block are thin bottleneck layers, while an intermediate layer is an expanded representation. For example, the expanded representation can include one or more convolutional layers, such as, for example, one or more depthwise separable convolutional layers. A residual shortcut connection can exist between the thin bottleneck layers that play a role of an input and output of the inverted residual block.

    Parameter-Efficient Multi-Task and Transfer Learning

    公开(公告)号:US20250061333A1

    公开(公告)日:2025-02-20

    申请号:US18939318

    申请日:2024-11-06

    Applicant: Google LLC

    Abstract: The present disclosure provides systems and methods that enable parameter-efficient transfer learning, multi-task learning, and/or other forms of model re-purposing such as model personalization or domain adaptation. In particular, as one example, a computing system can obtain a machine-learned model that has been previously trained on a first training dataset to perform a first task. The machine-learned model can include a first set of learnable parameters. The computing system can modify the machine-learned model to include a model patch, where the model patch includes a second set of learnable parameters. The computing system can train the machine-learned model on a second training dataset to perform a second task that is different from the first task, which may include learning new values for the second set of learnable parameters included in the model patch while keeping at least some (e.g., all) of the first set of parameters fixed.

Patent Agency Ranking