-
公开(公告)号:US12033073B2
公开(公告)日:2024-07-09
申请号:US17156464
申请日:2021-01-22
Applicant: Google LLC
Inventor: Sergey Ioffe , Corinna Cortes
CPC classification number: G06N3/08 , G06F18/10 , G06F18/2415 , G06N3/04 , G06N3/084 , G06V10/70 , G06V10/82 , G06T2207/20081
Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for processing inputs using a neural network system that includes a batch normalization layer. One of the methods includes receiving a respective first layer output for each training example in the batch; computing a plurality of normalization statistics for the batch from the first layer outputs; normalizing each component of each first layer output using the normalization statistics to generate a respective normalized layer output for each training example in the batch; generating a respective batch normalization layer output for each of the training examples from the normalized layer outputs; and providing the batch normalization layer output as an input to the second neural network layer.
-
公开(公告)号:US11887004B2
公开(公告)日:2024-01-30
申请号:US16854352
申请日:2020-04-21
Applicant: Google LLC
Inventor: Sergey Ioffe
Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for implementing a neural network. In one aspect, the neural network includes a batch renormalization layer between a first neural network layer and a second neural network layer. The first neural network layer generates first layer outputs having multiple components. The batch renormalization layer is configured to, during training of the neural network on a current batch of training examples, obtain respective current moving normalization statistics for each of the multiple components and determine respective affine transform parameters for each of the multiple components from the current moving normalization statistics. The batch renormalization layer receives a respective first layer output for each training example in the current batch and applies the affine transform to each component of a normalized layer output to generate a renormalized layer output for the training example.
-
公开(公告)号:US11853885B2
公开(公告)日:2023-12-26
申请号:US17723007
申请日:2022-04-18
Applicant: Google LLC
Inventor: Sergey Ioffe , Corinna Cortes
CPC classification number: G06N3/08 , G06F18/10 , G06F18/2415 , G06N3/04 , G06N3/084 , G06V10/70 , G06V10/82 , G06T2207/20081
Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for processing images or features of images using an image classification system that includes a batch normalization layer. One of the systems includes a convolutional neural network configured to receive an input comprising an image or image features of the image and to generate a network output that includes respective scores for each object category in a set of object categories, the score for each object category representing a likelihood that that the image contains an image of an object belonging to the category, and the convolutional neural network comprising: a plurality of neural network layers, the plurality of neural network layers comprising a first convolutional neural network layer and a second neural network layer; and a batch normalization layer between the first convolutional neural network layer and the second neural network layer.
-
公开(公告)号:US11281973B2
公开(公告)日:2022-03-22
申请号:US17390768
申请日:2021-07-30
Applicant: Google LLC
Inventor: Sergey Ioffe , Corinna Cortes
Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for processing inputs using a neural network system that includes a batch normalization layer. One of the methods includes receiving a respective first layer output for each training example in the batch; computing a plurality of normalization statistics for the batch from the first layer outputs; normalizing each component of each first layer output using the normalization statistics to generate a respective normalized layer output for each training example in the batch; generating a respective batch normalization layer output for each of the training examples from the normalized layer outputs; and providing the batch normalization layer output as an input to the second neural network layer.
-
公开(公告)号:US20210334605A1
公开(公告)日:2021-10-28
申请号:US17372090
申请日:2021-07-09
Applicant: Google LLC
Inventor: Vincent O. Vanhoucke , Christian Szegedy , Sergey Ioffe
Abstract: A neural network system that includes: multiple subnetworks that includes: a first subnetwork including multiple first modules, each first module including: a pass-through convolutional layer configured to process the subnetwork input for the first subnetwork to generate a pass-through output; an average pooling stack of neural network layers that collectively processes the subnetwork input for the first subnetwork to generate an average pooling output; a first stack of convolutional neural network layers configured to collectively process the subnetwork input for the first subnetwork to generate a first stack output; a second stack of convolutional neural network layers that are configured to collectively process the subnetwork input for the first subnetwork to generate a second stack output; and a concatenation layer configured to concatenate the pass-through output, the average pooling output, the first stack output, and the second stack output to generate a first module output for the first module.
-
公开(公告)号:US20210216870A1
公开(公告)日:2021-07-15
申请号:US17156464
申请日:2021-01-22
Applicant: Google LLC
Inventor: Sergey Ioffe , Corinna Cortes
Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for processing inputs using a neural network system that includes a batch normalization layer. One of the methods includes receiving a respective first layer output for each training example in the batch; computing a plurality of normalization statistics for the batch from the first layer outputs; normalizing each component of each first layer output using the normalization statistics to generate a respective normalized layer output for each training example in the batch; generating a respective batch normalization layer output for each of the training examples from the normalized layer outputs; and providing the batch normalization layer output as an input to the second neural network layer.
-
公开(公告)号:US10685278B1
公开(公告)日:2020-06-16
申请号:US16730855
申请日:2019-12-30
Applicant: Google LLC
Inventor: Sergey Ioffe , Raymond Wensley Smith
Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for implementing long-short term memory cells with saturating gating functions. One of the systems includes a first Long Short-Term Memory (LSTM) cell, wherein the first LSTM cell is configured to, for each of the plurality of time steps, generate a new cell state and a new cell output by applying a plurality of gates to a current cell input, a current cell state, and a current cell output, each of the plurality of gates being configured to, for each of the plurality of time steps: receive a gate input vector, generate a respective intermediate gate output vector from the gate input, and apply a respective gating function to each component of the respective intermediate gate output vector, wherein the respective gating function for at least one of the plurality of gates is a saturating gating function.
-
公开(公告)号:US10671922B2
公开(公告)日:2020-06-02
申请号:US16459057
申请日:2019-07-01
Applicant: Google LLC
Inventor: Sergey Ioffe
Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for implementing a neural network. In one aspect, the neural network includes a batch renormalization layer between a first neural network layer and a second neural network layer. The first neural network layer generates first layer outputs having multiple components. The batch renormalization layer is configured to, during training of the neural network on a current batch of training examples, obtain respective current moving normalization statistics for each of the multiple components and determine respective affine transform parameters for each of the multiple components from the current moving normalization statistics. The batch renormalization layer receives a respective first layer output for each training example in the current batch and applies the affine transform to each component of a normalized layer output to generate a renormalized layer output for the training example.
-
公开(公告)号:US20190325315A1
公开(公告)日:2019-10-24
申请号:US16459057
申请日:2019-07-01
Applicant: Google LLC
Inventor: Sergey Ioffe
Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for implementing a neural network. In one aspect, the neural network includes a batch renormalization layer between a first neural network layer and a second neural network layer. The first neural network layer generates first layer outputs having multiple components. The batch renormalization layer is configured to, during training of the neural network on a current batch of training examples, obtain respective current moving normalization statistics for each of the multiple components and determine respective affine transform parameters for each of the multiple components from the current moving normalization statistics. The batch renormalization layer receives a respective first layer output for each training example in the current batch and applies the affine transform to each component of a normalized layer output to generate a renormalized layer output for the training example.
-
公开(公告)号:US10229326B2
公开(公告)日:2019-03-12
申请号:US16125045
申请日:2018-09-07
Applicant: Google LLC
Inventor: Matthias Grundmann , Alexandra Ivanna Hawkins , Sergey Ioffe
Abstract: Methods, systems, and media for summarizing a video with video thumbnails are provided. In some embodiments, the method comprises: receiving a plurality of video frames corresponding to the video and associated information associated with each of the plurality of video frames; extracting, for each of the plurality of video frames, a plurality of features; generating candidate clips that each includes at least a portion of the received video frames based on the extracted plurality of features and the associated information; calculating, for each candidate clip, a clip score based on the extracted plurality of features from the video frames associated with the candidate clip; calculating, between adjacent candidate clips, a transition score based at least in part on a comparison of video frame features between frames from the adjacent candidate clips; selecting a subset of the candidate clips based at least in part on the clip score and the transition score associated with each of the candidate clips; and automatically generating an animated video thumbnail corresponding to the video that includes a plurality of video frames selected from each of the subset of candidate clips.
-
-
-
-
-
-
-
-
-