SPEECH CODING USING AUTO-REGRESSIVE GENERATIVE NEURAL NETWORKS

    公开(公告)号:US20210366495A1

    公开(公告)日:2021-11-25

    申请号:US17332898

    申请日:2021-05-27

    Applicant: Google LLC

    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for coding speech using neural networks. One of the methods includes obtaining a bitstream of parametric coder parameters characterizing spoken speech; generating, from the parametric coder parameters, a conditioning sequence; generating a reconstruction of the spoken speech that includes a respective speech sample at each of a plurality of decoder time steps, comprising, at each decoder time step: processing a current reconstruction sequence using an auto-regressive generative neural network, wherein the auto-regressive generative neural network is configured to process the current reconstruction to compute a score distribution over possible speech sample values, and wherein the processing comprises conditioning the auto-regressive generative neural network on at least a portion of the conditioning sequence; and sampling a speech sample from the possible speech sample values.

    Speech coding using auto-regressive generative neural networks

    公开(公告)号:US12062380B2

    公开(公告)日:2024-08-13

    申请号:US18144413

    申请日:2023-05-08

    Applicant: Google LLC

    CPC classification number: G10L19/0204 G10L25/30

    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for coding speech using neural networks. One of the methods includes obtaining a bitstream of parametric coder parameters characterizing spoken speech; generating, from the parametric coder parameters, a conditioning sequence; generating a reconstruction of the spoken speech that includes a respective speech sample at each of a plurality of decoder time steps, comprising, at each decoder time step: processing a current reconstruction sequence using an auto-regressive generative neural network, wherein the auto-regressive generative neural network is configured to process the current reconstruction to compute a score distribution over possible speech sample values, and wherein the processing comprises conditioning the auto-regressive generative neural network on at least a portion of the conditioning sequence; and sampling a speech sample from the possible speech sample values.

    Speech coding using auto-regressive generative neural networks

    公开(公告)号:US11676613B2

    公开(公告)日:2023-06-13

    申请号:US17332898

    申请日:2021-05-27

    Applicant: Google LLC

    CPC classification number: G10L19/0204 G10L25/30

    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for coding speech using neural networks. One of the methods includes obtaining a bitstream of parametric coder parameters characterizing spoken speech; generating, from the parametric coder parameters, a conditioning sequence; generating a reconstruction of the spoken speech that includes a respective speech sample at each of a plurality of decoder time steps, comprising, at each decoder time step: processing a current reconstruction sequence using an auto-regressive generative neural network, wherein the auto-regressive generative neural network is configured to process the current reconstruction to compute a score distribution over possible speech sample values, and wherein the processing comprises conditioning the auto-regressive generative neural network on at least a portion of the conditioning sequence; and sampling a speech sample from the possible speech sample values.

    Speech coding using auto-regressive generative neural networks

    公开(公告)号:US11024321B2

    公开(公告)日:2021-06-01

    申请号:US16206823

    申请日:2018-11-30

    Applicant: Google LLC

    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for coding speech using neural networks. One of the methods includes obtaining a bitstream of parametric coder parameters characterizing spoken speech; generating, from the parametric coder parameters, a conditioning sequence; generating a reconstruction of the spoken speech that includes a respective speech sample at each of a plurality of decoder time steps, comprising, at each decoder time step: processing a current reconstruction sequence using an auto-regressive generative neural network, wherein the auto-regressive generative neural network is configured to process the current reconstruction to compute a score distribution over possible speech sample values, and wherein the processing comprises conditioning the auto-regressive generative neural network on at least a portion of the conditioning sequence; and sampling a speech sample from the possible speech sample values.

    SPEECH CODING USING AUTO-REGRESSIVE GENERATIVE NEURAL NETWORKS

    公开(公告)号:US20200176004A1

    公开(公告)日:2020-06-04

    申请号:US16206823

    申请日:2018-11-30

    Applicant: Google LLC

    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for coding speech using neural networks. One of the methods includes obtaining a bitstream of parametric coder parameters characterizing spoken speech; generating, from the parametric coder parameters, a conditioning sequence; generating a reconstruction of the spoken speech that includes a respective speech sample at each of a plurality of decoder time steps, comprising, at each decoder time step: processing a current reconstruction sequence using an auto-regressive generative neural network, wherein the auto-regressive generative neural network is configured to process the current reconstruction to compute a score distribution over possible speech sample values, and wherein the processing comprises conditioning the auto-regressive generative neural network on at least a portion of the conditioning sequence; and sampling a speech sample from the possible speech sample values.

    SPECIFYING LOUDNESS IN AN IMMERSIVE AUDIO PACKAGE

    公开(公告)号:US20240329915A1

    公开(公告)日:2024-10-03

    申请号:US18353037

    申请日:2023-07-14

    Applicant: GOOGLE LLC

    CPC classification number: G06F3/165 G06F3/162

    Abstract: A method including generating an audio stream including a first substream as first audio data and a second substream as second audio data, generating a first loudness parameter associated with playback of the first substream, generating a second loudness parameter associated with playback of the second substream, and generating an audio package including an identification corresponding to the first audio data, an identification corresponding to the second audio data, and a codec agnostic container including the first loudness parameter, and the second loudness parameter.

    Identifying salient features for generative networks

    公开(公告)号:US12242567B2

    公开(公告)日:2025-03-04

    申请号:US17250506

    申请日:2019-05-16

    Applicant: Google LLC

    Abstract: Implementations identify a small set of independent, salient features from an input signal. The salient features may be used for conditioning a generative network, making the generative network robust to noise. The salient features may facilitate compression and data transmission. An example method includes receiving an input signal and extracting salient features for the input signal by providing the input signal to an encoder trained to extract salient features. The salient features may be independent and have a sparse distribution. The encoder may be configured to generate almost identical features from two input signals a system designer deems equivalent. The method also includes conditioning a generative network using the salient features. In some implementations, the method may also include extracting a plurality of time sequences from the input signal and extracting the salient features for each time sequence.

    IMMERSIVE AUDIO PACKAGE
    10.
    发明公开

    公开(公告)号:US20240331709A1

    公开(公告)日:2024-10-03

    申请号:US18355928

    申请日:2023-07-20

    Applicant: GOOGLE LLC

    CPC classification number: G10L19/02

    Abstract: A method including receiving first audio data, receiving second audio data, compressing the first audio data as first compressed audio data, compressing the second audio data as second compressed audio data, generating a codec dependent container including a parameter associated with compressing the first audio data, compressing the second audio data, a reference to the first compressed audio data, and a reference to the second compressed audio data, generating a codec agnostic container including at least one parameter representing time-varying data associated with playback of the first audio data and the second audio data, and generating an audio package including the codec dependent container and the codec agnostic container.

Patent Agency Ranking