Techniques and Models for Multilingual Text Rewriting

    公开(公告)号:US20230274100A1

    公开(公告)日:2023-08-31

    申请号:US17682282

    申请日:2022-02-28

    Applicant: Google LLC

    CPC classification number: G06F40/58 G06F40/197 G06F40/166 G06F40/253 G06N3/08

    Abstract: The technology provides a model-based approach for multilingual text rewriting that is applicable across many languages and across different styles including formality levels or other textual attributes. The model is configured to manipulate both language and textual attributes jointly. This approach supports zero-shot formality-sensitive translation, with no labeled data in the target language. An encoder-decoder architectural approach with attribute extraction is used to train rewriter models that can thus be used in “universal” textual rewriting across many different languages. A cross-lingual learning signal can be incorporated into the training approach. Certain training processes do not employ any exemplars. This approach enables not just straight translation, but also the ability to create new sentences with different attributes.

    Parameter Efficient Prompt Tuning for Efficient Models at Scale

    公开(公告)号:US20230325725A1

    公开(公告)日:2023-10-12

    申请号:US17718738

    申请日:2022-04-12

    Applicant: Google LLC

    CPC classification number: G06N20/20 G06V10/764 G06V10/7747

    Abstract: Systems and methods for natural language processing can leverage trained prompts to condition a large pre-trained machine-learned model to generate an output for a specific task. For example, a subset of parameters may be trained for the particular task to then be input with a set of input data into the pre-trained machine-learned model to generate the task-specific output. During the training of the prompt, the parameters of the pre-trained machine-learned model can be frozen, which can reduce the computational resources used during training while still leveraging the previously learned data from the pre-trained machine-learned model.

    Techniques and models for multilingual text rewriting

    公开(公告)号:US12210848B2

    公开(公告)日:2025-01-28

    申请号:US17682282

    申请日:2022-02-28

    Applicant: Google LLC

    Abstract: The technology provides a model-based approach for multilingual text rewriting that is applicable across many languages and across different styles including formality levels or other textual attributes. The model is configured to manipulate both language and textual attributes jointly. This approach supports zero-shot formality-sensitive translation, with no labeled data in the target language. An encoder-decoder architectural approach with attribute extraction is used to train rewriter models that can thus be used in “universal” textual rewriting across many different languages. A cross-lingual learning signal can be incorporated into the training approach. Certain training processes do not employ any exemplars. This approach enables not just straight translation, but also the ability to create new sentences with different attributes.

Patent Agency Ranking