-
公开(公告)号:US20250156871A1
公开(公告)日:2025-05-15
申请号:US18507044
申请日:2023-11-11
Inventor: Yada Zhu , Xiaojie Guo , David Cox , Shuaicheng Zhang
IPC: G06Q20/40
Abstract: A computer-implemented method for task-guided graph augmentation and editing includes receiving an input graph in an observed financial transaction network. A data augmentation function is learned, where the data augmentation function maintains a true data distribution of the input graph. An augmented financial transaction network is generated that enhances performance of a downstream task and preserves topological and temporal properties of the input graph.
-
2.
公开(公告)号:US20240185270A1
公开(公告)日:2024-06-06
申请号:US17972167
申请日:2022-10-24
Applicant: International Business Machines Corporation
Inventor: YADA ZHU , Zixuan Yuan , David Cox , Anna Wanda Topol
IPC: G06Q30/02 , G06F40/126 , G06F40/166 , G06F40/30
CPC classification number: G06Q30/0202 , G06F40/126 , G06F40/166 , G06F40/30
Abstract: Unsupervised cross-domain data augmentation techniques for long-text document based prediction and explanation are provided. In one aspect, a system for long-document based prediction includes: an encoder for creating embeddings of long-document texts with hierarchical sparse self-attention, and making predictions using the embeddings of the long-document texts; and a multi-source counterfactual augmentation module for generating perturbed long-document texts using unlabeled sentences from at least one external source to train the encoder. A method for long-document based prediction is also provided.
-
公开(公告)号:US20220392429A1
公开(公告)日:2022-12-08
申请号:US17337518
申请日:2021-06-03
Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
Inventor: Kaizhi Qian , Yang Zhang , Shiyu Chang , Jinjun Xiong , Chuang Gan , David Cox
IPC: G10L13/10 , G06N20/00 , G10L21/013 , G10L17/04 , G10L25/63
Abstract: A computer-implemented method is provided of using a machine learning model for disentanglement of prosody in spoken natural language. The method includes encoding, by a computing device, the spoken natural language to produce content code. The method further includes resampling, by the computing device without text transcriptions, the content code to obscure the prosody by applying an unsupervised technique to the machine learning model to generate prosody-obscured content code. The method additionally includes decoding, by the computing device, the prosody-obscured content code to synthesize speech indirectly based upon the content code.
-
公开(公告)号:US11295762B2
公开(公告)日:2022-04-05
申请号:US16852617
申请日:2020-04-20
Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
Inventor: Kaizhi Qian , Yang Zhang , Shiyu Chang , Chuang Gan , David Cox
Abstract: A method, a structure, and a computer system for decomposing speech. The exemplary embodiments may include one or more encoders for generating one or more encodings of a speech input comprising rhythm information, pitch information, timbre information, and content information, and a decoder for decoding the one or more encodings.
-
公开(公告)号:US20210327460A1
公开(公告)日:2021-10-21
申请号:US16852617
申请日:2020-04-20
Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
Inventor: Kaizhi Qian , Yang Zhang , Shiyu Chang , Chuang Gan , David Cox
IPC: G10L25/90
Abstract: A method, a structure, and a computer system for decomposing speech. The exemplary embodiments may include one or more encoders for generating one or more encodings of a speech input comprising rhythm information, pitch information, timbre information, and content information, and a decoder for decoding the one or more encodings.
-
公开(公告)号:US20250103875A1
公开(公告)日:2025-03-27
申请号:US18372661
申请日:2023-09-25
Inventor: Rameswar Panda , Peihao Wang , LEONID KARLINSKY , Rogerio Schmidt Feris , David Cox , Yoon Hyung Kim
IPC: G06N3/08 , G06N3/0455
Abstract: Parameters of a first transformer are accessed, and size dimensions of a second transformer that is to be trained and is larger than the first transformer are received. The parameters of the first transformer are linearly transformed using a combination of a width-growth operator and a depth-growth operator, wherein the linear transformation produces a set of new parameters, the set corresponding to the size dimensions of the second transformer. The second transformer is initialized with the set of new parameters.
-
公开(公告)号:US11996083B2
公开(公告)日:2024-05-28
申请号:US17337518
申请日:2021-06-03
Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
Inventor: Kaizhi Qian , Yang Zhang , Shiyu Chang , Jinjun Xiong , Chuang Gan , David Cox
IPC: G10L13/10 , G06N20/00 , G10L17/04 , G10L21/013 , G10L25/63
CPC classification number: G10L13/10 , G06N20/00 , G10L17/04 , G10L21/013 , G10L25/63
Abstract: A computer-implemented method is provided of using a machine learning model for disentanglement of prosody in spoken natural language. The method includes encoding, by a computing device, the spoken natural language to produce content code. The method further includes resampling, by the computing device without text transcriptions, the content code to obscure the prosody by applying an unsupervised technique to the machine learning model to generate prosody-obscured content code. The method additionally includes decoding, by the computing device, the prosody-obscured content code to synthesize speech indirectly based upon the content code.
-
8.
公开(公告)号:US20240153007A1
公开(公告)日:2024-05-09
申请号:US17978486
申请日:2022-11-01
Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
Inventor: Yada Zhu , Yang Zhang , Tian Gao , David Cox
IPC: G06Q40/06
CPC classification number: G06Q40/06
Abstract: A method includes: creating a training data set based on user input, the training data set including time series data of a price of an asset and stochastic event data of events related to the asset; creating an event intensity model that models an event intensity parameter of one of the events related to the asset, wherein the event intensity model comprises a proximal graphical event model (PGEM), and the creating the event intensity model includes learning parameters of the PGEM using machine learning and the training data set; creating a probabilistic time series model that predicts a probability distribution of a return of the asset, wherein the creating the probabilistic time series model includes learning parameters of the probabilistic time series model using machine learning and the training data set; and predicting a future return of the asset for a future time period using the probabilistic time series model.
-
公开(公告)号:US20210133539A1
公开(公告)日:2021-05-06
申请号:US16672996
申请日:2019-11-04
Applicant: International Business Machines Corporation
Inventor: Akash Srivastava , Jessie Carrigan Rosenberg , Dan Gutfreund , David Cox
Abstract: A generator network of a variational autoencoder can be trained to approximate a simulator and generate a first result. The simulator is associated with input data, based on which the simulator outputs output data. A training data set for the generator network can include the simulator's input data and output data. Based on the simulator's output data and the first result of the generator network, an inference network of the variational autoencoder can be trained to generate a second result. The second result of the trained inference network inverts the first result of the generator and approximates the simulator's input data. The trained inference network can function as an inverted simulator.
-
公开(公告)号:US20240127005A1
公开(公告)日:2024-04-18
申请号:US17954845
申请日:2022-09-28
Inventor: Rameswar Panda , Yi Li , Richard Chen , Rogerio Schmidt Feris , Yoon Hyung Kim , David Cox
IPC: G06F40/58 , G06F40/284 , G06N3/08
CPC classification number: G06F40/58 , G06F40/284 , G06N3/08
Abstract: Methods, systems, and computer program products for translating text using generated visual representations and artificial intelligence are provided herein. A computer-implemented method includes generating a tokenized form of at least a portion of input text in a first language; generating at least one visual representation of at least a portion of the input text using a first set of artificial intelligence techniques; generating a tokenized form of at least a portion of the at least one visual representation; and generating an output including a translated version of the input text into at least a second language by processing, using a second set of artificial intelligence techniques, at least a portion of the tokenized form of the at least a portion of the input text and at least a portion of the tokenized form of the at least a portion of the at least one visual representation.
-
-
-
-
-
-
-
-
-