Abstract:
Systems and methods for Deep Learning techniques based multi-purpose conversational agents for processing natural language queries. The traditional systems and methods provide for conversational systems for processing natural language queries but do not employ Deep Learning techniques, and thus are unable to process large number of intents. Embodiments of the present disclosure provide for Deep Learning techniques based multi-purpose conversational agents for processing the natural language queries by defining and logically integrating a plurality of components comprising of multi-purpose conversational agents, identifying an appropriate agent to process one or more natural language queries by a High Level Intent Identification technique, predicting a probable user intent, classifying the query, and generate a set of responses by querying or updating one or more knowledge graphs.
Abstract:
In automated assistant systems, a deep-learning model in form of a long short-term memory (LSTM) classifier is used for mapping questions to classes, with each class having a manually curated answer. A team of experts manually create the training data used to train this classifier. Relying on human curation often results in such linguistic training biases creeping into training data, since every individual has a specific style of writing natural language and uses some words in specific context only. Deep models end up learning these biases, instead of the core concept words of the target classes. In order to correct these biases, meaningful sentences are automatically generated using a generative model, and then used for training a classification model. For example, a variational autoencoder (VAE) is used as the generative model for generating novel sentences and a language model (LM) is utilized for selecting sentences based on likelihood.
Abstract:
The present subject matter discloses system and method for executing prescriptive analytics. Simulation is performed from an input data (xinput) and simulation parameters (μ) to generate simulating data (D). Further, forecast data may be predicted by processing the simulating data (D) using predictive model (M). Further, prescriptive value (x′) may be determined based on the forecast data by using optimization model. The prescriptive value (x′) may be determined such that an objective function associated with the optimization model is optimized, whereby the optimization of the objective function indicates business objective being achieved. Further, the steps of simulating, predicting and determining may be iteratively performed until the objective function is not further optimized, satisfying predefined condition. Further, at each iteration, except the first iteration, the input data (xinput) is the prescriptive value (x′) determined at immediate previous iteration, whereby at the first iteration, the input data (xinput) is a reference data.
Abstract:
This disclosure relates generally to data meta model and meta file generation for feature engineering and training of machine learning models thereof. Conventional methods do not facilitate appropriate relevant data identification for feature engineering and also do not implement standardization for use of solution across domains. Embodiments of the present disclosure provide systems and methods wherein datasets from various sources/domains are utilized for meta file generation that is based on mapping of the dataset with a data meta model based on the domains, the meta file comprises meta data and information pertaining to action(s) being performed. Further functions are generated using the meta file and the functions are assigned to corresponding data characterized in the meta file. Further functions are invoked to generate feature vector set and machine learning model(s) are trained using the features vector set. Implementation of the generated data meta-model enables re-using of feature engineering code.
Abstract:
Traditional systems and methods have implemented hand-crafted feature extraction from varying length time series that results in complexity and requires domain knowledge. Building classification models requires large labeled data and is computationally expensive. Embodiments of the present disclosure implement learning models for classification tasks in multi-dimensional time series by performing feature extraction from entity's parameters via unsupervised encoder and build a non-temporal linear classifier model. A fixed-dimensional feature vector is outputted using a pre-trained unsupervised encoder, which acts as off-the shelf feature extractor. Extracted features are concatenated to learn a non-temporal linear classification model and weight is assigned to each extracted feature during learning which helps to determine relevant parameters for each class. Mapping from parameters to target class is considered while constraining the linear model to use only subset of large number of features.
Abstract:
Estimating Remaining Useful Life (RUL) from multi-sensor time series data is difficult through manual inspection. Current machine learning and data analytics methods, for RUL estimation require large number of failed instances for training, which are rarely available in practice, and these methods cannot use information from currently operational censored instances since their failure time is unknown. Embodiments of the present disclosure provide systems and methods for estimating RUL using time series data by implementing an LSTM-RNN based ordinal regression technique, wherein during training RUL value of failed instance(s) is encoded into a vector which is given as a target to the model. Unlike a failed instance, the exact RUL for a censored instance is unknown. For using the censored instances, target vectors are generated and the objective function is modified for training wherein the trained LSTM-RNN based ordinal regression is applied on an input test time series for RUL estimation.
Abstract:
Organizations are constantly flooded with questions, ranging from mundane to the unanswerable. It is therefore respective department that actively looks for automated assistance, especially to alleviate the burden of routine, but time-consuming tasks. The embodiments of the present disclosure provide BiLSTM-Siamese Network based Classifier for identifying target class of queries and providing responses to queries pertaining to the identified target class, which acts as an automated assistant that alleviates burden of answering queries in well-defined domains. Siamese Model (SM) is trained for a epochs, and then the same Base-Network is used to train Classification Model (CM) for b epochs iteratively until best accuracy is observed on validation test, wherein SM ensures it learns which sentences are similar/dissimilar semantically while CM learns to predict target class of every user query. Here a and b are assumed to be hyper parameters and are tuned for best performance on the validation set.
Abstract:
Systems and methods for predictive reliability mining are provided that enable predicting of unexpected emerging failures in future without waiting for actual failures to start occurring in significant numbers. Sets of discriminative Diagnostic Trouble Codes (DTCs) from connected machines in a population are identified before failure of the associated parts. A temporal conditional dependence model based on the temporal dependence between the failure of the parts from past failure data and the identified sets of discriminative DTCs is generated. Future failures are predicted based on the generated temporal conditional dependence and root cause analysis of the predicted future failures is performed for predictive reliability mining. The probability of failure is computed based on both occurrence and non-occurrence of DTCs. The root cause analysis enables identifying a subset of the population when an early warning is generated and also when an early warning is not generated.
Abstract:
The present disclosure relates to business data processing and facilitates fusing business data spanning disparate sources for processing distributional queries for enterprise business intelligence application. Particularly, the method comprises defining a Bayesian network based on one or more attributes associated with raw data spanning a plurality of disparate sources; pre-processing the raw data based on the Bayesian network to compute conditional probabilities therein as parameters; joining the one or more attributes in the raw data using the conditional probabilities; and executing probabilistic inference from a database of the parameters by employing an SQL engine. The Bayesian Network may be validated based on estimation error computed by comparing results of processing a set of validation queries on the raw data and the Bayesian Network.
Abstract:
Embodiments disclosed herein model lifelong intent detection as a class-incremental learning where a new set of intents/classes are added at each incremental step. To address the issue of catastrophic forgetting during lifelong intent detection (LID), an incremental learner is provided with Prompt Augmented Generative Replay, wherein unlike existing approaches that store real samples in replay memory, only concept words obtained from old intents are stored, which reduces memory consumption and speeds up incremental training still enabling not forgetting the old intents. Joint training of an incremental learner is carried out for LID and a pseudo-labeled utterance generation with objective is to classify a user utterance into one of multiple pre-defined intents by minimizing a total Loss function comprising a LID loss function, a Labeled Utterance Generation loss function, a Supervised Contrastive Training loss function, and a Knowledge Distillation loss function.