摘要:
Systems, methods, and other embodiments associated with associated with automatic two-way generation and synchronization of notebook and pipeline are described. In one embodiment, a computer-implemented method includes: identifying a difference between pipeline and notebook representations of application logic, wherein the difference is one of (i) a notebook paragraph without a corresponding pipeline node and (ii) a pipeline node without a corresponding notebook paragraph; synchronizing the pipeline representation and notebook representation by, for the notebook paragraph that does not have a corresponding pipeline node, automatically generating a new pipeline node in the pipeline representation, and for the pipeline node that does not have a corresponding notebook paragraph, automatically generating a new notebook paragraph in the notebook representation; and updating either a pipeline user interface to show the new pipeline node or a notebook user interface to show the new notebook paragraph.
摘要:
Systems, methods, and other embodiments associated with associated with unified pipeline flow with common and phase-specific paths are described. In one embodiment, a method includes accepting, through a graphical user interface, a setting of a phase-specific link type for a link between nodes of a pipeline, wherein the phase-specific link type indicates that the link is associated with a particular phase; accepting, through the graphical user interface, a selection to execute the pipeline for the phase; parsing the pipeline to determine an execution set of nodes for execution in the phase based on the nodes being connected with links having either the link type or a default link type common to all phases; and executing the pipeline for the phase by executing the execution set of nodes, and not executing nodes not included in the set of nodes.
摘要:
Techniques for handling bulk file processing. One technique includes receiving a request to process transactions within a bulk file, consolidating the transactions into batches based on parameters used to define the transactions, processing a first set of exception validations for each of the batches, storing information for each of the batches that satisfies the first set of exception validations within a set of tables, processing, using JMS Queues and the set of tables, a second set of exception validations for each of the transactions within the batches that satisfy the first set of exception validations, collating, using a timer job and the set of tables, each of the transactions into subsequent batches based on whether each of the transactions satisfies or does not satisfies the second set of exception validations, and accounting each of the transactions in the subsequent batches that satisfy the second set of exception validations.
摘要:
Systems, methods, and other embodiments associated with decision tree machine learning native to a graph database are described. In one embodiment, a method includes receiving an objective property value of a training graph for a graph-native decision tree to be trained to predict; determining feasible ranges of values from the training graph; automatically generating a graph query language grammar for graph-native decision trees based on the feasible ranges; progressively evolving output decision trees in a graph query language from the graph query language grammar to generate a final decision tree that accurately predicts the objective property value; and storing the final decision tree for subsequent execution against a target graph.
摘要:
Techniques are disclosed for attribution analysis in analytical workflows. A data processing system (DPS) obtains an overall model comprising one or more sub-models. The DPS selects an output variable of the overall model for which attribution of changes is to be performed, and a plurality of input variables against which changes are to be attributed to. The overall model is initially executed with respect to a data set of values for the plurality of input variables to generate a base result for the output variable. The overall model is iteratively executed based on a condition associated with the plurality of input variables to obtain a new result for the output variable. In each iteration, a value of an input variable is changed with respect to the data set of values and a change in the output variable with respect to the base result is attributed to the corresponding input variable.
摘要:
Techniques for retaining in-memory dataframes beyond an in-memory processing session. One technique includes receiving a request to execute a first run having a first set of tasks, creating a first session to execute the first run, and executing the first run in the first session using a dataframe constructed for a dataset defined as a component of the first run. The executing the first run generates an updated dataframe. The technique further includes receiving a request to execute a second run having a second set of tasks. A dependency exists between the first run and the second run based on a condition that the dataset is defined as a component of the first run and the second run. The technique further includes creating a second session to execute the second run, and executing the second run in the second session using the updated dataframe for the dataset.
摘要:
An aspect of the present disclosure generates instruction sets implementing business rules designed to update business objects of financial applications. In one embodiment, business rules designed to update business objects used by a financial application are processed to form a set of buckets (of rules) such that each bucket contains rules that do not have inter-dependency in execution order. Then, for each bucket, subsets of rules designed to update a common business object are determined, and a corresponding single instruction set is generated for each determined subset of rules. The instruction sets generated for the subsets of rules contained in each bucket are then executed concurrently.
摘要:
Systems, methods, and other embodiments associated with controlling an execution pipeline are described. In one embodiment, a method includes generating an execution pipeline for executing a plurality of tasks. The example method may also include evaluating execution definitions of the tasks to identify execution properties of the plurality of tasks. The example method may also include assigning each task to an execution environment selected from a set of execution environments based upon execution properties of the task matching execution properties of the execution environments. The example method may also include controlling the execution pipeline to execute each task within the assigned execution environments.
摘要:
A system and method for determining outsourcing suitability of at least one business process in an enterprise. The method comprising screening at least one business process in the enterprise to obtain a first set of business processes, wherein the first set of business processes can potentially be outsourced. The method further comprising evaluating each potential-business-process corresponding to the first set of business process on a predefined criterion to obtain a second set of business process, wherein the second set of business processes can be outsourced. Also, a migration plan is prepared corresponding to potential-business-processes corresponding to the second set of business processes.
摘要:
A computer-implemented method includes: obtaining a record including text narration; inputting the text narration into an NLP model; generating, by the NLP model, encoded narration based on the text narration; determining similarity index between the encoded narration and each of encoded reference topics; comparing the similarity index between the encoded narration and each encoded reference topic to similarity threshold; and determining whether the similarity index between the encoded narration and each encoded reference topic is equal to or greater than the similarity threshold. When the similarity index is equal to or greater than the similarity threshold, a corresponding reference topic is added to a result group, and, when the similarity index is less than the similarity threshold, the corresponding reference topic is left out of the result group. The record is classified based on the reference topic included in the result group and corresponding to the encoded reference topic having greatest similarity index.