Abstract:
A method for detecting deviations between an event log and a process model includes converting the process model into a probability process model, the probability process model comprising multiple nodes in multiple hierarchies and probability distribution associated with the multiple nodes, a leaf node among the multiple nodes corresponding to an activity in the process model; detecting differences between at least one event sequence contained in the event log and the probability process model according to a correspondence relationship; and identifying the differences as the deviations in response to the differences exceeding a predefined threshold; wherein the correspondence relationship describes a correspondence relationship between an event in one event sequence of the at least one event sequence and a leaf node in the probability process model.
Abstract:
A method, computer program product, and system are provided. A resource exhaustion is predicted based on the available resource and a consumption speed of the resource in a bare-metal server. Then it is determined whether the resource exhaustion will have an impact on the Service Level Agreement (SLA) and, responsive to the determination that the resource exhaustion will have an impact on the SLA, a control of the resource consumption speed is enacted to slow down the resource exhaustion and to maintain the SLA.
Abstract:
Patient treatment may be optimized using Recurrent Neural Network (RNN) based state simulation and Reinforcement learning (RL) techniques to simulate future states and actions. A RNN state simulator and a RL action generator may be trained using patient data such as historical states and actions. The RL action generator may be optimized by applying the RNN state simulator to simulating future states and applying the RL action generator to generate recommended actions based on the simulated future states. This process may be iteratively performed until a computational convergence is reached by the RL action generator which may indicate that the RL action generator has been optimized. A patient state may be fed into the optimized RL action generator to generate an optimal recommended treatment action.
Abstract:
The subject disclosure relates to employing grouping and selection components to facilitate a grouping of failure data associated with oil and gas exploration equipment into one or more equipment failure type groups. In an example, a method comprises grouping, by a system operatively coupled to a processor, training data of a set of equipment failure data into one or more failure type groups based on one or more determined failure criteria, wherein the one or more failure type groups represent equipment failure classifications associated with energy exploration processes; and selecting, by the system, first ungrouped data from the set of equipment failure data based on a level of similarity between the first ungrouped data and the training data.
Abstract:
The present disclosure provides a method, apparatus and system for processing a case management model (CMM). According to an embodiment, there is provided a method for processing a CMM, the method includes: obtaining an existing CMM having a plurality of elements; obtaining a new CMM having at least one element; aligning an element of the new CMM to an element of the existing CMM according to match costs between the element of the new CMM and the plurality of elements of the existing CMM; and fusing the new CMM into the existing CMM based on the match cost between the aligned elements.
Abstract:
A method for distributed file transfers with high performance and reliability includes creating, on a first Trivial File Transfer Protocol (TFTP) server, a global cache, where the global cache is used to store up to a first portion of a data file. The method further includes storing, on the first TFTP server, in the global cache, one or more continuous data blocks that have exceeded a defined first request rate threshold, where the one or more continuous data blocks make up a subset of the data blocks of the first portion of the data file. The method further includes predicting, on the first TFTP server, a next data block in the data file to be stored in the global cache and in response to predicting the next data block, storing, on the first TFTP server, the next data block in the global cache.
Abstract:
The subject disclosure relates to employing grouping and selection components to facilitate a grouping of failure data associated with oil and gas exploration equipment into one or more equipment failure type groups. In an example, a method comprises grouping, by a system operatively coupled to a processor, training data of a set of equipment failure data into one or more failure type groups based on one or more determined failure criteria, wherein the one or more failure type groups represent equipment failure classifications associated with energy exploration processes; and selecting, by the system, first ungrouped data from the set of equipment failure data based on a level of similarity between the first ungrouped data and the training data.
Abstract:
A method for distributed file transfers with high performance and reliability includes creating, on a first Trivial File Transfer Protocol (TFTP) server, a global cache, where the global cache is used to store up to a first portion of a data file. The method further includes storing, on the first TFTP server, in the global cache, one or more continuous data blocks that have exceeded a defined first request rate threshold, where the one or more continuous data blocks make up a subset of the data blocks of the first portion of the data file. The method further includes predicting, on the first TFTP server, a next data block in the data file to be stored in the global cache and in response to predicting the next data block, storing, on the first TFTP server, the next data block in the global cache.
Abstract:
A memory array and a method for electrically coupling memory cell access devices to a word line. The memory array includes a source line electrically coupled to each source terminal of the memory cell access devices. The memory array also includes a first set of at least two vertical pillars positioned above and electrically coupled to the source line. A second set of vertical pillars electrically isolated from the source line and positioned such that the source line does not extend below the second set of vertical pillars is also included. Furthermore, gate terminals of the memory cell access devices laterally surround the first set of vertical pillars and the second set of vertical pillars. Finally, a first word line contact is positioned between two of the second set of vertical pillars. The first word line contact is electrically coupled to the gate terminals.
Abstract:
A memory system, system including the memory system and method of reducing memory system power consumption. The memory system includes multiple memory units allocable to one of a number of processor units, e.g., processors or processor cores. A memory controller receives requests for memory from the processor units and allocates sufficient space from the memory to each requesting processor unit. Allocated memory can include some Single Level per Cell (SLC) memory units storing a single bit per cell and other memory units storing more than one bit per cell. Thus, two processor units may be assigned identical memory space, while half, or fewer, than the number of cells of one are assigned to the other.