摘要:
Various embodiments monitor system noise in a parallel computing system. In one embodiment, at least one set of system noise data is stored in a shared buffer during a first computation interval. The set of system noise data is detected during the first computation interval and is associated with at least one parallel thread in a plurality of parallel threads. Each thread in the plurality of parallel threads is a thread of a program. The set of system noise data is filtered during a second computation interval based on at least one filtering condition creating a filtered set of system noise data. The filtered set of system noise data is then stored.
摘要:
A tool for autoscaling applications in a shared cloud resource environment. The tool registers, by one or more computer processors, one or more trigger conditions. The tool initiates, by one or more computer processors, a scaling event based, at least in part, on at least one of the one or more trigger conditions. The tool determines, by one or more computer processors, a scaling decision for the scaling event based, at least in part, on one or more scaling rules related to the one or more trigger conditions.
摘要:
Composite service provisioning is provided. One or more processors pre-provisions a first pool of service instances of a first composite service. One or more processors pre-provisions a second pool of service instances of a sub-service of the first composite service, wherein instances of the first pool of service instances have placeholder credentials identifying the second pool of service instances.
摘要:
Composite service provisioning is provided. A processor receives a first demand value of a first composite service. The processor identifies a sub-service based on the first composite service. The processor pre-provisions a first pool of service instances corresponding to the first composite service, the first pool of service instances having a quantity of service instances based, at least in part, on the first demand value. The processor determines a second demand value of the sub-service based, at least in part, on the quantity of service instances of the first pool of service instances. The processor pre-provisions a second pool of service instances corresponding to the sub-service, the second pool of service instances having a quantity of service instances based, at least in part, on the second demand value. The processor modifies each of the first pool of service instances with placeholder credentials that identify the second pool of service instances.
摘要:
A method for executing native code in a distributed Java Virtual Machine (JVM) is disclosed herein. The method may include receiving, in a first thread executing in a remote execution container, a first native code-generated call, such as a Java Native Interface (JNI) call, to a second thread, the first call including a first array write request. The first call may be stored in an instruction cache and bundled with a second native code-generated call and sent to the second thread. The calls are unbundled and executed in the second thread. An opaque handle to an array returned by the second call is bundled with corresponding array data and returned to the first thread. The array data of the bundle is stored in a data cache and retrieved in response to requests for the array data addressed to the second thread. A corresponding computer program product is also disclosed.
摘要:
Systems and techniques that facilitate backend quantum runtimes are provided. In various embodiments, a system can comprise a memory that can store computer-executable components. The system can further comprise a processor that can be operably coupled to the memory and that can execute the computer-executable components stored in the memory. In various embodiments, the computer-executable components can comprise an execution orchestration engine component that can parse a computer program into classical and quantum portions and that can host the computer program by instantiating a classical computing resource. In various embodiments, the computer-executable components can further comprise a classical computation resource that can perform a classical computation on an adaptively scalable classical computing environment, and a quantum computing node that can include a runtime application and a quantum processor, wherein the execution orchestration engine component can orchestrate both classical execution of the computer program and quantum execution of a quantum computation.
摘要:
Embodiments relate to a system, program product, and method for supporting elastic execution of a machine learning (ML) workload using application based profiling. A joint profile comprised of both ML application execution and resource usage data is generated. One or more feature(s) and signature(s) from the joint profile are identified, and a ML execution model for ML application execution and resource usage is built. The ML execution model leverages the feature(s) and signature(s) and is applied to provide one or more directives to subsequent application execution. The application of the ML execution model supports and enables the ML execution to elastically allocate and request one or more resources from a resource management component, with the elastic allocation supporting application execution.
摘要:
A method for reducing reactivation time of services that includes examining page faults that occur during processing of a service after the service has been inactive to provide a plurality of prefetch groups, and formulating a prefetch decision tree from page fault data in the prefetch groups. Pages from an initial page table for the service following a reactivated service request are then compared with the prefetched pages in the resident memory in accordance with the prefetch decision tree. Pages in the page table that are not included in said prefetched pages are paged in. A process to provide to provide the service is executed using the page table. Executing the process substantially avoids page faults.
摘要:
A method, information processing system, and computer program product are provided for managing operating system interference on applications in a parallel processing system. A mapping of hardware multi-threading threads to at least one processing core is determined, and first and second sets of logical processors of the at least one processing core are determined. The first set includes at least one of the logical processors of the at least one processing core, and the second set includes at least one of a remainder of the logical processors of the at least one processing core. A processor schedules application tasks only on the logical processors of the first set of logical processors of the at least one processing core. Operating system interference events are scheduled only on the logical processors of the second set of logical processors of the at least one processing core.
摘要:
Various embodiments monitor system noise in a parallel computing system. In one embodiment, at least one set of system noise data is stored in a shared buffer during a first computation interval. The set of system noise data is detected during the first computation interval and is associated with at least one parallel thread in a plurality of parallel threads. Each thread in the plurality of parallel threads is a thread of a program. The set of system noise data is filtered during a second computation interval based on at least one filtering condition creating a filtered set of system noise data. The filtered set of system noise data is then stored.