Abstract:
A resiliency system detects and corrects memory errors reported by a memory system of a computing system using previously stored error correction information. When a program stores data into a memory location, the resiliency system executing on the computing system generates and stores error correction information. When the program then executes a load instruction to retrieve the data from the memory location, the load instruction completes normally if there is no memory error. If, however, there is a memory error, the computing system passes control to the resiliency system (e.g., via a trap) to handle the memory error. The resiliency system retrieves the error correction information for the memory location and re-creates the data of the memory location. The resiliency system stores the data as if the load instruction had completed normally and passes control to the next instruction of the program.
Abstract:
Systems and methods for multi-threaded server control automation for disaster recovery are described. A method may include initiating a disaster recovery sequence on two or more processors, wherein the disaster recovery sequence comprises a plurality of subsequences. The method may also include implementing the disaster recovery sequence on the two or more processors in parallel, wherein one or more subsequences of the disaster recovery sequence are implemented on the two or more processors in parallel. Upon completion of the disaster recovery sequence, at least one server partition is repurposed from a first configuration, such as a test configuration, to a second configuration, such as a production configuration.
Abstract:
Methods, systems, and articles of manufacture for injecting faults at select execution points of distributed applications are provided herein. A method includes monitoring a run-time state of each of multiple components of a distributed application to determine one or more sequence of events that triggers a fault injection point at one of the multiple components; defining a fault injection scenario in a specification based on said monitoring, wherein said fault injection scenario comprises a description of one or more sequence of events during which an intended fault is to be injected to a target component of the multiple components at one selected event; and executing the fault injection defined in the specification to perform injection of the intended fault during run-time of the distributed application.
Abstract:
The present disclosure is directed to an application store on board a vehicle. The application store contains, in one configuration, a plurality of applications for installation on an on board computer of the vehicle, with the applications provided to the vehicle operator being provided to the operator being based on predetermined types of information related to the vehicle, its state, operation, and/or configuration, vehicle location, vehicle type, make, model, and/or year of manufacture, and/or occupant(s) and/or occupant(s) of other vehicles.
Abstract:
Systems and methods are provided to track cluster nodes and provide high availability in a computing system. A computer system includes hosts, a cluster manager, and a cluster database. The cluster database includes entries corresponding to the hosts which identify the physical location of a corresponding host. The cluster manager uses the data to select at least two hosts and assign the selected hosts to a service group for executing an application. The cluster manager selects hosts via an algorithm that determines which hosts are least likely to share a single point of failure. The data includes a hierarchical group of location attributes describing two or more of a host's country, state, city, building, room, enclosure, and radio frequency identifier (RFID). The location-based algorithm identifies a group of selected hosts whose smallest shared location attribute is highest in the hierarchical group. The system updates the data whenever a physical location of a host changes.
Abstract:
A service appliance is installed between production servers running service applications and service users. The production servers and their service applications provide services to the service users. In the event that a production server is unable to provide its service to users, the service appliance can transparently intervene to maintain service availability. To maintain transparency to service users and service applications, service users are located on a first network and production servers are located on a second network. The service appliance assumes the addresses of the service users on the second network and the addresses of the production servers on the first network. Thus, the service appliance obtains all network traffic sent between the production server and service users. While the service application is operating correctly, the service appliance forwards network traffic between the two networks using various network layers.
Abstract:
A system, method and program product that provides a disaster recovery (DR) brokerage infrastructure. A system is described that includes: a network of intermediate storage nodes and peering points; and a brokerage manager that periodically evaluates a set of DR as a service (DRaaS) providers and selects a cost optimal DRaaS provider for a subscriber, wherein the brokerage manager replicates DR related data for the subscriber in an intermediate storage node prior to delivery to the cost optimal DRaaS provider.
Abstract:
A system that implements a data storage service may maintain tables in a data store on behalf of clients. The service may maintain table data in multiple replicas of partitions of the data that are stored on respective computing nodes in the system. In response to detecting a failure or fault condition, or receiving a service request from a client to move or copy a partition replica, the data store may copy a partition replica to another computing node using a physical copy mechanism. The physical copy mechanism may copy table data from physical storage locations in which it is stored to physical storage locations allocated to a destination replica on the other computing node. During copying, service requests to modify table data may be logged and applied to the replica being copied. A catch-up operation may be performed to apply modification requests received during copying to the destination replica.
Abstract:
In one aspect of the disclosure, progress identifiers (e.g., a plurality of flags) and/or generation numbers are utilized to determine a “correct” non-volatile random access memory (NVRAM) replay source to use during a retried switchover or early switchback. Initially, a switchover sequence is initiated. The switchover sequence includes a replay of logged information to storage devices of an aggregate to ensure data consistency of the aggregate. Progress of the switchover sequence is tracked using the progress identifiers and/or generation numbers. After an interruption during the switchover sequence, switchover is retried or an early switchback is performed. The retried switchover or early switchback decides whether to replay logged information, and, if so, determines an NVRAM replay source to replay logged information from based on states of the progress identifiers and/or comparison of the generation numbers.
Abstract:
Providing high availability multi-way conferencing. Separate signaling and media components may be provided within an MCU or among a cluster of MCUs. A signaling server may control signaling aspects of a conference while a media server may provide media support for the conference. In the event of media server failure, the signaling server may assign a new media server to provide media support for the conference. A backup signaling server may also monitor the signaling server and may provide signaling support for the conference in the event of signaling server failure.