Abstract:
Storage system workload data associated with a storage system is analyzed. The workload data comprises indications of input and output operations associated with the storage system. A storage system configuration is determined based, at least in part, on said analyzing of the storage system workload data. An implementation plan comprising one or more operations for implementing the storage system configuration is generated.
Abstract:
In response to a request to perform an analysis of a storage system associated with a device, it is determined that the request indicates at least one of a proposed storage configuration, a configuration validation request, or a set of performance goals. Configuration data associated with the device is determined. The configuration data includes configuration data for an additional device. The analysis of the storage system is performed based, at least in part, on the configuration data associated with the device. Performing the analysis of the storage system comprises querying a database for entries associated with at least one of the device or the additional device.
Abstract:
Systems and methods that result in a stable storage system are provided. In the storage system, the latency spikes may be reduced when multiple volumes are aggregated into transfer sets according to system characteristics. The storage system transfers ownership of volumes in each transfer set as a single transaction. In the storage system, connectivity between the host and the storage controller is re-established based on the connectivity in a physical transport layer and a single path. In the storage system, pre-mature failback is also avoided when ownership of volumes is transferred back to a preferred storage controller when the same number of paths existed between the host and the preferred storage controller before and after a failover operation. Further, the storage system generates connectivity reports that display connectivity paths between hosts, storage controllers, and volumes.
Abstract:
A method, a computing device, and a non-transitory machine-readable medium for performing a multipath selection based on a determined quality of service for the paths. An example method includes a host computing device periodically polling a storage system for path information including an indication of a recommended storage controller. The host computing device periodically determines a quality of service information corresponding to a plurality of paths between the host computing device and a storage volume of the storage system, where at least one of the plurality of paths including the recommended storage controller. The host computing device identifies a fault corresponding to a path of the plurality of paths that routes I/O from the host computing device to the storage volume. The host computing device re-routes the I/O from the path to a different path of the plurality of paths, where the different path is selected for the re-routing based on the quality of service information and the path information.
Abstract:
Systems and methods that select a cache flushing algorithm are provided. A stripe that spans multiple storage devices and includes a plurality of segments is provided. The stripe also includes dirty data stored in a picket-fence pattern in at least a subset of segments in the plurality of segments. A memory cache that stores data separately from the plurality of storage devices and a metadata cache that stores metadata associated with the dirty data are also provided. A cache flushing algorithm is selected using the metadata. The selected cache flushing algorithm flushes data from the memory cache to the stripe.
Abstract:
A method, a computing device, and a non-transitory machine-readable medium for performing a multipath selection based on a determined quality of service for the paths. An example method includes a host computing device periodically polling a storage system for path information including an indication of a recommended storage controller. The host computing device periodically determines a quality of service information corresponding to a plurality of paths between the host computing device and a storage volume of the storage system, where at least one of the plurality of paths including the recommended storage controller. The host computing device identifies a fault corresponding to a path of the plurality of paths that routes I/O from the host computing device to the storage volume. The host computing device re-routes the I/O from the path to a different path of the plurality of paths, where the different path is selected for the re-routing based on the quality of service information and the path information.
Abstract:
Embodiments of the present disclosure enable high availability and performance in view of storage controller failure. A storage system includes three or more controllers that may be distributed in a plurality of enclosures. The controllers are in high availability pairs on a per volume basis, with volumes and corresponding mirror targets distributed throughout the storage system. When a controller fails, other controllers in the system detect the failure and assess whether one or more volumes and/or mirror targets are affected. If no volumes/mirror targets are affected, then write-back caching continues. If volume ownership is affected, then a new volume owner is selected so that write-back caching may continue. If mirror target ownership is affected, then a new mirror target is selected so that write-back caching may continue. As a result, write-back caching availability is increased to provide low latency and high throughput in degraded mode as in other modes.
Abstract:
A system, method, and computer program product is described for providing dynamic enabling and/or disabling of protection information (PI) in array systems during operation. A storage system receives a request to transition a volume from PI disabled to PI enabled during regular operation. The storage system synchronizes and purges the cache associated with the target volume. The storage system initiates an immediate availability format (IAF-PI) process to initialize PI for the associated data blocks of the volume's storage devices. The storage system continues receiving I/O requests as the IAF-PI process sweeps through the storage devices. The storage system inserts and checks PI for the write data as it is written to the storage devices. The storage system inserts PI for requested data above the IAF-PI boundary and checks PI for requested data below the IAF-PI boundary. The transition remains an online process that avoids downtime.
Abstract:
A clustered network-based storage system includes a host server, multiple high availability system controller pairs, and multiple storage devices across multiple arrays. Two independent storage array subsystems each include a quorum drive copy and are each controlled by a HA pair, with remote volume mirroring links coupling the separate HA pairs. The host server includes a virtualization agent that identifies and prioritizes communication paths, and also determines capacity across all system nodes. A system storage management agent determines an overall storage profile across the system. The virtualization agent, storage management agent, quorum drive copies and remote volume mirroring link all operate to provide increased redundancy, load sharing, or both between the separate first and second arrays of storage devices.
Abstract:
The storage device receives a write request from a disk controller to write data to a storage array. The storage device determines that one or more blocks are marked for deletion. In response to receiving the write request and determining that blocks are marked for deletion, the storage device issues a write command on a first media access channel for a first location of the storage array, and issues an erase command on a second media access channel for a different storage location of the storage array. Thus, the commands are issued concurrently on different channels.