JOURNAL REPLAY OPTIMIZATION
    721.
    发明公开

    公开(公告)号:US20240143210A1

    公开(公告)日:2024-05-02

    申请号:US18399555

    申请日:2023-12-28

    Applicant: NetApp Inc.

    CPC classification number: G06F3/064 G06F3/0619 G06F3/0656 G06F3/0659 G06F3/067

    Abstract: Techniques are provided for journal replay optimization. A distributed storage architecture can implement a journal within memory for logging write operations into log records. Latency of executing the write operations is improved because the write operations can be responded back to clients as complete once logged within the journal without having to store the data to higher latency disk storage. If there is a failure, then a replay process is performed to replay the write operations logged within the journal in order to bring a file system up-to-date. The time to complete the replay of the write operations is significantly reduced by caching metadata (e.g., indirect blocks, checksums, buftree identifiers, file block numbers, and consistency point counts) directly into log records. Replay can quickly access this metadata for replaying the write operations because the metadata does not need to be retrieved from the higher latency disk storage into memory.

    Defragmentation for log structured merge tree to improve read and write amplification

    公开(公告)号:US11971859B2

    公开(公告)日:2024-04-30

    申请号:US17732046

    申请日:2022-04-28

    Applicant: NetApp Inc.

    CPC classification number: G06F16/1748 G06F16/182

    Abstract: Techniques are provided for implementing a defragmentation process during a merge operation performed by a re-compaction process upon a log structured merge tree. The log structured merge tree is used to store keys of key-value pairs within a key-value store. As the log structured merge tree fills with keys over time, the re-compaction process is performed to merge keys down to lower levels of the log structured merge tree to re-compact the keys. Re-compaction can result in fragmentation because there is a lack of spatial locality of where the re-compaction operations re-writes the keys within storage. Fragmentation increases read and write amplification when accessing the keys stored in different locations within the storage. Accordingly, the defragmentation process is performed during a last merge operation of the re-compaction process in order to store keys together within the storage, thus reducing read and write amplification when accessing the keys.

    Methods for handling storage devices with different zone sizes and devices thereof

    公开(公告)号:US11966611B2

    公开(公告)日:2024-04-23

    申请号:US18332304

    申请日:2023-06-09

    Applicant: NetApp, Inc.

    CPC classification number: G06F3/0644 G06F3/0604 G06F3/0659 G06F3/0689

    Abstract: The disclosed technology relates determining a first subset of a plurality drives having a first zone size and a second subset of the plurality of drives having a second zone size different from the first zone size, within a redundant array of independent disks (RAID) group. A prevailing zone size between the first zone size and the second zone size is determined. One or more logical zones within the determined first subset of the plurality of drives and the determined second subset of the plurality of drives for a received input-output operation is reserved based on the determined prevailing zone size. The received input-output operation is completed within the reserved one or more logical zones within the determined first subset of the plurality of drives and the determined second subset of the plurality of drives.

    Re-aligning data replication configuration of primary and secondary data serving entities of a cross-site storage solution after a failover event

    公开(公告)号:US11966307B2

    公开(公告)日:2024-04-23

    申请号:US17751410

    申请日:2022-05-23

    Applicant: NetApp, Inc.

    CPC classification number: G06F11/2069 G06F11/2092 G06F2201/85 H04L67/1097

    Abstract: Systems and methods for re-aligning data replication configuration of a cross-site storage solution after a failover are provided. According to one embodiment, after a failover, the new primary distributed storage system orchestrates flipping of the data replication configuration of a peered consistency group (CG) to reestablish zero RPO and zero RTO protections for the peered CG. The primary causes the secondary distributed storage system to perform an atomic database operation on its remote configuration database to (i) delete an existing source configuration that identifies the secondary as a source of data replication; and (ii) persist a new destination configuration identifying the secondary as a destination of data replication. Then, the primary performs an atomic database operation on its local configuration database to (i) delete an existing destination configuration identifying the primary as the destination; and (ii) persist a new source configuration identifying the distributed storage system as the source.

    SHARING EPHEMERAL STORAGE OF A VIRTUAL MACHINE FOR USE AS VICTIM CACHES FOR USE BY VIRTUAL STORAGE APPLIANCES IN A CLOUD ENVIRONMENT

    公开(公告)号:US20240103771A1

    公开(公告)日:2024-03-28

    申请号:US18522756

    申请日:2023-11-29

    Applicant: NetApp, Inc.

    CPC classification number: G06F3/0664 G06F3/0613 G06F3/0631 G06F3/064 G06F3/067

    Abstract: Systems and methods are provided for sharing ephemeral storage of a virtual machine (VM) for use as victim caches for virtual storage appliances running on the VM. According to one embodiment, a central service may run within the VM and be responsible for managing allocation and reclamation of ephemeral storage space of the VM to/from the virtual storage appliances. Responsive to startup of a new virtual storage appliance on the VM, the new virtual storage appliance may request space from the central service to inform creation of its victim cache. In connection with servicing the request, the central service may take into consideration various factors including one or more of the total aggregate size of multiple local ephemeral drives associated with the VM, remaining available ephemeral storage space, the number of active virtual storage appliances, and the SLO of the virtual storage appliance seeking to establish its victim cache.

    Performing various operations at the granularity of a consistency group within a cross-site storage solution

    公开(公告)号:US11934670B2

    公开(公告)日:2024-03-19

    申请号:US17219759

    申请日:2021-03-31

    Applicant: NetApp, Inc.

    CPC classification number: G06F3/0631 G06F3/0604 G06F3/067

    Abstract: Systems and methods are described for efficiently performing various operations at the granularity of a consistency group (CG) within a cross-site storage solution. An example of one of the various operations includes an independent and parallel resynchronization approach that independently brings individual volumes of a CG to a steady state of in-synchronization (InSync), thereby contributing to scalability of CGs by supporting CGs having a large number of member volumes without requiring a change to the resynchronization process. Another example includes preserving dependent write-order consistency when a remote mirror copy goes out-of-synchronization (OOS) for any reason by driving all member volumes OOS responsive to any member volume becoming OOS. Yet another example includes independent creation of snapshots by member volumes to support efficient and on-demand creation by an application of a common snapshots of all or a subset of peered member volumes of a CG with which the application is associated.

    Garbage collection and bin synchronization for distributed storage architecture

    公开(公告)号:US11934656B2

    公开(公告)日:2024-03-19

    申请号:US17717454

    申请日:2022-04-11

    Applicant: NetApp Inc.

    CPC classification number: G06F3/0608 G06F3/0652 G06F3/067

    Abstract: Techniques are provided for implementing garbage collection and bin synchronization for a distributed storage architecture of worker nodes managing distributed storage composed of bins of blocks. As the distributed storage architecture scales out to accommodate more storage and worker nodes, garbage collection used to free unused blocks becomes unmanageable and slow. Accordingly garbage collection is improved by utilizing heuristics to dynamically speed up or down garbage collection and set sizes for subsets of a bin to process instead of the entire bin. This ensures that garbage collection does not use stale information about what blocks are in-use, and ensures garbage collection does not unduly impact client I/O processing or conversely falls behind on garbage collection. Garbage collection can be incorporated into a bin sync process to improve the efficiency of the bin sync process so that unused blocks are not needlessly copied by the bin sync process.

Patent Agency Ranking