Efficient differential techniques for metafiles

    公开(公告)号:US10242025B2

    公开(公告)日:2019-03-26

    申请号:US14924484

    申请日:2015-10-27

    Applicant: NETAPP, INC.

    Inventor: Rithin Shetty

    Abstract: Exemplary embodiments provide methods, mediums, and systems for replicating metafiles between a source and a destination. The metafile may be subdivided into blocks. The contents of the metafile may be transferred by locating the blocks which are changed between the source version of the metafile and the destination version of the metafile. The changed blocks may be examined to retrieve the contents of the changed blocks. The records in the changed blocks may be evaluated to determine whether to create a corresponding record at the destination, delete a corresponding record at the destination, or update a corresponding record at the destination. Accordingly, the metafile may be replicated in a logical manner, by transferring only changed records rather than the entirety of a changed block. Moreover, the transfer is conducted efficiently because unchanged blocks are eliminated from consideration at the outset.

    Storage layer based orchestration method for virtual machine migration across disparate virtualization environments

    公开(公告)号:US10228964B2

    公开(公告)日:2019-03-12

    申请号:US15387416

    申请日:2016-12-21

    Applicant: NetApp, Inc.

    Abstract: A storage layer based orchestration method can efficiently migrate a virtualized, enterprise scale system across disparate virtualization environments. A copy of a source logical storage container with multiple virtual disks of virtual machines (VMs) can be created in a public cloud destination as a destination logical storage container. Each of the VMs is associated with at least one virtual disk that includes boot data (“boot disk”) for the VM. With application programming interface function calls and/or scripted task automation and configuration management commands, the orchestration method coordinates different applications and tools to convert the boot disks into canonical storage representations (e.g., logical unit numbers (LUNs)), to instantiate VMs in the destination environment, and to chain load the boot disks to launch the VMs in a different virtualization environment.

    USE OF PREDEFINED BLOCK POINTERS TO REDUCE DUPLICATE STORAGE OF CERTAIN DATA IN A STORAGE SUBSYSTEM OF A STORAGE SERVER

    公开(公告)号:US20190018605A1

    公开(公告)日:2019-01-17

    申请号:US16133284

    申请日:2018-09-17

    Applicant: NETAPP, INC.

    Abstract: A method and system for eliminating the redundant allocation and deallocation of special data on disk, wherein the redundant allocation and deallocation of special data on disk is eliminated by providing an innovate technique for specially allocating special data of a storage system. Specially allocated data is data that is pre-allocated on disk and stored in memory of the storage system. “Special data” may include any pre-decided data, one or more portions of data that exceed a pre-defined sharing threshold, and/or one or more portions of data that have been identified by a user as special. For example, in some embodiments, a zero-filled data block is specially allocated by a storage system. As another example, in some embodiments, a data block whose contents correspond to a particular type document header is specially allocated.

    Granular consistency group replication

    公开(公告)号:US10176064B2

    公开(公告)日:2019-01-08

    申请号:US15054513

    申请日:2016-02-26

    Applicant: NetApp Inc.

    Abstract: One or more techniques and/or computing devices are provided for granular replication for data protection. For example, a first storage controller may host a first volume. A consistency group, comprising a subset of files, logical unit numbers, and/or other data of the first volume, is defined through a consistency group configuration. A baseline transfer, using a baseline snapshot of the first volume, is used to create a replicated consistency group within a second volume hosted by a second storage controller. In this way, an arbitrary level of granularity is used to synchronize/replicate a subset of the first volume to the second volume. If a synchronous replication relationship is specified, then one or more incremental transfer are performed and a synchronous replication engine is implemented. If an asynchronous replication relationship is specified, then snapshots are used to identify delta data of the consistency group for updating the replication consistency group.

    AUTOMATIC INCREMENTAL REPAIR OF GRANULAR FILESYSTEM OBJECTS

    公开(公告)号:US20180373596A1

    公开(公告)日:2018-12-27

    申请号:US15631477

    申请日:2017-06-23

    Applicant: NetApp, Inc.

    Abstract: Presented herein are methods, non-transitory computer readable media, and devices triggering a metadata recovery process within a network storage system, which include: dividing metadata into metadata segments, wherein each of the metadata segments is tasked to perform a specific file system operation function, validating each of the metadata segments during the specific file system operation function; upon failure to validate at least one of the metadata segments, triggering an automatic repair process while maintaining the operation function tasked to the metadata segment, and upon finalizing the automatic repair process, resuming the specific file system operation function tasked to the metadata segment.

    Cache affinity and processor utilization technique

    公开(公告)号:US10162686B2

    公开(公告)日:2018-12-25

    申请号:US15806852

    申请日:2017-11-08

    Applicant: NetApp, Inc.

    Abstract: A cache affinity and processor utilization technique efficiently load balances work in a storage input/output (I/O) stack among a plurality of processors and associated processor cores of a node. The storage I/O stack employs one or more non-blocking messaging kernel (MK) threads that execute non-blocking message handlers (i.e., non-blocking services). The technique load balances work between the processor cores sharing a last level cache (LLC) (i.e., intra-LLC processor load balancing), and load balances work between the processors having separate LLCs (i.e., inter-LLC processor load balancing). The technique may allocate a predetermined number of logical processors for use by an MK scheduler to schedule the non-blocking services within the storage I/O stack, as well as allocate a remaining number of logical processors for use by blocking services, e.g., scheduled by an operating system kernel scheduler.

Patent Agency Ranking