BACKGROUND DELETION OF LARGE DIRECTORIES
    1.
    发明公开

    公开(公告)号:US20240111722A1

    公开(公告)日:2024-04-04

    申请号:US18057384

    申请日:2022-11-21

    Applicant: VMware, Inc.

    CPC classification number: G06F16/162 G06F16/185 G06F16/188

    Abstract: Deleting directories in a virtual distributed file system (VDFS), and non-virtual file systems, involves changing the name of a selected directory to a unique object identifier (UID) and moving the selected directory, named according to the UID, to a deletion target directory. A recursive process, implemented using a background deletion thread, starts in the current directory and identifies objects in the current directory. For an object that is a file or an empty directory, the object is added to a deletion queue. For an object that is a directory that is not empty, the recursion drops down into that directory as the new current directory. When the recursion has exhausted the selected directory, or some maximum object count has been reached, the objects identified in the deletion queue are deleted. This approach can also be used for file operations other than deletion, such as compression, encryption, and hashing.

    OBJECT-BASED LOAD BALANCING APPROACHES IN DISTRIBUTED STORAGE SYSTEM

    公开(公告)号:US20210224096A1

    公开(公告)日:2021-07-22

    申请号:US16820735

    申请日:2020-03-17

    Applicant: VMware, Inc.

    Abstract: One example method to perform object-based load balancing in a distributed storage system of a virtualization system supported by a cluster of host machines may include determining, by a first host machine in the cluster, whether any host machine in the cluster has affinity to a site. The method may also include determining, by the first host machine, whether to distribute affinity Internet small computer system interface (iSCSI) targets owned by the first host machine to at least the second machine based on a first balance objective associated with the site, and after having considered the first balance objective, determining, by the first host machine, whether to distribute iSCSI targets without affinity owned by the first host machine to other host machines in the cluster based on a second balance objective associated with the cluster.

    ROUTING TRAFFIC OF A LOGICAL UNIT TO MULTIPLE BACKEND DATA OBJECTS BASED ON METADATA MAPPING

    公开(公告)号:US20210109862A1

    公开(公告)日:2021-04-15

    申请号:US16653548

    申请日:2019-10-15

    Applicant: VMware, Inc.

    Abstract: The disclosure herein describes enabling use of a logical unit for data storage in a distributed storage system using a plurality of backend data objects. Based on receiving instructions to create a logical unit of a logical unit size, a target backend object size to be used with the logical unit is determined, and a plurality of backend objects for allocation to the logical unit is calculated. The backend objects are allocated to the logical unit and a metadata mapping associated with the logical unit is generated. The metadata mapping associates logical block addresses of the logical unit to the allocated backend objects. The logical unit is linked with the metadata mapping in an input/output (I/O) service and, based on the linked metadata mapping, I/O traffic is routed to and from the logical unit. Using multiple backend objects enhances flexibility and efficiency of data storage on the distributed storage system.

    SYSTEM AND METHOD OF A HIGHLY CONCURRENT CACHE REPLACEMENT ALGORITHM

    公开(公告)号:US20210141728A1

    公开(公告)日:2021-05-13

    申请号:US16679570

    申请日:2019-11-11

    Applicant: VMware, Inc.

    Abstract: Disclosed are a method and system for managing multi-threaded concurrent access to a cache data structure. The cache data structure includes a hash table and three queues. The hash table includes a list of elements for each hash bucket with each hash bucket containing a mutex object and elements in each of the queues containing lock objects. Multiple threads can each lock a different hash bucket to have access to the list, and multiple threads can each lock a different element in the queues. The locks permit highly concurrent access to the cache data structure without conflict. Also, atomic operations are used to obtain pointers to elements in the queues so that a thread can safely advance each pointer. Race conditions that are encountered with locking an element in the queues or entering an element into the hash table are detected, and the operation encountering the race condition is retried.

Patent Agency Ranking