Abstract:
When a WORM cartridge is formatted for Tape File System (LTFS) in advance, such as prior to shipment, Linear Tape File System Library Edition (LTFS LE) is expanded by software to reduce consumption of the index partition and to support elimination of the appending of unnecessary data. More specifically, instead of recording metadata in the index partition during normal unmounting, the metadata is recorded in separate local storage such as on hard disk drive (HDD), and the index partition is updated only when the cartridge is ejected from the library. In this way, the present invention is able to significantly reduce the frequency of index partition updates. Because an update occurs only when the user intentionally ejects a cartridge, overflow of the index partition before overflow of the data partition can be prevented.
Abstract:
An archival storage cluster of preferably symmetric nodes includes a data protection management system that periodically organizes the then-available nodes into one or more protection sets, with each set comprising a set of n nodes, where “n” refers to a configurable “data protection level” (DPL). At the time of its creation, a given protection set is closed in the sense that each then available node is a member of one, and only one, protection set. When an object is to be stored within the archive, the data protection management system stores the object in a given node of a given protection set and then constrains the distribution of copies of that object to other nodes within the given protection set. As a consequence, all DPL copies of an object are all stored within the same protection set, and only that protection set. This scheme significantly improves MTDL for the cluster as a whole, as the data can only be lost if multiple failures occur within nodes of a given protection set. This is far more unlikely than failures occurring across any random distribution of nodes within the cluster.
Abstract:
A method for providing for write once read many (WORM) times from at least some addresses of a storage drive that is otherwise manufactured for multiple writes to individual addresses. In at least one embodiment, a WORM area(s) is defined by a START_LBA and an END_LBA and the method uses a HWM_LBA to determine whether a LBA in the WORM area has been written to previously and to prevent previously written to LBA(s) in the WORM area from being rewritten. In at least one embodiment where there are multiple WORM areas, each WORM area has its own respective START_LBA, END_LBA and HWM_LBA.
Abstract:
When a WORM cartridge is formatted for Tape File System (LTFS) in advance, such as prior to shipment, Linear Tape File System Library Edition (LTFS LE) is expanded by software to reduce consumption of the index partition and to support elimination of the appending of unnecessary data. More specifically, instead of recording metadata in the index partition during normal unmounting, the metadata is recorded in separate local storage such as on hard disk drive (HDD), and the index partition is updated only when the cartridge is ejected from the library. In this way, the present invention is able to significantly reduce the frequency of index partition updates. Because an update occurs only when the user intentionally ejects a cartridge, overflow of the index partition before overflow of the data partition can be prevented.
Abstract:
A Write-Once Read-Many (WORM) memory controller receives data from a processing system that is addressed to a location in a storage device, stores the data and a tag at the location, receives second data from the processing system that is addressed to the location, determines that the location includes the tag, and prevents the second data from being stored at the location based upon the presence of the tag. A WORM memory device sends a reply to a controller in response to an initialization command. The reply includes an address number that corresponds with the storage capacity of the WORM memory device. The WORM memory device sends another reply to another controller in response to another initialization command. The initialization commands are different from each other. The other initialization reply includes an address number of zero.
Abstract:
For automatically committing a file of a file system to a WORM committed state when the file has not been modified for a certain period of time, a file server maintains a list of file identifiers and periodically scans the list to check whether or not the certain period of time has expired for each file on the list. If so, the file is committed to the WORM state. The file server also enforces an in-place lock in response to a client request to modify or delete a file when the certain period of time is found to have expired for the file. A minimum modify time and a maximum modified time are included in each entry of the list of file identifiers in order to skip entries and to reclaim memory of entries during the periodic scan.
Abstract:
A write-once-read-many (WORM) storage system that employs large-capacity and relatively inexpensive disks in connection with a file system on a file server is provided. The file system contains most or all of the required WORM functionality so as to impose a minimal footprint on client applications, client operating systems and open protocols if desired. The system is organized around WORM storage volumes that contain files that, when committed to WORM storage, cannot be deleted or modified. Any file path or directory tree structure used to identify the file within the WORM volume is locked and cannot be deleted. In one embodiment, the administrator creates a WORM volume, capable of storing designated WORM files. The client then creates an appropriate WORM file using the appropriate protocol semantics. The file is written to the volume and committed by transitioning the file attributes from a not-read-only to a read-only state. The file system recognizes the persistently stored WORM attribute of any file in a WORM volume as WORM file. Henceforth, any attempt to modify the file attributes, write to the file, or delete the file, by clients, administrators or other entities is rejected and a request denied message is returned to the attempting party.
Abstract:
An archive cluster application runs across a redundant array of independent nodes. Each node runs an archive cluster application instance comprising a set of software processes: a request manager, a storage manager, a metadata manager, and a policy manager. The request manager manages requests for data, the storage manager manages data read/write functions, and the metadata manager facilitates metadata transactions and recovery. The policy manager implements policies, which are operations that determine the behavior of an “archive object” within the cluster. The archive cluster application provides object-based storage. It associates metadata and policies with the raw archived data, which together comprise an archive object. Object policies govern the object's behavior in the archive. The archive manages itself independently of client applications, acting automatically to ensure that object policies are valid.
Abstract:
Techniques for providing write-once-read-many (WORM) storage are described herein. According to one embodiment, in response to a command to set a time attribute of a file to a first predetermined value, it is determined whether the first predetermined value is outside of an ordinary range of values associated with the time attribute in accordance with a file system associated with the file. The file is designated as a WORM file and a WORM retention period is set for the file based on the first predetermined value, if the first predetermined value is outside of an ordinary range of values associated with the time attribute. The designation of the file as a WORM file and setting the WORM retention period are performed in response to the command.
Abstract:
An archive cluster application runs in a distributed manner across a redundant array of independent nodes. Each node preferably runs a complete archive cluster application instance. A given nodes provides a data repository, which stores up to a large amount (e.g., a terabyte) of data, while also acting as a portal that enables access to archive files. Each symmetric node has a set of software processes, e.g., a request manager, a storage manager, a metadata manager, and a policy manager. The request manager manages requests to the node for data (i.e., file data), the storage manager manages data read/write functions from a disk associated with the node, and the metadata manager facilitates metadata transactions and recovery across the distributed database. The policy manager implements one or more policies, which are operations that determine the behavior of an “archive object” within the cluster. The archive cluster application provides object-based storage. Preferably, the application permanently associates metadata and policies with the raw archived data, which together comprise an archive object. Object policies govern the object's behavior in the archive. As a result, the archive manages itself independently of client applications, acting automatically to ensure that all object policies are valid.