摘要:
Architecture that addresses an end-to-end solution for logical transactional replication from a shared-nothing clustered database management system, which uses adaptive cloning for high availability. This can be time based using a global logical timestamp. The disclosed architecture, used for refreshing stale clones, does not preserve user transaction boundaries, which is a more complex situation than where the boundaries are preserved. In such a scenario it is probable that for a given data segment no clone of the segment may contain the complete user transaction history, and hence, the history has to be pieced together from the logs of multiple different clones. This is accomplished such that log harvesting is coordinated with the clone state transitions to ensure the correctness of logical replication.
摘要:
The subject invention relates to systems and methods that enable pausing and/or resuming a backup of file system items. The items are backed up in an order based on a unique property associated with each of the items. When the backup is paused, the unique property associated with a last backed up item prior to the pause is stored. Additionally, a timestamp associated with a snapshot from which the backup was taken can be stored with the unique property. Upon being resumed, the backup can continue from the paused location via utilizing the stored unique property. The backup can be resumed either employing an original snapshot associated with a time prior to the pause or a second snapshot related to a subsequent time. The subject invention can perform an initial backup of file system items and/or an incremental backup of such items. Additionally, the subject invention can support backing up a file system at disparate levels of granularity such as, for example, an item level.
摘要:
A “lazy commit” allows database transactions to be committed faster by not waiting for log records of the transaction to be written to disk. When a system crash occurs during the commit process, transactions may be lost because the log data is not available to recreate the transaction. While lazy commits speed up processing, they also increase the potential for data inconsistency. This invention introduces the concept of “durable reads”—transactions that require isolation from lazy commit transactions and which must be guaranteed to read only durably committed data. When durable read transaction attempts to read data changed by a lazy commit transaction, the system ensures the lazy commit transaction's changes are first made durable.
摘要:
To ensure that logs representative of data changes are durably written, localized storage media cache is flushed to force a write to stable storage. Log sequence number tracking is leveraged to determine if log data in a localized storage media cache is likely to have not been written to durable storage. When this condition is likely, a FLUSH_CACHE command is issued to a storage media controller to force a durable write. This allows recovery of data changes to a transactional system even when its associated storage media does not provide write-ordering guarantees. Moreover, flushing of the storage media cache can be accomplished at a frequency that provides an upper bound on a maximum time between a transaction commit and data being made durable.
摘要:
A technique is described for storing data from a database across a plurality of data storage devices, wherein each data storage device is capable of being accessed only by a corresponding computer system in a group of interconnected computer systems. In accordance with the technique, an identifier of the database is received. An identifier of a storage tier instance is also received, wherein the storage tier instance comprises a logical representation of one or more storage locations within each of the data storage devices. Responsive to the receipt of the identifier of the database and the identifier of the storage tier instance, data from the database is stored in two or more of the storage locations logically represented by the storage tier instance, wherein each of the two or more storage locations in which data is stored is within a corresponding one of the data storage devices.
摘要:
A system and method for serialization and/or de-serialization of file system item(s) and associated entity(ies)is provided. A file system “item” comprises a core class which can include property(ies). An item can be simple or compound (e.g., includes other item(s) embedded in it). Associated with an item can be entity(ies) such as fragment(s), link(s) with other item(s) and/or extension(s). Through serialization, a consistent copy of the item and associated entity(ies), if any, can be captured (e.g., for transporting of the item and to reconstruct the item on a destination system). The serialization system includes an identification component that. identities entity(ies) associated with an item and a serialization component that serializes the item and associated entity(ies). The serialization component can further serialize a header that includes information associated with the item and associated entity(ies). The header can facilitate random access to the item and associated entity(ies) (e.g., allowing a reader to interpret/parse only the parts in which it is interested). The serialization system can expose application program interface(s) (API's) that facilitate the copying, moving and/or transfer of an item and its associated entity(ies) from one location to another location.
摘要:
Methods, systems, and computer-readable media are disclosed for transitioning clones and clone data access maps in response to node or media failure without blocking queries. In a system, a data request interface is configured to receive a query to access data at a storage device. Data management logic is configured to access clone state information of a plurality of clones including clones in a static state and clones in a transitory state. The data management logic is further configured to process the query to access the data according to the dynamic clone data access map, with functional correctness maintained.
摘要:
Architecture that addresses an end-to-end solution for logical transactional replication from a shared-nothing clustered database management system, which uses adaptive cloning for high availability. This can be time based using a global logical timestamp. The disclosed architecture, used for refreshing stale clones, does not preserve user transaction boundaries, which is a more complex situation than where the boundaries are preserved. In such a scenario it is probable that for a given data segment no clone of the segment may contain the complete user transaction history, and hence, the history has to be pieced together from the logs of multiple different clones. This is accomplished such that log harvesting is coordinated with the clone state transitions to ensure the correctness of logical replication.
摘要:
Methods, systems, and computer-readable media are disclosed for transitioning clones and clone data access maps in response to node or media failure without blocking queries. In a system, a data request interface is configured to receive a query to access data at a storage device. Data management logic is configured to access clone state information of a plurality of clones including clones in a static state and clones in a transitory state. The data management logic is further configured to process the query to access the data according to the dynamic clone data access map, with functional correctness maintained.
摘要:
To ensure that logs representative of data changes are durably written, localized storage media cache is flushed to force a write to stable storage. Log sequence number tracking is leveraged to determine if log data in a localized storage media cache is likely to have not been written to durable storage. When this condition is likely, a FLUSH_CACHE command is issued to a storage media controller to force a durable write. This allows recovery of data changes to a transactional system even when its associated storage media does not provide write-ordering guarantees. Moreover, flushing of the storage media cache can be accomplished at a frequency that provides an upper bound on a maximum time between a transaction commit and data being made durable.