摘要:
A method for generating and communicating file system hints. The method may include receiving an I/O request from a file system layer and checking the I/O request for file system contextual information. The method may also include accessing the file system layer to determine attributes of the file system contextual information and receiving the attributes of the file system contextual information from the file system layer. The method may further include analyzing attributes of the file system contextual information and generating a hint based upon analyzing the attributes of the file system contextual information. The method may include a mechanism to provide weight-age of data passed from the application. The hint may be associated with the I/O request, and the hint may comprise hotness information to increase efficiency of data accessibility at a lower level storage tier. The method may include sending the hint to the lower level storage tier.
摘要:
A method for implementing selective file system caching. The method includes receiving I/O requests from an application and comparing each of the I/O requests with a configurable cache map, wherein the configurable cache map controls a selective caching based on an I/O type and a file type. Each of the I/O requests are processed to render caching decisions based on the configurable cache map. Selective caching is then implemented on each of the I/O requests in accordance with the caching decisions.
摘要:
The described implementations relate to processing of electronic data. One implementation is manifested as a system that can include a cache module and at least one processing device configured to execute the cache module. The cache module can be configured to store data items in slots of a cache structure, receive a request for an individual data item that maps to an individual slot of the cache structure, and, when the individual slot of the cache structure is not available, return without further processing the request. For example, the request can be received from a calling application or thread that can proceed without blocking irrespective of whether the request is fulfilled by the cache module.
摘要:
Methods and systems for populating a cache memory that services a media composition system. Caching priorities are based on a state of the media composition system, such as media currently within a media composition timeline, a composition playback location, media playback history, and temporal location within clips that are included in the composition. Caching may also be informed by descriptive metadata and media search results within a media composition client or a within a media asset management system accessed by the client. Additional caching priorities may be based on a project workflow phase or a client project schedule. Media may be partially written to or read from cache in order to meet media request deadlines. Caches may be local to a media composition system or remote, and may be fixed or portable.
摘要:
A method, apparatus and computer-readable storage medium distribute a non-live content stream in a network. An initial meta-file is transmitted in response to a request for the content, which identifies a division of the content stream into blocks, and available sources for delivery of the blocks. The initial meta-file can identify a first multicast and a second multicast server, assigning a first and second portion of the blocks for delivery using the first and second multicast source server, respectively. The first and second portions are transmitted using the first and second multicast source servers, respectively. The first and second portions correspond to distinct non-overlapping portions of the non-live content stream. The initial meta-file can also identify a unicast source server, assigning a third portion of the blocks for delivery using the unicast source server, the third portion being transmitted by the unicast source server.
摘要:
Systems. Methods, and Computer Program Products are provided managing a global cache coherency and reducing messaging traffic for coordination of access permissions in a distributed shared caching for a clustered file systems (CFS). The CFS manages access permissions to an entire space of the data segments by using the DSM module. In response to processing a request message from a remote DSM module to access one of the data segments, including in a response message permission to access the one of the data segments, and deciding if one of a transfer of ownership, and most recent contents of the one of the data segments should be included in the response message for the permission to access the one of the data segments. The last two components of the response message depend upon availability of the one of the data segments that is requested in a local external cache.
摘要:
Methods and systems are presented for prefetching music for devices. The disclosed method includes detecting an access to a music library through a device. The music library is associated with a user music account and is managed in cloud music storage. The disclose method further includes monitoring use activities for the user music account and the use activities of selected songs in the music library are adjustable for relevance based on assigned weightings. One or more songs from the music library that satisfy a prefetching constraint based on the monitored use activities are identified. The songs that satisfy the prefetching constraint are tagged for transferring to the device. The tagged songs are transferred to the device without receiving a user initiated request for the tagged songs.
摘要:
A distributed storage system may store data object instances in persistent storage and may cache keymap information for those data object instances. The system may cache a latest symbolic key entry for some user keys of the data object instances. When a request is made for the latest version of stored data object instances having a specified user key, the latest version may be determined dependent on whether a latest symbolic key entry exists for the specified user key, and keymap information for the latest version may be returned. When storing keymap information, a flag may be set to indicate that a corresponding latest symbolic key entry should be updated. The system may delete a latest symbolic key entry for a particular user key from the cache in response to determining that no other requests involving the keymap information for data object instances having the particular user key are pending.
摘要:
A method and system for large data and distributed disk cache processing in a Pneuron platform 100. The system and method include three specific interoperable but distributed functions: the adapter/cache Pneuron 14 and distributed disk files 34, a dynamic memory mapping tree 50, and distributed disk file cleanup 28. The system allows for large data processing considerations and the ability to access and acquire information from large data files 102 and rapidly distribute and provide the information to subsequent Pneurons 104 for processing. The system also provides the ability to store large result sets, the ability to deal with sequential as well as asynchronous parallel processing, the ability to address large unstructured data; web logs, email, web pages, etc., as well as the ability to handle failures to large block processing.
摘要:
Users can model changes to entities in large scale sets of data for various portfolio holdings in different respective sandbox caches. A method includes loading, sets of data in a database and organizing the loaded data in entity caches according to an entity model, each entity cache corresponding to one or more entities associated with respective portfolio holdings. Further steps include creating an initial report of information drawn from the loaded data for manipulation by a user through a uses-interface, storing information in the initial report in a respective sandbox cache having data organized according to the entity model, and enabling the user to manipulate the respective sandbox cache to change values in the data organized according to the entity model in the respective sandbox cache without changing values of data in other sandbox caches or in the database.