摘要:
A sync manager invokes a scheduled sync task that includes filter attributes for filtering a plurality of memory pages into a subset of memory pages. The sync manager selects file pointers that correspond to a plurality of files and traverses through the files using the selected file pointers. While traversing, the sync manager extracts file attributes for each of the files and, in turn, filters the files by comparing the filter attributes with the extracted file attributes. The sync manager then extracts memory page identifiers that correspond to the filtered files and synchronizes a subset of memory pages using the extracted memory page identifiers.
摘要:
A sync manager invokes a scheduled sync task that includes filter attributes for filtering a plurality of memory pages into a subset of memory pages. The sync manager selects file pointers that correspond to a plurality of files and traverses through the files using the selected file pointers. While traversing, the sync manager extracts file attributes for each of the files and, in turn, filters the files by comparing the filter attributes with the extracted file attributes. The sync manager then extracts memory page identifiers that correspond to the filtered files and synchronizes a subset of memory pages using the extracted memory page identifiers.
摘要:
In accordance with the invention, it has been recognized that for an application that is of a particular type and has certain characteristics, a benefit can be achieved by presenting multiple files to the application in a particular pre-planned or pre-specified order. Accordingly, an embodiment of the invention is directed to a method in a data processing system, wherein information pertaining to multiple data files is located in a container, and a plurality of applications of different types are each disposed to request a presentation of the files. Responsive to a request from a given application for presentation of the files, it is determined that the given application is of a particular type. The method further includes specifying an order for the presentation of the files, wherein the specified order is pre-selected to achieve an objective associated with applications of the given type. The files are then presented to the given application in the specified order.
摘要:
Page faults during partition migration from a source computing system to a destination computing system are reduced by assigning each page used by a process as being hot or cold according to their frequency of use by the process. During a live partition migration, the cold or coldest (least frequently used) pages are copied to the destination server first, followed copying the warmer (less frequently used) and concluded by copying the hottest (most frequently used) pages. After all dirtied pages have been refreshed, cutover from the instance on the source server to the destination server is made. By transferring the warm and hot pages last (or later) in the migration process, the number of dirtied pages is reduced, thereby reducing page faults subsequent to the cutover.
摘要:
Personal digital assistant devices are provided that query power subsystems to determine a level of battery power currently available for operating processes, and further determine a number of screen display changes that may be effected on a persistent display device at that determined level. If the determined number of screen display changes that may be effected at the determined level of available battery power fails to meet a specified minimum threshold number of images, the processing unit powers down subsystems of the personal digital assistant device and causes a highest listed image of a prioritized list of screen image instances to be displayed on the persistent display device, which persistently displays the image after a loss of power.
摘要:
Page faults during partition migration from a source computing system to a destination computing system are reduced by assigning each page used by a process as being hot or cold according to their frequency of use by the process. During a live partition migration, the cold or coldest (least frequently used) pages are copied to the destination server first, followed copying the warmer (less frequently used) and concluded by copying the hottest (most frequently used) pages. After all dirtied pages have been refreshed, cutover from the instance on the source server to the destination server is made. By transferring the warm and hot pages last (or later) in the migration process, the number of dirtied pages is reduced, thereby reducing page faults subsequent to the cutover.
摘要:
Provided are techniques for the tracing of computer processes and, more specifically, to techniques that enable a work unit to be traced across multiple processing components. A data structure is created and associated with a work unit. The first time a unit of work is processed a tracing context is retrieved form a pool and associated with the data structure. As the unit of work is passed from component to component, the data structure is passed so that each component has access and is able to augment the same tracing context associated with the unit of work.
摘要:
Provided are techniques for the tracing of computer processes and, more specifically, to techniques that enable a work unit to be traced across multiple processing components. A data structure is created and associated with a work unit. The first time a unit of work is processed a tracing context is retrieved form a pool and associated with the data structure. As the unit of work is passed from component to component, the data structure is passed so that each component has access and is able to augment the same tracing context associated with the unit of work.
摘要:
A system for balancing component load. In response to receiving a request, data is updated to reflect a current number of pending requests. In response to analyzing the updated data, it is determined whether throttling is necessary. In response to determining that throttling is not necessary, a corresponding request to the received request is created and a flag is set in the corresponding request. Then, the corresponding request is sent to one of a plurality of lower level components of an input/output stack of an operating system for processing based on the analyzed data to balance component load in the input/output stack of the operating system.
摘要:
A mechanism is provided for automatically closing a file or a device. A service routine monitor monitors a request received from either an application that opened the file or a device driver that readied the device. The service routine monitor determines whether the file or the device has been accessed within a predetermined time interval. Responsive to the file or the device failing to be accessed within the predetermined time interval, the service routine monitor sends a call to the application that opened the file or the application or a higher level device driver that requested that the device driver ready the device. Responsive to a response from the application or the higher level device driver indicating that the use of the file or the device is no longer needed, the service routine monitor closes the file or quiesces the device.