摘要:
The disclosed massive multi-player on-line gaming environment and method eliminate the coordination and synchronization problems between servers by assigning each game user to a specific client proxy server for game duration and utilizing a space information distribution center to indicate client specific changes in game space. Such configuration eliminates the need to use “grey” areas around partitioning boundaries and servers having to “own” client objects thereby improving massive multi-player on-line gaming scalability.
摘要:
The initial partitioning of a distributed computing problem can be critical, and is often a source of tedium for the user. A method is provided that automatically segments the problem into fixed sized collections of original program cells (OPCs) based on the complexity of the problem specified, and the combination of computing agents of various caliber available for the overall job. The OPCs that are on the edge of a collection can communicate with OPCs on the edges of neighboring collections, and are indexed separately from OPCs that are within the ‘core’ or inner non-edge portion of a collection. Consequently, core OPCs can iterate independently of whether any communication occurs between collections and groups of collections (VPPs). All OPCs on an edge have common dependencies on remote information (i.e., their neighbors are all on the same edge of a neighboring collection).
摘要:
The initial partitioning of a distributed computing problem can be critical, and is often a source of tedium for the user. A method is provided that automatically segments the problem into fixed sized collections of original program cells (OPCs) based on the complexity of the problem specified, and the combination of computing agents of various caliber available for the overall job. The OPCs that are on the edge of a collection can communicate with OPCs on the edges of neighboring collections, and are indexed separately from OPCs that are within the ‘core’ or inner non-edge portion of a collection. Consequently, core OPCs can iterate independently of whether any communication occurs between collections and groups of collections (VPPs). All OPCs on an edge have common dependencies on remote information (i.e., their neighbors are all on the same edge of a neighboring collection).
摘要:
A distributed parallel computing system actively takes advantage of problem partitioning to balance the computing load among computing resources continually during processing. Variable problem partitions (VPPs) are initially defined as groups of original problem cells (OPCs). VPPs may be redefined and redistributed during execution, if necessary, to optimize performance based on the actual computing agent parameters and costs observed or reported through self-tests. For example, a good rule for efficient execution of a computing problem may be that the time required to perform a computation sequence (iteration) of all OPCs in a VPP should be comparable to the time required to share results via edge OPCs at the VPP collection perimeters. The rules that yield cost-efficient execution may be saved and re-used to generate initial partitionings for subsequent computing problem execution runs.
摘要:
A method is described for automatically segmenting and populating a distributed problem, including determining a bounding space that is at least equal in size to a problem space volume, partitioning the bounding space into a plurality of sub-spaces capable of holding collections of original problem cells, mapping the distributed problem into the partitioned bounding space to populate the sub-spaces with data from the problem, and determining boundaries between the sub-spaces. The method further includes determining coupling interdependencies between sub-spaces, determining complexities associated with boundaries, calculating densities of the data from the problem, and if the densities of two or more contiguous collections fall below a lower level threshold, dynamically joining the two or more contiguous collections, if a particular collection density exceeds an upper threshold, dynamically subdividing the particular collection into a finer granularity, and processing the problem based on the boundaries, the coupling interdependencies, the complexities, and the densities.
摘要:
A relational data base management system that stores and retrieves large data objects (LOBs) from data base storage includes a data manager that maintains LOB data space allocation according to the buddy system with two types of allocation pages, called regular buddy space (RBS) pages and super buddy space (SBS) pages. The RBS pages control buddy spaces comprising equal-size areas of the LOB data space in which LOB buddy segments are stored. The SBS pages control the RBS pages. The RBS pages include a pointer array that contains the forward-most location of where searching for a free storage space should begin and a counter array that indicates the number of free buddy segments in a buddy space for a particular buddy segment size. The SBS pages include analogous pointer and counter arrays that refer to the RBS pages and indicate the location and extent of free storage space, and also include a divided-space array and an SBS page directory.
摘要:
An “active” calendar automatically analyzes a user's calendar entries and sends machine-readable messages to destinations appropriate to content of the calendar entry. A group of event categories is established, each category specifying one class of anticipated calendar entry. An action rule database pre-associates each event category with one or more message formats each having a content and a destination. The action rule database also contains data identifying sources containing the content and destination for each message format. These sources include records of the action rule database itself, subparts of calendar entries of the pre-associated event category, one or more other databases, or a combination of the foregoing. After the calendar receives a user-submitted computer calendar entry describing a planned event, it identifies one of the event categories of the established group that classifies the planned event. For each message format pre-associated with the identified event category, the calendar determines the content and destination for the message as specified by the action rule database, and transmits the message to the destination.
摘要:
A method, apparatus, and article of manufacture for exchanging information in a computer-implemented database system. The present invention implements this exchange using a new operator known as a Rhonda operator. A Rhonda operator includes a tuple and template as arguments and, when performed, atomically swaps its tuple with a tuple from another Rhonda operator when both their templates match. More specifically, if two processes perform Rhonda operations, and each process' template argument matches the other process' tuple argument, then each process receives the other process' tuple as a result. This atomic synchronization can be performed for two or more Rhonda operators at a time.
摘要:
A method, apparatus, and article of manufacture for dynamically adding functionality to a server. A first operator is received at the server from an attached client, wherein the first operator indicates that new functionality is to be added to the server. A first handler is located for the first operator. The first handler is executed in the server, wherein the first handler registers a second operator associated with the new functionality and installs a second handler for the second operator to perform the new functionality.
摘要:
A message pathway selection system dynamically selects an optimum message pathway for transmitting messages. The system dynamically optimizes a message pathway according to various criteria such as, for example, efficiency, economy, data requirements, auditing requirements, security, data size, etc. The system can direct a message to bypass an infrastructure messaging server, using a direct message pathway. The system can also switch from an infrastructure messaging server to a direct method. The system can also utilize an infrastructure messaging pathway either as an alternative or in parallel with the direct message pathway. The system allows an application to use a single communication system for both a direct mode and an infrastructure mode of data transfer. The present system can bypass the infrastructure message pathway, thus reducing message latency, number of messages sent, and improving overall bandwidth.