摘要:
A method, apparatus and programmed storage device for routing data through a communications network. More specifically, a programmable traffic manager is programmed with at least one application level directive and the data is routed through the network to one of the network nodes using the programmable traffic manager, which is programmed in accordance with the application level directive. In a particular example of this invention, a request from a client is routed by the programmable traffic manager to at least one a plurality of servers hosting an application, where the programmable traffic manager is routed in accordance with the application level directive.
摘要:
A method, apparatus and programmed storage device for routing data through a communications network. More specifically, a programmable traffic manager is programmed with at least one application level directive and the data is routed through the network to one of the network nodes using the programmable traffic manager, which is programmed in accordance with the application level directive. In a particular example of this invention, a request from a client is routed by the programmable traffic manager to at least one a plurality of servers hosting an application, where the programmable traffic manager is routed in accordance with the application level directive.
摘要:
A distributable and serializable finite state machine and methods for using the distributable and serializable finite state machine are provided wherein finite state machine instance can be location-shifted, time-shifted or location-shift and time-shifted, for example by serializing and deserializing each instance. Each instance can be located-shifted between agents, and a persistent memory storage location is provided to facilitate both location-shifting and time-shifting. Finite state machine instances and the actions that make up each instance can be run in a distributed fashion among a plurality of agents.
摘要:
A system and method are provided for sharing and caching information in a data processing system and for efficiently managing a cacheable state shared among processes and clones. In one aspect, a method for managing a plurality of caches distributed in a network comprises maintaining, by each cache, a plurality of statistics associated with a cacheable object, wherein the statistics associated with the cacheable object comprise an access frequency (A(o)), an update frequency (U(o)); an update cost (C(o)), and a cost to fetch the cacheable object from remote source (F(o)); computing, by each cache, a metric using said statistics, wherein the metric quantitatively assesses the desirability of caching the cacheable object; and utilizing the metric, by each cache, to make caching decisions associated with the cacheable object.
摘要:
Methods and systems are provided for unified job processing of interdependent heterogeneous tasks within a domain containing a plurality of nodes. Jobs containing components to be processed in the domain are submitted and are identified by type. A job control flow associated with the job type is identified and used to process the components of the job on various nodes within the domain. Multiple job control flows are handled simultaneously, and provisions are made for sharing common job components among jobs. The job control flow utilizes a finite state machine where a given instance of the finite state machine applies to a specific job type. The finite state machine can be expressed using extensible mark-up language schema.
摘要:
Plug-in configurable middleware for managing distributed applications includes at least one core runtime engine configured as a plurality of concurrent instantiations on one or more hosts within a distributed architecture. These hosts can represent separate nodes or a single node within the architecture. Each core runtime engine instance provides the minimum amount of functionality required to support plug-in architecture, that is to support the instantiation of one or more plug-ins within that core runtime engine instance. Each core runtime engine instance is in communication with other concurrent core runtime engine instances and can share the functionality of plug-in instances with the other core runtime engine instances, for example through the use of proxies. A plurality of personalities representing pre-defined functions is defined and one of more of these personalities is associated with each core runtime engine instance. A plurality of pre-defined plug-ins are defined and associated with the personalities.
摘要:
A distributable and serializable finite state machine and methods for using the distributable and serializable finite state machine are provided wherein finite state machine instance can be location-shifted, time-shifted or location-shift and time-shifted, for example by serializing and deserializing each instance. Each instance can be located-shifted between agents, and a persistent memory storage location is provided to facilitate both location-shifting and time-shifting. Finite state machine instances and the actions that make up each instance can be run in a distributed fashion among a plurality of agents.
摘要:
A system and method for managing cachable entities (i.e., entities stored in a cache and/or entities which may be stored in a cache) in a data processing application. In one aspect, a method for managing cachable entities comprises the steps of analyzing program code to determine if there is at least one statement which affects a desirability of performing at least one cache transaction; and performing the at least one cache transaction if it is desired. In another aspect, a program analysis tool statically analyzes program code to locate points where object state changes occur, where objects are created and where objects are deleted, and then generates regularized dependencies at such points for and employing the dependencies to invalidate dependent cached queries. Query specific keys are generated to insert query results into and retrieve query results from a dependency managed cache. Regularized dependencies are generated at the object query points and are attached to query results inserted into a dependency managed cache, which facilitates consistency of query results from the dependency managed cache.
摘要:
Improved load balancing techniques are disclosed. For example, in one illustrative aspect of the invention, a method of satisfying requests in a system comprised of a plurality of servers comprises the following steps. At least one load balancer is provided for routing requests to the plurality of servers. At the at least one load balancer, a request sent from a client is obtained. At the at least one load balancer, the request is examined. Costs of satisfying the request by at least two of the plurality of servers are estimated. The estimation is based on at least one of a number and a cost of at least one remote access for satisfying the request. The request is routed to a server of the plurality of servers with a low estimated cost of satisfying the request.
摘要:
A distributable and serializable finite state machine and methods for using the distributable and serializable finite state machine are provided wherein finite state machine instance can be location-shifted, time-shifted or location-shift and time-shifted, for example by serializing and deserializing each instance. Each instance can be located-shifted between agents, and a persistent memory storage location is provided to facilitate both location-shifting and time-shifting. Finite state machine instances and the actions that make up each instance can be run in a distributed fashion among a plurality of agents.