摘要:
The present invention provides a method, system, and computer program product for supporting a large number of intermittently used application clusters without consuming all the resources needed to run server processes for all the application clusters all the time. A method in accordance with the present invention comprises: holding an application request for a dormant application cluster in a holding zone; changing a status of the dormant application cluster to active; changing a status of an unused active application cluster to dormant; stopping server processes for the newly dormant application cluster; starting server processes for the newly active application cluster; and directing the application request held in the holding zone to the server processes for the newly active application cluster.
摘要:
A method and apparatus route hypertext protocol requests to one of a plurality of application servers, which share a database through a backend database management system. The application servers store session data in the database. Hence, if a subsequent request is routed to a different application server, the session data is available through the backend database management system. One or more web servers perform routing of requests to the application server. When a request is received that is accompanied by a session ID, routing is performed by utilizing a hash function on the session ID. The resulting hash value is mapped to an application server. A hash function on a session ID will always result in the same hash value; therefore, the request will always be routed to the same application server. However, if an application server is non-functional, a new hash based on the previous hash is computed until a functional application server is selected.
摘要:
Thread pools in a multithreaded server are programmatically adjusted, based on observed statistics from the server's inbound workload. In a multithreaded server environment, response time to end users is improved while increasing the efficiency of software execution and resource usage. Execution time and wait/queued time are tracked, for various types of requests being serviced by a server. Multiple logical pools of threads are used to service these requests, and inbound requests are directed to a selected one of these pools such that requests of similar execution-time requirements are serviced by the threads in that pool. The number and size of thread pools may be adjusted programmatically, and the distribution calculation (i.e., determining which inbound requests should be assigned to which pools) is a programmatic determination. In preferred embodiments, only one of these variables is adjusted at a time, and the results are monitored to determine whether the effect was positive or negative. The disclosed techniques also apply to tracking and classifying requests by method name (and, optionally, parameters).
摘要:
An object state caching method can include augmenting a compiled object having at least one referenced method with conditional caching logic. The conditional caching logic can be executed in an execution environment to determine whether to execute the referenced method. If the conditional caching logic determines not to execute the referenced method, the execution of the referenced method can be bypassed. Additionally, the stored object state can be retrieved from a communicatively coupled object cache. Finally, the retrieved object state can be forwarded to a requesting process. If, however, the conditional caching logic determines to execute the referenced method, the referenced method can be executed producing the resultant object state, and, the resultant object state can be cached in the coupled object cache.
摘要:
A method of caching contextually variant objects in a common cache. The method can include identifying an object type for a requested object and determining whether the requested object has an object type which is specified among an enumerated set of cacheable object types which can be stored in the common cache. Importantly, each cacheable object type can have an associated context. If the requested object has an object type which is specified among the enumerated set of cacheable object types, a cache key can be computed for the requested object using cache key formulation rules for the associated context. Finally, the requested object can be retrieved from the common cache using the formulated cache key. Notably, in one aspect of the invention, the method also can include the step of invalidating individual objects in the common cache according to corresponding cache policies of associated contexts.
摘要:
The overall management task required to determine data dissemination in multi-user applications is reduced herein. The management task required at any one host in the multi-user application is also limited. In addition, the number of simultaneous connections that any one participating host must support in a large multi-user application is reduced. Each user is assigned to a single management partition.
摘要:
Illustrative embodiments disclose receiving a command to play a selected audio visual media on a client device. The client device determines portions of audio visual media from elected audio visual media and a sequence identifying each portion of the portions in a particular order for playing the portions. The portions and the sequence are determined according to a policy for playing each portion on the client device. The client device retrieves the portions to play in sequence and plays at least a partially retrieved first portion of the portions of the selected audio visual media on the client device. The first portion is identified based on the particular order in the sequence.
摘要:
A network system includes multiple network resource information handling systems (IHSs) for managing applications and application communications. An IHS operating system initializes an application optimizer to provide application acceleration capability to application optimizers, such as application delivery controllers (ADCs) and wide area network (WAN) optimizer controllers (WOCs) within the network system. Upon receipt of a server application request message (SARM), a network system server responds with a restful application optimizer message (RAOM) that includes protocol, policy, and other application optimizer information that pertains to the requesting SARM. Application optimizers may include clients, ADCs and WOCs that reside within the message communication path between client and server. Application optimizers may store protocol, policy, and other information from RAOM 280 to populate application table data. Application optimizers intercept messages between network resources of the network system and apply message policies to improve message performance thereby improving application performance within the network system. Application acceleration provides improvements in quality of experience (QoE) and quality of service (QoS).
摘要:
Embodiments of the present invention provide a method, system and computer program product for wireless tower caching. In an embodiment of the invention, a method for wireless tower caching in a radio access network (RAN) can include receiving a response to a request from an end user device wirelessly coupled to a base station of the RAN, determining geographic positioning data for the response and identifying a most geographically wireless tower coupled to the RAN relative to the determined geographic positioning data (such as the closest wireless tower to the determined geographic positioning data). Thereafter, the response can be cached at a cache in a base station of the identified wireless tower. In this way, the cached content can remain relevant for a significant period of time in as much as the correlation between a likely position of the end user device and the content cached at the base station of the wireless tower at the time of use of the cached content is strong.
摘要:
A method for application-specific chargeback of content cached at a wireless tower of a RAN includes receiving a request for content from an end user device in a base station of a RAN and routing the request to an application in a packet switched data communications network from the base station by way of a CN. Content from the application can be received in the base station within a response to the request, the response can be cached in a local cache of the base station in connection with the application and the response can be forwarded to the end user device. Subsequently, in response to a new request for the content received in the base station, the content can be served in a response from the local cache. Further, an identity of the application can be transmitted to a remote charging service external to the base station but within the RAN for charging the application in consequence of having served the content from the local cache of the base station.