摘要:
A custom class loader configured to dynamically locate and load classes in a virtual machine in accordance with an associated dependency specification. The custom class loader can include class loading logic configured to specifically and dynamically locate, define and load a class specified by name. The custom class loader also can include a list of peer class loaders arranged in a peer-to-peer fashion in accordance with the associated dependency specification in the virtual machine. In this regard, list generation logic can be included which can generate the list when the specified class has been replaced. Finally, the custom class loader can include deference logic configured to defer the location, definition and loading of the specified class to the peer class loaders in the list.
摘要:
A method and apparatus route hypertext protocol requests to one of a plurality of application servers, which share a database through a backend database management system. The application servers store session data in the database. Hence, if a subsequent request is routed to a different application server, the session data is available through the backend database management system. One or more web servers perform routing of requests to the application server. When a request is received that is accompanied by a session ID, routing is performed by utilizing a hash function on the session ID. The resulting hash value is mapped to an application server. A hash function on a session ID will always result in the same hash value; therefore, the request will always be routed to the same application server. However, if an application server is non-functional, a new hash based on the previous hash is computed until a functional application server is selected.
摘要:
An apparatus, program product and method are capable of dynamically assigning different code base identifiers to different classes defined in the same code base during class load operations to enable different code base-correlated policies to be assigned to such classes, thereby effectively granting different access permissions to classes that are defined within the same code base.
摘要:
A distributed cache management system that minimizes invalid cache notification events is provided. A cache management system in a sending device processes outgoing cache notification events by adding information about the source server's clock. A cache management system in the receiving device then uses this information to adjust event information once the event is received.
摘要:
In a method and system for monitoring events occurring at respective servers of a configuration of nodes, a first server located at a first node receives information from a messaging system pertaining to events at servers located at other nodes. The messaging system usefully comprises a highly available (HA) bulletin board or the like. When the first server receives a start event notification pertaining to a second server located at a second node, a direct communication path is established between the first and second servers. The first server identifies events in the second server that affect or are of interest to services of the first server. The first server then registers with the second server, to receive notification through the direct communication path when respective identified events occur.
摘要:
An apparatus, system, and storage medium that, in an embodiment, determine the classes that are needed during a first invocation of an application and preload those classes during a subsequent invocation of the application. In various embodiments, the classes may be preloaded if they are within a specified package, if they were needed greater than a threshold number of times, if the class path did not change between the first invocation and the subsequent invocation, or if the jar files the contain the classes are in a class path. In this way, an embodiment of the invention may increase the performance of class preloading.
摘要:
In a method and system for monitoring events occurring at respective servers of a configuration of nodes, a first server located at a first node receives information from a messaging system pertaining to events at servers located at other nodes. The messaging system usefully comprises a highly available (HA) bulletin board or the like. When the first server receives a start event notification pertaining to a second server located at a second node, a direct communication path is established between the first and second servers. The first server identifies events in the second server that affect or are of interest to services of the first server. The first server then registers with the second server, to receive notification through the direct communication path when respective identified events occur.
摘要:
A method, apparatus, system, and signal-bearing medium that, in an embodiment, determine the classes that are needed during a first invocation of an application and preload those classes during a subsequent invocation of the application. In various embodiments, the classes may be preloaded if they are within a specified package, if they were needed greater than a threshold number of times, if the class path did not change between the first invocation and the subsequent invocation, or if the jar files the contain the classes are in a class path. In this way, an embodiment of the invention may increase the performance of class preloading.
摘要:
A method for securing cached data in an enterprise environment. The method can include processing a request to locate data in a query cache. If the data can be located in the query cache, the data can be retrieved from the query cache. Additionally, at least one encrypted portion of the retrieved data can be decrypted. Finally, the decrypted portion and any remaining unencrypted portion of the retrieved data can be forwarded to a requesting client. By comparison, if the data cannot be located in the query cache, the data can be retrieved from a back-end data source over a computer communications network, and forwarded to the requesting client. Additionally, at least a portion of the retrieved data can be encrypted and both the encrypted portion and any remaining unencrypted portion can be stored in the query cache.
摘要:
Prioritization of fragmented messages in an ORB is achieved by intercepting the flow of message data within the ORB, marking the current data stream location and issuing sufficient read commands to gather the message header and control area information. Then classification for priority can be performed using the data in the control area to classify the request. The classification information is stored for the message for later use by queuing and prioritization processes and the data stream marker is reset to the original location. When the queuing and prioritization processes are ready to dispatch the message, the message can be sent on to the higher levels of processing within the ORB where the higher levels of message processing code will process the message, totally unaware of the fact that the message has been classified and/or queued or prioritized prior to delivery to the higher level.