Abstract:
A system and method for monitoring liveliness includes a management device which has an application layer where applications are executed. A connection monitor is located other than on the application layer, and the connection monitor is configured to receive requests from clients and deliver the requests to components on the application layer. The components include a generic application monitor which responds to liveliness monitor requests from the clients for all applications monitored, and one or more applications which response to requests to that application.
Abstract:
A system and method for scalable status change detection in a network of nodes includes embedding node coordinate positions into a virtual map environment. A cluster of nodes are represented with a representative node to form a hierarchy of nodes and clusters where changes in a status of the representative node imply a status change in the nodes of the cluster. By selectively monitoring the representative nodes for the changes in status, the representative nodes create a capability for scalable monitoring of the network.
Abstract:
The present invention broadly contemplates addressing QoS concerns in overlay design to account for the last mile problem. In accordance with the present invention, a simple queuing network model for bandwidth usage in the last-mile bottlenecks is used to capture the effects of the asymmetry, the contention for bandwidth on the outgoing link, and to provide characterization of network throughput and latency. Using this characterization computationally inexpensive heuristics are preferably used for organizing end-systems into a multicast overlay which meets specified latency and packet loss bounds, given a specific packet arrival process.
Abstract:
A number of improvements in network adapters that offload protocol processing from the host processor are provided. Specifically, an improved mechanism for handling receipt of data packets in a system utilizing an offload network adapter. The offload network adapter may include logic that permits the offload network adapter to delay notification of data reception to the host system in different ways. The advantage of delaying the notice of data packet reception to the host system is the potential for aggregation of several data packets, which can arrive immediately after the first one, for example, in a single notification. Given a stream with continuous data packet arrival, a value may be set, either statically or dynamically, for notification delay and this value may be configurable for the host system per communication socket.
Abstract:
In one embodiment, the present invention is a method and an apparatus for bandwidth efficient distribution of content over a computer network. In one embodiment, a server cooperates with clients to reduce the usage of server bandwidth, thereby reducing bandwidth-associated distribution costs. The server controls the way that content is distributed within the system, and the clients/receivers may act as intermediate forwarders of data packets. In one embodiment, this is accomplished by sending data items to a first group of receivers on a packet-by-packet basis using unicast communication. The first group of receivers then sends the packets on to a second set of receivers, and the process continues until the packet is forwarded to all intended receivers.
Abstract:
A mechanism for bottleneck avoidance is provided in an intelligent adapter. The mechanism allows the adapter to be used such that host/adapter system throughput is optimized. The bottleneck avoidance mechanism of the present invention determines when the adapter becomes a bottleneck. If certain conditions exist, then new connections are refused so that the adapter can process packets for existing connections. If certain other conditions exist, the adapter may migrate workload to the host processor for processing. These conditions may be determined by comparing memory usage or central processing unit usage to predetermined thresholds. Alternatively, the conditions may be determined by comparing adapter response time to host response time.
Abstract:
Access is obtained to a plurality of intermediately transformed electronic documents (with a plurality of sections and subsections) which have been transformed, by topical analysis and text summarization techniques, from a plurality of original electronic documents comprising at least some unstructured electronic documents. Audit and retrieval agent code is appended to the sections and subsections to create a plurality of finally transformed electronic documents. Users are allowed to access the finally transformed electronic documents. The users are provided with accountability reminders contemporaneous with the access. The access of the users to the sections and subsections of the finally transformed electronic documents is logged. An audit report is provided based on the logging. Also provided is a cloud service for enterprise-level sensitive data protection with variable data granularity, using one or more one guest virtual machine images.
Abstract:
A system and computer program product for enabling monitoring across one or more domains include identifying information technology resources across one or more domains, collecting information technology dependency information relating to the information technology resources identified across one or more domains, collecting monitoring data from the information technology resources identified across one or more domains, and using the dependency information to aggregate the monitoring data into one or more pre-determined metrics.
Abstract:
In an illustrative embodiment, a computing system comprises a first computer coupled to one or more additional computers. Each of the one or more additional computers manages a respective set of one or more virtualized computing resources. The first computer manages the one or more additional computers.
Abstract:
A semiconductor device includes a first layer, first and second active areas disposed on the first layer; a trench disposed between the first and second active areas, an insulating oxide that fills the trench to a level below a surface of the first and second active layers, and a nitride cap disposed on top of the insulating oxide so that the first and second active areas can be cleaned without damaging the insulating oxide. A top surface of the nitride cap in regions adjacent to the first and second active areas in aligned with a top surface of the first and second active areas, a top surface of the nitride cap in a center region of the nitride cap is stepped below the top surface of the adjacent regions, and a void is formed between the top surface regions adjacent to the first and second active areas.