摘要:
Content delivery by a network node is optimized. The network node is communicatively coupled between multiple end user devices and at least one content service provider. A request for a media streaming service available from the content service provider is intercepted from a first requestor device. It is determined if the media streaming service is registered for delivery optimization based on at least one service rule specific to the media streaming service. It is determined if a second requestor device is currently receiving the media streaming service from the content service provider. In response to the media streaming service being received by the second requestor device, delivery optimization of the media streaming service is performed by replicating the media streaming service for delivery to the first requestor device. The replicated media streaming service is sent to the first requestor device.
摘要:
Content delivery by a network node is optimized. The network node is communicatively coupled between multiple end user devices and at least one content service provider. A request for a media streaming service available from the content service provider is intercepted from a first requestor device. It is determined if the media streaming service is registered for delivery optimization based on at least one service rule specific to the media streaming service. It is determined if a second requestor device is currently receiving the media streaming service from the content service provider. In response to the media streaming service being received by the second requestor device, delivery optimization of the media streaming service is performed by replicating the media streaming service for delivery to the first requestor device. The replicated media streaming service is sent to the first requestor device.
摘要:
Content delivery by a network node in a network is optimized. The network node is communicatively coupled between multiple client devices and at least one content service provider. A media streaming service provided by a content service provider is received at the network node over a first unicast link. The service is transmitted from the network node to a first requestor device via a second unicast link. A request from a second requestor device for the service is intercepted by the network node. If it is determined that a multicast optimization threshold has been reached and/or exceeded, the service is transmitted from the network node to the first and second requestor devices using a single multicast link, while the service is received from the content service provider over the first unicast link.
摘要:
Content delivery by a network node in a network is optimized. The network node is communicatively coupled between multiple client devices and at least one content service provider. A media streaming service provided by a content service provider is received at the network node over a first unicast link. The service is transmitted from the network node to a first requestor device via a second unicast link. A request from a second requestor device for the service is intercepted by the network node. If it is determined that a multicast optimization threshold has been reached and/or exceeded, the service is transmitted from the network node to the first and second requestor devices using a single multicast link, while the service is received from the content service provider over the first unicast link.
摘要:
A dynamic predictive and/or exact caching mechanism is provided in various stages of a microprocessor pipeline so that various control signals can be stored and memorized in the course of program execution. Exact control signal vector caching may be done. Whenever an issue group is formed following instruction decode, register renaming, and dependency checking, an encoded copy of the issue group information can be cached under the tag of the leading instruction. The resulting dependency cache or control vector cache can be accessed right at the beginning of the instruction issue logic stage of the microprocessor pipeline the next time the corresponding group of instructions come up for re-execution. Since the encoded issue group bit pattern may be accessed in a single cycle out of the cache, the resulting microprocessor pipeline with this embodiment can be seen as two parallel pipes, where the shorter pipe is followed if there is a dependency cache or control vector cache hit.
摘要:
This invention pertains to apparatus, method and a computer program stored on a computer readable medium. The computer program includes instructions for use with an instruction unit having a code page, and has computer program code for partitioning the code page into at least two sections for storing in a first section thereof a plurality of instruction words and, in association with at least one instruction word, for storing in a second section thereof an extension to each instruction word in the first section. The computer program further includes computer program code for setting a state of at least one page table entry bit for indicating, on a code page by code page basis, whether the code page is partitioned into the first and second sections for storing instruction words and their extensions, or whether the code page is comprised instead of a single section storing only instruction words.
摘要:
A design structure embodied in a machine readable storage medium for at least one of designing, manufacturing, and testing a design is provided. The design structure includes a page table walker. The page table walker is moved from its conventional location in the memory management unit associated with the data processor to a location in main memory i.e. the main memory controller. As a result, wherein the processing of requests for data could selectively avoid or bypass cumbersome caches associated with the data processor.
摘要:
A method and a system for operating a plurality of processors that each includes an execution pipeline for processing dependence chains, the method comprising: configuring the plurality of processors to execute the dependence chains on execution pipelines; implementing a Super Re-Order Buffer (SuperROB) in which received instructions are re-ordered after out-of-order execution when at least one of the plurality of processors is in an Instruction Level Parallelism (ILP) mode and at least one of the plurality of processors has a Thread Level Parallelism (TLP) core; detecting an imbalance in a dispatch of instructions of a first dependence chain compared to a dispatch of instructions of a second dependence chain with respect to dependence chain priority; determining a source of the imbalance; and activating the ILP mode when the source of the imbalance has been determined.
摘要:
A single unified level one instruction cache in which some lines may contain traces and other lines in the same congruence class may contain blocks of instructions consistent with conventional cache lines. A mechanism is described for indexing into the cache, and selecting the desired line. Control is exercised over which lines are contained within the cache. Provision is made for selection between a trace line and a conventional line when both match during a tag compare step.
摘要:
A single unified level one instruction cache in which some lines may contain traces and other lines in the same congruence class may contain blocks of instructions consistent with conventional cache lines. Power is conserved by guiding access to lines stored in the cache and lowering cache clock speed relative to the central processor clock speed.