摘要:
Devices and techniques for out-of-band platform tuning and configuration are described herein. A device can include a telemetry interface to a telemetry collection system and a network interface to network adapter hardware. The device can receive platform telemetry metrics from the telemetry collection system, and network adapter silicon hardware statistics over the network interface, to gather collected statistics. The device can apply a heuristic algorithm using the collected statistics to determine processing core workloads generated by operation of a plurality of software systems communicatively coupled to the device. The device can provide a reconfiguration message to instruct at least one software system to switch operations to a different processing core, responsive to detecting an overload state on at least one processing core, based on the processing core workloads. Other embodiments are also described.
摘要:
In embodiments, apparatuses, methods and storage media (transitory and non-transitory) are described that are associated with end-to-end datacenter performance control. In various embodiments, an apparatus for computing may receive a datacenter performance target, determine an end-to-end datacenter performance level based at least in part on quality of service data collected from a plurality of nodes, and send a mitigation command based at least in part on a result of a comparison of the end-to-end datacenter performance level determined to the datacenter performance target. In various embodiments, the apparatus for computing may include one or more processors, a memory, a datacenter performance monitor to receive a datacenter performance target corresponding to a service level agreement, and a mitigation module to send a mitigation command based at least in part on a result of a comparison of an end-to-end datacenter performance level to a datacenter performance target.
摘要:
In embodiments, apparatuses, methods and storage media (transitory and non-transitory) are described that are associated with end-to-end datacenter performance control. In various embodiments, an apparatus for computing may receive a datacenter performance target, determine an end-to-end datacenter performance level based at least in part on quality of service data collected from a plurality of nodes, and send a mitigation command based at least in part on a result of a comparison of the end-to-end datacenter performance level determined to the datacenter performance target. In various embodiments, the apparatus for computing may include one or more processors, a memory, a datacenter performance monitor to receive a datacenter performance target corresponding to a service level agreement, and a mitigation module to send a mitigation command based at least in part on a result of a comparison of an end-to-end datacenter performance level to a datacenter performance target.
摘要:
Technologies for identifying service functions that may be performed in parallel in a service function chain include a computing device for running one or more virtual machines for each of a plurality of service functions based on a preferred service function chain being selected. To identify which service functions may be performed in parallel, the computing device may determine which service functions are not required to be performed on a critical path of the service function chain and/or which service functions are not required to be performed in real-time. Additionally, selecting the preferred service function chain may be based on selection criteria.
摘要:
Methods and apparatus implementing Hardware/Software co-optimization to improve performance and energy for inter-VM communication for NFVs and other producer-consumer workloads. The apparatus include multi-core processors with multi-level cache hierarchies including and L1 and L2 cache for each core and a shared last-level cache (LLC). One or more machine-level instructions are provided for proactively demoting cachelines from lower cache levels to higher cache levels, including demoting cachelines from L1/L2 caches to an LLC. Techniques are also provided for implementing hardware/software co-optimization in multi-socket NUMA architecture system, wherein cachelines may be selectively demoted and pushed to an LLC in a remote socket. In addition, techniques are disclosure for implementing early snooping in multi-socket systems to reduce latency when accessing cachelines on remote sockets.
摘要:
Technologies for secure inter-virtual-machine shared memory communication include a computing device with hardware virtualization support. A virtual machine monitor (VMM) authenticates a view switch component of a target virtual machine. The VMM adds configures a secure memory view to access a shared memory segment. The shared memory segment may include memory pages of a source virtual machine or the VMM. The view switch component switches to the secure memory view without generating a virtual machine exit event, using the hardware virtualization support. The view switch component may switch to the secure memory view by modifying an extended page table (EPT) pointer. The target virtual machine accesses the shared memory segment via the secure memory view. The target virtual machine and the source virtual machine may coordinate ownership of memory pages using a secure view control structure stored in the shared memory segment. Other embodiments are described and claimed.
摘要:
Technologies to monitor and manage platform, device, processor and power characteristics throughout a system utilizing a remote entity such as controller node. By remotely monitoring and managing system operation and performance over time, future system performance requirements may be anticipated, allowing system parameters to be adjusted proactively in a more coordinated way. The controller node may monitor, control and predict traffic flows in the system and provide performance modification instructions to any of the computer nodes and a network switch to better optimize performance. The target systems collaborate with the controller node by respectively monitoring internal resources, such as resource availability and performance requirements to provide necessary resources for optimizing operating parameters of the system. The controller node may collect local system information for one or all of the computer nodes to dynamically steer traffic to a specific set of computers for processing to meet desired performance and power requirements.
摘要:
Embodiments of the present disclosure provide for determination of transmit power allocations and modulation and coding schemes for multiuser orthogonal frequency division multiple access downlink transmissions. Other embodiments may be described and claimed.
摘要:
Embodiments of a station (STA) and method of communication are generally described herein. The STA may be included in a first plurality of STAs affiliated with a first multi-link logical entity (MLLE). A plurality of links may be established between the first MLLE and a second MLLE, wherein the second MLLE may be affiliated with a second plurality of STAs. The STA may receive a first subset of a sequence of MAC protocol data units (MPDUs). A second subset of the sequence of MPDUs may be transmitted by another STA of the first plurality of STAs. The STA may transmit a block acknowledgement (BA) frame that includes: a number of BA bitmaps, configurable to values greater than or equal to one; and BA control information for each of the BA bitmaps.
摘要:
Embodiments of an access point (AP), station (STA), and method of communication are generally described herein. The AP may perform channel sensing in a first channel of a first frequency band and a second channel of a second frequency band to obtain access for transmission of a physical layer convergence procedure (PLCP) protocol data unit (PPDU) that comprises multiple subframes. The transmission of the PPDU may be configurable to use a multi-band layer-1 aggregation of the first and second channels. While the second channel is unavailable, the AP may transmit one or more subframes on the first channel without usage of the multi-band layer-1 aggregation. When it is determined that the second channel has become available, the AP may switch to synchronized transmission of the subframes on the first and second channels in accordance with the multi-band layer-1 aggregation.