-
11.
公开(公告)号:US20200310865A1
公开(公告)日:2020-10-01
申请号:US16370248
申请日:2019-03-29
Applicant: Intel Corporation
Inventor: Andrew Herdrich , Ian Steiner , Leeor Peled , Michael Prinke , Eylon Toledano
Abstract: A simultaneous multi-threading (SMT) processor core capable of thread-based biasing with respect to execution resources. The SMT processor includes priority controller circuitry to determine a thread priority value for each of a plurality of threads to be executed by the SMT processor core and to generate a priority vector comprising the thread priority value of each of the plurality of threads. The SMT processor further includes thread selector circuitry to make execution cycle assignments of a pipeline by assigning to each of the plurality of threads a portion of the pipeline's execution cycles based on each thread's priority value in the priority vector. The thread selector circuitry is further to select, from the plurality of threads, tasks to be processed by the pipeline based on the execution cycle assignments.
-
12.
公开(公告)号:US20150095675A1
公开(公告)日:2015-04-02
申请号:US14564436
申请日:2014-12-09
Applicant: INTEL CORPORATION
Inventor: Andrew Herdrich , Ramesh Illikkal , Donald Newell , Ravishankar Iyer , Vineet Chadha
IPC: G06F1/32
CPC classification number: G06F9/44505 , G06F1/10 , G06F1/32 , G06F1/3203 , G06F1/3206 , G06F1/3234 , G06F1/3237 , G06F1/324 , G06F1/3296 , G06F9/50 , G06F9/5094 , G06F12/0891 , Y02D10/128
Abstract: An apparatus comprises a plurality of cores and a controller coupled to the cores. The controller is to lower an operating point of a first core if a first number based on processor clock cycles per instruction (CPI) associated with a second core is higher than a first threshold. The controller is operable to increase the operating point of the first core if the first number is lower than a second threshold.
Abstract translation: 一种装置包括多个核心和耦合到核心的控制器。 如果基于与第二核心相关联的每个指令(CPI)的处理器时钟周期的第一数量高于第一阈值,则控制器将降低第一核心的工作点。 如果第一数量低于第二阈值,则控制器可操作以增加第一核心的工作点。
-
公开(公告)号:US12229069B2
公开(公告)日:2025-02-18
申请号:US17083200
申请日:2020-10-28
Applicant: Intel Corporation
Inventor: Pratik Marolia , Andrew Herdrich , Rajesh Sankaran , Rahul Pal , David Puffer , Sayantan Sur , Ajaya Durg
Abstract: Methods and apparatus for an accelerator controller hub (ACH). The ACH may be a stand-alone component or integrated on-die or on package in an accelerator such as a GPU. The ACH may include a host device link (HDL) interface, one or more Peripheral Component Interconnect Express (PCIe) interfaces, one or more high performance accelerator link (HPAL) interfaces, and a router, operatively coupled to each of the HDL interface, the one or more PCIe interfaces, and the one or more HPAL interfaces. The HDL interface is configured to be coupled to a host CPU via an HDL link and the one or more HPAL interfaces are configured to be coupled to one or more HPALs that are used to access high performance accelerator fabrics (HPAFs) such as NVlink fabrics and CCIX (Cache Coherent Interconnect for Accelerators) fabrics. Platforms including ACHs or accelerators with integrated ACHs support RDMA transfers using RDMA semantics to enable transfers between accelerator memory on initiators and targets without CPU involvement.
-
公开(公告)号:US12210434B2
公开(公告)日:2025-01-28
申请号:US16914305
申请日:2020-06-27
Applicant: Intel Corporation
Inventor: Bin Li , Ren Wang , Kshitij Arun Doshi , Francesc Guim Bernat , Yipeng Wang , Ravishankar Iyer , Andrew Herdrich , Tsung-Yuan Tai , Zhu Zhou , Rasika Subramanian
Abstract: An apparatus and method for closed loop dynamic resource allocation. For example, one embodiment of a method comprises: collecting data related to usage of a plurality of resources by a plurality of workloads over one or more time periods, the workloads including priority workloads associated with one or more guaranteed performance levels and best effort workloads not associated with guaranteed performance levels; analyzing the data to identify resource reallocations from one or more of the priority workloads to one or more of the best effort workloads in one or more subsequent time periods while still maintaining the guaranteed performance levels; reallocating the resources from the priority workloads to the best effort workloads for the subsequent time periods; monitoring execution of the priority workloads with respect to the guaranteed performance level during the subsequent time periods; and preemptively reallocating resources from the best effort workloads to the priority workloads during the subsequent time periods to ensure compliance with the guaranteed performance level and responsive to detecting that the guaranteed performance level is in danger of being breached.
-
公开(公告)号:US12093100B2
公开(公告)日:2024-09-17
申请号:US17033753
申请日:2020-09-26
Applicant: Intel Corporation
Inventor: Vivek Garg , Ankush Varma , Krishnakanth Sistla , Nikhil Gupta , Nikethan Shivanand Baligar , Stephen Wang , Nilanjan Palit , Timothy Yee-Kwong Kam , Adwait Purandare , Ujjwal Gupta , Stanley Chen , Dorit Shapira , Shruthi Venugopal , Suresh Chemudupati , Rupal Parikh , Eric Dehaemer , Pavithra Sampath , Phani Kumar Kandula , Yogesh Bansal , Dean Mulla , Michael Tulanowski , Stephen Paul Haake , Andrew Herdrich , Ripan Das , Nazar Syed Haider , Aman Sewani
CPC classification number: G06F1/28 , G06F1/30 , G06F13/20 , G06F2213/40
Abstract: Hierarchical Power Management (HPM) architecture considers the limits of scaling on a power management controller, the autonomy at each die, and provides a unified view of the package to a platform. At a simplest level, HPM architecture has a supervisor and one or more supervisee power management units (PMUs) that communicate via at least two different communication fabrics. Each PMU can behave as a supervisor for a number of supervisee PMUs in a particular domain. HPM addresses these needs for products that comprise a collection of dice with varying levels of power and thermal management capabilities and needs. HPM serves as a unified mechanism than can span collection of dice of varying capability and function, which together form a traditional system-on-chip (SoC). HPM provides a basis for managing power and thermals across a diverse set of dice.
-
公开(公告)号:US11531562B2
公开(公告)日:2022-12-20
申请号:US17077796
申请日:2020-10-22
Applicant: Intel Corporation
Inventor: Matthew Fleming , Edwin Verplanke , Andrew Herdrich , Ravishankar Iyer
Abstract: Systems, methods, and apparatuses for resource monitoring identification reuse are described. In an embodiment, a system comprising a hardware processor core to execute instructions storage for a resource monitoring identification (RMID) recycling instructions to be executed by a hardware processor core, a logical processor to execute on the hardware processor core, the logical processor including associated storage for a RMID and state, are described.
-
17.
公开(公告)号:US11121940B2
公开(公告)日:2021-09-14
申请号:US15470664
申请日:2017-03-27
Applicant: Intel Corporation
Inventor: Francesc Guim Bernat , Karthik Kumar , Thomas Willhalm , Andrew Herdrich , Edwin Verplanke
IPC: H04L12/851 , H04L12/701 , H04L12/911 , H04L12/24 , H04L12/947 , H04L5/00
Abstract: Examples include techniques to meet quality of service (QoS) requirements for a fabric point to point connection. Examples include an application hosted by a compute node coupled with a fabric requesting bandwidth for a point to point connection through the fabric and the request being granted or not granted based at least partially on whether bandwidth is available for allocation to meet one or more QoS requirements.
-
公开(公告)号:US10929323B2
公开(公告)日:2021-02-23
申请号:US16601137
申请日:2019-10-14
Applicant: Intel Corporation
Inventor: Ren Wang , Yipeng Wang , Andrew Herdrich , Jr-Shian Tsai , Tsung-Yuan C. Tai , Niall D. McDonnell , Hugh Wilkinson , Bradley A. Burres , Bruce Richardson , Namakkal N. Venkatesan , Debra Bernstein , Edwin Verplanke , Stephen R. Van Doren , An Yan , Andrew Cunningham , David Sonnier , Gage Eads , James T. Clee , Jamison D. Whitesell , Jerry Pirog , Jonathan Kenny , Joseph R. Hasting , Narender Vangati , Stephen Miller , Te K. Ma , William Burroughs
IPC: G06F13/37 , G06F9/54 , G06F12/0868 , G06F12/0811 , G06F13/16 , G06F12/04 , G06F9/38
Abstract: Apparatus and methods implementing a hardware queue management device for reducing inter-core data transfer overhead by offloading request management and data coherency tasks from the CPU cores. The apparatus include multi-core processors, a shared L3 or last-level cache (“LLC”), and a hardware queue management device to receive, store, and process inter-core data transfer requests. The hardware queue management device further comprises a resource management system to control the rate in which the cores may submit requests to reduce core stalls and dropped requests. Additionally, software instructions are introduced to optimize communication between the cores and the queue management device.
-
公开(公告)号:US20160077844A1
公开(公告)日:2016-03-17
申请号:US14947321
申请日:2015-11-20
Applicant: INTEL CORPORATION
Inventor: Andrew Herdrich , Ramesh Illikkal , Donald Newell , Ravishankar Iyer , Vineet Chadha
CPC classification number: G06F9/44505 , G06F1/10 , G06F1/32 , G06F1/3203 , G06F1/3206 , G06F1/3234 , G06F1/3237 , G06F1/324 , G06F1/3296 , G06F9/50 , G06F9/5094 , G06F12/0891 , Y02D10/128
Abstract: An apparatus comprises a plurality of cores and a controller coupled to the cores. The controller is to lower an operating point of a first core if a first number based on processor clock cycles per instruction (CPI) associated with a second core is higher than a first threshold. The controller is operable to increase the operating point of the first core if the first number is lower than a second threshold.
-
公开(公告)号:US12198186B2
公开(公告)日:2025-01-14
申请号:US17401575
申请日:2021-08-13
Applicant: Intel Corporation
Inventor: Andrew Herdrich , Edwin Verplanke , Ravishankar Iyer , Christopher Gianos , Jeffrey D. Chamberlain , Ronak Singhal , Julius Mandelblat , Bret Toll
IPC: G06Q40/03 , G06F12/0875 , G06F12/0897
Abstract: Systems, methods, and apparatuses for resource bandwidth monitoring and control are described. For example, in some embodiments, an apparatus comprising a requestor device to send a credit based request, a receiver device to receive and consume the credit based request, and a delay element in a return path between the requestor and receiver devices, the delay element to delay a credit based response from the receiver to the requestor are detailed.
-
-
-
-
-
-
-
-
-