-
公开(公告)号:US12204755B2
公开(公告)日:2025-01-21
申请号:US17853123
申请日:2022-06-29
Applicant: Nutanix, Inc.
Inventor: Daniel Chilton , Gaurav Gangalwar , Manoj Premanand Naik , Pradeep Thomas , Will Strickland
IPC: G06F3/06
Abstract: An elastic request handling technique limits a number of threads used to service input/output (I/O) requests of a low-latency I/O workload received by a file system server executing on a cluster having a plurality of nodes deployed in a virtualization environment. The limited number of threads (server threads) is constantly maintained as “active” and running on virtual central processing units (vCPUs) of a node. The file system server spawns and organizes the active server threads as one or more pools of threads. The server prioritizes the low-latency I/O requests by loading them onto the active threads and allowing the requests to run on those active threads to completion, thereby obviating overhead associated with lock contention and vCPU migration after a context switch (i.e., to avoid rescheduling a thread on a different vCPU after execution of the thread was suspended).
-
公开(公告)号:US20230359359A1
公开(公告)日:2023-11-09
申请号:US17853123
申请日:2022-06-29
Applicant: Nutanix, Inc.
Inventor: Daniel Chilton , Gaurav Gangalwar , Manoj Premanand Naik , Pradeep Thomas , Will Strickland
IPC: G06F3/06
CPC classification number: G06F3/0611 , G06F3/0655 , G06F3/0679
Abstract: An elastic request handling technique limits a number of threads used to service input/output (I/O) requests of a low-latency I/O workload received by a file system server executing on a cluster having a plurality of nodes deployed in a virtualization environment. The limited number of threads (server threads) is constantly maintained as “active” and running on virtual central processing units (vCPUs) of a node. The file system server spawns and organizes the active server threads as one or more pools of threads. The server prioritizes the low-latency I/O requests by loading them onto the active threads and allowing the requests to run on those active threads to completion, thereby obviating overhead associated with lock contention and vCPU migration after a context switch (i.e., to avoid rescheduling a thread on a different vCPU after execution of the thread was suspended).
-