Elastic request handling technique for optimizing workload performance

    公开(公告)号:US12204755B2

    公开(公告)日:2025-01-21

    申请号:US17853123

    申请日:2022-06-29

    Applicant: Nutanix, Inc.

    Abstract: An elastic request handling technique limits a number of threads used to service input/output (I/O) requests of a low-latency I/O workload received by a file system server executing on a cluster having a plurality of nodes deployed in a virtualization environment. The limited number of threads (server threads) is constantly maintained as “active” and running on virtual central processing units (vCPUs) of a node. The file system server spawns and organizes the active server threads as one or more pools of threads. The server prioritizes the low-latency I/O requests by loading them onto the active threads and allowing the requests to run on those active threads to completion, thereby obviating overhead associated with lock contention and vCPU migration after a context switch (i.e., to avoid rescheduling a thread on a different vCPU after execution of the thread was suspended).

    ELASTIC REQUEST HANDLING TECHNIQUE FOR OPTIMIZING WORKLOAD PERFORMANCE

    公开(公告)号:US20230359359A1

    公开(公告)日:2023-11-09

    申请号:US17853123

    申请日:2022-06-29

    Applicant: Nutanix, Inc.

    CPC classification number: G06F3/0611 G06F3/0655 G06F3/0679

    Abstract: An elastic request handling technique limits a number of threads used to service input/output (I/O) requests of a low-latency I/O workload received by a file system server executing on a cluster having a plurality of nodes deployed in a virtualization environment. The limited number of threads (server threads) is constantly maintained as “active” and running on virtual central processing units (vCPUs) of a node. The file system server spawns and organizes the active server threads as one or more pools of threads. The server prioritizes the low-latency I/O requests by loading them onto the active threads and allowing the requests to run on those active threads to completion, thereby obviating overhead associated with lock contention and vCPU migration after a context switch (i.e., to avoid rescheduling a thread on a different vCPU after execution of the thread was suspended).

Patent Agency Ranking