INPUT/OUTPUT STUTTER WAKE ALIGNMENT
    181.
    发明公开

    公开(公告)号:US20240004721A1

    公开(公告)日:2024-01-04

    申请号:US17853294

    申请日:2022-06-29

    CPC classification number: G06F9/5083 G06F9/5038 G06F9/5033 G06F9/5016

    Abstract: An apparatus and method for efficiently performing power management for a multi-client computing system. In various implementations, a computing system includes multiple clients that process tasks corresponding to applications. The clients store generated requests of a particular type while processing tasks. A client receives an indication specifying that another client is having requests of the particular type being serviced. In response to receiving this indication, the client inserts a first urgency level in one or more stored requests of the particular type prior to sending the requests for servicing. When the client determines a particular time interval has elapsed, the client sends an indication to other clients specifying that requests of the particular type are being serviced. The client also inserts a second urgency level different from the first urgency level in one or more stored requests of the particular type prior to sending the requests for servicing.

    ENCODED DATA DEPENDENCY MATRIX FOR POWER EFFICIENCY SCHEDULING

    公开(公告)号:US20240004657A1

    公开(公告)日:2024-01-04

    申请号:US17855621

    申请日:2022-06-30

    CPC classification number: G06F9/30145 G06F9/3838

    Abstract: The disclosed system may include a processor configured to encode, using an encoding scheme that reduces a number of bits needed to represent one or more instructions from a set of instructions in an instruction buffer represented by a dependency matrix, a dependency indicating that a child instruction represented in the dependency matrix depends on a parent instruction represented in the dependency matrix. The processor may also be configured to store the encoded dependency in the dependency matrix and dispatch instructions in the instruction buffer based at least on decoding one or more dependencies stored in the dependency matrix for the instructions. Various other methods, systems, and computer-readable media are also disclosed.

    ACCELERATING PREDICATED INSTRUCTION EXECUTION IN VECTOR PROCESSORS

    公开(公告)号:US20240004656A1

    公开(公告)日:2024-01-04

    申请号:US17853790

    申请日:2022-06-29

    CPC classification number: G06F9/30145 G06F9/3851 G06F9/3887

    Abstract: Methods and systems are disclosed for processing a vector by a vector processor. Techniques disclosed include receiving predicated instructions by a scheduler, each of which is associated with an opcode, a vector of elements, and a predicate. The techniques further include executing the predicated instructions. Executing a predicated instruction includes compressing, based on an index derived from a predicate of the instruction, elements in a vector of the instruction, where the elements in the vector are contiguously mapped, then, after the mapped elements are processed, decompressing the processed mapped elements, where the processed mapped elements are reverse mapped based on the index.

    PLATFORM EFFICIENCY TRACKER
    184.
    发明公开

    公开(公告)号:US20240004448A1

    公开(公告)日:2024-01-04

    申请号:US17853759

    申请日:2022-06-29

    CPC classification number: G06F1/28 G06F11/3062

    Abstract: Systems, apparatuses, and methods for dynamically estimating power losses in a computing system. A system management circuit tracks a state of a computing system and dynamically estimates power losses in the computing system based in part on the state. Based on the estimated power losses, power consumption of the computing system is estimated. In response to detecting reduced power losses in at least a portion of the computing system, the system management circuit is configured to increase a power-performance state of one or more circuits of the computing system while remaining within a power allocation limit of the computing system.

    Hardware assisted memory profiling aggregator

    公开(公告)号:US11860755B2

    公开(公告)日:2024-01-02

    申请号:US17861435

    申请日:2022-07-11

    CPC classification number: G06F11/3037 G06F11/076 G06F11/3471 G06F11/3476

    Abstract: An approach is provided for implementing memory profiling aggregation. A hardware aggregator provides memory profiling aggregation by controlling the execution of a plurality of hardware profilers that monitor memory performance in a system. For each hardware profiler of the plurality of hardware profilers, a hardware counter value is compared to a threshold value. When a threshold value is satisfied, execution of a respective hardware profiler of the plurality of hardware profilers is initiated to monitor memory performance. Multiple hardware profilers of the plurality of hardware profilers may execute concurrently and each generate a result counter value. The result counter values generated by each hardware profiler of the plurality of hardware profilers are aggregated to generate an aggregate result counter value. The aggregate result counter value is stored in memory that is accessible by a software processes for use in optimizing memory-management policy decisions.

    Sense amplifier sleep state for leakage savings without bias mismatch

    公开(公告)号:US11854652B2

    公开(公告)日:2023-12-26

    申请号:US17984796

    申请日:2022-11-10

    CPC classification number: G11C7/065

    Abstract: A sense amplifier is biased to reduce leakage current equalize matched transistor bias during an idle state. A first read select transistor couples a true bit line and a sense amplifier true (SAT) signal line and a second read select transistor couples a complement bit line and a sense amplifier complement (SAC) signal line. The SAT and SAC signal lines are precharged during a precharge state. An equalization circuit shorts the SAT and SAC signal lines during the precharge state. A differential sense amplifier circuit for latching the memory cell value is coupled to the SAT signal line and the SAC signal line. The precharge circuit and the differential sense amplifier circuit are turned off during a sleep state to cause the SAT and SAC signal lines to float. A sleep circuit shorts the SAT and SAC signal lines during the sleep state.

    Neural Network Activation Scaled Clipping Layer

    公开(公告)号:US20230409868A1

    公开(公告)日:2023-12-21

    申请号:US17844204

    申请日:2022-06-20

    CPC classification number: G06N3/04 G06N3/08

    Abstract: Activation scaled clipping layers for neural networks are described. An activation scaled clipping layer processes an output of a neuron in a neural network using a scaling parameter and a clipping parameter. The scaling parameter defines how numerical values are amplified relative to zero. The clipping parameter specifies a numerical threshold that causes the neuron output to be expressed as a value defined by the numerical threshold if the neuron output satisfies the numerical threshold. In some implementations, the scaling parameter is linear and treats numbers within a numerical range as being equivalent, such that any number in the range is scaled by a defined magnitude, regardless of value. Alternatively, the scaling parameter is nonlinear, which causes the activation scaled clipping layer to amplify numbers within a range by different magnitudes. Each scaling and clipping parameter is learnable during training of a machine learning model implementing the neural network.

    Approach for reducing side effects of computation offload to memory

    公开(公告)号:US11847055B2

    公开(公告)日:2023-12-19

    申请号:US17364854

    申请日:2021-06-30

    CPC classification number: G06F12/084 G06F12/0238 G06F12/0811 G06F12/0862

    Abstract: A technical solution to the technical problem of how to reduce the undesirable side effects of offloading computations to memory uses read hints to preload results of memory-side processing into a processor-side cache. A cache controller, in response to identifying a read hint in a memory-side processing instruction, causes results of the memory-side processing to be preloaded into a processor-side cache. Implementations include, without limitation, enabling or disabling the preloading based upon cache thrashing levels, preloading results, or portions of results, of memory-side processing to particular destination caches, preloading results based upon priority and/or degree of confidence, and/or during periods of low data bus and/or command bus utilization, last stores considerations, and enforcing an ordering constraint to ensure that preloading occurs after memory-side processing results are complete.

Patent Agency Ranking