Multi-die system performance optimization

    公开(公告)号:US11709536B2

    公开(公告)日:2023-07-25

    申请号:US17029852

    申请日:2020-09-23

    CPC classification number: G06F1/28 G05F1/625

    Abstract: A multi-die semiconductor package includes a first integrated circuit (IC) die having a first intrinsic performance level and a second IC die having a second intrinsic performance level different from the first intrinsic performance level. A power management controller distributes, based on a determined die performance differential between the first IC die and the second IC die, a level of power allocated to the semiconductor chip package between the first IC die and the second IC die. In this manner, the first IC die receives and operates at a first level of power resulting in performance exceeding its intrinsic performance level. The second IC die receives and operates at a second level of power resulting in performance below its intrinsic performance level, thereby reducing performance differentials between the IC dies.

    Die-stacked memory device with reconfigurable logic
    4.
    发明授权
    Die-stacked memory device with reconfigurable logic 有权
    具有可重构逻辑的堆叠式存储器件

    公开(公告)号:US08922243B2

    公开(公告)日:2014-12-30

    申请号:US13726145

    申请日:2012-12-23

    Abstract: A die-stacked memory device incorporates a reconfigurable logic device to provide implementation flexibility in performing various data manipulation operations and other memory operations that use data stored in the die-stacked memory device or that result in data that is to be stored in the die-stacked memory device. One or more configuration files representing corresponding logic configurations for the reconfigurable logic device can be stored in a configuration store at the die-stacked memory device, and a configuration controller can program a reconfigurable logic fabric of the reconfigurable logic device using a selected one of the configuration files. Due to the integration of the logic dies and the memory dies, the reconfigurable logic device can perform various data manipulation operations with higher bandwidth and lower latency and power consumption compared to devices external to the die-stacked memory device.

    Abstract translation: 芯片堆叠的存储器件包括可重构逻辑器件,以在执行各种数据操作操作和使用存储在管芯堆叠的存储器件中的数据的其他存储器操作中提供实现灵活性,或者导致要存储在管芯堆叠存储器件中的数据。 堆叠式存储设备。 代表可重配置逻辑器件的相应逻辑配置的一个或多个配置文件可被存储在管芯堆叠的存储器件的配置存储器中,并且配置控制器可使用所选择的一个存储器件对可重新配置的逻辑器件进行编程 配置文件。 由于逻辑管芯和存储器管芯的集成,与可堆叠存储器件外部的器件相比,可重构逻辑器件可以执行具有更高带宽和更低延迟和功耗的各种数据操作操作。

    Computation Memory Operations in a Logic Layer of a Stacked Memory
    6.
    发明申请
    Computation Memory Operations in a Logic Layer of a Stacked Memory 有权
    堆叠存储器逻辑层中的计算存储器操作

    公开(公告)号:US20140181483A1

    公开(公告)日:2014-06-26

    申请号:US13724506

    申请日:2012-12-21

    CPC classification number: G06F15/7821 Y02D10/12 Y02D10/13

    Abstract: Some die-stacked memories will contain a logic layer in addition to one or more layers of DRAM (or other memory technology). This logic layer may be a discrete logic die or logic on a silicon interposer associated with a stack of memory dies. Additional circuitry/functionality is placed on the logic layer to implement functionality to perform various computation operations. This functionality would be desired where performing the operations locally near the memory devices would allow increased performance and/or power efficiency by avoiding transmission of data across the interface to the host processor.

    Abstract translation: 除了一层或多层DRAM(或其他存储器技术)之外,一些堆叠堆叠的存储器将包含逻辑层。 该逻辑层可以是与存储器管芯堆叠相关联的硅插入器上的离散逻辑管芯或逻辑。 额外的电路/功能被放置在逻辑层上以实现执行各种计算操作的功能。 通过避免通过接口向主机处理器传输数据,执行本地在存储器件附近的操作将允许提高性能和/或功率效率将需要该功能。

    Compound Memory Operations in a Logic Layer of a Stacked Memory
    7.
    发明申请
    Compound Memory Operations in a Logic Layer of a Stacked Memory 审中-公开
    堆叠存储器的逻辑层中的复合存储器操作

    公开(公告)号:US20140181427A1

    公开(公告)日:2014-06-26

    申请号:US13724338

    申请日:2012-12-21

    CPC classification number: G06F9/3004 G06F9/3455 G06F15/7821

    Abstract: Some die-stacked memories will contain a logic layer in addition to one or more layers of DRAM (or other memory technology). This logic layer may be a discrete logic die or logic on a silicon interposer associated with a stack of memory dies. Additional circuitry/functionality is placed on the logic layer to implement functionality to perform various data movement and address calculation operations. This functionality would allow compound memory operations—a single request communicated to the memory that characterizes the accesses and movement of many data items. This eliminates the performance and power overheads associated with communicating address and control information on a fine-grain, per-data-item basis from a host processor (or other device) to the memory. This approach also provides better visibility of macro-level memory access patterns to the memory system and may enable additional optimizations in scheduling memory accesses.

    Abstract translation: 除了一层或多层DRAM(或其他存储器技术)之外,一些堆叠堆叠的存储器将包含逻辑层。 该逻辑层可以是与存储器管芯堆叠相关联的硅插入器上的离散逻辑管芯或逻辑。 额外的电路/功能被放置在逻辑层上以实现执行各种数据移动和地址计算操作的功能。 该功能将允许复合存储器操作 - 传达到存储器的单个请求,其表征许多数据项的访问和移动。 这消除了与从主处理器(或其他设备)到存储器的以细粒度,每数据项为基础传送地址和控制信息相关联的性能和功耗开销。 这种方法还提供了对存储器系统的宏级存储器访问模式的更好的可见性,并且可以在调度存储器访问中实现附加优化。

    PROCESSING ENGINE FOR COMPLEX ATOMIC OPERATIONS
    8.
    发明申请
    PROCESSING ENGINE FOR COMPLEX ATOMIC OPERATIONS 有权
    加工发动机用于复杂的原子操作

    公开(公告)号:US20140181421A1

    公开(公告)日:2014-06-26

    申请号:US13725724

    申请日:2012-12-21

    CPC classification number: G06F9/50 G06F9/526 G06F2209/521 G06F2209/522

    Abstract: A system includes an atomic processing engine (APE) coupled to an interconnect. The interconnect is to couple to one or more processor cores. The APE receives a plurality of commands from the one or more processor cores through the interconnect. In response to a first command, the APE performs a first plurality of operations associated with the first command. The first plurality of operations references multiple memory locations, at least one of which is shared between two or more threads executed by the one or more processor cores.

    Abstract translation: 系统包括耦合到互连的原子处理引擎(APE)。 互连将耦合到一个或多个处理器内核。 APE通过互连从一个或多个处理器核接收多个命令。 响应于第一命令,APE执行与第一命令相关联的第一多个操作。 第一组多个操作引用多个存储器位置,其中至少一个在一个或多个处理器核心执行的两个或多个线程之间共享。

Patent Agency Ranking