-
公开(公告)号:US11296681B2
公开(公告)日:2022-04-05
申请号:US16726020
申请日:2019-12-23
Applicant: Intel Corporation
Inventor: Amit Agarwal , Steven Hsu , Simeon Realov , Mahesh Kumashikar , Ram Krishnamurthy
IPC: H03K3/00 , H03K3/037 , G01R31/3177 , H03K3/038 , H03K19/20
Abstract: A fast Mux-D scan flip-flop is provided, which bypasses a scan multiplexer to a master keeper side path, removing delay overhead of a traditional Mux-D scan topology. The design is compatible with simple scan methodology of Mux-D scan, while preserving smaller area and small number of inputs/outputs. Since scan Mux is not in the forward critical path, circuit topology has similar high performance as level-sensitive scan flip-flop and can be easily converted into bare pass-gate version. The new fast Mux-D scan flip-flop combines the advantages of the conventional LSSD and Mux-D scan flip-flop, without the disadvantages of each.
-
公开(公告)号:US11151046B2
公开(公告)日:2021-10-19
申请号:US16921685
申请日:2020-07-06
Applicant: Intel Corporation
Inventor: Amrita Mathuriya , Sasikanth Manipatruni , Victor Lee , Huseyin Sumbul , Gregory Chen , Raghavan Kumar , Phil Knag , Ram Krishnamurthy , Ian Young , Abhishek Sharma
Abstract: The present disclosure is directed to systems and methods of implementing a neural network using in-memory mathematical operations performed by pipelined SRAM architecture (PISA) circuitry disposed in on-chip processor memory circuitry. A high-level compiler may be provided to compile data representative of a multi-layer neural network model and one or more neural network data inputs from a first high-level programming language to an intermediate domain-specific language (DSL). A low-level compiler may be provided to compile the representative data from the intermediate DSL to multiple instruction sets in accordance with an instruction set architecture (ISA), such that each of the multiple instruction sets corresponds to a single respective layer of the multi-layer neural network model. Each of the multiple instruction sets may be assigned to a respective SRAM array of the PISA circuitry for in-memory execution. Thus, the systems and methods described herein beneficially leverage the on-chip processor memory circuitry to perform a relatively large number of in-memory vector/tensor calculations in furtherance of neural network processing without burdening the processor circuitry.
-
公开(公告)号:US20210263100A1
公开(公告)日:2021-08-26
申请号:US17240877
申请日:2021-04-26
Applicant: Intel Corporation
Inventor: Amit Agarwal , Ram Krishnamurthy , Satish Damaraju , Steven Hsu , Simeon Realov
IPC: G01R31/317 , G01R31/3177 , H03K3/037 , G01R31/3185
Abstract: An apparatus is provided which comprises: a multi-bit quad latch with an internally coupled level sensitive scan circuitry; and a combinational logic coupled to an output of the multi-bit quad latch. Another apparatus is provided which comprises: a plurality of sequential logic circuitries; and a clocking circuitry comprising inverters, wherein the clocking circuitry is shared by the plurality of sequential logic circuitries.
-
公开(公告)号:US20210194469A1
公开(公告)日:2021-06-24
申请号:US16726020
申请日:2019-12-23
Applicant: Intel Corporation
Inventor: Amit Agarwal , Steven Hsu , Simeon Realov , Mahesh Kumashikar , Ram Krishnamurthy
IPC: H03K3/037 , H03K19/20 , H03K3/038 , G01R31/3177
Abstract: A fast Mux-D scan flip-flop is provided, which bypasses a scan multiplexer to a master keeper side path, removing delay overhead of a traditional Mux-D scan topology. The design is compatible with simple scan methodology of Mux-D scan, while preserving smaller area and small number of inputs/outputs. Since scan Mux is not in the forward critical path, circuit topology has similar high performance as level-sensitive scan flip-flop and can be easily converted into bare pass-gate version. The new fast Mux-D scan flip-flop combines the advantages of the conventional LSSD and Mux-D scan flip-flop, without the disadvantages of each.
-
公开(公告)号:US20210117197A1
公开(公告)日:2021-04-22
申请号:US17132895
申请日:2020-12-23
Applicant: Intel Corporation
Inventor: Steven Hsu , Amit Agarwal , Debabrata Mohapatra , Arnab Raha , Moongon Jung , Gautham Chinya , Ram Krishnamurthy
Abstract: Systems, apparatuses and methods identify a plurality of registers that are associated with a system-on-chip. The plurality of registers includes a first portion dedicated to write operations and a second portion dedicated to read operations. The technology writes data to the first portion of the plurality of registers, and transfers the data from the first portion to the second portion.
-
公开(公告)号:US10860682B2
公开(公告)日:2020-12-08
申请号:US16839013
申请日:2020-04-02
Applicant: Intel Corporation
Inventor: Phil Knag , Gregory K. Chen , Raghavan Kumar , Huseyin Ekin Sumbul , Abhishek Sharma , Sasikanth Manipatruni , Amrita Mathuriya , Ram Krishnamurthy , Ian A. Young
IPC: G06F17/16 , G11C11/419 , G11C11/418 , G11C7/12 , G11C8/08 , G06G7/16 , G06G7/22 , G11C11/56 , G06F9/30 , G11C7/10 , G06N3/063
Abstract: A binary CIM circuit enables all memory cells in a memory array to be effectively accessible simultaneously for computation using fixed pulse widths on the wordlines and equal capacitance on the bitlines. The fixed pulse widths and equal capacitance ensure that a minimum voltage drop in the bitline represents one least significant bit (LSB) so that the bitline voltage swing remains safely within the maximum allowable range. The binary CIM circuit maximizes the effective memory bandwidth of a memory array for a given maximum voltage range of bitline voltage.
-
公开(公告)号:US20200334161A1
公开(公告)日:2020-10-22
申请号:US16921685
申请日:2020-07-06
Applicant: Intel Corporation
Inventor: Amrita Mathuriya , Sasikanth Manipatruni , Victor Lee , Huseyin Sumbul , Gregory Chen , Raghavan Kumar , Phil Knag , Ram Krishnamurthy , IAN YOUNG , Abhishek Sharma
Abstract: The present disclosure is directed to systems and methods of implementing a neural network using in-memory mathematical operations performed by pipelined SRAM architecture (PISA) circuitry disposed in on-chip processor memory circuitry. A high-level compiler may be provided to compile data representative of a multi-layer neural network model and one or more neural network data inputs from a first high-level programming language to an intermediate domain-specific language (DSL). A low-level compiler may be provided to compile the representative data from the intermediate DSL to multiple instruction sets in accordance with an instruction set architecture (ISA), such that each of the multiple instruction sets corresponds to a single respective layer of the multi-layer neural network model. Each of the multiple instruction sets may be assigned to a respective SRAM array of the PISA circuitry for in-memory execution. Thus, the systems and methods described herein beneficially leverage the on-chip processor memory circuitry to perform a relatively large number of in-memory vector/tensor calculations in furtherance of neural network processing without burdening the processor circuitry.
-
公开(公告)号:US10565138B2
公开(公告)日:2020-02-18
申请号:US16146534
申请日:2018-09-28
Applicant: Intel Corporation
Inventor: Jack Kavalieros , Ram Krishnamurthy , Sasikanth Manipatruni , Gregory Chen , Van Le , Amrita Mathuriya , Abhishek Sharma , Raghavan Kumar , Phil Knag , Huseyin Sumbul , Ian Young
IPC: G11C8/00 , G06F13/16 , H01L25/18 , H03K19/21 , G11C11/408 , H01L23/522 , G11C11/419
Abstract: Techniques and mechanisms for providing data to be used in an in-memory computation at a memory device. In an embodiment a memory device comprises a first memory array and circuitry, coupled to the first memory array, to perform a data computation based on data stored at the first memory array. Prior to the computation, the first memory array receives the data from a second memory array of the memory device. The second memory array extends horizontally in parallel with, but is offset vertically from, the first memory array. In another embodiment, a single integrated circuit die includes both the first memory array and the second memory array.
-
公开(公告)号:US10248906B2
公开(公告)日:2019-04-02
申请号:US15392407
申请日:2016-12-28
Applicant: Intel Corporation
Inventor: Huseyin E. Sumbul , Gregory K. Chen , Raghavan Kumar , Phil Christopher Knag , Ram Krishnamurthy
Abstract: A neuromorphic computing system is provided which comprises: a synapse core; and a pre-synaptic neuron, a first post-synaptic neuron, and a second post-synaptic neuron coupled to the synaptic core, wherein the synapse core is to: receive a request from the pre-synaptic neuron, generate, in response to the request, a first address of the first post-synaptic neuron and a second address of the second post-synaptic neuron, wherein the first address and the second address are not stored in the synapse core prior to receiving the request.
-
公开(公告)号:US09843441B2
公开(公告)日:2017-12-12
申请号:US14035508
申请日:2013-09-24
Applicant: Intel Corporation
Inventor: Sanu Mathew , Vikram Suresh , Sudhir Satpathy , Mark Anders , Himanshu Kaul , Ram Krishnamurthy
CPC classification number: H04L9/0631 , H04L2209/24
Abstract: Embodiments of an invention for a compact, low power Advanced Encryption Standard circuit are disclosed. In one embodiment, an apparatus includes an encryption unit having a substitution box and an accumulator. The substitution box is to perform a substitution operation on one byte per clock cycle. The accumulator is to accumulate four bytes and perform a mix-column operation in four clock cycles. The encryption unit is implemented using optimum Galois Field polynomial arithmetic for minimum area.
-
-
-
-
-
-
-
-
-