Patents by Inventor Gregory K. Chen

Gregory K. Chen has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20200034148
    Abstract: A compute near memory (CNM) convolution accelerator enables a convolutional neural network (CNN) to use dedicated acceleration to achieve efficient in-place convolution operations with less impact on memory and energy consumption. A 2D convolution operation is reformulated as 1D row-wise convolution. The 1D row-wise convolution enables the CNM convolution accelerator to process input activations row-by-row, while using the weights one-by-one. Lightweight access circuits provide the ability to stream both weights and input row as vectors to MAC units, which in turn enables modules of the CNM convolution accelerator to implement convolution for both [1×1] and chosen [n×n] sized filters.
    Type: Application
    Filed: September 28, 2019
    Publication date: January 30, 2020
    Inventors: Huseyin Ekin SUMBUL, Gregory K. CHEN, Phil KNAG, Raghavan KUMAR, Ram KRISHNAMURTHY
  • Publication number: 20200026498
    Abstract: A memory circuit includes a number (X) of multiply-accumulate (MAC) circuits that are dynamically configurable. The MAC circuits can either compute an output based on computations of X elements of the input vector with the weight vector, or to compute the output based on computations of a single element of the input vector with the weight vector, with each element having a one bit or multibit length. A first memory can hold the input vector having a width of X elements and a second memory can store the weight vector. The MAC circuits include a MAC array on chip with the first memory.
    Type: Application
    Filed: September 27, 2019
    Publication date: January 23, 2020
    Inventors: Huseyin Ekin SUMBUL, Gregory K. CHEN, Phil KNAG, Raghavan KUMAR, Ram KRISHNAMURTHY
  • Publication number: 20200019847
    Abstract: An apparatus is described. The apparatus includes a circuit to process a binary neural network. The circuit includes an array of processing cores, wherein, processing cores of the array of processing cores are to process different respective areas of a weight matrix of the binary neural network. The processing cores each include add circuitry to add only those weights of an i layer of the binary neural network that are to be effectively multiplied by a non zero nodal output of an i?1 layer of the binary neural network.
    Type: Application
    Filed: September 25, 2019
    Publication date: January 16, 2020
    Inventors: Ram KRISHNAMURTHY, Gregory K. CHEN, Raghavan KUMAR, Phil KNAG, Huseyin Ekin SUMBUL, Deepak Vinayak KADETOTAD
  • Publication number: 20200019846
    Abstract: An apparatus is described. The apparatus includes a long short term memory (LSTM) circuit having a multiply accumulate circuit (MAC). The MAC circuit has circuitry to rely on a stored product term rather than explicitly perform a multiplication operation to determine the product term if an accumulation of differences between consecutive, preceding input values has not reached a threshold.
    Type: Application
    Filed: September 25, 2019
    Publication date: January 16, 2020
    Inventors: Ram KRISHNAMURTHY, Gregory K. CHEN, Raghavan KUMAR, Phil KNAG, Huseyin Ekin SUMBUL
  • Patent number: 10482372
    Abstract: Systems and methods for an interconnection scheme for reconfigurable neuromorphic hardware are disclosed. A neuromorphic processor may include a plurality of corelets, each corelet may include a plurality of synapse arrays and a neuron array. Each synapse array may include a plurality of synapses and a synapse array router coupled to synapse outputs in a synapse array. Each synapse may include a synapse input, synapse output; and a synapse memory. A neuron array may include a plurality of neurons, each neuron may include a neuron input and a neuron output. Each synapse array router may include a first logic to route one or more of the synapse outputs to one or more of the neuron inputs.
    Type: Grant
    Filed: December 23, 2015
    Date of Patent: November 19, 2019
    Assignee: Intel Corporation
    Inventors: Gregory K. Chen, Jae-Sun Seo
  • Publication number: 20190303750
    Abstract: Examples described herein relate to a neural network whose weights from a matrix are selected from a set of weights stored in a memory on-chip with a processing engine for generating multiply and carry operations. The number of weights in the set of weights stored in the memory can be less than a number of weights in the matrix thereby reducing an amount of memory used to store weights in a matrix. The weights in the memory can be generated in training using gradients from back propagation. Weights in the memory can be selected using a tabulation hash calculation on entries in a table.
    Type: Application
    Filed: June 17, 2019
    Publication date: October 3, 2019
    Inventors: Raghavan KUMAR, Gregory K. CHEN, Huseyin Ekin SUMBUL, Phil KNAG, Ram KRISHNAMURTHY
  • Patent number: 10360496
    Abstract: An apparatus and method are described for a neuromorphic processor design in which neuron timing information is duplicated on a neuromorphic core.
    Type: Grant
    Filed: April 1, 2016
    Date of Patent: July 23, 2019
    Assignee: Intel Corporation
    Inventors: Gregory K. Chen, Jae-Sun Seo, Thomas C Chen, Raghavan Kumar
  • Publication number: 20190205730
    Abstract: A neuromorphic computing system is provided which comprises: a synapse core; and a pre-synaptic neuron, a first post-synaptic neuron, and a second post-synaptic neuron coupled to the synaptic core, wherein the synapse core is to: receive a request from the pre-synaptic neuron, generate, in response to the request, a first address of the first post-synaptic neuron and a second address of the second post-synaptic neuron, wherein the first address and the second address are not stored in the synapse core prior to receiving the request.
    Type: Application
    Filed: March 11, 2019
    Publication date: July 4, 2019
    Applicant: Intel Corporation
    Inventors: Huseyin E. SUMBUL, Gregory K. Chen, Raghavan Kumar, Phil Christopher Knag, Ram Krishnamurthy
  • Publication number: 20190138893
    Abstract: An apparatus is described. The apparatus includes a compute-in-memory (CIM) circuit for implementing a neural network disposed on a semiconductor chip. The CIM circuit includes a mathematical computation circuit coupled to a memory array. The memory array includes an embedded dynamic random access memory (eDRAM) memory array. Another apparatus is described. The apparatus includes a compute-in-memory (CIM) circuit for implementing a neural network disposed on a semiconductor chip. The CIM circuit includes a mathematical computation circuit coupled to a memory array. The mathematical computation circuit includes a switched capacitor circuit. The switched capacitor circuit includes a back-end-of-line (BEOL) capacitor coupled to a thin film transistor within the metal/dielectric layers of the semiconductor chip. Another apparatus is described. The apparatus includes a compute-in-memory (CIM) circuit for implementing a neural network disposed on a semiconductor chip.
    Type: Application
    Filed: September 28, 2018
    Publication date: May 9, 2019
    Inventors: Abhishek SHARMA, Jack T. KAVALIEROS, Ian A. YOUNG, Ram KRISHNAMURTHY, Sasikanth MANIPATRUNI, Uygar AVCI, Gregory K. CHEN, Amrita MATHURIYA, Raghavan KUMAR, Phil KNAG, Huseyin Ekin SUMBUL, Nazila HARATIPOUR, Van H. LE
  • Publication number: 20190102170
    Abstract: A compute-in-memory (CIM) circuit that enables a multiply-accumulate (MAC) operation based on a current-sensing readout technique. An operational amplifier coupled with a bitline of a column of bitcells included in a memory array of the CIM circuit to cause the bitcells to act like ideal current sources for use in determining an analog voltage value outputted from the operational amplifier for given states stored in the bitcells and for given input activations for the bitcells. The analog voltage value sensed by processing circuitry of the CIM circuit and converted to a digital value to compute a multiply-accumulate (MAC) value.
    Type: Application
    Filed: September 28, 2018
    Publication date: April 4, 2019
    Inventors: Gregory K. CHEN, Raghavan KUMAR, Huseyin Ekin SUMBUL, Phil KNAG, Ram KRISHNAMURTHY, Sasikanth MANIPATRUNI, Amrita MATHURIYA, Abhishek SHARMA, Ian A. YOUNG
  • Publication number: 20190102669
    Abstract: In one embodiment, a processor comprises a first neuromorphic core to implement a plurality of neural units of a neural network, the first neuromorphic core comprising a memory to store a current time-step of the first neuromorphic core; and a controller to track current time-steps of neighboring neuromorphic cores that receive spikes from or provide spikes to the first neuromorphic core; and control the current time-step of the first neuromorphic core based on the current time-steps of the neighboring neuromorphic cores.
    Type: Application
    Filed: September 29, 2017
    Publication date: April 4, 2019
    Inventors: Gregory K. Chen, Kshitij Bhardwaj, Raghavan Kumar, Huseyin E. Sumbul, Phil Knag, Ram K. Krishnamurthy, Himanshu Kaul
  • Publication number: 20190103156
    Abstract: A full-rail digital-read CIM circuit enables a weighted read operation on a single row of a memory array. A weighted read operation captures a value of a weight stored in the single memory array row without having to rely on weighted row-access. Rather, using full-rail access and a weighted sampling capacitance network, the CIM circuit enables the weighted read operation even under process variation, noise and mismatch.
    Type: Application
    Filed: September 28, 2018
    Publication date: April 4, 2019
    Inventors: Huseyin Ekin SUMBUL, Gregory K. CHEN, Raghavan KUMAR, Phil Ekin KNAG, Abhishek SHARMA, Sasikanth MANIPATRUNI, Amrita MATHURIYA, Ram A. KRISHNAMURTHY, Ian A. YOUNG
  • Publication number: 20190102359
    Abstract: A binary CIM circuit enables all memory cells in a memory array to be effectively accessible simultaneously for computation using fixed pulse widths on the wordlines and equal capacitance on the bitlines. The fixed pulse widths and equal capacitance ensure that a minimum voltage drop in the bitline represents one least significant bit (LSB) so that the bitline voltage swing remains safely within the maximum allowable range. The binary CIM circuit maximizes the effective memory bandwidth of a memory array for a given maximum voltage range of bitline voltage.
    Type: Application
    Filed: September 28, 2018
    Publication date: April 4, 2019
    Inventors: Phil KNAG, Gregory K. CHEN, Raghavan KUMAR, Huseyin Ekin SUMBUL, Abhishek Sharma, Sasikanth Manipatruni, Amrita Mathuriya, Ram A. Krishnamurthy, Ian A. Young
  • Patent number: 10248906
    Abstract: A neuromorphic computing system is provided which comprises: a synapse core; and a pre-synaptic neuron, a first post-synaptic neuron, and a second post-synaptic neuron coupled to the synaptic core, wherein the synapse core is to: receive a request from the pre-synaptic neuron, generate, in response to the request, a first address of the first post-synaptic neuron and a second address of the second post-synaptic neuron, wherein the first address and the second address are not stored in the synapse core prior to receiving the request.
    Type: Grant
    Filed: December 28, 2016
    Date of Patent: April 2, 2019
    Assignee: Intel Corporation
    Inventors: Huseyin E. Sumbul, Gregory K. Chen, Raghavan Kumar, Phil Christopher Knag, Ram Krishnamurthy
  • Publication number: 20190065151
    Abstract: A memory device that includes a plurality subarrays of memory cells to store static weights and a plurality of digital full-adder circuits between subarrays of memory cells is provided. The digital full-adder circuit in the memory device eliminates the need to move data from a memory device to a processor to perform machine learning calculations. Rows of full-adder circuits are distributed between sub-arrays of memory cells to increase the effective memory bandwidth and reduce the time to perform matrix-vector multiplications in the memory device by performing bit-serial dot-product primitives in the form of accumulating m 1-bit x n-bit multiplications.
    Type: Application
    Filed: September 28, 2018
    Publication date: February 28, 2019
    Inventors: Gregory K. CHEN, Raghavan KUMAR, Huseyin Ekin SUMBUL, Phil KNAG, Ram KRISHNAMURTHY, Sasikanth MANIPATRUNI, Amrita MATHURIYA, Abhishek SHARMA, Ian A. YOUNG
  • Publication number: 20190042949
    Abstract: A semiconductor chip is described. The semiconductor chip includes a compute-in-memory (CIM) circuit to implement a neural network in hardware. The semiconductor chip also includes at least one output that presents samples of voltages generated at a node of the CIM circuit in response to a range of neural network input values applied to the CIM circuit to optimize the CIM circuit for the neural network.
    Type: Application
    Filed: September 28, 2018
    Publication date: February 7, 2019
    Inventors: Ian A. YOUNG, Ram KRISHNAMURTHY, Sasikanth MANIPATRUNI, Gregory K. CHEN, Amrita MATHURIYA, Abhishek SHARMA, Raghavan KUMAR, Phil KNAG, Huseyin Ekin SUMBUL
  • Publication number: 20190043560
    Abstract: A memory circuit has compute-in-memory circuitry that enables a multiply-accumulate (MAC) operation based on shared charge. Row access circuitry drives multiple rows of a memory array to multiply a first data word with a second data word stored in the memory array. The row access circuitry drives the multiple rows based on the bit pattern of the first data word. Column access circuitry drives a column of the memory array when the rows are driven. Accessed rows discharge the column line in an accumulative fashion. Sensing circuitry can sense voltage on the column line. A processor in the memory circuit computes a MAC value based on the voltage sensed on the column.
    Type: Application
    Filed: September 28, 2018
    Publication date: February 7, 2019
    Inventors: Huseyin Ekin Sumbul, Gregory K. Chen, Raghavan Kumar, Phil Knag, Abhishek Sharma, Sasikanth Manipatruni, Amrita Mathuriya, Ram Krishnamurthy, Ian A. Young
  • Publication number: 20190042199
    Abstract: Compute-in memory circuits and techniques are described. In one example, a memory device includes an array of memory cells, the array including multiple sub-arrays. Each of the sub-arrays receives a different voltage. The memory device also includes capacitors coupled with conductive access lines of each of the multiple sub-arrays and circuitry coupled with the capacitors, to share charge between the capacitors in response to a signal. In one example, computing device, such as a machine learning accelerator, includes a first memory array and a second memory array. The computing device also includes an analog processor circuit coupled with the first and second memory arrays to receive first analog input voltages from the first memory array and second analog input voltages from the second memory array and perform one or more operations on the first and second analog input voltages, and output an analog output voltage.
    Type: Application
    Filed: September 28, 2018
    Publication date: February 7, 2019
    Inventors: Huseyin Ekin SUMBUL, Phil KNAG, Gregory K. CHEN, Raghavan KUMAR, Abhishek SHARMA, Sasikanth MANIPATRUNI, Amrita MATHURIYA, Ram KRISHNAMURTHY, Ian A. YOUNG
  • Publication number: 20190042909
    Abstract: In one embodiment, a processor comprises a first neuro-synaptic core comprising first circuitry to configure the first neuro-synaptic core as a neuron core responsive to a first value specified by a configuration parameter; and configure the first neuro-synaptic core as a synapse core responsive to a second value specified by the configuration parameter.
    Type: Application
    Filed: November 22, 2017
    Publication date: February 7, 2019
    Applicant: Intel Corporation
    Inventors: Huseyin E. Sumbul, Gregory K. Chen, Phil Knag, Raghavan Kumar, Ram K. Krishnamurthy
  • Publication number: 20190042928
    Abstract: An apparatus is described. The apparatus includes a compute in memory circuit. The compute in memory circuit includes a memory circuit and an encoder. The memory circuit is to provide 2m voltage levels on a read data line where m is greater than 1. The memory circuit includes storage cells sufficient to store a number of bits n where n is greater than m. The encoder is to receive an m bit input and convert the m bit input into an n bit word that is to be stored in the memory circuit, where, the m bit to n bit encoding performed by the encoder creates greater separation between those of the voltage levels that demonstrate wider voltage distributions on the read data line than others of the voltage levels.
    Type: Application
    Filed: September 28, 2018
    Publication date: February 7, 2019
    Inventors: Ian A. YOUNG, Ram KRISHNAMURTHY, Sasikanth MANIPATRUNI, Gregory K. CHEN, Amrita MATHURIYA, Abhishek SHARMA, Raghavan KUMAR, Phil KNAG, Huseyin Ekin SUMBUL