Patents by Inventor Huseyin E. Sumbul
Huseyin E. Sumbul has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11195079Abstract: In one embodiment, a processor comprises a first neuro-synaptic core comprising first circuitry to configure the first neuro-synaptic core as a neuron core responsive to a first value specified by a configuration parameter; and configure the first neuro-synaptic core as a synapse core responsive to a second value specified by the configuration parameter.Type: GrantFiled: November 22, 2017Date of Patent: December 7, 2021Assignee: Intel CorporationInventors: Huseyin E. Sumbul, Gregory K. Chen, Phil Knag, Raghavan Kumar, Ram K. Krishnamurthy
-
Patent number: 11157799Abstract: A neuromorphic computing system is provided which comprises: a synapse core; and a pre-synaptic neuron, a first post-synaptic neuron, and a second post-synaptic neuron coupled to the synaptic core, wherein the synapse core is to: receive a request from the pre-synaptic neuron, generate, in response to the request, a first address of the first post-synaptic neuron and a second address of the second post-synaptic neuron, wherein the first address and the second address are not stored in the synapse core prior to receiving the request.Type: GrantFiled: March 11, 2019Date of Patent: October 26, 2021Assignee: Intel CorporationInventors: Huseyin E. Sumbul, Gregory K. Chen, Raghavan Kumar, Phil Christopher Knag, Ram Krishnamurthy
-
Patent number: 11100385Abstract: Apparatus and method for a scalable, free running neuromorphic processor. For example, one embodiment of a neuromorphic processing apparatus comprises: a plurality of neurons; an interconnection network to communicatively couple at least a subset of the plurality of neurons; a spike controller to stochastically generate a trigger signal, the trigger signal to cause a selected neuron to perform a thresholding operation to determine whether to issue a spike signal.Type: GrantFiled: December 30, 2016Date of Patent: August 24, 2021Assignee: INTEL CORPORATIONInventors: Raghavan Kumar, Gregory K. Chen, Huseyin E. Sumbul, Ram K. Krishnamurthy, Phil Knag
-
Patent number: 10884957Abstract: Techniques and mechanisms for performing in-memory computations with circuitry having a pipeline architecture. In an embodiment, various stages of a pipeline each include a respective input interface and a respective output interface, distinct from said input interface, to couple to different respective circuitry. These stages each further include a respective array of memory cells and circuitry to perform operations based on data stored by said array. A result of one such in-memory computation may be communicated from one pipeline stage to a respective next pipeline stage for use in further in-memory computations. Control circuitry, interconnect circuitry, configuration circuitry or other logic of the pipeline precludes operation of the pipeline as a monolithic, general-purpose memory device. In other embodiments, stages of the pipeline each provide a different respective layer of a neural network.Type: GrantFiled: October 15, 2018Date of Patent: January 5, 2021Assignee: Intel CorporationInventors: Amrita Mathuriya, Sasikanth Manipatruni, Victor W. Lee, Abhishek Sharma, Huseyin E. Sumbul, Gregory Chen, Raghavan Kumar, Phil Knag, Ram Krishnamurthy, Ian Young
-
Publication number: 20190205730Abstract: A neuromorphic computing system is provided which comprises: a synapse core; and a pre-synaptic neuron, a first post-synaptic neuron, and a second post-synaptic neuron coupled to the synaptic core, wherein the synapse core is to: receive a request from the pre-synaptic neuron, generate, in response to the request, a first address of the first post-synaptic neuron and a second address of the second post-synaptic neuron, wherein the first address and the second address are not stored in the synapse core prior to receiving the request.Type: ApplicationFiled: March 11, 2019Publication date: July 4, 2019Applicant: Intel CorporationInventors: Huseyin E. SUMBUL, Gregory K. Chen, Raghavan Kumar, Phil Christopher Knag, Ram Krishnamurthy
-
Publication number: 20190102669Abstract: In one embodiment, a processor comprises a first neuromorphic core to implement a plurality of neural units of a neural network, the first neuromorphic core comprising a memory to store a current time-step of the first neuromorphic core; and a controller to track current time-steps of neighboring neuromorphic cores that receive spikes from or provide spikes to the first neuromorphic core; and control the current time-step of the first neuromorphic core based on the current time-steps of the neighboring neuromorphic cores.Type: ApplicationFiled: September 29, 2017Publication date: April 4, 2019Inventors: Gregory K. Chen, Kshitij Bhardwaj, Raghavan Kumar, Huseyin E. Sumbul, Phil Knag, Ram K. Krishnamurthy, Himanshu Kaul
-
Patent number: 10248906Abstract: A neuromorphic computing system is provided which comprises: a synapse core; and a pre-synaptic neuron, a first post-synaptic neuron, and a second post-synaptic neuron coupled to the synaptic core, wherein the synapse core is to: receive a request from the pre-synaptic neuron, generate, in response to the request, a first address of the first post-synaptic neuron and a second address of the second post-synaptic neuron, wherein the first address and the second address are not stored in the synapse core prior to receiving the request.Type: GrantFiled: December 28, 2016Date of Patent: April 2, 2019Assignee: Intel CorporationInventors: Huseyin E. Sumbul, Gregory K. Chen, Raghavan Kumar, Phil Christopher Knag, Ram Krishnamurthy
-
Publication number: 20190057727Abstract: Techniques and mechanisms for configuring a memory device to perform a sequence of in-memory computations. In an embodiment, a memory device includes a memory array and circuitry, coupled thereto, to perform data computations based on the data stored at the memory array. Based on instructions received at the memory device, control circuitry is configured to enable an automatic performance of a sequence of operations. In another embodiment, the memory device is coupled in an in-series arrangement of other memory devices to provide a pipeline circuit architecture. The memory devices each function as a respective stage of the pipeline circuit architecture, where the stages each perform respective in-memory computations. Some or all such stages each provide a different respective layer of a neural network.Type: ApplicationFiled: October 15, 2018Publication date: February 21, 2019Inventors: Amrita Mathuriya, Sasikanth Manipatruni, Victor W. Lee, Abhishek Sharma, Huseyin E. Sumbul, Gregory Chen, Raghavan Kumar, Phil Knag, Ram Krishnamurthy, Ian Young
-
Publication number: 20190057050Abstract: Techniques and mechanisms for performing in-memory computations with circuitry having a pipeline architecture. In an embodiment, various stages of a pipeline each include a respective input interface and a respective output interface, distinct from said input interface, to couple to different respective circuitry. These stages each further include a respective array of memory cells and circuitry to perform operations based on data stored by said array. A result of one such in-memory computation may be communicated from one pipeline stage to a respective next pipeline stage for use in further in-memory computations. Control circuitry, interconnect circuitry, configuration circuitry or other logic of the pipeline precludes operation of the pipeline as a monolithic, general-purpose memory device. In other embodiments, stages of the pipeline each provide a different respective layer of a neural network.Type: ApplicationFiled: October 15, 2018Publication date: February 21, 2019Inventors: Amrita Mathuriya, Sasikanth Manipatruni, Victor W. Lee, Abhishek Sharma, Huseyin E. Sumbul, Gregory Chen, Raghavan Kumar, Phil Knag, Ram Krishnamurthy, Ian Young
-
Publication number: 20190042909Abstract: In one embodiment, a processor comprises a first neuro-synaptic core comprising first circuitry to configure the first neuro-synaptic core as a neuron core responsive to a first value specified by a configuration parameter; and configure the first neuro-synaptic core as a synapse core responsive to a second value specified by the configuration parameter.Type: ApplicationFiled: November 22, 2017Publication date: February 7, 2019Applicant: Intel CorporationInventors: Huseyin E. Sumbul, Gregory K. Chen, Phil Knag, Raghavan Kumar, Ram K. Krishnamurthy
-
Publication number: 20180189646Abstract: Apparatus and method for configuring large numbers of fan-in and fan-out connections in a neuromorphic computer. For example, one embodiment of an apparatus comprises: a plurality of neurons, each neuron uniquely identifiable with a neuron identifier (ID); at least one memory to store neuron addresses with wildcard values to establish fan-in and/or fan-out connections between the neurons; and a router to translate at least one neuron address containing wildcard values into two or more neuron IDs to establish the fan-in and/or fan-out connections between the neurons.Type: ApplicationFiled: December 30, 2016Publication date: July 5, 2018Inventors: Raghavan Kumar, Huseyin E. Sumbul, Gregory K. Chen, Phil Knag
-
Publication number: 20180181861Abstract: A neuromorphic computing system is provided which comprises: a synapse core; and a pre-synaptic neuron, a first post-synaptic neuron, and a second post-synaptic neuron coupled to the synaptic core, wherein the synapse core is to: receive a request from the pre-synaptic neuron, generate, in response to the request, a first address of the first post-synaptic neuron and a second address of the second post-synaptic neuron, wherein the first address and the second address are not stored in the synapse core prior to receiving the request.Type: ApplicationFiled: December 28, 2016Publication date: June 28, 2018Inventors: Huseyin E. SUMBUL, Gregory K. CHEN, Raghavan KUMAR, Phil C. Knag, Ram Krishnamurthy
-
Publication number: 20170286829Abstract: Systems and methods for event-driven learning with spike timing dependent plasticity in neuromorphic computers are disclosed. A neuromorphic processor includes a synapse coupled to a pre-synaptic neuron and coupled to a post-synaptic neuron, the synapse including a synapse memory to store a synapse weight and synapse spike timing dependent plasticity (STDP) circuit coupled to the synapse memory. The pre-synaptic neuron includes a pre-synaptic neuron memory to store a pre-synaptic neuron spike history and a pre-synaptic neuron STDP circuit coupled to the pre-synaptic neuron memory, the pre-synaptic neuron STDP circuit to, in response to the pre-synaptic neuron firing, initiate performing long term potentiation. The post-synaptic neuron includes a post-synaptic neuron memory storing a post-synaptic neuron spike history and a post-synaptic neuron STDP circuit coupled to the post-synaptic neuron memory to, in response to the post-synaptic neuron firing, initiate performing long term depression.Type: ApplicationFiled: April 1, 2016Publication date: October 5, 2017Inventors: Gregory K. Chen, Raghavan Kumar, Huseyin E. Sumbul
-
Patent number: 9524767Abstract: A method includes receiving a data bit value at a buffer in a bitcell based on a first state of a write bitline connected to the buffer, and transferring the data bit value from the buffer to a first magnetic switching cell in the bitcell for a later read operation at least by holding the write bitline to a reference value different from the first state, and asserting first and second predetermined voltage levels on respective first and second write wordlines connected to the buffer.Type: GrantFiled: June 20, 2014Date of Patent: December 20, 2016Assignee: Carnegie Mellon UniversityInventors: Lawrence Pileggi, David M. Bromberg, Huseyin E. Sumbul
-
Publication number: 20160125928Abstract: A method includes receiving a data bit value at a buffer in a bitcell based on a first state of a write bitline connected to the buffer, and transferring the data bit value from the buffer to a first magnetic switching cell in the bitcell for a later read operation at least by holding the write bitline to a reference value different from the first state, and asserting first and second predetermined voltage levels on respective first and second write wordlines connected to the buffer.Type: ApplicationFiled: June 20, 2014Publication date: May 5, 2016Inventors: Lawrence Pileggi, David M. Bromberg, Huseyin E. Sumbul