Patents by Inventor Christophe Piveteau
Christophe Piveteau has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11436302Abstract: The present disclosure relates to an electronic system for computing items of an outer product matrix, for each item of at least part of the items of the matrix. The system is configured to receive a pair of real numbers of two vectors, the pair corresponding to the item. The system is further configured to compute a stochastic representation of the real numbers resulting in two sets of bits, the set of bits comprising a subset of bits representing the real number and a sign bit indicative of the sign of the real number. The system is further configured to perform a sequence of digital operations using the two sets of bits to provide a representation of the item.Type: GrantFiled: June 7, 2019Date of Patent: September 6, 2022Assignee: International Business Machines CorporationInventors: Vinay Manikrao Joshi, Abu Sebastian, Manuel Le Gallo-Bourdeau, Irem Boybat Kara, Christophe Piveteau
-
Patent number: 11386319Abstract: Methods and apparatus are provided for training an artificial neural network, having a succession of neuron layers with interposed synaptic layers each storing a respective set of weights {w} for weighting signals propagated between its adjacent neuron layers, via an iterative cycle of signal propagation and weight-update calculation operations. Such a method includes, for at least one of the synaptic layers, providing a plurality Pl of arrays of memristive devices, each array storing the set of weights of that synaptic layer Sl in respective memristive devices, and, in a signal propagation operation, supplying respective subsets of the signals to be weighted by the synaptic layer Sl in parallel to the Pl arrays. The method also includes, in a weight-update calculation operation, calculating updates to respective weights stored in each of the Pl arrays in dependence on signals propagated by the neuron layers.Type: GrantFiled: March 14, 2019Date of Patent: July 12, 2022Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Manuel Le Gallo-Bourdeau, Nandakumar Sasidharan Rajalekshmi, Christophe Piveteau, Irem Boybat Kara, Abu Sebastian, Evangelos Stavros Eleftheriou
-
Patent number: 11373092Abstract: Methods are provided for training weights of an artificial neural network to be implemented by inference computing apparatus in which the trained weights are stored as programmed conductance states of respective predetermined memristive devices. Such a method includes deriving for the memristive devices a probability distribution indicating distribution of conductance errors for the devices in the programmed conductance states. The method further comprises, in a digital computing apparatus: training the weights via an iterative training process in which the weights are repeatedly updated in response to processing by the network of training data which is propagated over the network via the weights; and applying noise dependent on said probability distribution to weights used in the iterative training process.Type: GrantFiled: April 10, 2019Date of Patent: June 28, 2022Assignee: International Business Machines CorporationInventors: Christophe Piveteau, Abu Sebastian, Manuel Le Gallo-Bourdeau, Vinay Manikrao Joshi
-
Patent number: 11308252Abstract: Techniques that combine quantum error correction and quantum error mitigation are used to simulate a fault-tolerant T-gate with low sampling overhead using the quasiprobability decomposition method. In some embodiments, the T-gate can be simulated using two logical bits and a magic state preparation that mitigates the need for magic state distillation and consequently has a low sampling overhead. Alternatively, the T-gate can be simulated based on code deformation performed on the surface code. Noise is removed from the T-gate using quasiprobability decomposition based on a learned logical error rate.Type: GrantFiled: November 13, 2020Date of Patent: April 19, 2022Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Christophe Piveteau, David Sutter, Paul Kristan Temme, Sergey Bravyi, Jay Michael Gambetta, Stefan Woerner
-
Publication number: 20220092460Abstract: Techniques facilitating error mitigation for quantum computing devices. In one example, a system can comprise a process that executes computer executable components stored in memory. The computer executable components comprise: an approximation component; a budget component; and an optimization component. The approximation component can generate an approximate decomposition of a quantum gate. The budget component can set a budget value (Cbudget) for a C-factor that is a metric for increase in variance of quasi-probability sampling. The optimization component can determine an optimal decomposition for the quantum gate as a function of Cbudget.Type: ApplicationFiled: September 23, 2020Publication date: March 24, 2022Inventors: Christophe Piveteau, David Sutter, Stefan Woerner
-
Patent number: 11250107Abstract: The present disclosure relates to a method for executing a computation task composed of at least one set of operations where subsets of pipelineable operations of the set of operations are determined in accordance with a pipelining scheme. A single routine may be created for enabling execution of the determined subsets of operations by a hardware accelerator. The routine has, as arguments, a value indicative of input data and values of configuration parameters of the computation task, where a call of the routine causes a scheduling of the subsets of operations on the hardware accelerator in accordance with the values of the configuration parameters. Upon receiving input data of the computation task, the routine may be called to cause the hardware accelerator to perform by the computation task in accordance with the scheduling.Type: GrantFiled: July 15, 2019Date of Patent: February 15, 2022Assignee: International Business Machines CorporationInventors: Christophe Piveteau, Nikolas Ioannou, Igor Krawczuk, Manuel Le Gallo-Bourdeau, Abu Sebastian, Evangelos Stavros Eleftheriou
-
Patent number: 11042715Abstract: A system can include a memristive crossbar array, which can include row lines and column lines intersecting the row lines. Resistive memory elements can be coupled between the row lines and the column lines at the junctions formed by the row and column lines. The resistive memory elements represent the values of the matrix. The system can further include an analogue circuit. The system can be configured to perform an exponentiation of the values of the vector in accordance with a first exponent. The crossbar array can be configured to apply the resulting values of the vector to the resistive elements thereby generating currents. The analogue circuit can be configured to perform an exponentiation of the generated currents in accordance with a second exponent.Type: GrantFiled: April 11, 2019Date of Patent: June 22, 2021Assignee: International Business Machines CorporationInventors: Christophe Piveteau, Abu Sebastian, Manuel Le Gallo-Bourdeau, Riduan Khaddam-Aljameh
-
Publication number: 20210019362Abstract: The present disclosure relates to a method for executing a computation task composed of at least one set of operations where subsets of pipelineable operations of the set of operations are determined in accordance with a pipelining scheme. A single routine may be created for enabling execution of the determined subsets of operations by a hardware accelerator. The routine has, as arguments, a value indicative of input data and values of configuration parameters of the computation task, where a call of the routine causes a scheduling of the subsets of operations on the hardware accelerator in accordance with the values of the configuration parameters. Upon receiving input data of the computation task, the routine may be called to cause the hardware accelerator to perform by the computation task in accordance with the scheduling.Type: ApplicationFiled: July 15, 2019Publication date: January 21, 2021Inventors: Christophe Piveteau, Nikolas loannou, Igor Krawczuk, Manuel Le Gallo-Bourdeau, Abu Sebastian, Evangelos Stavros Eleftheriou
-
Publication number: 20200387563Abstract: The present disclosure relates to an electronic system for computing items of an outer product matrix, for each item of at least part of the items of the matrix. The system is configured to receive a pair of real numbers of two vectors, the pair corresponding to said item. The system is further configured to compute a stochastic representation of the real numbers resulting in two sets of bits, the set of bits comprising a subset of bits representing the real number and a sign bit indicative of the sign of the real number. The system is further configured to perform a sequence of digital operations using the two sets of bits to provide a representation of said item.Type: ApplicationFiled: June 7, 2019Publication date: December 10, 2020Inventors: Vinay Manikrao Joshi, Abu Sebastian, Manuel Le Gallo-Bourdeau, Irem Boybat Kara, Christophe Piveteau
-
Publication number: 20200327287Abstract: A system can include a memristive crossbar array, which can include row lines and column lines intersecting the row lines. Resistive memory elements can be coupled between the row lines and the column lines at the junctions formed by the row and column lines. The resistive memory elements represent the values of the matrix. The system can further include an analogue circuit. The system can be configured to perform an exponentiation of the values of the vector in accordance with a first exponent. The crossbar array can be configured to apply the resulting values of the vector to the resistive elements thereby generating currents. The analogue circuit can be configured to perform an exponentiation of the generated currents in accordance with a second exponent.Type: ApplicationFiled: April 11, 2019Publication date: October 15, 2020Inventors: Christophe Piveteau, Abu Sebastian, Manuel Le Gallo-Bourdeau, Riduan Khaddam-Aljameh
-
Publication number: 20200327406Abstract: Methods are provided for training weights of an artificial neural network to be implemented by inference computing apparatus in which the trained weights are stored as programmed conductance states of respective predetermined memristive devices. Such a method includes deriving for the memristive devices a probability distribution indicating distribution of conductance errors for the devices in the programmed conductance states. The method further comprises, in a digital computing apparatus: training the weights via an iterative training process in which the weights are repeatedly updated in response to processing by the network of training data which is propagated over the network via the weights; and applying noise dependent on said probability distribution to weights used in the iterative training process.Type: ApplicationFiled: April 10, 2019Publication date: October 15, 2020Inventors: Christophe Piveteau, Abu Sebastian, Manuel Le Gallo-Bourdeau, Vinay Manikrao Joshi
-
Publication number: 20200293855Abstract: Methods and apparatus are provided for training an artificial neural network, having a succession of neuron layers with interposed synaptic layers each storing a respective set of weights {w} for weighting signals propagated between its adjacent neuron layers, via an iterative cycle of signal propagation and weight-update calculation operations. Such a method includes, for at least one of the synaptic layers, providing a plurality P1 of arrays of memristive devices, each array storing the set of weights of that synaptic layer S1 in respective memristive devices, and, in a signal propagation operation, supplying respective subsets of the signals to be weighted by the synaptic layer S1 in parallel to the P1 arrays. The method also includes, in a weight-update calculation operation, calculating updates to respective weights stored in each of the P1 arrays in dependence on signals propagated by the neuron layers.Type: ApplicationFiled: March 14, 2019Publication date: September 17, 2020Inventors: Manuel Le Gallo-Bourdeau, Nandakumar Sasidharan Rajalekshmi, Christophe Piveteau, Irem Boybat Kara, Abu Sebastian, Evangelos Stavros Eleftheriou