Patents by Inventor Christophe Piveteau

Christophe Piveteau has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11436302
    Abstract: The present disclosure relates to an electronic system for computing items of an outer product matrix, for each item of at least part of the items of the matrix. The system is configured to receive a pair of real numbers of two vectors, the pair corresponding to the item. The system is further configured to compute a stochastic representation of the real numbers resulting in two sets of bits, the set of bits comprising a subset of bits representing the real number and a sign bit indicative of the sign of the real number. The system is further configured to perform a sequence of digital operations using the two sets of bits to provide a representation of the item.
    Type: Grant
    Filed: June 7, 2019
    Date of Patent: September 6, 2022
    Assignee: International Business Machines Corporation
    Inventors: Vinay Manikrao Joshi, Abu Sebastian, Manuel Le Gallo-Bourdeau, Irem Boybat Kara, Christophe Piveteau
  • Patent number: 11386319
    Abstract: Methods and apparatus are provided for training an artificial neural network, having a succession of neuron layers with interposed synaptic layers each storing a respective set of weights {w} for weighting signals propagated between its adjacent neuron layers, via an iterative cycle of signal propagation and weight-update calculation operations. Such a method includes, for at least one of the synaptic layers, providing a plurality Pl of arrays of memristive devices, each array storing the set of weights of that synaptic layer Sl in respective memristive devices, and, in a signal propagation operation, supplying respective subsets of the signals to be weighted by the synaptic layer Sl in parallel to the Pl arrays. The method also includes, in a weight-update calculation operation, calculating updates to respective weights stored in each of the Pl arrays in dependence on signals propagated by the neuron layers.
    Type: Grant
    Filed: March 14, 2019
    Date of Patent: July 12, 2022
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Manuel Le Gallo-Bourdeau, Nandakumar Sasidharan Rajalekshmi, Christophe Piveteau, Irem Boybat Kara, Abu Sebastian, Evangelos Stavros Eleftheriou
  • Patent number: 11373092
    Abstract: Methods are provided for training weights of an artificial neural network to be implemented by inference computing apparatus in which the trained weights are stored as programmed conductance states of respective predetermined memristive devices. Such a method includes deriving for the memristive devices a probability distribution indicating distribution of conductance errors for the devices in the programmed conductance states. The method further comprises, in a digital computing apparatus: training the weights via an iterative training process in which the weights are repeatedly updated in response to processing by the network of training data which is propagated over the network via the weights; and applying noise dependent on said probability distribution to weights used in the iterative training process.
    Type: Grant
    Filed: April 10, 2019
    Date of Patent: June 28, 2022
    Assignee: International Business Machines Corporation
    Inventors: Christophe Piveteau, Abu Sebastian, Manuel Le Gallo-Bourdeau, Vinay Manikrao Joshi
  • Patent number: 11308252
    Abstract: Techniques that combine quantum error correction and quantum error mitigation are used to simulate a fault-tolerant T-gate with low sampling overhead using the quasiprobability decomposition method. In some embodiments, the T-gate can be simulated using two logical bits and a magic state preparation that mitigates the need for magic state distillation and consequently has a low sampling overhead. Alternatively, the T-gate can be simulated based on code deformation performed on the surface code. Noise is removed from the T-gate using quasiprobability decomposition based on a learned logical error rate.
    Type: Grant
    Filed: November 13, 2020
    Date of Patent: April 19, 2022
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Christophe Piveteau, David Sutter, Paul Kristan Temme, Sergey Bravyi, Jay Michael Gambetta, Stefan Woerner
  • Publication number: 20220092460
    Abstract: Techniques facilitating error mitigation for quantum computing devices. In one example, a system can comprise a process that executes computer executable components stored in memory. The computer executable components comprise: an approximation component; a budget component; and an optimization component. The approximation component can generate an approximate decomposition of a quantum gate. The budget component can set a budget value (Cbudget) for a C-factor that is a metric for increase in variance of quasi-probability sampling. The optimization component can determine an optimal decomposition for the quantum gate as a function of Cbudget.
    Type: Application
    Filed: September 23, 2020
    Publication date: March 24, 2022
    Inventors: Christophe Piveteau, David Sutter, Stefan Woerner
  • Patent number: 11250107
    Abstract: The present disclosure relates to a method for executing a computation task composed of at least one set of operations where subsets of pipelineable operations of the set of operations are determined in accordance with a pipelining scheme. A single routine may be created for enabling execution of the determined subsets of operations by a hardware accelerator. The routine has, as arguments, a value indicative of input data and values of configuration parameters of the computation task, where a call of the routine causes a scheduling of the subsets of operations on the hardware accelerator in accordance with the values of the configuration parameters. Upon receiving input data of the computation task, the routine may be called to cause the hardware accelerator to perform by the computation task in accordance with the scheduling.
    Type: Grant
    Filed: July 15, 2019
    Date of Patent: February 15, 2022
    Assignee: International Business Machines Corporation
    Inventors: Christophe Piveteau, Nikolas Ioannou, Igor Krawczuk, Manuel Le Gallo-Bourdeau, Abu Sebastian, Evangelos Stavros Eleftheriou
  • Patent number: 11042715
    Abstract: A system can include a memristive crossbar array, which can include row lines and column lines intersecting the row lines. Resistive memory elements can be coupled between the row lines and the column lines at the junctions formed by the row and column lines. The resistive memory elements represent the values of the matrix. The system can further include an analogue circuit. The system can be configured to perform an exponentiation of the values of the vector in accordance with a first exponent. The crossbar array can be configured to apply the resulting values of the vector to the resistive elements thereby generating currents. The analogue circuit can be configured to perform an exponentiation of the generated currents in accordance with a second exponent.
    Type: Grant
    Filed: April 11, 2019
    Date of Patent: June 22, 2021
    Assignee: International Business Machines Corporation
    Inventors: Christophe Piveteau, Abu Sebastian, Manuel Le Gallo-Bourdeau, Riduan Khaddam-Aljameh
  • Publication number: 20210019362
    Abstract: The present disclosure relates to a method for executing a computation task composed of at least one set of operations where subsets of pipelineable operations of the set of operations are determined in accordance with a pipelining scheme. A single routine may be created for enabling execution of the determined subsets of operations by a hardware accelerator. The routine has, as arguments, a value indicative of input data and values of configuration parameters of the computation task, where a call of the routine causes a scheduling of the subsets of operations on the hardware accelerator in accordance with the values of the configuration parameters. Upon receiving input data of the computation task, the routine may be called to cause the hardware accelerator to perform by the computation task in accordance with the scheduling.
    Type: Application
    Filed: July 15, 2019
    Publication date: January 21, 2021
    Inventors: Christophe Piveteau, Nikolas loannou, Igor Krawczuk, Manuel Le Gallo-Bourdeau, Abu Sebastian, Evangelos Stavros Eleftheriou
  • Publication number: 20200387563
    Abstract: The present disclosure relates to an electronic system for computing items of an outer product matrix, for each item of at least part of the items of the matrix. The system is configured to receive a pair of real numbers of two vectors, the pair corresponding to said item. The system is further configured to compute a stochastic representation of the real numbers resulting in two sets of bits, the set of bits comprising a subset of bits representing the real number and a sign bit indicative of the sign of the real number. The system is further configured to perform a sequence of digital operations using the two sets of bits to provide a representation of said item.
    Type: Application
    Filed: June 7, 2019
    Publication date: December 10, 2020
    Inventors: Vinay Manikrao Joshi, Abu Sebastian, Manuel Le Gallo-Bourdeau, Irem Boybat Kara, Christophe Piveteau
  • Publication number: 20200327287
    Abstract: A system can include a memristive crossbar array, which can include row lines and column lines intersecting the row lines. Resistive memory elements can be coupled between the row lines and the column lines at the junctions formed by the row and column lines. The resistive memory elements represent the values of the matrix. The system can further include an analogue circuit. The system can be configured to perform an exponentiation of the values of the vector in accordance with a first exponent. The crossbar array can be configured to apply the resulting values of the vector to the resistive elements thereby generating currents. The analogue circuit can be configured to perform an exponentiation of the generated currents in accordance with a second exponent.
    Type: Application
    Filed: April 11, 2019
    Publication date: October 15, 2020
    Inventors: Christophe Piveteau, Abu Sebastian, Manuel Le Gallo-Bourdeau, Riduan Khaddam-Aljameh
  • Publication number: 20200327406
    Abstract: Methods are provided for training weights of an artificial neural network to be implemented by inference computing apparatus in which the trained weights are stored as programmed conductance states of respective predetermined memristive devices. Such a method includes deriving for the memristive devices a probability distribution indicating distribution of conductance errors for the devices in the programmed conductance states. The method further comprises, in a digital computing apparatus: training the weights via an iterative training process in which the weights are repeatedly updated in response to processing by the network of training data which is propagated over the network via the weights; and applying noise dependent on said probability distribution to weights used in the iterative training process.
    Type: Application
    Filed: April 10, 2019
    Publication date: October 15, 2020
    Inventors: Christophe Piveteau, Abu Sebastian, Manuel Le Gallo-Bourdeau, Vinay Manikrao Joshi
  • Publication number: 20200293855
    Abstract: Methods and apparatus are provided for training an artificial neural network, having a succession of neuron layers with interposed synaptic layers each storing a respective set of weights {w} for weighting signals propagated between its adjacent neuron layers, via an iterative cycle of signal propagation and weight-update calculation operations. Such a method includes, for at least one of the synaptic layers, providing a plurality P1 of arrays of memristive devices, each array storing the set of weights of that synaptic layer S1 in respective memristive devices, and, in a signal propagation operation, supplying respective subsets of the signals to be weighted by the synaptic layer S1 in parallel to the P1 arrays. The method also includes, in a weight-update calculation operation, calculating updates to respective weights stored in each of the P1 arrays in dependence on signals propagated by the neuron layers.
    Type: Application
    Filed: March 14, 2019
    Publication date: September 17, 2020
    Inventors: Manuel Le Gallo-Bourdeau, Nandakumar Sasidharan Rajalekshmi, Christophe Piveteau, Irem Boybat Kara, Abu Sebastian, Evangelos Stavros Eleftheriou