Patents by Inventor AMBILI VENGALLUR

AMBILI VENGALLUR has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11640537
    Abstract: An apparatus to facilitate execution of non-linear functions operations is disclosed. The apparatus comprises accelerator circuitry including a compute grid having a plurality of processing elements to execute neural network computations, store values resulting from the neural network computations, and perform piecewise linear (PWL) approximations of one or more non-linear functions using the stored values as input data.
    Type: Grant
    Filed: April 8, 2019
    Date of Patent: May 2, 2023
    Assignee: Intel Corporation
    Inventors: Bharat Daga, Krishnakumar Nair, Pradeep Janedula, Aravind Babu Srinivasan, Bijoy Pazhanimala, Ambili Vengallur
  • Patent number: 11544191
    Abstract: Hardware accelerators for accelerated grouped convolution operations. A first buffer of a hardware accelerator may receive a first row of an input feature map (IFM) from a memory. A first group comprising a plurality of tiles may receive a first row of the IFM. A plurality of processing elements of the first group may compute a portion of a first row of an output feature map (OFM) based on the first row of the IFM and a kernel. A second buffer of the accelerator may receive a third row of the IFM from the memory. A second group comprising a plurality of tiles may receive the third row of the IFM. A plurality of processing elements of the second group may compute a portion of a third row of the OFM based on the third row of the IFM and the kernel as part of a grouped convolution operation.
    Type: Grant
    Filed: March 26, 2020
    Date of Patent: January 3, 2023
    Assignee: INTEL CORPORATION
    Inventors: Ambili Vengallur, Bharat Daga, Pradeep K. Janedula, Bijoy Pazhanimala, Aravind Babu Srinivasan
  • Publication number: 20220101091
    Abstract: A DNN accelerator includes a multiplication controller controlling whether to perform matrix computation based on weight values. The multiplication controller reads a weight matrix from a WRAM in the DNN accelerator and determines a row value for a row in the weight matrix. In an embodiment where the row value is one, a first switch sends a read request to the WRAM to read weights in the row and a second switch forms a data transmission path from an IRAM in the DNN accelerator to a PE in the DNN accelerator. The PE receives the weights and input data stored in the IRAM and performs MAC operations. In an embodiment where the row value is zero, the first and second switches are not triggered. No read request is sent to the WRAM and the data transmission path is not formed. The PE will not perform any MAC operations.
    Type: Application
    Filed: December 14, 2021
    Publication date: March 31, 2022
    Applicant: Intel Corporation
    Inventors: Srivatsa Rangachar Srinivasa, Jainaveen Sundaram Priya, Bradley A. Jackson, Ambili Vengallur, Dileep John Kurian, Tanay Karnik
  • Publication number: 20200320403
    Abstract: An apparatus to facilitate execution of non-linear functions operations is disclosed. The apparatus comprises accelerator circuitry including a compute grid having a plurality of processing elements to execute neural network computations, store values resulting from the neural network computations, and perform piecewise linear (PWL) approximations of one or more non-linear functions using the stored values as input data.
    Type: Application
    Filed: April 8, 2019
    Publication date: October 8, 2020
    Applicant: Intel Corporation
    Inventors: Bharat Daga, Krishnakumar Nair, Pradeep Janedula, Aravind Babu Srinivasan, Bijoy Pazhanimala, Ambili Vengallur
  • Patent number: 10769526
    Abstract: An apparatus to facilitate acceleration of machine learning operations is disclosed. The apparatus comprises accelerator circuitry including a first set of processing elements to perform first computations including matrix multiplication operations, a second set of processing elements to perform second computations including sum of elements of weights and offset multiply operations and a third set of processing elements to perform third computations including sum of elements of inputs and offset multiply operations, wherein the second and third computations are performed in parallel with the first computations.
    Type: Grant
    Filed: April 24, 2018
    Date of Patent: September 8, 2020
    Assignee: Intel Corporation
    Inventors: Bharat Daga, Pradeep Janedula, Aravind Babu Srinivasan, Ambili Vengallur
  • Publication number: 20200233803
    Abstract: Hardware accelerators for accelerated grouped convolution operations. A first buffer of a hardware accelerator may receive a first row of an input feature map (IFM) from a memory. A first group comprising a plurality of tiles may receive a first row of the IFM. A plurality of processing elements of the first group may compute a portion of a first row of an output feature map (OFM) based on the first row of the IFM and a kernel. A second buffer of the accelerator may receive a third row of the IFM from the memory. A second group comprising a plurality of tiles may receive the third row of the IFM. A plurality of processing elements of the second group may compute a portion of a third row of the OFM based on the third row of the IFM and the kernel as part of a grouped convolution operation.
    Type: Application
    Filed: March 26, 2020
    Publication date: July 23, 2020
    Applicant: Intel Corporation
    Inventors: AMBILI VENGALLUR, BHARAT DAGA, PRADEEP K. JANEDULA, BIJOY PAZHANIMALA, ARAVIND BABU SRINIVASAN
  • Publication number: 20200097799
    Abstract: Heterogeneous multiplier circuitry is provided with an interface to a configuration register to access configuration information, where the configuration information identifies respective data formats of a first operand and a second operand to be used in a first multiplication operation, where the first operand is in a first data format including a first numerical representation and the second operand is in a different, second data format including a different, second numerical representation. The heterogeneous multiplier circuitry includes an operand modifier to modify the second operand to generate a modified second operand, and further includes a multiplier to perform multiplication of the first operand and the modified second operand to generate a result in the first data format.
    Type: Application
    Filed: June 30, 2017
    Publication date: March 26, 2020
    Applicant: Intel Corporation
    Inventors: Dilin Divakar, Ambili Vengallur, Ajit Singh
  • Publication number: 20190325303
    Abstract: An apparatus to facilitate acceleration of machine learning operations is disclosed. The apparatus comprises accelerator circuitry including a first set of processing elements to perform first computations including matrix multiplication operations, a second set of processing elements to perform second computations including sum of elements of weights and offset multiply operations and a third set of processing elements to perform third computations including sum of elements of inputs and offset multiply operations, wherein the second and third computations are performed in parallel with the first computations.
    Type: Application
    Filed: April 24, 2018
    Publication date: October 24, 2019
    Applicant: Intel Corporation
    Inventors: BHARAT DAGA, PRADEEP JANEDULA, ARAVIND BABU SRINIVASAN, AMBILI VENGALLUR