Patents by Inventor Ujval Kapasi

Ujval Kapasi has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20210256348
    Abstract: Aspects of the present invention are directed to computer-implemented techniques for performing data compression and conversion between data formats of varying degrees of precision, and more particularly for improving the inferencing (application) of artificial neural networks using a reduced precision (e.g., INT8) data format. Embodiments of the present invention generate candidate conversions of data output, then employ a relative measure of quality to identify the candidate conversion with the greatest accuracy (i.e., least divergence from the original higher precision values). The representation can be then be used during inference to perform computations on the resulting output data.
    Type: Application
    Filed: May 3, 2021
    Publication date: August 19, 2021
    Inventors: Szymon Migacz, Hao Wu, Dilip Sequeira, Ujval Kapasi, Maxim Milakov, Slawomir Kierat, Zacky Zhou, Yilin Zhang, Alex Fit-Florea
  • Patent number: 10997492
    Abstract: Aspects of the present invention are directed to computer-implemented techniques for performing data compression and conversion between data formats of varying degrees of precision, and more particularly for improving the inferencing (application) of artificial neural networks using a reduced precision (e.g., INT8) data format. Embodiments of the present invention generate candidate conversions of data output, then employ a relative measure of quality to identify the candidate conversion with the greatest accuracy (i.e., least divergence from the original higher precision values). The representation can be then be used during inference to perform computations on the resulting output data.
    Type: Grant
    Filed: December 11, 2017
    Date of Patent: May 4, 2021
    Assignee: Nvidia Corporation
    Inventors: Szymon Migacz, Hao Wu, Dilip Sequeira, Ujval Kapasi, Maxim Milakov, Slawomir Kierat, Zacky Zhou, Yilin Zhang, Alex Fit-Florea
  • Publication number: 20180211152
    Abstract: Aspects of the present invention are directed to computer-implemented techniques for performing data compression and conversion between data formats of varying degrees of precision, and more particularly for improving the inferencing (application) of artificial neural networks using a reduced precision (e.g., INT8) data format. Embodiments of the present invention generate candidate conversions of data output, then employ a relative measure of quality to identify the candidate conversion with the greatest accuracy (i.e., least divergence from the original higher precision values). The representation can be then be used during inference to perform computations on the resulting output data.
    Type: Application
    Filed: December 11, 2017
    Publication date: July 26, 2018
    Inventors: Szymon Migacz, Hao Wu, Dilip Sequeira, Ujval Kapasi, Maxim Milakov, Slawomir Kierat, Zacky Zhou, Yilin Zhang, Alex Fit-Florea
  • Publication number: 20080117978
    Abstract: A method of estimating motion is disclosed. A first plurality of candidates is identified in a reference frame, wherein the total area occupied by the first plurality of candidates is substantially smaller than that of the reference frame. A first refinement search is then performed based, at least in part, on the first plurality of candidates. One or more best candidates are then identified based, at least in part, on the first refinement search. Finally, motion data is encoded based, at least in part, on the one or more best candidates.
    Type: Application
    Filed: October 9, 2007
    Publication date: May 22, 2008
    Inventors: Ujval Kapasi, Amit Gulati, John Sievers, Yipeng Liu, Dan Miller
  • Publication number: 20070150700
    Abstract: A processor implements conditional vector operations in which, for example, an input vector containing multiple operands to be used in conditional operations is divided into two or more output vectors based on a condition vector. Each output vector can then be processed at full processor efficiency without cycles wasted due to branch latency. Data to be processed are divided into two groups based on whether or not they satisfy a given condition by e.g., steering each to one of the two index vectors. Once the data have been segregated in this way, subsequent processing can be performed without conditional operations, processor cycles wasted due to branch latency, incorrect speculation or execution of unnecessary instructions due to predication. Other examples of conditional operations include combining one or more input vectors into a single output vector based on a condition vector, conditional vector switching, conditional vector combining, and conditional vector load balancing.
    Type: Application
    Filed: August 28, 2006
    Publication date: June 28, 2007
    Applicants: The Board of Trustees of the Leland Stanford Junior University, The Massachusetts Institute of Technology
    Inventors: William Dally, Scott Rixner, John Owens, Ujval Kapasi