Patents by Inventor Alex Fit-Florea

Alex Fit-Florea has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 12299577
    Abstract: Aspects of the present invention are directed to computer-implemented techniques for improving the training of artificial neural networks using a reduced precision (e.g., float16) data format. Embodiments of the present invention rescale tensor values prior to performing matrix operations (such as matrix multiplication or matrix addition) to prevent overflow and underflow. To preserve accuracy throughout the performance of the matrix operations, the scale factors are defined using a novel data format to represent tensors, wherein a matrix is represented by the tuple X, where X=(a, v[.]), wherein a is a float scale factor and v[.] are scaled values stored in the float16 format. The value of any element X[i] according to this data format would be equal to a*v[i].
    Type: Grant
    Filed: June 15, 2017
    Date of Patent: May 13, 2025
    Assignee: NVIDIA Corporation
    Inventors: Boris Ginsburg, Sergei Nikolaev, Ahmad Kiswani, Hao Wu, Amir Gholaminejad, Slawomir Kierat, Michael Houston, Alex Fit-Florea
  • Publication number: 20230418726
    Abstract: Methods and systems for comparing information obtained during execution of a workload to a set of inefficiency patterns, and determining the workload includes a potential inefficiency when the information matches at least one of the set of inefficiency patterns.
    Type: Application
    Filed: June 27, 2022
    Publication date: December 28, 2023
    Inventors: Szymon Migacz, Pawel Morkisz, Alex Fit-Florea, Maciej Bala, Jakub Zakrzewski, Trivikram Krishnamurthy, Nitin Nitin, Sangkug Lym, Shang Wang, Chenhan Yu, Alexandre Milesi
  • Publication number: 20210256348
    Abstract: Aspects of the present invention are directed to computer-implemented techniques for performing data compression and conversion between data formats of varying degrees of precision, and more particularly for improving the inferencing (application) of artificial neural networks using a reduced precision (e.g., INT8) data format. Embodiments of the present invention generate candidate conversions of data output, then employ a relative measure of quality to identify the candidate conversion with the greatest accuracy (i.e., least divergence from the original higher precision values). The representation can be then be used during inference to perform computations on the resulting output data.
    Type: Application
    Filed: May 3, 2021
    Publication date: August 19, 2021
    Inventors: Szymon Migacz, Hao Wu, Dilip Sequeira, Ujval Kapasi, Maxim Milakov, Slawomir Kierat, Zacky Zhou, Yilin Zhang, Alex Fit-Florea
  • Publication number: 20210232366
    Abstract: A method, computer readable medium, and system are disclosed for rounding floating point values. Dynamic directional rounding is a rounding technique for floating point operations. A floating point operation (addition, subtraction, multiplication, etc.) is performed on an operand to compute a floating point result. A sign (positive or negative) of the operand is identified. In one embodiment, the sign determines a direction in which the floating point result is rounded (towards negative or positive infinity). When used for updating parameters of a neural network during backpropagation, dynamic directional rounding ensures that rounding is performed in the direction of the gradient.
    Type: Application
    Filed: February 1, 2021
    Publication date: July 29, 2021
    Inventors: Alex Fit-Florea, Boris Ginsburg, Pooya Davoodi, Amir Gholaminejad
  • Patent number: 10997492
    Abstract: Aspects of the present invention are directed to computer-implemented techniques for performing data compression and conversion between data formats of varying degrees of precision, and more particularly for improving the inferencing (application) of artificial neural networks using a reduced precision (e.g., INT8) data format. Embodiments of the present invention generate candidate conversions of data output, then employ a relative measure of quality to identify the candidate conversion with the greatest accuracy (i.e., least divergence from the original higher precision values). The representation can be then be used during inference to perform computations on the resulting output data.
    Type: Grant
    Filed: December 11, 2017
    Date of Patent: May 4, 2021
    Assignee: Nvidia Corporation
    Inventors: Szymon Migacz, Hao Wu, Dilip Sequeira, Ujval Kapasi, Maxim Milakov, Slawomir Kierat, Zacky Zhou, Yilin Zhang, Alex Fit-Florea
  • Patent number: 10908878
    Abstract: A method, computer readable medium, and system are disclosed for rounding floating point values. Dynamic directional rounding is a rounding technique for floating point operations. A floating point operation (addition, subtraction, multiplication, etc.) is performed on an operand to compute a floating point result. A sign (positive or negative) of the operand is identified. In one embodiment, the sign determines a direction in which the floating point result is rounded (towards negative or positive infinity). When used for updating parameters of a neural network during backpropagation, dynamic directional rounding ensures that rounding is performed in the direction of the gradient.
    Type: Grant
    Filed: November 26, 2018
    Date of Patent: February 2, 2021
    Assignee: NVIDIA Corporation
    Inventors: Alex Fit-Florea, Boris Ginsburg, Pooya Davoodi, Amir Gholaminejad
  • Publication number: 20200167125
    Abstract: A method, computer readable medium, and system are disclosed for rounding floating point values. Dynamic directional rounding is a rounding technique for floating point operations. A floating point operation (addition, subtraction, multiplication, etc.) is performed on an operand to compute a floating point result. A sign (positive or negative) of the operand is identified. In one embodiment, the sign determines a direction in which the floating point result is rounded (towards negative or positive infinity). When used for updating parameters of a neural network during backpropagation, dynamic directional rounding ensures that rounding is performed in the direction of the gradient.
    Type: Application
    Filed: November 26, 2018
    Publication date: May 28, 2020
    Inventors: Alex Fit-Florea, Boris Ginsburg, Pooya Davoodi, Amir Gholaminejad
  • Publication number: 20180211152
    Abstract: Aspects of the present invention are directed to computer-implemented techniques for performing data compression and conversion between data formats of varying degrees of precision, and more particularly for improving the inferencing (application) of artificial neural networks using a reduced precision (e.g., INT8) data format. Embodiments of the present invention generate candidate conversions of data output, then employ a relative measure of quality to identify the candidate conversion with the greatest accuracy (i.e., least divergence from the original higher precision values). The representation can be then be used during inference to perform computations on the resulting output data.
    Type: Application
    Filed: December 11, 2017
    Publication date: July 26, 2018
    Inventors: Szymon Migacz, Hao Wu, Dilip Sequeira, Ujval Kapasi, Maxim Milakov, Slawomir Kierat, Zacky Zhou, Yilin Zhang, Alex Fit-Florea
  • Publication number: 20170372202
    Abstract: Aspects of the present invention are directed to computer-implemented techniques for improving the training of artificial neural networks using a reduced precision (e.g., float16) data format. Embodiments of the present invention rescale tensor values prior to performing matrix operations (such as matrix multiplication or matrix addition) to prevent overflow and underflow. To preserve accuracy throughout the performance of the matrix operations, the scale factors are defined using a novel data format to represent tensors, wherein a matrix is represented by the tuple X, where X=(a, v[.]), wherein a is a float scale factor and v[.] are scaled values stored in the float16 format. The value of any element X[i] according to this data format would be equal to a*v[i].
    Type: Application
    Filed: June 15, 2017
    Publication date: December 28, 2017
    Inventors: Boris GINSBURG, Sergei NIKOLAEV, Ahmad KISWANI, Hao WU, Amir GHOLAMINEJAD, Slawomir KIERAT, Michael HOUSTON, Alex FIT-FLOREA
  • Patent number: 9448806
    Abstract: A floating-point unit and a method of identifying exception cases in a floating-point unit. In one embodiment, the floating-point unit includes: (1) a floating-point computation circuit having a normal path and an exception path and operable to execute an operation on an operand and (2) a decision circuit associated with the normal path and the exception path and configured to employ a flush-to-zero mode of the floating-point unit to determine which one of the normal path and the exception path is appropriate for carrying out the operation on the operand.
    Type: Grant
    Filed: September 25, 2012
    Date of Patent: September 20, 2016
    Assignee: Nvidia Corporation
    Inventors: Marcin Andrychowicz, Alex Fit-Florea
  • Publication number: 20140089644
    Abstract: A floating-point unit and a method of identifying exception cases in a floating-point unit. In one embodiment, the floating-point unit includes: (1) a floating-point computation circuit having a normal path and an exception path and operable to execute an operation on an operand and (2) a decision circuit associated with the normal path and the exception path and configured to employ a flush-to-zero mode of the floating-point unit to determine which one of the normal path and the exception path is appropriate for carrying out the operation on the operand.
    Type: Application
    Filed: September 25, 2012
    Publication date: March 27, 2014
    Applicant: NVIDIA CORPORATION
    Inventors: Marcin Andrychowicz, Alex Fit-Florea