Patents by Inventor DMITRI VAINBRAND

DMITRI VAINBRAND has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20220365882
    Abstract: Apparatuses, systems, and techniques to control operation of a memory cache. In at least one embodiment, cache guidance is specified within application source code by associating guidance with declaration of a memory block, and then applying specified guidance to source code statements that access said memory block.
    Type: Application
    Filed: August 5, 2021
    Publication date: November 17, 2022
    Inventors: Harold Carter Edwards, Luke David Durant, Stephen Jones, Jack H. Choquette, Ronny Krashinsky, Dmitri Vainbrand, Olivier Giroux, Olivier Francois Joseph Harel, Shirish Gadre, Ze Long, Matthieu Tardy, David Dastous St Hilaire, Gokul Ramaswamy Hirisave Chandra Shekhara, Jaydeep Marathe, Jaewook Shin, Jayashree Venkatesh, Girish Bhaskar Bharambe
  • Publication number: 20220027704
    Abstract: Methods and apparatus relating to techniques for incremental network quantization. In an example, an apparatus comprises logic, at least partially comprising hardware logic to determine a plurality of weights for a layer of a convolutional neural network (CNN) comprising a plurality of kernels; organize the plurality of weights into a plurality of clusters for the plurality of kernels; and apply a K-means compression algorithm to each of the plurality of clusters. Other embodiments are also disclosed and claimed.
    Type: Application
    Filed: July 2, 2021
    Publication date: January 27, 2022
    Applicant: Intel Corporation
    Inventors: Yonatan Glesner, Gal Novik, Dmitri Vainbrand, Gal Leibovich
  • Patent number: 11055604
    Abstract: Methods and apparatus relating to techniques for incremental network quantization. In an example, an apparatus comprises logic, at least partially comprising hardware logic to determine a plurality of weights for a layer of a convolutional neural network (CNN) comprising a plurality of kernels; organize the plurality of weights into a plurality of clusters for the plurality of kernels; and apply a K-means compression algorithm to each of the plurality of clusters. Other embodiments are also disclosed and claimed.
    Type: Grant
    Filed: September 12, 2017
    Date of Patent: July 6, 2021
    Assignee: INTEL CORPORATION
    Inventors: Yonatan Glesner, Gal Novik, Dmitri Vainbrand, Gal Leibovich
  • Patent number: 10726583
    Abstract: Embodiments described herein provide a processing apparatus comprising compute logic to generate output feature map data for a convolutional neural network (CNN) and write the feature map data to a memory buffer; a direct memory access (DMA) controller including a feature map encoder, the DMA controller to read the feature map data from the memory buffer, encode the feature map data using one of multiple encode algorithms, and write encoded feature map data to memory coupled with the processing apparatus; and wherein the compute logic is to read the encoded feature map data from the memory in an encoded format and decode the encoded feature map data while reading the encoded feature map data.
    Type: Grant
    Filed: December 30, 2016
    Date of Patent: July 28, 2020
    Assignee: INTEL CORPORATION
    Inventors: Ajit Singh, Bharat Daga, Oren Agam, Michael Behar, Dmitri Vainbrand
  • Publication number: 20190080222
    Abstract: Methods and apparatus relating to techniques for incremental network quantization. In an example, an apparatus comprises logic, at least partially comprising hardware logic to determine a plurality of weights for a layer of a convolutional neural network (CNN) comprising a plurality of kernels; organize the plurality of weights into a plurality of clusters for the plurality of kernels; and apply a K-means compression algorithm to each of the plurality of clusters. Other embodiments are also disclosed and claimed.
    Type: Application
    Filed: September 12, 2017
    Publication date: March 14, 2019
    Inventors: Yonatan Glesner, Gal Novik, Dmitri Vainbrand, Gal Leibovich
  • Patent number: 10210137
    Abstract: A processor, including: decode circuitry to decode instructions; a data cache unit including circuitry to cache data for the processor; and an approximate matrix multiplication (AMM) circuit including: a data receptor circuit to receive a weight vector w and an input vector x, both of size N, and a compression regulating parameter n; a factorizer circuit to factorize w into w?B·s, by computing a binary factorized matrix B of size N×n, and a dictionary vector s of size n; and a binary multiplier circuit to compute w^T x?(B·s)^T x=s^T(B^T x), the binary multiplier circuit comprising a hardware accelerator circuit to compute an array product B^T x).
    Type: Grant
    Filed: June 28, 2017
    Date of Patent: February 19, 2019
    Assignee: Intel Corporation
    Inventors: Ehud Cohen, Daniel David Ben-Dayan Rubin, Michael Behar, Dmitri Vainbrand
  • Publication number: 20190004997
    Abstract: A processor, including: decode circuitry to decode instructions; a data cache unit including circuitry to cache data for the processor; and an approximate matrix multiplication (AMM) circuit including: a data receptor circuit to receive a weight vector w and an input vector x, both of size N, and a compression regulating parameter n; a factorizer circuit to factorize w into w?B·s, by computing a binary factorized matrix B of size N×n, and a dictionary vector s of size n; and a binary multiplier circuit to compute w?T x?(B·s)?T x=s?T (B)?T x), the binary multiplier circuit comprising a hardware accelerator circuit to compute an array product B?T x).
    Type: Application
    Filed: June 28, 2017
    Publication date: January 3, 2019
    Applicant: Intel Corporation
    Inventors: Ehud Cohen, Daniel David Ben-Dayan Rubin, Michael Behar, Dmitri Vainbrand
  • Publication number: 20180189981
    Abstract: Embodiments described herein provide a processing apparatus comprising compute logic to generate output feature map data for a convolutional neural network (CNN) and write the feature map data to a memory buffer; a direct memory access (DMA) controller including a feature map encoder, the DMA controller to read the feature map data from the memory buffer, encode the feature map data using one of multiple encode algorithms, and write encoded feature map data to memory coupled with the processing apparatus; and wherein the compute logic is to read the encoded feature map data from the memory in an encoded format and decode the encoded feature map data while reading the encoded feature map data.
    Type: Application
    Filed: December 30, 2016
    Publication date: July 5, 2018
    Inventors: AJIT SINGH, BHARAT DAGA, OREN AGAM, MICHAEL BEHAR, DMITRI VAINBRAND