Patents by Inventor Avi Baum

Avi Baum has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11263512
    Abstract: A novel and useful neural network (NN) processing core adapted to implement artificial neural networks (ANNs) and incorporating strictly separate control and data planes. The NN processor is constructed from self-contained computational units organized in a hierarchical architecture. The homogeneity enables simpler management and control of similar computational units, aggregated in multiple levels of hierarchy. Computational units are designed with minimal overhead as possible, where additional features and capabilities are aggregated at higher levels in the hierarchy. On-chip memory provides storage for content inherently required for basic operation at a particular hierarchy and is coupled with the computational resources in an optimal ratio. Lean control provides just enough signaling to manage only the operations required at a particular hierarchical level. Dynamic resource assignment agility is provided which can be adjusted as required depending on resource availability and capacity of the device.
    Type: Grant
    Filed: April 3, 2018
    Date of Patent: March 1, 2022
    Inventors: Avi Baum, Or Danon, Hadar Zeitlin, Daniel Ciubotariu, Rami Feig
  • Patent number: 11263077
    Abstract: Novel and useful system and methods of several functional safety mechanisms for use in an artificial neural network (ANN) processor. The mechanisms can be deployed individually or in combination to provide a desired level of safety in neural networks. Multiple strategies are applied involving redundancy by design, redundancy through spatial mapping as well as self-tuning procedures that modify static (weights) and monitor dynamic (activations) behavior. The various mechanisms of the present invention address ANN system level safety in situ, as a system level strategy that is tightly coupled with the processor architecture. The NN processor incorporates several functional safety concepts which reduce its risk of failure that occurs during operation from going unnoticed. The mechanisms function to detect and promptly flag and report the occurrence of an error with some mechanisms capable of correction as well.
    Type: Grant
    Filed: September 29, 2020
    Date of Patent: March 1, 2022
    Inventors: Roi Seznayov, Guy Kaminitz, Daniel Chibotero, Ori Katz, Amir Shmul, Yuval Adelstein, Nir Engelberg, Or Danon, Avi Baum
  • Patent number: 11237894
    Abstract: Novel and useful system and methods of several functional safety mechanisms for use in an artificial neural network (ANN) processor. The mechanisms can be deployed individually or in combination to provide a desired level of safety in neural networks. Multiple strategies are applied involving redundancy by design, redundancy through spatial mapping as well as self-tuning procedures that modify static (weights) and monitor dynamic (activations) behavior. The various mechanisms of the present invention address ANN system level safety in situ, as a system level strategy that is tightly coupled with the processor architecture. The NN processor incorporates several functional safety concepts which reduce its risk of failure that occurs during operation from going unnoticed. The mechanisms function to detect and promptly flag and report the occurrence of an error with some mechanisms capable of correction as well.
    Type: Grant
    Filed: September 29, 2020
    Date of Patent: February 1, 2022
    Inventors: Avi Baum, Roi Seznayov, Daniel Chibotero, Ori Katz, Guy Kaminitz, Nir Engelberg, Yuval Adelstein, Itai Resh, Or Danon
  • Patent number: 11238331
    Abstract: A novel and useful augmented artificial neural network (ANN) incorporating an existing artificial neural network (ANN) coupled to a supplemental ANN and a first-in first-out (FIFO) stack for storing historical output values of the network. The augmented ANN exploits the redundant nature of information present in an input data stream. The addition of the supplemental ANN along with a FIFO enables the augmented network to look back into the past in making a decision for the current frame. It provides context aware object presence as well as lowers the rate of false detections and misdetections. The output of the existing ANN is stored in a FIFO to create a lookahead system in which both past output values of the supplemental ANN and ‘future’ values of the output of the existing ANN are used in making a decision for the current frame. In addition, the mechanism does not require retraining the entire neural network nor does it require data set labeling.
    Type: Grant
    Filed: October 26, 2017
    Date of Patent: February 1, 2022
    Inventors: Avi Baum, Or Danon
  • Patent number: 11238334
    Abstract: A novel and useful system and method of input alignment for streamlining vector operations that reduce the required memory read bandwidth. The input aligner as deployed in the NN processor, functions to facilitate the reuse of data read from memory and to avoid having to re-read that data in the context of neural network calculations. The input aligner functions to distribute input data (or weights) to the appropriate compute elements while consuming input data in a single cycle. Thus, the input aligner is operative to lower the required read bandwidth of layer input in an ANN. This reflects the fact that normally in practice, a vector multiplication is performed every time instance. This considers the fact that in many native calculations that take place in an ANN, the same data point is involved in multiple calculations.
    Type: Grant
    Filed: September 12, 2019
    Date of Patent: February 1, 2022
    Inventors: Avi Baum, Or Danon, Daniel Ciubotariu
  • Patent number: 11221929
    Abstract: Novel and useful system and methods of several functional safety mechanisms for use in an artificial neural network (ANN) processor. The mechanisms can be deployed individually or in combination to provide a desired level of safety in neural networks. Multiple strategies are applied involving redundancy by design, redundancy through spatial mapping as well as self-tuning procedures that modify static (weights) and monitor dynamic (activations) behavior. The various mechanisms of the present invention address ANN system level safety in situ, as a system level strategy that is tightly coupled with the processor architecture. The NN processor incorporates several functional safety concepts which reduce its risk of failure that occurs during operation from going unnoticed. The mechanisms function to detect and promptly flag and report the occurrence of an error with some mechanisms capable of correction as well.
    Type: Grant
    Filed: September 29, 2020
    Date of Patent: January 11, 2022
    Inventors: Ori Katz, Avi Baum, Guy Kaminitz, Daniel Chibotero, Or Danon, Roi Seznayov, Itai Resh
  • Patent number: 11216717
    Abstract: A novel and useful neural network (NN) processing core adapted to implement artificial neural networks (ANNs). The NN processor is constructed from self-contained computational units organized in a hierarchical architecture. The homogeneity enables simpler management and control of similar computational units, aggregated in multiple levels of hierarchy. Computational units are designed with minimal overhead as possible, where additional features and capabilities are aggregated at higher levels in the hierarchy. On-chip memory provides storage for content inherently required for basic operation at a particular hierarchy and is coupled with the computational resources in an optimal ratio. Lean control provides just enough signaling to manage only the operations required at a particular hierarchical level. Dynamic resource assignment agility is provided which can be adjusted as required depending on resource availability and capacity of the device.
    Type: Grant
    Filed: April 3, 2018
    Date of Patent: January 4, 2022
    Inventors: Avi Baum, Or Danon, Hadar Zeitlin, Daniel Ciubotariu, Rami Feig
  • Publication number: 20210160139
    Abstract: A Wi-Fi device includes a controller coupled to a writeable memory implementing a MAC and PHY layer and to a transceiver. Connection data stored in the writeable memory includes Wi-Fi connection parameters including ?1 router MAC level information or a most recently utilized (MRU) channel used, and IP addresses including ?1 of an IP address of the Wi-Fi device, IP address of the MRU router, an IP address of a MRU target server, and an IP address of a network connected device. An accelerated reconnecting to a Wi-Fi network algorithm is implemented by the processor is for starting from being in a network disconnected state, establishing current connection parameters for a current Wi-Fi network connection using the Wi-Fi connection parameters for at least one MAC layer parameter for the MAC layer.
    Type: Application
    Filed: February 3, 2021
    Publication date: May 27, 2021
    Inventors: YANIV TZOREFF, GILBOA SHVEKI, AVI BAUM, BARAK CHERCHES
  • Patent number: 10944632
    Abstract: A Wi-Fi device includes a controller coupled to a writeable memory implementing a MAC and PHY layer and to a transceiver. Connection data stored in the writeable memory includes Wi-Fi connection parameters including ?1 router MAC level information or a most recently utilized (MRU) channel used, and IP addresses including ?1 of an IP address of the Wi-Fi device, IP address of the MRU router, an IP address of a MRU target server, and an IP address of a network connected device. An accelerated reconnecting to a Wi-Fi network algorithm is implemented by the processor is for starting from being in a network disconnected state, establishing current connection parameters for a current Wi-Fi network connection using the Wi-Fi connection parameters for at least one MAC layer parameter for the MAC layer.
    Type: Grant
    Filed: June 22, 2017
    Date of Patent: March 9, 2021
    Assignee: TEXAS INSTRUMENTS INCORPORATED
    Inventors: Yaniv Tzoreff, Gilboa Shveki, Avi Baum, Barak Cherches
  • Publication number: 20200305079
    Abstract: In a described example, an integrated circuit includes an input coupled to receive a plurality of beacon frames, the beacon frames include an indication of data transmissions available for a device that includes the integrated circuit. The integrated circuit also includes a controller configured to compare the plurality of beacon frames to determine a plurality of bytes prior to the indication of data transmission available that is present in each of the plurality of beacon frames and is configured to provide a signal indicating a low power mode in which the device does not receive transmissions and to provide a signal indicating a wake mode at a selected time before transmission of the plurality of bytes in a subsequent beacon transmission.
    Type: Application
    Filed: June 9, 2020
    Publication date: September 24, 2020
    Inventors: Oran Naftali, Avi Baum, Yuval Jakira, Asaf Even-Chen
  • Publication number: 20200285892
    Abstract: A novel and useful system and method of improved power performance and lowered memory requirements for an artificial neural network based on packing memory utilizing several structured sparsity mechanisms. The invention applies to neural network (NN) processing engines adapted to implement mechanisms to search for structured sparsity in weights and activations, resulting in a considerably reduced memory usage. The sparsity guided training mechanism synthesizes and generates structured sparsity weights A compiler mechanism within a software development kit (SDK), manipulates structured weight domain sparsity to generate a sparse set of static weights for the NN. The structured sparsity static weights are loaded into the NN after compilation and utilized by both the structured weight domain sparsity mechanism and the structured activation domain sparsity mechanism.
    Type: Application
    Filed: May 21, 2020
    Publication date: September 10, 2020
    Applicant: Hailo Technologies Ltd.
    Inventors: Avi Baum, Or Danon, Daniel Chibotero, Gilad Nahor
  • Publication number: 20200285950
    Abstract: A novel and useful system and method of improved power performance and lowered memory requirements for an artificial neural network based on packing memory utilizing several structured sparsity mechanisms. The invention applies to neural network (NN) processing engines adapted to implement mechanisms to search for structured sparsity in weights and activations, resulting in a considerably reduced memory usage. The sparsity guided training mechanism synthesizes and generates structured sparsity weights A compiler mechanism within a software development kit (SDK), manipulates structured weight domain sparsity to generate a sparse set of static weights for the NN. The structured sparsity static weights are loaded into the NN after compilation and utilized by both the structured weight domain sparsity mechanism and the structured activation domain sparsity mechanism.
    Type: Application
    Filed: May 21, 2020
    Publication date: September 10, 2020
    Applicant: Hailo Technologies Ltd.
    Inventors: Avi Baum, Or Danon, Daniel Chibotero
  • Publication number: 20200285949
    Abstract: A novel and useful system and method of improved power performance and lowered memory requirements for an artificial neural network based on packing memory utilizing several structured sparsity mechanisms. The invention applies to neural network (NN) processing engines adapted to implement mechanisms to search for structured sparsity in weights and activations, resulting in a considerably reduced memory usage. The sparsity guided training mechanism synthesizes and generates structured sparsity weights A compiler mechanism within a software development kit (SDK), manipulates structured weight domain sparsity to generate a sparse set of static weights for the NN. The structured sparsity static weights are loaded into the NN after compilation and utilized by both the structured weight domain sparsity mechanism and the structured activation domain sparsity mechanism.
    Type: Application
    Filed: May 21, 2020
    Publication date: September 10, 2020
    Applicant: Hailo Technologies Ltd.
    Inventors: Avi Baum, Or Danon, Daniel Chibotero, Gilad Nahor
  • Publication number: 20200279133
    Abstract: A novel and useful system and method of improved power performance and lowered memory requirements for an artificial neural network based on packing memory utilizing several structured sparsity mechanisms. The invention applies to neural network (NN) processing engines adapted to implement mechanisms to search for structured sparsity in weights and activations, resulting in a considerably reduced memory usage. The sparsity guided training mechanism synthesizes and generates structured sparsity weights A compiler mechanism within a software development kit (SDK), manipulates structured weight domain sparsity to generate a sparse set of static weights for the NN. The structured sparsity static weights are loaded into the NN after compilation and utilized by both the structured weight domain sparsity mechanism and the structured activation domain sparsity mechanism.
    Type: Application
    Filed: May 21, 2020
    Publication date: September 3, 2020
    Applicant: Hailo Technologies Ltd.
    Inventors: Avi Baum, Or Danon, Daniel Chibotero
  • Patent number: 10708859
    Abstract: In a described example, an integrated circuit includes an input coupled to receive a plurality of beacon frames, the beacon frames include an indication of data transmissions available for a device that includes the integrated circuit. The integrated circuit also includes a controller configured to compare the plurality of beacon frames to determine a plurality of bytes prior to the indication of data transmission available that is present in each of the plurality of beacon frames and is configured to provide a signal indicating a low power mode in which the device does not receive transmissions and to provide a signal indicating a wake mode at a selected time before transmission of the plurality of bytes in a subsequent beacon transmission.
    Type: Grant
    Filed: November 28, 2016
    Date of Patent: July 7, 2020
    Assignee: TEXAS INSTRUMENTS INCORPORATED
    Inventors: Oran Naftali, Avi Baum, Yuval Jakira, Asaf Even-Chen
  • Publication number: 20200005127
    Abstract: A novel and useful system and method of input alignment for streamlining vector operations that reduce the required memory read bandwidth. The input aligner as deployed in the NN processor, functions to facilitate the reuse of data read from memory and to avoid having to re-read that data in the context of neural network calculations. The input aligner functions to distribute input data (or weights) to the appropriate compute elements while consuming input data in a single cycle. Thus, the input aligner is operative to lower the required read bandwidth of layer input in an ANN. This reflects the fact that normally in practice, a vector multiplication is performed every time instance. This considers the fact that in many native calculations that take place in an ANN, the same data point is involved in multiple calculations.
    Type: Application
    Filed: September 12, 2019
    Publication date: January 2, 2020
    Inventors: Avi Baum, Or Danon, Daniel Ciubotariu
  • Patent number: 10387298
    Abstract: A novel and useful artificial neural network that incorporates emphasis and focus techniques to extract more information from one or more portions of an input image compared to the rest of the image. The ANN recognizes that valuable information in an input image is typically not distributed throughout the image but rather is concentrated in one or more regions. Rather than implement CNN layers sequentially (i.e. row by row) on the input domain of each layer, the present invention leverages the fact that valuable information is focused in one or more regions of the image where it is desirable to apply more attention and for which it is desired to apply more elaborate evaluation. Precision dilution can be applied to those portions of the input image that are not the center of focus and emphasis. A spatial aware function determines the location(s) of the ears of focus and is applied to the first convolutional layer.
    Type: Grant
    Filed: August 6, 2017
    Date of Patent: August 20, 2019
    Assignee: Hailo Technologies Ltd
    Inventors: Avi Baum, Or Danon, Mark Grobman, Hadar Zeitlin
  • Publication number: 20180375729
    Abstract: A Wi-Fi device includes a controller coupled to a writeable memory implementing a MAC and PHY layer and to a transceiver. Connection data stored in the writeable memory includes Wi-Fi connection parameters including ?1 router MAC level information or a most recently utilized (MRU) channel used, and IP addresses including ?1 of an IP address of the Wi-Fi device, IP address of the MRU router, an IP address of a MRU target server, and an IP address of a network connected device. An accelerated reconnecting to a Wi-Fi network algorithm is implemented by the processor is for starting from being in a network disconnected state, establishing current connection parameters for a current Wi-Fi network connection using the Wi-Fi connection parameters for at least one MAC layer parameter for the MAC layer.
    Type: Application
    Filed: June 22, 2017
    Publication date: December 27, 2018
    Inventors: YANIV TZOREFF, GILBOA SHVEKI, AVI BAUM, BARAK CHERCHES
  • Publication number: 20180285678
    Abstract: A novel and useful artificial neural network that incorporates emphasis and focus techniques to extract more information from one or more portions of an input image compared to the rest of the image. The ANN recognizes that valuable information in an input image is typically not distributed throughout the image but rather is concentrated in one or more regions. Rather than implement CNN layers sequentially (i.e. row by row) on the input domain of each layer, the present invention leverages the fact that valuable information is focused in one or more regions of the image where it is desirable to apply more attention and for which it is desired to apply more elaborate evaluation. Precision dilution can be applied to those portions of the input image that are not the center of focus and emphasis. A spatial aware function determines the location(s) of the ears of focus and is applied to the first convolutional layer.
    Type: Application
    Filed: August 6, 2017
    Publication date: October 4, 2018
    Applicant: Hailo Technologies Ltd.
    Inventors: Avi Baum, Or Danon, Mark Grobman, Hadar Zeitlin
  • Publication number: 20180285718
    Abstract: A novel and useful neural network (NN) processing core adapted to implement artificial neural networks (ANNs). The NN processor is constructed from self-contained computational units organized in a hierarchical architecture. The homogeneity enables simpler management and control of similar computational units, aggregated in multiple levels of hierarchy. Computational units are designed with minimal overhead as possible, where additional features and capabilities are aggregated at higher levels in the hierarchy. On-chip memory provides storage for content inherently required for basic operation at a particular hierarchy and is coupled with the computational resources in an optimal ratio. Lean control provides just enough signaling to manage only the operations required at a particular hierarchical level. Dynamic resource assignment agility is provided which can be adjusted as required depending on resource availability and capacity of the device.
    Type: Application
    Filed: April 3, 2018
    Publication date: October 4, 2018
    Applicant: Hailo Technologies Ltd.
    Inventors: Avi Baum, Or Danon, Hadar Zeitlin, Daniel Ciubotariu, Rami Feig