Patents Examined by Michael D Yaary
  • Patent number: 11269629
    Abstract: Many signal processing, machine learning and scientific computing applications require a large number of multiply-accumulate (MAC) operations. This type of operation is demanding in both computation and memory. Process in memory has been proposed as a new technique that computes directly on a large array of data in place, to eliminate expensive data movement overhead. To enable parallel multi-bit MAC operations, both width- and level-modulating memory word lines are applied. To improve performance and provide tolerance against process-voltage-temperature variations, a delay-locked loop is used to generate fine unit pulses for driving memory word lines and a dual-ramp Single-slope ADC is used to convert bit line outputs. The concept is prototyped in a 180 nm CMOS test chip made of four 320×64 compute-SRAMs, each supporting 128× parallel 5 b×5 b MACs with 32 5 b output ADCs and consuming 16.6 mW at 200 MHz.
    Type: Grant
    Filed: November 29, 2018
    Date of Patent: March 8, 2022
    Assignee: The Regents of the University of Michigan
    Inventors: Zhengya Zhang, Thomas Chen, Jacob Christopher Botimer, Shiming Song
  • Patent number: 11262982
    Abstract: A computation circuit includes a plurality of processing elements and a common accumulator. The plurality of processing elements are sequentially coupled in series, and performs a multiply and accumulate (MAC) operation on a weight signal and at least one of two or more input signals received in each unit cycle. The common accumulator is sequentially and cyclically coupled to first to Kth processing elements among the plurality of processing elements, and configured to receive a computation value outputted from a processing element coupled thereto among the first to Kth processing elements, and store computation information. The K is decided based on values of the two or more input signals and the number of guard bits included in one processing element.
    Type: Grant
    Filed: July 22, 2019
    Date of Patent: March 1, 2022
    Assignees: SK hynix Inc., SK Telecom Co., Ltd.
    Inventors: Yong Sang Park, Seok Joong Hwang
  • Patent number: 11250103
    Abstract: A system for determining the frequency coefficients of a one or multi-dimensional signal that is sparse in the frequency domain includes determining the locations of the non-zero frequency coefficients, and then determining values of the coefficients using the determined locations. If N is total number of frequency coefficients across the one or more dimension of the signal, and if R is an upper bound of the number of non-zero ones of these frequency coefficients, the systems requires up to (O (R log(R) (N))) samples and has a computation complexity of up to O (R log2(R) log (N). The system and the processing technique are stable to low-level noise and can exhibit only a small probability of failure. The frequency coefficients can be real and positive or they can be complex numbers.
    Type: Grant
    Filed: January 25, 2017
    Date of Patent: February 15, 2022
    Assignee: Reservoir Labs, Inc.
    Inventor: Pierre-David Letourneau
  • Patent number: 11243744
    Abstract: A method (40) is provided for performing a trustworthiness test on a random number generator, RNG, (20) comprising a physical unclonable function, PUF-module (21). The trustworthiness test is implemented as a known answer test, KAT, and the method (40) comprises: receiving (41), in the PUF-module (21), an input based on test data, T, received from a verifier (11) provided with at least one test data-test result pair, (T, R), providing (42) an output from the PUF-module (21), determining (43) a test result, R?, based on the output from the PUF-module (21), and providing (44) the test result, R?, to the verifier (11). A random number generator (20), computer program and computer program products and a method performed by or in a verifier are also provided.
    Type: Grant
    Filed: November 15, 2016
    Date of Patent: February 8, 2022
    Assignee: Telefonaktiebolaget LM Ericsson (publ)
    Inventors: Mats Näslund, Elena Dubrova, Karl Norrman
  • Patent number: 11232174
    Abstract: Techniques and systems for solving Boolean satisfiability (SAT) problems are described. Some embodiments solve SAT problems using efficient construction of truth tables. Some embodiments can improve performance of SAT solvers by using truth tables instead of incurring the overhead of Conjunctive Normal Form (CNF) conversion.
    Type: Grant
    Filed: September 12, 2019
    Date of Patent: January 25, 2022
    Assignee: Synopsys, Inc.
    Inventor: Dmitry Korchemny
  • Patent number: 11226791
    Abstract: An arithmetic processing device has, when any or both of a first operand and a second operand included in a multiply-add operation instruction is or are zero, an exponent setting circuit sets an exponent of the first operand to a first set value, and sets an exponent of the second operand to a second set value. An exponent calculation circuit calculates an exponent obtained by a multiply-add operation, based on the exponents of the first and second operands outputted by the exponent setting circuit and an exponent of a third operand included in the multiply-add operation instruction. The sum of the first set value and the second set value is set so that a bit position of the third operand is located on a higher-order bit side than the most significant bit of the sum of the first operand and the second operand.
    Type: Grant
    Filed: October 2, 2019
    Date of Patent: January 18, 2022
    Assignee: FUJITSU LIMITED
    Inventors: Takio Ono, Hiroyuki Wada
  • Patent number: 11222257
    Abstract: Some embodiments provide a neural network inference circuit (NNIC) for executing a neural network. The NNIC includes a first circuit that outputs dot products for computation nodes of a first set of neural network layers, that include dot product computations of sets of weight values with sets of input values. The NNIC also includes a second circuit that outputs values for computation nodes of a second set of neural network layers, that apply a set of calculations that do not include dot products to sets of input values. The NNIC also includes a selection circuit that selects a dot product output from the first circuit when a current layer being processed by the NNIC belongs to the first set of layers, and selects a non-dot product output from the second circuit when the current layer belongs to the second set of layers.
    Type: Grant
    Filed: August 21, 2019
    Date of Patent: January 11, 2022
    Assignee: PERCEIVE CORPORATION
    Inventors: Jung Ko, Kenneth Duong, Steven L. Teig
  • Patent number: 11216534
    Abstract: An information processing apparatus includes a data acquisition unit that acquires data including a missing value, a missing rate calculation unit that calculates a missing rate indicating a ratio of missing values included in the data, and a covariance matrix estimation unit that estimates a covariance matrix based on the missing rate. According to the information processing apparatus, since the covariance matrix is estimated based on the missing rate, the estimation accuracy of the covariance matrix can be improved.
    Type: Grant
    Filed: March 8, 2019
    Date of Patent: January 4, 2022
    Assignee: KABUSHIKI KAISHA TOSHIBA
    Inventors: Masaaki Takada, Hironori Fujisawa, Takeichiro Nishikawa
  • Patent number: 11216532
    Abstract: The present disclosure relates generally to techniques for efficiently performing operations associated with artificial intelligence (AI), machine learning (ML), and/or deep learning (DL) applications, such as training and/or interference calculations, using an integrated circuit device. More specifically, the present disclosure relates to an integrated circuit design implemented to perform these operations with low latency and/or a high bandwidth of data. For example, embodiments of a computationally dense digital signal processing (DSP) circuitry, implemented to efficiently perform one or more arithmetic operations (e.g., a dot-product) on an input are disclosed. Moreover, embodiments described herein may relate to layout, design, and data scheduling of a processing element array implemented to compute matrix multiplications (e.g., systolic array multiplication).
    Type: Grant
    Filed: March 29, 2019
    Date of Patent: January 4, 2022
    Assignee: Intel Corporation
    Inventors: Martin Langhammer, Andrei-Mihai Hagiescu-Miriste
  • Patent number: 11216275
    Abstract: The embodiments herein describe a conversion engine that converts floating point data into integer data using a dynamic scaling factor. To select the scaling factor, the conversion engine compares a default (or initial) scaling factor value to an exponent portion of the floating point value to determine a shift value with which to bit shift a mantissa of the floating point value. After bit shifting the mantissa, the conversion engine determines whether the shift value caused an overflow or an underflow and whether that overflow or underflow violates a predefined policy. If the policy is violated, the conversion engine adjusts the scaling factor and restarts the conversion process. In this manner, the conversion engine can adjust the scaling factor until identifying a scaling factor that converts all the floating point values in the batch without violating the policy.
    Type: Grant
    Filed: August 5, 2019
    Date of Patent: January 4, 2022
    Assignee: XILINX, INC.
    Inventors: Philip B. James-Roxby, Eric F. Dellinger
  • Patent number: 11210586
    Abstract: Some embodiments provide a method for a neural network inference circuit that executes a neural network including multiple computation nodes at multiple layers. Each computation node of a set of the computation nodes includes a dot product of input values and weight values. The method reads a set of encoded weight data for a set of weight values from a memory of the neural network inference circuit. The method decodes the encoded weight data to generate decoded weight data for the set of weight values. The method stores the decoded weight data in a buffer. The method uses the decoded weight data to execute a set of computation nodes. Each computation node of the set of computation nodes includes a dot product between the set of weight values and a different set of input values.
    Type: Grant
    Filed: June 28, 2019
    Date of Patent: December 28, 2021
    Assignee: PERCEIVE CORPORATION
    Inventors: Kenneth Duong, Jung Ko, Steven L. Teig
  • Patent number: 11204740
    Abstract: The neuromorphic arithmetic device performs a multiply-accumulate (MAC) calculation using a multiplier and an accumulator. The neuromorphic arithmetic device includes an offset accumulator configured to receive a plurality of offset data measured a plurality of times and accumulate the plurality of offset data, a bit extractor configured to obtain average offset data by extracting at least one first bit from the plurality of accumulated offset data, and a cumulative synapse array configured to accumulate a plurality of multiplication values generated by the multiplier and output a cumulative result of the plurality of multiplication values corrected according to the average offset data.
    Type: Grant
    Filed: November 26, 2019
    Date of Patent: December 21, 2021
    Assignee: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE
    Inventors: Min-Hyung Cho, Young-deuk Jeon, Ki Hyuk Park, Joo Hyun Lee
  • Patent number: 11194886
    Abstract: Various arrangements for performing vector-matrix multiplication are provided here. Digital input vectors that include binary-encoded values can be converted into a plurality of analog signals using a plurality of one-bit digital to analog converters (DACs). Using an analog vector matrix multiplier, a vector-matrix multiplication operation can be performed using a weighting matrix for each bit-order of the plurality of analog signals. For each performed vector-matrix multiplication operation, a bit-ordered indication of an output of the analog vector matrix multiplier may be stored. A bit-order weighted summation of the sequentially performed vector-matrix multiplication operation may be performed.
    Type: Grant
    Filed: May 9, 2019
    Date of Patent: December 7, 2021
    Assignee: Applied Materials, Inc.
    Inventors: She-Hwa Yen, Frank Tzen-Wen Guo
  • Patent number: 11188305
    Abstract: A computation device includes: a data multiplexer configured to output first high-order data as first output data and fifth output data, output first low-order data as third output data and seventh output data, output second high-order data as second output data, output second low-order data as fourth output data, output third high-order data, which is high-order data having a second bit number out of third input data, as sixth output data, and output third low-order data, which is low-order data having the second bit number out of the third input data, as eighth output data when a mode signal indicates a second computation mode; and first to fourth multipliers each of which multiplies two output data.
    Type: Grant
    Filed: May 11, 2018
    Date of Patent: November 30, 2021
    Assignees: Preferred Networks, Inc., Riken
    Inventors: Junichiro Makino, Takayuki Muranushi, Miyuki Tsubouchi, Ken Namura
  • Patent number: 11188618
    Abstract: An apparatus to facilitate acceleration of matrix multiplication operations. The apparatus comprises a systolic array including matrix multiplication hardware to perform multiply-add operations on received matrix data comprising data from a plurality of input matrices and sparse matrix acceleration hardware to detect zero values in the matrix data and perform one or more optimizations on the matrix data to reduce multiply-add operations to be performed by the matrix multiplication hardware.
    Type: Grant
    Filed: September 5, 2019
    Date of Patent: November 30, 2021
    Assignee: Intel Corporation
    Inventors: Subramaniam Maiyuran, Mathew Nevin, Jorge Parra, Ashutosh Garg, Shubra Marwaha, Shubh Shah
  • Patent number: 11182127
    Abstract: Techniques facilitating binary floating-point multiply and scale operation for compute-intensive numerical applications and apparatuses are provided. An embodiment relates to a system that can comprise a memory that stores computer executable components and a processor that executes the computer executable components stored in the memory. The computer executable components can comprise a receiver component that receives an instruction to perform a multiply and scale operation of the first floating point operand value, the second floating point operand value, and the integer operand value, wherein the multiplication component obtains the floating-point product in response to the instruction to perform the multiply and scale operation. The multiplication can be performed as a single instruction.
    Type: Grant
    Filed: March 25, 2019
    Date of Patent: November 23, 2021
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Silvia Melitta Mueller, Bruce Fleischer, Ankur Agrawal, Kailash Gopalakrishnan
  • Patent number: 11182457
    Abstract: Matrix factorization based gradient compression may be applied to an allreduce operation to improve efficiency including the elimination of unnecessary meta data while maintaining accuracy in training of deep learning (DL) of Artificial Intelligence. This compression may include generating a predetermined matrix and a degree of data compression k as a dimension of the predetermined matrix for a plurality of computing nodes. Each computing node may receive a corresponding matrix of matrices to be allreduced, and each corresponding matrix may be decomposed into a plurality of non-fixed matrices and the predetermined matrix. The plurality of non-fixed matrices may be summed to provide an optimized matrix, which may be multiplied by the predetermined matrix to provide a result matrix. The optimized matrix may be designated as a predetermined matrix. These operations may be repeated until all of the matrices received by the computing nodes have been allreduced.
    Type: Grant
    Filed: March 28, 2019
    Date of Patent: November 23, 2021
    Assignee: International Business Machines Corporation
    Inventors: Minsik Cho, Vinod Muthusamy
  • Patent number: 11182126
    Abstract: Computationally efficient mixed precision floating point waveform generation takes advantage of the high-speed generation of waveforms with single-precision floating point numbers while reducing the generally unacceptable loss of precision of pure single-precision floating point to generate any waveform that repeats in 2?. This approaches computes a reference phase in double precision as the modulus of the phase with 2? and then computes offsets to that value in single precision. The double precision reference phase is recomputed as needed depending on how quickly the phase grows and how large a machine epsilon is desired.
    Type: Grant
    Filed: June 25, 2019
    Date of Patent: November 23, 2021
    Assignee: Raytheon Company
    Inventors: Ender Barillas, Brian Filarsky
  • Patent number: 11182128
    Abstract: A multiply-accumulate calculation device, a multiply-accumulate calculation method, and a system for efficiently performing a multiply-accumulate calculation are provided. The multiply-accumulate operation device includes a plurality of memory blocks that store a plurality of multiplied elements and performs a multiply-accumulate operation on input data. Each of memory blocks includes stores one bit value of the same bit digit of a plurality of multiplied elements. An input data generation unit generates input data by extracting data of a same bit digit from the plurality of multiplication elements. A control unit that accumulates and adds value of the multiply-accumulate operation result, and a data memory that stores the accumulated addition value as a multiplication element of the next multiply-accumulate operation.
    Type: Grant
    Filed: August 20, 2019
    Date of Patent: November 23, 2021
    Assignee: RENESAS ELECTRONICS CORPORATION
    Inventor: Shunsuke Okumura
  • Patent number: 11175891
    Abstract: Disclosed embodiments relate to performing floating-point addition with selected rounding. In one example, a processor includes circuitry to decode and execute an instruction specifying locations of first and second floating-point (FP) sources, and an opcode indicating the processor is to: bring the FP sources into alignment by shifting a mantissa of the smaller source FP operand to the right by a difference between their exponents, generating rounding controls based on any bits that escape; simultaneously generate a sum of the FP sources and of the FP sources plus one, the sums having a fuzzy-Jbit format having an additional Jbit into which a carry-out, if any, select one of the sums based on the rounding controls, and generate a result comprising a mantissa-wide number of most-significant bits of the selected sum, starting with the most significant non-zero Jbit.
    Type: Grant
    Filed: March 30, 2019
    Date of Patent: November 16, 2021
    Assignee: Intel Corporation
    Inventors: Simon Rubanovich, Amit Gradstein, Zeev Sperber, Mrinmay Dutta