Patents Examined by Eric Coleman
  • Patent number: 11816062
    Abstract: The invention relates to a control unit for controlling a data transfer between a classical processor and a quantum processor with a plurality of qubits. The control unit comprises a plurality of control and read-out circuits configured for controlling and reading out the plurality of qubits. Each of the control and read-out circuits is assigned to one or more of the qubits. A controlling of the quantum processor by the control unit comprises selectively powering on a subset of the control and read-out circuits during an instruction cycle, while ensuring that the remaining control and read-out circuits are powered off during the instruction cycle. The powered-on subset of control and read-out circuits is used to control a subset of the qubits and to read out data from the subset of qubits.
    Type: Grant
    Filed: November 4, 2021
    Date of Patent: November 14, 2023
    Assignee: International Business Machines Corporation
    Inventors: Thomas Morf, Cezar Bogdan Zota, Peter Mueller, Pier Andrea Francese, Marcel A. Kossel, Matthias Braendli, Mridula Prathapan
  • Patent number: 11816480
    Abstract: A computing unit is disclosed, comprising a first memory bank for storing input activations and a second memory bank for storing parameters used in performing computations. The computing unit includes at least one cell comprising at least one multiply accumulate (“MAC”) operator that receives parameters from the second memory bank and performs computations. The computing unit further includes a first traversal unit that provides a control signal to the first memory bank to cause an input activation to be provided to a data bus accessible by the MAC operator. The computing unit performs one or more computations associated with at least one element of a data array, the one or more computations being performed by the MAC operator and comprising, in part, a multiply operation of the input activation received from the data bus and a parameter received from the second memory bank.
    Type: Grant
    Filed: August 22, 2022
    Date of Patent: November 14, 2023
    Assignee: Google LLC
    Inventors: Olivier Temam, Ravi Narayanaswami, Harshit Khaitan, Dong Hyuk Woo
  • Patent number: 11816560
    Abstract: The technology disclosed relates to allocating available physical compute units (PCUs) and/or physical memory units (PMUs) of a reconfigurable data processor to operation units of an operation unit graph for execution thereof. In particular, it relates to selecting, for evaluation, an intermediate stage compute processing time between lower and upper search bounds of a generic stage compute processing time, determining a pipeline number of the PCUs and/or the PMUs required to process the operation unit graph, and iteratively, initializing new lower and upper search bounds of the generic stage compute processing time and selecting, for evaluation in a next iteration, a new intermediate stage compute processing time taking into account whether the pipeline number of the PCUs and/or the PMUs produced for a prior intermediate stage compute processing time in a previous iteration is lower or higher than the available PCUs and/or PMUs.
    Type: Grant
    Filed: August 8, 2022
    Date of Patent: November 14, 2023
    Assignee: SambaNova Systems, Inc.
    Inventors: Zhuo Chen, Sumti Jairath
  • Patent number: 11803509
    Abstract: A merge sort circuit can include a parallel merge sort core that performs a partial merge on two input tuples, each containing a number P of data elements sorted according to a sort key, to produce a sorted output tuple of P data elements. Input data blocks to be merged can be stored in first and second block buffers. The block buffers can receive data from a vector memory read interface that reads groups of at least P data elements at a time. Loading of data elements into the block buffers can be based on respective fill levels of the block buffers.
    Type: Grant
    Filed: May 23, 2022
    Date of Patent: October 31, 2023
    Assignee: Apple Inc.
    Inventors: Xiaoning Nie, Mathias Kohlenz, Jin-Soo Yoo
  • Patent number: 11790267
    Abstract: An architecture and associated techniques of an apparatus for hardware accelerated machine learning are disclosed. The architecture features multiple memory banks storing tensor data. The tensor data may be concurrently fetched by a number of execution units working in parallel. Each operational unit supports an instruction set specific to certain primitive operations for machine learning. An instruction decoder is employed to decode a machine learning instruction and reveal one or more of the primitive operations to be performed by the execution units, as well as the memory addresses of the operands of the primitive operations as stored in the memory banks. The primitive operations, upon performed or executed by the execution units, may generate some output that can be saved into the memory banks. The fetching of the operands and the saving of the output may involve permutation and duplication of the data elements involved.
    Type: Grant
    Filed: October 14, 2020
    Date of Patent: October 17, 2023
    Assignee: Intel Corporation
    Inventors: Jeremy Bruestle, Choong Ng
  • Patent number: 11789897
    Abstract: A data processing circuit, a data processing method, and an electronic device are provided. The data processing circuit includes a first data processing sub-circuit and a second data processing sub-circuit. An output terminal of the first data processing sub-circuit is connected to an input terminal of the second data processing sub-circuit. The first data processing sub-circuit is configured to receive an original sequence to generate a first processed sequence. Each of first processed numbers in the first processed sequence is calculated from at least two pieces of original data in the original data. The second data processing sub-circuit is configured to receive the first processed sequence to generate a second processed sequence.
    Type: Grant
    Filed: October 27, 2020
    Date of Patent: October 17, 2023
    Assignees: BOE TECHNOLOGY DEVELOPMENT CO., LTD., BOE TECHNOLOGY GROUP CO., LTD.
    Inventors: Xiangye Wei, Liming Xiu
  • Patent number: 11782724
    Abstract: The present disclosure provides a parallel decision system and method for distributed data processing. The system includes: an initial logical node generation assembly, a logical node traversal assembly, a predetermined configuration cost computation assembly, and a parallel decision assembly. The initial logical node generation assembly is configured to receive task configuration data input by a user to generate an initial logical node topology for the distributed data processing system. The logical node traversal assembly is configured to traverse the initial logical node topology to obtain a predetermined configuration in the initial logical node topology. The predetermined configuration cost computation assembly is configured to compute a transmission cost of each predetermined configuration and a cost sum.
    Type: Grant
    Filed: March 22, 2023
    Date of Patent: October 10, 2023
    Assignee: BEIJING ONEFLOW TECHNOLOGY CO., LTD
    Inventors: Yipeng Li, Juncheng Liu, Xinqi Li, Cheng Cheng, Jinhui Yuan
  • Patent number: 11782679
    Abstract: A circuit for use in a processor includes a first processing channel having a first logic unit, a second processing channel having a second logic unit, and multiplexing circuitry. The multiplexing circuitry includes an input multiplexer arranged to switch between a first state in which an input of the first logic unit is coupled to an input line of the first processing channel, and a respective second state in which the input of the first logic unit is instead coupled to an input line of the second processing channel; and an output multiplexer arranged to switch between a first state in which an output line of the second processing channel is coupled to an output of the second logic unit, and a second state in which the output line of the second processing channel is instead coupled to an output of the first logic unit.
    Type: Grant
    Filed: June 20, 2022
    Date of Patent: October 10, 2023
    Assignee: Imagination Technologies Limited
    Inventor: Kenneth C. Rovers
  • Patent number: 11782872
    Abstract: Logical processor distribution across physical processors is provided. A set of logical processors of a number of logical processors defined for a particular logical partition of a plurality of active logical partitions is assigned to a physical processor chip having a greatest logical processor entitlement for the particular logical partition until no more logical processors can be assigned to that physical processor chip based on a logical processor entitlement of that physical processor chip being exhausted. Remaining logical processors of the number of logical processors defined for the particular logical partition are assigned to other physical processor chips of a plurality of physical processor chips assigned to the particular logical partition until all of the remaining logical processors have been assigned to a physical processor chip.
    Type: Grant
    Filed: March 7, 2022
    Date of Patent: October 10, 2023
    Assignee: International Business Machines Corporation
    Inventors: Jeffrey G. Chan, Seth E. Lederer, Jerry A. Moody, Hunter J. Kauffman
  • Patent number: 11782871
    Abstract: In one implementation a vector processor unit having preload registers for at least some of vector length, vector constant, vector address, and vector stride. Each preload register has an input and an output. All the preload register inputs are coupled to receive a new vector parameters. Each of the preload registers' outputs are coupled to a first input of a respective multiplexor, and the second input of all the respective multiplexors are coupled to the new vector parameters.
    Type: Grant
    Filed: March 22, 2022
    Date of Patent: October 10, 2023
    Assignee: Microchip Technology Inc.
    Inventor: Christopher I. W. Norrie
  • Patent number: 11775312
    Abstract: A processing element includes a PIM cluster configured to read data from and write data to an adjacent DRAM subarray, wherein the PIM cluster has a plurality of processing cores, each processing core of the plurality of processing cores containing a look-up table, and a router connected to each processing core, wherein the router is configured to communicate data among each processing core; and a controller unit configured to communicate with the router, wherein the controller unit contains an executable program of operational decomposition algorithms. The look-up tables can be programmable. A DRAM chip including a plurality of DRAM banks, each DRAM bank having a plurality of interleaved DRAM subarrays and a plurality of the PIM clusters configured to read data from and write data to an adjacent DRAM subarray is disclosed.
    Type: Grant
    Filed: April 11, 2022
    Date of Patent: October 3, 2023
    Assignee: Rochester Institute of Technology
    Inventors: Amlan Ganguly, Sai Manoj Pudukotai Dinakarrao, Mark Connolly, Purab Ranjan Sutradhar, Sathwika Bavikadi, Mark Allen Indovina
  • Patent number: 11775295
    Abstract: A processing-in-memory (PIM) device includes a data storage region and an arithmetic circuit. The data storage region is configured to store first data and second data. The arithmetic circuit includes a zero-detection circuit and multiplier. The zero-detection circuit including first transfer gates, second transfer gates, and an output control logic circuit. The first controller and the second controller are configured to receive a first output value and a second output value generated by inverting the first output value. The first output value is having a value of “1” when all bits of the first data or the second data have a value of “0”. The output control logic circuit is configured to generate zero data including bits having a value of ‘0’ as output data of the multiplier when the second output value is a value of ‘0’.
    Type: Grant
    Filed: July 14, 2022
    Date of Patent: October 3, 2023
    Assignee: SK hynix Inc.
    Inventors: Mun Gyu Son, Choung Ki Song
  • Patent number: 11768683
    Abstract: A processor includes a front end, an execution unit, a retirement stage, a counter, and a performance monitoring unit. The front end includes logic to receive an event instruction to enable supervision of a front end event that will delay execution of instructions. The execution unit includes logic to set a register with parameters for supervision of the front end event. The front end further includes logic to receive a candidate instruction and match the candidate instruction to the front end event. The counter includes logic to generate the front end event upon retirement of the candidate instruction.
    Type: Grant
    Filed: February 18, 2022
    Date of Patent: September 26, 2023
    Assignee: Intel Corporation
    Inventor: Ahmad Yasin
  • Patent number: 11762803
    Abstract: Systems and methods are provided to enable parallelized multiply-accumulate operations in a systolic array. Each column of the systolic array can include multiple busses enabling independent transmission of input partial sums along the respective bus. Each processing element of a given columnar bus can receive an input partial sum from a prior element of the given columnar bus, and perform arithmetic operations on the input partial sum. Each processing element can generate an output partial sum based on the arithmetic operations, provide the output partial sum to a next processing element of the given columnar bus, without the output partial sum being processed by a processing element of the column located between the two processing elements that uses a different columnar bus. Use of columnar busses can enable parallelization to increase speed or enable increased latency at individual processing elements.
    Type: Grant
    Filed: April 18, 2022
    Date of Patent: September 19, 2023
    Assignee: Amazon Technologies, Inc.
    Inventors: Thomas A Volpe, Sundeep Amirineni, Thomas Elmer
  • Patent number: 11741044
    Abstract: The present disclosure relates to a mechanism for issuing instructions in a processor (e.g., a vector processor) implemented as an overlay on programmable hardware (e.g., a field programmable gate array (FPGA) device). Implementations described herein include features for optimizing resource availability on programmable hardware units and enabling superscalar execution when coupled with a temporal single-instruction multiple data (SIMD). Systems described herein involve an issue component of a processor controller (e.g., a vector processor controller) that enables fast and efficient instruction issue while verifying that structural and data hazards between instructions have been resolved.
    Type: Grant
    Filed: December 30, 2021
    Date of Patent: August 29, 2023
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Aaron Michael Landy, Skand Hurkat
  • Patent number: 11740904
    Abstract: A vector friendly instruction format and execution thereof. According to one embodiment of the invention, a processor is configured to execute an instruction set. The instruction set includes a vector friendly instruction format. The vector friendly instruction format has a plurality of fields including a base operation field, a modifier field, an augmentation operation field, and a data element width field, wherein the first instruction format supports different versions of base operations and different augmentation operations through placement of different values in the base operation field, the modifier field, the alpha field, the beta field, and the data element width field, and wherein only one of the different values may be placed in each of the base operation field, the modifier field, the alpha field, the beta field, and the data element width field on each occurrence of an instruction in the first instruction format in instruction streams.
    Type: Grant
    Filed: November 11, 2021
    Date of Patent: August 29, 2023
    Assignee: Intel Corporation
    Inventors: Robert C. Valentine, Jesus Corbal San Adrian, Roger Espasa Sans, Robert D. Cavin, Bret L. Toll, Santiago Galan Duran, Jeffrey G. Wiedemeier, Sridhar Samudrala, Milind Baburao Girkar, Edward Thomas Grochowski, Jonathan Cannon Hall, Dennis R. Bradford, Elmoustapha Ould-Ahmed-Vall, James C Abel, Mark Charney, Seth Abraham, Suleyman Sair, Andrew Thomas Forsyth, Lisa Wu, Charles Yount
  • Patent number: 11741043
    Abstract: This invention provides a generalized electronic computer architecture with multiple cores, memory distributed amongst the cores (a core-local memory). This arrangement provides predictable, low-latency memory response time, as well as a flexible, code-supplied flow of memory from one specific operation to another (using an operation graph). In one instantiation, the operation graph consists of a set of math operations, each accompanied by an ordered list of one or more input addresses. Input addresses may be specific addresses in memory, references to other math operations in the graph, or references to the next item in a particular data stream, where data streams are iterators through a continuous block of memory. The arrangement can also be packaged as a PCIe daughter card, which can be selectively plugged into a host server/PC constructed/organized according to traditional von Neumann architecture.
    Type: Grant
    Filed: January 28, 2022
    Date of Patent: August 29, 2023
    Assignee: The Trustees of Dartmouth College
    Inventors: Elijah F. W. Bowen, Richard H. Granger, Jr.
  • Patent number: 11734003
    Abstract: The present disclosure relates to a compiler for causing a computer to execute a process. The process includes generating a first program, wherein the first program includes a first code that determines whether a first area of a memory that a process inside a loop included in a second program refers to in a first execution time of the loop is in duplicate with a second area of the memory that the process refers to in a second execution time of the loop, a second code that executes the process in an order of the first and second execution times when it is determined that the first and the second areas are duplicate, and a third code that executes the process for the first execution time and the process for the second execution time in parallel when it is determined that the first and the second areas are not duplicate.
    Type: Grant
    Filed: December 16, 2021
    Date of Patent: August 22, 2023
    Assignee: FUJITSU LIMITED
    Inventor: Yuta Mukai
  • Patent number: 11726791
    Abstract: Examples of the present disclosure provide apparatuses and methods related to generating and executing a control flow. An example apparatus can include a first device configured to generate control flow instructions, and a second device including an array of memory cells, an execution unit to execute the control flow instructions, and a controller configured to control an execution of the control flow instructions on data stored in the array.
    Type: Grant
    Filed: May 12, 2022
    Date of Patent: August 15, 2023
    Assignee: Micron Technology, Inc.
    Inventors: Kyle B. Wheeler, Richard C. Murphy, Troy A. Manning, Dean A. Klein
  • Patent number: 11726951
    Abstract: A packet transmission apparatus includes a processor such as a central processing unit (CPU), a first processing chip, and a second processing chip. The second processing chip is separately connected to the processor and the first processing chip. The second processing chip is disposed between the processor and the first processing chip. The first processing chip is a non-programmable chip such as an application-specific integrated circuit (ASIC) chip, and the second processing chip is a programmable chip such as a field-programmable gate array (FPGA) chip. The second processing chip supports a second function, and the second function is updatable. Both the processor and the first processing chip are configured to exchange a packet with the second processing chip. The second processing chip is configured to process a received packet based on the second function and send the processed packet to the processor or the first processing chip.
    Type: Grant
    Filed: December 20, 2021
    Date of Patent: August 15, 2023
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Zhongtian Guo, Tao Li