Operation Patents (Class 712/42)
  • Patent number: 11977885
    Abstract: An apparatus to facilitate utilizing structured sparsity in systolic arrays is disclosed. The apparatus includes a processor comprising a systolic array to receive data from a plurality of source registers, the data comprising unpacked source data, structured source data that is packed based on sparsity, and metadata corresponding to the structured source data; identify portions of the unpacked source data to multiply with the structured source data, the portions of the unpacked source data identified based on the metadata; and output, to a destination register, a result of multiplication of the portions of the unpacked source data and the structured source data.
    Type: Grant
    Filed: November 30, 2020
    Date of Patent: May 7, 2024
    Assignee: INTEL CORPORATION
    Inventors: Subramaniam Maiyuran, Jorge Parra, Ashutosh Garg, Chandra Gurram, Chunhui Mei, Durgesh Borkar, Shubra Marwaha, Supratim Pal, Varghese George, Wei Xiong, Yan Li, Yongsheng Liu, Dipankar Das, Sasikanth Avancha, Dharma Teja Vooturi, Naveen K. Mellempudi
  • Patent number: 11960884
    Abstract: An embodiment of the invention is a processor including execution circuitry to calculate, in response to a decoded instruction, a result of a complex multiplication of a first complex number and a second complex number. The calculation includes a first operation to calculate a first term of a real component of the result and a first term of the imaginary component of the result. The calculation also includes a second operation to calculate a second term of the real component of the result and a second term of the imaginary component of the result. The processor also includes a decoder, a first source register, and a second source register. The decoder is to decode an instruction to generate the decoded instruction. The first source register is to provide the first complex number and the second source register is to provide the second complex number.
    Type: Grant
    Filed: November 2, 2021
    Date of Patent: April 16, 2024
    Assignee: Intel Corporation
    Inventors: Robert Valentine, Mark Charney, Raanan Sade, Elmoustapha Ould-Ahmed-Vall, Jesus Corbal, Roman S. Dubtsov
  • Patent number: 11900107
    Abstract: Disclosed embodiments relate to instructions for fused multiply-add (FMA) operations with variable-precision inputs. In one example, a processor to execute an asymmetric FMA instruction includes fetch circuitry to fetch an FMA instruction having fields to specify an opcode, a destination, and first and second source vectors having first and second widths, respectively, decode circuitry to decode the fetched FMA instruction, and a single instruction multiple data (SIMD) execution circuit to process as many elements of the second source vector as fit into an SIMD lane width by multiplying each element by a corresponding element of the first source vector, and accumulating a resulting product with previous contents of the destination, wherein the SIMD lane width is one of 16 bits, 32 bits, and 64 bits, the first width is one of 4 bits and 8 bits, and the second width is one of 1 bit, 2 bits, and 4 bits.
    Type: Grant
    Filed: March 25, 2022
    Date of Patent: February 13, 2024
    Assignee: Intel Corporation
    Inventors: Dipankar Das, Naveen K. Mellempudi, Mrinmay Dutta, Arun Kumar, Dheevatsa Mudigere, Abhisek Kundu
  • Patent number: 11893398
    Abstract: Methods, apparatuses, and systems for implementing data flows in a processor are described herein. A data flow manager may be configured to generate a configuration packet for a compute operation based on status information regarding multiple processing elements of the processor. Accordingly, multiple processing elements of a processor may concurrently process data flows based on the configuration packet. For example, the multiple processing elements may implement a mapping of processing elements to memory, while also implementing identified paths, through the processor, for the data flows. After executing the compute operation at certain processing elements of the processor, the processing results may be provided. In speech signal processing operations, the processing results may be compared to phonemes to identify such components of human speech in the processing results.
    Type: Grant
    Filed: February 16, 2022
    Date of Patent: February 6, 2024
    Assignee: MICRON TECHNOLOGY, INC.
    Inventors: Jeremy Chritz, Tamara Schmitz, Fa-Long Luo, David Hulton
  • Patent number: 11853658
    Abstract: The present disclosure provides an information processing apparatus capable of improving the accuracy of solving a problem when an annealing algorithm is executed. An information processing apparatus 1 includes a calculation unit 2 and an arithmetic unit 4. The calculation unit 2 calculates at least one eigenvector of a coupling coefficient matrix including a coupling coefficient indicating a strength of interaction between each of a plurality of binary variables that indicate states of a plurality of respective spins in an Ising model, the coupling coefficient matrix being given in advance in accordance with a problem to be solved by an annealing algorithm. The arithmetic unit 4 executes the annealing algorithm from an initial state set based on the at least one eigenvector.
    Type: Grant
    Filed: December 17, 2020
    Date of Patent: December 26, 2023
    Assignee: NEC CORPORATION
    Inventor: Yoshihiro Nambu
  • Patent number: 11782601
    Abstract: Embodiments of the present invention are directed to an instruction set of an object memory fabric. This object memory fabric instruction set can be used to define arbitrary, parallel functionality such as: direct object address manipulation and generation without the overhead of complex address translation and software layers to manage differing address space; direct object authentication with no runtime overhead that can be set based on secure 3rd party authentication software; object related memory computing in which, as objects move between nodes, the computing can move with them; and parallelism that is dynamically and transparent based on scale and activity. These instructions are divided into three conceptual classes: memory reference including load, store, and special memory fabric instructions; control flow including fork, join, and branches; and execute including arithmetic and comparison instructions.
    Type: Grant
    Filed: January 20, 2016
    Date of Patent: October 10, 2023
    Assignee: Ultrata, LLC
    Inventors: Steven J. Frank, Larry Reback
  • Patent number: 11775302
    Abstract: A digital data processor includes an instruction memory storing instructions each specifying a data processing operation and at least one data operand field, an instruction decoder coupled to the instruction memory for sequentially recalling instructions from the instruction memory and determining the data processing operation and the at least one data operand, and at least one operational unit coupled to a data register file and to an instruction decoder to perform a data processing operation upon at least one operand corresponding to an instruction decoded by the instruction decoder and storing results of the data processing operation. The operational unit is configured to increment histogram values in response to a histogram instruction by incrementing a bin entry at a specified location in a specified number of at least one histogram.
    Type: Grant
    Filed: October 25, 2021
    Date of Patent: October 3, 2023
    Assignee: Texas Instruments Incorporated
    Inventors: Naveen Bhoria, Duc Bui, Rama Venkatasubramanian, Dheera Balasubramanian Samudrala, Alan Davis
  • Patent number: 11755202
    Abstract: Embodiments of the invention provide systems and methods to implement a hardware-based multi-node processing system of an object memory fabric. Hardware-based processing nodes operatively coupled may each include object memory modules storing and managing memory objects, each memory object being created natively within the memory module and managed by the memory module at a memory layer, and each memory object including memory object data and memory object metadata. The memory object metadata may include triggers that specify additional operations to be executed by any object memory module of the hardware-based processing nodes when the respective memory object is located at the respective object memory module and accessed as part of the respective object memory module processing requests.
    Type: Grant
    Filed: January 20, 2016
    Date of Patent: September 12, 2023
    Assignee: Ultrata, LLC
    Inventors: Steven J. Frank, Larry Reback
  • Patent number: 11755201
    Abstract: Embodiments of the invention provide systems and methods to implement an object memory fabric including hardware-based processing nodes having memory modules storing and managing memory objects created natively within the memory modules and managed by the memory modules at a memory layer, where physical address of memory and storage is managed with the memory objects based on an object address space that is allocated on a per-object basis with an object addressing scheme. Each node may utilize the object addressing scheme to couple to additional nodes to operate as a set of nodes so that all memory objects of the set are accessible based on the object addressing scheme, which defines invariant object addresses for the memory objects that are invariant with respect to physical memory storage locations and storage location changes of the memory objects within the memory module and across all modules interfacing the object memory fabric.
    Type: Grant
    Filed: January 20, 2016
    Date of Patent: September 12, 2023
    Assignee: Ultrata, LLC
    Inventors: Steven J. Frank, Larry Reback
  • Patent number: 11755474
    Abstract: Tile based interleaving and de-interleaving of row-column interleaved data is described. In one example, the de-interleaving is divided into two memory transfer stages, the first from an on-chip memory to a DRAM and the second from the DRAM to an on-chip memory. Each stage operates on part of a row-column interleaved block of data and re-orders the data items, such that the output of the second stage comprises de-interleaved data. In the first stage, data items are read from the on-chip memory according to a non-linear sequence of memory read addresses and written to the DRAM. In the second stage, data items are read from the DRAM according to bursts of linear address sequences which make efficient use of the DRAM interface and written back to on-chip memory according to a non-linear sequence of memory write addresses.
    Type: Grant
    Filed: November 18, 2021
    Date of Patent: September 12, 2023
    Assignee: Imagination Technologies Limited
    Inventors: Paul Murrin, Adrian J. Anderson, Mohammed El-Hajjar
  • Patent number: 11650792
    Abstract: A tile of an FPGA includes a multiple mode arithmetic circuit. The multiple mode arithmetic circuit is configured by control signals to operate in an integer mode, a floating-point mode, or both. In some example embodiments, multiple integer modes (e.g., unsigned, two's complement, and sign-magnitude) are selectable, multiple floating-point modes (e.g., 16-bit mantissa and 8-bit sign, 8-bit mantissa and 6-bit sign, and 6-bit mantissa and 6-bit sign) are supported, or any suitable combination thereof. The tile may also fuse a memory circuit with the arithmetic circuits. Connections directly between multiple instances of the tile are also available, allowing multiple tiles to be treated as larger memories or arithmetic circuits. By using these connections, referred to as cascade inputs and outputs, the input and output bandwidth of the arithmetic circuit is further increased.
    Type: Grant
    Filed: January 6, 2022
    Date of Patent: May 16, 2023
    Assignee: Achronix Semiconductor Corporation
    Inventors: Daniel Pugh, Raymond Nijssen, Michael Philip Fitton, Marcel Van der Goot
  • Patent number: 11645206
    Abstract: A method for using a distributed memory device in a memory augmented neural network system includes receiving, by a controller, an input query to access data stored in the distributed memory device, the distributed memory device comprising a plurality of memory banks. The method further includes determining, by the controller, a memory bank selector that identifies a memory bank from the distributed memory device for memory access, wherein the memory bank selector is determined based on a type of workload associated with the input query. The method further includes computing, by the controller and by using content based access, a memory address in the identified memory bank. The method further includes generating, by the controller, an output in response to the input query by accessing the memory address.
    Type: Grant
    Filed: September 13, 2021
    Date of Patent: May 9, 2023
    Assignee: International Business Machines Corporation
    Inventors: Ahmet Serkan Ozcan, Tomasz Kornuta, Carl Radens, Nicolas Antoine
  • Patent number: 11605015
    Abstract: A hybrid quantum-classical (HQC) computer prepares a quantum Boltzmann machine (QBM) in a pure state. The state is evolved in time according to a chaotic, tunable quantum Hamiltonian. The pure state locally approximates a (potentially highly correlated) quantum thermal state at a known temperature. With the chaotic quantum Hamiltonian, a quantum quench can be performed to locally sample observables in quantum thermal states. With the samples, an inverse temperature of the QBM can be approximated, as needed for determining the correct sign and magnitude of the gradient of a loss function of the QBM.
    Type: Grant
    Filed: October 24, 2019
    Date of Patent: March 14, 2023
    Assignee: Zapata Computing, Inc.
    Inventors: Eric R. Anschuetz, Yudong Cao
  • Patent number: 11586575
    Abstract: There is disclosed an example of an artificial intelligence (AI) system, including: a first hardware platform; a fabric interface configured to communicatively couple the first hardware platform to a second hardware platform; a processor hosted on the first hardware platform and programmed to operate on an AI problem; and a first training accelerator, including: an accelerator hardware; a platform inter-chip link (ICL) configured to communicatively couple the first training accelerator to a second training accelerator on the first hardware platform without aid of the processor; a fabric ICL to communicatively couple the first training accelerator to a third training accelerator on a second hardware platform without aid of the processor; and a system decoder configured to operate the fabric ICL and platform ICL to share data of the accelerator hardware between the first training accelerator and second and third training accelerators without aid of the processor.
    Type: Grant
    Filed: January 25, 2022
    Date of Patent: February 21, 2023
    Assignee: Intel Corporation
    Inventors: Francesc Guim Bernat, Da-Ming Chiang, Kshitij A. Doshi, Suraj Prabhakaran, Mark A. Schmisseur
  • Patent number: 11579580
    Abstract: The present invention provides a technology for invoking a non-periodic-execution function module from a periodic-execution control program. A control system that comprises first and second control parts (C1, C2) and a storage device that stores control programs (210, 211) for a controller. The control programs (210, 211) include a periodic-execution function module (55B) that invokes a non-periodic-execution function module (55A). The first control part (C1) reflects the value of an input variable for the periodic-execution function module (55B) in an argument for the non-periodic-execution function module (55A) and outputs an execution start command for the function modules to the second control part (C2). The second control part (C2) outputs a return value for the non-periodic-execution function module (55A) to the first control part (C1). The first control part (C1) reflects the return value in an output variable for the periodic-execution function module (55B).
    Type: Grant
    Filed: October 2, 2019
    Date of Patent: February 14, 2023
    Assignee: OMRON Corporation
    Inventor: Takuya Hirota
  • Patent number: 11579884
    Abstract: Techniques for performing instruction fetch operations are provided. The techniques include determining instruction addresses for a primary branch prediction path; requesting that a level 0 translation lookaside buffer (“TLB”) caches address translations for the primary branch prediction path; determining either or both of alternate control flow path instruction addresses and lookahead control flow path instruction addresses; and requesting that either the level 0 TLB or an alternative level TLB caches address translations for either or both of the alternate control flow path instruction addresses and the lookahead control flow path instruction addresses.
    Type: Grant
    Filed: June 26, 2020
    Date of Patent: February 14, 2023
    Assignee: Advanced Micro Devices, Inc.
    Inventors: Ashok Tirupathy Venkatachar, Steven R. Havlir, Robert B. Cohen
  • Patent number: 11579774
    Abstract: Embodiments of the invention provide systems and methods for managing processing, memory, storage, network, and cloud computing to significantly improve the efficiency and performance of processing nodes. More specifically, embodiments of the present invention are directed to an instruction set of an object memory fabric. This object memory fabric instruction set can include trigger instructions defined in metadata for a particular memory object. Each trigger instruction can comprise a single instruction and action based on reference to a specific object to initiate or perform defined actions such as pre-fetching other objects or executing a trigger program.
    Type: Grant
    Filed: January 20, 2016
    Date of Patent: February 14, 2023
    Assignee: Ultrata, LLC
    Inventors: Steven J. Frank, Larry Reback
  • Patent number: 11568431
    Abstract: Methods and apparatus to compensate impression data for misattribution and non-coverage are disclosed.
    Type: Grant
    Filed: October 12, 2020
    Date of Patent: January 31, 2023
    Assignee: THE NIELSEN COMPANY (US), LLC
    Inventors: Kumar Nagaraja Rao, Tianjue Luo, Albert Ronald Perez, Stephen S. Bell, Mimi Zhang, David Wong, Jennifer Haskell
  • Patent number: 11487921
    Abstract: A method and system perform quantum-assisted finite elements-based, design optimization of an object to minimize a shape-specific quantity by manipulating the shape of the object using a processing unit, for example, a Quantum Processing Unit (QPU). As a result, a shape-specific quantity, such as an approximation of sound pressure at a specific position around an object, can be minimized by manipulating the object shape using the QPU.
    Type: Grant
    Filed: November 1, 2019
    Date of Patent: November 1, 2022
    Inventors: Florian Neukart, Dyon Van Vreumingen, David Von Dollen, Arne-Christian Voigt, Michael Hartmann, Carsten Othmer
  • Patent number: 11455169
    Abstract: A digital data processor includes an instruction memory storing instructions specifying data processing operations and a data operand field, an instruction decoder coupled to the instruction memory for recalling instructions from the instruction memory and determining the operation and the data operand, and an operational unit coupled to a data register file and an instruction decoder to perform an operation upon an operand corresponding to an instruction decoded by the instruction decoder and storing results of the operation. The operational unit is configured to perform a table recall in response to a look up table read instruction by recalling data elements from a specified location and adjacent location to the specified location, in a specified number of at least one table and storing the recalled data elements in successive slots in a destination register. Recalled data elements include at least one interpolated data element in the adjacent location.
    Type: Grant
    Filed: September 13, 2019
    Date of Patent: September 27, 2022
    Assignee: Texas Instruments Incorporated
    Inventors: Naveen Bhoria, Dheera Balasubramanian Samudrala, Duc Bui, Alan Davis
  • Patent number: 11321052
    Abstract: The present invention provides an environment that makes it possible to incorporate a non-periodic-execution function module into a periodic-execution control program. A development support program that makes a computer execute steps that generate a periodic-execution second function module that invokes a non-periodic-execution first function module. Said steps include: a step in which a data type that is in a periodic-execution programming language and corresponds to the data type of an argument for the first function module is specified, and an input variable of said data type is established for the second function module; and a step in which a data type that is in a periodic-execution programming language and corresponds to the data type of a return value for the first function module is specified, and an output variable of said data type is established for the second function module.
    Type: Grant
    Filed: October 2, 2019
    Date of Patent: May 3, 2022
    Assignee: OMRON Corporation
    Inventor: Takuya Hirota
  • Patent number: 11269801
    Abstract: There is disclosed an example of an artificial intelligence (AI) system, including: a first hardware platform; a fabric interface configured to communicatively couple the first hardware platform to a second hardware platform; a processor hosted on the first hardware platform and programmed to operate on an AI problem; and a first training accelerator, including: an accelerator hardware; a platform inter-chip link (ICL) configured to communicatively couple the first training accelerator to a second training accelerator on the first hardware platform without aid of the processor; a fabric ICL to communicatively couple the first training accelerator to a third training accelerator on a second hardware platform without aid of the processor; and a system decoder configured to operate the fabric ICL and platform ICL to share data of the accelerator hardware between the first training accelerator and second and third training accelerators without aid of the processor.
    Type: Grant
    Filed: December 17, 2020
    Date of Patent: March 8, 2022
    Assignee: Intel Corporation
    Inventors: Francesc Guim Bernat, Da-Ming Chiang, Kshitij A. Doshi, Suraj Prabhakaran, Mark A. Schmisseur
  • Patent number: 11263162
    Abstract: There is disclosed an example of an artificial intelligence (AI) system, including: a first hardware platform; a fabric interface configured to communicatively couple the first hardware platform to a second hardware platform; a processor hosted on the first hardware platform and programmed to operate on an AI problem; and a first training accelerator, including: an accelerator hardware; a platform inter-chip link (ICL) configured to communicatively couple the first training accelerator to a second training accelerator on the first hardware platform without aid of the processor; a fabric ICL to communicatively couple the first training accelerator to a third training accelerator on a second hardware platform without aid of the processor; and a system decoder configured to operate the fabric ICL and platform ICL to share data of the accelerator hardware between the first training accelerator and second and third training accelerators without aid of the processor.
    Type: Grant
    Filed: December 20, 2017
    Date of Patent: March 1, 2022
    Assignee: Intel Corporation
    Inventors: Francesc Guim Bernat, Da-Ming Chiang, Kshitij A. Doshi, Suraj Prabhakaran, Mark A. Schmisseur
  • Patent number: 11250107
    Abstract: The present disclosure relates to a method for executing a computation task composed of at least one set of operations where subsets of pipelineable operations of the set of operations are determined in accordance with a pipelining scheme. A single routine may be created for enabling execution of the determined subsets of operations by a hardware accelerator. The routine has, as arguments, a value indicative of input data and values of configuration parameters of the computation task, where a call of the routine causes a scheduling of the subsets of operations on the hardware accelerator in accordance with the values of the configuration parameters. Upon receiving input data of the computation task, the routine may be called to cause the hardware accelerator to perform by the computation task in accordance with the scheduling.
    Type: Grant
    Filed: July 15, 2019
    Date of Patent: February 15, 2022
    Assignee: International Business Machines Corporation
    Inventors: Christophe Piveteau, Nikolas Ioannou, Igor Krawczuk, Manuel Le Gallo-Bourdeau, Abu Sebastian, Evangelos Stavros Eleftheriou
  • Patent number: 11062227
    Abstract: A hybrid computer generates samples for machine learning. The hybrid computer includes a processor that implements a Boltzmann machine, e.g., a quantum Boltzmann machine, which returns equilibrium samples from eigenstates of a quantum Hamiltonian. Subsets of samples are provided to training and validations modules. Operation can include: receiving a training set; preparing a model described by an Ising Hamiltonian; initializing model parameters; segmenting the training set into subsets; creating a sample set by repeatedly drawing samples until the determined number of samples has been drawn; and updating the model. Operation can include partitioning the training set into input and output data sets, and determining a conditional probability distribution that describes a probability of observing an output vector given a selected input vector, e.g.
    Type: Grant
    Filed: October 14, 2016
    Date of Patent: July 13, 2021
    Assignee: D-WAVE SYSTEMS INC.
    Inventors: Mohammad H. S. Amin, Evgeny Andriyash, Jason Rolfe
  • Patent number: 11023233
    Abstract: A processor of an aspect includes a decode unit to decode a user-level suspend thread instruction that is to indicate a first alternate state. The processor also includes an execution unit coupled with the decode unit. The execution unit is to perform the instruction at a user privilege level. The execution unit in response to the instruction, is to: (a) suspend execution of a user-level thread, from which the instruction is to have been received; (b) transition a logical processor, on which the user-level thread was to have been running, to the indicated first alternate state; and (c) resume the execution of the user-level thread, when the logical processor is in the indicated first alternate state, with a latency that is to be less than half a latency that execution of a thread can be resumed when the logical processor is in a halt processor power state.
    Type: Grant
    Filed: February 9, 2016
    Date of Patent: June 1, 2021
    Assignee: INTEL CORPORATION
    Inventors: Michael Mishaeli, Jason W. Brandt, Gilbert Neiger, Asit K. Mallick, Rajesh M. Sankaran, Raghunandan Makaram, Benjamin C. Chaffin, James B. Crossland, H. Peter Anvin
  • Patent number: 10924127
    Abstract: In some aspects, a control system interacts with a quantum system. In some instances, the quantum system includes qubits that respond to a control signal generated by the control system, and the control system is configured to generate the control signal in response to an input signal. A control sequence (which may include, for example, a sequence of values for the input signal) can be generated by a computing system based on a target operation to be applied to the qubits. The control sequence can be generated based on the target operation, a quantum system model, a distortion model and possibly other information. The quantum system model represents the quantum system and includes a control parameter representing the control signal. The distortion model represents a nonlinear relationship between the control signal and the input signal. The control sequence is applied to the quantum system by operation of the control system.
    Type: Grant
    Filed: March 5, 2020
    Date of Patent: February 16, 2021
    Assignee: Quantum Valley Investment Fund LP
    Inventors: Ian N. Hincks, Chris E. Granade, Troy W. Borneman, David G. Cory
  • Patent number: 10896241
    Abstract: An information processing apparatus manufactured at low cost and with ease and that is capable of making a search for a ground state of an arbitrary Ising model. An information processing unit containing a plurality of semiconductor chips, each retains a value of one spin or values of a plurality of spins and simulates interactions among the spins, inter-chip wiring between the necessary semiconductor chips, and a control unit that cause each semiconductor chip to perform interaction computation. The control unit converts data of a problem into data of a lattice-shaped Ising model, which is possibly expressed by the plurality of semiconductor chips, without causing a spin arrangement, in a ground state of an Ising model for the problem, to be changed. The data of the lattice-shaped Ising model is divided for allocation to the plurality of semiconductor chips, and causes each semiconductor chip to perform the interaction computation.
    Type: Grant
    Filed: June 9, 2015
    Date of Patent: January 19, 2021
    Assignee: HITACHI, LTD.
    Inventors: Takuya Okuyama, Masanao Yamaoka, Chihiro Yoshimura, Masato Hayashi, Akihito Akai
  • Patent number: 10748078
    Abstract: Techniques relate to operating a quantum processing device is provided. The device includes at least two fixed-frequency quantum circuits coupled to a frequency-tunable coupler. The frequency of the coupler can be modulated so as to drive at least two selectively addressable energy transitions in the quantum processing device. The method includes modulating the frequency of the coupler so as to drive two first-order energy transitions. This is done so as to transfer (at least partly) an excitation of one of the quantum circuits to at least another one of the quantum circuits, via the tunable coupler. Related quantum processing devices are also provided.
    Type: Grant
    Filed: May 23, 2019
    Date of Patent: August 18, 2020
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Stefan Filipp, Nikolaj Moll, Daniel J. Egger, Marc Ganzhorn, Andreas Fuhrer, Gian R. Von Salis
  • Patent number: 10347323
    Abstract: A semiconductor memory device includes a memory core that performs reading and writing of data, data delivery and training blocks that are connected between first pads and the memory core, and at least one data delivery, clock generation and training block that is connected between at least one second pad and the memory core. In a first training operation, the data delivery and training blocks output first training data, received through the first pads, through the first pads as second training data. In a second training operation, at least one of the data delivery and training blocks outputs third training data, received through the at least one second pad, through at least one of the first pads as fourth training data. The second training data and the fourth training data are output in synchronization with read data strobe signals output through the at least one second pad.
    Type: Grant
    Filed: May 21, 2018
    Date of Patent: July 9, 2019
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Young-Hun Kim, Sihong Kim
  • Patent number: 10242141
    Abstract: A computer-implemented method includes receiving a first circuit design for an integrated circuit device, determining when multiple power-drawing events are to occur at substantially the same time via one or more circuitry components of the integrated circuit device, which would have a disruptive effect on a power distribution network of the integrated circuit device, based on the first circuit design, and generating logic that schedules the more than one event so that the more than one event do not occur simultaneously. The logic is included in an event sequencer. The method also includes inserting the event sequencer into the first circuit design during compilation to create a second circuit design and outputting the second circuit design to be implemented on the integrated circuit device.
    Type: Grant
    Filed: September 27, 2016
    Date of Patent: March 26, 2019
    Assignee: Altera Corporation
    Inventors: Jakob Raymond Jones, Tim Tri Hoang, Ben Chunben Wang
  • Patent number: 10002586
    Abstract: Display data used in display frame generation are compressed for efficient storage in a local memory within a graphics processing unit. The compression technique used is difference encoding and before performing difference encoding, display data in RGB format are converted into YCbCr format. Since the component values of adjacent pixels in YCbCr format typically vary less than the component values of the same adjacent pixels in RGB format, converting the display data to YCbCr format before performing difference encoding improves the compression efficiency.
    Type: Grant
    Filed: December 13, 2006
    Date of Patent: June 19, 2018
    Assignee: NVIDIA CORPORATION
    Inventors: Sreenivas Krishnan, Koen Bennebroek, Karthik Bhat, Stefano A. Pescador, David G. Reed, Brad W. Simeral, Edward M. Veeser
  • Patent number: 9529600
    Abstract: Provided are an application booting system and method that control a booting time and booting order of each application installed in a terminal (for example, an AVN) according to an importance of each application. The application booting system and method generate or update priority information of each application on the basis of an execution time and the number of executions of each application installed in the terminal, and when power is supplied to the terminal, the application booting system and method control an application to be booted according to the priority information while a system of the terminal is being booted. Accordingly, the application booting system and method enhance a booting speed of a whole system, and provide an execution convenience of an application having a high importance after booting of the system is completed.
    Type: Grant
    Filed: June 2, 2014
    Date of Patent: December 27, 2016
    Assignee: HYUNDAI MOBIS Co., Ltd.
    Inventor: Hyun Tae Kim
  • Patent number: 9501327
    Abstract: Provided are a computer program product, system, and method for concurrently processing parts of cells of a data structure with multiple processes. Information is provided to indicate a partitioning of the cells of the data structure into a plurality of parts, and having a cursor pointing to a cell in the part. Processes concurrently process different parts of the data structure by performing: determining from the cursor for the part one of the cells in the part to process; processing the cells from the cursor to determine whether to process the unit of work corresponding to the cell; and setting the cursor to identify one of the cells from which processing is to continue in a subsequent iteration in response to processing the units of work for a plurality of the processed cells.
    Type: Grant
    Filed: May 13, 2014
    Date of Patent: November 22, 2016
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Theresa M. Brown, Nicolas M. Clayton, Lokesh M. Gupta, Theodore T. Harris, Jr., Brian D. Hatfield, Matthew Sanchez, David B. Schreiber
  • Patent number: 9361109
    Abstract: A system and method to evaluate a data value as an instruction is disclosed. For example, an apparatus configured to execute program code includes an execute unit configured to execute a first instruction associated with a location of a second instruction. The first instruction is identified by a program counter. The apparatus also includes a decode unit configured to receive the second instruction from the location and to decode the second instruction to generate a decoded second instruction without changing the program counter to point to the second instruction.
    Type: Grant
    Filed: May 24, 2010
    Date of Patent: June 7, 2016
    Assignee: QUALCOMM Incorporated
    Inventors: Lucian Codrescu, Erich James Plondke, Suresh Venkumahanti
  • Patent number: 9032186
    Abstract: Augmented processor hardware contains a microcode interpreter. When encrypted microcode is included in a message from a service, the microcode may be passed to the microcode interpreter. Based on decryption and execution of the microcode taking place at the processor hardware, extended functionality may be realized.
    Type: Grant
    Filed: July 8, 2011
    Date of Patent: May 12, 2015
    Assignee: BlackBerry Limited
    Inventors: Ian Robertson, Roger Paul Bowman, Robert Henderson Wood
  • Publication number: 20150113247
    Abstract: A method for securing a data processing system having a processing unit is disclosed. At least a group of N1 digital words of m1 bits is selected from among the set of M1 digital words. N1 is less than M1. These words are selected in such a way that each selected digital word differs from all the other selected digital words by a number of bits at least equal to an integer p which is at least equal to 2. The group of N1 digital words of m1 bits form at least one group of N1 executable digital instructions. The processing unit is configured to make it capable of executing each instruction of the at least one group of N1 executable digital instructions.
    Type: Application
    Filed: October 10, 2014
    Publication date: April 23, 2015
    Inventor: Philippe Escalona
  • Publication number: 20150095614
    Abstract: An apparatus and method are described for the efficient migration of architectural state between processor cores. For example, a processor according to one embodiment comprises: a first processing core having a first instruction execution pipeline including first register set for storing a first architectural state of a first thread being executed thereon; a second processing core having a second instruction execution pipeline including a second register set for storing a second architectural state of a second thread being executed thereon; and architectural state migration logic to perform a direct, simultaneous swap of the first architectural state from the first register set with the second architectural state from the second register set responsive to detecting that the execution of the first thread is to be migrated from the first core to the second core.
    Type: Application
    Filed: September 27, 2013
    Publication date: April 2, 2015
    Inventors: Bret L. Toll, Scott D. Hahn, Jason W. Brandt, Thomas F. Toll
  • Publication number: 20150067301
    Abstract: A microprocessor includes a plurality of registers that holds an architectural state of the microprocessor and an indicator that indicates a boot instruction set architecture (ISA) of the microprocessor as either the x86 ISA or the Advanced RISC Machines (ARM) ISA. The microprocessor also includes a hardware instruction translator that translates x86 ISA instructions and ARM ISA instructions into microinstructions. The hardware instruction translator translates, as instructions of the boot ISA, the initial ISA instructions that the microprocessor fetches from architectural memory space after receiving a reset signal. The microprocessor also includes an execution pipeline, coupled to the hardware instruction translator. The execution pipeline executes the microinstructions to generate results defined by the x86 ISA and ARM ISA instructions.
    Type: Application
    Filed: October 28, 2014
    Publication date: March 5, 2015
    Inventors: G. GLENN HENRY, TERRY PARKS, RODNEY E. HOOKER
  • Publication number: 20150046681
    Abstract: Quantum processor architectures employ unit cells tiled over an area. A unit cell may include first and second sets of qubits where each qubit in the first set crosses at least one qubit in the second set. Angular deviations between qubits in one set may allow qubits in the same set to cross one another. Each unit cell is positioned proximally adjacent at least one other unit cell. Communicatively coupling between qubits is realized through respective intra-cell and inter-cell coupling devices.
    Type: Application
    Filed: August 7, 2014
    Publication date: February 12, 2015
    Inventor: Andrew Douglas King
  • Patent number: 8954941
    Abstract: Method of generating respective instruction compaction schemes for subsets of instructions to be processed by a programmable processor, comprising the steps of a) receiving at least one input code sample representative for software to be executed on the programmable processor, the input code comprising a plurality of instructions defining a first set of instructions (S1), b) initializing a set of removed instructions as empty (S3), c) determining the most compact representation of the first set of instructions (S4) d) comparing the size of said most compact representation with a threshold value (S5), e) carrying out steps e1 to e3 if the size is larger than said threshold value, e1) determining which instruction of the first set of instructions has a highest coding cost (S6), e2) removing said instruction having the highest coding cost from the first set of instructions and (S7), e3) adding said instruction to the set of removed instructions (S8), f) repeating steps b-f, wherein the first set of instructions
    Type: Grant
    Filed: September 3, 2010
    Date of Patent: February 10, 2015
    Assignee: Intel Corporation
    Inventors: Hendrik Tjeerd Joannes Zwartenkot, Alexander Augusteijn, Yuanging Guo, Jürgen Von Oerthel, Jeroen Anton Johan Leijten, Erwan Yann Maurice Le Thenaff
  • Publication number: 20150032993
    Abstract: Achieving orthogonal control of non-orthogonal qubit parameters of a logical qubit allows for increasing the length of a qubit chain thereby increasing the effective connectivity of the qubit chain. A hybrid qubit is formed by communicatively coupling a dedicated second qubit to a first qubit. By tuning a programmable parameter of the second qubit of a hybrid qubit, an effective programmable parameter of the hybrid qubit is adjusted without affecting another effective programmable parameter of the hybrid qubit thereby achieving orthogonal control of otherwise non-orthogonal qubit parameters. The length of the logical qubit may thus be increased by communicatively coupling a plurality of such hybrid qubits together.
    Type: Application
    Filed: July 23, 2014
    Publication date: January 29, 2015
    Inventors: Mohammad H.S. Amin, Trevor Michael Lanting, Colin Enderud
  • Publication number: 20150032994
    Abstract: Techniques for improving the performance of a quantum processor are described. Some techniques employ improving the processor topology through design and fabrication, reducing intrinsic/control errors, reducing thermally-assisted errors and methods of encoding problems in the quantum processor for error correction.
    Type: Application
    Filed: July 24, 2014
    Publication date: January 29, 2015
    Inventors: Fabian Ariel Chudak, Christopher B. Rich, Paul I. Bunyk
  • Publication number: 20140359256
    Abstract: The present invention provides systems and methods for coupled dynamical systems for chaos computing. For example, a system for the coupled dynamical system comprises a first, second, and third circuit. The first circuit comprising a plurality of single dynamical systems forms a coupled dynamical system that reduces local noises in the plurality of single dynamical systems by diffusing the local noises across the coupled dynamical system. The second circuit, connected to the first circuit, receives the data and control inputs and builds an encoding map that translates the data and control inputs to an initial condition on an unstable manifold of the plurality of single dynamical systems in the coupled dynamical system. After the coupled dynamical system evolves, a third circuit, connected to the first circuit, samples a state of one of the plurality of single dynamical systems in the coupled dynamical system and builds a decoding map.
    Type: Application
    Filed: March 21, 2014
    Publication date: December 4, 2014
    Inventors: William Ditto, Behnam Kia, Sarvenaz Kia
  • Publication number: 20140325184
    Abstract: A processor saves micro-architectural contexts to increase the efficiency of code execution and power management. Power management hardware during runtime monitors execution of a code block. The code block has been compiled to have a reserved space appended to one end of the code block. The reserved space includes a metadata block associated with the code block or an identifier of the metadata block. The hardware stores a micro-architectural context of the processor in the metadata block. The micro-architectural context includes performance data resulting from a first execution of the code block. The hardware reads the metadata block upon a second execution of the code block and tunes the second execution based on the performance data.
    Type: Application
    Filed: March 30, 2012
    Publication date: October 30, 2014
    Inventors: Efraim Rotem, Eliezer Weissamann, Boris Ginzburg, Alon Naveh, Nadav Shulman, Ronny Ronen
  • Publication number: 20140325185
    Abstract: A method for operating a processor in which a first program comprising a first sequence of commands is provided, at least one second program is provided comprising a second sequence of commands, where the first program comprises a time-critical section with time-critical commands, commands from the first and second programs are processed in a processor pipeline, a start time is identified for the time-critical section in the first program, and a predefined interrupt program is incorporated into the at least one second program once the start time of the time critical section in the first program has been identified.
    Type: Application
    Filed: January 31, 2012
    Publication date: October 30, 2014
    Applicant: Siemens Aktiengesellschaff
    Inventors: Rene Graf, Wolfgang Hartmann
  • Patent number: 8874837
    Abstract: An integrated circuit can include a programmable circuitry operable according to a first clock frequency and a block random access memory. The block random access memory can include a random access memory (RAM) element having at least one data port and a memory processor coupled to the data port of the RAM element and to the programmable circuitry. The memory processor can be operable according to a second clock frequency that is higher than the first clock frequency. Further, the memory processor can be hardwired and dedicated to perform operations in the RAM element of the block random access memory.
    Type: Grant
    Filed: November 8, 2011
    Date of Patent: October 28, 2014
    Assignee: Xilinx, Inc.
    Inventors: Christopher E. Neely, Gordon J. Brebner
  • Publication number: 20140304491
    Abstract: It is provided a processor system comprising at least one processor core including a processor, a memory and an accelerator. The memory includes an instruction area, a synchronization flag area and a data area. The accelerator starts, even if the processor is executing another processing, acceleration processing and executes read instruction in a case where the read instruction is a flag checking instruction and a flag indicating the completion of predetermined processing has been written; and stores the data subjected to the acceleration processing after completion of the acceleration processing, and further writes a flag indicating the completion of the acceleration processing. The processor starts, even if the accelerator is executing another processing, read instruction corresponding to a flag in a case where the read instruction is the flag checking instruction and it is confirmed that the flag indicating the completion of the acceleration processing has been written.
    Type: Application
    Filed: October 30, 2012
    Publication date: October 9, 2014
    Inventors: Hironori Kasahara, Keiji Kimura
  • Publication number: 20140297993
    Abstract: A microprocessor includes a plurality of processing cores each comprises a corresponding memory physically located inside the core and readable by the core but not readable by the other cores (“core memory”). The microprocessor also includes a memory physically located outside all of the cores and readable by all of the cores (“uncore memory”). For each core, the uncore memory and corresponding core memory collectively provide M words of storage for microcode instructions fetchable by the core as follows: the uncore memory provides J of the M words of microcode instruction storage, and the corresponding core memory provides K of the M words of microcode instruction storage. J, K and M are counting numbers, and M=J+K. The memories are non-architecturally-visible and accessed using a fetch address provided by a non-architectural program counter, and the microcode instructions are non-architectural instructions that implement architectural instructions.
    Type: Application
    Filed: November 5, 2013
    Publication date: October 2, 2014
    Applicant: VIA Technologies, Inc.
    Inventors: G. Glenn Henry, Terry Parks, Rodney E. Hooker, John D. Bunda, Brent Bean
  • Patent number: RE49467
    Abstract: A semiconductor memory device includes a memory core that performs reading and writing of data, data delivery and training blocks that are connected between first pads and the memory core, and at least one data delivery, clock generation and training block that is connected between at least one second pad and the memory core. In a first training operation, the data delivery and training blocks output first training data, received through the first pads, through the first pads as second training data. In a second training operation, at least one of the data delivery and training blocks outputs third training data, received through the at least one second pad, through at least one of the first pads as fourth training data. The second training data and the fourth training data are output in synchronization with read data strobe signals output through the at least one second pad.
    Type: Grant
    Filed: December 22, 2020
    Date of Patent: March 21, 2023
    Inventors: Young-Hun Kim, Sihong Kim