Patents Assigned to Advanced Micro Devices
-
Publication number: 20140181457Abstract: A system, method, and memory device embodying some aspects of the present invention for remapping external memory addresses and internal memory locations in stacked memory are provided. The stacked memory includes one or more memory layers configured to store data. The stacked memory also includes a logic layer connected to the memory layer. The logic layer has an Input/Output (I/O) port configured to receive read and write commands from external devices, a memory map configured to maintain an association between external memory addresses and internal memory locations, and a controller coupled to the I/O port, memory map, and memory layers, configured to store data received from external devices to internal memory locations.Type: ApplicationFiled: December 21, 2012Publication date: June 26, 2014Applicant: Advanced Micro Devices, Inc.Inventors: Lisa R. HSU, Gabriel H. LOH, Michael IGNATOWSKI, Michael J. SCHULTE, Nuwan S. JAYASENA, James M. O'CONNOR
-
Publication number: 20140181483Abstract: Some die-stacked memories will contain a logic layer in addition to one or more layers of DRAM (or other memory technology). This logic layer may be a discrete logic die or logic on a silicon interposer associated with a stack of memory dies. Additional circuitry/functionality is placed on the logic layer to implement functionality to perform various computation operations. This functionality would be desired where performing the operations locally near the memory devices would allow increased performance and/or power efficiency by avoiding transmission of data across the interface to the host processor.Type: ApplicationFiled: December 21, 2012Publication date: June 26, 2014Applicant: Advanced Micro Devices, Inc.Inventors: James M. O'CONNOR, Nuwan S. Jayasena, Gabriel H. Loh, Michael Ignatowski, Michael J. Schulte
-
Publication number: 20140181427Abstract: Some die-stacked memories will contain a logic layer in addition to one or more layers of DRAM (or other memory technology). This logic layer may be a discrete logic die or logic on a silicon interposer associated with a stack of memory dies. Additional circuitry/functionality is placed on the logic layer to implement functionality to perform various data movement and address calculation operations. This functionality would allow compound memory operations—a single request communicated to the memory that characterizes the accesses and movement of many data items. This eliminates the performance and power overheads associated with communicating address and control information on a fine-grain, per-data-item basis from a host processor (or other device) to the memory. This approach also provides better visibility of macro-level memory access patterns to the memory system and may enable additional optimizations in scheduling memory accesses.Type: ApplicationFiled: December 21, 2012Publication date: June 26, 2014Applicant: Advanced Micro Devices, Inc.Inventors: Nuwan S. JAYASENA, James M. O'Connor, Gabriel H. Loh, Michael J. Schulte, Bradford M. Beckmann, Michael Ignatowski
-
Publication number: 20140181453Abstract: A system, method, and computer program product are provided for a memory device system. One or more memory dies and at least one logic die are disposed in a package and communicatively coupled. The logic die comprises a processing device configurable to manage virtual memory and operate in an operating mode. The operating mode is selected from a set of operating modes comprising a slave operating mode and a host operating mode.Type: ApplicationFiled: December 20, 2012Publication date: June 26, 2014Applicant: Advanced Micro Devices, Inc.Inventors: Nuwan S. JAYASENA, Gabriel H. LOH, Bradford M. BECKMANN, James M. O'CONNOR, Lisa R. HSU
-
Publication number: 20140176187Abstract: A die-stacked memory device incorporates a reconfigurable logic device to provide implementation flexibility in performing various data manipulation operations and other memory operations that use data stored in the die-stacked memory device or that result in data that is to be stored in the die-stacked memory device. One or more configuration files representing corresponding logic configurations for the reconfigurable logic device can be stored in a configuration store at the die-stacked memory device, and a configuration controller can program a reconfigurable logic fabric of the reconfigurable logic device using a selected one of the configuration files. Due to the integration of the logic dies and the memory dies, the reconfigurable logic device can perform various data manipulation operations with higher bandwidth and lower latency and power consumption compared to devices external to the die-stacked memory device.Type: ApplicationFiled: December 23, 2012Publication date: June 26, 2014Applicant: Advanced Micro Devices, Inc.Inventors: Nuwan S. Jayasena, Michael J. Schulte, Gabriel H. Loh, Michael Ignatowski
-
Publication number: 20140181822Abstract: A system, method and a computer-readable medium for task scheduling using fragmented channels is provided. A plurality of fragmented channels are stored in memory accessible to a plurality of compute units. Each fragmented channel is associated with a particular compute unit. Each fragmented channel also stores a plurality of data items from tasks scheduled for processing on the associated compute unit and links to another fragmented channel in the plurality of fragmented channels.Type: ApplicationFiled: December 20, 2012Publication date: June 26, 2014Applicant: Advanced Micro Devices, Inc.Inventors: Bradford M. BECKMANN, Marc S. Orr
-
Publication number: 20140181421Abstract: A system includes an atomic processing engine (APE) coupled to an interconnect. The interconnect is to couple to one or more processor cores. The APE receives a plurality of commands from the one or more processor cores through the interconnect. In response to a first command, the APE performs a first plurality of operations associated with the first command. The first plurality of operations references multiple memory locations, at least one of which is shared between two or more threads executed by the one or more processor cores.Type: ApplicationFiled: December 21, 2012Publication date: June 26, 2014Applicant: Advanced Micro Devices, Inc.Inventors: James M. O'CONNOR, Michael J. Schulte, Nuwan S. Jayasena, Gabriel H. Loh
-
Publication number: 20140177349Abstract: Embodiments are described for a method for regulating sleep mode of a plurality of sub-banks in an SRAM array by isolating all of the sub-banks from a regulator upon access to at least one sub-bank that causes the accessed sub-bank to go to an operating voltage, and isolating the accessed sub-bank from non accessed sub-banks, while maintaining a sleep voltage on a load memory array and each of the sub-banks through the regulator; comparing a voltage on the non accessed sub-banks to a voltage output of the regulator; and providing a sleep voltage level to all of the sub-banks through the regulator when the voltage on the non accessed sub-banks is less than the sleep voltage.Type: ApplicationFiled: December 21, 2012Publication date: June 26, 2014Applicant: Advanced Micro Devices Inc.Inventors: Michael DREESEN, Stephen GREENWOOD, Bruce DOYLE
-
Publication number: 20140177362Abstract: A memory and a method of storing data in a memory are provided. The memory comprises a memory block comprising data bits and additional bits. The memory includes logic which, when receiving a first command, writes data into the data bits of the memory block, wherein the data is masked according to a first input. The logic, in response to a second command, writes data into the data bits of the memory block and writes a second input into the additional bits of the memory block.Type: ApplicationFiled: December 20, 2012Publication date: June 26, 2014Applicant: Advanced Micro Devices, Inc.Inventors: James O'CONNOR, Warren Kruger
-
Publication number: 20140181410Abstract: In response to a processor core exiting a low-power state, a cache is set to a minimum size so that fewer than all of the cache's entries are available to store data, thus reducing the cache's power consumption. Over time, the size of the cache can be increased to account for heightened processor activity, thus ensuring that processing efficiency is not significantly impacted by a reduced cache size. In some embodiments, the cache size is increased based on a measured processor performance metric, such as an eviction rate of the cache. In some embodiments, the cache size is increased at regular intervals until a maximum size is reached.Type: ApplicationFiled: December 20, 2012Publication date: June 26, 2014Applicant: Advanced Micro Devices, Inc.Inventors: John Kalamatianos, Edward J. McLellan, Paul Keltcher, Srilatha Manne, Richard E. Klass, James M. O'Connor
-
Publication number: 20140181414Abstract: A system and method for efficiently limiting storage space for data with particular properties in a cache memory. A computing system includes a cache array and a corresponding cache controller. The cache array includes multiple banks, wherein a first bank is powered down. In response a write request to a second bank for data indicated to be stored in the powered down first bank, the cache controller determines a respective bypass condition for the data. If the bypass condition exceeds a threshold, then the cache controller invalidates any copy of the data stored in the second bank. If the bypass condition does not exceed the threshold, then the cache controller stores the data with a clean state in the second bank. The cache controller writes the data in a lower-level memory for both cases.Type: ApplicationFiled: October 16, 2013Publication date: June 26, 2014Applicant: Advanced Micro Devices, Inc.Inventors: Yasuko Eckert, Gabriel H. Loh, Mauricio Breternitz, James M. O'Connor, Srilatha Manne, Nuwan S. Jayasena, Mithuna S. Thottethodi
-
Publication number: 20140181594Abstract: A system and method for optimizing redundant output verification, are provided. A hardware-based store fingerprint buffer receives multiple instances of output from multiple instances of computation. The store fingerprint buffer generates a signature from the content included in the multiple instances of output. When a barrier is reached, the store fingerprint buffer uses the signature to verify the content is error-free.Type: ApplicationFiled: December 21, 2012Publication date: June 26, 2014Applicant: Advanced Micro Devices, Inc.Inventors: Vilas SRIDHARAN, Sudhanva Gurumurthi
-
Publication number: 20140181458Abstract: A die-stacked memory device incorporates a data translation controller at one or more logic dies of the device to provide data translation services for data to be stored at, or retrieved from, the die-stacked memory device. The data translation operations implemented by the data translation controller can include compression/decompression operations, encryption/decryption operations, format translations, wear-leveling translations, data ordering operations, and the like. Due to the tight integration of the logic dies and the memory dies, the data translation controller can perform data translation operations with higher bandwidth and lower latency and power consumption compared to operations performed by devices external to the die-stacked memory device.Type: ApplicationFiled: December 23, 2012Publication date: June 26, 2014Applicant: Advanced Micro Devices, Inc.Inventors: Gabriel H. Loh, Bradford M. Beckmann, James M. O'Connor, Michael Ignatowski, Michael J. Schulte, Lisa R. Hsu, Nuwan S. Jayasena
-
Publication number: 20140177626Abstract: An electronic assembly includes horizontally-stacked die disposed at an interposer, and may also include vertically-stacked die. The stacked die are interconnected via a multi-hop communication network that is partitioned into a link partition and a router partition. The link partition is at least partially implemented in the metal layers of the interposer for horizontally-stacked die. The link partition may also be implemented in part by the intra-die interconnects in a single die and by the inter-die interconnects connecting vertically-stacked sets of die. The router partition is implemented at some or all of the die disposed at the interposer and comprises the logic that supports the functions that route packets among the components of the processing system via the interconnects of the link partition. The router partition may implement fixed routing, or alternatively may be configurable using programmable routing tables or configurable logic blocks.Type: ApplicationFiled: December 23, 2012Publication date: June 26, 2014Applicant: Advanced Micro Devices, Inc.Inventors: Mithuna S. Thottethodi, Gabriel H. Loh
-
Publication number: 20140181384Abstract: A system, method and computer program product to store tag blocks in a tag buffer in order to provide early row-buffer miss detection, early page closing, and reductions in tag block transfers. A system comprises a tag buffer, a request buffer, and a memory controller. The request buffer stores a memory request having an associated tag. The memory controller compares the associated tag to a plurality of tags stored in the tag buffer and issues the memory request stored in the request buffer to either a memory cache or a main memory based on the comparison.Type: ApplicationFiled: December 21, 2012Publication date: June 26, 2014Applicant: Advanced Micro Devices, Inc.Inventors: Gabriel LOH, Jaewoong Sim
-
Patent number: 8760450Abstract: A graphics-processing unit is used to perform mesh simplification. A vertex shader receives a dataset for an input mesh that portrays a three-dimensional graphics object. The vertex shader generates from the dataset vertices for primitives that make up the input mesh. The input mesh is divided into a grid of cells. A geometry shader receives the vertices from the vertex shader and generates from the received vertices a simplified mesh that portrays the three-dimensional graphics object in less detail than the input mesh. Before the input mesh is divided into grid cells, a warping function can be applied to the input mesh based on a weighting function to warp the input mesh, thereby increasing sampling at a region of interest. A projective warping can be performed on the grid to produce grid cells of different volumes in accordance with a camera position.Type: GrantFiled: October 29, 2008Date of Patent: June 24, 2014Assignee: Advanced Micro Devices, Inc.Inventors: Christopher DeCoro, Natalya Tatarchuk
-
Patent number: 8760946Abstract: Various method and apparatus embodiments for training a delay for enabling a data strobe signal in a memory subsystem are disclosed. In one embodiment, a system includes a memory controller configured to receive a data strobe signal. The memory controller includes a training circuit. The training circuit includes a first storage circuit coupled to receive the data strobe signal on a data input and an enable signal on a clock input, and a training unit configured to, based on an output signal received from the first flip-flop, adjust a phase of the enable signal until an assertion of the enable signal coincides with a preamble indication in the data strobe signal.Type: GrantFiled: May 22, 2012Date of Patent: June 24, 2014Assignee: Advanced Micro DevicesInventors: Glenn A Dearth, Warren R Anderson, Anwar P Kashem, Richard W Reeves, Edoardo Prete, Gerald R Talbot
-
Patent number: 8759962Abstract: Various methods and apparatus for establishing thermal pathways for a semiconductor device are disclosed. In one aspect, a method of manufacturing is provided that includes providing a first semiconductor chip that has a substrate and a first active circuitry portion extending a first distance into the substrate. A barrier is formed in the first semiconductor chip that surrounds but is laterally separated from the first active circuitry portion and extends into the substrate a second distance greater than the first distance.Type: GrantFiled: October 27, 2012Date of Patent: June 24, 2014Assignee: Advanced Micro Devices, Inc.Inventor: Michael Z. Su
-
Patent number: 8760452Abstract: A method and apparatus for processing data in a system comprising a central processing unit (CPU), a system memory, and a graphics processing unit (GPU) includes determining whether the GPU is an integrated graphics processor (IGP). Based upon a determination that the GPU is an IGP, data stored in the system memory is accessed by the GPU without copying the data to a memory on the GPU. Processing is performed on the data in the GPU.Type: GrantFiled: July 1, 2010Date of Patent: June 24, 2014Assignee: Advanced Micro Devices, Inc.Inventors: Gary C. Tiggs, Earl M. Stahl
-
Patent number: 8760196Abstract: A circuit for detecting out-of-band signals is disclosed. In one embodiment, the circuit includes a first differential circuit configured to level shift and positively rectify a differential input signal to produce a first output component of a differential output signal. The detector further includes a second differential circuit configured to level shift and negatively rectify the differential input signal to produce a second output component of the differential output signal. A third differential circuit is configured to level shift and output first and second fixed voltages based on an input reference voltage and a ground voltage. The circuit is configured to provide the differential output signal and the first and second fixed voltages to an indicator circuit configured to assert an indication responsive to detecting that a differential voltage of the differential output signal is greater than a differential voltage of the first and second fixed voltages.Type: GrantFiled: December 12, 2011Date of Patent: June 24, 2014Assignee: Advanced Micro DevicesInventors: Xin Liu, Arvind Bomdica