Patents Assigned to Advanced Micros Devices, Inc.
-
Patent number: 10074600Abstract: Various resistor circuits and methods of making and using the same are disclosed. In one aspect, a method of manufacturing is provided that includes forming a resistor onboard an interposer. The resistor is adapted to dampen a capacitive network. The capacitive network has at least one capacitor positioned external to the interposer.Type: GrantFiled: March 30, 2012Date of Patent: September 11, 2018Assignees: ATI Technologies ULC, Advanced Micro Devices, Inc.Inventors: Fei Guo, Feng Zhu, Julius Din, Anwar Kashem, Sally Yeung
-
Patent number: 10067710Abstract: A processing apparatus is provided that includes a plurality of memory regions each corresponding to a memory address and configured to store data associated with the corresponding memory address. The processing apparatus also includes an accelerated processing device in communication with the memory regions and configured to determine a request to allocate an initial memory buffer comprising a number of contiguous memory regions, create a new memory buffer comprising one or more additional memory regions adjacent to the contiguous memory regions of the initial memory buffer, assign one or more values to the one or more additional memory regions and detect a change to the one or more values at the one or more additional memory regions.Type: GrantFiled: November 23, 2016Date of Patent: September 4, 2018Assignee: ADVANCED MICRO DEVICES, INC.Inventors: Joseph L. Greathouse, Christopher D. Erb, Michael G. Collins
-
Patent number: 10068794Abstract: A system and method for fabricating non-planar devices while managing semiconductor processing yield and cost are described. A semiconductor device fabrication process forms a stack of alternating semiconductor layers. A trench is etched and filled with at least an oxide layer with a length at least that of a device channel length while being bounded by sites for a source region and a drain region. The process places a second silicon substrate on top of both the oxide layer in the trench and the top-most semiconducting layer of the stack. The two surfaces making contact by wafer bonding use the same type of semiconducting layer. The device is flipped such that the first substrate and the stack are on top of the second substrate. The process forms nanowires of a gate region from the stack in the top first substrate.Type: GrantFiled: January 31, 2017Date of Patent: September 4, 2018Assignee: Advanced Micro Devices, Inc.Inventor: Richard T. Schultz
-
Patent number: 10067718Abstract: Dynamic random access memory (DRAM) chips in memory modules include multi-purpose registers (MPRs) having pre-defined data patterns which, when selected, are accessed with read commands and output on data lines for performing read training. The MPRs are accessed by issuing read commands to specific register addresses to request reads from specific MPR locations. In some embodiments, read training for memory modules includes addressing, for a first half of a memory module, a read command to a first register address and performing read training using a first set of bit values received in response to addressing using the first register address. For a second half of the memory module, the same read command is used, but read training is performed using a second set of bit values received in response to addressing using the first register address.Type: GrantFiled: September 23, 2016Date of Patent: September 4, 2018Assignee: Advanced Micro Devices, Inc.Inventors: Glennis Eliagh Covington, Kevin M. Brandl, Nienchi Hu, Shannon T. Kesner
-
Patent number: 10067709Abstract: Systems, apparatuses, and methods for accelerating page migration using a two-level bloom filter are disclosed. In one embodiment, a system includes a GPU and a CPU and a multi-level memory hierarchy. When a memory request misses in a first memory, the GPU is configured to check a first level of a two-level bloom filter to determine if a page targeted by the memory request is located in a second memory. If the first level of the two-level bloom filter indicates that the page is not in the second memory, then the GPU generates a page fault and sends the memory request to a third memory. If the first level of the two-level bloom filter indicates that the page is in the second memory, then the GPU sends the memory request to the CPU.Type: GrantFiled: September 19, 2016Date of Patent: September 4, 2018Assignee: Advanced Micro Devices, Inc.Inventors: Leonardo Piga, Mauricio Breternitz
-
Patent number: 10067555Abstract: An apparatus and a method for controlling power consumption associated with a computing device having first and second processors configured to perform different types of operations includes providing a user interface that allows, during normal operation of the computing device, at least one of: (i) a user selection of desired performance levels of the first and second processors relative to one another, such that higher desired performance levels of one processor correspond to lower desired performance levels of the other processor, and (ii) a user selection of a desired performance level of the first processor and a user selection of a desired performance level of the second processor, the two user selections being made independently of one another. The apparatus and method control, during normal operation of the computing device, performance levels of the processors in response to the one or more user selections of the desired performance levels.Type: GrantFiled: February 20, 2014Date of Patent: September 4, 2018Assignee: Advanced Micro Devices, Inc.Inventor: I-Ming Lin
-
Patent number: 10067911Abstract: Systems, apparatuses, and methods for performing in-place matrix transpose operations are disclosed. Operations for transposing tiles of a matrix are scheduled in an order determined by moving diagonally through tiles of the matrix. When a diagonal line hits a boundary, then a tile on a new diagonal line of the matrix is selected and operations are scheduled for transposing this tile. Only tiles within a triangular region of the matrix are scheduled for being transposed. This allows memory access operations to be performed in parallel, expediting the matrix transpose operation compared to linear tile indexing.Type: GrantFiled: July 26, 2016Date of Patent: September 4, 2018Assignee: Advanced Micro Devices, Inc.Inventors: Amir Gholaminejad, Bragadeesh Natarajan
-
Patent number: 10068372Abstract: A method, a system, and a computer-readable storage medium directed to performing high-speed parallel tessellation of 3D surface patches are disclosed. The method includes generating a plurality of primitives in parallel. Each primitive in the plurality is generated by a sequence of functional blocks, in which each sequence acts independently of all the other sequences.Type: GrantFiled: December 30, 2015Date of Patent: September 4, 2018Assignee: Advanced Micro Devices, Inc.Inventors: Timour T. Paltashev, Boris Prokopenko, Vladimir V. Kibardin
-
Patent number: 10067872Abstract: A plurality of memory modules, which may be used to form a heterogeneous memory system, are connected to a plurality of prefetchers. Each prefetcher is independently configured to prefetch information from a corresponding one of the plurality of memory modules in response to feedback from the corresponding one of the plurality of memory modules.Type: GrantFiled: June 22, 2015Date of Patent: September 4, 2018Assignee: Advanced Micro Devices, Inc.Inventor: David A. Roberts
-
Publication number: 20180246657Abstract: Techniques for handling data compression in which metadata that indicates which portions of data are compressed are which portions of data are not compressed are disclosed. Segments of a buffer referred to as block groups store compressed blocks of data along with uncompressed blocks of data and hash blocks. If a block group includes a block that is a hash of another block in the block group, then the other block is considered to be compressed. If the block group does not include a block that is a hash of another block in the block group, then the blocks in the block group are uncompressed. The hash function to generate the hash is selected to prevent “collisions,” which occur when the data being stored in the buffer is such that it is possible for a hash block and an uncompressed block to be the same.Type: ApplicationFiled: February 24, 2017Publication date: August 30, 2018Applicant: Advanced Micro Devices, Inc.Inventor: Greg Sadowski
-
Publication number: 20180246655Abstract: Improvements in compute shader programs executed on parallel processing hardware are disclosed. An application or other entity defines a sequence of shader programs to execute. Each shader program defines inputs and outputs which would, if unmodified, execute as loads and stores to a general purpose memory, incurring high latency. A compiler combines the shader programs into groups that can operate in a lower-latency, but lower-capacity local data store memory. The boundaries of these combined shader programs are defined by several aspects including where memory barrier operations are to execute, whether combinations of shader programs can execute using only the local data store and not the global memory (except for initial reads and writes) and other aspects.Type: ApplicationFiled: February 24, 2017Publication date: August 30, 2018Applicant: Advanced Micro Devices, Inc.Inventors: Michael L. Schmit, Radhakrishna Giduthuri
-
Publication number: 20180246815Abstract: Techniques are provided for managing address translation request traffic where memory access requests can be made with differing quality-of-service levels, which specify latency and/or bandwidth requirements. The techniques involve translation lookaside buffers. Within the translation lookaside buffers, certain resources are reserved for specific quality-of-service levels. More specifically, translation lookaside buffer slots, which store the actual translations, as well as finite state machines in a work queue, are reserved for specific quality-of-service levels. The translation lookaside buffer receives multiple requests for address translation. The translation lookaside buffer selects requests having the highest quality-of-service level for which an available finite state machine is available.Type: ApplicationFiled: February 24, 2017Publication date: August 30, 2018Applicants: Advanced Micro Devices, Inc., ATI Technologies ULCInventors: Wade K. Smith, Kostantinos Danny Christidis
-
Publication number: 20180246816Abstract: Techniques are provided for using a translation lookaside buffer to provide low latency memory address translations for data streams. Clients of a memory system first prepare the address translation cache hierarchy by requesting that a translation pre-fetch stream is initialized. After the translation pre-fetch stream is initialized, the cache hierarchy returns an acknowledgment of completion to the client, which then begins to access memory. Pre-fetch streams are specified in terms of address ranges and are performed for large contiguous portions of the virtual memory address space.Type: ApplicationFiled: February 24, 2017Publication date: August 30, 2018Applicants: Advanced Micro Devices, Inc., ATI Technologies ULCInventors: Wade K. Smith, Kostantinos Danny Christidis
-
Patent number: 10062143Abstract: A method and apparatus for real time compressing randomly accessed data includes extracting a block of randomly accessed data from a memory hierarchy. One or more individual portions of the randomly accessed data are independently compressed in real time to create a lossless compressed image surface. The compressed image surface includes data of independently compressed image blocks for reading and decompressing in a random order. The method further includes storing structured information relating to the dynamically compressed randomly accessed data.Type: GrantFiled: September 12, 2016Date of Patent: August 28, 2018Assignee: Advanced Micro Devices, Inc.Inventors: Chris Brennan, Timour T. Paltashev
-
Patent number: 10060955Abstract: A processing system includes one or more power supply monitors (PSMs) to measure one or more first voltages corresponding to one or more locations in the processing system. The measurements are performed concurrently with the processing system executing one or more code loops. The processing system also includes calibration logic to modify a second voltage provided to the processing system based on a comparison of a reference voltage and the one or more first voltages. The reference voltage is determined based on previous execution of the one or more code loops by the processing system.Type: GrantFiled: June 25, 2014Date of Patent: August 28, 2018Assignee: Advanced Micro Devices, Inc.Inventors: Aaron J. Grenat, Robert A. Hershberger, Sriram Sambamurthy, Samuel D. Naffziger, Christopher E. Tressler, Sho-Chien Kang, Joseph P. Shannon, Krishna Sai Bernucho, Ashwin Chincholi, Michael J. Austin, Steven F. Liepe, Umair B. Cheema
-
Patent number: 10062206Abstract: A parallel adaptable graphics rasterization system in which a primitive assembler includes a router to selectively route a primitive to a first rasterizer or one of a plurality of second rasterizers. The second rasterizers concurrently operate on different primitives and the primitive is selectively routed based on an area of the primitive. In some variations, a bounding box of the primitive is reduced to a predetermined number of pixels prior to providing the primitive to the one of the plurality of second rasterizers. Reducing the bounding box can include subtracting an origin of the bounding box from coordinates of points that represent the primitive.Type: GrantFiled: August 30, 2016Date of Patent: August 28, 2018Assignee: Advanced Micro Devices, Inc.Inventors: Boris Prokopenko, Timour T. Paltashev, Vladimir V. Kibardin
-
Publication number: 20180239722Abstract: A method, computer program product, and system are provided for associating one or more memory buffers in a computing system with a plurality of memory channels. The method can include associating a first memory buffer to a first plurality of memory banks, where the first plurality of memory banks spans over a first set of one or more memory channels. Similarly, the method can include associating a second memory buffer to a second plurality of memory banks, where the second plurality of memory banks spans over a second set of one or more memory channels. The method can also include associating a first sequence identifier and a second sequence identifier with the first memory buffer and the second memory buffer, respectively. Further, the method can include accessing the first and second memory buffers based on the first and second sequence identifiers.Type: ApplicationFiled: April 20, 2018Publication date: August 23, 2018Applicant: Advanced Micro Devices, Inc.Inventors: Greg Sadowski, Philip J. Rogers
-
Patent number: 10055370Abstract: A method of and device for removing a processor from a low power mode. The method includes and the device provides for performing multiple processor start-up tasks in parallel. Memory interface training between the processor and memory and restoration and initialization of the processor are performed in parallel with each other and with a serial bus controller entering serial bus training to facilitate communication between the processor and a system controller.Type: GrantFiled: July 9, 2014Date of Patent: August 21, 2018Assignee: Advanced Micro Devices, Inc.Inventor: Greg Sadowski
-
Patent number: 10055359Abstract: The described embodiments include a computer system having a multi-level memory hierarchy with two or more levels of memory, each level being one of two or more types of memory. The computer system handles storing objects in the multi-level memory hierarchy. During operation, a system runtime in the computer system identifies an object to be stored in the multi-level memory hierarchy. The system runtime then determines, based on one or more attributes of the object, that the object is to be pinned in a level of the multi-level memory hierarchy. The system runtime then pins the object in the level of the multi-level memory hierarchy. In the described embodiments, the pinning includes hard pinning and soft pinning, which are each associated with corresponding retention policies for pinned objects.Type: GrantFiled: February 10, 2016Date of Patent: August 21, 2018Assignee: ADVANCED MICRO DEVICES, INC.Inventors: Sergey Blagodurov, Gabriel H. Loh, John R. Slice
-
Patent number: 10048741Abstract: Systems, apparatuses, and methods for implementing performance estimation mechanisms are disclosed. In one embodiment, a computing system includes at least one processor and a memory subsystem. During a characterization phase, the system utilizes a memory intensive workload to detect when the memory subsystem reaches its saturation point. Then, the system collects performance counter values during a sampling phase of a target application to determine the memory bandwidth. If the memory bandwidth is greater than the saturation point, then the system generates a prediction of the memory time which is based on a ratio of the memory bandwidth over the saturation point. Otherwise, if the memory bandwidth is less than the saturation point, the system assumes memory time is constant versus processor frequency. Then, the system uses the memory time and an estimate of the compute time to estimate a phase time for the target application at different processor frequencies.Type: GrantFiled: January 26, 2017Date of Patent: August 14, 2018Assignee: Advanced Micro Devices, Inc.Inventors: Md Abdullah Shahneous Bari, Leonardo Piga, Indrani Paul