Patents Assigned to Advanced Micro Device, Inc.
-
Patent number: 11860797Abstract: Restricting peripheral device protocols in confidential compute architectures, the method including: receiving a first address translation request from a peripheral device supporting a first protocol, wherein the first protocol supports cache coherency between the peripheral device and a processor cache; determining that a confidential compute architecture is enabled; and providing, in response to the first address translation request, a response including an indication to the peripheral device to not use the first protocol.Type: GrantFiled: December 30, 2021Date of Patent: January 2, 2024Assignees: ADVANCED MICRO DEVICES, INC., ATI TECHNOLOGIES ULCInventors: Philip Ng, Nippon Raval, David A. Kaplan, Donald P. Matthews, Jr.
-
Patent number: 11860784Abstract: A technique for operating a cache is disclosed. The technique includes recording access data for a first set of memory accesses of a first frame; identifying parameters for a second set of memory accesses of a second frame subsequent to the first frame, based on the access data; and applying the parameters to the second set of memory accesses.Type: GrantFiled: June 27, 2022Date of Patent: January 2, 2024Assignee: Advanced Micro Devices, Inc.Inventors: Christopher J. Brennan, Akshay Lahiry
-
Patent number: 11864344Abstract: A computing device chassis for a common cooling solution for die packages comprising: a chassis base comprising: an internal cavity; a cooling element housed in the internal cavity; and one or more thermal interfaces to the cooling element.Type: GrantFiled: September 27, 2021Date of Patent: January 2, 2024Assignee: ADVANCED MICRO DEVICES, INC.Inventor: Christopher M. Jaggers
-
Patent number: 11860787Abstract: Methods, devices, and systems for retrieving information based on cache miss prediction. A prediction that a cache lookup for the information will miss a cache is made based on a history table. The cache lookup for the information is performed based on the request. A main memory fetch for the information is begun before the cache lookup completes, based on the prediction that the cache lookup for the information will miss the cache. In some implementations, the prediction includes comparing a first set of bits stored in the history table with a second set of bits stored in the history table. In some implementations, the prediction includes comparing at least a portion of an address of the request for the information with a set of bits in the history table.Type: GrantFiled: September 30, 2021Date of Patent: January 2, 2024Assignee: Advanced Micro Devices, Inc.Inventors: Ciji Isen, Paul J. Moyer
-
Patent number: 11860755Abstract: An approach is provided for implementing memory profiling aggregation. A hardware aggregator provides memory profiling aggregation by controlling the execution of a plurality of hardware profilers that monitor memory performance in a system. For each hardware profiler of the plurality of hardware profilers, a hardware counter value is compared to a threshold value. When a threshold value is satisfied, execution of a respective hardware profiler of the plurality of hardware profilers is initiated to monitor memory performance. Multiple hardware profilers of the plurality of hardware profilers may execute concurrently and each generate a result counter value. The result counter values generated by each hardware profiler of the plurality of hardware profilers are aggregated to generate an aggregate result counter value. The aggregate result counter value is stored in memory that is accessible by a software processes for use in optimizing memory-management policy decisions.Type: GrantFiled: July 11, 2022Date of Patent: January 2, 2024Assignee: Advanced Micro Devices, Inc.Inventors: Sergey Blagodurov, Jinyoung Choi
-
Patent number: 11862640Abstract: A system and method for efficiently creating layout for memory bit cells are described. In various implementations, cells of a library use Cross field effect transistors (FETs) that include vertically stacked gate all around (GAA) transistors with conducting channels oriented in an orthogonal direction between them. The channels of the vertically stacked transistors use opposite doping polarities. One or more of these cells use a dual polarity local interconnect power connection to receive a voltage reference level from a backside bus. For example, a power supply reference voltage level is received by a p-type device from a backside bus where the connection traverses both a p-type local interconnect layer and an n-type local interconnect layer.Type: GrantFiled: September 29, 2021Date of Patent: January 2, 2024Assignee: Advanced Micro Devices, Inc.Inventor: Richard T. Schultz
-
Patent number: 11861781Abstract: The graphics processing unit (GPU) of a processing system transitions to a low-power state between frame rendering operations according to an inter-frame power off process, where GPU state information is stored on retention hardware. The retention hardware can include retention random access memory (RAM) or retention flip-flops. The retention hardware is operable in an active mode and a retention mode, where read/write operations are enabled at the retention hardware in the active mode and disabled in the retention mode, but data stored on the retention hardware is still retained in the retention mode. The retention hardware is placed in the retention state between frame rendering operations. The GPU transitions from its low-power state to its active state upon receiving an indication that a new frame is ready to be rendered and is restored using the GPU state information stored at the retention hardware.Type: GrantFiled: December 28, 2020Date of Patent: January 2, 2024Assignees: SAMSUNG ELECTRONICS CO., LTD., Advanced Micro Devices, Inc., ATI TECHNOLOGIES ULCInventors: Sreekanth Godey, Ashkan Hosseinzadeh Namin, Seunghun Jin, Teik-Chung Tan
-
Patent number: 11863769Abstract: A system configured to perform scalable video encoding is provided. The system includes a memory; and a processing unit, wherein the processing unit is configured to: receive inter-layer data and a current picture, wherein the current picture has a base layer; upsample the inter-layer data to generate residual data and reconstruction data, wherein the inter-layer data includes a base mode flag; and encode the current picture to an enhanced layer using the upsampled inter-layer data based on a block type of the base layer and the base mode flag.Type: GrantFiled: August 16, 2021Date of Patent: January 2, 2024Assignees: Advanced Micro Devices, Inc., ATI Technologies ULCInventors: Lei Zhang, Ji Zhou, Zhen Chen, Min Yu
-
Patent number: 11860685Abstract: A system and method for efficiently generating clock signals are described. In various implementations, an integrated circuit includes multiple clock frequency dividers both at its I/O boundaries and across its die. A clock frequency divider utilizes a first clock divider and a second clock divider that receive input clock signals with an initial phase difference between them. The first clock divider and the second clock divider generate output clock signals that have frequencies that are a fraction of the frequencies of the received input clock signals. The second clock divider uses a combined multiplexer and flip-flop (combined mux-flop) circuit. The combined mux-flop circuit receives a reset signal that is asserted asynchronously with respect to an input clock signal received by the second clock divider. The second clock divider generates an output clock signal that has the initial phase difference with an output clock signal of the first clock divider.Type: GrantFiled: October 29, 2021Date of Patent: January 2, 2024Assignee: Advanced Micro Devices, Inc.Inventors: Luke Jereme Whitaker, Edoardo Prete
-
Publication number: 20230418772Abstract: A data processor accesses a memory having a first pseudo channel and a second pseudo channel. The data processor includes at least one memory accessing agent for generating a memory access request, a memory controller for providing a memory command to the memory in response to a normalized request selectively using a first pseudo channel pipeline circuit and a second pseudo channel pipeline circuit, and a data fabric for converting the memory access request into the normalized request selectively for the first pseudo channel and the second pseudo channel.Type: ApplicationFiled: June 24, 2022Publication date: December 28, 2023Applicant: Advanced Micro Devices, Inc.Inventors: Hideki Kanayama, YuBin Yao
-
Publication number: 20230418753Abstract: A technique for operating a cache is disclosed. The technique includes based on a workload change, identifying a first allocation permissions policy; operating the cache according to the first allocation permissions policy; based on set sampling, identifying a second allocation permissions policy; and operating the cache according to the second allocation permissions policy.Type: ApplicationFiled: June 28, 2022Publication date: December 28, 2023Applicant: Advanced Micro Devices, Inc.Inventors: Chintan S. Patel, Alexander J. Branover, Benjamin Tsien, Edgar Munoz, Vydhyanathan Kalyanasundharam
-
Publication number: 20230420036Abstract: A fine-grained dynamic random-access memory (DRAM) includes a first memory bank, a second memory bank, and a dual mode I/O circuit. The first memory bank includes a memory array divided into a plurality of grains, each grain including a row buffer and input/output (I/O) circuitry. The dual-mode I/O circuit is coupled to the I/O circuitry of each grain in the first memory bank, and operates in a first mode in which commands having a first data width are routed to and fulfilled individually at each grain, and a second mode in which commands having a second data width different from the first data width are fulfilled by at least two of the grains in parallel.Type: ApplicationFiled: August 31, 2023Publication date: December 28, 2023Applicant: Advanced Micro Devices, Inc.Inventors: Sriseshan Srikanth, Vignesh Adhinarayanan, Jagadish B. Kotra, Sergey Blagodurov
-
Publication number: 20230418744Abstract: A technique for operating a cache is disclosed. The technique includes recording access data for a first set of memory accesses of a first frame; identifying parameters for a second set of memory accesses of a second frame subsequent to the first frame, based on the access data; and applying the parameters to the second set of memory accesses.Type: ApplicationFiled: June 27, 2022Publication date: December 28, 2023Applicant: Advanced Micro Devices, Inc.Inventors: Christopher J. Brennan, Akshay Lahiry
-
Publication number: 20230418782Abstract: Methods and systems are disclosed for performing dataflow execution by an accelerated processing unit (APU). Techniques disclosed include decoding information from one or more dataflow instructions. The decoded information is associated with dataflow execution of a computational task. Techniques disclosed further include configuring, based on the decoded information, dataflow circuitry, and, then, executing the dataflow execution of the computational task using the dataflow circuitry.Type: ApplicationFiled: June 28, 2022Publication date: December 28, 2023Applicant: Advanced Micro Devices, Inc.Inventors: Johnathan Robert Alsop, Karthik Ramu Sangaiah, Anthony T. Gutierrez
-
Publication number: 20230418745Abstract: A technique for operating a cache is disclosed. The technique includes utilizing a first portion of a cache in a directly accessed manner; and utilizing a second portion of the cache as a cache.Type: ApplicationFiled: June 28, 2022Publication date: December 28, 2023Applicant: Advanced Micro Devices, Inc.Inventors: Chintan S. Patel, Vydhyanathan Kalyanasundharam, Benjamin Tsien, Alexander J. Branover
-
Publication number: 20230420018Abstract: A data processor is for accessing a memory having a first pseudo channel and a second pseudo channel. The data processor includes at least one memory accessing agent, a memory controller, and a data fabric. The at least one memory accessing agent generates generating memory access requests including first memory access requests that access the memory. The memory controller provides memory commands to the memory in response to the first memory access requests. The data fabric routes the first memory access requests to a first downstream port in response to a corresponding first memory request accessing the first pseudo channel, and to a second downstream port in response to the corresponding first memory request accessing the second pseudo channel. The memory controller has first and second upstream ports coupled to the first and second downstream ports of the data fabric, respectively, and a downstream port coupled to the memory.Type: ApplicationFiled: June 24, 2022Publication date: December 28, 2023Applicants: ATI Technologies ULC, Advanced Micro Devices, Inc.Inventors: Xuan Chen, Chih-Hua Hsu, Pradeep Jayaraman, Abdussalam Aburwein
-
Publication number: 20230421203Abstract: An integrated circuit includes a first terminal for receiving a data signal, a second terminal for receiving an external reference voltage, a receiver, and a reference voltage generation circuit. The receiver is powered by a power supply voltage with respect to ground and has a first input coupled to the first terminal, a second input for receiving a shared reference voltage, and an output for providing a data input signal. The reference voltage generation circuit is coupled to the second terminal and receives the power supply voltage. The reference voltage generation circuit is operable to form the shared reference voltage by mixing noise from the power supply voltage and noise from the second terminal.Type: ApplicationFiled: September 7, 2023Publication date: December 28, 2023Applicant: Advanced Micro Devices, Inc.Inventors: Ramon Mangaser, Karthik Gopalakrishnan, Andy Huei Chu, Pradeep Jayaraman
-
Patent number: 11854139Abstract: A processing unit employs a hardware traversal engine to traverse an acceleration structure such as a ray tracing structure. The hardware traversal engine includes one or more memory modules to store state information and other data used for the structure traversal, and control logic to execute a traversal process based on the stored data and based on received information indicating a source node of the acceleration structure to be used for the traversal process. By employing a hardware traversal engine, the processing unit is able to execute the traversal process more quickly and efficiently, conserving processing resources and improving overall processing efficiency.Type: GrantFiled: December 28, 2021Date of Patent: December 26, 2023Assignee: Advanced Micro Devices, Inc.Inventors: Konstantin Igorevich Shkurko, Michael Mantor
-
Patent number: 11854138Abstract: Described herein is a technique for modifying a bounding volume hierarchy. The techniques include combining preferred orientations of child nodes of a first bounding box node to generate a first preferred orientation; based on the first preferred orientation, converting one or more child nodes of the first bounding box node into one or more oriented bounding box nodes; combining preferred orientations of child nodes of a second bounding box node to generate a second preferred orientation; and based on the second preferred orientation, maintaining one or more children of the second bounding box node as non-oriented bounding box nodes.Type: GrantFiled: July 23, 2021Date of Patent: December 26, 2023Assignee: Advanced Micro Devices, Inc.Inventors: Matthäus G Chajdas, Michael A. Kern, David Ronald Oldcorn
-
Patent number: 11853734Abstract: A processing system includes a compiler that automatically identifies sequences of instructions of tileable source code that can be replaced with tensor operations. The compiler generates enhanced code that replaces the identified sequences of instructions with tensor operations that invoke a special-purpose hardware accelerator. By automatically replacing instructions with tensor operations that invoke the special-purpose hardware accelerator, the compiler makes the performance improvements achievable through the special-purpose hardware accelerator available to programmers using high-level programming languages.Type: GrantFiled: May 10, 2022Date of Patent: December 26, 2023Assignee: Advanced Micro Devices, Inc.Inventors: Gregory P. Rodgers, Joseph L. Greathouse