Patents Assigned to Advanced Micro Devices, Incs.
-
Publication number: 20250110647Abstract: Non-blocking processing system are described. In accordance with the described techniques, a pending range store receives, at a start of a bulk memory operation, a pending memory range of the bulk memory operation. A logic unit includes at least one of check conflict logic or check address logic. The logic unit detects a conflicting memory access based on a target address of the pending memory range conflicting with a memory access request separate from the bulk memory operation. The logic unit performs at least a portion of the bulk memory operation associated with the target address before the memory access request is allowed to proceed.Type: ApplicationFiled: September 29, 2023Publication date: April 3, 2025Applicant: Advanced Micro Devices, Inc.Inventors: Johnathan Robert Alsop, Shaizeen Dilawarhusen Aga, Khaled Hamidouche
-
Publication number: 20250110861Abstract: In accordance with the described techniques for data compression using reconfigurable hardware based on data redundancy patterns, a computing device includes a memory, processing-in-memory units, a host processing unit, and a compression unit having reconfigurable logic for performing multiple compression algorithms. The host processing unit issues processing-in-memory requests instructing the processing-in-memory units to scan a block of the memory for one or more data redundancy patterns, and to identify a compression algorithm of the multiple compression algorithms based on the one or more data redundancy patterns. Further, the host processing unit issues a memory request to access a memory address in the block of the memory. The memory request causes data of the memory address to be communicated from the block of the memory to the compression unit to be compressed using the compression algorithm.Type: ApplicationFiled: September 29, 2023Publication date: April 3, 2025Applicant: Advanced Micro Devices, Inc.Inventors: Moumita Dey, Varun Agrawal
-
Publication number: 20250112639Abstract: An apparatus can include: a processor; a voltage regulator configured to provide a processor voltage and a processor current to the processor; and a voltage regulator controller that can include a current sensor comprising an analog-to-digital converter (ADC) having an ADC input range and configured to provide current data based on an ADC input voltage, and a configuration manager configured to receive processor power data and adjust the ADC input range based on the processor power data. Various other methods, systems, and computer-readable media are also disclosed.Type: ApplicationFiled: September 29, 2023Publication date: April 3, 2025Applicant: Advanced Micro Devices, Inc.Inventors: Wei Han, Meeta Surendramohan Srivastav, LiLi Chen, Indrani Paul
-
Publication number: 20250110773Abstract: The disclosed device includes a heterogeneous processor architecture having heterogeneous processors, and a control circuit that can assign, in response to an interrupt, the interrupt to one of the heterogenous processors that is selected based on power efficiency. Various other methods, systems, and computer-readable media are also disclosed.Type: ApplicationFiled: September 29, 2023Publication date: April 3, 2025Applicant: Advanced Micro Devices, Inc.Inventors: Pravesh Gupta, Benjamin Tsien
-
Publication number: 20250111601Abstract: A technique for rendering is provided. The technique includes performing a visibility pass that designates portions of shade space textures visible in a scene, wherein the visibility pass generates tiles that cover shade space textures visible in the scene; performing a rate controller operation on output of the visibility pass using spatiotemporal adaptive sampling; performing a shade space shading operation on the tiles that cover the shade space textures visible in the scene based on a result of the spatiotemporal adaptive sampling; performing a regularization operation based on an output of the shade space shading operation; and performing a reconstruction operation using output from the regularization operation to produce a final scene.Type: ApplicationFiled: September 29, 2023Publication date: April 3, 2025Applicants: Advanced Micro Devices, Inc., ATI Technologies ULCInventors: Guennadi Riguer, Michal Adam Wozniak
-
Publication number: 20250110664Abstract: A memory controller includes a command queue for receiving memory access requests and an arbiter. The arbiter is operable to allow cross-mode activations during a streak of accesses of a current mode in response to a number of cross-mode accesses present in the command queue exceeding an adaptive threshold.Type: ApplicationFiled: September 28, 2023Publication date: April 3, 2025Applicant: Advanced Micro Devices, Inc.Inventor: Guanhao Shen
-
Patent number: 12265732Abstract: A data processor that is operable to be coupled to a memory includes a memory operation array, a controller, a refresh logic circuit, and a selector. The memory operation array is for storing memory operations for a first power state of the memory. The controller is responsive to a power state change request to execute a plurality of memory operations from the memory operation array when the first power state is selected. The refresh logic circuit generates refresh cycles periodically for the memory. The selector is for multiplexing the refresh cycles with the memory operations during a power state change to the first power state.Type: GrantFiled: September 29, 2023Date of Patent: April 1, 2025Assignee: Advanced Micro Devices, Inc.Inventors: Jean J. Chittilappilly, Kevin M. Brandl, Jing Wang, Kedarnath Balakrishnan
-
Patent number: 12265908Abstract: Systems, apparatuses, and methods for achieving higher cache hit rates for machine learning models are disclosed. When a processor executes a given layer of a machine learning model, the processor generates and stores activation data in a cache subsystem a forward or reverse manner. Typically, the entirety of the activation data does not fit in the cache subsystem. The processor records the order in which activation data is generated for the given layer. Next, when the processor initiates execution of a subsequent layer of the machine learning model, the processor processes the previous layer's activation data in a reverse order from how the activation data was generated. In this way, the processor alternates how the layers of the machine learning model process data by either starting from the front end or starting from the back end of the array.Type: GrantFiled: August 31, 2020Date of Patent: April 1, 2025Assignees: Advanced Micro Devices, Inc., ATI Technologies ULCInventors: Benjamin Thomas Sander, Swapnil Sakharshete, Ashish Panday
-
Patent number: 12265496Abstract: An apparatus and method for efficiently supporting multiple peripheral communication protocols in a computing system. A computing system includes multiple servers with one or more of the servers using multiple connectors for connecting to multiple peripheral devices such as data storage devices. At least one of the connectors is able to support multiple communication protocols, rather than a single communication protocol. A processor of the server determines a peripheral device has been attached to a connector that supports multiple communication protocols, and the processor determines whether one of the multiple communication protocols supported by the particular connector matches the attached peripheral device's communication protocol. If so, the processor configures the connector with the matching communication protocol. Otherwise, the processor generates an indication that specifies that there is no match.Type: GrantFiled: June 30, 2022Date of Patent: April 1, 2025Assignee: Advanced Micro Devices, Inc.Inventors: Brian Mitchell, George D. Azevedo
-
Patent number: 12265915Abstract: A technique for manipulating a generic tensor is provided. The technique includes receiving a first request to perform a first operation on a generic tensor descriptor associated with the generic tensor, responsive to the first request, performing the first operation on the generic tensor descriptor, receiving a second request to perform a second operation on generic tensor raw data associated with the generic tensor, and responsive to the second request, performing the second operation on the generic tensor raw data, the performing the second operation including mapping a tensor coordinate specified by the second request to a memory address, the mapping including evaluating a delta function to determine an address delta value to add to a previously determined address for a previously processed tensor coordinate.Type: GrantFiled: December 30, 2020Date of Patent: April 1, 2025Assignee: Advanced Micro Devices, Inc.Inventors: Chao Liu, Daniel Isamu Lowell, Wen Heng Chung, Jing Zhang
-
Patent number: 12265484Abstract: An accelerated processing device is provided which comprises a plurality of compute units each including a plurality of SIMD units, and each SIMD unit comprises a register file. The accelerated processing device also comprises LDS in communication with each of the SIMD units. The accelerated processing device also comprises a first portion of cache memory, in communication with each of the SIMD units and a second cache portion of memory shared by the compute units. The compute units are configured to execute a program in which a storage portion of at least one of the register file of a SIMD unit, the first portion of cache memory and the LDS is reserved as part of another of the register file, the first portion of cache memory and the LDS.Type: GrantFiled: September 3, 2021Date of Patent: April 1, 2025Assignee: Advanced Micro Devices, Inc.Inventor: Maxim V. Kazakov
-
Patent number: 12265467Abstract: Enhanced methods for memory context restore are described. A device may include a physical layer (PHY) having an interface to support communication of command signals and data with a physical memory. The PHY implements a training mode to train the interface, detect values of a plurality of parameters as part of training the interface, and store the detected values as initial training data. The PHY also implements a retraining mode to use the initial training data as seed data to retrain the interface.Type: GrantFiled: September 29, 2023Date of Patent: April 1, 2025Assignee: Advanced Micro Devices, Inc.Inventors: Anwar Parvez Kashem, Alicia Wen Ju Yurie Leong, Glennis Eliagh Covington
-
Patent number: 12266139Abstract: A method and apparatus for integrating data compression in a computer system includes receiving first data at a first system level. Based upon a number of planes of the first data being less than or equal to a threshold, the data is compressed with a first data compression scheme, and transferred to a second system level for processing. Based upon the number of planes of the first data exceeding the threshold, the first data is transferred uncompressed to the second system level for processing. Based upon the received data at the second system level being compressed with the first compression scheme, the data is transferred to a third system level, and based upon the received data at the second system level being uncompressed with the first compression scheme, compressing the data with a second compression scheme, and transferring the compressed data to the third system level.Type: GrantFiled: December 14, 2021Date of Patent: April 1, 2025Assignee: Advanced Micro Devices, Inc.Inventors: Christopher J. Brennan, Pazhani Pillai
-
Patent number: 12266030Abstract: Systems and methods related to priority-based and performance-based selection of a render mode, such as a two-level binning mode, in which to execute workloads with a graphics processing unit (GPU) of a system are provided. A user mode driver (UMD) or kernel mode driver (KMD) executed at a central processing unit (CPU) configures low and medium priority workloads to be executed in a two-level binning mode and selects a binning mode for high priority workloads based on whether performance heuristics indicate that one or more binning conditions or override conditions have been met. High priority workloads are maintained in a high priority queue, while low and medium priority workloads are maintained in a low/medium priority queue, such that execution of low and medium priority workloads at the GPU can be preempted in favor of executing high priority workloads.Type: GrantFiled: April 15, 2021Date of Patent: April 1, 2025Assignee: Advanced Micro Devices, Inc.Inventors: Anirudh R. Acharya, Ruijin Wu, Young In Yeo
-
Patent number: 12265470Abstract: Selectively bypassing cache directory lookups for processing-in-memory instructions is described. In one example, a system maintains information describing a status—clean or dirty—of a memory address, where a dirty status indicates that the memory address is modified in a cache and thus different than the memory address as represented in system memory. A processing-in-memory request involving the memory address is assigned a cache directory bypass bit based on the status of the memory address. The cache directory bypass bit for a processing-in-memory request controls whether a cache directory lookup is performed after the processing-in-memory request is issued by a processor core and before the processing-in-memory request is executed by a processing-in-memory component.Type: GrantFiled: September 29, 2023Date of Patent: April 1, 2025Assignee: Advanced Micro Devices, Inc.Inventors: Travis Henry Boraten, Jagadish B. Kotra, David Andrew Werner
-
Patent number: 12265735Abstract: An approach is provided for processing near-memory processing commands, e.g., PIM commands, using PIM register definition data that defines multiple combinations of source and/or destination registers to be used to process PIM commands. A particular combination of source and/or destination registers to be used to process a PIM command is specified by the PIM command or determined by a near-memory processing element processing the PIM command. According to another implementation, the PIM register definition data specifies an initial combination of source and/or destination registers and one or more update functions for each PIM command. A near-memory processing element processes a PIM command using the initial combination of source and/or destination registers and uses the one or more update functions to update the combination of source and/or destination registers to be used the next time the PIM command is processed.Type: GrantFiled: June 21, 2022Date of Patent: April 1, 2025Assignee: Advanced Micro Devices, Inc.Inventors: Shaizeen Aga, Nuwan Jayasena
-
Publication number: 20250106508Abstract: A technique for generating video is provided. The technique includes obtaining a plurality of source frames with a wide-angle camera and a narrow-angle camera; identifying a plurality of central portions and a plurality of peripheral portions of the plurality of source frames based on image stabilization; and combining the plurality of central portions and the plurality of peripheral portions to generate a plurality of resulting frames of an output video.Type: ApplicationFiled: September 27, 2023Publication date: March 27, 2025Applicant: Advanced Micro Devices, Inc.Inventor: Po-Min Wang
-
Publication number: 20250103650Abstract: Graph analytics system are described. In accordance with the described techniques, a graph having vertices that include a first vertex and a second vertex that are associated with access control metadata are received. An updated graph is output based on a merging of the first vertex and the second vertex into a merged vertex of a group of vertices based on the first vertex and the second vertex being associated with access control metadata common to the first vertex and the second vertex and based on a reordering technique. A single copy of the access control metadata is stored for the first vertex and the second vertex.Type: ApplicationFiled: September 21, 2023Publication date: March 27, 2025Applicant: Advanced Micro Devices, Inc.Inventors: Kishore Punniyamurthy, Jagadish B. Kotra
-
Publication number: 20250103340Abstract: A computer-implemented method for resynchronization at execution time can include detecting, by at least one processor and during an execution time of an instruction, a resynchronization. The method can additionally include regenerating, by the at least one processor and in response to the detection, an instruction pointer. The method can also include performing, by the at least one processor and during the execution time of the instruction, the resynchronization by using the regenerated instruction pointer. Various other methods and systems are also disclosed.Type: ApplicationFiled: September 27, 2023Publication date: March 27, 2025Applicant: Advanced Micro Devices, Inc.Inventors: Upamanyu Banerjee, Arun A. Nair
-
Publication number: 20250103395Abstract: A computer-implemented method for dynamic resource management can include evaluating, by at least one processor, whether a priority of one or more processes associated with a request for one or more shared resources meets a threshold condition. The method can additionally include determining, by the at least one processor and in response to an evaluation that the priority meets the threshold condition, whether the one or more shared resources is available to meet the request. The method can further include completing, by the at least one processor and in response to a determination that the one or more shared resources is available, execution of the one or more processes. Various other methods, systems, and computer-readable media are also disclosed.Type: ApplicationFiled: September 27, 2023Publication date: March 27, 2025Applicant: Advanced Micro Devices, Inc.Inventors: Bradford Beckmann, Matthew David Sinclair, Vinay Bharadwaj Ramakrishnaiah, William Peter Ehrett