Patents Assigned to Advanced Micro Devices
-
Publication number: 20250117523Abstract: A method can include overriding settings of an integrated circuit device by reading one or more settings from a setting record that correspond to a part number of the integrated circuit device. The method can also include performing an override of the settings of the integrated circuit device based on the one or more settings of the setting record that correspond to the part number of the integrated circuit device. Various other methods and systems are also disclosed.Type: ApplicationFiled: October 9, 2024Publication date: April 10, 2025Applicants: ATI Technologies ULC, Advanced Micro Devices, Inc.Inventors: Valeri Kirischian, Steven Leonard Roberts, Ruchir Badola
-
Patent number: 12271244Abstract: Systems, methods, and devices for integrated circuit power management. A mode of a power management state is entered, from the power management state, in response to an entry condition of the mode. A device that is otherwise powered off in the power management state is powered on in the mode of the power management state. In some implementations, the device includes a communications path between a second device and a third device. In some implementations, the device is in a power domain that is powered off in the power management state. In some implementations, the power domain is powered off in the mode. In some implementations, the device is powered on in the mode via a power rail that is specific to the mode. In some implementations, the entry condition of the mode includes an amount of data stored for display in a display buffer falling below a threshold amount.Type: GrantFiled: July 30, 2021Date of Patent: April 8, 2025Assignee: Advanced Micro Devices, Inc.Inventors: Benjamin Tsien, Indrani Paul, Alexander J. Branover, Thomas J. Gibney, Mihir Shaileshbhai Doctor, John P. Petry, Stephen V. Kosonocky, Christopher T. Weaver
-
Patent number: 12272000Abstract: Concurrently with performing a visibility pass to generate visibility data for two or more bins of an image, a processing system determines whether a primitive to be rendered covers at least a predetermined threshold percentage of a tile of the image. In response to the primitive coving at least the predetermined threshold percentage of the tile, the processing system stores the depth data of the primitive in a depth buffer for pixel-based rendering. In response to the primitive not covering at least the predetermined threshold percentage of the tile, the processing system fuses the primitive with one or more preceding primitives sharing an edge with the primitive in the tile to generate a fused primitive. In response to the fused primitive being valid in the tile, the processing system passes the depth data of the fused primitive to the depth buffer.Type: GrantFiled: June 29, 2022Date of Patent: April 8, 2025Assignee: Advanced Micro Devices, Inc.Inventors: Kiia K. Kallio, Jan Achrenius
-
Patent number: 12271627Abstract: An apparatus and method for efficiently managing performance among multiple integrated circuits in separate semiconductor chips. In various implementations, a computing system includes at least a first processing node and a second processing node. While processing tasks, the first processing node accesses a first memory and the second processing node accesses a second memory. A first communication channel transfers data between the first and second processing nodes. The first processing node accesses the second memory using a second communication channel different from the first communication channel and supports point-to-point communication. The second memory services access requests from the first and second processing nodes as the access requests are received while foregoing access conflict detection. The first processing node accesses the second memory after a particular amount of time has elapsed after reception of an indication from the second processing node specifying that a particular task has begun.Type: GrantFiled: September 30, 2022Date of Patent: April 8, 2025Assignees: Advanced Micro Devices, Inc., ATI Technologies ULCInventors: Michael John Austin, Dmitri Tikhostoup
-
Patent number: 12274046Abstract: A system and method for efficiently creating layout for memory bit cells are described. In various implementations, a memory bit cell uses Cross field effect transistors (FETs) that include vertically stacked gate all around (GAA) transistors with conducting channels oriented in an orthogonal direction between them. The channels of the vertically stacked transistors use opposite doping polarities. The memory bit cell includes one of a read bit line and a write word line routed in no other metal layer other than a local interconnect layer. In addition, a six transistor (6T) random access data storage of the given memory bit cell consumes a planar area above a silicon substrate of four transistors.Type: GrantFiled: October 3, 2023Date of Patent: April 8, 2025Assignee: Advanced Micro Devices, Inc.Inventor: Richard T. Schultz
-
Patent number: 12273277Abstract: Systems and methods for allocating computing resources within a distributed computing system are disclosed. Computing resources such as CPUs, GPUs, network cards, and memory are allocated to jobs submitted to the system by a scheduler. System configuration and interconnectivity information is gathered by a mapper and used to create a graph. Resource allocation is optimized based on one or more quality of service (QoS) levels determined for the job. Job performance characterization, affinity models, computer resource power consumption, and policies may also be used to optimize the allocation of computing resources.Type: GrantFiled: May 9, 2023Date of Patent: April 8, 2025Assignee: Advanced Micro Devices, Inc.Inventors: Max Alt, Paulo Roberto Pereira de Souza filho
-
Patent number: 12271588Abstract: The disclosed device includes a memory-semantic fabric comprising memory components accessible by multiple processors and a controller for the memory-semantic fabric. The controller receives, from multiple processes, memory requests for a memory-semantic fabric. The controller also identifies, within the processes, a source process for each of the memory requests and prioritizes forwarding the memory requests to the memory-semantic fabric based on the source processes. Various other methods, systems, and computer-readable media are also disclosed.Type: GrantFiled: March 30, 2023Date of Patent: April 8, 2025Assignee: Advanced Micro Devices, Inc.Inventors: Atul Kumar Sujayendra Sandur, Sergey Blagodurov, Nathaniel Morris
-
Patent number: 12271597Abstract: A memory package includes first, second, third, and fourth channels arranged consecutively in a clockwise direction on the memory package, each of the first, second, third, and fourth channels having access circuitry and memory arrays. In a first mode, the first channel controls access to the memory arrays in the second channel and the fourth channel controls access to the memory arrays in the third channel.Type: GrantFiled: November 7, 2022Date of Patent: April 8, 2025Assignees: ATI Technologies ULC, Advanced Micro Devices, Inc.Inventors: Xuan Chen, Ross V. La Fetra, Michael John Litt
-
Patent number: 12271318Abstract: Method and apparatus monitor eviction conflicts among cache directory entries in a cache directory and produce cache directory victim entry information for a memory manager. In some examples, the memory manager reduces future cache directory conflicts by changing a page level physical address assignment for a page of memory based on the produced cache directory victim entry information. In some examples, a scalable data fabric includes hardware control logic that performs the monitoring of the eviction conflicts among cache directory entries in the cache directory and produces the cache directory victim entry information.Type: GrantFiled: December 28, 2020Date of Patent: April 8, 2025Assignee: ADVANCED MICRO DEVICES, INC.Inventors: Brandon K. Potter, Marko Scrbak, Sergey Blagodurov, Kishore Punniyamurthy, Nathaniel Morris
-
Publication number: 20250110538Abstract: The disclosed device includes a processing component having various compute blocks, and a control circuit that switches at least one of the compute blocks from a normal voltage rail for the processing component to a second voltage rail in response to power gating a normal voltage rail. Various other methods, systems, and computer-readable media are also disclosed.Type: ApplicationFiled: September 28, 2023Publication date: April 3, 2025Applicant: Advanced Micro Devices, Inc.Inventors: Meeta Surendramohan Srivastav, Indrani Paul, Akila Subramaniam
-
Publication number: 20250110655Abstract: Efficient memory operation using a destructive read memory array is described. In accordance with the described techniques, a system may include a memory configured to store data of a first logic state in a ferroelectric capacitor when an electric polarization of the ferroelectric capacitor is in a first direction. A system may include a controller configured to erase the data from the memory by commanding the electric polarization of the ferroelectric capacitor in a second direction, opposite of the first direction and skipping a subsequent write operation of a null value to the memory.Type: ApplicationFiled: September 28, 2023Publication date: April 3, 2025Applicant: Advanced Micro Devices, Inc.Inventors: Jagadish B. Kotra, Divya Madapusi Srinivas Prasad
-
Publication number: 20250110878Abstract: Selectively bypassing cache directory lookups for processing-in-memory instructions is described. In one example, a system maintains information describing a status—clean or dirty—of a memory address, where a dirty status indicates that the memory address is modified in a cache and thus different than the memory address as represented in system memory. A processing-in-memory request involving the memory address is assigned a cache directory bypass bit based on the status of the memory address. The cache directory bypass bit for a processing-in-memory request controls whether a cache directory lookup is performed after the processing-in-memory request is issued by a processor core and before the processing-in-memory request is executed by a processing-in-memory component.Type: ApplicationFiled: September 29, 2023Publication date: April 3, 2025Applicant: Advanced Micro Devices, Inc.Inventors: Travis Henry Boraten, Jagadish B. Kotra, David Andrew Werner
-
Publication number: 20250110647Abstract: Non-blocking processing system are described. In accordance with the described techniques, a pending range store receives, at a start of a bulk memory operation, a pending memory range of the bulk memory operation. A logic unit includes at least one of check conflict logic or check address logic. The logic unit detects a conflicting memory access based on a target address of the pending memory range conflicting with a memory access request separate from the bulk memory operation. The logic unit performs at least a portion of the bulk memory operation associated with the target address before the memory access request is allowed to proceed.Type: ApplicationFiled: September 29, 2023Publication date: April 3, 2025Applicant: Advanced Micro Devices, Inc.Inventors: Johnathan Robert Alsop, Shaizeen Dilawarhusen Aga, Khaled Hamidouche
-
Publication number: 20250110894Abstract: Scratchpad memory translation lookaside buffer techniques are described. In an implementation, the techniques described herein relate to a device including a memory management unit implemented in hardware of an integrated circuit to receive a mapping instruction from a mapping instruction source, the mapping instruction specifying a mapping between a virtual memory address and a physical memory address of a scratchpad memory and store a virtual-to-physical mapping entry in a translation lookaside buffer based on the mapping instruction.Type: ApplicationFiled: September 28, 2023Publication date: April 3, 2025Applicant: Advanced Micro Devices, Inc.Inventor: Mark Evan Wilkening
-
Publication number: 20250110864Abstract: Enhanced methods for memory context restore are described. A device may include a physical layer (PHY) having an interface to support communication of command signals and data with a physical memory. The PHY implements a training mode to train the interface, detect values of a plurality of parameters as part of training the interface, and store the detected values as initial training data. The PHY also implements a retraining mode to use the initial training data as seed data to retrain the interface.Type: ApplicationFiled: September 29, 2023Publication date: April 3, 2025Applicant: Advanced Micro Devices, Inc.Inventors: Anwar Parvez Kashem, Alicia Wen Ju Yurie Leong, Glennis Eliagh Covington
-
Publication number: 20250110773Abstract: The disclosed device includes a heterogeneous processor architecture having heterogeneous processors, and a control circuit that can assign, in response to an interrupt, the interrupt to one of the heterogenous processors that is selected based on power efficiency. Various other methods, systems, and computer-readable media are also disclosed.Type: ApplicationFiled: September 29, 2023Publication date: April 3, 2025Applicant: Advanced Micro Devices, Inc.Inventors: Pravesh Gupta, Benjamin Tsien
-
Publication number: 20250111582Abstract: A technique for rendering is provided. The technique includes determining a level of detail for a shade space texture and a screen space; shading the shade space texture having a resolution based on the level of detail; and for a reconstruction operation, performing sampling from the shade space texture, the sampling including a high frequency attenuation of samples of the shade space texture.Type: ApplicationFiled: September 28, 2023Publication date: April 3, 2025Applicants: Advanced Micro Devices, Inc., ATI Technologies ULCInventors: Boris Ivanovic, Guennadi Riguer, Michal Adam Wozniak
-
Publication number: 20250110898Abstract: Method and apparatus for collaborative memory accesses is described. A system includes a memory controller that receives a command from a host. The command is associated with at least one of a plurality of data elements. The memory controller causes execution of data casting operations that adjust a bit size of the plurality of data elements to generate casted data elements. The system includes an interface for communicating data between the host and a memory.Type: ApplicationFiled: September 29, 2023Publication date: April 3, 2025Applicant: Advanced Micro Devices, Inc.Inventors: Shaizeen Dilawarhusen Aga, Mahzabeen Islam, Nuwan S. Jayasena
-
Publication number: 20250110861Abstract: In accordance with the described techniques for data compression using reconfigurable hardware based on data redundancy patterns, a computing device includes a memory, processing-in-memory units, a host processing unit, and a compression unit having reconfigurable logic for performing multiple compression algorithms. The host processing unit issues processing-in-memory requests instructing the processing-in-memory units to scan a block of the memory for one or more data redundancy patterns, and to identify a compression algorithm of the multiple compression algorithms based on the one or more data redundancy patterns. Further, the host processing unit issues a memory request to access a memory address in the block of the memory. The memory request causes data of the memory address to be communicated from the block of the memory to the compression unit to be compressed using the compression algorithm.Type: ApplicationFiled: September 29, 2023Publication date: April 3, 2025Applicant: Advanced Micro Devices, Inc.Inventors: Moumita Dey, Varun Agrawal
-
Publication number: 20250110884Abstract: Systems and techniques for selectively transferring one or more portions of a cache block in response to a request are described. Computing system components are informed as to instances where data transfer operations involve moving less than an entirety of data included in a cache block cache block. In one example, executable code for a computational task includes hints that identify when memory requests involve accessing and transmitting less than an entirety of a cache block and cause system components to communicate a subset of the cache block during a memory access. In another example, a data differentiator unit is implemented to analyze a cache block and return a portion of the cache block that is selected based on one or more criteria specified for a computational task. The described techniques thus overcome conventional drawbacks facing systems that transmit an entire cache block when only a portion is needed.Type: ApplicationFiled: September 29, 2023Publication date: April 3, 2025Applicant: Advanced Micro Devices, Inc.Inventors: Shaizeen Dilawarhusen Aga, Nuwan S. Jayasena, Michael J. Schulte, Srilatha Manne