Overlapped Cache Accessing, E.g., Pipeline, Etc. (epo) Patents (Class 711/E12.049)
-
Patent number: 10216516Abstract: A processing device includes a store instruction identification unit to identify a pair of store instructions among a plurality of instructions in an instruction queue. The pair of store instructions include a first store instruction and a second store instruction. The first data of the first store instruction corresponds to a first memory region adjacent to a second memory region, and a second data of the second store instruction corresponds to the second memory region. The processing device to include a store instruction fusion unit to fuse the first store instruction with the second store instruction resulting in a fused store instruction.Type: GrantFiled: September 30, 2016Date of Patent: February 26, 2019Assignee: Intel CorporationInventors: Sebastian Winkel, Jamison D. Collins, Tyler Sondag
-
Patent number: 10146545Abstract: Embodiments related to fetching instructions and alternate versions achieving the same functionality as the instructions from an instruction cache included in a microprocessor are provided. In one example, a method is provided, comprising, at an example microprocessor, fetching an instruction from an instruction cache. The example method also includes hashing an address for the instruction to determine whether an alternate version of the instruction which achieves the same functionality as the instruction exists. The example method further includes, if hashing results in a determination that such an alternate version exists, aborting fetching of the instruction and retrieving and executing the alternate version.Type: GrantFiled: March 13, 2012Date of Patent: December 4, 2018Assignee: Nvidia CorporationInventors: Ross Segelken, Alex Klaiber, Nathan Tuck, David Dunn
-
Patent number: 8966221Abstract: A lookup operation is performed in a translation look aside buffer based on a first translation request as current translation request, wherein a respective absolute address is returned to a corresponding requestor for the first translation request as translation result in case of a hit. A translation engine is activated to perform at least one translation table fetch in case the current translation request does not hit an entry in the translation look aside buffer, wherein the translation engine is idle waiting for the at least one translation table fetch to return data, reporting the idle state of the translation engine as lookup under miss condition and accepting a currently pending translation request as second translation request, wherein a lookup under miss sequence is performed in the translation look aside buffer based on said second translation request.Type: GrantFiled: June 21, 2011Date of Patent: February 24, 2015Assignee: International Business Machines CorporationInventors: Ute Gaertner, Thomas Koehler
-
Patent number: 8683125Abstract: A tier identification (TID) is to indicate a characteristic of a memory region associated with a virtual address in a tiered memory system. A thread may be serviced according to a first path based on the TID indicating a first characteristic. The thread may be serviced according to a second path based on the TID indicating a second characteristic.Type: GrantFiled: November 1, 2011Date of Patent: March 25, 2014Assignee: Hewlett-Packard Development Company, L.P.Inventors: Jichuan Chang, Kevin T Lim, Parthasarathy Ranganathan
-
Patent number: 8589656Abstract: Queuing of received transactions that have a resource conflict is disclosed. A first node receives a first transaction from a second node, where the first transaction relates to a resource of the first node. The transaction may be a request relating to a memory line of the first node, for instance. It is determined that a second transaction that relates to this resource of the first node is already being processed by the first node. Therefore, the first transaction is enqueued in a conflict queue within the first node. The queuing may be a linked list, a priority queue, or another type of queue. Once the second transaction has been processed, the first transaction is restarted for processing by the first node. The first transaction is then processed by the first node.Type: GrantFiled: June 25, 2011Date of Patent: November 19, 2013Assignee: International Business Machines CorporationInventors: Donald R. DeSota, Robert Joersz, Davis A. Miller, Maged M. Michael
-
Patent number: 8578117Abstract: Write-through-read (WTR) comparator circuits and related WTR processes and memory systems are disclosed. The WTR comparator circuits can be configured to perform WTR functions for a multiple port file having one or more read and write ports. One or more WTR comparators in the WTR comparator circuit are configured to compare a read index into a file with a write index corresponding to a write-back stage selected write port among a plurality of write ports that can write data to the entry in the file. The WTR comparators then generate a WTR comparator output indicating whether the write index matches the read index to control a WTR function. In this manner, the WTR comparator circuit can employ less WTR comparators than the number of read and write port combinations. Providing less WTR comparators can reduce power consumption, cost, and area required on a semiconductor die for the WTR comparator circuit.Type: GrantFiled: February 10, 2010Date of Patent: November 5, 2013Assignee: QUALCOMM IncorporatedInventors: Gregory Christopher Burda, Michael Scott McIlvaine, Nathan Samuel Nunamker, Yeshwant Nagaraj Kolla
-
Patent number: 8566539Abstract: A method, system, and computer usable program product for managing thermal condition of a memory are provided in the illustrative embodiments. A condition that a threshold value of a thermal condition of the memory has been exceeded or is likely to be exceeded is identified. A portion of a first workload is identified as being a cause of exceeding the threshold. A second portion of a second workload is identified, the second portion not causing the threshold to be exceeded when executed. A set of operations corresponding to the first portion is interleaved with a second set of operations corresponding to the second portion. The interleaved first and second portions of the first and second workloads are executed, causing the thermal condition of the memory to remain below the threshold. The second portion may use a second memory, a second area of the memory, or a combination thereof when executing.Type: GrantFiled: January 14, 2009Date of Patent: October 22, 2013Assignee: International Business Machines CorporationInventors: Diane Garza Flemming, Ghadir Robert Gholami, Octavian Florin Herescu, William A Maron, Mysore Sathyanarayana Srinivas
-
Patent number: 8464016Abstract: A method and device for switching over in a memory for a control device, a first storage area in the control unit being overlaid by a second storage area; the second storage area including at least one memory page, and each of the memory pages being able to overlay the first storage area; switching over being able to be performed between the memory pages and the overlaying being able to be switched on/switched off, and the switching over of the memory pages and the switching on/switching off of the overlaying of the second storage area being automatically carried out and/or triggered by the software of the control unit.Type: GrantFiled: May 22, 2006Date of Patent: June 11, 2013Assignee: Robert Bosch GmbHInventors: Claus Moessner, Gert Maier, Claus Spizig, Frank Lustig, Jens Schneider
-
Patent number: 8335892Abstract: One embodiment of the present invention sets forth a technique for arbitrating requests received by an L1 cache from multiple clients. The L1 cache outputs bubble requests to a first one of the multiple clients that cause the first one of the multiple clients to insert bubbles into the request stream, where a bubble is the absence of a request. The bubbles allow the L1 cache to grant access to another one of the multiple clients without stalling the first one of the multiple clients. The L1 cache services multiple clients with diverse latency and bandwidth requirements and may be reconfigured to provide memory spaces for clients executing multiple parallel threads, where the memory spaces each have a different scope.Type: GrantFiled: December 30, 2009Date of Patent: December 18, 2012Assignee: NVIDIA CorporationInventors: Alexander L. Minkin, Steven J. Heinrich, Rajeshwaran Selvanesan, Charles McCarver, Stewart Glenn Carlton, Anjana Rajendran
-
Patent number: 8266382Abstract: One embodiment of the present invention sets forth a technique for arbitrating requests received from one of the multiple clients of an L1 cache and for providing hints to the client to assist in arbitration. The L1 cache services multiple clients with diverse latency and bandwidth requirements and may be reconfigured to provide memory spaces for clients executing multiple parallel threads, where the memory spaces each have a different scope.Type: GrantFiled: December 30, 2009Date of Patent: September 11, 2012Assignee: NVIDIA CorporationInventors: Alexander L. Minkin, Steven J. Heinrich, Rajeshwaran Selvanesan, Charles McCarver, Stewart Glenn Carlton, Anjana Rajendran, Yan Yan Tang
-
Patent number: 8230175Abstract: A system and method optimizing data throughput to a processor from a storage device having sequential data access capabilities where the processor enables its data cache for memory operations involving the storage device. The system includes a processor coupled to the data storage device, e.g., a NAND flash memory. The processor establishes an address window used as a cache (CW) for reading data from the flash memory and also establishes a non-cacheable address window (NCW) for commands, address delivery and writes to the flash memory. The CW is sized to be larger than the processor data cache to ensure that reads from the flash memory always encounter a cache-miss so that read data is obtained directly from the flash memory. By reading through the CW from the flash memory, the processor takes advantage of bursting, pipelining and data prefetch efficiencies which significantly increase data throughput.Type: GrantFiled: August 9, 2005Date of Patent: July 24, 2012Assignee: Hewlett-Packard Development Company, L.P.Inventors: Nicholas Vaccaro, Mostafa Kashi
-
Patent number: 8166246Abstract: A computer implemented method, a processor chip, a data processing system, and computer program product in a data processing system process information in a store cache of a data processing system. The store cache receives a first entry that includes a first address indicating a first segment of a cache line. The store cache then receives a second entry including a second address indicating a second segment of the cache line. Responsive to the first segment not being equal to the second segment, the first entry is chained to the second entry.Type: GrantFiled: January 31, 2008Date of Patent: April 24, 2012Assignee: International Business Machines CorporationInventors: Guy Lynn Guthrie, Thomas Leo Jeremiah, William Lloyd McNeil, Hugh Shen, William John Starke
-
Publication number: 20100095068Abstract: With a view to reducing the congestion of a pipeline for cache memory access in, for example, a multi-core system, a cache memory control device includes: a determination unit for determining whether or not a command provided from, for example, each core is to access cache memory during the execution of the command; and a path switch unit for putting a command determined as accessing the cache memory in pipeline processing, and outputting a command determined as not accessing the cache memory directly to an external unit without putting the command in the pipeline processing.Type: ApplicationFiled: December 11, 2009Publication date: April 15, 2010Applicant: FUJITSU LIMITEDInventors: Koken Shimizuno, Naoya Ishimura
-
Publication number: 20090063779Abstract: A cache memory that includes: (i) an arbitrator, connected to multiple access generator, the arbitrator is adapted to receive different types of access requests from the multiple access generators and to select a single access request per arbitration cycle; (ii) a sequence of pipeline stages, the sequence comprises an input pipeline stage that is connected to the arbiter; and (iii) multiple cache resources, connected to the sequence of pipeline stages; wherein each cache resource can be read only by a small portion of the sequence of pipeline stages and can be written to only by a small portion of the sequence of pipeline stages.Type: ApplicationFiled: September 4, 2007Publication date: March 5, 2009Applicant: Freescale Semiconductor Inc.Inventors: Shai Koren, Alon Eldar, Amit Gur, Itay Peled, Rotem Porat
-
Publication number: 20090006763Abstract: An arrangement and method for update of configuration cache data in a disk storage subsystem in which a cache memory (110) is updated using two-phase (220, 250) commit technique. This provides the advantage that known changes to the subsystem do not require an invalidate/rebuild style operation on the cache. This is especially important where a change will invalidate the entire cache.Type: ApplicationFiled: July 11, 2008Publication date: January 1, 2009Inventors: David John Carr, Michael John Jones, Andrew Key, Robert Bruce Nicholson, William James Scales, Barry Douglas Whyte