Multiple Simultaneous Or Quasi-simultaneous Cache Accessing (epo) Patents (Class 711/E12.044)
  • Patent number: 11966363
    Abstract: Disclosed is a new identifier allocation approach for allocating repository object identifiers beyond a maximum number usually allowed for a single repository. The new identifier allocation approach includes an enhanced identifier structure in which a portion usually reserved for a docbase identifier is modified to combine with a partition identifier. The actual docbase identifier is stored in a server configuration file. When the maximum number of object identifiers has been allocated, a new partition is created and a sequence table is updated with a new entry that reflects the new partition. A new repository object identifier is allocated using the docbase identifier combined with the new partition identifier. In this way, more than 4 billion objects per object type can be addressed uniquely within a docbase by creating new partitions. This new approach is very unique and flexible in accommodating both on premises and multi-tenant environments.
    Type: Grant
    Filed: November 29, 2021
    Date of Patent: April 23, 2024
    Assignee: OPEN TEXT HOLDINGS, INC.
    Inventors: Pulla Rayudu Garaga, Satya Sai Chaitanya Patchigolla
  • Patent number: 11934310
    Abstract: In one embodiment, a microprocessor, comprising: plural cores, each of the cores comprising a level 1 (L1) cache and a level 2 (L2) cache; and a shared level 3 (L3) cache comprising plural L3 tag array entries, wherein a first portion of the plural L3 tag array entries is associated with data and a second portion of the plural L3 tag array entries is decoupled from data, wherein each L3 tag array entry comprises tag information and data zero information, the data zero information indicating whether any data associated with the tag information is known to be zero or not.
    Type: Grant
    Filed: January 21, 2022
    Date of Patent: March 19, 2024
    Assignee: CENTAUR TECHNOLOGY, INC.
    Inventors: Douglas Raye Reed, Al Loper, Terry Parks
  • Patent number: 8924678
    Abstract: A memory card is attached to a host device, and includes a data control circuit which transfers data with respect to the host device in synchronism with a rise edge and a fall edge of a clock signal.
    Type: Grant
    Filed: January 13, 2011
    Date of Patent: December 30, 2014
    Assignee: Kabushiki Kaisha Toshiba
    Inventor: Takafumi Ito
  • Patent number: 8762633
    Abstract: A device for recording data emitted from a motor vehicle management system, including a volatile memory, a non-volatile memory, and a data recording module configured to receive a signal concerning activation status emitted by the management system and to record the data in a first zone of the volatile memory on a rising edge of the activation status signal and in a second zone of the volatile memory on a falling edge of the activation status signal, and including a record management module configured to receive the activation status signal and to activate a command to record on a falling edge with the activation status signal, the recording module being further configured to receive the record command and to record the content of the two zones of the volatile memory into a zone of the non-volatile memory when the record command is activated.
    Type: Grant
    Filed: June 12, 2009
    Date of Patent: June 24, 2014
    Assignee: Renault S.A.S.
    Inventors: Alessandro Monti, Arnaud Losq, Eric Mounier
  • Patent number: 8732387
    Abstract: A flash memory controller with automatic interface mode switching is applied to a flash memory apparatus with a plurality of flash memories and the controller contains: a memory interface, a microprocessor, and an interface mode controller. The microprocessor recognizes the supported interface mode of every flash memory connected with the memory interface in an initial setting process, and individually sets the corresponding interface mode setting value into the interface mode controller. Thus, when the flash memory apparatus is operating in a normal operation state, the interface mode controller can output the corresponding interface mode setting value according to the present enabled flash memory, and the memory interface can adjust and switch the interface mode according to the interface mode setting value outputted by the interface mode controller. Thereby, the present invention can achieve the purpose whereby the flash memory apparatus can speed up accessing and increase efficiency.
    Type: Grant
    Filed: September 24, 2008
    Date of Patent: May 20, 2014
    Assignee: A-Data Technology Co., Ltd.
    Inventors: Ming-Dar Chen, Chuan-Sheng Lin, Hsiang-An Hsieh
  • Publication number: 20140129775
    Abstract: Disclosed is a method of pre-fetching NFA instructions to an NFA cell array. The method and system fetch instructions for use in an L1 cache during NFA instruction execution. Successive instructions from a current active state are fetched and loaded in the L1 cache. Disclosed is a system comprising an external memory, a cache line fetcher, and an L1 cache where the L1 cache is accessible and searchable by an NFA cell array and where successive instructions from a current active state in the NFA are fetched from external memory in an atomic cache line manner into a plurality of banks in the L1 cache.
    Type: Application
    Filed: November 6, 2012
    Publication date: May 8, 2014
    Applicant: LSI CORPORATION
    Inventor: Michael Ruehle
  • Patent number: 8661194
    Abstract: A cache control method for a hybrid hard disk drive (HDD) comprising a nonvolatile cache (NVC) and a hard disk. When the hybrid HDD is operating in a non-parallel mode of operation, the control method sequentially searches the NVC and then reads the hard disk for requested data, but when the hybrid HDD is operating in a parallel mode of operation, the control method simultaneously searches the NVC and reads hard disk for the data requested.
    Type: Grant
    Filed: February 5, 2007
    Date of Patent: February 25, 2014
    Assignee: Seagate Technology LLC
    Inventor: Hye-jeong Nam
  • Patent number: 8533389
    Abstract: A method and system for controlling multiple client access to cache memory and a single CAM device. Each client has a corresponding integrity controller in communication with the CAM device and all the other integrity controllers associated with the other clients in the system. Each integrity controller monitors states of the other clients, and inhibits its respective client from executing any operation when a common lookup index is detected during a co-pending operation with a first client. Once the operations of the first client are completed, its integrity controller signals the integrity controller of other clients to exit their inhibit or hold states, thereby allowing the other clients to resume their operations. Another advantage of the integrity controller is that its algorithms also prevents multiple host memory fetches of the same key, thereby saving time and improving system performance.
    Type: Grant
    Filed: September 15, 2009
    Date of Patent: September 10, 2013
    Assignee: PMC-Sierra, Inc.
    Inventors: Gregg Goyins, Jonathan Bradley Sadowsky
  • Patent number: 8504796
    Abstract: A system and method for providing high-speed memory operations is disclosed. The technique uses virtualization of memory space to map a virtual address space to a larger physical address space wherein no memory bank conflicts will occur. The larger physical address space is used to prevent memory bank conflicts from occurring by moving the virtualized memory addresses of data being written to memory to a different location in physical memory that will eliminate a memory bank conflict. A changeable mapping table that maps the virtualized memory addresses to physical memory addresses is stored in the same memory system.
    Type: Grant
    Filed: August 8, 2012
    Date of Patent: August 6, 2013
    Assignee: Memoir Systems, Inc.
    Inventors: Sundar Iyer, Shang-Tse Chuang
  • Publication number: 20130145103
    Abstract: In a database system having a plurality of concurrently executing session processes, the method commences by establishing a master list of sequences, the master list comprising a plurality of sequence objects which in turn define a sequence of values used for numbering and other identification within the database system. To reduce sequence cache latch access contention, multiple tiers of latches are provided. Methods of the system provide a first tier having a first tier “global” latch to serialize access to the master list such that at any point in time, only one of the concurrently executing session processes is granted access to the master list, from which master list are allocated sequences on demand. A second tier of latches is provided, the second tier having multiple second tier latches to serialize access to corresponding allocated sequences of values such that at any point in time, only one of the concurrently executing session processes is granted access to the allocated sequence.
    Type: Application
    Filed: December 1, 2011
    Publication date: June 6, 2013
    Applicant: Oracle International Corporation
    Inventors: Fulu LI, Vineet MARWAH, Amit GANESH
  • Patent number: 8447931
    Abstract: One embodiment of the present invention provides a processor that supports multiple-issue execution. This processor includes a register file, which contains an array of memory cells, wherein the memory cells contain bits for architectural registers of the processor. The register file also includes multiple read ports and multiple write ports to support multiple-issue execution. During operation, if multiple read ports simultaneously read from a given register, the register file is configured to: read each bit of the given register out of the array of memory cells through a single bitline associated with the bit; and to use a driver located outside of the array of memory cells to drive the bit to the multiple read ports. In this way, each memory cell only has to drive a single bitline (instead of multiple bitlines) during a multiple-port read operation, thereby allowing memory cells to use smaller and more power-efficient drivers for read operations.
    Type: Grant
    Filed: July 1, 2005
    Date of Patent: May 21, 2013
    Assignee: Oracle America, Inc.
    Inventors: Shailender Chaudhry, Paul Caprioli, Marc Tremblay
  • Patent number: 8307173
    Abstract: A technique for reducing reader overhead when referencing a shared data element while facilitating realtime-safe detection of a grace period for deferring destruction of the shared data element. The grace period is determined by a condition in which all readers that are capable of referencing the shared data element have reached a quiescent state subsequent to a request for a quiescent state. Common case local quiescent state tracking may be performed using only local per-reader state information for all readers that have not blocked while in a read-side critical section in which the data element is referenced. Uncommon case non-local quiescent state tracking may be performed using non-local multi-reader state information for all readers that have blocked while in their read-side critical section. The common case local quiescent state tracking requires less processing overhead than the uncommon case non-local quiescent state tracking.
    Type: Grant
    Filed: February 13, 2012
    Date of Patent: November 6, 2012
    Assignee: International Business Machines Corporation
    Inventor: Paul E. McKenney
  • Publication number: 20120233398
    Abstract: The present invention comprises a CHA 110 which transmits/receives data to/from an external device, a DKA 140 which transmits/receives data to/from an HDD unit 200, a primary cache unit 120 which has a primary cache memory 124, a secondary cache unit 130 which is installed between the primary cache unit 120 and the DKA 140 and has a secondary cache memory 134, a CCP 121 which stores write target data received by the CHA 110 in the primary cache memory 124, and a CCP 131 which stores the write target data in the secondary cache memory 134, and transfers the write target data stored in the secondary cache memory 134 to the DKA 140.
    Type: Application
    Filed: May 24, 2012
    Publication date: September 13, 2012
    Applicant: HITACHI, LTD.
    Inventors: Tatsuya NINOMIYA, Kazuo TANAKA
  • Patent number: 8266408
    Abstract: A system and method for providing high-speed memory operations is disclosed. The technique uses virtualization of memory space to map a virtual address space to a larger physical address space wherein no memory bank conflicts will occur. The larger physical address space is used to prevent memory bank conflicts from occurring by moving the virtualized memory addresses of data being written to memory to a different location in physical memory that will eliminate a memory bank conflict.
    Type: Grant
    Filed: December 15, 2009
    Date of Patent: September 11, 2012
    Assignee: Memoir Systems, Inc.
    Inventors: Sundar Iyer, Shang-Tse Chuang
  • Patent number: 8214596
    Abstract: In some embodiments, a non-volatile cache memory may include a segmented non-volatile cache memory configured to be located between a system memory and a mass storage device of an electronic system and a controller coupled to the segmented non-volatile cache memory, wherein the controller is configured to control utilization of the segmented non-volatile cache memory. The segmented non-volatile cache memory may include a file cache segment, the file cache segment to store complete files in accordance with a file cache policy, and a block cache segment, the block cache segment to store one or more blocks of one or more files in accordance with a block cache policy, wherein the block cache policy is different from the file cache policy. The controller may be configured to utilize the file cache segment in accordance with information related to the block cache segment and to utilize the block cache segment in accordance with information related to the file cache segment.
    Type: Grant
    Filed: September 30, 2008
    Date of Patent: July 3, 2012
    Assignee: Intel Corporation
    Inventors: Dale Juenemann, R. Scott Tetrick, Oscar Pinto
  • Publication number: 20120159077
    Abstract: A method and apparatus to reduce unnecessary write backs of cached data to a main memory and to optimize the usage of a cache memory tag directory. In one embodiment of the invention, the power consumption of a processor can be saved by eliminating write backs of cache memory lines that has information that has reached its end-of-life. In one embodiment of the invention, when a processing unit is required to clear one or more cache memory lines, it uses a write-zero command to clear the one or more cache memory lines. The processing unit does not perform a write operation to move or pass data values of zero to the one or more cache memory lines. By doing so, it reduces the power consumption of the processing unit.
    Type: Application
    Filed: December 21, 2010
    Publication date: June 21, 2012
    Inventors: SIMON C. STEELY, JR., Joel S. Emer, William C. Hasenplaugh
  • Publication number: 20120144129
    Abstract: A technique for reducing reader overhead when referencing a shared data element while facilitating realtime-safe detection of a grace period for deferring destruction of the shared data element. The grace period is determined by a condition in which all readers that are capable of referencing the shared data element have reached a quiescent state subsequent to a request for a quiescent state. Common case local quiescent state tracking may be performed using only local per-reader state information for all readers that have not blocked while in a read-side critical section in which the data element is referenced. Uncommon case non-local quiescent state tracking may be performed using non-local multi-reader state information for all readers that have blocked while in their read-side critical section. The common case local quiescent state tracking requires less processing overhead than the uncommon case non-local quiescent state tracking.
    Type: Application
    Filed: February 13, 2012
    Publication date: June 7, 2012
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventor: Paul E. McKenney
  • Patent number: 8176253
    Abstract: Various technologies and techniques are disclosed for using transactional memory hardware to accelerate virtualization or emulation. A central processing unit is provided with the transactional memory hardware. Code backpatching can be facilitated by providing transactional memory hardware that supports a facility to maintain private memory state and an atomic commit feature. Changes made to certain code are stored in the private state facility. Backpatching changes are enacted by attempting to commit all the changes to memory at once using the atomic commit feature. An efficient call return stack can be provided by using transactional memory hardware. A call return cache stored in the private state facility captures a host address to return to after execution of a guest function completes. A direct-lookup hardware-based hash table is used for the call return cache.
    Type: Grant
    Filed: June 27, 2007
    Date of Patent: May 8, 2012
    Assignee: Microsoft Corporation
    Inventors: Martin Taillefer, Darek Mihocka, Bruno Silva
  • Patent number: 8145869
    Abstract: A single data bus to a memory device can be split up into a number of data bus portions, each of which is managed by a different respective controller chip of multiple controller chips. During a memory access to a respective memory device, each of the multiple controller chips controls a different corresponding portion of the data bus to retrieve data from or store data to the memory device depending on whether the access is a read or write. To perform the data access, a synchronizer circuit (internal and/or external to the memory controller chips) synchronizes the multiple memory controller chips such that one of the memory controller chips drives the address bus and/or control signals to the memory device. After setting the address to the memory device, the memory controller chips either read data from or write data to the memory device based on the address.
    Type: Grant
    Filed: January 12, 2007
    Date of Patent: March 27, 2012
    Assignee: Broadbus Technologies, Inc.
    Inventors: Matthew G. Sargeant, Michael A. Kahn, Francis J. Stifter, Jr., Jason P. Colangelo
  • Patent number: 8078820
    Abstract: A method, and corresponding system and software, is described for writing data to a plurality of queues, each portion of the data being written to a corresponding one of the queues. The method includes, without requiring concurrent locking of more than one queue, determining if a space is available in each queue for writing a corresponding portion of the data, and if available, reserving the spaces in the queues. The method includes writing each portion of the data to a corresponding one of the queues.
    Type: Grant
    Filed: December 6, 2010
    Date of Patent: December 13, 2011
    Assignee: Ab Initio Technology LLC
    Inventors: Spiro Michaylov, Sanjeev Banerji, Craig W. Stanfill
  • Patent number: 7899999
    Abstract: Various technologies and techniques are disclosed for detecting falsely doomed parent transactions of nested children in transactional memory systems. When rolling back nested transactions, a release count is tracked each time that a write lock is released due to rollback for a given nested transaction. For example, a write abort compensation map can be used to track the release count for each nested transaction. The number of times the nested transactions releases a write lock is recorded in their respective write abort compensation map. The release counts can be used during a validation of a parent transaction to determine if a failed optimistic read is really valid. If an aggregated release count for the nested children transactions accounts for the difference in version numbers exactly, then the optimistic read is valid.
    Type: Grant
    Filed: June 27, 2007
    Date of Patent: March 1, 2011
    Assignee: Microsoft Corporation
    Inventors: Michael M. Magruder, David Detlefs, John Joseph Duffy, Goetz Graefe, Vinod K. Grover
  • Patent number: 7890707
    Abstract: Various technologies and techniques are disclosed for implementing retrying transactions in a transactional memory system. The system allows a transaction to execute a retry operation. The system registers for waits on every read in a read set of the retrying transaction. The retrying transaction waits for notification that something in the read set has changed. A transaction knows if notification is required in one of two ways. If the transactional memory word contained a waiters bit during write lock acquisition, then during release the transactional memory word is looked up in an object waiters map, and waiting transactions are signaled. If a writing transaction finds a global count of waiting transactions to be greater than zero after releasing write locks, a transaction waiters map is used to determine which waiting transactions need to be signaled. In each case, the write lock is released using a normal store operation.
    Type: Grant
    Filed: June 27, 2007
    Date of Patent: February 15, 2011
    Assignee: Microsoft Corporation
    Inventors: Michael M. Magruder, David Detlefs, John Joseph Duffy, Goetz Graefe, Vinod K. Grover
  • Patent number: 7882306
    Abstract: A memory device capable of sequentially outputting multiple pages of cached data while mitigating any interruption typically caused by fetching and transferring operations. The memory device outputs cached data from a first page while data from a second page is fetched into sense amplifier circuitry. When the outputting of the first page reaches a predetermined transfer point, a portion of the fetched data from the second page is transferred into the cache at the same time the remainder of the cached first page is being output. The remainder of the second page is transferred into the cache after all of the data from the first page is output while the outputting of the first portion of the second page begins with little or no interruption.
    Type: Grant
    Filed: May 7, 2009
    Date of Patent: February 1, 2011
    Assignee: Micron Technology, Inc.
    Inventor: Aaron Yip
  • Patent number: 7865684
    Abstract: A method, and corresponding system and software, is described for writing data to a plurality of queues, each portion of the data being written to a corresponding one of the queues. The method includes, without requiring concurrent locking of more than one queue, determining if a space is available in each queue for writing a corresponding portion of the data, and if available, reserving the spaces in the queues. The method includes writing each portion of the data to a corresponding one of the queues.
    Type: Grant
    Filed: June 27, 2005
    Date of Patent: January 4, 2011
    Assignee: Ab Initio Technology LLC
    Inventors: Spiro Michaylov, Sanjeev Banerji, Craig W. Stanfill
  • Patent number: 7856536
    Abstract: Provided are a method, system, and article of manufacture for providing a process exclusive access to a page including a memory address to which a lock is granted to the process. A request is received for a memory address in a memory device from a requesting process. A lock is granted to the requested memory address to the requesting process. The requesting process is provided exclusive access to a page including the requested memory address for a page access time period. The exclusive access to the page provided to the requesting process is released in response to an expiration of the page access time period.
    Type: Grant
    Filed: October 5, 2007
    Date of Patent: December 21, 2010
    Assignee: International Business Machines Corporation
    Inventors: Philippe Bergheaud, Dinesh Kumar Subhraveti, Marc Philippe Vertes
  • Publication number: 20100299477
    Abstract: According to one embodiment, a method for storing data on a magnetic tape comprises receiving data from two different hosts and simultaneously writing the data from the hosts to the magnetic tape using multiple transducers. In another approach, a method for storing data on a magnetic tape comprises receiving requests to establish a concurrent reservation from multiple hosts and allocating a unique stripe in a wrap to each of the hosts that sent the requests, wherein the wrap is a collection of data tracks to be written simultaneously in one direction of tape movement by multiple transducers of a tape head, and the wrap is logically divided into the stripes. Also, the method includes receiving data from the hosts and simultaneously writing the data from the hosts to the magnetic tape using the multiple transducers. Other systems and methods concerning storing data on magnetic tapes are described as well.
    Type: Application
    Filed: May 20, 2009
    Publication date: November 25, 2010
    Inventors: Shawn O. Brume, Fahnmusa C. Jangaba, Christine R. Knibloe, David L. Swanson
  • Publication number: 20100268890
    Abstract: An information handling system (IHS) includes a processor with a cache memory system. The processor includes a processor core with an L1 cache memory that couples to an L2 cache memory. The processor includes an arbitration mechanism that controls load and store requests to the L2 cache memory. The arbitration mechanism includes control logic that enables a load request to interrupt a store request that the L2 cache memory is currently servicing. The L2 cache memory includes dual data banks so that one bank may perform a load operation while the other bank performs a store operation. The cache system provides a single dispatch point into the data flow to the dual cache banks of the L2 cache memory.
    Type: Application
    Filed: April 15, 2009
    Publication date: October 21, 2010
    Applicants: INTERNATIONAL BUSINESS MACHINES CORPORATION, IBM Corporation
    Inventors: Sanjeev Ghai, Guy Lynn Guthrie, Hugh Shen, William John Starke
  • Patent number: 7752399
    Abstract: Disclosed is an information processing apparatus that has an update procedure semaphore, and a generation management information as management information of a shared data area that requires exclusion control. The generation management information specifies one item of generation information of the shared data area. As generation information provided for every generation, the apparatus has a reference-count measuring counter, a semaphore for updating generation information, a pointer for pointing to old generation information, and a pointer for pointing to the substance of the shared data area. In a case where the latest shared data is updated, a duplicate of the latest shared data area is created, new generation information corresponding to the duplicated shared data area is generated, data in the duplicated shared data area is updated and generation information, which corresponds to the shared data area after the updating thereof, is registered as the latest generation information.
    Type: Grant
    Filed: March 24, 2006
    Date of Patent: July 6, 2010
    Assignee: NEC Corporation
    Inventor: Hiroaki Oyama
  • Publication number: 20100161912
    Abstract: A mechanism for simultaneous multiple host access to shared centralized memory space via a virtualization protocol utilizing a network transport. The invention combines local memory interfacing with the handling of multiple hosts implementing virtualized memory-mapped I/O systems, such that the memory becomes a global resource. The end result is a widely distributed memory-mapped computer cluster, sharing a 2?64 byte memory space.
    Type: Application
    Filed: December 23, 2009
    Publication date: June 24, 2010
    Inventor: David A. Daniel
  • Publication number: 20100106913
    Abstract: According to an aspect of the embodiment, an FP includes a plurality of entries which holds requests to be processed, and each of the plurality of entries includes a requested flag indicating that data transfer is once requested. An FP-TOQ holds information indicating an entry holding the oldest request. A data transfer request prevention determination circuit checks the requested flag of a request to be processed and the FP-TOQ, and when a transfer request of data as a target of the request to be processed has already been issued and the entry holding the request to be processed is not the entry indicated by the FP-TOQ, transmits a signal which prevents the transfer request of the data to a data transfer request control circuit. Even when a cache miss occurs in a primary cache RAM, the data transfer request control circuit does not issue a data transfer request when the signal which prevents the transfer request is received.
    Type: Application
    Filed: December 16, 2009
    Publication date: April 29, 2010
    Applicant: FUJITSU LIMITED
    Inventor: Naohiro Kiyota
  • Publication number: 20090216948
    Abstract: A memory device capable of sequentially outputting multiple pages of cached data while mitigating any interruption typically caused by fetching and transferring operations. The memory device outputs cached data from a first page while data from a second page is fetched into sense amplifier circuitry. When the outputting of the first page reaches a predetermined transfer point, a portion of the fetched data from the second page is transferred into the cache at the same time the remainder of the cached first page is being output. The remainder of the second page is transferred into the cache after all of the data from the first page is output while the outputting of the first portion of the second page begins with little or no interruption.
    Type: Application
    Filed: May 7, 2009
    Publication date: August 27, 2009
    Inventor: Aaron Yip
  • Publication number: 20080201531
    Abstract: A design structure embodied in a machine readable storage medium for designing, manufacturing, and/or testing a design is provided. The design structure includes an apparatus for administering an access conflict in a cache. The apparatus includes the cache, a cache controller, and a superscalar computer processor. The cache controller is capable of receiving a write address and write data from the superscalar computer processor's store memory instruction execution unit and a read address for read data from the superscalar computer processor's load memory instruction execution unit, for writing and reading data from a same cache line in the cache simultaneously on a current clock cycle; storing the write data in the same cache line on the current clock cycle; stalling, in the load memory instruction execution unit, a corresponding load microinstruction; and reading from the cache on a subsequent clock cycle read data from the read address.
    Type: Application
    Filed: April 18, 2008
    Publication date: August 21, 2008
    Inventors: Marcus L. Kornegay, Ngan N. Pham
  • Publication number: 20080168230
    Abstract: Color-based caching allows each cache line to be distinguished by a specific color, and enables the manipulation of cache behavior based upon the colors of the cache lines. When multiple threads are able to share a cache, effective cache management is critical to overall performance. Color-based caching provides an effective method to better utilize a cache and avoid unnecessary cache thrashing and/or pollution. The color based caching can be monitored to improve memory performance and guarantee Quality-Of-Service of cache utilization.
    Type: Application
    Filed: January 5, 2007
    Publication date: July 10, 2008
    Inventors: Xiaowei Shen, David F. Bacon, Robert W. Wisniewski, Orran Krieger
  • Publication number: 20080120466
    Abstract: A method and system for accessing a single port multi-way cache includes an address multiplexer that simultaneously addresses a set of data and a set of program instructions in the multi-way cache. Duplicate output way multiplexers respectively select data and program instructions read from the cache responsive to the address multiplexer.
    Type: Application
    Filed: November 20, 2006
    Publication date: May 22, 2008
    Inventor: Klaus Oberlaender
  • Publication number: 20080077740
    Abstract: A cache memory logically associates a cache line with at least two cache sectors of a cache array wherein different sectors have different output latencies and, for a load hit, selectively enables the cache sectors based on their latency to output the cache line over successive clock cycles. Larger wires having a higher transmission speed are preferably used to output the cache line corresponding to the requested memory block. In the illustrative embodiment the cache is arranged with rows and columns of the cache sectors, and a given cache line is spread across sectors in different columns, with at least one portion of the given cache line being located in a first column having a first latency, and another portion of the given cache line being located in a second column having a second latency greater than the first latency.
    Type: Application
    Filed: November 29, 2007
    Publication date: March 27, 2008
    Inventors: Leo Clark, Guy Guthrie, Kirk Livingston, William Starke