Using Clearing, Invalidating, Or Resetting Means (epo) Patents (Class 711/E12.022)
  • Publication number: 20110113202
    Abstract: A processing node tracks probe activity level associated with its cache. The processing node and/or processing system further predicts an idle duration. If the probe activity level increases above a threshold probe activity level, and the idle duration prediction is above a threshold idle duration threshold, the processing node flushes its cache to prevent probes to the cache. If the probe activity level is above the threshold probe activity level but the predicted idle duration is too short, the performance state of the processing node is increased above its current performance state to provide enhanced performance capability in responding to the probe requests.
    Type: Application
    Filed: February 8, 2010
    Publication date: May 12, 2011
    Inventors: Alexander Branover, Maurice B. Steinman
  • Patent number: 7941608
    Abstract: A method and data processing apparatus comprise a cache having a plurality of data entries; an eviction buffer comprising an information portion and a data portion; and eviction logic to transfer information associated with a first of the plurality of data entries from the cache to the information portion and to determine, with reference to the information, whether the first of the plurality of data entries should be written to a memory by examining the information.
    Type: Grant
    Filed: March 17, 2009
    Date of Patent: May 10, 2011
    Assignee: ARM Limited
    Inventors: Florent Begon, Philippe Luc, Elodie Charra, Nicolas Chaussade
  • Patent number: 7937531
    Abstract: In one embodiment, a processor regularly writes one or more cache entries back to memory to reduce the likelihood of cache soft errors. The regularly occurring write backs operate independently of Least Recently Used (LRU) status of the entries so that all entries are flushed.
    Type: Grant
    Filed: February 1, 2007
    Date of Patent: May 3, 2011
    Assignee: Cisco Technology, Inc.
    Inventor: Somnath Mitra
  • Publication number: 20110099152
    Abstract: Described is caching classification-related metadata for a file in an alternate data stream of that file. When a file is classified (e.g., for data management), the classification properties are cached in association with the file, along with classification-related metadata that indicates the state of the file at the time of caching. The classification-related metadata in the alternate data stream is then useable in determining whether the classification properties are valid and up-to-date when next accessed, or whether the file needs to be reclassified. If the properties are valid and up-to-date, they may be used without requiring the computationally costly steps of reclassification. Also described is using more than one alternate data stream for the cache, and extending the classification-related metadata through a defined extension mechanism.
    Type: Application
    Filed: October 26, 2009
    Publication date: April 28, 2011
    Applicant: Microsoft Corporation
    Inventors: Clyde Law, Paul Adrian Oltean, Ran Kalach, Nir Ben-Zvi, Matthias H. Wollnik
  • Publication number: 20110093654
    Abstract: A data processing apparatus 1 comprises data processing circuitry 2, a memory 8 for storing data and a cache memory 5 for storing cached data from the memory 8. The cache memory 5 is partitioned into cache segments 12 which may be individually placed in a power saving state by power supply circuitry 15 under control of power control circuitry 22. The number of segments which are active at any time may be dynamically adjusted in dependence upon operating requirements of the processor 2. An eviction selection mechanism 35 is provided to select evictable cached data for eviction from the cache. A cache compacting mechanism 40 is provided to evict evictable cached data from the cache and to store non-evictable cached data in fewer cache segments than were used to store the cached data prior to eviction of the evictable cached data.
    Type: Application
    Filed: October 20, 2009
    Publication date: April 21, 2011
    Applicant: The regents of the University of Michigan
    Inventors: David Andrew Roberts, Trevor Nigel Mudge, Thomas Friedric Wenisch
  • Publication number: 20110093646
    Abstract: A method for processing a read request identifying an address. The method includes receiving, at a module including a flash memory and a memory buffer, the read request from a requesting processor, mapping, using a coherence directory controller within the module, the address to a cache line in a cache memory associated with a remote processor, and sending a coherency message from the module to the remote processor to change a state of the cache line in the cache memory. The method further includes receiving, at the module, the cache line from the remote processor, sending, using processor bus and in response to the read request, the cache line to the requesting processor, identifying a requested page stored within the flash memory based on the address, storing a copy of the requested page in the memory buffer, and writing the cache line to the copy of the requested page.
    Type: Application
    Filed: October 16, 2009
    Publication date: April 21, 2011
    Applicant: SUN MICROSYSTEMS, INC.
    Inventors: Pranay Koka, Michael Oliver McCracken, Herbert Dewitt Schwetman, JR., Jan Lodewijk Bonebakker
  • Publication number: 20110087845
    Abstract: The present disclosure generally relates to cache memory systems and/or techniques to identify dead cache blocks in cache memory systems. Example systems may include a cache memory that is accessible by a cache client. The cache memory may include a plurality of storage locations for a first cache block, with a most recently used position location in the cache memory. A cache controller may be configured to predict whether the first cache block stored in the cache memory is identified as a dead cache block based on a cache burst of the first cache block. The cache burst may comprise a first access of the first cache block by a cache client and any subsequent contiguous accesses of the first cache block following the first access by the cache client while the first cache block is in a most recently used position of the cache set.
    Type: Application
    Filed: October 14, 2009
    Publication date: April 14, 2011
    Inventors: Doug Burger, Haiming Liu
  • Publication number: 20110078381
    Abstract: A method for managing a parallel cache hierarchy in a processing unit. The method including receiving an instruction that includes a cache operations modifier that identifies a level of the parallel cache hierarchy in which to cache data associated with the instruction; and implementing a cache replacement policy based on the cache operations modifier.
    Type: Application
    Filed: September 24, 2010
    Publication date: March 31, 2011
    Inventors: Steven James HEINRICH, Alexander L. Minkin, Brett W. Coon, Rajeshwaran Selvanesan, Robert Steven Glanville, Charles McCarver, Anjana Rajendran, Stewart Glenn Carlton, John R. Nickolls, Brian Fahs
  • Publication number: 20110060881
    Abstract: A method and system to refresh a data entry in a cache before the data entry expires. The system includes a client computing system coupled to a server via a network connection. In response to a request for data access, the client computing system locates a data entry in a cache and determines whether the data entry in the cache has exceeded a refresh timeout since a last update of the data entry. If the data entry in the cache has exceeded the refresh timeout, the client computing system retrieves the data entry found in the cache in response to the request without waiting for the data entry to be refreshed, and requests a refresh of the data entry from the server via the network connection.
    Type: Application
    Filed: September 10, 2009
    Publication date: March 10, 2011
    Applicant: Red Hat, Inc.
    Inventor: Stephen J. Gallagher
  • Publication number: 20110055488
    Abstract: A processor includes multiple processor core units, each including a processor core and a cache memory. Victim lines evicted from a first processor core unit's cache may be stored in another processor core unit's cache, rather than written back to system memory. If the victim line is later requested by the first processor core unit, the victim line is retrieved from the other processor core unit's cache. The processor has low latency data transfers between processor core units. The processor transfers victim lines directly between processor core units' caches or utilizes a victim cache to temporarily store victim lines while searching for their destinations. The processor evaluates cache priority rules to determine whether victim lines are discarded, written back to system memory, or stored in other processor core units' caches. Cache priority rules can be based on cache coherency data, load balancing schemes, and architectural characteristics of the processor.
    Type: Application
    Filed: June 30, 2010
    Publication date: March 3, 2011
    Applicant: MIPS Technologies, Inc.
    Inventor: Sanjay Vishin
  • Publication number: 20110055485
    Abstract: An apparatus for allocating entries in a set associative cache memory includes an array that provides a first pseudo-least-recently-used (PLRU) vector in response to a first allocation request from a first functional unit. The first PLRU vector specifies a first entry from a set of the cache memory specified by the first allocation request. The first vector is a tree of bits comprising a plurality of levels. Toggling logic receives the first vector and toggles predetermined bits thereof to generate a second PLRU vector in response to a second allocation request from a second functional unit generated concurrently with the first allocation request and specifying the same set of the cache memory specified by the first allocation request. The second vector specifies a second entry different from the first entry from the same set. The predetermined bits comprise bits of a predetermined one of the levels of the tree.
    Type: Application
    Filed: July 6, 2010
    Publication date: March 3, 2011
    Applicant: VIA TECHNOLOGIES, INC.
    Inventors: Colin Eddy, Rodney E. Hooker
  • Patent number: 7899992
    Abstract: In the cache circuit, an instruction cache hit counter counts the number of cache hits, and an instruction memory access counter counts the number of times of instruction access. An instruction cache hit rate computation/entry disabling control circuit computes the ratio of the cache hit count to the instruction access count (cache hit rate). If the cache hit rate exceeds an instruction cache entry disabling threshold, an instruction cache control circuit disables contents of instruction cache memory.
    Type: Grant
    Filed: January 11, 2006
    Date of Patent: March 1, 2011
    Assignee: Panasonic Corporation
    Inventor: Nobuhiro Tsuboi
  • Publication number: 20110035554
    Abstract: A method and an apparatus for determining a usage level of a memory device to notify a running application to perform memory reduction operations selected based on the memory usage level are described. An application calls APIs (Application Programming Interface) integrated with the application codes in the system to perform memory reduction operations. A memory usage level is determined according to a memory usage status received from the kernel of a system. A running application is associated with application priorities ranking multiple running applications statically or dynamically. Selecting memory reduction operations and notifying a running application are based on application priorities. Alternatively, a running application may determine a mode of operation to directly reduce memory usage in response to a notification for reducing memory usage without using API calls to other software.
    Type: Application
    Filed: October 21, 2010
    Publication date: February 10, 2011
    Inventors: Matt Watson, James M. Magee
  • Publication number: 20110035553
    Abstract: Systems and methods for managing cached content are disclosed. More particularly, embodiments disclosed herein may allow cached content to be updated (e.g. regenerated or replaced) in response to a notification. Specifically, embodiments disclosed herein may process a notification pertaining to content stored in a cache. Processing the notification may include locating cached content associated with the notification. After the cached content which corresponds to the notification is found, an appropriate action may be taken. For example, the cached content may be flushed from the cache or a request may be regenerated. As a result of the action, new content is generated. This new content is then used to replace or update the cached content.
    Type: Application
    Filed: October 14, 2010
    Publication date: February 10, 2011
    Inventors: Lee Shepstone, Conleth S. O'Connell, JR., Mark R. Scheevel, Newton Isaac Rajkumar, Jamshid Afshar, JR., Puhong You, Brett J. Larsen, David Dean Caldwell
  • Publication number: 20110029737
    Abstract: In a method of synchronizing with a separated disk cache, the separated cache is configured to transfer cache data to a staging area of a storage device. An atomic commit operation is utilized to instruct the storage device to atomically commit the cache data to a mapping scheme of the storage device.
    Type: Application
    Filed: October 14, 2010
    Publication date: February 3, 2011
    Applicant: MICROSOFT CORPORATION
    Inventors: Ruston Panabaker, Cenk Ergan, Michael R. Fortin
  • Publication number: 20110016276
    Abstract: Aspects of the invention relate to improvements to the Least Recently Used (LRU) cache replacement method. Weighted LRU (WLRU) and Compact Weighted LRU (CWLRU) are CPU cache replacement methods that have superior hit rates to LRU replacement for programs with poor locality, such as network protocols and applications. WLRU assigns weights to cache lines and makes replacement decision by comparing weights. When a cache line is first brought into the cache, it is assigned an initial weight. Weights of cache lines in WLRU increase when hit and decrease when not hit. Weights in WLRU also have upper limits, and the weight of a cache line never increases beyond the upper limit. CWLRU is a more space-efficient implementation of WLRU. Compared to WLRU, CWLRU uses fewer bits per cache line to store the weight.
    Type: Application
    Filed: July 21, 2010
    Publication date: January 20, 2011
    Inventor: Qufei Wang
  • Patent number: 7873791
    Abstract: A cache management system and method monitors and controls the contents of cache memory coupled to at least one host and at least one data storage device where data-units are prefetched and stored into a cache memory in a data-set format, comprising the data-units and meta-data associated with the data units. The meta-data contain time stamp information encoded with information whether the cached prefetched information is available for reuse by additional prefetched data. Prefetched data-units are controlled by executing a first prefetch task to prefetch a first prefetch series of data-units from off-cache. A first prefetch operation is executed to prefetch and store a first selected set of data units. The prefetch task contains three dedicated pointers into cache memory.
    Type: Grant
    Filed: September 28, 2007
    Date of Patent: January 18, 2011
    Assignee: EMC Corporation
    Inventor: Rong Yu
  • Publication number: 20110010502
    Abstract: In an embodiment, a cache stores tags for cache blocks stored in the cache. Each tag may include an indication identifying which of two or more replacement policies supported by the cache is in use for the corresponding cache block, and a replacement record indicating the status of the corresponding cache block in the replacement policy. Requests may include a replacement attribute that identifies the desired replacement policy for the cache block accessed by the request. If the request is a miss in the cache, a cache block storage location may be allocated to store the corresponding cache block. The tag associated with the cache block storage location may be updated to include the indication of the desired replacement policy, and the cache may manage the block in accordance with the policy. For example, in an embodiment, the cache may support both an LRR and an LRU policy.
    Type: Application
    Filed: July 10, 2009
    Publication date: January 13, 2011
    Inventors: James Wang, Zongjian Chen, James B. Keller, Timothy J. Millet
  • Publication number: 20110010505
    Abstract: A resource management cache of a computing device receives a request for an item. The item may include any type of content, such as an image or a video. A rendition for the item is determined. The item may be stored in a plurality of renditions for retrieval. The resource management cache can send one or more requests to one or more sources for the rendition. The sources may include remote sources and also a local source. If a source responds with an indication the rendition is available, the rendition is sent to and received at the computing device. If no sources respond with an indication the rendition is available, the resource management cache may send a message asking if a source can generate the rendition from another rendition of the item. The rendition may be generated and it is sent to and received at the resource management cache.
    Type: Application
    Filed: July 13, 2009
    Publication date: January 13, 2011
    Applicants: Sony Corporation, Sony Electronics Inc
    Inventors: Steve Burnap, Joshua Shagam, Stephan Szabo, Nicholas Trotta
  • Publication number: 20110004730
    Abstract: A cache memory device that connects an instruction controlling unit outputting a memory access request for requesting data and a storage device storing data, the cache memory device including: a data memory unit that holds data for each cache line, a tag memory unit that holds, for each cache line linked with a cache line of the data memory unit, tag addresses specifying storage positions of data covered by the memory access request at the storage device and status data indicating states of the data of the data memory unit corresponding to the tag addresses, a search unit that searches for a cache line of the tag memory unit corresponding to an index address included in the memory access request, a comparison unit that compares a tag address held in the found cache line of the tag memory unit and a tag address included in the memory access request and, when the two do not match, detects a “cache miss” and reads out the status information of the found cache line, and a controlling unit that, when the compariso
    Type: Application
    Filed: June 30, 2010
    Publication date: January 6, 2011
    Applicant: FUJITSU LIMITED
    Inventor: Hiroaki Kimura
  • Publication number: 20100332686
    Abstract: Systems and methods of processing write transactions provide for combining write transactions on an input/output (I/O) hub according to a protocol between the I/O hub and a processor. Data associated with the write transactions can be flushed to an I/O device without the need for proprietary software and specialized registers within the I/O device.
    Type: Application
    Filed: January 19, 2010
    Publication date: December 30, 2010
    Inventors: Kenneth C. Creta, Aaron T. Spink, Lance E. Hacking, Sridhar Muthrasanallur, Jasmin Ajanovic
  • Publication number: 20100332716
    Abstract: Storing metadata that is disjoint from corresponding data by storing the metadata to the same address as the corresponding data but in a different address space. A metadata store instruction includes a storage address for the metadata. The storage address is the same address as that for data corresponding to the metadata, but the storage address when used for the metadata is implemented in a metadata address space while the storage address, when used for the corresponding data is implemented in a different data address space. As a result of executing the metadata store instruction, the metadata is stored at the storage address. A metadata load instruction includes the storage address for the metadata. As a result of executing the metadata load instruction, the metadata stored at the address is received. Some embodiments may further implement a metadata clear instruction which clears any entries in the metadata address space.
    Type: Application
    Filed: June 26, 2009
    Publication date: December 30, 2010
    Applicant: MICROSOFT CORPORATION
    Inventors: Gad Sheaffer, David Callahan, Jan Gray, Ali-Reza Adl-Tabatabai, Shlomo Raikin
  • Publication number: 20100325363
    Abstract: A method and system for hierarchical caching of objects of an ERP system is provided. A caching system comprises a server cache component that is executed by a server and a client cache component that is executed by each client of the server. The server cache component maintains a server cache at the server, and the client cache component maintains a client cache at each client. The client cache components also cache the objects in local client caches. Upon opening an object, the client cache component checks its local client cache to determine whether the object is cached. If so, then the client cache component need not retrieve the object from the server. Thus, the caching system is hierarchical in that each server and client maintains its own cache.
    Type: Application
    Filed: June 22, 2009
    Publication date: December 23, 2010
    Applicant: Microsoft Corporation
    Inventors: Anders Olesen, Mehmet K. Kiziltunc, Thomas Hejlsberg, Kim C. Olsen, Dean McCrae
  • Publication number: 20100325362
    Abstract: Systems and methods are provided for providing users at remote access devices with conditional access to server-based applications. Requests for access to server-based applications (e.g., requests to launch or obtain data associated with the server-based applications) by remote access devices may be prevented or allowed based on device compliance with one or more policies including whether data-retention prevention code can be downloaded to and operational on the remote access devices. The data-retention prevention code may be used to both determine whether data can be automatically deleted from a cache or file directory at the remote access device and to delete potentially retention-sensitive data once the data is downloaded to the remote access device from the server-based application.
    Type: Application
    Filed: August 31, 2010
    Publication date: December 23, 2010
    Applicant: Microsoft Corporation
    Inventor: Lior Kohavi
  • Publication number: 20100325365
    Abstract: An improved sectored cache replacement algorithm is implemented via a method and computer program product. The method and computer program product select a cache sector among a plurality of cache sectors for replacement in a computer system. The method may comprise selecting a cache sector to be replaced that is not the most recently used and that has the least amount of modified data. In the case in which there is a tie among cache sectors, the sector to be replaced may be the sector among such cache sectors with the least amount of valid data. In the case in which there is still a tie among cache sectors, the sector to be replaced may be randomly selected among such cache sectors. Unlike conventional sectored cache replacement algorithms, the improved algorithm implemented by the method and computer program product accounts for both hit rate and bus utilization.
    Type: Application
    Filed: June 17, 2009
    Publication date: December 23, 2010
    Applicant: International Business Machines Corporation
    Inventor: Daniel J. Colglazier
  • Publication number: 20100318742
    Abstract: In a particular embodiment, a circuit device includes a translation look-aside buffer (TLB) configured to receive a virtual address and to translate the virtual address to a physical address of a cache having at least two partitions. The circuit device also includes a control logic circuit adapted to identify a partition replacement policy associated with the identified one of the at least two partitions based on a partition indicator. The control logic circuit controls replacement of data within the cache according to the identified partition replacement policy in response to a cache miss event.
    Type: Application
    Filed: June 11, 2009
    Publication date: December 16, 2010
    Applicant: QUALCOMM INCORPORATED
    Inventors: Erich James Plondke, Lucian Codrescu, Ajay Ingle
  • Publication number: 20100318744
    Abstract: A method for allocating space in a cache based on media I/O speed is disclosed herein. In certain embodiments, such a method may include storing, in a read cache, cache entries associated with faster-responding storage devices and cache entries associated with slower-responding storage devices. The method may further include implementing an eviction policy in the read cache. This eviction policy may include demoting, from the read cache, the cache entries of faster-responding storage devices faster than the cache entries of slower-responding storage devices, all other variables being equal. In certain embodiments, the eviction policy may further include demoting, from the read cache, cache entries having a lower read-hit ratio faster than cache entries having a higher read-hit ratio, all other variables being equal. A corresponding computer program product and apparatus are also disclosed and claimed herein.
    Type: Application
    Filed: June 15, 2009
    Publication date: December 16, 2010
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Michael T. Benhase, Lawrence Y. Chiu, Lokesh M. Gupta, Yu-Cheng Hsu
  • Publication number: 20100309915
    Abstract: Various exemplary embodiments relate to a provider edge node and a method performed on the node including one or more of the following: storing, in a cache on the node, an association between an Internet Protocol (IP) address of the customer edge device and a Media Access Control (MAC) address of the customer edge device; storing, in a backup memory in the node, a copy of the last-known IP address of the customer edge device; removing the stored association from the cache in response to a cache-clearing event, while maintaining the copy of the last-known IP address in the backup memory; sending an address discovery message, the address discovery message requesting an updated MAC address corresponding to the last-known IP address maintained in the backup memory; and updating the cache to reflect the updated MAC address upon receipt of an additional address update message from the customer edge device.
    Type: Application
    Filed: June 8, 2009
    Publication date: December 9, 2010
    Applicants: Alcatel-Lucent Canada Inc., Alcatel-Lucent USA Inc.
    Inventors: Shafiq Pirbhai, Ron Haberman, Mustapha Aissaoui
  • Publication number: 20100312970
    Abstract: The illustrative embodiments provide a method, apparatus, and computer program product for managing a number of cache lines in a cache. In one illustrative embodiment, it is determined whether activity on a memory bus in communication with the cache exceeds a threshold activity level. A least important cache line is located in the cache responsive to a determination that the threshold activity level is exceeded, wherein the least important cache line is located using a cache replacement scheme. It is determined whether the least important cache line is clean responsive to the determination that the threshold activity level is exceeded. The least important cache line is selected for replacement in the cache responsive to a determination that the least important cache line is clean.
    Type: Application
    Filed: June 4, 2009
    Publication date: December 9, 2010
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Gordon B. Bell, Anil Krishna, Brian M. Rogers, Ken V. Vu
  • Publication number: 20100312971
    Abstract: In one embodiment, an integrated circuit includes a processor, an internal memory, and a memory controller coupled to an external memory. The integrated circuit may support two or more modes of operation, with different operating points. To switch from one operating point to another, code executed by the processor may copy switch code from the external memory into the internal memory, and may jump to the switch code. Executing out of the internal memory, the switch code may communicate with the memory controller to cause the external memory to enter into self-refresh mode. The operating point may be altered, and the switch code may reinitialize the memory controller after the integrated circuit has stabilized at the new operating point. After the memory controller's physical interface circuit has relocked, the external memory may exit self-refresh mode.
    Type: Application
    Filed: June 5, 2009
    Publication date: December 9, 2010
    Inventors: Maziar H. Moallem, Sanjay Mansingh, Richard F. Avra
  • Publication number: 20100293337
    Abstract: The disclosure is related to data storage systems having multiple cache and to management of cache activity in data storage systems having multiple cache. In a particular embodiment, a data storage device includes a volatile memory having a first read cache and a first write cache, a non-volatile memory having a second read cache and a second write cache and a controller coupled to the volatile memory and the non-volatile memory. The memory can be configured to selectively transfer read data from the first read cache to the second read cache based on a least recently used indicator of the read data and selectively transfer write data from the first write cache to the second write cache based on a least recently written indicator of the write data.
    Type: Application
    Filed: May 13, 2009
    Publication date: November 18, 2010
    Applicant: SEAGATE TECHNOLOGY LLC
    Inventors: Robert D. Murphy, Robert W. Dixon, Steven S. Williams
  • Publication number: 20100293335
    Abstract: A method for cache management in an environment based on Common Information Model is described. Cache elements in the cache are associated with a time attribute and historical data. Cache elements having a time attribute lying in a certain range are polled for from the server and updated at predetermined time points. A new time attribute is calculated for each cache element based on its historical data and this new time attribute assists in adapting the polling frequency for the cache element to its importance and change characteristics. Asynchronous notifications from the server preempt the polling based on the time attribute for a cache element and instead, polling for the cache element is based on the asynchronous notification. A system for cache management includes a client and a server, the client having a cache that is managed based on each cache element's importance and change characteristics.
    Type: Application
    Filed: May 13, 2009
    Publication date: November 18, 2010
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Muthu Annamalai Muthiah, Jayesh V. Rane, Sanket S. Sangwikar
  • Publication number: 20100293420
    Abstract: System and method for using flash memory in a memory hierarchy. A computer system includes a processor coupled to a memory hierarchy via a memory controller. The memory hierarchy includes a cache memory, a first memory region of random access memory coupled to the memory controller via a first buffer, and an auxiliary memory region of flash memory coupled to the memory controller via a flash controller. The first buffer and the flash controller are coupled to the memory controller via a single interface. The memory controller receives a request to access a particular page in the first memory region. The processor detects a page fault corresponding to the request and in response, invalidates cache lines in the cache memory that correspond to the particular page, flushes the invalid cache lines, and swaps a page from the auxiliary memory region to the first memory region.
    Type: Application
    Filed: May 15, 2009
    Publication date: November 18, 2010
    Inventors: Sanjiv Kapil, Ricky C. Hetherington
  • Publication number: 20100293338
    Abstract: A low priority queue can be configured to list low priority removal candidates to be removed from a cache, with the low priority removal candidates being sorted in an order of priority for removal. A high priority queue can be configured to list high priority removal candidates to be removed from the cache. In response to receiving a request for one or more candidates for removal from the cache, one or more high priority removal candidates from the high priority queue can be returned if the high priority queue lists any high priority removal candidates. If no more high priority removal candidates remain in the high priority queue, then one or more low priority removal candidates from the low priority queue can be returned in the order of priority for removal. Write-only latches can also be used during write operations in a cache lookup data structure.
    Type: Application
    Filed: May 20, 2009
    Publication date: November 18, 2010
    Applicant: Microsoft Corporation
    Inventors: Muralidhar Krishnaprasad, Sudhir Mohan Jorwekar, Sharique Muhammed, Subramanian Muralidhar, Anil K. Nori
  • Publication number: 20100287333
    Abstract: A data storage device comprises a plurality of memory devices, a buffer memory, and a controller. The plurality of memory devices are connected to a plurality of channels and a plurality of ways. The buffer memory temporarily stores data to be written in the memory devices. The controller stores the data in the buffer memory based on channel and way information of the memory devices.
    Type: Application
    Filed: May 3, 2010
    Publication date: November 11, 2010
    Applicant: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Jin-Hyuk LEE, Jang hwan KIM, Han-Chan JO, Yeong-Jae WOO, Dong hyun SONG
  • Publication number: 20100287336
    Abstract: In an LSI that determines timing of DRAM refresh by a refresh timer to synchronize an external I/O signal and DRAM refresh timing with each other, a circuit configuration capable of controlling a value of the refresh timer by a CPU at arbitrary timing is employed. Alternatively, a circuit configuration capable of controlling the value of the refresh timer at arbitrary timing by an external terminal, or a circuit configuration capable of controlling the refresh timing directly from the external terminal without through the refresh timer.
    Type: Application
    Filed: July 26, 2010
    Publication date: November 11, 2010
    Applicant: PANASONIC CORPORATION
    Inventors: Shinsuke TANAKA, Daigo SENOO
  • Publication number: 20100281218
    Abstract: A method for replacing cached data is disclosed. The method in one aspect associates an importance value to each block of data in the cache. When a new entry needs to be stored in the cache, a cache block for replacing is selected based on the importance values associated with cache blocks. In another aspect, the importance values are set according to the hardware and/or software's knowledge of the memory access patterns. The method in one aspect may also include varying the importance value over time over different processing requirements.
    Type: Application
    Filed: July 11, 2006
    Publication date: November 4, 2010
    Applicant: International Business Machines Corporation
    Inventors: Xiaowei Shen, Balaram Sinharoy, Robert W. Wisniewski
  • Publication number: 20100281222
    Abstract: A cache system and a method for controlling the cache system are provided. The cache system includes a plurality of caches, a buffer module, and a migration selector. Each of the caches is accessed by a corresponding processor. Each of the caches includes a plurality of cache sets and each of the cache sets includes a plurality of cache lines. The buffer module is coupled to the caches for receiving and storing data evicted due to conflict miss from a source cache line of a source cache set of a source cache among the caches. The migration selector is coupled to the caches and the buffer module. The migration selector selects, from all the cache sets, a destination cache set of a destination cache among the caches according to a predetermined condition and causing the evicted data to be sent from the buffer module to the destination cache set.
    Type: Application
    Filed: April 29, 2009
    Publication date: November 4, 2010
    Applicant: FARADAY TECHNOLOGY CORP.
    Inventors: Kuang-Chih Liu, Luen-Ming Shen
  • Publication number: 20100281204
    Abstract: A memory system includes a WC 21 from which data is read out and to which data is written in sector units by a host apparatus, an FS 12 from which data is read out and to which data is written in page units, an MS 11 from which data is read out and to which data written in track units, an FSIB 12a functioning as an input buffer for the FS 12, and an MSIB 11a functioning as an input buffer to the MS 11. An FSBB 12ac that has a capacity equal to or larger than a storage capacity of the WC 21 and stores data written in the WC 21 is provided in the FSIB12a. A data managing unit 120 that manages the respective storing units suspends, when it is judged that one kind of processing performed among the storing units exceeds predetermined time, the processing judged as exceeding the predetermined time and controls the data written in the WC 21 to be saved in the FSBB 12ac.
    Type: Application
    Filed: September 22, 2008
    Publication date: November 4, 2010
    Applicant: Kabushiki Kaisha Toshiba
    Inventors: Junji Yano, Hidenori Matsuzaki, Kosuke Hatsuda
  • Publication number: 20100274768
    Abstract: Aspects of the subject matter described herein relate to multi-log based replication. In aspects, database fragments are associated with different logs. Each change to a duplicated database record may be recorded in multiple logs. A history data structure is used to determine when duplication schemas are valid. A duplication schema indicates what database fragments duplicate one or more database records. For a particular time range, the duplication schema history is used to determine whether currently available logs include all changes. If multiple logs include the same change, one log may be selected to provide the change. Non-duplicative changes may be placed into a single data stream usable to update a remote database.
    Type: Application
    Filed: April 23, 2009
    Publication date: October 28, 2010
    Applicant: Microsoft Corporation
    Inventors: Rui Wang, Qun Guo, Yixue Zhu, Michael E. Habben
  • Publication number: 20100274962
    Abstract: The present disclosure relates to methods, devices and computer-readable medium for implementing a caching policy and/or a cache flushing policy in a peripheral non-volatile storage device operatively coupled to a host device. In some embodiments, data is stored to a cache area of a non-volatile memory within the peripheral non-volatile storage device in accordance with a historical rate at which other data was received by the peripheral storage device from the host device and/or a historical average time interval between successive host write requests received and/or an assessed rate at which data is required to be written to the non-volatile memory and/or a detecting by the peripheral non-volatile memory device that the host has read the storage ready/busy flag. In some embodiments, data is copied from a cache storage area of the non-volatile memory to a main storage area in accordance with the historical rate and/or the historical average time interval.
    Type: Application
    Filed: April 26, 2009
    Publication date: October 28, 2010
    Applicant: SanDisk IL Ltd.
    Inventors: Amir MOSEK, Menahem LASSER, Mark MURIN
  • Publication number: 20100274971
    Abstract: Technologies are generally described herein for maintaining cache coherency within a multi-core processor. A first cache entry to be evicted from a first cache may be identified. The first cache entry may include a block of data and a first tag indicating an owned state. An owner eviction message for the first cache entry may be broadcasted from the first cache. A second cache entry in a second cache may be identified. The second cache entry may include the block of data and a second tag indicating a shared state. The broadcasted owner eviction message may be detected with the second cache. An ownership acceptance message for the second cache entry may be broadcasted from the second cache. The broadcasted ownership acceptance message may be detected with the first cache. The second tag in the second cache entry may be transformed from the shared state to the owned state.
    Type: Application
    Filed: April 23, 2009
    Publication date: October 28, 2010
    Inventor: Yan Solihin
  • Publication number: 20100274974
    Abstract: A system and method for replacing data in a cache utilizes cache block validity information, which contains information that indicates that data in a cache block is no longer needed for processing, to maintain least recently used information of cache blocks in a cache set of the cache, identifies the least recently used cache block of the cache set using the least recently used information of the cache blocks in the cache set, and replaces data in the least recently used cache block of the cache set with data from main memory.
    Type: Application
    Filed: April 24, 2009
    Publication date: October 28, 2010
    Applicant: NXP B.V.
    Inventors: JAN-WILLEM VAN DE WAERDT, JOHAN GERARD WILLEM MARIA JANSSEN, MAURICE PENNERS
  • Publication number: 20100268881
    Abstract: A method to associate a storage policy with a cache region is disclosed. In this method, a cache region associated with an application is created. The application runs on virtual machines, and where a first virtual machine has a local memory cache that is private to the first virtual machine. The first virtual machine additionally has a shared memory cache that is shared by the first virtual machine and a second virtual machine. Additionally, the cache region is associated with a storage policy. Here, the storage policy specifies that a first copy of an object to be stored in the cache region is to be stored in the local memory cache and that a second copy of the object to be stored in the cache region is to be stored in the shared memory cache.
    Type: Application
    Filed: July 7, 2010
    Publication date: October 21, 2010
    Inventors: Galin Galchev, Frank Kilian, Oliver Luik, Dirk Marwinski, Petio G. Petev
  • Publication number: 20100262817
    Abstract: A data storage device includes a computer-readable medium encoded with a computer program that, when executed communicates with a basic input/output system (BIOS), receives a user selection from the BIOS to wipe the data storage device and performs a wipe of the data storage device. In an embodiment, the wipe of the data storage device includes writing a series of 1s and/or 0s to substantially all data bits of the data storage device.
    Type: Application
    Filed: April 9, 2009
    Publication date: October 14, 2010
    Applicant: DELL PRODUCTS L.P.
    Inventors: Kurt Edmiston, Donald Gage, Ryan Holland
  • Publication number: 20100262771
    Abstract: According to one embodiment, a data storage system includes a controller which accesses a first storage device using a first module on startup and accesses the first storage device using a second module after the startup. The first module records, when the write-target data is written to the first storage device, trace information indicating the write command in a second storage device. The second module determines, when taking over a reception of a command instructing writing/reading of data from the first module, whether or not unupdated data to be updated as a result of a writing of the first module is cached in the second storage device based on the trace information, and invalidates a data block including the unupdated data when the unupdated data is cached.
    Type: Application
    Filed: April 1, 2010
    Publication date: October 14, 2010
    Inventor: Takehiko Kurashige
  • Publication number: 20100257317
    Abstract: A data processing system includes an interconnect fabric, a system memory coupled to the interconnect fabric and including a virtual barrier synchronization region allocated to storage of virtual barrier synchronization registers (VBSRs), and a plurality of processing units coupled to the interconnect fabric and operable to access the virtual barrier synchronization region of the system memory. Each of the plurality of processing units includes a processor core and a cache memory including a cache array that caches VBSR lines from the virtual barrier synchronization region of the system memory and a cache controller. The cache controller, responsive to a store request from the processor core to update a particular VBSR line, performs a non-blocking update of the cache array in each other of the plurality of processing units contemporaneously holding a copy of the particular VBSR line by transmitting a VBSR update command on the interconnect fabric.
    Type: Application
    Filed: April 7, 2009
    Publication date: October 7, 2010
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Ravi K. Arimilli, Guy L. Guthrie, Robert A. Cargnoni, William J. Starke, Derek E. Williams
  • Publication number: 20100257318
    Abstract: Executable computer code sections can be effectively evicted from secondary memory (e.g., instruction cache) during execution time in order to reduce the observable changes to the state of the secondary memory, thereby enhancing the security of computing systems that use secondary memory in addition the primary (main) memory to support execution of computer code. In particular, codes sections considered to be critical to security can be identified and effectively mapped to the same section of an instruction cache (I-cache) as provided in more modern computing systems in order to improve the efficiency of execution, thereby allowing use of the I-cache in a more secure manner.
    Type: Application
    Filed: April 3, 2009
    Publication date: October 7, 2010
    Applicant: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Onur Aciicmez, Xinwen Zhang, Jean-Pierre Seifert
  • Publication number: 20100257320
    Abstract: Techniques for replacing one or more blocks in a cache, the one or more blocks being associated with a plurality of data streams, are provided. The one or more blocks in the cache are grouped into one or more groups. Each group corresponding to one of the plurality of data streams. One or more incoming blocks are received. To free space, the one or more blocks of the one or more groups in the cache are invalidated in accordance with at least one of an inactivity of a given data stream corresponding to the one or more groups and a length of the one or more groups. The one or more incoming blocks are stored in the cache. A number of data streams maintained within the cache is maximized.
    Type: Application
    Filed: April 7, 2009
    Publication date: October 7, 2010
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Brian Bass, Giora Biran, Hubertus Franke, Amit Golander, Hao Yu
  • Publication number: 20100250857
    Abstract: A technique for managing a cache memory for temporarily retaining data read out from a main memory so as to be used by a processing section is disclosed. The cache memory is managed using a tag memory and utilized by a write-through method. The cache controlling apparatus includes a supervising section adapted to supervise accessing time to the cache memory, and a refreshing section adapted to read out data on one or more cache lines of the cache memory from the main memory again in response to a result of the supervision by the supervising section and retain the read out data into the cache memory.
    Type: Application
    Filed: February 22, 2010
    Publication date: September 30, 2010
    Applicant: FUJITSU LIMITED
    Inventor: Noriyuki MATSUI