Using Directory Methods (epo) Patents (Class 711/E12.027)
-
Patent number: 12164422Abstract: Disclosed in some examples are methods, systems, memory devices, memory controllers, and machine-readable mediums which provide for reserving physical memory device resources to specific execution units. Execution units may include processes, threads, virtual machines, functions, procedures, or the like. Physical memory device resources may include channels, modules, ranks, banks, bank groups, and the like. For example, a physical memory device resource that is reservable may be a smallest unit that allows for parallel access with another of the same size unit.Type: GrantFiled: November 4, 2022Date of Patent: December 10, 2024Assignee: Micron Technology, Inc.Inventor: Patrick Michael Sheridan
-
Patent number: 12061552Abstract: Example implementations relate to cache coherency protocols as applied to a memory block range. Exclusive ownership of a range of blocks of memory in a default shared state may be tracked by a directory. The directory may be associated with a first processor of a set of processors. When a request is received from a second processor of the set of processors to read one or more blocks of memory absent from the directory, one or more blocks may be transmitted in the default shared state to the second processor. The blocks absent from the directory may not be tracked in the directory.Type: GrantFiled: May 11, 2023Date of Patent: August 13, 2024Assignee: Hewlett Packard Enterprise Development LPInventors: Michael Malewicki, Thomas McGee, Michael S. Woodacre
-
Patent number: 12056251Abstract: A method for protecting a folder from unauthorized file modification may include receiving, from a remote device, a modify request for a target file in a folder and determining whether the folder is a protected folder. The method may also include determining, in response to determining the folder is the protected folder, whether the remote device is a trusted host. The method may further include allowing, in response to determining that the remote device is the trusted host, the modify request for the target file.Type: GrantFiled: March 18, 2020Date of Patent: August 6, 2024Assignee: Veritas Technologies LLCInventors: Narayan Subramanian, Arindam Panna, Srineet Sridharan
-
Patent number: 12019894Abstract: A method for providing a coresident copy of external container data, that includes identifying, via a cluster storage manager, an access request to a cluster storage for external container data, where the access request is sent from a container executing on a first host, making a determination that the external container data is not stored on the first host, and based on the determination, copying the external container data from a second host to the first host to provide the coresident copy of the external container data.Type: GrantFiled: August 10, 2022Date of Patent: June 25, 2024Assignee: DELL PRODUCTS L.P.Inventors: Alan Barnett, David Bowden, Merry Globin
-
Patent number: 11892949Abstract: A method and a system detects a cache line as a potential or confirmed hot cache line based on receiving an intervention of a processor associated with a fetch of the cache line. The method and system include suppressing an action of operations associated with the hot cache line. A related method and system detect an intervention and, in response, communicates an intervention notification to another processor. An alternative method and system detect a hot data object associated with an intervention event of an application. The method and system can suppress actions of operations associated with the hot data object. An alternative method and system can detect and communicate an intervention associated with a data object.Type: GrantFiled: January 13, 2023Date of Patent: February 6, 2024Assignee: International Business Machines CorporationInventors: Christian Zoellin, Christian Jacobi, Chung-Lung K. Shum, Martin Recktenwald, Anthony Saporito, Aaron Tsai
-
Patent number: 11886284Abstract: A method for data redistribution of a job data in a first datanode (DN) to at least one additional DN in a Massively Parallel Processing (MPP) Database (DB) is provided. The method includes recording a snapshot of the job data, creating a first data portion in the first DN and a redistribution data portion in the first DN, collecting changes to a job data copy stored in a temporary table, and initiating transfer of the redistribution data portion to the at least one additional DN.Type: GrantFiled: May 17, 2022Date of Patent: January 30, 2024Assignee: Futurewei Technologies, Inc.Inventors: Le Cai, QingQing Zhou, Yang Sun
-
Patent number: 11880327Abstract: A coherent connection and a non-coherent connection are provided between system-on-chips (SoCs). The coherent connection can be coupled to coherent interconnects on the SoCs, and the non-coherent connection can be coupled to non-coherent interconnects on the SoCs. An input/output (I/O) transaction from an I/O device on a first SoC that is targeted to a second SoC can be transmitted via the non-coherent connection, and a processor transaction from the first SoC that is targeted to the second SoC can be transmitted via the coherent connection.Type: GrantFiled: December 7, 2021Date of Patent: January 23, 2024Assignee: Amazon Technologies, Inc.Inventors: Guy Nakibly, Barak Wasserstrom, Yaniv Shapira, Erez Izenberg, Adi Habusha
-
Patent number: 11853210Abstract: Provided are systems, methods, and apparatuses for providing a storage resource. The method can include: operating a first controller coupled to a network interface in accordance with a cache coherent protocol; performing at least one operation on data associated with a cache using a second controller coupled to the first controller and coupled to a first memory; and storing the data on a second memory coupled to one of the first controller or the second controller.Type: GrantFiled: April 30, 2021Date of Patent: December 26, 2023Assignee: Samsung Electronics Co., Ltd.Inventors: Krishna T. Malladi, Andrew Chang, Ehsan Najafabadi
-
Patent number: 11741012Abstract: A system includes a plurality of host processors and a plurality of hybrid memory cube (HMC) devices configured as a distributed shared memory for the host processors. An HMC device includes a plurality of integrated circuit memory die including at least a first memory die arranged on top of a second memory die, and at least a portion of the memory of the memory die is mapped to include at least a portion of a memory coherence directory; and a logic base die including at least one memory controller configured to manage three-dimensional (3D) access to memory of the plurality of memory die by at least one second device, and logic circuitry configured to implement a memory coherence protocol for data stored in the memory of the plurality of memory die.Type: GrantFiled: October 8, 2020Date of Patent: August 29, 2023Assignee: Micron Technology, Inc.Inventors: John Leidel, Richard C Murphy
-
Patent number: 11675706Abstract: A programmable switch includes at least one memory configured to store a cache directory for a distributed cache, and circuitry configured to receive a cache line request from a client device to obtain a cache line. The cache directory is updated based on the received cache line request, and the cache line request is sent to a memory device to obtain the requested cache line. An indication of the cache directory update is sent to a controller for the distributed cache to update a global cache directory. In one aspect, the controller sends at least one additional indication of the update to at least one other programmable switch to update at least one backup cache directory stored at the at least one other programmable switch.Type: GrantFiled: June 30, 2020Date of Patent: June 13, 2023Assignee: Western Digital Technologies, Inc.Inventors: Marjan Radi, Dejan Vucinic
-
Patent number: 11586564Abstract: A method of a buffer memory device, a storage system, and a buffer memory device are provided. The method of the buffer memory device, the buffer memory device having a lower tier memory and a higher tier memory, may include receiving a new entry request, determining that the new entry request includes an HOL entry, selecting an entry on the higher tier memory to be tiered down to the lower tier memory in response to determining that the new entry request includes an HOL entry, removing the selected entry from the higher tier memory, storing the HOL entry in the higher tier memory of the buffer memory device, and outputting the HOL entry to an arbiter.Type: GrantFiled: May 7, 2021Date of Patent: February 21, 2023Inventors: Chun-Chu Chen-Jhy Archie Wu, Fnu Vikram Singh, Syed Kaiser
-
Patent number: 11586542Abstract: A method and a system detects a cache line as a potential or confirmed hot cache line based on receiving an intervention of a processor associated with a fetch of the cache line. The method and system include suppressing an action of operations associated with the hot cache line. A related method and system detect an intervention and, in response, communicates an intervention notification to another processor. An alternative method and system detect a hot data object associated with an intervention event of an application. The method and system can suppress actions of operations associated with the hot data object. An alternative method and system can detect and communicate an intervention associated with a data object.Type: GrantFiled: April 9, 2021Date of Patent: February 21, 2023Assignee: International Business Machines CorporationInventors: Christian Zoellin, Christian Jacobi, Chung-Lung K. Shum, Martin Recktenwald, Anthony Saporito, Aaron Tsai
-
Patent number: 11360906Abstract: The devices within an inter-device processing system maintain data coherency in the last level caches of the system as a cache line of data is shared between the devices by utilizing a directory in one of the devices that tracks the coherency protocol states of the memory addresses in the last level caches of the system.Type: GrantFiled: August 14, 2020Date of Patent: June 14, 2022Assignee: Alibaba Group Holding LimitedInventors: Lide Duan, Hongyu Liu, Hongzhong Zheng, Yen-Kuang Chen
-
Patent number: 9009416Abstract: A method, computer program product, and computing system for reclassifying a first assigned cache portion associated with a first machine as a public cache portion associated with the first machine and at least one additional machine after the occurrence of a reclassifying event. The public cache portion includes a plurality of pieces of content received by the first machine. A content identifier for each of the plurality of pieces of content included within the public cache portion is compared with content identifiers for pieces of content included within a portion of a data array associated with the at least one additional machine to generate a list of matching data portions. The list of matching data portions is provided to at least one additional assigned cache portion within the cache system that is associated with the at least one additional machine.Type: GrantFiled: December 30, 2011Date of Patent: April 14, 2015Assignee: EMC CorporationInventors: Philip Derbeko, Anat Eyal, Roy E. Clark
-
Patent number: 8972664Abstract: Embodiments relate to accessing a cache line on a multi-level cache system having a system memory. Based on a request for exclusive ownership of a specific cache line at the local node, requests are concurrently sent to the system memory and remote nodes of the plurality of nodes for the specific cache line by the local node. The specific cache line is found in a specific remote node. The specific remote node is one of the remote nodes. The specific cache line is removed from the specific remote node for exclusive ownership by another node. Based on the specified node having the specified cache line in ghost state, any subsequent fetch request is initiated for the specific cache line from the specific node encounters the ghost state. When the ghost state is encountered, the subsequent fetch request is directed only to nodes of the plurality of nodes.Type: GrantFiled: March 11, 2013Date of Patent: March 3, 2015Assignee: International Business Machines CorporationInventors: Timothy C. Bronson, Garrett M. Drapala, Michael A. Blake, Craig R. Walters, Pak-Kin Mak
-
Patent number: 8918587Abstract: Embodiments relate to accessing a cache line on a multi-level cache system having a system memory. Based on a request for exclusive ownership of a specific cache line at the local node, requests are concurrently sent to the system memory and remote nodes of the plurality of nodes for the specific cache line by the local node. The specific cache line is found in a specific remote node. The specific remote node is one of the remote nodes. The specific cache line is removed from the specific remote node for exclusive ownership by another node. Based on the specified node having the specified cache line in ghost state, any subsequent fetch request is initiated for the specific cache line from the specific node encounters the ghost state. When the ghost state is encountered, the subsequent fetch request is directed only to nodes of the plurality of nodes.Type: GrantFiled: June 13, 2012Date of Patent: December 23, 2014Assignee: International Business Machines CorporationInventors: Timothy C. Bronson, Garrett M. Drapala, Michael A. Blake, Craig R. Walters, Pak-Kin Mak
-
Patent number: 8886885Abstract: Apparatus having corresponding methods and computer-readable media comprise: a plurality of flash modules, wherein each of the flash modules comprises a cache memory; a flash memory; and a flash controller in communication with the cache memory and the flash memory; wherein the flash controller of a first one of the flash modules is configured to operate the cache memories together as a global cache; wherein the flash controller of a second one of the flash modules is configured to operate a second one of the flash modules as a directory controller for the flash memories.Type: GrantFiled: November 5, 2010Date of Patent: November 11, 2014Assignee: Marvell World Trade Ltd.Inventors: Wei Zhou, Chee Hoe Chu, Po-Chien Chang
-
Patent number: 8874855Abstract: Techniques are generally described for methods, systems, data processing devices and computer readable media related to multi-core parallel processing directory-based cache coherence. Example systems may include one multi-core processor or multiple multi-core processors. An example multi-core processor includes a plurality of processor cores, each of the processor cores having a respective cache. The system may further include a main memory coupled to each multi-core processor. A directory descriptor cache may be associated with the plurality of the processor cores, where the directory descriptor cache may be configured to store a plurality of directory descriptors. Each of the directory descriptors may provide an indication of the cache sharing status of a respective cache-line-sized row of the main memory.Type: GrantFiled: December 28, 2009Date of Patent: October 28, 2014Assignee: Empire Technology Development LLCInventor: Tom Conte
-
Patent number: 8868837Abstract: In a multiprocessor system, with conflict checking implemented in a directory lookup of a shared cache memory, a reader set encoding permits dynamic recordation of read accesses. The reader set encoding includes an indication of a portion of a line read, for instance by indicating boundaries of read accesses. Different encodings may apply to different types of speculative execution.Type: GrantFiled: January 18, 2011Date of Patent: October 21, 2014Assignee: International Business Machines CorporationInventors: Alan Gara, Martin Ohmacht
-
Patent number: 8862857Abstract: A data access processing method and apparatus, the method comprising: copying a kernel code and a global descriptor table on a memory of each of nodes respectively (101); making base addresses of kernel code segments on the respective nodes in the global descriptor table respectively point to linear addresses of the kernel code corresponding to the respective nodes based on a virtual address of the kernel code (102); and recording a mapping relation between the linear addresses of the kernel code corresponding to the respective nodes and physical addresses of the respective nodes in a kernel page table respectively, to enable a process to access the kernel code locally in the respective nodes (103). The apparatus comprises a copying module (401), a modifying module (402) and a recording module (403).Type: GrantFiled: December 29, 2011Date of Patent: October 14, 2014Assignee: Huawei Technologies Co., Ltd.Inventors: Wei Wang, Xiaofeng Zhang
-
Patent number: 8812786Abstract: A system and method of providing directory cache coherence are disclosed. The system and method may include tracking the coherence state of at least one cache block contained within a region using a global directory, providing at least one region level sharing information about the least one cache block in the global directory, and providing at least one block level sharing information about the at least one cache block in the global directory. The tracking of the provided at least one region level sharing information and the provided at least one block level sharing information may organize the coherence state of the at least one cache block and the region.Type: GrantFiled: October 18, 2011Date of Patent: August 19, 2014Assignee: Advanced Micro Devices, Inc.Inventors: Bradfod M. Beckmann, Arkaprava Basu, Steven K. Reinhardt
-
Patent number: 8799586Abstract: Methods and apparatus relating to memory mirroring and migration at a Home Agent (HA) are described. In one embodiment, a home agent may mirror its data at a slave agent. In some embodiments, a bit in a directory may indicate status of cache lines. Other embodiments are also disclosed.Type: GrantFiled: September 30, 2009Date of Patent: August 5, 2014Assignee: Intel CorporationInventors: Ganesh Kumar, Dheemanth Nagaraj, Vincent R. Freytag, Eric Delano, Gregory S. Averill
-
Patent number: 8782344Abstract: A cache layer leverages a logical address space and storage metadata of a storage layer (e.g., storage layer) to cache data of a backing store. The cache layer maintains access metadata to track data characteristics of logical identifiers in the logical address space, including accesses pertaining to data that is not in the cache. The access metadata may be separate and distinct from the storage metadata maintained by the storage layer. The cache layer determines whether to admit data into the cache using the access metadata. Data may be admitted into the cache when the data satisfies cache admission criteria, which may include an access threshold and/or a sequentiality metric. Time-ordered history of the access metadata is used to identify important/useful blocks in the logical address space of the backing store that would be beneficial to cache.Type: GrantFiled: January 12, 2012Date of Patent: July 15, 2014Assignee: Fusion-io, Inc.Inventors: Nisha Talagala, Swaminathan Sundararaman, Amar Mudrankit
-
Patent number: 8751748Abstract: In a parallel processing system with speculative execution, conflict checking occurs in a directory lookup of a cache memory that is shared by all processors. In each case, the same physical memory address will map to the same set of that cache, no matter which processor originated that access. The directory includes a dynamic reader set encoding, indicating what speculative threads have read a particular line. This reader set encoding is used in conflict checking. A bitset encoding is used to specify particular threads that have read the line.Type: GrantFiled: January 18, 2011Date of Patent: June 10, 2014Assignee: International Business Machines CorporationInventors: Daniel Ahn, Luis H. Ceze, Alan Gara, Martin Ohmacht, Zhuang Xiaotong
-
Patent number: 8719507Abstract: Parallel computing environments, where threads executing in neighboring processors may access the same set of data, may be designed and configured to share one or more levels of cache memory. Before a processor forwards a request for data to a higher level of cache memory following a cache miss, the processor may determine whether a neighboring processor has the data stored in a local cache memory. If so, the processor may forward the request to the neighboring processor to retrieve the data. Because access to the cache memories for the two processors is shared, the effective size of the memory is increased. This may advantageously decrease cache misses for each level of shared cache memory without increasing the individual size of the caches on the processor chip.Type: GrantFiled: January 4, 2012Date of Patent: May 6, 2014Assignee: International Business Machines CorporationInventors: Miguel Comparan, Robert A. Shearer
-
Patent number: 8719508Abstract: Parallel computing environments, where threads executing in neighboring processors may access the same set of data, may be designed and configured to share one or more levels of cache memory. Before a processor forwards a request for data to a higher level of cache memory following a cache miss, the processor may determine whether a neighboring processor has the data stored in a local cache memory. If so, the processor may forward the request to the neighboring processor to retrieve the data. Because access to the cache memories for the two processors is shared, the effective size of the memory is increased. This may advantageously decrease cache misses for each level of shared cache memory without increasing the individual size of the caches on the processor chip.Type: GrantFiled: December 10, 2012Date of Patent: May 6, 2014Assignee: International Business Machines CorporationInventors: Miguel Comparan, Robert A. Shearer
-
Patent number: 8631208Abstract: In one embodiment, the present invention includes a method for receiving a memory request from a device coupled to an input/output (IO) interconnect, accessing a mapping table associated with the IO interconnect to determine if an address range including an address of the memory request is coherent, and if so, sending the memory request and a coherency indicator to indicate the coherent state of data at the address, otherwise sending the memory request and the coherency indicator to indicate a non-coherent state. Other embodiments are described and claimed.Type: GrantFiled: January 27, 2009Date of Patent: January 14, 2014Assignee: Intel CorporationInventors: Zhen Fang, David J. Harriman, Michael W. Leddige
-
Patent number: 8402247Abstract: Described herein are method and apparatus for using an LLRRM device as a storage device in a storage system. At least three levels of data structures may be used to remap storage system addresses to LLRRM addresses for read requests, whereby a first-level data structure is used to locate a second-level data structure corresponding to the storage system address, which is used to locate a third-level data structure corresponding to the storage system address. An LLRRM address may comprise a segment number determined from the second-level data structure and a page number determined from the third-level data structure. Update logs may be produced and stored for each new remapping caused by a write request. An update log may specify a change to be made to a particular data structure. The stored update logs may be performed on the data structures upon the occurrence of a predetermined event.Type: GrantFiled: March 7, 2012Date of Patent: March 19, 2013Assignee: NetApp, Inc.Inventors: Garth R. Goodson, Rahul N. Iyer
-
Patent number: 8209491Abstract: Techniques for directory server integration are disclosed. In one particular exemplary embodiment, the techniques may be realized as a method for directory server integration comprising setting one or more parameters determining a range of permissible expiration times for a plurality of cached directory entries, creating, in electronic storage, a cached directory entry from a directory server, assigning a creation time to the cached directory entry, and assigning at least one random value to the cached directory entry, the random value determining an expiration time for the cached directory entry within the range of permissible expiration times, wherein randomizing the expiration time for the cached directory entry among the range of permissible expiration times for a plurality of cached directory entries reduces an amount of synchronization required between cache memory and the directory server at a point in time.Type: GrantFiled: April 27, 2010Date of Patent: June 26, 2012Assignee: Symantec CorporationInventors: Ayman Mobarak, Nathan Moser, Chad Jamart
-
Patent number: 8190852Abstract: System for controlling data transfer between a host system and storage devices. A virtualization controller implements the data transfer and includes first ports for connection with the storage devices, a second port for connection with the host system, a processor, and a memory configured to store volume mapping information which correlates first identification information used by the host system to access a first storage area in one of the storage devices, with second identification information for identifying the first storage area, the correlation being used by the processor to access the first storage area. When data stored in the first storage area is transferred to a second storage area, the processor correlates the first identification information with a third identification information for identifying the second storage area and registers the first identification information and the third identification information in the volume mapping information.Type: GrantFiled: March 19, 2007Date of Patent: May 29, 2012Assignee: Hitachi, Ltd.Inventors: Kiyoshi Honda, Naoko Iwami, Kazuyoshi Serizawa
-
Patent number: 8156305Abstract: Described herein are method and apparatus for using an LLRRM device as a storage device in a storage system. At least three levels of data structures may be used to remap storage system addresses to LLRRM addresses for read requests, whereby a first-level data structure is used to locate a second-level data structure corresponding to the storage system address, which is used to locate a third-level data structure corresponding to the storage system address. An LLRRM address may comprise a segment number determined from the second-level data structure and a page number determined from the third-level data structure. Update logs may be produced and stored for each new remapping caused by a write request. An update log may specify a change to be made to a particular data structure. The stored update logs may be performed on the data structures upon the occurrence of a predetermined event.Type: GrantFiled: October 1, 2008Date of Patent: April 10, 2012Assignee: NetApp, Inc.Inventors: Garth R. Goodson, Rahul N. Iyer
-
Patent number: 8131936Abstract: A method and apparatus for implementing a combined data/coherency cache for a shared memory multi-processor. The combined data/coherency cache includes a system cache with a number of entries. The method includes building a system cache directory with a number of entries equal to the number of entries of the system cache. The building includes designating a number of sub-entries for each entry which is determined by a number of sub-entries operable for performing system cache coherency functions. The building also includes providing a sub-entry logic designator for each entry, and mapping one of the sub-entries for each entry to the system cache via the sub-entry logic designator.Type: GrantFiled: February 11, 2005Date of Patent: March 6, 2012Assignee: International Business Machines CorporationInventors: Keith N. Langston, Pak-kin Mak, Bruce A. Wagar
-
Patent number: 7941610Abstract: Coherency directory updating is provided in a multiprocessor computing system. A plurality of memory resources have a directory, and are operably connected to an interconnect fabric. A cell is operably connected to the interconnect fabric. The cell has a cache including an entry for each of a plurality of coherency units, each coherency unit included in a memory block representing a contiguous portion of the plurality of memory resources. A controller is operably connected to the interconnect fabric. The controller is configured to control a portion of the plurality of memory resources, and has a comparator configured to identify whether a memory block is local. If the memory block is local, the controller is configured to set a state of the directory to exclusive for a write transaction. If the memory block is not local, the controller is configured to set the state to invalid for a write transaction.Type: GrantFiled: April 27, 2006Date of Patent: May 10, 2011Assignee: Hewlett-Packard Development Company, L.P.Inventors: Erin A. Handgen, Patrick Knebel
-
Patent number: 7937530Abstract: A method and apparatus for accessing a processor cache. The method includes executing an access instruction in a processor core of the processor. The access instruction provides an untranslated effective address of data to be accessed by the access instruction. The method also includes determining whether a level one cache for the processor core includes the data corresponding to the effective address of the access instruction. The effective address of the access instruction is used without address translation to determine whether the level one cache for the processor core includes the data corresponding to the effective address. If the level one cache includes the data corresponding to the effective address, the data for the access instruction is provided from the level one cache.Type: GrantFiled: June 28, 2007Date of Patent: May 3, 2011Assignee: International Business Machines CorporationInventor: David Arnold Luick
-
Patent number: 7831775Abstract: Method, system and computer program product for tracking changes in an L1 data cache directory. A method for tracking changes in an L1 data cache directory determines if data to be written to the L1 data cache is to be written to an address to be changed from an old address to a new address. If it is determined that the data to be written is to be written to an address to be changed, a determination is made if the data to be written is associated with the old address or the new address. If it is determined that the data is to be written to the new address, the data is allowed to be written to the new address following a prescribed delay after the address to be changed is changed. The method is preferably implemented in a system that provides a Store Queue (STQU) design that includes a Content Addressable Memory (CAM)-based store address tracking mechanism that includes early and late write CAM ports. The method eliminates time windows and the need for an extra copy of the L1 data cache directory.Type: GrantFiled: June 2, 2008Date of Patent: November 9, 2010Assignee: International Business Machines CorporationInventors: Sheldon B. Levenstein, Anthony Saporito
-
Patent number: 7774551Abstract: A method for maintaining cache coherence comprises coordinating operations among a plurality of processors distributed among a plurality of nodes coupled by an interconnect fabric and managing cache coherence in a plurality of memory directories respectively associated with the processor plurality in combination with a node controller directory cache associated with a node controller coupled between the processor plurality and the interconnect fabric. The method further comprises maintaining memory coherence directory information comprising identifying processors within a node in a first portion of bits of a memory directory entry coupled to an associated processor in the node and identifying subsets of processors external to the node in the system in a second portion of bits.Type: GrantFiled: October 6, 2006Date of Patent: August 10, 2010Assignee: Hewlett-Packard Development Company, L.P.Inventors: Blaine D. Gaither, Verna Knapp