Abstract: Exemplary embodiments of the disclosure provide a memory management method for a rewritable non-volatile memory module including the following steps. A host write operation is performed to receive a write command from a host system and store a first data corresponding to the write command to a first physical unit. A first updating data corresponding to the host write operation is recorded. A data merge operation is performed to read a second data from a second physical unit and store the second data to a third physical unit. A second updating data corresponding to the data merge operation is recorded. A management information is read from the rewritable non-volatile memory module to a buffer memory and updated in the buffer memory according to the first updating data and the second updating data.
Abstract: Systems and methods for pre-fetching data in a memory device are disclosed. The method may include receiving a current read command and determining whether the current read command is a random read command, for example based on a data chunk length identified by the current read command. The method may further include updating a prior read command data structure with the current read command, for random read commands; determining a predicted next read command from the prior read command data structure based on the current read command; and pre-fetching data associated with the predicted next read command. Functionality for prediction of next read commands, or pre-fetch of predicted next read commands, may be turned on or off based on resource availability or prediction success rate measurements.
Abstract: A storage device includes a descramble module configured to descramble at least a portion of a read command, which includes logical block address (LBA) information and first meta information, into first signature information and first physical address (PA) information, for accessing a flash memory. A compare module is provided, which is configured to compare the first signature information against stored signature information to thereby determine an equivalency or discrepancy therebetween. An access module is provided, which is configured to use the first PA information to access a data region of the flash memory, upon determination of the equivalency by said compare module.
Type:
Grant
Filed:
September 18, 2018
Date of Patent:
May 4, 2021
Inventors:
Dong-Woo Kim, Jae Sun No, Song Ho Yoon, Kyoung Back Lee, Wook Han Jeong
Abstract: A storage system includes a host device and a storage device. The host device generates a write command, a logical address of write data corresponding to the write command, and a selective purge tag indicating that the write data are targeted for selective purge. The storage device receives the write command, the logical address, and the selective purge tag, stores write data, and logically erases the stored write data upon receiving an erase command from the host device. In addition, the storage device physically erases the stored write data upon receiving a selective purge request from the host device.
Abstract: Methods, non-transitory computer readable media, and computing devices that execute a storage operation, without journaling the storage operation in a log, and withhold from a file system a list of freed inodes including an indication of an inode freed as a result of the execution of the storage operation. A consistency point operation is then initiated that retrieves storage operations logged as journal entries in the log and commits a result of each of the storage operations to data storage devices. A list of available inodes is updated based on the list of freed inodes, when the consistency point operation is determined to be complete. This technology reduces the number of storage operations that are required to be journaled to maintain consistency of a file system, thereby reducing the runtime resources required to facilitate the journaling and replay resource required to replay the storage operations following a recovery.
Type:
Grant
Filed:
January 30, 2020
Date of Patent:
April 20, 2021
Assignee:
NETAPP, INC.
Inventors:
Ram Kesavan, Ananthan Subramanian, Hiroshi Ishii, Abdul Basit, Joseph Brown, Rohit Singh
Abstract: In one embodiment, a method is operable in an over-provisioned storage device comprising a cache region and a main storage region. The method includes compressing incoming data, generating a compression parameter for the compressed data, and storing at least a portion of the compressed data in chunks in the main storage region of the storage device. The method also includes predicting when to store other chunks of the compressed data in the cache region based on the compression parameter.
Abstract: A method, apparatus and product for accelerating concurrent access to a file in a memory-based file system. The method comprising receiving a request issued by a program, for accessing a file stored in a memory-based file system; and subject to the request being associated with data modification of data within the file, and subject to the modification not necessitating change in a structure of a data structure used for content lookup for the file, acquiring a lock to the file to the program, wherein the lock is acquired in a shared mode.
Type:
Grant
Filed:
September 27, 2019
Date of Patent:
March 23, 2021
Assignee:
NetApp Inc.
Inventors:
Amit Golander, Sagi Manole, Boaz Harrosh
Abstract: A processor core associated with a first cache initiates entry into a powered-down state. In response, information representing a set of entries of the first cache are stored in a retention region that receives a retention voltage while the processor core is in a powered-down state. Information indicating one or more invalidated entries of the set of entries is also stored in the retention region. In response to the processor core initiating exit from the powered-down state, entries of the first cache are restored using the stored information representing the entries and the stored information indicating the at least one invalidated entry.
Type:
Grant
Filed:
November 1, 2017
Date of Patent:
March 23, 2021
Assignee:
ADVANCED MICRO DEVICES, INC.
Inventors:
William L. Walker, Michael L. Golden, Marius Evers
Abstract: An aspect of dynamic physical capacity allocation in an unbalanced content-addressable storage system includes initiating a scale up for a storage cluster. An aspect further includes determining a target number of hash-to-physical-location-on-disk (h2d) slices to be assigned to each data module (D module), and reassigning h2d slices assigned to D modules having greater than the target number of h2d slices to D modules having less h2d slices than the target number.
Type:
Grant
Filed:
July 31, 2018
Date of Patent:
March 23, 2021
Assignee:
EMC IP Holding Company LLC
Inventors:
Svetlana Kronrod, Zvi Schneider, Anton Kucherov
Abstract: Provided herein may be a memory system and a method of operating the memory system. The method includes: allocating a first buffer region to a first workload group; allocating a second buffer region to a second workload group; monitoring a first workload group latency and a second workload group latency; and dynamically adjusting a memory space of each of the first and second buffer regions based on a result of the monitoring.
Abstract: An apparatus and method are provided for issuing access requests to a memory controller for a memory device whose memory structure consists of a plurality of sub-structures. The apparatus has a request interface for issuing access requests to the memory controller, each access request identifying a memory address. Within the apparatus static abstraction data is stored providing an indication of one or more of the sub-structures of the memory device, and the apparatus also stores an indication of outstanding access requests issued from the request interface. Next access request selection circuitry is then arranged to select from a plurality of candidate access requests a next access request to issue from the request interface. That selection is dependent on sub-structure indication data that is derived from application of an abstraction data function, using the static abstraction data, to the memory addresses of the candidate access requests and the outstanding access requests.
Abstract: A cache stores, along with data that is being transferred from a higher level cache to a lower level cache, information indicating the higher level cache location from which the data was transferred. Upon receiving a request for data that is stored at the location in the higher level cache, a cache controller stores the higher level cache location information in a status tag of the data. The cache controller then transfers the data with the status tag indicating the higher level cache location to a lower level cache. When the data is subsequently updated or evicted from the lower level cache, the cache controller reads the status tag location information and transfers the data back to the location in the higher level cache from which it was originally transferred.
Abstract: Methods, systems, and devices are described for wireless communications. A request for data located in a memory page of a memory array may be received at a device, and a value of a prefetch counter associated with the memory page may be identified. A portion of the memory page that includes the requested data may then be communicated between a memory array and memory bank of the device based on the value of the prefetch counter. For instance, the portion of the memory page may be selected based on the value of the prefetch counter. A second portion of the memory page may be communicated to a buffer of the device, and the value of the prefetch counter may be modified based on a relationship between the first portion of the memory page and the second portion of the memory page.
Type:
Grant
Filed:
August 20, 2018
Date of Patent:
March 23, 2021
Assignee:
Micron Technology, Inc.
Inventors:
Robert Nasry Hasbun, Dean D. Gans, Sharookh Daruwalla
Abstract: A technique is described for managing a cache structure in a system employing transactional memory. The apparatus comprises processing circuitry to perform data processing operations in response to instructions, the processing circuitry comprising transactional memory support circuitry to support execution of a transaction, and a cache structure comprising a plurality of cache entries for storing data for access by the processing circuitry. Each cache entry has associated therewith an allocation tag, and allocation tag control circuitry is provided to control use of a plurality of allocation tags and to maintain an indication of a current state of each of those allocation tags.
Type:
Grant
Filed:
April 2, 2019
Date of Patent:
March 23, 2021
Assignee:
Arm Limited
Inventors:
Damien Guillaume Pierre Payet, Lucas Garcia, Natalya Bondarenko, Stefano Ghiggini
Abstract: The present invention provides a method performed by a secure digital (SD) card supporting both an SD mode and a peripheral component interconnect express (PCIe) mode for initializing the SD card. The method comprises: (a) after receiving a first supply voltage through a first voltage supply pin from a host coupled to the SD card, entering the SD mode if the SD card is not in the PCIe mode and a CMD0 command for entering the SD mode is received through a command pin from the host coupled to the SD card; and (b) after receiving the first supply voltage through the first voltage supply pin from the host coupled to the SD card, performing a PCIe linkup process if the SD card is not in the SD mode and a second supply voltage is received through a second voltage supply pin from the host coupled to the SD card. The SD card enters the PCIe mode if the PCIe linkup process succeeds.
Abstract: A memory management system, such as a virtual memory manager that manages a virtual memory space that includes volatile memory (e.g. DRAM) and non-volatile memory (e.g., flash memory) creates a reserved portion of memory in the volatile memory for at least one user application in one embodiment, and that reserved portion can also store content that it restricted to read only permission within the non-volatile memory.
Type:
Grant
Filed:
September 6, 2016
Date of Patent:
March 9, 2021
Assignee:
Apple Inc.
Inventors:
Francois Barbou-Des-Places, Joseph Sokol, Jr., Simon Douglas
Abstract: An apparatus includes a host device and a data storage device. The host device is configured to store a first translation map for converting a logical sector to a logical erase unit. The data storage device includes a plurality of flash memory devices and a memory controller operationally coupled with the flash memory devices, each of the flash memory devices being arranged into a plurality of erase units, each of the erase units having a plurality of pages for storing data. The memory controller is configured to receive a second translation map from the host device, the second translation map for converting a logical erase unit to a physical erase unit within the flash memory devices, and store the second translation map in a memory module on the data storage device.
Type:
Grant
Filed:
November 18, 2019
Date of Patent:
March 9, 2021
Assignee:
Google LLC
Inventors:
Christopher John Sabol, Slava Pestov, Thomas Wyatt Craig, Manuel Enrique Benitez, Monish Shah, Daniel Ari Ehrenberg
Abstract: A data storage device includes a storage unit; and a controller configured to select a write mode by analyzing a tendency of commands received from a host device, and operate in the selected write mode to write data to the storage or to read data from the storage.
Abstract: Methods, systems, and devices are described for wireless communications. A request for data located in a memory page of a memory array may be received at a device, and a value of a prefetch counter associated with the memory page may be identified. A portion of the memory page that includes the requested data may then be communicated between a memory array and memory bank of the device based on the value of the prefetch counter. For instance, the portion of the memory page may be selected based on the value of the prefetch counter. A second portion of the memory page may be communicated to a buffer of the device, and the value of the prefetch counter may be modified based on a relationship between the first portion of the memory page and the second portion of the memory page.
Type:
Grant
Filed:
May 9, 2018
Date of Patent:
March 9, 2021
Assignee:
Micron Technology, Inc.
Inventors:
Robert Nasry Hasbun, Dean D. Gans, Sharookh Daruwalla
Abstract: Apparatuses and methods for managing a coherent memory are described. These may include one or more algorithmic logic units (ALUs) and an input/output (IO) interface. The I/O interface may receive one or more commands and retrieve data from or write data to a memory device. Each command may contain a memory address portion associated with a memory device. The apparatus may also include a memory mapping unit and a device controller. The memory mapping unit may map the memory address to a memory portion of the memory device, and the device controller may communicate with the memory device to retrieve data from or write data to the memory device. The apparatus may be implemented as a processing element in a configurable logic block network, which may additionally include a control logic unit that receives programming instructions from an application and generate the one or more commands based on the instructions.