Patents Examined by Tracy C Chan
-
Patent number: 11797442Abstract: An integrated circuit and a method for executing a cache management operation are provided. The integrated circuit includes a master interface, a slave interface, and a link. The link is connected between the master interface and the slave interface, and the link includes an A-channel, a B-channel, a C-channel, a D-channel, and an E-channel. The A-channel is configured to transmit a cache management operation message of the master interface to the slave interface, and the cache management operation message is configured to manage data consistency between different data caches. The D-channel is configured to transmit a cache management operation acknowledgement message of the slave interface to the master interface.Type: GrantFiled: October 18, 2021Date of Patent: October 24, 2023Assignee: ANDES TECHNOLOGY CORPORATIONInventors: Zhong-Ho Chen, Yu-Lin Hsiao, Hsin Ming Chen
-
Patent number: 11797441Abstract: An exempt portion of a data cache of a memory sub-system is identified. The exempt portion includes a first set of data blocks comprising first data written by a host system to the data cache. A collected portion of the data cache of the memory sub-system is identified. The collected portion includes a second set of data blocks comprising second data written by the host system. A media management operation is performed on the collected portion of the data cache to relocate the second data to a storage area of the memory sub-system that is at a higher data density than the data cache, wherein the exempt portion of the data cache is exempt from the media management operation.Type: GrantFiled: November 24, 2020Date of Patent: October 24, 2023Assignee: Micron Technology, Inc.Inventors: Ashutosh Malshe, Sampath K. Ratnam, Kishore Kumar Muchherla, Peter Feeley
-
Patent number: 11790981Abstract: Provided is a method of performing an internal processing operation of a memory device in a system including a host device and the memory device. The memory device includes a memory cell array and a processor-in-memory (PIM) performing an internal processing operation. In an internal processing mode, by the PIM, the memory device performs the internal processing operation based on internal processing information stored in the memory cell array. When the internal processing information is an internal processing operation command indicating a type of the internal processing operation, the memory device outputs the internal processing operation command including an internal processing read command and an internal processing write command to the host device. The host device issues to the memory device a priority command determined from among a data transaction command and the internal processing operation command.Type: GrantFiled: August 8, 2022Date of Patent: October 17, 2023Assignee: SAMSUNG ELECTRONICS CO., LTD.Inventors: Pavan Kumar Kasibhatla, Seong-il O, Hak-soo Yu
-
Patent number: 11782875Abstract: A plurality of computing devices are communicatively coupled to each other via a network, and each of the plurality of computing devices is operably coupled to one or more flash storage devices. Each computing device is operable to access one or more memory blocks within the flash storage devices and maintain a directory structure for managing access to the memory. The directory structure may be adaptively resized according to the addition or removal of one or more associated files stored in memory.Type: GrantFiled: September 5, 2018Date of Patent: October 10, 2023Inventors: Maor Ben Dayan, Omri Palmon, Liran Zvibel, Kanael Arditti
-
Patent number: 11776591Abstract: Various implementations described herein refer to a method for providing single port memory with multiple different banks having a first bank and a second bank that is different than the first bank. The method may include coupling multiple wordlines to the single port memory including coupling a first wordline to the first bank and coupling a second wordline to the second bank. The method may include performing multiple memory access operations concurrently in the single port memory.Type: GrantFiled: September 26, 2019Date of Patent: October 3, 2023Assignee: Arm LimitedInventors: Lalit Gupta, Bo Zheng, El Mehdi Boujamaa, Fakhruddin Ali Bohra
-
Patent number: 11768602Abstract: Embodiments of the invention provide systems and methods for managing processing, memory, storage, network, and cloud computing to significantly improve the efficiency and performance of processing nodes. More specifically, embodiments of the present invention are directed to an instruction set of an object memory fabric. This object memory fabric instruction set can be used to provide a unique instruction model based on triggers defined in metadata of the memory objects. This model represents a dynamic dataflow method of execution in which processes are performed based on actual dependencies of the memory objects. This provides a high degree of memory and execution parallelism which in turn provides tolerance of variations in access delays between memory objects. In this model, sequences of instructions are executed and managed based on data access. These sequences can be of arbitrary length but short sequences are more efficient and provide greater parallelism.Type: GrantFiled: August 6, 2021Date of Patent: September 26, 2023Assignee: Ultrata, LLCInventors: Steven J. Frank, Larry Reback
-
Patent number: 11755495Abstract: A processing device receives a request specifying a logical address associated with a host-initiated operation directed at a first portion of a memory device. The processing device accesses a second L2P table comprising a mapping between logical addresses and physical addresses in a second portion of the memory device. A physical location within the second portion of the memory device is identified based on the second L2P table. The physical location corresponds to a portion of a first L2P table that specifies a physical address within the first portion of the memory device that corresponds to the logical address. The physical address is identified based on the portion of the first L2P table and the host-initiated operation is performed at the physical address.Type: GrantFiled: March 16, 2021Date of Patent: September 12, 2023Assignee: Micron Technology, Inc.Inventor: Sanjay Subbarao
-
Patent number: 11755475Abstract: An information handling system includes first and second memory modules, and a central processing unit. The first memory module includes one or more memory ranks of memory devices, and a first plurality of thermal sensors. The second memory module includes one or more memory ranks of memory devices, and a second plurality of thermal sensors. The central processing unit receives first thermal telemetry data for the first memory module from the first thermal sensors, and second thermal telemetry data for the second memory module from the second thermal sensors. In response to the reception of the first thermal telemetry data, the central processing unit determines a first localized temperature of a first memory rank. In response to the first localized temperature exceeding a threshold temperature, the central processing unit re-maps access of data from the first memory rank to a second memory rank.Type: GrantFiled: January 31, 2020Date of Patent: September 12, 2023Assignee: Dell Products L.P.Inventors: Balaji Bapu Gururaja Rao, Jordan Chin, Stuart Allen Berke
-
Patent number: 11749339Abstract: Provided is a method of performing an internal processing operation of a memory device in a system including a host device and the memory device. The memory device includes a memory cell array and a processor-in-memory (PIM) performing an internal processing operation. In an internal processing mode, by the PIM, the memory device performs the internal processing operation based on internal processing information stored in the memory cell array. When the internal processing information is an internal processing operation command indicating a type of the internal processing operation, the memory device outputs the internal processing operation command including an internal processing read command and an internal processing write command to the host device. The host device issues to the memory device a priority command determined from among a data transaction command and the internal processing operation command.Type: GrantFiled: August 8, 2022Date of Patent: September 5, 2023Assignee: SAMSUNG ELECTRONICS CO., LTD.Inventors: Pavan Kumar Kasibhatla, Seong-il O, Hak-soo Yu
-
Patent number: 11734180Abstract: A method may use memory efficiently to extend cache. A processor receives a request to write data. The size of the data in the write request is compared to a threshold. When the size of the data exceeds the threshold, the data is stored on a solid state device. Page descriptors for the data on the solid state device are stored in a metadata log, and a reference to a first page descriptor of the page descriptors in the metadata log is stored in a first hash table in memory.Type: GrantFiled: July 28, 2021Date of Patent: August 22, 2023Assignee: EMC IP Holding Company LLCInventors: Oran Baruch, Vamsi K. Vankamamidi
-
Patent number: 11734189Abstract: A request that specifies a logical address associated with a host-initiated operation directed at a first portion of a memory device is received. A logical to physical (L2P) table is accessed. The L2P table comprises a mapping between logical addresses and physical addresses in a second portion of the memory device. An entry in the L2P table that corresponds to the logical address is identified and is determined to point to an entry in a read cache table. Based on an entry number of the entry in the read cache table, a chunk address of a chunk from among multiple chunks of a read cache is calculated. A physical address that corresponds to the logical address specified by the request is identified by accessing the chunk of read cache. The host-initiated operation is performed at a physical location within the first portion of the memory device corresponding the physical address.Type: GrantFiled: March 16, 2021Date of Patent: August 22, 2023Assignee: Micron Technology, Inc.Inventor: Sanjay Subbarao
-
Patent number: 11734167Abstract: The present technology relates to an electronic device. According to the present technology, a storage device that manages map data using a volatile memory device having a limited capacity may include a nonvolatile memory device, a memory controller, and the volatile memory device which includes a map chunk buffer, a map chunk status table, a journal buffer, and a meta slice buffer.Type: GrantFiled: April 9, 2021Date of Patent: August 22, 2023Assignee: SK hynix Inc.Inventors: Ju Hyun Kim, Jin Yeong Kim, Jae Wan Yeon
-
Patent number: 11734178Abstract: A storage device includes: a memory device including a plurality of planes, and a plurality of cache buffers and data buffers; and a memory controller for controlling the memory device to transmit first data and second data from first plane and second plane into the respective first cache buffer and second cache buffer, and control the first cache buffer and the second cache buffer to transmit the first data and the second data to the memory controller. In response to a read request for third data from a host while the first data is transmitting from the first cache buffer to the memory controller, the memory controller transmits a cache read command to the memory device such that the memory device reads the third data after the first data is completely transmitted to the memory controller, before the second data is transmitted from the second cache buffer.Type: GrantFiled: June 25, 2021Date of Patent: August 22, 2023Assignee: SK hynix Inc.Inventor: Chung Un Na
-
Patent number: 11726916Abstract: A method, computer program product, and computing system for defining a normal IO write mode for writing data to a storage system including: writing the data to a cache memory system of a first storage node, writing the data to a journal of the first storage node, sending a notification concerning the data to a second storage node, writing one or more metadata entries concerning the data to a journal of the second storage node, sending an acknowledgment signal to the host device, and writing the data to the storage array. A request may be received to enter a testing IO write mode. In response to receiving the request, the data may be written to the cache memory system. The writing of the data to the journal may be bypassed. The acknowledgment signal may be sent to the host device. The data may be written to the storage array.Type: GrantFiled: April 27, 2022Date of Patent: August 15, 2023Assignee: EMC IP Holding Company, LLCInventors: Geng Han, Vladimir Shveidel, Uri Shabi
-
Patent number: 11720489Abstract: A variety of applications can include apparatus and/or methods of operating the apparatus in which storage in the memory device is managed. An allocation can include conducting a garbage collection procedure to free up one or more blocks. In various embodiments, execution of a garbage collection procedure can be based on operation of two tables with respect to a logical to physical mapping table split into logical to physical mapping table regions saved in the memory device. The first table can maintain counts of valid pages in blocks for a logical to physical mapping table region. The second table can include bits to identify logical to physical mapping table regions involved in the garbage collection procedure based on the entries in the first table. Search of the second table can determine logical to physical mapping table regions involved in the garbage collection. Additional apparatus, systems, and methods are disclosed.Type: GrantFiled: September 2, 2022Date of Patent: August 8, 2023Assignee: Micron Technology, Inc.Inventors: Xinghui Duan, Giuseppe D'Eliseo, Lalla Fatima Drissi, Giuseppe Ferrari, Eric Kwok Fung Yuen, Massimo Iaculo
-
Patent number: 11681442Abstract: An example method may include performing a first wear leveling operation on a group of data blocks based on a write counter associated with the group of data blocks, wherein the first wear leveling operation comprises including the group of data blocks in a plurality of groups of mapped data blocks, responsive to including the group of data blocks in the plurality of groups of mapped data blocks, performing a second wear leveling operation on the group of data blocks, wherein performing the second wear leveling operation comprises determining a base address of the group of data blocks, the base address indicating a location at which the group of data blocks begins, and accessing a data block in the group of data blocks based on the base address of the group of data blocks and a logical address associated with the data block.Type: GrantFiled: June 13, 2022Date of Patent: June 20, 2023Assignee: Micron Technology, Inc.Inventors: Fangfang Zhu, Jiangli Zhu, Ning Chen, Ying Yu Tai
-
Patent number: 11681626Abstract: A device including: a processor executing a program; a first cache memory; a second cache memory belonging to a memory hierarchy lower than that of the first cache memory; a determination unit that determines, based on first information indicating a virtual address of information accessed in the second cache memory when the program is executed, second information indicating a virtual address of target information to be prefetched; and a prefetch unit that prefetches the target information and stores the prefetched target information in the second cache memory, wherein the second cache memory includes a conversion unit that converts, by using correspondence information indicating a correspondence relationship between the physical address of the target information and the virtual address of the target information, the second information into third information indicating a physical address of the target information, and the prefetch unit prefetches the target information using the third information.Type: GrantFiled: January 10, 2022Date of Patent: June 20, 2023Assignee: FUJITSU LIMITEDInventors: Shiho Nakahara, Takahide Yoshikawa
-
Patent number: 11675498Abstract: One or more usage parameter values associated with a host system are obtained. The one or more parameter values correspond to one or more operations associated with a memory sub-system. An expected time period during which a set of host data will be received from the host system is determined in view of the one or more usage parameter values. In response to a determination, in view of an indication received from the host system, that the set of host data will not be received at the expected time period, a media management operation is performed at memory units of the memory sub-system.Type: GrantFiled: March 8, 2022Date of Patent: June 13, 2023Assignee: Micron Technology, Inc.Inventors: Poorna Kale, Ashok Sahoo
-
Patent number: 11675700Abstract: A method includes receiving, by a level two (L2) controller, a first request for a cache line in a shared cache coherence state; mapping, by the L2 controller, the first request to a second request for a cache line in an exclusive cache coherence state; and responding, by the L2 controller, to the second request.Type: GrantFiled: February 7, 2022Date of Patent: June 13, 2023Assignee: Texas Instruments IncorporatedInventors: Abhijeet Ashok Chachad, David Matthew Thompson, Timothy David Anderson, Kai Chirca
-
Patent number: 11663120Abstract: Devices and techniques for controlling NAND operation latency are described herein. A controlled can receive a write request. The controller can then calculate a number of garbage collection operations to perform on a physical block that is closed. Here, the calculation includes adding a logical-to-physical (L2P) region search ratio to a cadence calculation for garbage collection. Garbage collection operations can be performed on the physical block in accordance with the calculated number of garbage collection operations to perform. Then, the controller can perform the write request in response to completing the calculated number of garbage collection operations.Type: GrantFiled: November 8, 2021Date of Patent: May 30, 2023Assignee: Micron Technology, Inc.Inventors: Giuseppe D'Eliseo, Luigi Esposito, Xinghui Duan, Lucia Santojanni, Massimo Iaculo