Multiple Caches Patents (Class 711/119)
  • Patent number: 12277059
    Abstract: The present application discloses a method and apparatus for reducing a mirror data transmission amount by a dual layer cache, and a device and a medium. The method includes: after receiving an input/output (IO) request, writing, by a first node, the IO request into a first upper-layer cache space; writing, by the first node, first cached data corresponding to the IO request into a first lower-layer cache space according to the IO request, and generating, by the first node, first index information for the first cached data; writing mirror data of the IO request into a second upper-layer cache space of a second node; and writing mirror data of the first index information into a second lower-layer cache space of the second node.
    Type: Grant
    Filed: December 20, 2024
    Date of Patent: April 15, 2025
    Assignee: Suzhou MetaBrain Intelligent Technology Co., Ltd.
    Inventors: Xiangfei Kong, Yonggang Wang
  • Patent number: 12271709
    Abstract: The present application relates to a device for selecting top values from a set of raw values, comprising: an output queue, a loop queue, a top value storage module and a control module.
    Type: Grant
    Filed: December 27, 2022
    Date of Patent: April 8, 2025
    Assignee: MONTAGE ELECTRONICS (SHANGHAI) CO
    Inventors: Zhijie Liu, Jinfeng Ji, Jie Dai
  • Patent number: 12261603
    Abstract: A System-on-Chip includes a data processing engine array. The data processing engine array includes a plurality of data processing engines organized in a grid. The plurality of data processing engines are partitioned into at least a first partition and a second partition. The first partition includes one or more first data processing engines of the plurality of data processing engines. The second partition includes one or more second data processing engines of the plurality of data processing engines. Each partition is configured to implement an application that executes independently of the other partition.
    Type: Grant
    Filed: May 18, 2023
    Date of Patent: March 25, 2025
    Assignee: Xilinx, Inc.
    Inventors: Sagheer Ahmad, Jaideep Dastidar, Brian C. Gaide, Juan J. Noguera Serra, Ian A. Swarbrick
  • Patent number: 12248401
    Abstract: An agent may be configured to invoke a first eviction operation that is interruptible by probe operations when receiving a first type of eviction message and invoke a second eviction operation in which probe operations are interruptible by the second eviction operation when receiving a second type of eviction message. In some implementations, the agent may maintain a data storage that is inclusive of at least one of unique or dirty cache blocks in a cache maintained by an agent that transmits the second type of eviction message. In some implementations, the agent may prevent a cache block from transitioning from a modified state to an exclusive state when the agent invokes the second eviction operation to evict the cache block. In some implementations, the agent may convert from the second eviction operation to the first eviction operation when receiving the second type of eviction message.
    Type: Grant
    Filed: June 26, 2023
    Date of Patent: March 11, 2025
    Assignee: SiFive, Inc.
    Inventors: Michael Klinglesmith, Eric Andrew Gouldey, Wesley Waylon Terpstra
  • Patent number: 12248402
    Abstract: A high bandwidth memory system. In some embodiments, the system includes: a memory stack having a plurality of memory dies and eight 128-bit channels; and a logic die, the memory dies being stacked on, and connected to, the logic die; wherein the logic die may be configured to operate a first channel of the 128-bit channels in: a first mode, in which a first 64 bits operate in pseudo-channel mode, and a second 64 bits operate as two 32-bit fine-grain channels, or a second mode, in which the first 64 bits operate as two 32-bit fine-grain channels, and the second 64 bits operate as two 32-bit fine-grain channels.
    Type: Grant
    Filed: November 28, 2022
    Date of Patent: March 11, 2025
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Krishna T. Malladi, Mu-Tien Chang, Dimin Niu, Hongzhong Zheng
  • Patent number: 12235756
    Abstract: Near-memory compute elements perform memory operations and temporarily store at least a portion of address information for the memory operations in local storage. A broadcast memory command is then issued to the near-memory compute elements that causes the near-memory compute elements to perform a subsequent memory operation using their respective address information stored in the local storage. This allows a single broadcast memory command to be used to perform memory operations across multiple memory elements, such as DRAM banks, using bank-specific address information. In one implementation, the approach is used to process workloads with irregular updates to memory while consuming less command bus bandwidth than conventional approaches. Implementations include using conditional flags to selectively designate address information in local storage that is to be processed with the broadcast memory command.
    Type: Grant
    Filed: December 21, 2021
    Date of Patent: February 25, 2025
    Assignee: Advanced Micro Devices, Inc.
    Inventors: Shaizeen Aga, Johnathan Alsop, Nuwan Jayasena
  • Patent number: 12223188
    Abstract: A memory interface for interfacing with a memory device includes a control circuit configured to determine whether a trigger event has occurred for initializing one or more memory locations in the memory device, and initialize the one or more memory locations in the memory device with pre-defined data upon determining the trigger event has occurred.
    Type: Grant
    Filed: May 19, 2022
    Date of Patent: February 11, 2025
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Raghu Vamsi Krishna Talanki, Archita Khare, Rahul Tarikere Ravikumar, Jinin So, Jonggeon Lee
  • Patent number: 12222867
    Abstract: A technology flushing a hierarchical cache structure based on a designated key identification code and a designated address. A processor includes a first core and a last level cache (LLC). The first core includes a decoder, a memory ordering buffer, and a first in-core cache module. In response to an Instruction Set Architecture (ISA) instruction that requests to flush a hierarchical cache structure according to a designated key identification code and a designated address, the decoder outputs at least one microinstruction. According to the at least one microinstruction, a flushing request with the designated key identification code and the designated address is provided to the first in-core cache module through the memory ordering buffer, and then the first in-core cache module provides the LLC with the flushing request, so that the LLC flushes its matching cache line which matches the designated key identification code and the designated address.
    Type: Grant
    Filed: October 14, 2022
    Date of Patent: February 11, 2025
    Assignee: SHANGHAI ZHAOXIN SEMICONDUCTOR CO., LTD.
    Inventors: Weilin Wang, Yingbing Guan, Minfang Zhu
  • Patent number: 12222873
    Abstract: Embodiments described herein may include apparatus, systems, techniques, or processes that are directed to PCIe Address Translation Service (ATS) to allow devices to have a DevTLB that caches address translation (per page) information in conjunction with a Device ProcessInfoCache (DevPIC) that will store process specific information. Other embodiments may be described and/or claimed.
    Type: Grant
    Filed: September 13, 2021
    Date of Patent: February 11, 2025
    Assignee: Intel Corporation
    Inventors: Rupin Vakharwala, Vedvyas Shanbhogue
  • Patent number: 12222953
    Abstract: Example embodiments described herein pertain to a geographic information system (GIS), configured to obtain geospatial data representing a geographic area, assign a projection and coordinate system to the geospatial data, apply a transformation to the geospatial data, and generate a tile cache based on the transformed geospatial data, the tile cache including the determined projection and coordinate system.
    Type: Grant
    Filed: June 8, 2023
    Date of Patent: February 11, 2025
    Assignee: Palantir Technologies Inc.
    Inventor: Peter Hong
  • Patent number: 12222853
    Abstract: Provided is a memory resource sharing system including a first memory, a first memory subsystem configured to identify and distribute resources of the first memory and to control data transmission of the first memory, and a first processor unit including a first processor connected to the first memory subsystem, a second memory, a second memory subsystem configured to identify and distribute resources of the second memory and to control data transmission of the second memory, and a second processor unit including a second processor connected to the second memory subsystem, wherein the first memory subsystem and the second memory subsystem are communicatively connected to each other through a memory bus.
    Type: Grant
    Filed: May 4, 2023
    Date of Patent: February 11, 2025
    Assignee: COSIGNON
    Inventor: Jang Ho Park
  • Patent number: 12222913
    Abstract: One example method includes receiving at a dedupe system, from a client, a request that comprises a set of fingerprints, where each fingerprint in the set corresponds to a particular data segment, filtering, at the dedupe system, the set of fingerprints into a set of unique fingerprints and a set of non-unique fingerprints, reading, at the dedupe system, from a container where copies of the non-unique fingerprints are stored, an additional set of non-unique fingerprints, sending, from the dedupe system to the client, a single response that comprises both the set of unique fingerprints and the additional set of non-unique fingerprints, and receiving from the client, at the dedupe system, data segments that respectively correspond to the unique fingerprints in the set of unique fingerprints, but no data segments corresponding to the non-unique fingerprints in the set of non-unique fingerprints are received by the dedupe system from the client.
    Type: Grant
    Filed: November 30, 2023
    Date of Patent: February 11, 2025
    Assignee: EMC IP Holding Company LLC
    Inventors: Kalyan C. Gunda, Jagannathdas Rath
  • Patent number: 12210457
    Abstract: A network processor includes a memory subsystem serving a plurality of processor cores. The memory subsystem includes a hierarchy of caches. A mid-level instruction cache provides for caching instructions for multiple processor cores. Likewise, a mid-level data cache provides for caching data for multiple cores, and can optionally serve as a point of serialization of the memory subsystem. A low-level cache is partitionable into partitions that are subsets of both ways and sets, and each partition can serve an independent process and/or processor core.
    Type: Grant
    Filed: July 8, 2021
    Date of Patent: January 28, 2025
    Assignee: MARVELL ASIA PTE, LTD.
    Inventors: Shubhendu S. Mukherjee, David H. Asher, Richard E. Kessler, Srilatha Manne
  • Patent number: 12205000
    Abstract: Methods and systems for cross-platform user profiling based on disparate datasets using machine learning models. Specifically, the methods and systems comprising retrieving a cross-platform profile, wherein the cross-platform profile comprises a profile linked to an account, for a user, that is used across multiple assets. The methods and system may then update a status of the cross-platform profile based on incidents detected using machine learning models. The methods and system may then generate for presentation, in a user interface for the account, the status of cross-platform profile.
    Type: Grant
    Filed: December 31, 2020
    Date of Patent: January 21, 2025
    Assignee: GGWP, Inc.
    Inventors: Dennis Fong, Kun Gao, George Ng, Ling Xiao
  • Patent number: 12197342
    Abstract: An arithmetic processing device includes: an arithmetic circuit that executes an instruction; a first cache which is coupled to the arithmetic circuit and which has a plurality of first entries each including a first tag region and a first data region that holds cache line data; a second tag region; a processor which controls the first cache based on information held in the second tag region; and a second cache which is coupled to the first cache via the processor and which includes a plurality of second entries each of which includes a third tag region and a second data region that holds cache line data. The second tag region includes a first region that holds first information which specifies whether or not the second data region holds cache line data which has the same address as the address of cache line data held in the first data region.
    Type: Grant
    Filed: March 6, 2023
    Date of Patent: January 14, 2025
    Assignee: FUJITSU LIMITED
    Inventor: Toru Hikichi
  • Patent number: 12197334
    Abstract: This invention involves a cache system in a digital data processing apparatus including: a central processing unit core; a level one instruction cache; and a level two cache. The cache lines in the second level cache are twice the size of the cache lines in the first level instruction cache. The central processing unit core requests additional program instructions when needed via a request address. Upon a miss in the level one instruction cache that causes a hit in the upper half of a level two cache line, the level two cache supplies the upper half level cache line to the level one instruction cache. On a following level two cache memory cycle, the level two cache supplies the lower half of the cache line to the level one instruction cache. This cache technique thus prefetches the lower half level two cache line employing fewer resources than an ordinary prefetch.
    Type: Grant
    Filed: September 8, 2022
    Date of Patent: January 14, 2025
    Assignee: Texas Instruments Incorporated
    Inventors: Oluleye Olorode, Ramakrishnan Venkatasubramanian, Hung Ong
  • Patent number: 12197938
    Abstract: An input/output (I/O) device can initiate data migration of a virtual machine (VM) instance from a source device to a target device. The data migration of the VM instance may include migrating the data for the VM instance and tag data associated with the data. The data for the VM instance and the tag data may be stored together in a source memory. A first read request from the I/O device can enable a memory controller in the source device to read the data for the VM instance and the tag data together, store the tag data in a tag data buffer, and transmit the data for the VM instance to the target device. A second read request from the I/O device can read the stored tag data from the tag data buffer and transmit to the target device. The target device can write the data for the VM instance together with the tag data in the target memory.
    Type: Grant
    Filed: December 10, 2021
    Date of Patent: January 14, 2025
    Assignee: Amazon Technologies, Inc.
    Inventors: Ali Ghassan Saidi, Adi Habusha
  • Patent number: 12182421
    Abstract: In at least one embodiment, processing can include: receiving write operations; persistently recording, in a write cache or log, the write operations using page descriptors (PDESCs) of a PDESC pool and page buffers (PBs) of a PB pool; selecting, in accordance with criteria, write data pages stored in the PB pool for demotion to a physical large block (PLB) pool included in backend non-volatile storage, wherein each write data page selected denotes content written by a corresponding one of the write operations; responsive to the selecting, persistently storing the write data pages of the PB pool in the PLB pool; and updating PDESCs associated with the write data pages to reference corresponding storage locations in the PLB pool rather than in the PB pool.
    Type: Grant
    Filed: August 10, 2023
    Date of Patent: December 31, 2024
    Assignee: Dell Products L.P.
    Inventors: Vamsi K. Vankamamidi, Geng Han, Vikram A. Prabhakar
  • Patent number: 12182396
    Abstract: Techniques for performing memory operations are disclosed herein. The techniques include generating a plurality of performance log entries based on observed operations; generating a plurality of memory access log entries based on the observed operations, wherein each performance log entry of the plurality of performance log entries are associated with one or more memory access log entries of the plurality of memory access log entries, wherein each performance log entry is associated with an epoch; and wherein each memory access log entry is associated with an epoch and a memory address range.
    Type: Grant
    Filed: March 30, 2023
    Date of Patent: December 31, 2024
    Assignees: Advanced Micro Devices, Inc., ATI Technologies ULC
    Inventors: Christopher J. Brennan, Akshay Lahiry, Guennadi Riguer
  • Patent number: 12182038
    Abstract: Methods, apparatus, systems and articles of manufacture are disclosed for allocation in a victim cache system. An example apparatus includes a first cache storage, a second cache storage, a cache controller coupled to the first cache storage and the second cache storage and operable to receive a memory operation that specifies an address, determine, based on the address, that the memory operation evicts a first set of data from the first cache storage, determine that the first set of data is unmodified relative to an extended memory, and cause the first set of data to be stored in the second cache storage.
    Type: Grant
    Filed: January 8, 2024
    Date of Patent: December 31, 2024
    Assignee: Texas Instruments Incorporated
    Inventors: Naveen Bhoria, Timothy David Anderson, Pete Michael Hippleheuser
  • Patent number: 12182405
    Abstract: A data access system has host computers having front-end controllers nFE_SAN connected via a bus or network interconnect to back-end storage controllers nBE_SAN, and physical disk drives connected via network interconnect to the nBE_SANs to provide a distributed, high performance, policy based or dynamically reconfigurable, centrally managed, data storage acceleration system. The hardware and software architectural solutions eliminate BE_SAN controller bottlenecks and improve performance and scalability. In an embodiment, the nBE_SAN (BE_SAN) firmware recognize controller overload conditions, informs Distributed Resource Manager (DRM), and, based on the DRM provided optimal topology information, delegates part of its workload to additional controllers. The nFE_SAN firmware and additional hardware using functionally independent and redundant CPUs and memory that mitigate single points of failure and accelerates write performance.
    Type: Grant
    Filed: July 26, 2023
    Date of Patent: December 31, 2024
    Inventor: Branislav Radovanovic
  • Patent number: 12164385
    Abstract: An assigned subgroup that includes a plurality of entries is traversed by a prefetcher. It is determined that an expected number of entries associated with the assigned subgroup have been traversed. In response to determining that expected number of entries associated with the assigned subgroup have been traversed, it is determined that a last read entry associated with the assigned subgroup does not correspond to a last entry associated with the assigned subgroup. The prefetcher is preempted by stopping the prefetcher from obtaining a list of entries associated with a remaining portion of the assigned subgroup.
    Type: Grant
    Filed: January 17, 2024
    Date of Patent: December 10, 2024
    Assignee: Cohesity, Inc.
    Inventors: Amandeep Gautam, Venkata Ranga Radhanikanth Guturi
  • Patent number: 12111762
    Abstract: An embodiment of an integrated circuit may comprise a core, and a cache controller coupled to the core, the cache controller including circuitry to identify data from a working set for dynamic inclusion in a next level cache based on an amount of re-use of the next level cache, send a shared copy of the identified data to a requesting core of one or more processor cores, and maintain a copy of the identified data in the next level cache. Other embodiments are disclosed and claimed.
    Type: Grant
    Filed: December 22, 2020
    Date of Patent: October 8, 2024
    Assignee: Intel Corporation
    Inventors: Ayan Mandal, Leon Polishuk, Oz Shitrit, Joseph Nuzman
  • Patent number: 12099400
    Abstract: This invention is a streaming engine employed in a digital signal processor. A fixed data stream sequence is specified by a control register. The streaming engine fetches stream data ahead of use by a central processing unit and stores it in a stream buffer. Upon occurrence of a fault reading data from memory, the streaming engine identifies the data element triggering the fault preferably storing this address in a fault address register. The streaming engine defers signaling the fault to the central processing unit until this data element is used as an operand. If the data element is never used by the central processing unit, the streaming engine never signals the fault. The streaming engine preferably stores data identifying the fault in a fault source register. The fault address register and the fault source register are preferably extended control registers accessible only via a debugger.
    Type: Grant
    Filed: February 6, 2023
    Date of Patent: September 24, 2024
    Assignee: TEXAS INSTRUMENTS INCORPORATED
    Inventors: Joseph Zbiciak, Timothy D. Anderson, Duc Bui, Kai Chirca
  • Patent number: 12099443
    Abstract: Techniques are provided for implementing and managing a multi-modal write cache for a data storage system. For example, a storage control system is configured to perform a write caching method which comprises the storage control system receiving an input/output (I/O) write request from a client application to write data to a primary storage volume, comparing a current I/O workload associated with the client application to an I/O workload threshold, and writing the data of the I/O write request to one of (i) a persistent write cache in a persistent storage volume and (ii) a non-persistent write cache in a non-persistent storage volume, based at least in part on a result of comparing the current I/O workload to the I/O workload threshold.
    Type: Grant
    Filed: July 13, 2023
    Date of Patent: September 24, 2024
    Assignee: Dell Products L.P.
    Inventors: Doron Tal, Yosef Shatsky
  • Patent number: 12086066
    Abstract: A cache architecture for an array of identical cores arranged in a grid. Each of the cores include interconnections to neighboring cores in the grid, a memory, and an algorithmic logic unit. A first core of the array is configured to receive a memory access request for data from at least one core of the array of cores configured to perform a computational operation. A second core of the array is configured to determine whether the requested data is present in a cache memory via a cache index including addresses in the cache memory. A third core of the array is configured as the cache memory. The memory of the third core is used as the cache memory. An address of the requested data from the cache index is passed to the third core to output the requested data.
    Type: Grant
    Filed: March 15, 2023
    Date of Patent: September 10, 2024
    Assignee: Cornami, Inc.
    Inventor: Martin Alan Franz, II
  • Patent number: 12061551
    Abstract: An access counter associated with a segment of a memory device is maintained. An access notification for a first line of the segment is received. An access type associated with the access notification is identified. A first value of the access counter is changed by a second value based on the access type. Based on the first value of the access counter, a memory management scheme is implemented.
    Type: Grant
    Filed: August 26, 2022
    Date of Patent: August 13, 2024
    Assignee: Micron Technology, Inc.
    Inventor: David Andrew Roberts
  • Patent number: 12047477
    Abstract: A network device includes a statelet storage storing statelets that retain state information associated with a packet flow through the network device and that the network device can interact with to control processing performed on packets of the data flow. The network device implements a set of instructions that interpret commands in the data packets to manage and interact with statelets. The statelets in the statelet storage are organized by a statelet key that is derived from information identifying the packet flow. Responsive to the commands in the packets, the network device can create, read, write, or delete statelets from the statelet storage. The statelet storage includes multiple statelets each statelet including multiple fields. The network device may access the statelets to control/monitor a packet flow using information in a network data plane without receiving control information from a network control plane.
    Type: Grant
    Filed: August 31, 2020
    Date of Patent: July 23, 2024
    Assignee: Huawei Technologies Co., Ltd.
    Inventors: Alexander Clemm, Uma S. Chunduri, Renwei Li
  • Patent number: 12045498
    Abstract: Disclosed are a solid state drive and a write operation method. The solid state drive comprises: a controller, receiving write data from outside and comprising a first cache unit for storing the write data; a Flash memory, receiving the write data sent by the first cache unit according to a first instruction of the controller; a second cache unit, storing the write data from the first cache unit as backup data, and sending the backup data to the Flash memory according to a second instruction of the controller. The second instruction is obtained after the write data fails to be written into the Flash memory under the first instruction, so that the backup data can continue to be called if write operation fails. By combining advantages of the first and second cache units, efficiency and quality of write operations are improved and bandwidth requirements are lowered.
    Type: Grant
    Filed: April 25, 2022
    Date of Patent: July 23, 2024
    Assignee: MAXIO TECHNOLOGY (HANGZHOU) CO., LTD.
    Inventors: Wei Xu, Zihua Xiao, Hui Jiang, Zhengliang Chen
  • Patent number: 12038860
    Abstract: Systems, methods and computer software are disclosed for fronthaul. In one embodiment a method is disclosed, comprising: providing a virtual Radio Access Network (vRAN) having a centralized unit (CU) and a distributed unit (DU); and interconnecting the CU and DU over an Input/Output (I/O) bus using Peripheral Component Interconnect-Express (PCIe); wherein the CU and the DU include a PCI to optical converter and an optical to PCI converter.
    Type: Grant
    Filed: January 3, 2023
    Date of Patent: July 16, 2024
    Assignee: Parallel Wireless, Inc.
    Inventors: Ofir Ben Ari Katzav, David Johnston, Steven Paul Papa
  • Patent number: 12038901
    Abstract: Methods and system for a database management system (DBMS) in which a leader thread is elected from concurrent transaction threads stored in one or more data nodes. While the leader thread copies its own thread transaction log onto a reserved portion of the shared log buffer, the leader thread permits other transaction threads to attach to a thread chain starting with the leader thread. Once the leader has completed copying its thread transaction log onto the shared log buffer, it then reserves a portion of the shared log buffer, and copies the member thread transaction logs onto the shared log buffer to reduce the contention for shared buffer may be reduced.
    Type: Grant
    Filed: February 7, 2022
    Date of Patent: July 16, 2024
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Ronen Grosman, Ping Chen
  • Patent number: 12038913
    Abstract: Disclosed is a method for managing a database, which is performed by a first database server including at least one processor constituting a cluster jointly with at least one second database server. The method for managing a database may include loading, on a buffer cache, a first data block based on a first transaction for modifying the first data block located in a sharing storage shared jointly with the at least one second database server. The method may include modifying the first data block loaded on the buffer cache. The method may include determining flushing a first log generated by the modification of the first data block to the sharing storage.
    Type: Grant
    Filed: October 26, 2022
    Date of Patent: July 16, 2024
    Assignee: TmaxTibero Co., Ltd.
    Inventors: Jaemin Oh, Hakju Kim, Dongyun Yang, Sangyoung Park
  • Patent number: 12032683
    Abstract: Log entries and baseline log entries have timestamps, and can be structured over columns of respective data types. Temporal inconsistency can be identified by comparing a probability distribution of time differences between the timestamps of the log entries with a probability distribution of time differences between the timestamps of the baseline log entries. Data type inconsistency can be identified by comparing a data type of each column of the log entries with a data type of a corresponding column of the baseline log entries. Columnar inconsistency can be identified by comparing a number of the columns of the log entries with a number of the columns of the baseline log entries. In response to identification of temporal, data type, and/or columnar inconsistency, that an abnormality exists in collecting the log entries is detected.
    Type: Grant
    Filed: July 29, 2021
    Date of Patent: July 9, 2024
    Assignee: Micro Focus LLC
    Inventors: Manish Marwah, Martin Arlitt
  • Patent number: 12014206
    Abstract: A method includes receiving, by a first stage in a pipeline, a first transaction from a previous stage in pipeline; in response to first transaction comprising a high priority transaction, processing high priority transaction by sending high priority transaction to a buffer; receiving a second transaction from previous stage; in response to second transaction comprising a low priority transaction, processing low priority transaction by monitoring a full signal from buffer while sending low priority transaction to buffer; in response to full signal asserted and no high priority transaction being available from previous stage, pausing processing of low priority transaction; in response to full signal asserted and a high priority transaction being available from previous stage, stopping processing of low priority transaction and processing high priority transaction; and in response to full signal being de-asserted, processing low priority transaction by sending low priority transaction to buffer.
    Type: Grant
    Filed: October 3, 2022
    Date of Patent: June 18, 2024
    Assignee: Texas Instruments Incorporated
    Inventors: Abhijeet Ashok Chachad, David Matthew Thompson
  • Patent number: 11977486
    Abstract: A computer system includes a processor core and a memory system in signal communication with the processor core. The memory system includes a first cache and a second cache. The first cache is arranged at a first level of a hierarchy in the memory system and is configured to store a plurality of first-cache entries. The second cache is arranged at a second level of the hierarchy that is lower than the first level, and stores a plurality of second-cache entries. The first cache maintains a directory that contains information for each of the first-cache entries. The second cache maintains a shadow pointer directory (SPD) that includes one or more SPD entries that maps each of the first-cache entries to a corresponding second cache entry at a lower-level cache location.
    Type: Grant
    Filed: April 4, 2022
    Date of Patent: May 7, 2024
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Ashraf ElSharif, Richard Joseph Branciforte, Gregory William Alexander, Deanna Postles Dunn Berger, Timothy Bronson, Aaron Tsai, Taylor J. Pritchard, Markus Kaltenbach, Christian Jacobi, Michael A. Blake
  • Patent number: 11971855
    Abstract: Methods, apparatus, and processor-readable storage media for supporting multiple operations in transaction logging for a cloud enabled file system are provided herein. An example computer-implemented method includes obtaining a plurality of file system operations to be performed on a cloud enabled file system; executing the plurality of file system operations as a single file system transaction; and maintaining a transaction log for the single transaction, the transaction log comprising information for one or more sub-transactions that were completed in conjunction with said executing, wherein the one or more sub-transactions correspond to at least a portion of the plurality of file system operations.
    Type: Grant
    Filed: May 19, 2020
    Date of Patent: April 30, 2024
    Assignee: EMC IP Holding Company LLC
    Inventor: Priyamrita Ghosh
  • Patent number: 11972034
    Abstract: A computer system and associated methods are disclosed for mitigating side-channel attacks using a shared cache. The computer system includes a host having a main memory and a shared cache. The host executes a virtual machine manager (VMM) that determines respective security keys for a plurality of co-located virtual machines (VMs). A cache controller for the shared cache includes a scrambling function that scrambles addresses of memory accesses performed by threads of the VMs according to the respective security keys. Different cache tiers may implement different scrambling functions optimized to the architecture of each cache tier. Security keys may be periodically updated to further reduce predictability of shared cache to memory address mappings.
    Type: Grant
    Filed: October 29, 2020
    Date of Patent: April 30, 2024
    Assignee: Amazon Technologies, Inc.
    Inventors: Martin Pohlack, Pawel Wieczorkiewicz, Uwe Dannowski
  • Patent number: 11966385
    Abstract: In various examples, there is provided a computer-implemented method for writing transaction log entries to a transaction log for a database system. At least part of the database system is configured to be executed within a trusted execution environment. The transaction log is stored outside of the trusted execution environment. The method maintains a first secure count representing a number of transaction log entries which have been written to the transaction log for transactions which have been committed to the database and writes a transaction log entry to the transaction log. In other examples, there is also provided is a computer-implemented method for restoring a database system using transaction log entries received from the transaction log and a current value of the first secure count.
    Type: Grant
    Filed: August 25, 2021
    Date of Patent: April 23, 2024
    Assignee: Microsoft Technology Licensing, LLC.
    Inventors: Christian Priebe, Kapil Vaswani, Manuel Silverio da Silva Costa
  • Patent number: 11966398
    Abstract: A method for storing video data includes, when receiving the I-frame data to be stored, detecting whether the written data exists in the video cache space; when detecting that the written data exists in the video cache space, reading a target writing position of the I-frame data to be stored and determining whether the target writing position is located within a position range corresponding to the written data in the first cache space; when determining the target writing position is located within the position range, writing, based on the target writing position, the I-frame data to be stored to the first cache space for caching and detecting whether the first cache space is full; and when detecting that the first cache space is full, writing all the video data in the video cache space to a memory space of the terminal device for storage and emptying the video cache space.
    Type: Grant
    Filed: October 11, 2019
    Date of Patent: April 23, 2024
    Assignee: ZHEJIANG UNIVIEW TECHNOLOGIES CO., LTD.
    Inventors: Zuohua Wu, Qiang Ding
  • Patent number: 11954022
    Abstract: Provided are a storage device, system, and method for throttling host writes in a host buffer to a storage device. The storage device is coupled to a host system having a host buffer that includes reads and writes to pages of the storage device. Garbage collection consolidates valid data from pages in the storage device to fewer pages. A determination is made as to whether a processing measurement at the storage device satisfies a threshold. A timer value is set to a positive value in response to determining that the processing measurement satisfies the threshold. The timer is started to run for the timer value. Writes from the host buffer are blocked while the timer is running. Writes remain in the host buffer while the timer is running. A write is accepted from the host buffer to process in response to expiration of the timer.
    Type: Grant
    Filed: March 2, 2022
    Date of Patent: April 9, 2024
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Matthew S. Reuter, Timothy J. Fisher, Aaron Daniel Fry, Jenny L. Brown, John Carrington Cates, Austin Eberle
  • Patent number: 11954028
    Abstract: There is disclosed a method of storing an encoded block of data in memory comprising encoding a block of data elements and determining a memory location (26) at which the encoded block of data is to be stored. The memory location (26) at which the encoded block of data is stored is then indicated in a header (406) for the encoded block of data by including in the header a memory address value (407) together with a modifier value (500) representing a modifier that is to be applied to the memory address value (407) when determining the memory location (26). When the encoded block of data is to be retrieved, the header (406) is read and processed to determine the memory location (26).
    Type: Grant
    Filed: March 31, 2022
    Date of Patent: April 9, 2024
    Assignee: Arm Limited
    Inventors: Edvard Fielding, Jian Wang, Jakob Axel Fries, Carmelo Giliberto
  • Patent number: 11947581
    Abstract: A plurality of personalized news feeds are generated from input feeds including digital content items based on a dynamic taxonomy data structure. Entities are extracted from the input feeds and relationship strengths are obtained for the extracted entities and the digital content items. The dynamic taxonomy data structure is updated with the extracted entities and entries for the digital content news items are included at the corresponding branches based on the relationship strengths. Attributes are obtained for the entities and those entities corresponding to the trending topics are identified. Personalized news feeds are generated including the digital content items listed under the entities. Digital content items are added or removed from the digital content feeds based on one or more entity attributes.
    Type: Grant
    Filed: May 18, 2021
    Date of Patent: April 2, 2024
    Assignee: ACCENTURE GLOBAL SOLUTIONS LIMITED
    Inventors: Srikanth G Rao, Tarun Singhal, Mathangi Sandilya, Issac Abraham Alummoottil, Raja Sekhar Velagapudi, Rahel James Kale, Ankur Garg, Jayaprakash Nooji Shekar, Omkar Sudhakar Deorukhkar, Veera Raghavan Valayaputhur
  • Patent number: 11940911
    Abstract: Techniques are provided for implementing a persistent key-value store for caching client data, journaling, and/or crash recovery. The persistent key-value store may be hosted as a primary cache that provides read and write access to key-value record pairs stored within the persistent key-value store. The key-value record pairs are stored within multiple chains in the persistent key-value store. Journaling is provided for the persistent key-value store such that incoming key-value record pairs are stored within active chains, and data within frozen chains is written in a distributed manner across distributed storage of a distributed cluster of nodes. If there is a failure within the distributed cluster of nodes, then the persistent key-value store may be reconstructed and used for crash recovery.
    Type: Grant
    Filed: December 17, 2021
    Date of Patent: March 26, 2024
    Assignee: NetApp, Inc.
    Inventors: Sudheer Kumar Vavilapalli, Asif Imtiyaz Pathan, Parag Sarfare, Nikhil Mattankot, Stephen Wu, Amit Borase
  • Patent number: 11934307
    Abstract: An apparatus and method are provided for receiving a request from a plurality of processing units, where multiple of those processing units have associated cache storage. A snoop unit is used to implement a cache coherency protocol when a request is received that identifies a cacheable memory address. The snoop unit has snoop filter storage comprising a plurality of snoop filter tables organized in a hierarchical arrangement. The snoop filter tables comprise a primary snoop filter table at a highest level in the hierarchy, and each snoop filter table at a lower level in the hierarchy forms a backup snoop filter table for an adjacent snoop filter table at a higher level in the hierarchy. Each snoop filter table is arranged as a multi-way set associative storage structure, and each backup snoop filter table has a different number of sets than are provided in the adjacent snoop filter table.
    Type: Grant
    Filed: January 18, 2021
    Date of Patent: March 19, 2024
    Assignee: Arm Limited
    Inventors: Joshua Randall, Jesse Garrett Beu
  • Patent number: 11937164
    Abstract: A method for processing a data packet at a node in a Bluetooth Mesh network, comprising: (a) determining a one-hop device cache list of the node, wherein the one-hop device cache list comprises an address of one or more one-hop nodes; (b) when the node sends a data packet, checking whether a destination address of the data packet is the same as an address stored in the one-hop device cache list; if yes, setting a TTL value of the data packet to 0 and sending the data packet; otherwise, setting the TTL value of the data packet to be greater than a specified TTL threshold, and sending the data packet; and (c) when the node forwards a data packet, checking whether the destination address of the data packet is the same as an address stored in the one-hop device cache list; if yes, setting the TTL value of the data packet to 1 and forwarding the data packet; otherwise, deducting the TTL value of the data packet by 1 and forwarding the data packet.
    Type: Grant
    Filed: February 25, 2020
    Date of Patent: March 19, 2024
    Assignee: ESPRESSIF SYSTEMS (SHANGHAI) CO., LTD.
    Inventors: Yizan Zhou, Swee Ann Teo
  • Patent number: 11928352
    Abstract: Systems and methods are described for performing persistent inflight tracking of operations (Ops) within a cross-site storage solution. According to one embodiment, a method comprises maintaining state information regarding a data synchronous replication status for a first storage object of a primary storage cluster and a second storage object of a secondary storage cluster. The state information facilitates automatic triggering of resynchronization for data replication between the first storage object and the second storage object. The method includes performing persistent inflight tracking of I/O operations with a first Op log of the primary storage cluster and a second Op log of the secondary storage cluster, establishing and comparing Op ranges for the first and second Op logs, and determining a relation between the Op range of the first Op log and the Op range of the second Op log to prevent divergence of Ops in the first and second Op logs and to support parallel split of the Ops.
    Type: Grant
    Filed: October 26, 2021
    Date of Patent: March 12, 2024
    Assignee: NetApp, Inc.
    Inventors: Krishna Murthy Chandraiah Setty Narasingarayanapeta, Preetham Shenoy, Divya Kathiresan, Rakesh Bhargava
  • Patent number: 11908546
    Abstract: A system includes a plurality of host processors and a plurality of HMC devices configured as a distributed shared memory for the host processors. An HMC device includes a plurality of integrated circuit memory die including at least a first memory die arranged on top of a second memory die and at least a portion of the memory of the memory die is mapped to include at least a portion of a memory coherence directory; and a logic base die including at least one memory controller configured to manage three-dimensional (3D) access to memory of the plurality of memory die by at least one second device, and logic circuitry configured to determine memory coherence state information for data stored in the memory of the plurality of memory die, communicate information regarding the access to memory, and include the memory coherence information in the communicated information.
    Type: Grant
    Filed: October 8, 2020
    Date of Patent: February 20, 2024
    Assignee: Micron Technology, Inc.
    Inventor: Richard C Murphy
  • Patent number: 11907528
    Abstract: Techniques for loading data, comprising receiving a memory management command to perform a memory management operation to load data into the cache memory before execution of an instruction that requests the data, formatting the memory management command into one or more instruction for a cache controller associated with the cache memory, and outputting an instruction to the cache controller to load the data into the cache memory based on the memory management command.
    Type: Grant
    Filed: July 20, 2021
    Date of Patent: February 20, 2024
    Assignee: Texas Instruments Incorporated
    Inventors: Kai Chirca, Daniel Wu, Matthew David Pierson
  • Patent number: 11899937
    Abstract: Systems and methods of a memory allocation buffer to reduce heap fragmentation. In one embodiment, the memory allocation buffer structures a memory arena dedicated to a target region that is one of a plurality of regions in a server in a database cluster such as an HBase cluster. The memory area has a chunk size (e.g., 2 MB) and an offset pointer. Data objects in write requests targeted to the region are received and inserted to the memory arena at a location specified by the offset pointer. When the memory arena is filled, a new one is allocated. When a MemStore of the target region is flushed, the entire memory arenas for the target region are freed up. This reduces heap fragmentation that is responsible for long and/or frequent garbage collection pauses.
    Type: Grant
    Filed: March 3, 2020
    Date of Patent: February 13, 2024
    Assignee: Cloudera, Inc.
    Inventor: Todd Lipcon
  • Patent number: 11892955
    Abstract: System and method for analyzing CXL flits at read bypass detection logic to identify bypass memory read requests and transmitting the identified bypass memory read requests over a read request bypass path directly to a transaction/application layer of the CXL memory controller, wherein the read request bypass path does not include an arbitration/multiplexing layer and a link layer of the CXL memory controller, thereby reducing the latency inherent in a CXL memory controller.
    Type: Grant
    Filed: May 10, 2022
    Date of Patent: February 6, 2024
    Assignee: Microchip Technology Inc.
    Inventors: Sanjay Goyal, Larrie Simon Carr, Patrick Bailey