Patents Examined by Richard B. Franklin
-
Patent number: 11960755Abstract: A semiconductor memory device comprises: first storage logic configured to store, as first addresses, âKâ addresses having different values among input addresses applied during the enable period of a reference signal, second storage logic configured to store, as second addresses, âLâ addresses corresponding to a time point at which the enable period of the reference signal is ended among the input addresses, an order controller configured to determine a first output order of each of the first addresses based on a number of times each of the first addresses is repeatedly input, and to determine a second output order for outputting mixed addresses obtained by mixing the first addresses based on the first output order and the second addresses together, and refresh operation logic configured to apply the mixed addresses according to the second output order, to a target refresh operation.Type: GrantFiled: December 13, 2021Date of Patent: April 16, 2024Assignee: SK hynix Inc.Inventors: Woongrae Kim, Kwi Dong Kim, Chul Moon Jung, Jeong Tae Hwang
-
Patent number: 11947840Abstract: Described systems, apparatuses, and methods relate to volatile memories that are refreshed to maintain data integrity, such as dynamic random-access memory (DRAM) and synchronous DRAM (SDRAM). A memory device includes multiple dies, with each die having a memory array to be refreshed. The multiple dies may be interconnected via at least one inter-die bus of the memory device. A memory controller sends a command to the memory device to enter a self-refresh mode. In response, a die of the multiple dies can enter the self-refresh mode and initiate or otherwise coordinate refresh operations of the other dies. To do so, the die may transmit at least one refresh-related command to at least one other die using the inter-die bus. Multiple different signaling schemes and timing approaches are disclosed. The described inter-die refresh control principles may be implemented in energy-efficient applications, such as in low-power double data rate (LPDDR) SDRAM.Type: GrantFiled: October 28, 2021Date of Patent: April 2, 2024Assignee: Micron Technology, Inc.Inventors: Kang-Yong Kim, Hyun Yoo Lee
-
Patent number: 11947833Abstract: A method and apparatus for training data in a computer system includes reading data stored in a first memory address in a memory and writing it to a buffer. Training data is generated for transmission to the first memory address. The data is transmitted to the first memory address. Information relating to the training data is read from the first memory address and the stored data is read from the buffer and written to the memory area where the training data was transmitted.Type: GrantFiled: June 21, 2022Date of Patent: April 2, 2024Assignee: Advanced Micro Devices, Inc.Inventors: Anwar Kashem, Craig Daniel Eaton, Pouya Najafi Ashtiani, Tsun Ho Liu
-
Patent number: 11934673Abstract: A data storage device includes at least one data storage medium. The data storage device also includes a workload rating associated with data access operations carried out on the at least one data storage medium. The data storage device further includes a controller configured to enable performance of the data access operations, and change a rate of consumption of the workload rating by internal device management operations carried out in the data storage device in response to a change in a workload consumed by host commands serviced by the data storage device.Type: GrantFiled: August 11, 2022Date of Patent: March 19, 2024Assignee: Seagate Technology LLCInventors: Abhay T. Kataria, Praveen Viraraghavan, Mark A. Gaertner
-
Patent number: 11934654Abstract: An integrated circuit (IC) memory device includes an array of storage cells configured into multiple banks. Interface circuitry receives refresh commands from a host memory controller to refresh the multiple banks for a first refresh mode. On-die refresh control circuitry selectively generates local refresh commands to refresh the multiple banks in cooperation with the host memory controller during a designated hidden refresh interval in a second refresh mode. Mode register circuitry stores a value indicating whether the on-die refresh control circuitry is enabled for use during the second refresh mode. The interface circuitry includes backchannel control circuitry to transmit a corrective action control signal during operation in the second refresh mode.Type: GrantFiled: December 7, 2021Date of Patent: March 19, 2024Assignee: Rambus Inc.Inventors: Michael Raymond Miller, Steven C. Woo, Thomas Vogelsang
-
Patent number: 11934684Abstract: A data storage device includes a memory device and a controller coupled to the memory device. The controller is configured to determine a maximum bandwidth of an interface, allocate a portion of the maximum bandwidth to one or more tenants, either: determine a maximum data transfer size (MDTS) setting based on quality of service (QoS) requirements, determine an aggregated queue depth (QD) setting based on QoS requirements, or determine a combined MDTS and aggregated QD setting based on QoS requirements, and provide the determined settings to the one or more tenants.Type: GrantFiled: December 14, 2021Date of Patent: March 19, 2024Assignee: Western Digital Technologies, Inc.Inventors: Amir Segev, Shay Benisty
-
Patent number: 11928062Abstract: Some embodiments provide a method for performing data message processing at a smart NIC of a computer that executes a software forwarding element (SFE). The method determines whether a received data message matches an entry in a data message classification cache stored on the smart NIC based on data message classification results of the SFE. When the data message matches an entry, the method determines whether the matched entry is valid by comparing a timestamp of the entry to a set of rules stored on the smart NIC. When the matched entry is valid, the method processes the data message according to the matched entry without providing the data message to the SFE executing on the computer.Type: GrantFiled: June 21, 2022Date of Patent: March 12, 2024Assignee: VMware LLCInventors: Shay Vargaftik, Alex Markuze, Yaniv Ben-Itzhak, Igor Golikov, Avishay Yanai
-
Patent number: 11928920Abstract: A server system for electronic games includes a memory and a processor configured to execute instructions stored in the memory. When the instructions are executed, the instructions cause the processor to receive from a communication device, a plurality of first signals generated in response to the communication device entering one or more predefined zones associated with the electronic games, and generate, based on the plurality of first signals, a heat map that defines one or more cells based upon a magnitude of a data element of the first signals.Type: GrantFiled: January 13, 2023Date of Patent: March 12, 2024Assignee: VIDEO GAMING TECHNOLOGIES, INC.Inventors: Ryan Christopher Johnson, Lawrence Acosta Hysler, III
-
Patent number: 11914902Abstract: Shared memory access in a distributed system, including: receiving a memory access request associated with a time value; determining, based on the time value, an entry in a translation lookaside buffer (TLB); and determining, based on the entry, whether to allow the memory access request.Type: GrantFiled: February 28, 2023Date of Patent: February 27, 2024Assignee: GHOST AUTONOMY INC.Inventors: John Hayes, Volkmar Uhlig, Richard A. Swetz, Daniel P. Potts, Aaron Carroll
-
Patent number: 11892957Abstract: A system is disclosed. An upstream interface enables communication with a processor; a downstream interface enables communication with a storage device. The system may also include an acceleration module implemented using hardware to execute an acceleration instruction. The storage device may include an endpoint of the storage device for communicating with the acceleration module, a controller to manage operations of the storage device, storage for data, and a storage device acceleration module to assist the acceleration module in executing the acceleration instruction.Type: GrantFiled: February 26, 2021Date of Patent: February 6, 2024Inventors: Ramdas P. Kachare, Fred Worley, Harry Rogers, Wentao Wu, Nagarajan Subramaniyan
-
Patent number: 11886357Abstract: A memory includes: a control chip; and a plurality of storage chips, in which the plurality of storage chips are electrically connected with the control chip via a common communication channel, the plurality of storage chips include a first storage chip set and a second storage chip set, the storage chips in the first storage chip set are configured to perform information interaction with the control chip by adopting a first clock signal, the storage chips in the second storage chip set are configured to perform information interaction with the control chip by adopting a second clock signal, and phase of the first clock signal is different from phase of the second clock signal.Type: GrantFiled: August 23, 2021Date of Patent: January 30, 2024Assignee: CHANGXIN MEMORY TECHNOLOGIES, INC.Inventors: Shu-Liang Ning, Jun He, Zhan Ying, Jie Liu
-
Patent number: 11880329Abstract: The present application relates to a network-on-chip data processing method. The method is applied to a network-on-chip processing system, the network-on-chip processing system is used for executing machine learning calculation, and the network-on-chip processing system comprises a storage device and a calculation device. The method comprises: accessing the storage device in the network-on-chip processing system by means of a first calculation device in the network-on-chip processing system and obtaining first operation data; performing an operation on the first operation data by means of the first calculation device to obtain a first operation result; and sending the first operation result to a second calculation device in the network-on-chip processing system. According to the method, operation overhead can be reduced and data read/write efficiency can be improved.Type: GrantFiled: December 29, 2021Date of Patent: January 23, 2024Assignee: SHANGHAI CAMBRICON INFORMATION TECHNOLOGY CO., LTD.Inventors: Shaoli Liu, Zhen Li, Yao Zhang
-
Patent number: 11880328Abstract: The present application relates to a network-on-chip data processing method. The method is applied to a network-on-chip processing system, the network-on-chip processing system is used for executing machine learning calculation, and the network-on-chip processing system comprises a storage device and a calculation device. The method comprises: accessing the storage device in the network-on-chip processing system by means of a first calculation device in the network-on-chip processing system, and obtaining first operation data; performing an operation on the first operation data by means of the first calculation device to obtain a first operation result; and sending the first operation result to a second calculation device in the network-on-chip processing system. According to the method, operation overhead can be reduced and data read/write efficiency can be improved.Type: GrantFiled: December 29, 2021Date of Patent: January 23, 2024Assignee: SHANGHAI CAMBRICON INFORMATION TECHNOLOGY CO., LTD.Inventors: Yao Zhang, Shaoli Liu, Jun Liang, Yu Chen
-
Patent number: 11880330Abstract: The present application relates to a network-on-chip data processing method. The method is applied to a network-on-chip processing system, the network-on-chip processing system is used for executing machine learning calculation, and the network-on-chip processing system comprises a storage device and a calculation device. The method comprises: accessing the storage device in the network-on-chip processing system by means of a first calculation device in the network-on-chip processing system and obtaining first operation data; performing an operation on the first operation data by means of the first calculation device to obtain a first operation result; and sending the first operation result to a second calculation device in the network-on-chip processing system. According to the method, operation overhead can be reduced and data read/write efficiency can be improved.Type: GrantFiled: December 29, 2021Date of Patent: January 23, 2024Assignee: SHANGHAI CAMBRICON INFORMATION TECHNOLOGY CO., LTD.Inventors: Shaoli Liu, Zhen Li, Yao Zhang
-
Patent number: 11880301Abstract: Techniques for enabling efficient guest OS access to PCIe configuration space are provided. In one set of embodiments, a hypervisor can reserve a single host physical memory page in the host physical memory of a host system and can populate the single host physical memory page with a value indicating non-presence of PCIe device functions. The hypervisor can then create, for each guest physical memory page in a guest physical memory of a virtual machine (VM) corresponding to a PCIe configuration space of an absent PCIe device function in the VM, a mapping in the hypervisor's second-level page tables that maps the guest physical memory page to the single host physical memory page.Type: GrantFiled: December 21, 2022Date of Patent: January 23, 2024Assignee: VMware LLCInventors: Andrei Warkentin, Alexander Fainkichen, Ye Li, Regis Duchesne, Cyprien Laplace, Shruthi Hiriyuru, Sunil Kotian
-
Patent number: 11874784Abstract: A memory device of a memory module includes a CA buffer that receives a command/address (CA) signal through a bus shared by a memory device different from the memory device of the memory module, and a calibration logic circuit that identifies location information of the memory device on the bus. The memory device recognizes its own location on a bus in a memory module to perform self-calibration, and thus, the memory device appropriately operates even under an operation condition varying depending on a location in the memory module.Type: GrantFiled: December 27, 2022Date of Patent: January 16, 2024Assignee: SAMSUNG ELECTRONICS CO., LTD.Inventors: Heon Su Jeong, Hangi Jung, Wangsoo Kim, Hae Young Chung
-
Patent number: 11876702Abstract: A network interface controller (NIC) capable of facilitating efficient memory address translation is provided. The NIC can be equipped with a host interface, a cache, and an address translation unit (ATU). During operation, the ATU can determine an operating mode. The operating mode can indicate whether the ATU is to perform a memory address translation at the NIC. The ATU can then determine whether a memory address indicated in the memory access request is available in the cache. If the memory address is not available in the cache, the ATU can perform an operation on the memory address based on the operating mode.Type: GrantFiled: March 23, 2020Date of Patent: January 16, 2024Assignee: Hewlett Packard Enterprise Development LPInventors: Abdulla M. Bataineh, Thomas L. Court, Hess M. Hodge
-
Patent number: 11868654Abstract: A semiconductor device includes: a nonvolatile memory cell including first memory cells and second memory cells; a bit latch; and a saved register. In a first writing operation, first writing data are stored in the bit latch and the saved register, and writing to the first memory cells is executed based on the first writing data. During the first writing operation, the first writing operation is interrupted based on a suspension command, and a second writing operation is executed. In the second writing operation, second writing data are stored in the bit latch, and writing to the second memory cells is executed based on the second writing data. After the second writing operation is ended, the first writing data is reset to the bit latch based on a resume command, and the interrupted first writing operation is restarted based on the first writing data reset to the bit latch.Type: GrantFiled: May 12, 2021Date of Patent: January 9, 2024Assignee: RENESAS ELECTRONICS CORPORATIONInventors: Takanori Moriyasu, Kazuo Yoshihara, Takayuki Nishiyama
-
Patent number: 11869113Abstract: Apparatuses including general-purpose graphics processing units and graphics multiprocessors that exploit queues or transitional buffers for improved low-latency high-bandwidth on-die data retrieval are disclosed. In one embodiment, a graphics multiprocessor includes at least one compute engine to provide a request, a queue or transitional buffer, and logic coupled to the queue or transitional buffer. The logic is configured to cause a request to be transferred to a queue or transitional buffer for temporary storage without processing the request and to determine whether the queue or transitional buffer has a predetermined amount of storage capacity.Type: GrantFiled: December 7, 2021Date of Patent: January 9, 2024Assignee: Intel CorporationInventors: Aravindh Anantaraman, Altug Koker, Varghese George, Subramaniam Maiyuran, SungYe Kim, Valentin Andrei
-
Patent number: 11861227Abstract: A method of operating a storage device including a non-volatile memory and a multi-core processor with at least two cores includes the following steps: receiving, by a host interface of the storage device, a first command from a host for requesting the non-volatile memory to perform a predetermined memory operation; generating, by a task scheduler of the storage device, first and second tasks from the first memory command; selecting, by the task scheduler, a first core from among the at least two cores based on execution times of the at least two cores; assigning, by the task scheduler, the first and second tasks to the first core; and requesting, by the first core, a subsequent task from the scheduler while the first core processes the first assigned task and loads code for processing the second task.Type: GrantFiled: September 17, 2021Date of Patent: January 2, 2024Assignee: SAMSUNG ELECTRONICS CO., LTD.Inventors: Wan-Soo Choi, Young Wook Kim, Do Hyeon Park