Patents by Inventor Yakov Tokar

Yakov Tokar has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11178054
    Abstract: A network device includes a memory configured to store a plurality of entries in respective locations in the memory, the plurality of entries corresponding to a trie data structure for performing a longest prefix match search. The network device also includes: a memory access engine configured to retrieve from a location in the memory, in a single memory lookup operation, i) longest prefix match information for a node corresponding to a network address in a header of a packet, and ii) pointer information that indicates a child node in the trie data structure. The network device also includes: a child node address calculator configured to use i) the longest prefix match information, and ii) the pointer information, to calculate a memory address of another location in the memory corresponding to the child node.
    Type: Grant
    Filed: August 22, 2019
    Date of Patent: November 16, 2021
    Assignee: Marvell Israel (M.I.S.L) Ltd.
    Inventors: Ziv Zamsky, Ilan Mayer-Wolf, Yakov Tokar
  • Patent number: 11159148
    Abstract: A first-in/first-out (FIFO) buffer includes at least one latch-based FIFO storage line, an input flip-flop stage upstream of the at least one latch-based storage line, an output flip-flop stage downstream of the at least one latch-based storage line. The output flip-flop stage functions as an additional storage line. Clock-gating circuitry separate from the device clock controls timing of the at least one latch-based FIFO storage line, the input flip-flop stage, and the output flip-flop stage. The input flip-flop stage functions as a second additional storage line, or as an input sampling stage. Optional bypass circuitry between the input flip-flop stage and the output flip-flop stage passes data for a storage line directly to the output flip-flop stage, without passing through the at least one latch-based storage line, when the buffer is empty.
    Type: Grant
    Filed: January 27, 2020
    Date of Patent: October 26, 2021
    Assignee: Marvell Israel (M.I.S.L) Ltd.
    Inventors: Lior Moheban, Alex Pinskiy, Yakov Tokar
  • Patent number: 10621122
    Abstract: Embodiments described herein provide a dual-line FIFO structure without the use of any multiplexer. Instead, the dual-line FIFO described herein uses a selectively transparent latch and a flip-flop serially connected to the latch, such that the combination of the serially connected latch and the flip-flop can temporarily store up to two data units at two clock cycles.
    Type: Grant
    Filed: May 14, 2018
    Date of Patent: April 14, 2020
    Assignee: Marvell Israel (M.I.S.L) Ltd.
    Inventors: Lior Moheban, Ronen Goldberg, Yakov Tokar, Gregory Kovishaner, Alex Pinskiy
  • Patent number: 9436624
    Abstract: A circuitry for a computing system comprising a first load/store unit, LSU, and a second LSU as well as a memory arrangement. The first LSU is connected to the memory arrangement via a first bus arrangement comprising a first write bus and a first read bus. The second LSU is connected to the memory arrangement via a second bus arrangement comprising a second write bus and a second read bus. The computing system is arranged to carry out a multiple load instruction to read data via the first read bus and the second read bus and/or to carry out a multiple store instruction to write data via the first write bus and the second write bus.
    Type: Grant
    Filed: July 26, 2013
    Date of Patent: September 6, 2016
    Assignee: FREESCALE SEMICONDUCTOR, INC.
    Inventors: Ziv Zamsky, Moshe Anschel, Itay Keidar, Itay S. Peled, Doron Schupper, Yakov Tokar
  • Publication number: 20150032929
    Abstract: A circuitry for a computing system comprising a first load/store unit, LSU, and a second LSU as well as a memory arrangement. The first LSU is connected to the memory arrangement via a first bus arrangement comprising a first write bus and a first read bus. The second LSU is connected to the memory arrangement via a second bus arrangement comprising a second write bus and a second read bus. The computing system is arranged to carry out a multiple load instruction to read data via the first read bus and the second read bus and/or to carry out a multiple store instruction to write data via the first write bus and the second write bus.
    Type: Application
    Filed: July 26, 2013
    Publication date: January 29, 2015
    Inventors: ZIV ZAMSKY, MOSHE ANSCHEL, ITAY KEIDAR, ITAY S. PELED, DORON SCHUPPER, YAKOV TOKAR
  • Patent number: 8041899
    Abstract: A write back allocate system that includes: (i) a store request circuit; (ii) a processor, adapted to generate a store request that comprises an information unit and an information unit address; and (iii) a cache module, connected to the store request circuit and to a high level memory unit. A single cache module line includes multiple segments, each segment is adapted to store a single information unit. A content of a cache module line is retrieved from the high level memory unit by generating a fetch burst that includes multiple segment fetch operations. The store request circuit includes a snooper and a controller. The snooper detects a portion of an address of a cache segment of a cache line that is being fetched during a fetch burst. The controller is adapted to request from the cache module to receive the information unit before a completion of the fetch burst if the portion of the address of the cache segment matches a corresponding portion of the information unit address.
    Type: Grant
    Filed: July 29, 2008
    Date of Patent: October 18, 2011
    Assignee: Freescale Semiconductor, Inc.
    Inventors: Kostantin Godin, Roman Landa, Itay Peled, Yakov Tokar, Ziv Zamsky
  • Patent number: 7865691
    Abstract: A virtual address cache and a method for sharing data. The virtual address cache includes: a memory, adapted to store virtual addresses, task identifiers and data associated with the virtual addresses and the task identifiers; and a comparator, connected to the memory, adapted to determine that data associated with a received virtual address and a received task identifier is stored in the memory if at least a portion of the received virtual address equals at least a corresponding portion of a certain stored virtual address and a stored task identifier associated with the certain stored virtual address indicates that the data is shared between multiple tasks.
    Type: Grant
    Filed: August 31, 2004
    Date of Patent: January 4, 2011
    Assignee: Freescale Semiconductor, Inc.
    Inventors: Itay Peled, Moshe Anschel, Moshe Bachar, Jacob Efrat, Alon Eldar, Yakov Tokar
  • Publication number: 20100030974
    Abstract: A write back allocate system that includes: (i) a store request circuit; (ii) a processor, adapted to generate a store request that comprises an information unit and an information unit address; and (iii) a cache module, connected to the store request circuit and to a high level memory unit. A single cache module line includes multiple segments, each segment is adapted to store a single information unit. A content of a cache module line is retrieved from the high level memory unit by generating a fetch burst that includes multiple segment fetch operations. The store request circuit includes a snooper and a controller. The snooper detects a portion of an address of a cache segment of a cache line that is being fetched during a fetch burst. The controller is adapted to request from the cache module to receive the information unit before a completion of the fetch burst if the portion of the address of the cache segment matches a corresponding portion of the information unit address.
    Type: Application
    Filed: July 29, 2008
    Publication date: February 4, 2010
    Inventors: Kostantin Godin, Roman Landa, Itay Peled, Yakov Tokar, Ziv Zamsky
  • Publication number: 20070266199
    Abstract: A virtual address cache comprising a comparator arranged to receive a virtual address for addressing data associated with a task and a memory, wherein the comparator is arranged to make a determination as to whether data associated with the received virtual address is stored in the memory based upon an indication that the virtual address is associated with data shared between a first task and a second task and a comparison of the received virtual address with an address associated with data stored in memory.
    Type: Application
    Filed: September 7, 2004
    Publication date: November 15, 2007
    Applicant: Freescale Semiconductor, Inc.
    Inventors: Itay Peled, Moshe Anschel, Moshe Bachar, Jacob Efrat, Alon Eldar, Yakov Tokar
  • Publication number: 20050246498
    Abstract: Read/write conflicts in an instruction cache memory (11) are reduced by configuring the memory as two even and odd array sub-blocks (12,13) and adding an input buffer (10) between the memory (11) and an update (16). Contentions between a memory read and a memory write are minimised by the buffer (10) shifting the update sequence with respect to the read sequence. The invention can adapt itself for use in digital signal processing systems with different external memory behaviour as far as latency and burst capability is concerned.
    Type: Application
    Filed: March 3, 2003
    Publication date: November 3, 2005
    Inventors: Doron Schupper, Yakov Tokar, Jacob Efrat
  • Patent number: 6848030
    Abstract: A processing system has a processor, a cache, and a fetch unit. If there is a miss in the cache, the fetch unit generates a fetch address for the miss in the cache for the purpose of retrieving the requested data from external memory and providing the data to the processor and loading the data in a location in a line in the cache. The fetch unit also generates additional prefetch addresses for addresses consecutive with the fetch address. The prefetch addresses continue to be generated for all of the locations in the line in the cache that are consecutive with the fetch address. The generation of prefetch addresses will be stopped if another requests arrives that is not part of the already generated prefetched addresses. Further, the outstanding prefetches will be terminated if the external memory can handle such termination.
    Type: Grant
    Filed: July 20, 2001
    Date of Patent: January 25, 2005
    Assignee: Freescale Semiconductor, Inc.
    Inventors: Yakov Tokar, Amit Gur, Jacob Efrat, Doron Schupper
  • Publication number: 20030041213
    Abstract: A cache is used in the performance of one task that may be interrupted by another task. The first task results in the cache being loaded at least partially. The second task interrupts, but is preventing from thrashing the highest priority data. The highest priority data is not available for thrashing during the running of the second task. The second task may be interrupted as well. Similarly, the third task is prevented from thrashing the highest priority data of the second task and the first task. The third task can thrash all of the cache except that preserved for the first and second tasks. After the third task is completed, the second task can begin running again without having to reload the highest priority data. The first task is similarly completed.
    Type: Application
    Filed: August 24, 2001
    Publication date: February 27, 2003
    Inventors: Yakov Tokar, Yacov Efrat, Doron Schupper, Brett L. Lindsley
  • Publication number: 20030018853
    Abstract: A processing system has a processor, a cache, and a fetch unit. If there is a miss in the cache, the fetch unit generates a fetch address for the miss in the cache for the purpose of retrieving the requested data from external memory and providing the data to the processor and loading the data in a location in a line in the cache. The fetch unit also generates additional prefetch addresses for addresses consecutive with the fetch address. The prefetch addresses continue to be generated for all of the locations in the line in the cache that are consecutive with the fetch address. The generation of prefetch addresses will be stopped if another requests arrives that is not part of the already generated prefetched addresses. Further, the outstanding prefetches will be terminated if the external memory can handle such termination.
    Type: Application
    Filed: July 20, 2001
    Publication date: January 23, 2003
    Inventors: Yakov Tokar, Amit Gur, Jacob Efrat, Doron Schupper