Patents by Inventor Rojit Jacob
Rojit Jacob has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20200311217Abstract: A method and system for implementing memory changes in digital Integrated Circuits (ICs) includes a step of generating a plurality of memory wrappers based on a first library associated with a digital IC design requirement and a second library associated with a set of available memories. The method includes identifying at least one available memory from the set of available memories, for each of the plurality of memory wrappers, based on the associated width and depth requirement and width and depth details associated with each of the set of available memories. The method further includes managing port connections for the at least one available memory associated with each of the plurality of memory wrappers, based on the first library and the second library. The method includes validating each of the plurality of memory wrappers using a testbench generated for the digital IC design.Type: ApplicationFiled: March 29, 2019Publication date: October 1, 2020Inventors: JENTIL JOSE, ROJIT JACOB, SHOHIN KAKKOTH
-
Patent number: 9258565Abstract: A method and system are disclosed for managing cache memory in a dual-pipelined CABAC encoder. A request for a context model is received from both encoder pipelines. If the requested context model is not stored in cache, the requested context model is retrieved from a context table. At least one context model stored in cache is written to the context table. The retrieved context model is updated and written to the cache. If the requested context model is stored in cache, and if the requested context model was updated in the previous clock cycle, the requested context model is retrieved from the pipeline, updated, and written to cache. If the requested context model is not stored in cache, and if the requested contest model was not updated in the previous clock cycle, the requested context model retrieved from cache, updated, and written back to cache.Type: GrantFiled: June 29, 2011Date of Patent: February 9, 2016Assignee: FREESCALE SEMICONDUCTOR, INC.Inventor: Rojit Jacob
-
Patent number: 8798139Abstract: A method and system are disclosed for the lossless compression of video data in a synchronous pipelined environment. One or more syntax elements of video data are binarized into one or more ordered bins. A first context model associated with a first bin and a second context model associated with a second bin are received. The first bin is encoded based on the first context model and the second bin is encoded based on the second context model, both bins being encoded within the same clock cycle. One or more encoded bits are outputted based on encoding the first and second bin. In one embodiment, the first bin is encoded in a first pipeline and the second bin is encoded in a second pipeline. In this embodiment, two bins may be encoded every clock cycle, one per pipeline. Further, in one embodiment, multiple context models are received and one context model is selected by each pipeline for encoding. After encoding, one or more context models may be updated and stored.Type: GrantFiled: June 29, 2011Date of Patent: August 5, 2014Assignee: Zenverge, Inc.Inventor: Rojit Jacob
-
Patent number: 8411749Abstract: A system (and a method) are disclosed for intelligently fetch one or multiple reference blocks from memory for each block to be motion compensated or motion estimated within a video processing system. The system includes a reference block configuration evaluation unit and a motion compensation memory fetching unit. The reference block configuration evaluation unit analyzes the reference block configuration of the block being motion compensated with a plurality of reference block configurations of its neighboring blocks. In response to a reference block configuration evaluation result, the reference block configuration evaluation unit decides the configuration of reference blocks to be fetched from a memory. The motion vector memory fetching unit fetches the number of reference blocks for motion compensation accordingly.Type: GrantFiled: October 1, 2009Date of Patent: April 2, 2013Assignee: Zenverge, Inc.Inventors: Rojit Jacob, Alexander Kipnis, Behzad R. Sayyah
-
Patent number: 7865647Abstract: Resource requests are allocated by storing resource requests in a queue slots in a queue. A token is associated with one of the queue slots. During an arbitration cycle, the queue slot with the token is given the priority to the resource. If the queue slot with the token does not include a request, a different queue slot having the highest static priority and including a request is given access to the resource. The token is advanced to a different queue slot after one or more arbitration cycles. Requests are assigned to the highest priority queue slot, to random or arbitrarily selected queue slots, or based on the source and/or type of the request. One or more queue slots may be received for specific sources or types of requests. Resources include processor access, bus access, cache or system memory interface access, and internal or external interface access.Type: GrantFiled: December 27, 2006Date of Patent: January 4, 2011Assignee: MIPS Technologies, Inc.Inventor: Rojit Jacob
-
Patent number: 7840874Abstract: A cache tag comparison unit in a cache controller evaluates tag data and error correction codes to determine if there is a cache hit or miss. The cache tag comparison unit speculatively compares the tag data with the request tag without regard to error correction. The error correction code verifies whether this initial comparison is correct and provides a confirmed cache hit or miss signal. The tag data is compared with the request tag to determine a provisional cache hit or miss, and in parallel, the error correction code is evaluated. If the error code evaluation indicates errors in the tag data, a provisional cache hit is converted into a cache miss if errors are responsible for a false match. If the error code evaluation identifies the locations of errors, a provisional cache miss is converted into a cache hit if the errors are responsible for the mismatch.Type: GrantFiled: December 27, 2006Date of Patent: November 23, 2010Assignee: MIPS Technologies, Inc.Inventor: Rojit Jacob
-
Patent number: 7502915Abstract: The present invention provides an adaptive computing engine (ACE) that includes processing nodes having different capabilities such as arithmetic nodes, bit-manipulation nodes, finite state machine nodes, input/output nodes and a programmable scalar node (PSN). In accordance with one embodiment of the present invention, a common architecture is adaptable to function in either a kernel node, or k-node, or as general purpose RISC node. The k-node acts as a system controller responsible for adapting other nodes to perform selected functions. As a RISC node, the PSN is configured to perform computationally intensive applications such as signal processing.Type: GrantFiled: September 29, 2003Date of Patent: March 10, 2009Assignee: NVIDIA CorporationInventors: Rojit Jacob, Dan Minglun Chuang
-
Publication number: 20080162760Abstract: Resource requests are allocated by storing resource requests in a queue slots in a queue. A token is associated with one of the queue slots. During an arbitration cycle, the queue slot with the token is given the priority to the resource. If the queue slot with the token does not include a request, a different queue slot having the highest static priority and including a request is given access to the resource. The token is advanced to a different queue slot after one or more arbitration cycles. Requests are assigned to the highest priority queue slot, to random or arbitrarily selected queue slots, or based on the source and/or type of the request. One or more queue slots may be received for specific sources or types of requests. Resources include processor access, bus access, cache or system memory interface access, and internal or external interface access.Type: ApplicationFiled: December 27, 2006Publication date: July 3, 2008Inventor: Rojit Jacob
-
Publication number: 20080163008Abstract: A cache tag comparison unit in a cache controller evaluates tag data and error correction codes to determine if there is a cache hit or miss. The cache tag comparison unit speculatively compares the tag data with the request tag without regard to error correction. The error correction code verifies whether this initial comparison is correct and provides a confirmed cache hit or miss signal. The tag data is compared with the request tag to determine a provisional cache hit or miss, and in parallel, the error correction code is evaluated. If the error code evaluation indicates errors in the tag data, a provisional cache hit is converted into a cache miss if errors are responsible for a false match. If the error code evaluation identifies the locations of errors, a provisional cache miss is converted into a cache hit if the errors are responsible for the mismatch.Type: ApplicationFiled: December 27, 2006Publication date: July 3, 2008Inventor: Rojit Jacob
-
Patent number: 7194598Abstract: The present invention provides an adaptive computing engine (ACE) that includes processing nodes having different capabilities such as arithmetic nodes, bit-manipulation nodes, finite state machine nodes, input/output nodes and a programmable scalar node (PSN). In accordance with one embodiment of the present invention, a common architecture is adaptable to function in either a kernel node, or k-node, or as general purpose RISC node. The k-node acts as a system controller responsible for adapting other nodes to perform selected functions. As a RISC node, the PSN is configured to perform computationally intensive applications such as signal processing. The present invention further provides an interconnection scheme so that a plurality of ACE devices operates under the control of a single k-node.Type: GrantFiled: January 26, 2004Date of Patent: March 20, 2007Assignee: NVIDIA CorporationInventor: Rojit Jacob
-
Publication number: 20050166033Abstract: The present invention provides an adaptive computing engine (ACE) that includes processing nodes having different capabilities such as arithmetic nodes, bit-manipulation nodes, finite state machine nodes, input/output nodes and a programmable scalar node (PSN). In accordance with one embodiment of the present invention, a common architecture is adaptable to function in either a kernel node, or k-node, or as general purpose RISC node. The k-node acts as a system controller responsible for adapting other nodes to perform selected functions. As a RISC node, the PSN is configured to perform computationally intensive applications such as signal processing. The present invention further provides an interconnection scheme so that a plurality of ACE devices operates under the control of a single k-node.Type: ApplicationFiled: January 26, 2004Publication date: July 28, 2005Applicant: QuickSilver Technology, Inc.Inventor: Rojit Jacob
-
Publication number: 20040143724Abstract: The present invention provides an adaptive computing engine (ACE) that includes processing nodes having different capabilities such as arithmetic nodes, bit-manipulation nodes, finite state machine nodes, input/output nodes and a programmable scalar node (PSN). In accordance with one embodiment of the present invention, a common architecture is adaptable to function in either a kernel node, or k-node, or as general purpose RISC node. The k-node acts as a system controller responsible for adapting other nodes to perform selected functions. As a RISC node, the PSN is configured to perform computationally intensive applications such as signal processing.Type: ApplicationFiled: September 29, 2003Publication date: July 22, 2004Applicant: QuickSilver Technology, Inc.Inventors: Rojit Jacob, Dan MingLun Chuang