Patents by Inventor Dmitry Vyshetsky

Dmitry Vyshetsky has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 7562205
    Abstract: A virtual address translation table and an on-chip address cache are usable for translating virtual addresses to physical addresses. Address translation information is provided using a cluster that is associated with some range of virtual addresses and that can be used to translate any virtual address in its range to a physical address, where the sizes of the ranges mapped by different clusters may be different. Clusters are stored in an address translation table that is indexed by virtual address so that, starting from any valid virtual address, the appropriate cluster for translating that address can be retrieved from the translation table. Recently retrieved clusters are stored in an on-chip cache, and a cached cluster can be used to translate any virtual address in its range without accessing the address translation table again.
    Type: Grant
    Filed: August 23, 2007
    Date of Patent: July 14, 2009
    Assignee: Nvidia Corporation
    Inventors: Colyn S. Case, Dmitry Vyshetsky, Sean J. Treichler
  • Publication number: 20090150689
    Abstract: A data path controller, a computer device, an apparatus and a method are disclosed for integrating power management functions into a data path controller to manage power consumed by processors and peripheral devices. By embedding power management within the data path controller, the data path controller can advantageously modify its criteria in-situ so that it can adapt its power management actions in response to changes in processors and peripheral devices. In addition, the data path controller includes a power-managing interface that provides power-monitoring ports for monitoring and/or quantifying power consumption of various components. In one embodiment, the data path controller includes a power-monitoring interface for selectably monitoring power of a component.
    Type: Application
    Filed: December 23, 2008
    Publication date: June 11, 2009
    Inventors: Brad W. Simeral, David C. Reed, Dmitry Vyshetsky, Roman Surgutchick, Robert William Chapman, Joshua Titus, Anand Srinivasan, Hari U. Krishnan
  • Patent number: 7487371
    Abstract: A data path controller, a computer device, an apparatus and a method are disclosed for integrating power management functions into a data path controller to manage power consumed by processors and peripheral devices. By embedding power management within the data path controller, the data path controller can advantageously modify its criteria in-situ so that it can adapt its power management actions in response to changes in processors and peripheral devices. In addition, the data path controller includes a power-managing interface that provides power-monitoring ports for monitoring and/or quantifying power consumption of various components. In one embodiment, the data path controller includes a power-monitoring interface for selectably monitoring power of a component.
    Type: Grant
    Filed: December 16, 2005
    Date of Patent: February 3, 2009
    Assignee: Nvidia Corporation
    Inventors: Brad W. Simeral, David G. Reed, Dmitry Vyshetsky, Roman Surgutchick, Robert William Chapman, Joshua Titus, Anand Srinivasan, Hari U. Krishnan
  • Patent number: 7334108
    Abstract: A virtual address translation table and an on-chip address cache are usable for translating virtual addresses to physical addresses. Address translation information is provided using a cluster that is associated with some range of virtual addresses and that can be used to translate any virtual address in its range to a physical address, where the sizes of the ranges mapped by different clusters may be different. Clusters are stored in an address translation table that is indexed by virtual address so that, starting from any valid virtual address, the appropriate cluster for translating that address can be retrieved from the translation table. Recently retrieved clusters are stored in an on-chip cache, and a cached cluster can be used to translate any virtual address in its range without accessing the address translation table again.
    Type: Grant
    Filed: January 30, 2004
    Date of Patent: February 19, 2008
    Assignee: NVIDIA Corporation
    Inventors: Colyn S. Case, Dmitry Vyshetsky, Sean J. Treichler
  • Patent number: 7296139
    Abstract: A virtual address translation table and an on-chip address cache are usable for translating virtual addresses to physical addresses. Address translation information is provided using a cluster that is associated with some range of virtual addresses and that can be used to translate any virtual address in its range to a physical address, where the sizes of the ranges mapped by different clusters may be different. Clusters are stored in an address translation table that is indexed by virtual address so that, starting from any valid virtual address, the appropriate cluster for translating that address can be retrieved from the translation table. Recently retrieved clusters are stored in an on-chip cache, and a cached cluster can be used to translate any virtual address in its range without accessing the address translation table again.
    Type: Grant
    Filed: January 30, 2004
    Date of Patent: November 13, 2007
    Assignee: NVIDIA Corporation
    Inventors: Colyn S. Case, Dmitry Vyshetsky
  • Patent number: 7278008
    Abstract: A virtual address translation table and an on-chip address cache are usable for translating virtual addresses to physical addresses. Address translation information is provided using a cluster that is associated with some range of virtual addresses and that can be used to translate any virtual address in its range to a physical address, where the sizes of the ranges mapped by different clusters may be different. Clusters are stored in an address translation table that is indexed by virtual address so that, starting from any valid virtual address, the appropriate cluster for translating that address can be retrieved from the translation table. Recently retrieved clusters are stored in an on-chip cache, and a cached cluster can be used to translate any virtual address in its range without accessing the address translation table again.
    Type: Grant
    Filed: January 30, 2004
    Date of Patent: October 2, 2007
    Assignee: NVIDIA Corporation
    Inventors: Colyn S. Case, Dmitry Vyshetsky, Sean J. Treichler
  • Patent number: 7260686
    Abstract: A system, apparatus, and method are disclosed for storing predictions as well as examining and using one or more caches for anticipating accesses to a memory. In one embodiment, an exemplary apparatus is a prefetcher for managing predictive accesses with a memory. The prefetcher can include a speculator to generate a range of predictions, and multiple caches. For example, the prefetcher can include a first cache and a second cache to store predictions. An entry of the first cache is addressable by a first representation of an address from the range of predictions, whereas an entry of the second cache is addressable by a second representation of the address. The first and the second representations are compared in parallel against the stored predictions of either the first cache and the second cache, or both.
    Type: Grant
    Filed: August 17, 2004
    Date of Patent: August 21, 2007
    Assignee: Nvidia Corporation
    Inventors: Ziyad S. Hakura, Radoslav Danilak, Brad W. Simeral, Brian Keith Langendorf, Stefano A. Pescador, Dmitry Vyshetsky
  • Publication number: 20070143640
    Abstract: A data path controller, a computer device, an apparatus and a method are disclosed for integrating power management functions into a data path controller to manage power consumed by processors and peripheral devices. By embedding power management within the data path controller, the data path controller can advantageously modify its criteria in-situ so that it can adapt its power management actions in response to changes in processors and peripheral devices. In addition, the data path controller includes a power-managing interface that provides power-monitoring ports for monitoring and/or quantifying power consumption of various components. In one embodiment, the data path controller includes a power-monitoring interface for selectably monitoring power of a component.
    Type: Application
    Filed: December 16, 2005
    Publication date: June 21, 2007
    Inventors: Brad Simeral, David Reed, Dmitry Vyshetsky, Roman Surgutchick, Robert Chapman, Joshua Titus, Anand Srinivasan, Hari Krishnan
  • Patent number: 7055151
    Abstract: In a multi-tasking pipelined processor, consecutive instructions are executed by different tasks, eliminating the need to purge an instruction execution pipeline of subsequent instructions when a previous instruction cannot be completed. The tasks do not share registers which store task-specific values, thus eliminating the need to save or load registers when a new task is scheduled for execution. If an instruction accesses an unavailable resource, the instruction becomes suspended, allowing other tasks' instructions to be executed instead until the resource becomes available. Task scheduling is performed by hardware; no operating system is needed. Simple techniques are provided to synchronize shared resource access between different tasks.
    Type: Grant
    Filed: December 9, 1999
    Date of Patent: May 30, 2006
    Assignee: Applied Micro Circuits Corporation
    Inventors: Alexander Joffe, Dmitry Vyshetsky
  • Publication number: 20060041722
    Abstract: A system, apparatus, and method are disclosed for storing predictions as well as examining and using one or more caches for anticipating accesses to a memory. In one embodiment, an exemplary apparatus is a prefetcher for managing predictive accesses with a memory. The prefetcher can include a speculator to generate a range of predictions, and multiple caches. For example, the prefetcher can include a first cache and a second cache to store predictions. An entry of the first cache is addressable by a first representation of an address from the range of predictions, whereas an entry of the second cache is addressable by a second representation of the address. The first and the second representations are compared in parallel against the stored predictions of either the first cache and the second cache, or both.
    Type: Application
    Filed: August 17, 2004
    Publication date: February 23, 2006
    Inventors: Ziyad Hakura, Radoslav Danilak, Brad Simeral, Brian Langendorf, Stefano Pescador, Dmitry Vyshetsky
  • Publication number: 20040199916
    Abstract: In a multi-tasking pipelined processor, consecutive instructions are executed by different tasks, eliminating the need to purge an instruction execution pipeline of subsequent instructions when a previous instruction cannot be completed. The tasks do not share registers which store task-specific values, thus eliminating the need to save or load registers when a new task is scheduled for execution. If an instruction accesses an unavailable resource, the instruction becomes suspended, allowing other tasks' instructions to be executed instead until the resource becomes available. Task scheduling is performed by hardware; no operating system is needed. Simple techniques are provided to synchronize shared resource access between different tasks.
    Type: Application
    Filed: April 14, 2004
    Publication date: October 7, 2004
    Inventors: Alexander Joffe, Dmitry Vyshetsky
  • Patent number: 6330584
    Abstract: In a multi-tasking pipelined processor, consecutive instructions are executed by different tasks, eliminating the need to purge an instruction execution pipeline of subsequent instructions when a previous instruction cannot be completed. The tasks do not share registers which store task-specific values, thus eliminating the need to save or load registers when a new task is scheduled for execution. If an instruction accesses an unavailable resource, the instruction becomes suspended, allowing other tasks' instructions to be executed instead until the resource becomes available. Task scheduling is performed by hardware; no operating system is needed. Simple techniques are provided to synchronize shared resource access between different tasks.
    Type: Grant
    Filed: April 3, 1998
    Date of Patent: December 11, 2001
    Assignee: MMC Networks, Inc.
    Inventors: Alexander Joffe, Dmitry Vyshetsky
  • Patent number: 6307860
    Abstract: A processor system suitable to provide an interface between networks includes a software programmable processor and a channel processor that receives data from a network and transforms data at commands from the software programmable processor. The channel can execute only a few simple commands, but these commands are sufficient for a wide range of systems. The commands include (1) a command to transmit received data, perhaps skipping some data; and (2) a command to transmit data specified by the command itself rather than the received data. The channel is fast, simple and inexpensive.
    Type: Grant
    Filed: April 3, 1998
    Date of Patent: October 23, 2001
    Assignee: MMC Networks, Inc.
    Inventors: Alexander Joffe, Dmitry Vyshetsky