Patents Assigned to NVidia
-
Patent number: 9454792Abstract: A device, system and method for transferring network data are presented. The device includes: a data processing module configured to convert a data signal in a TMDS signal output via a DVI of a graphics card to network data for being transferred via a network; and a network transmitter for receiving the network data and transferring the network data to an external device via the network. The data signal is generated by encoding texture data generated by a GPU of the graphics card by a digital video sender of the graphics card. The texture data is generated by binding a pointer to general computing data stored in a device memory of the graphics card to a texture stored in the device memory by the GPU. The general computing data is generated by general computation executed by the GPU.Type: GrantFiled: April 25, 2013Date of Patent: September 27, 2016Assignee: NVIDIA CorporationInventors: Zhen Jia, Shu Zhang, Jun Qiu
-
Patent number: 9454806Abstract: A computer implemented method of performing an approximate-nearest-neighbor search is disclosed. The method comprises dividing an image into a plurality of tiles. Further, for each of the plurality of tiles, perform the following in parallel on a processor: (a) dividing image patches into a plurality of clusters, wherein each cluster comprises similar images patches, and wherein the dividing continues recursively until a size of a cluster is below a threshold value; (b) performing a nearest-neighbor query within each of the plurality of clusters; and (c) performing collaborative filtering in parallel for each image patch, wherein the collaborative filtering aggregates and processes nearest neighbor image patches from a same cluster containing a respective image patch to form an output image.Type: GrantFiled: February 26, 2015Date of Patent: September 27, 2016Assignee: NVIDIA CORPORATIONInventors: Dawid Stanislaw Pajak, Yun-Ta Tsai, Markus Steinberger
-
Patent number: 9454975Abstract: Voice trigger. In accordance with a first method embodiment, a long term average audio energy is determined based on a one-bit pulse-density modulation bit stream. A short term average audio energy is determined based on the one-bit pulse-density modulation bit stream. The long term average audio energy is compared to the short term average audio energy. Responsive to the comparing, a voice trigger signal is generated if the short term average audio energy is greater than the long term average audio energy. Determining the long term average audio energy may be performed independent of any decimation of the bit stream.Type: GrantFiled: November 7, 2013Date of Patent: September 27, 2016Assignee: NVIDIA CORPORATIONInventor: Anil W. Ubale
-
Patent number: 9454843Abstract: A system, method, and computer program product are provided for anti-aliasing. During a first processing pass of a plurality of graphics primitives, z data is computed for multiple samples of each pixel in an image to generate a multi-sample z buffer. During a second processing pass of the graphics primitives, computed color values corresponding to each pixel in a color buffer that stores one color value for each pixel are accumulated.Type: GrantFiled: February 5, 2013Date of Patent: September 27, 2016Assignee: NVIDIA CorporationInventors: Christian Jean Rouet, Eric Brian Lum, Rui Manuel Bastos
-
Patent number: 9448803Abstract: A method and a system are provided for hardware scheduling of barrier instructions. Execution of a plurality of threads to process instructions of a program that includes a barrier instruction is initiated, and when each thread reaches the barrier instruction during execution of program, it is determined whether the thread participates in the barrier instruction. The threads that participate in the barrier instruction are then serially executed to process one or more instructions of the program that follow the barrier instruction. A method and system are also provided for impatient scheduling of barrier instructions. When a portion of the threads that is greater than a minimum number of threads and less than all of the threads in the plurality of threads reaches the barrier instruction each of the threads in the portion is serially executed to process one or more instructions of the program that follow the barrier instruction.Type: GrantFiled: March 11, 2013Date of Patent: September 20, 2016Assignee: NVIDIA CorporationInventors: John Erik Lindholm, Tero Tapani Karras, Timo Oskari Aila, Samuli Matias Laine
-
Patent number: 9448779Abstract: One embodiment of the present invention sets forth a technique for translating application programs written using a parallel programming model for execution on multi-core graphics processing unit (GPU) for execution by general purpose central processing unit (CPU). Portions of the application program that rely on specific features of the multi-core GPU are converted by a translator for execution by a general purpose CPU. The application program is partitioned into regions of synchronization independent instructions. The instructions are classified as convergent or divergent and divergent memory references that are shared between regions are replicated. Thread loops are inserted to ensure correct sharing of memory between various threads during execution by the general purpose CPU.Type: GrantFiled: March 20, 2009Date of Patent: September 20, 2016Assignee: NVIDIA CorporationInventors: Vinod Grover, Bastiaan Joannes Matheus Aarts, Michael Murphy, Jayant B. Kolhe, John Bryan Pormann, Douglas Saylor
-
Patent number: 9448125Abstract: A method, in one embodiment, can include modeling and calibrating two types of sensors that are part of a semiconductor device. In addition, the method can include determining a temperature and voltage based on data received from the two sensors.Type: GrantFiled: January 25, 2012Date of Patent: September 20, 2016Assignee: NVIDIA CORPORATIONInventors: Abhishek Singh, Wojciech Jakub Poppe, Ilyas Elkin
-
Patent number: 9448766Abstract: An arithmetic logic stage in a graphics pipeline includes a number of arithmetic logic units (ALUs). The ALUs each include, for example, a multiplier and an adder. The ALUs are interconnected by circuitry that, for example, routes the output from the multiplier in one ALU to both the adder in that ALU and an adder in another ALU.Type: GrantFiled: August 27, 2013Date of Patent: September 20, 2016Assignee: NVIDIA CorporationInventors: Tyson Bergland, Michael J. M. Toksvig, Justin Michael Mahan
-
Patent number: 9448806Abstract: A floating-point unit and a method of identifying exception cases in a floating-point unit. In one embodiment, the floating-point unit includes: (1) a floating-point computation circuit having a normal path and an exception path and operable to execute an operation on an operand and (2) a decision circuit associated with the normal path and the exception path and configured to employ a flush-to-zero mode of the floating-point unit to determine which one of the normal path and the exception path is appropriate for carrying out the operation on the operand.Type: GrantFiled: September 25, 2012Date of Patent: September 20, 2016Assignee: Nvidia CorporationInventors: Marcin Andrychowicz, Alex Fit-Florea
-
Patent number: 9448837Abstract: Techniques are provided for restoring thread groups in a cooperative thread array (CTA) within a processing core. Each thread group in the CTA is launched to execute a context restore routine. Each thread group, executes the context restore routine to restore from a memory a first portion of context associated with the thread group, and determines whether the thread group completed an assigned function prior to executing the context restore routine. If the thread group completed an assigned function prior to executing the context restore routine, then the thread group exits the context restore routine. If the thread group did not complete the assigned function prior to executing the context restore routine, then the thread group executes one or more operations associated with a trap handler routine. One advantage of the disclosed techniques is that the trap handling routine operates efficiently in parallel processors.Type: GrantFiled: April 15, 2013Date of Patent: September 20, 2016Assignee: NVIDIA CorporationInventors: Gerald F. Luiz, Philip Alexander Cuadra, Luke Durant, Shirish Gadre, Robert Ohannessian, Lacky V. Shah, Nicholas Wang, Arthur Merlin Danskin
-
Patent number: 9451187Abstract: Embodiments of the present invention are directed to methods and systems for performing automatic lens shading correction using module-specific calibrations. According to one aspect of the invention, a method is provided that is performed over three main stages. During a first stage, radial symmetric component data is determined that is common to camera modules of the type to be calibrated. During the second stage, the actual measurement of the shading profile of one or more specific camera modules is performed. In the third and final stage is the extrapolation stage, the base measurement surfaces of a camera module type determined in the first stage are extrapolated and modified with the module-specific Bezier correction surface and calibration data of the second stage. The output surfaces of this third and final stage are used to correct for the shading profile in the camera module, depending on the light-source estimation.Type: GrantFiled: December 31, 2012Date of Patent: September 20, 2016Assignee: NVIDIA CORPORATIONInventors: Noy Cohen, Ricardo J. Motta
-
Patent number: 9448804Abstract: One embodiment of the present invention sets forth a technique for managing buffer entries in a tile-based architecture. The technique includes receiving a first plurality of graphics primitives and a first buffer address at which attributes associated with the first plurality of graphics primitives are stored. The technique further includes, for each tile included in a plurality of tiles, transmitting the first plurality of graphics primitives and the first buffer address to a screen space pipeline and receiving an acknowledgement from the screen space pipeline indicating that processing the first plurality of graphics primitives has completed. The technique further includes determining that processing the first plurality of graphics primitives has completed for a last tile included in the plurality of tiles and that the acknowledgement has been received for each tile included in the plurality of tiles, and, in response, releasing a buffer entry associated with the first buffer address.Type: GrantFiled: October 3, 2013Date of Patent: September 20, 2016Assignee: NVIDIA CorporationInventors: Ziyad S. Hakura, Cynthia Ann Edgeworth Allison, Dale L. Kirkland
-
Patent number: 9448935Abstract: Techniques are disclosed for performing memory access operations. A texture unit receives a memory access operation that includes a tuple associated with a first view in a plurality of views. The texture unit retrieves a first hash value associated with a first texture header in a plurality of texture headers, where the first texture header is related to the first view. The texture unit retrieves a second hash value associated with a second texture header in the plurality of texture headers, where the second texture header is related to a second view. The texture unit determines whether the first view is potentially aliased with the second view, based on the first and second hash values. If so, then the texture unit invalidates a cache entry in a cache memory associated with the second texture header. Otherwise, the texture unit maintains the cache entry.Type: GrantFiled: September 25, 2013Date of Patent: September 20, 2016Assignee: NVIDIA CorporationInventors: Jeff Bolz, Patrick R. Brown, Steven J. Heinrich, Dale L. Kirkland, Joel McCormack
-
Patent number: 9442755Abstract: A method and a system are provided for hardware scheduling of indexed barrier instructions. Execution of a plurality of threads to process instructions of a program that includes a barrier instruction is initiated and when each thread reaches the barrier instruction, the thread pauses execution of the instructions. A first sub-group of threads in the plurality of threads is associated with a first sub-barrier index and a second sub-group of threads in the plurality of threads is associated with a second sub-barrier index. When the barrier instruction can be scheduled for execution, threads in the first sub-group are executed serially and threads in the second sub-group are executed serially and at least one thread in the first sub-group is executed in parallel with at least one thread in the second sub-group.Type: GrantFiled: March 15, 2013Date of Patent: September 13, 2016Assignee: NVIDIA CorporationInventors: John Erik Lindholm, Tero Tapani Karras
-
Patent number: 9442759Abstract: A time slice group (TSG) is a grouping of different streams of work (referred to herein as “channels”) that share the same context information. The set of channels belonging to a TSG are processed in a pre-determined order. However, when a channel stalls while processing, the next channel with independent work can be switched to fully load the parallel processing unit. Importantly, because each channel in the TSG shares the same context information, a context switch operation is not needed when the processing of a particular channel in the TSG stops and the processing of a next channel in the TSG begins. Therefore, multiple independent streams of work are allowed to run concurrently within a single context increasing utilization of parallel processing units.Type: GrantFiled: December 9, 2011Date of Patent: September 13, 2016Assignee: NVIDIA CorporationInventors: Samuel H. Duncan, Lacky V. Shah, Sean J. Treichler, Daniel Elliot Wexler, Jerome F. Duluk, Jr., Philip Browning Johnson, Jonathon Stuart Ramsay Evans
-
Patent number: 9445242Abstract: A method for supporting broadcast transmission in a wireless communication system (200) that comprises a plurality of communication cells, with broadcast content being routed from a base station (210) to at least one wireless communication unit (225, 226) via at least one relay node (RN) (224) is described. The method comprises, at the base station (210) broadcasting the broadcast content from the base station (210) to at least one from a group consisting of: the at least one RN (224), the at least one wireless communication unit (226); and supplementing the broadcast transmission with at least one augmented unicast transmission associated with the broadcast content. A base station (210), an integrated circuit and a non-transitory computer program product comprising executable program code are also described. A relay node (224), an integrated circuit, a method performed at the relay node and a non-transitory computer program product comprising executable program code are also described.Type: GrantFiled: December 16, 2011Date of Patent: September 13, 2016Assignee: NVIDIA CORPORATIONInventor: Stephen Barrett
-
Patent number: 9436235Abstract: A cooling subsystem is provided for dissipating heat from processor. The cooling subsystem includes a heat sink comprising an upper portion having a plurality of fins formed therein and a base portion fixed to the upper portion to form a vapor chamber in an enclosed volume between the upper portion and the base portion.Type: GrantFiled: February 26, 2013Date of Patent: September 6, 2016Assignee: NVIDIA CorporationInventors: Srinivasa Rao Damaraju, Joseph Walters, Dalton Seth O'Connor
-
Patent number: 9436475Abstract: A system and method for executing sequential code in the context of a single-instruction, multiple-thread (SIMT) processor. In one embodiment, the system includes: (1) a pipeline control unit operable to create a group of counterpart threads of the sequential code, one of the counterpart threads being a master thread, remaining ones of the counterpart threads being slave threads and (2) lanes operable to: (2a) execute certain instructions of the sequential code only in the master thread, corresponding instructions in the slave threads being predicated upon the certain instructions and (2b) broadcast branch conditions in the master thread to the slave threads.Type: GrantFiled: December 21, 2012Date of Patent: September 6, 2016Assignee: NVIDIA CORPORATIONInventors: Gautam Chakrabarti, Yuan Lin, Jaydeep Marathe, Okwan Kwon, Amit Sabne
-
Patent number: 9437031Abstract: A large non-patterned noise texture occupies a relatively small physical memory space. Each of a small set of physical pages in physical memory includes noise texels forming part of a noise texture. A large “virtual” noise texture is created by mapping each one of a large number of pages in virtual address space to one of the small set of physical pages; multiple virtual pages may be mapped to the same physical page. The physical page that each virtual page maps to is randomly or pseudo-randomly selected such that the resulting noise texture appears to be non-repeating. When a noise texel is requested by reference to a virtual address during rendering, the virtual address of the virtual page is translated to the corresponding physical address, and the noise texel is retrieved.Type: GrantFiled: September 15, 2006Date of Patent: September 6, 2016Assignee: NVIDIA CorporationInventor: Cass W Everitt
-
Patent number: 9436243Abstract: A power source management system of a circuit board that comprises: a processor, comprising a core voltage input terminal; and a core voltage feedback terminal; and a voltage regulating member, comprising a setting terminal with a fixed reference voltage provided thereto; a detecting terminal connected to the core voltage feedback terminal to detect a feedback core voltage from the core voltage feedback terminal; and a core voltage output terminal connected to the core voltage input terminal to provide a core voltage thereto, wherein the core voltage is regulated by the voltage regulating member based on the feedback core voltage, such that the feedback core voltage is equal to the fixed reference voltage, wherein an offset voltage equal to a difference between a desired core voltage of the processor and the fixed reference voltage is provided between the core voltage input terminal and the core voltage feedback terminal by the processor.Type: GrantFiled: March 8, 2013Date of Patent: September 6, 2016Assignee: NVIDIA CorporationInventors: Yu Zhao, Fei Wang, Xiang Sun