Patents by Inventor Jeffrey L. Nye
Jeffrey L. Nye has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11082327Abstract: A system and method are disclosed for performing operations on data passing through the network to reduce latency. The overall system allows data transport to become an active component in the computation, thereby improving the overall system latency, bandwidth, and/or power.Type: GrantFiled: July 9, 2019Date of Patent: August 3, 2021Assignee: ARTERIS, INC.Inventor: Jeffrey L. Nye
-
Patent number: 10990394Abstract: An integrated circuit may include a mixed instruction multiple data (xIMD) computing system. The xIMD computing system may include a plurality of data processors, each data processor representative of a lane of a single instruction multiple data (SIMD) computing system, wherein the plurality of data processors are configured to use a first dominant lane for instruction execution and to fork a second dominant lane when a data dependency instruction that does not share a taken/not-taken state with the first dominant lane is encountered during execution of a program by the xIMD computing system.Type: GrantFiled: September 28, 2017Date of Patent: April 27, 2021Assignee: Intel CorporationInventor: Jeffrey L. Nye
-
Publication number: 20200213217Abstract: A system and method are disclosed for performing operations on data passing through the network to reduce latency. The overall system allows data transport to become an active component in the computation, thereby improving the overall system latency, bandwidth, and/or power.Type: ApplicationFiled: July 9, 2019Publication date: July 2, 2020Applicant: ARTERIS, INC.Inventor: Jeffrey L. NYE
-
Patent number: 10489193Abstract: Live migration of a hardware accelerated application may be orchestrated by cloud services to transfer the application from a source server to a destination server. The live migration may be triggered by probe circuitry that monitors quality of service metrics for migration conditions at the source server. When live migration is initiated by the cloud services, a snapshot of all state information relevant to the application at the source server may be saved to network attached storage accessible by the destination server. Changes to said state information at the source server may be mirrored onto the network attached storage. The destination server may copy the snapshot and subsequent changes and run the application in parallel before taking complete control of the application. After a handshake operation between the source and destination servers, the application may be shut down at the source server.Type: GrantFiled: November 20, 2018Date of Patent: November 26, 2019Assignee: Altera CorporationInventors: Jeffrey L. Nye, Shiva Rao
-
Publication number: 20190129744Abstract: Live migration of a hardware accelerated application may be orchestrated by cloud services to transfer the application from a source server to a destination server. The live migration may be triggered by probe circuitry that monitors quality of service metrics for migration conditions at the source server. When live migration is initiated by the cloud services, a snapshot of all state information relevant to the application at the source server may be saved to network attached storage accessible by the destination server. Changes to said state information at the source server may be mirrored onto the network attached storage. The destination server may copy the snapshot and subsequent changes and run the application in parallel before taking complete control of the application. After a handshake operation between the source and destination servers, the application may be shut down at the source server.Type: ApplicationFiled: November 20, 2018Publication date: May 2, 2019Inventors: Jeffrey L. Nye, Shiva Rao
-
Publication number: 20190095208Abstract: An integrated circuit may include a mixed instruction multiple data (xIMD) computing system. The xIMD computing system may include a plurality of data processors, each data processor representative of a lane of a single instruction multiple data (SIMD) computing system, wherein the plurality of data processors are configured to use a first dominant lane for instruction execution and to fork a second dominant lane when a data dependency instruction that does not share a taken/not-taken state with the first dominant lane is encountered during execution of a program by the xIMD computing system.Type: ApplicationFiled: September 28, 2017Publication date: March 28, 2019Inventor: Jeffrey L. Nye
-
Patent number: 10169065Abstract: Live migration of a hardware accelerated application may be orchestrated by cloud services to transfer the application from a source server to a destination server. The live migration may be triggered by probe circuitry that monitors quality of service metrics for migration conditions at the source server. When live migration is initiated by the cloud services, a snapshot of all state information relevant to the application at the source server may be saved to network attached storage accessible by the destination server. Changes to said state information at the source server may be mirrored onto the network attached storage. The destination server may copy the snapshot and subsequent changes and run the application in parallel before taking complete control of the application. After a handshake operation between the source and destination servers, the application may be shut down at the source server.Type: GrantFiled: June 29, 2016Date of Patent: January 1, 2019Assignee: Altera CorporationInventors: Jeffrey L. Nye, Shiva Rao
-
Patent number: 9552206Abstract: Traditionally, providing parallel processing within a multi-core system has been very difficult. Here, however, a system is provided where serial source code is automatically converted into parallel source code, and a processing cluster is reconfigured “on the fly” to accommodate the parallelized code based on an allocation of memory and compute resources. Thus, the processing cluster and its corresponding system programming tool provide a system that can perform parallel processing from a serial program that is transparent to a user. Generally, a control node connected to the address and data leads of a host processor uses messages to control the processing of data in a processing cluster. The cluster includes nodes of parallel processors, shared function memory, a global load/store, and hardware accelerators all connected to the control node by message busses. A crossbar data interconnect routes data to the cluster circuits separate from the message busses.Type: GrantFiled: September 14, 2011Date of Patent: January 24, 2017Assignee: Texas Instruments IncorporatedInventors: William M. Johnson, Murali S. Chinnakonda, Jeffrey L. Nye, Toshio Nagata, John W. Glotzbach, Hamid R. Sheikh, Ajay Jayaraj, Stephen Busch, Shalini Gupta, Robert J.P. Nychka, David H. Bartley, Ganesh Sundararajan
-
Publication number: 20120131309Abstract: Traditionally, providing parallel processing within a multi-core system has been very difficult. Here, however, a system in provided where serial source code is automatically converted into parallel source code, and a processing cluster is reconfigured “on the fly” to accommodate the parallelized code based on an allocation of memory and compute resources. Thus, the processing cluster and its corresponding system programming tool provide a system that can perform parallel processing from a serial program that is transparent to a user.Type: ApplicationFiled: September 14, 2011Publication date: May 24, 2012Applicant: Texas Instruments IncorporatedInventors: William M. Johnson, Murali S. Chinnakonda, Jeffrey L. Nye, Toshio Nagata, John W. Glotzbach, Hamid R. Sheikh, Ajay Jayaraj, Stephen Busch, Shalini Gupta, Robert J.P. Nychka, David H. Bartley, Ganesh Sundararajan
-
Patent number: 7752426Abstract: A processor (1700) for processing instructions has a pipeline (1710, 1736, 1740) including a fetch stage (1710) and an execute stage (1870), a first storing circuit (aGHR 2130) associated with said fetch stage (1710) and operable to store a history of actual branches, and a second storing circuit (wGHR 2140) associated with said fetch stage (1710) and operable to store a pattern of predicted branches, said second storing circuit (wGHR 2140) coupled to said first storing circuit (aGHR 2130), said execute stage (1870) coupled back to said first storing circuit (aGHR 2130). Other processors, wireless communications devices, systems, circuits, devices, branch prediction processes and methods of operation, processes of manufacture, and articles of manufacture, as disclosed and claimed.Type: GrantFiled: August 24, 2005Date of Patent: July 6, 2010Assignee: Texas Instruments IncorporatedInventors: Jeffrey L. Nye, Thang Tran
-
Patent number: 7587532Abstract: A method and apparatus for adaptive buffer sizing adjusts the size of the buffer at different levels using a “high water mark” to different levels for different system conditions. The high water mark is used by the buffer logic as an indication of when to assert the buffer “Full” flag. In turn, the full flag is used by the instruction fetch logic as an indication of when to stop fetching further instructions.Type: GrantFiled: January 31, 2005Date of Patent: September 8, 2009Assignee: Texas Instruments IncorporatedInventors: Jeffrey L. Nye, Sam B. Sandbote
-
Patent number: 7266648Abstract: A method and apparatus to speculatively allocate a combinable store request buffer for a cache lock operation by setting an extended lock indicator. The store request buffer is speculatively allocated and the load-lock micro-operation associated with the cache lock operation is sent to a replay loop. During the replay window, conflict detection processing is completed. When the load-lock micro-operation returns from replay, the store buffer is de-allocated if a conflict has been detected during the replay window.Type: GrantFiled: August 29, 2005Date of Patent: September 4, 2007Assignee: Intel CorporationInventors: Edwin R. Sutanto, Hoichi Cheong, Zhongru J. Lin, Jeffrey L. Nye
-
Publication number: 20070180310Abstract: Disclosed herein are a system and method for designing digital circuits. In some embodiments, the digital circuits include processors having dedicated messaging hardware that enable processor cores to minimize interrupt activity related to inter-core communications. The messaging hardware receives and parses any message in its entirety prior to passing the contents of the message on to the digital circuit. In other embodiments, the digital circuit functionalities are partitioned across individual cores to enable parallel execution. Each core may be provided with standardized messaging hardware that shields internal implementation details from all other cores. This modular approach accelerates development and testing, and renders parallel circuit design to more efficiently attain feasible speedups. These digital circuit cores may be homogenous or heterogeneous.Type: ApplicationFiled: January 26, 2007Publication date: August 2, 2007Applicant: TEXAS INSTRUMENTS, INC.Inventors: William M. Johnson, Jeffrey L. Nye
-
Patent number: 6986010Abstract: A method and apparatus to speculatively allocate a combinable store request buffer for a cache lock operation by setting an extended lock indicator. The store request buffer is speculatively allocated and the load-lock micro-operation associated with the cache lock operation is sent to a replay loop. During the replay window, conflict detection processing is completed. When the load-lock micro-operation returns from replay, the store buffer is de-allocated if a conflict has been detected during the replay window.Type: GrantFiled: December 13, 2002Date of Patent: January 10, 2006Assignee: Intel CorporationInventors: Edwin R. Sutanto, Hoichi Cheong, Zhongru J. Lin, Jeffrey L. Nye
-
Publication number: 20040117573Abstract: A method and apparatus to speculatively allocate a combinable store request buffer for a cache lock operation by setting an extended lock indicator. The store request buffer is speculatively allocated and the load-lock micro-operation associated with the cache lock operation is sent to a replay loop. During the replay window, conflict detection processing is completed. When the load-lock micro-operation returns from replay, the store buffer is de-allocated if a conflict has been detected during the replay window.Type: ApplicationFiled: December 13, 2002Publication date: June 17, 2004Inventors: Edwin R. Sutanto, Hoichi Cheong, Zhongru J. Lin, Jeffrey L. Nye
-
Patent number: 6189077Abstract: An access circuit for data swapping between two computers and a computer system including the access circuit. Each computer including an address bus for supplying addresses and a data bus for transferring data. The access circuit includes a register file and two address decoder circuits. The register file has a plurality of storage locations for storing data. The register file has dual data ports capable of simultaneous data transfer via the first data port with a first data storage location and via the second data port with a second, different storage location. Each address decoder is connected to the address bus of a corresponding computer and the register file. The address decoders translate an address received on the address bus to a storage location of the register file. Two handshakes circuits are connected to respective address decoders and digital computers. The first and second address decoders are connected to each other.Type: GrantFiled: June 7, 1995Date of Patent: February 13, 2001Assignee: Texas Instruments IncorporatedInventors: Iain C. Robertson, Jeffrey L. Nye, Michael D. Asal, Graham B. Short, Richard D. Simpson, James G. Littleton
-
Patent number: 6154824Abstract: A multifunction access circuit for use with first and second digital computers each having an address bus for supplying addresses and a data bus for supplying data. The access circuit has an address decoder with inputs for the address bus from the first computer, and an address translator circuit having address inputs for addresses supplied by the address bus of the first computer and outputs for translated addresses to the address bus of the second computer. The address translator circuit also has registers selectable by the address decoder and data inputs to program the registers so selected with data from the data bus from the first computer. Also in the access circuit is a port circuit with registers controlled by the address decoder for entry of address information from the data bus of the first computer and assertion of the address information on the address bus of the second computer.Type: GrantFiled: June 7, 1995Date of Patent: November 28, 2000Assignee: Texas Instruments IncorporatedInventors: Iain C. Robertson, Jeffrey L. Nye, Michael D. Asal, Graham B. Short, Richard D. Simpson, James G. Littleton
-
Patent number: 5940610Abstract: Multimedia information (e.g. graphics, video, sound, control information) passes through a system bus from a CPU main memory to a display memory in accordance with CPU commands. The information may be packetized with associated packet types identifying the different media. A media stream controller processes the information and passes the processed information to the display memory. Controllers in the media stream controller individually pass multimedia information to the display memory. A PACDAC controller in the media stream controller causes media (e.g. graphics, video) in the display memory to be transferred to a PACDAC for display. The format, sequence, and rate of this transfer may be flexibly controlled by software on a frame by frame basis. Arbitration logic establishes priorities for the different controllers in the media stream controller so they may share a single bus for accessing the display memory. A single interrupt controller coordinates interrupts (e.g.Type: GrantFiled: October 3, 1996Date of Patent: August 17, 1999Assignee: Brooktree CorporationInventors: David C. Baker, Michael D. Asal, Jonathan I. Siann, Paul B. Wood, Jeffrey L. Nye, Stephen G. Glennon, Matthew D. Bates
-
Patent number: 5696924Abstract: A memory access system for use with a graphics processor having an address bus, a data bus and a set of control lines. An address translator circuit connected to the address bus of the graphics processor supplies a translated address to a memory upon receipt of an address from the graphics processor. A logic circuit responds to a write signal to automatically increment the translated address and responds to a control signal to return to the translated address. Control circuitry connected to the logic circuit responds to a read signal to supply the control signal to the logic circuit.Type: GrantFiled: June 7, 1995Date of Patent: December 9, 1997Assignee: Texas Instruments IncorporatedInventors: Iain C. Robertson, Jeffrey L. Nye, Michael D. Asal, Graham B. Short, Richard D. Simpson, James G. Littleton
-
Patent number: 5696923Abstract: A computer graphics system includes a host computer and a graphics processor. The graphics processor includes a control register. When the graphics processor writes to the control register it simultaneously generates a predetermined address on a local address bus and supplies data on a local data bus identical to data to be written into the control register. A shadow register circuit connected to both the host computer and the graphics processor includes a shadow register and first and second address decoders. The first address decoder enables a write from a local data bus into the shadow register upon detection of the predetermined address. The second address decoder enables a read from the shadow register via a host data bus upon detection of the predetermined address on a host address bus.Type: GrantFiled: June 7, 1995Date of Patent: December 9, 1997Assignee: Texas Instruments IncorporatedInventors: Iain C. Robertson, Jeffrey L. Nye, Michael D. Asal, Graham B. Short, Richard D. Simpson, James G. Littleton