Patents by Inventor Phillip M. Jones

Phillip M. Jones has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 9804969
    Abstract: A method includes receiving an instruction to be executed by a processor. The method further includes performing a lookup in a page crossing buffer that includes one or more entries to determine if the instruction has an entry in the page crossing buffer. Each of the entries includes a physical address. The method further includes, when the page crossing buffer has the entry in the page crossing buffer, retrieving a particular physical address from the entry in the page crossing buffer.
    Type: Grant
    Filed: December 20, 2012
    Date of Patent: October 31, 2017
    Assignee: QUALCOMM Incorporated
    Inventors: Suresh K. Venkumahanti, Jiajin Tu, Phillip M. Jones
  • Patent number: 9122486
    Abstract: Each branch instruction having branch prediction support has branch prediction bits in architecture specified bit positions in the branch instruction. An instruction cache supports modifying the branch instructions with updated branch prediction bits that are dynamically determined when the branch instruction executes.
    Type: Grant
    Filed: November 8, 2010
    Date of Patent: September 1, 2015
    Assignee: QUALCOMM Incorporated
    Inventors: Suresh K. Venkumahanti, Lucian Codrescu, Stephen R. Shannon, Lin Wang, Phillip M. Jones, Daisy T. Palal, Jiajin Tu
  • Publication number: 20140181459
    Abstract: A method includes receiving an instruction to be executed by a processor. The method further includes performing a lookup in a page crossing buffer that includes one or more entries to determine if the instruction has an entry in the page crossing buffer. Each of the entries includes a physical address. The method further includes, when the page crossing buffer has the entry in the page crossing buffer, retrieving a particular physical address from the entry in the page crossing buffer.
    Type: Application
    Filed: December 20, 2012
    Publication date: June 26, 2014
    Applicant: QUAL COMM Incorporated
    Inventors: Suresh K. Venkumahanti, Jiajin Tu, Phillip M. Jones
  • Publication number: 20120117327
    Abstract: Each branch instruction having branch prediction support has branch prediction bits in architecture specified bit positions in the branch instruction. An instruction cache supports modifying the branch instructions with updated branch prediction bits that are dynamically determined when the branch instruction executes.
    Type: Application
    Filed: November 8, 2010
    Publication date: May 10, 2012
    Applicant: QUALCOMM INCORPORATED
    Inventors: Suresh K. Venkumahanti, Lucian Codrescu, Stephen R. Shannon, Lin Wang, Phillip M. Jones, Daisy T. Palal, Jiajin Tu
  • Patent number: 8078818
    Abstract: A system comprises a plurality of nodes coupled together via a switching device. Each node comprises a processor coupled to a memory. Migration logic in the switching device is configured to migrate segments of each memory to the switching device.
    Type: Grant
    Filed: February 25, 2005
    Date of Patent: December 13, 2011
    Assignee: Hewlett-Packard Development Company, L.P.
    Inventors: William J. Walker, Paras A. Shah, James K. Yu, Kenneth Jansen, Vasileios Balabanos, Andrew D. Olsen, Phillip M. Jones
  • Patent number: 7685411
    Abstract: An instruction memory unit comprises a first memory structure operable to store program instructions, and a second memory structure operable to store program instructions fetched from the first memory structure, and to issue stored program instructions for execution. The second memory structure is operable to identify a repeated issuance of a forward program redirect construct, and issue a next program instruction already stored in the second memory structure if a resolution of the forward branching instruction is identical to a last resolution of the same. The second memory structure is further operable to issue a backward program redirect construct, determine whether a target instruction is stored in the second memory structure, issue the target instruction if the target instruction is stored in the second memory structure, and fetch the target instruction from the first memory structure if the target instruction is not stored in the second memory structure.
    Type: Grant
    Filed: April 11, 2005
    Date of Patent: March 23, 2010
    Assignee: QUALCOMM Incorporated
    Inventors: Muhammad Ahmed, Lucian Codrescu, Erich Plondke, William C. Anderson, Robert Allan Lester, Phillip M. Jones
  • Patent number: 7120758
    Abstract: Method and apparatus for improving processor performance. In some embodiments, processing speed may be improved by reusing data stored in a buffer during an initial request by subsequent requests. Assignment of temporary storage buffers in a controller may be made to allow for the potential for reuse of the data. Further, a hot buffer may be designated to allow for reuse of the data stored in the hot buffer. On subsequent requests, data stored in the hot buffer may be sent to a requesting device without re-retrieving the data from memory.
    Type: Grant
    Filed: February 12, 2003
    Date of Patent: October 10, 2006
    Assignee: Hewlett-Packard Development Company, L.P.
    Inventors: Phillip M. Jones, Robert A. Lester, Jens K. Ramsey, William J. Walker, John E. Larson, James Andre, Paul Rawlins
  • Patent number: 6961800
    Abstract: Methods for improving processor performance. Specifically, by reducing some of the latency cycles within a host controller, request processing speed can be improved. One technique for improving processing speed involves initiating a deferred reply transaction before the data is available from a memory controller. A second technique involves anticipating the need to transition from a block next request (BNR) state to a bus priority request (BPRI) state, thereby eliminating the need to wait for a request check to determine if the BPRI state must be implemented.
    Type: Grant
    Filed: September 28, 2001
    Date of Patent: November 1, 2005
    Assignee: Hewlett-Packard Development Company, L.P.
    Inventors: Robert A. Lester, Kenneth T. Chin, Jim Blocker, John E. Larson, Phillip M. Jones, Paul B. Rawlins
  • Patent number: 6865647
    Abstract: A cache-based system is adapted for dynamic cache partitioning. A cache is partitioned into a plurality of cache partitions for a plurality of entities. Each cache partition can be assigned as a private cache for a different entity. If a first cache partition satisfying a first predetermined cache partition condition and a second cache partition satisfying a second predetermined cache partition condition are detected, then the size of the first cache partition is increased by a predetermined segment and the size of the second cache partition is decreased by the predetermined segment. An entity can perform cacheline replacement exclusively in its assigned cache partition, and also be capable of reading any cache partition.
    Type: Grant
    Filed: December 8, 2003
    Date of Patent: March 8, 2005
    Assignee: Hewlett-Packard Development Company, L.P.
    Inventors: Sompong P. Olarig, Phillip M. Jones, John E. Jenne
  • Patent number: 6848015
    Abstract: A computer system including multiple CPUs inform other logic in a computer system as to the priority level (e.g., task priority) associated with the CPU or software executing thereon. The logic makes arbitration decisions regarding CPU transactions based, at least in part, on the task priorities of the various CPUs. The logic that implements this technique may be a host bridge within a computer system having multiple CPUs or in a switch or router that interconnects multiple nodes or computer systems.
    Type: Grant
    Filed: November 30, 2001
    Date of Patent: January 25, 2005
    Assignee: Hewlett-Packard Development Company, L.P.
    Inventor: Phillip M. Jones
  • Patent number: 6829665
    Abstract: A technique for optimizing cycle time in maintaining cache coherency. Specifically, a method and apparatus are provided to optimize the processing of requests in a multi-processor-bus system which implements a snoop-based coherency scheme. The acts of snooping a bus for a first address and searching a posting queue for the next address to be snooped are performed simultaneously to minimize the request cycle time.
    Type: Grant
    Filed: September 28, 2001
    Date of Patent: December 7, 2004
    Assignee: Hewlett-Packard Development Company, L.P.
    Inventors: Phillip M. Jones, Paul B. Rawlins, Kenneth T. Chin
  • Patent number: 6823409
    Abstract: A mechanism for efficiently filtering snoop requests in a multi-processor bus system. Specifically, a snoop filter is provided to filter unnecessary snoops in a multi-bus system.
    Type: Grant
    Filed: September 28, 2001
    Date of Patent: November 23, 2004
    Assignee: Hewlett-Packard Development Company, L.P.
    Inventors: Phillip M. Jones, Paul B. Rawlins
  • Publication number: 20040158685
    Abstract: Method and apparatus for improving processor performance. In some embodiments, processing speed may be improved by reusing data stored in a buffer during an initial request by subsequent requests. Assignment of temporary storage buffers in a controller may be made to allow for the potential for reuse of the data. Further, a hot buffer may be designated to allow for reuse of the data stored in the hot buffer. On subsequent requests, data stored in the hot buffer may be sent to a requesting device without re-retrieving the data from memory.
    Type: Application
    Filed: February 12, 2003
    Publication date: August 12, 2004
    Inventors: Phillip M. Jones, Robert A. Lester, Jens K. Ramsey, William J. Walker, John E. Larson, James Andre, Paul Rawlins
  • Publication number: 20040143707
    Abstract: A cache-based system is adapted for dynamic cache partitioning. A cache is partitioned into a plurality of cache partitions for a plurality of entities. Each cache partition can be assigned as a private cache for a different entity. If a first cache partition satisfying a first predetermined cache partition condition and a second cache partition satisfying a second predetermined cache partition condition are detected, then the size of the first cache partition is increased by a predetermined segment and the size of the second cache partition is decreased by the predetermined segment. An entity can perform cacheline replacement exclusively in its assigned cache partition, and also be capable of reading any cache partition.
    Type: Application
    Filed: December 8, 2003
    Publication date: July 22, 2004
    Inventors: Sompong P. Olarig, Phillip M. Jones, John E. Jenne
  • Patent number: 6662272
    Abstract: A cache-based system is adapted for dynamic cache partitioning. A cache is partitioned into a plurality of cache partitions for a plurality of entities. Each cache partition can be assigned as a private cache for a different entity. If a first cache partition satisfying a first predetermined cache partition condition and a second cache partition satisfying a second predetermined cache partition condition are detected, then the size of the first cache partition is increased by a predetermined segment and the size of the second cache partition is decreased by the predetermined segment. An entity can perform cacheline replacement exclusively in its assigned cache partition, and also be capable of reading any cache partition.
    Type: Grant
    Filed: September 29, 2001
    Date of Patent: December 9, 2003
    Assignee: Hewlett-Packard Development Company, L.P.
    Inventors: Sompong P. Olarig, Phillip M. Jones, John E. Jenne
  • Publication number: 20030105911
    Abstract: A computer system including multiple CPUs inform other logic in a computer system as to the priority level (e.g., task priority) associated with the CPU or software executing thereon. The logic makes arbitration decisions regarding CPU transactions based, at least in part, on the task priorities of the various CPUs. The logic that implements this technique may be a host bridge within a computer system having multiple CPUs or in a switch or router that interconnects multiple nodes or computer systems.
    Type: Application
    Filed: November 30, 2001
    Publication date: June 5, 2003
    Inventor: Phillip M. Jones
  • Publication number: 20030070016
    Abstract: A mechanism for efficiently filtering snoop requests in a multi-processor bus system. Specifically, a snoop filter is provided to filter unnecessary snoops in a multi-bus system.
    Type: Application
    Filed: September 28, 2001
    Publication date: April 10, 2003
    Inventors: Phillip M. Jones, Paul B. Rawlins
  • Publication number: 20030065886
    Abstract: A cache-based system is adapted for dynamic cache partitioning. A cache is partitioned into a plurality of cache partitions for a plurality of entities. Each cache partition can be assigned as a private cache for a different entity. If a first cache partition satisfying a first predetermined cache partition condition and a second cache partition satisfying a second predetermined cache partition condition are detected, then the size of the first cache partition is increased by a predetermined segment and the size of the second cache partition is decreased by the predetermined segment. An entity can perform cacheline replacement exclusively in its assigned cache partition, and also be capable of reading any cache partition.
    Type: Application
    Filed: September 29, 2001
    Publication date: April 3, 2003
    Inventors: Sompong P. Olarig, Phillip M. Jones, John E. Jenne
  • Publication number: 20030065860
    Abstract: An internal bus structure for a multi-processor-bus system. More specifically, an internal bus protocol/structure is described. The internal bus structure includes unidirectional, point-to-point connections between control modules. The individual buses carry unique transactions corresponding to a request. Each transaction includes an identification tag. The present protocol provides for efficient communication between processors, peripheral devices, memory and coherency modules. The present protocol and design scheme is generic in that the techniques are scalable and re-usable.
    Type: Application
    Filed: September 28, 2001
    Publication date: April 3, 2003
    Inventors: Robert A. Lester, Kenneth T. Chin, Jim Blocker, John E. Larson, Phillip M. Jones, Paul B. Rawlins
  • Publication number: 20030065844
    Abstract: Methods for improving processor performance. Specifically, by reducing some of the latency cycles within a host controller, request processing speed can be improved. One technique for improving processing speed involves initiating a deferred reply transaction before the data is available from a memory controller. A second technique involves anticipating the need to transition from a block next request (BNR) state to a bus priority request (BPRI) state, thereby eliminating the need to wait for a request check to determine if the BPRI state must be implemented.
    Type: Application
    Filed: September 28, 2001
    Publication date: April 3, 2003
    Inventors: Robert A. Lester, Kenneth T. Chin, Jim Blocker, John E. Larson, Phillip M. Jones, Paul B. Rawlins