Processor utilizing novel architectural ordering scheme
Various methods, apparatuses, and systems in which a processor includes an issue engine and an in-order execution pipeline. The issue engine categorizes operations as at least one of either a speculative operation which perform computations or an architectural operation which has potential to fault or cause an exception. Each architectural operation issues with an associated architectural micro-operation. A first micro-operation checks whether a first speculative operation is dependent upon an intervening first architectural operation. The in-order execution pipeline executes the speculative operation, the architectural operation, and the associated architectural micro-operations.
The present application is a Divisional of U.S. Ser. No. 10/247,894, filed Sep. 19, 2002.
FIELD OF THE INVENTIONEmbodiments of the invention generally relate to the field of computer processor architecture. More one or more embodiments of the invention relate specifically, to methods and apparatus for addressing problems associated with the execution of different sets of architectural semantics.
BACKGROUND OF THE INVENTIONDue to the physical designs of processor architectures, two or more clock cycles may occur between when the issuing engine issues an operation and when the issuing engine receives feedback regarding whether that issued operation has been executed or retired. Problems can occur if a mechanism is not in place during the interim time period between the occurrence of the issuing of the operation and the feedback to the processor on the execution/retirement of that operation. For example, data corruption can occur if a first operation results in an irreversible data change or state change external to the processor and a second operation executes after the first operation but anticipated using the original data or state.
Also, the continued growth of the microprocessor industry has lead to the development of competing processor architectures. Several prior processor designs try to maintain compatibility between different machines operating according to different instruction set architectures (ISAs). However, a problems in the industry exist in designing a microprocessor architecture to provide architectural compatibility with prior sets of instructions, while introducing a new instruction set architectures such as the reduced instruction set computer (RISC) designs.
One of the difficulties in implementing such a machine is how to superimpose the older, for example, 32-bit instruction semantics on a new, 64-bit architecture having a completely different set of semantics while minimizing the use of special hardware in the execution core of the machine.
A previous processor used an additional piece of hardware called a memory order buffer to handle memory ordering semantics. The processor included an out-of-order engine wherein operations are issued to the execution core of the processor before all of the control dependencies for those operations had been resolved. These operations are known as speculative operations. In the event that a particular operation's control dependencies are resolved to be false, the results of the operation are ignored. However, some operations, such as STORE operations, cannot be performed speculatively as they update the architectural state external to the processor. This processor uses the memory order buffer to resolve this potential data corruption conflict.
For example, a STORE is not issued to the execution engine, but instead is placed into the memory order buffer to hold the STORE addresses and associated data. The STORE is then issued when all the control dependencies have been resolved for that particular operation. To provide correct data for speculative LOADs, the execute engine snoops the speculative store buffer for speculative STOREs to the LOAD address. If a match was found, data was provided from the speculative store buffer. If the Store address is unknown, the LOAD must wait until the STORE address computation result is available.
Thus, the memory order buffer is typically closely coupled with the processor. The memory complex continually receives requests and sends responses to the memory order buffer (MOB). The issue engine (e.g., for issuing instructions) also should couple with the MOB in order to indicate when a STORE is eligible for retirement, and hence, must be considered a committed STORE. The specific problem with this approach is that in an out-of-order machine handling different architectural semantics the issue engine is typically remote from the execute engine; therefore, any access of the machine's architectural state requires many clock cycles. The issue engine is thus unable to rely on architectural state or instruction results when making issuing decisions.
This problem is best illustrated by considering the problem encountered for LOAD operations. First, a determination of whether a LOAD should be blocked due to an unknown STORE address might typically require waiting 7-8 clocks after the address generation micro-operations (uops) have been issued from the issue engine. Again, this delay is due to the physical distance between the scheduling logic and the processor's execution units.
Other prior art processors add a piece of hardware to maintain a list of speculative LOAD addresses and issues STOREs non-speculatively, and in-order. If an address conflict occurs, the LOAD causes a machine flush and re-execution when it comes time for retirement.
Yet another approach is embodied in the HAL, out-of-order implementation of the SPARC™ V9 architecture. This machine sequentializes the address generation component of the memory hierarchy. The address generation component guarantees older STORE addresses are generated before any younger STORE address. Data is then forwarded between the older STOREs and the younger LOADs.
BRIEF DESCRIPTION OF THE DRAWINGSThe present embodiments of the invention will be understood more fully from the detailed description which follows and from the accompanying drawings, which, however, should not be taken to limit the invention to the specific embodiments shown, but rather are for explanation and understanding only.
In general, a processor having an out-of-order issue engine using two different sets of architectural semantics to insure architectural consistency is described. In the following description, numerous specific details are set forth, such as particular micro-operation sequences, pipeline stages, bit sizes, etc., in order to provide a thorough understanding the invention. Practitioners having ordinary skill in the data processing arts will understand that the embodiments of the invention may be practiced without many of these details. In other instances, well-known signals, components, and circuits have not been described in detail to avoid obscuring the embodiments of the invention.
The processor 10 further comprises a second engine, such as an execution engine 30, which processes instructions having a second set of architectural semantics. The second engine includes a data cache 34 which maintains an architectural state of the processor. The second engine also includes an in-order execution pipeline 33 which is coupled to the scheduling unit 23 of the first engine. The execution pipeline 33 executes the uops which have been dispatched by the scheduling unit 23. The results from the execution of the uops are then written into the data cache 34. In addition, the results are transmitted to the retirement unit 24 of the first engine.
Note, pipeline processing may be a category of techniques that provide simultaneous, or parallel, processing within the computer. Pipeline processing refers to overlapping operations by moving data or instructions into a conceptual pipe with all stages of the pipe processing simultaneously. For example, while one instruction is being executed, the computer is decoding the next instruction. In vector processors, several steps in a floating point operation can be processed simultaneously.
In an embodiment, the processor 10 provides architectural consistency in cases where the execution resources of the machine are many clock cycles away from the out of order issue engine 20, and where the out of order issue engine 20 cannot access the architectural state. This processor 10 may execute software that was written to run on its architecture as well as emulate another model and execute software that was written to run in the other machine.
In an embodiment, an architectural ordering model implemented in the processor 10 supports two categories of issuing semantics: speculative operations and architectural operations. Speculative operations are those that can be issued as soon as their data dependencies are satisfied but before their control dependencies are resolved. Architectural operations, on the other hand, can be issued only when all older operations in program order, such as speculative or other architectural operations, have been issued. Architectural operations include operations which cause the execution pipeline of the processor to be flushed if the operation faults.
Potentially excepting operations are decomposed into two separate uops. First, a speculative micro-operation may be used to generate the data results speculatively, so that the operations which are dependent upon the results can also be speculatively issued. This is followed by an architectural micro-operation, which signals the faulting condition for the operation. In accordance with the architectural ordering model, a STORE becomes an architectural operation, and all previous faulting conditions are guaranteed to have evaluated before the STORE is executed. However, the STORE operation can be issued speculatively before issue pipeline resolves all of the faults from operations issued earlier in program. In this way, architectural operations have no data dependancies. For pipelines having a period of many clock cycles between operation issue and execution, many operations may be being processed simultaneously. However, a STORE operation is not required to wait to issue until all of these operations execute or retire. The STORE operation may issue when all of these operations earlier in program order issue. This removes some issuing time constraints from the issue queue by allowing STORE operations to issue much quicker than in other prior methods.
Further, the processor 10 may provide significant performance benefits by reducing pressure on the retire queues of the retirement unit. Additionally, STOREs may be presented to the in-order machine faster, and therefore STOREs are passed by fewer speculative LOADs.
In an embodiment, processor 10 includes an out-of-order issue-engine such as an Intela architecture value engine (iVE) 20 which supports instructions written for the existing iA-32 Intelâ architecture. The primary execution engine 30 processes instructions written with differing architectural semantics for 64-bit instruction processing. Also, the execution engine 30 may be an enhanced mode (EM) engine.
It should be understood that in order to maintain compatibility with the older instruction architectures (iA) such as an iA-32 architecture, the out-of-order issue-engine 20 may be based on an out-of-order execution paradigm. Out-of-order execution implies executing an operation as soon as all resources (e.g., source operand inputs) to the operation are ready and available. This means that an out-of-order machine does not necessarily execute instructions in a traditional von-Neumann order as in the original instruction stream. For example, if an original program consisted of an in-order instruction sequence A, B, C, D, an out-of-order engine may execute this sequence as A, D, C, B. Essentially, the out-of-order engine of a processor attempts to find the longest critical path of a program and thereafter spends most of the time in this path, while other paths are evaluated in parallel. It also tries to remove artificial dependencies created by inefficiencies in programming or a given architecture, such as register shortages, control dependencies, cache misses, and other dynamic effects that limit pre-runtime compliance.
The out-of-order issue-engine 20 of processor 10 relies upon instruction cache 31 in execution engine 30 for feeding iA instructions to its issue pipeline. These iA instructions are decoded by decoder 21 which is located in out-of-order issue-engine 20. The decoding process takes each iA instruction and breaks it down into more primitive operations or steps—commonly referred to as micro-operations (uops). Renamer 22 performs well-known register renaming functions.
Following renaming, uops are fed into a scheduler where they are scheduled for dispatch to an available execution unit. Scheduling may be based on source data dependencies and execution resource availability. The scheduling and dispatch of uops operations is represented in
An aspect of processor 10 may be that out-of-order issue-engine 20 relies on execution engine 30 for register files, execution resources, and memory accesses through the cache and bus complex. For example,
Once a particular operation has been completed, out-of-order issue-engine 20 records this information and updates its data structures at the retirement phase. It should be understood that out-of-order issue-engine 20 does not maintain data. Instead, it controls manipulation of data which physically resides in execution engine 30. This manipulation of data may occur either by tracking, or monitoring, or other data processing operations.
Another aspect of processor 10 may be that execution engine 30 executes instructions in-order. This means that execution engine 30 relies on the software writer/compiler to perform necessary code scheduling in accordance with the instruction set architecture of the execution engine 30. On the other hand, out-of-order issue-engine 20 performs code scheduling dynamically at run time to extract as much performance as possible. This run time optimization occurs for iA code, as most of the code which already exists cannot be recompiled. Thus, the out-of-order issue-engine scheduler logic may be responsible for analyzing data dependencies of operations and dispatching them to execution engine execution pipeline 33 based on operating execution unit availability.
Due to the inevitability of branches and exceptions, out-of-order issue-engine 20 may schedule operations such that no system state is affected by rescheduled operations which cannot be rolled back when an exception/branch is taken on an earlier operation in program order. This condition is illustrated in
The processor implements an architectural ordering model in which the issue agent of out-of-order issue-engine 20 supports two different issuing semantics. These semantics results in operations being categorized in two different ways: either as speculative operations or as architectural operations.
Speculative operations are operations that can be issued as soon as their data dependencies are satisfied and before their control dependancies are satisfied. Speculative operations execute whenever their data is ready (e.g., source operands have been computed). For this category, an ordinary re-order buffer (ROB) may be utilized to place execution results in proper execution order. In addition, the ROB may be used to generate faults, if necessary.
Architectural operations issue when all older operations in program order—either speculative or architectural—have been issued. Thus, an architectural operation may be basically any instruction that can fault. These operations are constrained to execute in the original program order. Thus, architectural operations generate no data dependencies. To put it another way, architectural operations do not produce any data for a computation. Architectural operations can and will flush the pipeline if the operation faults. In this way, architectural operations maintain proper processor state in the case of exceptions.
Practitioners familiar with computer architecture will appreciate that in the absence of a memory order buffer, if the processor were to not release STORE operations until retirement, out of necessity, LOAD operations would also have to be blocked. The reason why may be because there would be no way to determine whether data associated with a particular LOAD operation is valid. In other words, functionality could not be guaranteed.
In accordance with an embodiment of the architectural ordering model, potentially expecting operations 306 are decomposed into two separate uops. A speculative uop may be used to generate data results speculatively, so that the operations dependent upon its results can be likewise speculatively issued. In the instruction stream this may be followed by an architectural micro-operation 307, which signals the faulting condition for the operation. A STORE operation 208 becomes an architectural operation in the processor, and all previous faulting conditions are guaranteed to have been evaluated before the STORE is issued.
Note that the architectural model deals with the issuing semantics. Because the execution pipeline is in-order, all operations that have been issued to the execution pipeline are evaluated in-order. In the issue domain, however, the concept of architectural ordering guarantees that all faults are resolved before STORE hits occur.
To reiterate, architectural operations produce no data, but merely signal the presence of faults, and are utilized to flush the pipeline. Those of ordinary skill in the art will appreciate that this allows the release of STORE operations 208 at issue time in out-of-order issue-engine 20. One of the consequences of the concept of architectural ordering in the processor may be that there are two different instruction streams coming out of the scheduler: speculative operations, (which perform all of the computations) and architectural operations (which both resolve faults and basically issue STOREs).
The architectural ordering model thus provides consistent LOAD/STORE behavior and scheduling without the drawbacks associated with having additional components such as a memory ordering buffer. Some of the advantages of architectural ordering may be as follows. Whereas previous architectures have deferred the execution of STORE operations 208 until retirement (i.e., resolution of all control dependencies), in the invented processor, architectural operations, such as a store operation 208, are issued as soon as all previous operations in program order have been issued, but not necessarily evaluated. This means that the out of order issue engine may be effectively de-coupled from the retirement engine. If the execute engine is multiple clock cycles from the issue engine, multiple unevaluated architectural operations can still remain in the pipeline. Those of ordinary skill will appreciate that this provides important performance benefits.
One benefit of architectural ordering may be that it reduces pressure on the retirement and issue queues. If STORE operations 208 were deferred until retirement, large delays would develop between issue time and retirement time (due to the long latencies associated with the execution engine). This would mean that the retirement pointer typically would have to wait until an operation has fully passed through the execution pipeline before it could be advanced. In the situation where two STORE operations 208 occurred consecutively, the issue pointer for the second STORE might be more than two times the length of the execution pipeline from the retirement pointer. Of course, this distance increases linearly with the number of consecutive STOREs being executed.
A second performance advantage that an embodiment of the processor 10 provides is that STORE operations 208 are presented to the in-order portion (execution engine 30) faster; therefore the STORE operations 208 are passed by fewer speculative LOADs. This is because the in-order portion of processor 10 does not support speculative STOREs. In other words, no forwarding is available for speculative LOADs that wish to use the STORE operation's 208 contents. When a STORE is made visible to the in-order execution portion of the machine, the overlapping addresses of speculative LOADs simply create faults. In an embodiment, the overlapping addresses of speculative loads create faults through the mechanism of advanced LOADs, discussed below. Because the out of issue engine does not need to rely upon architectural data, speculative LOADs and STOREs may be issued without performing address comparisons.
The processor issues STORE operations 308 309 to memory when every preceding operation in program order has successfully completed. Therefore, the architectural order issue model relies upon two separate issuing semantics. All potentially faulting (i.e., LOAD/STORE) and control (i.e., branch) instructions have an associated architectural operation, referred to as an “arch_op” 307. Updates which cannot be rolled back are a side effect of arch_ops 307. These include, for example, a STORE issued to memory operation.
Secondly, arch_ops 307 are issued when all older operations in program order (older instructions and older uops for a current instruction) have been issued. As mentioned above, arch_ops 307 are issued in strict program order with respect to one another. In an embodiment, only one arch_op 307 may be issued per clock cycle of the processor. The execution engine 30 of processor 10 flushes the execution pipeline 33, when an arch_op 307 with a fault is executed or a mis-predicted branch is encountered in the program). Execution pipeline 33 signals out-of-order issue-engine 20 when this happens.
It should be understood that instruction issuance with architectural ordering means that uops D2 306, STORE2 309, and arch_op 307, are issued in strict program order. Note that, in this example, even though STORE1 308 is issued earlier, it does not change or alter memory. However, the STORE1 308 uop does allow for address computation to be performed as early as possible.
Continuing with the example, when uop D2 306 is executed in the execution pipeline it flushes the pipeline, which also results in eliminating the STORE2 306 uop. In this example, uop D2 306 may represent a mis-predicted branch or a faulting instruction.
In an embodiment, one benefit of restricting the processor to issuing one arch_op 307 per clock cycle is that it simplifies scheduling in out-of-order issue-engine 20, as it does not impact performance. Note that the execution engine portion of processor 10 resolves more than one arch_op 307 in a bundle group according to a fixed order. For example, the fixed order may be left to right.
The gen_efa uop 406 and gen_la uop 408 generate effective and linear addresses, respectively. The gen_efa uop 406 may not always be needed as gen_la uop 408 can create linear addresses directly for all addressing modes in the architectural instructions except base+index+displacement and base+index modes. The adv_load uop 409 is an advanced LOAD operation which is performed speculatively. Essentially, the idea of an advanced LOAD 409 is to start the LOAD operation as early as possible, giving it as much time as possible to complete before any instructions which are dependent upon the LOAD are encountered. As explained earlier, STORE operations traditionally have been a barrier as to how far ahead a LOAD instruction could be moved. The reason why is because compilers often cannot determine if a LOAD and a STORE instruction possibly conflict, in which case they may be reading and writing data from the same memory location. The adv_load uop 409 allows the LOAD operation to pass in execution order the STORE, which allows greater parallelism.
The chk_load uop 412 may be a check LOAD operation that verifies if any intervening STORE happens to update any one of the bytes accessed by the associated LOAD. Note that all of the LOAD uops shown in
When an advanced LOAD is executed in pipeline 33, it may be logged into a structure known as an advanced load address table (ALAT).
The physical ALAT array 505 shown in
The speculative invalidation pipeline 504 keeps track of events and instructions that invalidate ALAT entries until they are non-speculative. In some cases, the comparison is made on fewer bits, in order to facilitate a high frequency pipeline. Prioritization logic block 503 prioritizes between the advanced LOADs, STOREs, and invalidations that are in the pipeline. According to the prioritization scheme, earlier instructions take precedence over later instructions.
The last basic sub-component of ALAT 500 is the check look-up logic 501, which responds to check requests being made. Logic Block 501 queries both the physical ALAT array 505 (for non-speculative accesses) and prioritization logic block 503 (for speculative accesses), using the register ID as the tag for the request. It reports information from the prioritization logic over information from the physical ALAT array, if both happen to respond.
Thus, ALAT 500 is basically a sixteen-deep, first-in-first-out (FIFO) stack that remembers linear addresses and destination register identifiers for the last sixteen advanced LOAD uops. When the ALAT is full, the oldest entry is discarded. For any LOAD that has been potentially boosted above an intervening STORE or STOREs, it is important to know if any one of the intervening STOREs overlap the address for the LOAD. One of the functions of ALAT 500 is to keep track of this information.
When a chk_load operation is encountered, ALAT 500 indicates that an earlier STORE did overlap with the address of the associated LOAD. In such an instance the original LOAD is re-executed to obtain the most recent data. The pipelines of both execution engine 30 and out-of-order issue-engine 20 are also flushed, and the instruction stream restarted from the instruction immediately following the LOAD. This is necessary as the data consumers of adv_load may have received incorrect data. Note that reissuing of the subsequent instructions (after the LOAD) requires flushing the pipeline of out-of-order issue-engine 20 to re-create dependency information.
Referring again to
Code sequence B 604 represents a traditional approach to optimizing code sequence A 602. This optimization may be implemented for example, by a compiler moving the LOAD operation 608 as far ahead in the code as possible. Note that in code sequence B 604, the LOAD 608 is two clocks away from the dependant AND operation 609. However, unless the compiler can determine that R9 (the exemplary address of the LOAD 608) and R4 (the exemplary address of the earlier STORE operation 607) refer to different memory address, it is not permitted to move the LOAD 608 pass the STORE 607. The reason why is because if the LOAD 608 and STORE 607 are to the same address, the LOAD 608 needs to obtain the data form the STORE 607. This requirement is violated if the LOAD 608 is earlier in the program order.
Code sequence C 606 represents how the processor allows the LOAD 608 operation to be boosted past the STORE 607. This type of passing is permitted as long as a later check LOAD (ld.c) instruction 612 is used to make sure that a dependency problem does not exist. If the LOAD check 612 fails, the LOAD 608 needs to be transparently re-performed, and the dependent instructions 609 should observe the dependency. In an embodiment, in order to accommodate high performance in a superscalar implementation, the check LOAD instruction 612 has virtually no affect on the architectural state of the processor assuming that the check succeeds.
Proceeding from left to right in
Branch prediction also takes place in the first four stages of the front-end pipeline 706. At the same time that a line fetch request is issued to execution engine 30, a branch target buffer (BTB) of the processor is consulted to determine if there is a known branch in the line being fetched. If a branch is present, it can be predicted.
Instruction cache 32 of execution engine 30 may be organized on, for example, a 32-byte line basis. Therefore, when the instruction bytes are returned to out-of-order issue-engine 20, either the upper or lower half of the line is selected before being transferred over signal lines 12. The line fetch request takes two clocks: one for instruction pointer generation and a second for instruction cache lookup. The lower or upper half is selected in a third clock cycle, shown as the Rotate/Transmit pipestage. All architectural instruction are byte aligned and can be between 1-15 bytes long.
Since the code parcels received from instruction cache 32 are 16-byte aligned, the iA instructions need to be extracted from these parcels before being decoded. This process is called instruction alignment/steering, and occurs in the ALN and LEN pipestages. Instructions are decoded in out-of-order issue-engine 20 at a rate of one instruction per clock. Decoding operations are shown occurring in the DE1-DE4 pipestages. At the end of the alignment (ALN) stage, a micro-ROM address is produced. This address starts a microcode sequence for the instruction currently being decoded. The microcode sequence is produced during the MS1-MS3 pipestages.
Signal flight stages SF0-SF1 are not required for functionality, and merely represent an artifact of the large chip floorplan for one embodiment. These stages cover the time required for transmission of signals and information across physically distant sections of the chip. The back-end of the machine 708 receives an in-order stream of uops and re-orders them based on information such as input data dependencies, operation latency, and execution resource availability. These operations are executed out-of-order based on actual dependencies.
Renaming operations take place in rename stages RN1-RN3, as shown in
Instruction scheduling and dispatch is performed in the Ready/Schedule (RDY/SCH) and dispatch stages. Following renaming, pairs of uops are written into a structure which is the equivalent of a reservation station. This occurs at the end of the RN3 pipestage. The structure that the uops are written into comprises a dependency matrix and a uop waiting buffer. The uop waiting buffer is simply a holding structure where uops are held until they can be dispatched to execution engine execution pipeline 33.
Every set of four uops that is transmitted by out-of-order issue-engine 20 to execution engine 30 enters execution pipeline 33 starting at the WLD stage. Once in execution pipeline 33, the set of four uops proceeds in a lock/step manner. Execution of micro-operations is performed after sources have been read, followed by exception detection and write/back into the execution engine register file. The write/back (WRB) stage of the execution engine pipeline is also used to transmit execution status (i.e., exception information) back to out-of-order issue-engine 20.
Because the instruction execution status information is stored in a location that is a considerable physical distance away from out-of-order issue-engine 20, a WRB 1 stage is needed to accommodate signal flight time. The execution status information is eventually recorded in re-order queue (ROQ) by the end of the same clock cycle. Retirement logic uses this information to update its data structures. It is appreciated that checks for exceptions and appropriate redirection is performed as part of the retirement process. For example, redirection of an exception may take up to three pipestages to complete.
Computer system 800 further comprises a random access memory (RAM), or other dynamic storage device 804 (referred to as main memory) coupled to bus 811 for storing information and instructions to be executed by processor 812. Main memory 804 also may be used for storing temporary variables or other intermediate information during execution of instructions by processor 812. In an embodiment, the Processor 812 may include a microprocessor, but is not limited to a microprocessor, such as a Pentium, PowerPC, etc.
Computer system 800 also comprises a read only memory (ROM) and/or other static storage device 806 coupled to bus 811 for storing static information and instructions for processor 812, and a mass storage memory 807, such as a magnetic disk or optical disk and its corresponding disk drive. Mass storage memory 807 is coupled to bus 811 for storing information and instructions.
While some specific embodiments of the invention have been shown the invention is not to be limited to these embodiments. For example, most functions performed by electronic hardware components may be duplicated by software emulation. Thus, a software program written to accomplish those same functions may emulate the functionality of the hardware components in input-output circuitry. The invention is to be understood as not limited by the specific embodiments described herein, but only by scope of the appended claims.
Claims
1. A processor comprising:
- a first engine to processes instructions having a first set of architectural semantics, the first engine including a decoder unit that decomposes each of the instructions into one or more micro-operations, a scheduling unit that dispatches the micro-operations in an out-of-order manner, and a retirement unit; and
- a second engine to processes instructions having a second set of architectural semantics, the second engine including a register to maintain an architectural state of the processor, and an in-order execution pipeline coupled to the scheduling unit of the first engine, wherein the execution pipeline to execute the micro-operations which have been dispatched by the scheduling unit, results from the execution of the micro-operations being written into the register, the results also transmitted to the retirement unit of the first engine.
2. The processor according to claim 1, further comprising:
- a cache memory complex associated with the second engine, the cache complex being coupled to the in-order execution pipeline; and
- a result bus coupling the cache memory complex to the retirement unit.
3. The processor of claim 2, wherein the scheduling unit dispatches the micro-operations according to source data dependencies and execution resource availability.
4. The processor of claim 2, wherein each instruction of a category of instructions having the first of architectural semantics is issued by the first engine with an associated architectural operation, execution of the associated architectural operation causing the second engine to flush the execution pipeline in response to a first condition.
5. The processor of claim 4, wherein the first condition comprises either a false or a mis-predicted branch.
6. The processor of claim 5, wherein the category of instructions includes a STORE.
7. The processor of claim 4, wherein architectural operations associated with the category of instructions are issued in-order by the first engine.
8. The processor of claim 4, wherein the architectural operations associated with the category of instructions are issued one per clock cycle of the processor.
9. A method of operating a processor, comprising:
- processing using a first engine instructions having a first set of architectural semantics, comprising: decomposing each of the instructions into one or more micro-operations; dispatching the micro-operations in an out-of-order manner; and
- processing using a second engine instructions having a second set of architectural semantics, comprising: maintaining an architectural state of the processor in a register; executing the micro-operations which have been dispatched by the scheduling unit using in-order execution pipeline; writing results from the execution of the micro-operations into the register; and transmitting the results from the execution of the micro-operations to a retirement unit of the first engine.
10. The method of claim 9, further comprising dispatching the micro-operations according to source data dependencies and execution resource availability.
11. The method of claim 9, further comprising:
- issuing each instruction of a category of instructions having the first of architectural semantics with an associated architectural operation; and
- causing the second engine to flush the in-order execution pipeline in response to a first condition upon execution of the associated architectural operation.
12. The method of claim 11, further comprising issuing architectural operations associated with the category of instructions in-order using the first engine.
13. The method of claim 11, further comprising issuing architectural operations associated with the category of instructions one per clock cycle of the processor.
14. A system, comprising:
- a processor having: a first engine to processes instructions having a first set of architectural semantics, the first engine including a decoder unit that decomposes each of the instructions into one or more micro-operations, a scheduling unit that dispatches the micro-operations in an out-of-order manner, and a retirement unit; and a second engine to processes instructions having a second set of architectural semantics, the second engine including a register to maintain an architectural state of the processor, and an in-order execution pipeline coupled to the scheduling unit of the first engine, wherein the execution pipeline to execute the micro-operations which have been dispatched by the scheduling unit, results from the execution of the micro-operations being written into the register, the results also transmitted to the retirement unit of the first engine;
- a non-volatile memory; and
- a system bus coupled to the processor and the non-volatile memory.
15. The system according to claim 14, wherein the processor further comprises:
- a cache memory complex associated with the second engine, the cache complex being coupled to the in-order execution pipeline; and
- a result bus coupling the cache memory complex to the retirement unit.
16. The system according to claim 14, wherein the non-volatile memory comprises read-only-memory (ROM).
Type: Application
Filed: Apr 18, 2006
Publication Date: Sep 21, 2006
Inventors: Jeffery Baxter (Los Gatos, CA), Gary Hammond (Fort Collins, CO), Nazar Zaidi (San Jose, CA)
Application Number: 11/407,184
International Classification: G06F 9/30 (20060101);