Presentation is loading. Please wait.

Presentation is loading. Please wait.

Chapter 12 CPU Structure and Function. CPU Sequence Fetch instructions Interpret instructions Fetch data Process data Write data.

Similar presentations


Presentation on theme: "Chapter 12 CPU Structure and Function. CPU Sequence Fetch instructions Interpret instructions Fetch data Process data Write data."— Presentation transcript:

1 Chapter 12 CPU Structure and Function

2 CPU Sequence Fetch instructions Interpret instructions Fetch data Process data Write data

3 CPU With Systems Bus

4 CPU Internal Structure

5 Registers CPU must have some working space (temporary or scratch pad storage) Top level of memory hierarchy Number and function vary between processor designs

6 User Visible Registers General Purpose Data Address Control

7 General Purpose Registers May be true general purpose May be used for data or addressing May be restricted Data —May include Accumulator Addressing —May include Segment Register(s)

8 General Purpose Registers Design Decision Make them general purpose ? —Increase flexibility and programmer options —Increase instruction size & complexity Make them specialized —Smaller (faster) instructions —Less flexibility

9 How Many GP Registers? Between 8 – 32 ? Fewer = more memory references More does not tend to reduce memory references and takes up processor real estate

10 How big? Large enough to hold full address Large enough to hold full word - Sometimes possible to combine two data registers

11 Control & Status Registers Program Counter Instruction Decoding Register Memory Address Register Memory Buffer Register Program Status Word

12 Control - Condition Code Registers CC: Sets of individual bits —e.g. result of last operation was zero Can be read (implicitly) by programs —e.g. Jump if zero Usually can not be set by programs

13 Program Status Word A set of status and/or control bits Includes Condition Codes Priority, Interrupt enable/disable Supervisor State information - Kernel Mode - Not available to user programs (Used by operating system) (Allows privileged instructions to execute)

14 Other Registers May have registers pointing to: —Process control blocks —Interrupt Vectors Note: CPU design and operating system design are closely linked

15 Example Register Organizations

16 Instruction Cycle with Indirect Note: ‘Indirect’ allows for fetching data with indirect addressing

17 Data Flow (Fetch Diagram)

18 Data Flow (Instruction Fetch) Fetch —PC contains address of next instruction —Address moved to MAR —Address placed on address bus —Control unit requests memory read —Result placed on data bus, copied to MBR, then to IR —Meanwhile PC incremented by 1 (or more)

19 Data Flow (Indirect Diagram)

20 Data Flow (Data Fetch) IR is examined If indirect addressing, indirect cycle is performed —N bits of MBR transferred to MAR —Control unit requests memory read —Result (address of operand) moved to MBR

21 Data Flow (Execute) May take many forms Depends on instruction being executed May include —Memory read/write —Input/Output —Register transfers —ALU operations

22 Data Flow (Data Store) If indirect addressing, indirect cycle is performed —N bits of MBR transferred to MAR —Control unit requests memory read —Result (address of operand) moved to MBR

23 Data Flow (Interrupt) Context Stored / Interrupt Acknowledged Vector Fetched & Intr Serv Routine Addr => PC Intr Serv Routine (Handler) executed …. Context Restored Continue execution of main program

24 Instruction Cycle State Diagram

25 Prefetch Consider the instruction sequence as: Fetch instruction Execution instruction (often does not access main memory) Can computer fetch next instruction during execution of current instruction ? Called instruction Prefetch What are the implications of Prefetch?

26 Improved Performance with Prefetch Improved but not doubled: —Fetch usually shorter than execution —Any jump or branch means that prefetched instructions are not the required instructions Could we Prefetch more than one instruction ? Could we add “more stages” to improve performance even more? This is Pipelining

27 Pipelining Consider the instruction sequence as: instruction fetch, decode instruction, fetch data, execute instruction, store result, check for interrupt Consider it as an “assembly line” of operations. Then we can begin the next instruction assembly line sequence before the last has finished. Actually we can fetch the next instruction while the present one is being decoded. This is pipelining This is pipelining.

28 Two Stage Instruction Pipeline

29 Define Pipeline “stations” Fetch instruction (FI) Decode Instruction (DI) Calculate Operand Addresses (CO) Fetch Operands (FO) Execute Instruction (EI) Write Operand (WO) Let’s define some possible Pipeline stations:

30 Timing Diagram for Instruction Pipeline Operation

31 The Effect of a Conditional Branch on Instruction Pipeline Operation Instruction 3 is a conditional branch to instruction 15:

32 Alternative Pipeline Depiction Instruction 3 is conditional branch to instruction 15:

33 Speedup Factors with Instruction Pipelining

34 Dealing with Branches – Possible approaches Multiple Streams Prefetch Branch Target Loop Buffer Branch Prediction Delayed Branching

35 Multiple Streams Have two pipelines Prefetch each branch into a separate pipeline Use appropriate pipeline Challenges: Leads to bus & register contention Multiple branches lead to further pipelines being needed

36 Prefetch Branch Target Target of branch is prefetched in addition to instructions following branch Keep target until branch is executed Used by IBM 360/91

37 Loop Buffer Use Very fast memory (“Loop Buffer Cache”) Maintained by fetch stage of pipeline Check buffer before fetching from memory Very good for small loops or jumps in small code sections Used by CRAY-1

38 Branch Prediction Predict branch never taken Predict branch always taken Predict by opcode Predict branch taken/not taken switch Maintain branch history table

39 Predict Branch Taken / Not taken Predict never taken —Assume that jump will not happen —Always fetch next instruction —68020 & VAX 11/780, VAX will not prefetch after branch if a page fault would result (O/S v CPU design) Predict always taken —Assume that jump will happen —Always fetch target instruction Which is better?

40 Branch Prediction by Opcode / Switch Predict by Opcode —Some instructions are more likely to result in a jump than others —Can get up to 75% success with this stategy Taken/Not taken switch —Based on previous history —Good for loops —Perhaps good to match programmer style

41 Maintain Branch Table Perhaps a cache table of three entries: - Address of branch - History of branching - Targets of branch

42 Intel Pipelining Fetch (Fetch) —From cache or external memory —Put in one of two 16-byte prefetch buffers —Fill buffer with new data as soon as old data consumed —Average 5 instructions fetched per load —Independent of other stages to keep buffers full Decode stage 1 (D1) —Opcode & address-mode info —At most first 3 bytes of instruction —Can direct D2 stage to get rest of instruction Decode stage 2 (D2) —Expand opcode into control signals —Computation of complex address modes Execute (EX) —ALU operations, cache access, register update Writeback (WB) —Update registers & flags —Results sent to cache & bus interface write buffers

43 80486 Instruction Pipeline Examples


Download ppt "Chapter 12 CPU Structure and Function. CPU Sequence Fetch instructions Interpret instructions Fetch data Process data Write data."

Similar presentations


Ads by Google