CPE 731 Advanced Computer Architecture ILP: Part V – Multiple Issue Dr. Gheith Abandah Adapted from the slides of Prof. David Patterson, University of.

Slides:



Advertisements
Similar presentations
CPE 731 Advanced Computer Architecture Instruction Level Parallelism Part I Dr. Gheith Abandah Adapted from the slides of Prof. David Patterson, University.
Advertisements

Lecture 8 Dynamic Branch Prediction, Superscalar and VLIW Advanced Computer Architecture COE 501.
1 COMP 206: Computer Architecture and Implementation Montek Singh Wed, Oct 19, 2005 Topic: Instruction-Level Parallelism (Multiple-Issue, Speculation)
A scheme to overcome data hazards
ENGS 116 Lecture 101 ILP: Software Approaches Vincent H. Berk October 12 th Reading for today: , 4.1 Reading for Friday: 4.2 – 4.6 Homework #2:
Instruction Level Parallelism 2. Superscalar and VLIW processors.
CPE 631: ILP, Static Exploitation Electrical and Computer Engineering University of Alabama in Huntsville Aleksandar Milenkovic,
Instruction Level Parallelism María Jesús Garzarán University of Illinois at Urbana-Champaign.
COMP4611 Tutorial 6 Instruction Level Parallelism
POLITECNICO DI MILANO Parallelism in wonderland: are you ready to see how deep the rabbit hole goes? ILP: VLIW Architectures Marco D. Santambrogio:
ILP: Loop UnrollingCSCE430/830 Instruction-level parallelism: Loop Unrolling CSCE430/830 Computer Architecture Lecturer: Prof. Hong Jiang Courtesy of Yifeng.
1 COMP 740: Computer Architecture and Implementation Montek Singh Tue, Feb 24, 2009 Topic: Instruction-Level Parallelism IV (Software Approaches/Compiler.
1 Lecture 18: VLIW and EPIC Static superscalar, VLIW, EPIC and Itanium Processor (First introduce fast and high- bandwidth L1 cache design)
Computer Architecture Instruction Level Parallelism Dr. Esam Al-Qaralleh.
CSE 502 Graduate Computer Architecture Lec 11 – More Instruction Level Parallelism Via Speculation Larry Wittie Computer Science, StonyBrook University.
Dynamic Branch PredictionCS510 Computer ArchitecturesLecture Lecture 10 Dynamic Branch Prediction, Superscalar, VLIW, and Software Pipelining.
CS152 Lec15.1 Advanced Topics in Pipelining Loop Unrolling Super scalar and VLIW Dynamic scheduling.
Pipelining 5. Two Approaches for Multiple Issue Superscalar –Issue a variable number of instructions per clock –Instructions are scheduled either statically.
1 Advanced Computer Architecture Limits to ILP Lecture 3.
Computer Architecture Lec 8 – Instruction Level Parallelism.
Lecture 3: Chapter 2 Instruction Level Parallelism Dr. Eng. Amr T. Abdel-Hamid CSEN 601 Spring 2011 Computer Architecture Text book slides: Computer Architec.
CS252 Graduate Computer Architecture Lecture 6 Static Scheduling, Scoreboard February 6 th, 2012 John Kubiatowicz Electrical Engineering and Computer Sciences.
1 Lecture 10: Static ILP Basics Topics: loop unrolling, static branch prediction, VLIW (Sections 4.1 – 4.4)
CPE432 Chapter 4C.1Dr. W. Abu-Sufah, UJ Chapter 4C: The Processor, Part C Read Section 4.10 Parallelism and Advanced Instruction-Level Parallelism Adapted.
CS136, Advanced Architecture Speculation. CS136 2 Outline Speculation Speculative Tomasulo Example Memory Aliases Exceptions VLIW Increasing instruction.
CPE 731 Advanced Computer Architecture ILP: Part IV – Speculative Execution Dr. Gheith Abandah Adapted from the slides of Prof. David Patterson, University.
DAP Spr.‘98 ©UCB 1 Lecture 6: ILP Techniques Contd. Laxmi N. Bhuyan CS 162 Spring 2003.
EECC551 - Shaaban #1 lec # 6 Fall Multiple Instruction Issue: CPI < 1 To improve a pipeline’s CPI to be better [less] than one, and to.
Prof. John Nestor ECE Department Lafayette College Easton, Pennsylvania ECE Computer Organization Lecture 19 - Pipelined.
COMP381 by M. Hamdi 1 Superscalar Processors. COMP381 by M. Hamdi 2 Recall from Pipelining Pipeline CPI = Ideal pipeline CPI + Structural Stalls + Data.
1 COMP 206: Computer Architecture and Implementation Montek Singh Mon., Oct. 9, 2002 Topic: Instruction-Level Parallelism (Multiple-Issue, Speculation)
Prof. John Nestor ECE Department Lafayette College Easton, Pennsylvania Computer Organization Pipelined Processor Design 3.
EECC551 - Shaaban #1 lec # 8 Winter Multiple Instruction Issue: CPI < 1 To improve a pipeline’s CPI to be better [less] than one, and to.
CIS 629 Fall 2002 Multiple Issue/Speculation Multiple Instruction Issue: CPI < 1 To improve a pipeline’s CPI to be better [less] than one, and to utilize.
COMP381 by M. Hamdi 1 Commercial Superscalar and VLIW Processors.
Ch2. Instruction-Level Parallelism & Its Exploitation 2. Dynamic Scheduling ECE562/468 Advanced Computer Architecture Prof. Honggang Wang ECE Department.
1 Overcoming Control Hazards with Dynamic Scheduling & Speculation.
1 Chapter 2: ILP and Its Exploitation Review simple static pipeline ILP Overview Dynamic branch prediction Dynamic scheduling, out-of-order execution Hardware-based.
Chapter 8 CPU and Memory: Design, Implementation, and Enhancement The Architecture of Computer Hardware and Systems Software: An Information Technology.
CSCE 614 Fall Hardware-Based Speculation As more instruction-level parallelism is exploited, maintaining control dependences becomes an increasing.
Advanced Pipelining 7.1 – 7.5. Peer Instruction Lecture Materials for Computer Architecture by Dr. Leo Porter is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike.
CS 5513 Computer Architecture Lecture 6 – Instruction Level Parallelism continued.
现代计算机体系结构 1 主讲教师:张钢 教授 天津大学计算机学院 通信邮箱: 提交作业邮箱: 2012 年.
Ch2. Instruction-Level Parallelism & Its Exploitation 2. Dynamic Scheduling ECE562/468 Advanced Computer Architecture Prof. Honggang Wang ECE Department.
Use of Pipelining to Achieve CPI < 1
IBM System 360. Common architecture for a set of machines
CS 352H: Computer Systems Architecture
/ Computer Architecture and Design
CPE 731 Advanced Computer Architecture ILP: Part V – Multiple Issue
COMP 740: Computer Architecture and Implementation
CS203 – Advanced Computer Architecture
CS203 – Advanced Computer Architecture
Henk Corporaal TUEindhoven 2009
CPSC 614 Computer Architecture Lec 5 – Instruction Level Parallelism
CPE 631 Lecture 13: Exploiting ILP with SW Approaches
David Patterson Electrical Engineering and Computer Sciences
Lecture 8: ILP and Speculation Contd. Chapter 2, Sections 2. 6, 2
Adapted from the slides of Prof
Lecture 23: Static Scheduling for High ILP
Henk Corporaal TUEindhoven 2011
Larry Wittie Computer Science, StonyBrook University and ~lw
CC423: Advanced Computer Architecture ILP: Part V – Multiple Issue
CPSC 614 Computer Architecture Lec 5 – Instruction Level Parallelism
Adapted from the slides of Prof
Chapter 3: ILP and Its Exploitation
CSC3050 – Computer Architecture
CPE 631 Lecture 14: Exploiting ILP with SW Approaches (2)
Overcoming Control Hazards with Dynamic Scheduling & Speculation
Lecture 5: Pipeline Wrap-up, Static ILP
Presentation transcript:

CPE 731 Advanced Computer Architecture ILP: Part V – Multiple Issue Dr. Gheith Abandah Adapted from the slides of Prof. David Patterson, University of California, Berkeley

4/15/2015CPE 731, ILP5 2 Outline Multiple Issue Statically-Scheduled Superscalar Processors Multiple Instruction Issue with Dynamic Scheduling VLIW Perspective

4/15/2015CPE 731, ILP5 3 Getting CPI below 1 CPI ≥ 1 if issue only 1 instruction every clock cycle Multiple-issue processors come in 3 flavors: 1.statically-scheduled superscalar processors, 2.dynamically-scheduled superscalar processors, and 3.VLIW (very long instruction word) processors 2 types of superscalar processors issue varying numbers of instructions per clock –use in-order execution if they are statically scheduled, or –out-of-order execution if they are dynamically scheduled VLIW processors, in contrast, issue a fixed number of instructions formatted either as one large instruction or as a fixed instruction packet with the parallelism among instructions explicitly indicated by the instruction (Intel/HP Itanium)

4/15/2015CPE 731, ILP5 4 Statically-Scheduled Superscalar Processors Example: Pipelined Instr. Fetch Unit Issue Packet FU 1 FU N … FDIntegerMW FDFPMW Issue 0, 1, or 2 instrs.

4/15/2015CPE 731, ILP5 5 Statically-Scheduled Superscalar Processors RAW hazards within issue packet: 1.Load FP followed by use FP 2.FP op followed by store FP Need additional FP RF ports Need more bypass paths

4/15/2015CPE 731, ILP5 6 Outline Multiple Issue Statically-Scheduled Superscalar Processors Multiple Instruction Issue with Dynamic Scheduling VLIW Perspective

4/15/2015CPE 731, ILP5 7 Multiple Instruction Issue with Dynamic Scheduling Issue multiple instructions involves 1.Assigning reservation stations 2.Updating the pipeline control tables May use split cycle or double HW Here we do branch prediction, but execution starts after branch resolution

4/15/2015CPE 731, ILP5 8 Superscalar Tomasulo Organization FP adders Add1 Add2 Add3 FP multipliers Mult1 Mult2 From Mem FP Registers Reservation Stations CDB1 and CDB2 To Mem FP Op Queue Load Buffers Store Buffers Load1 Load2 Load3 Load4 Load5 Load6

4/15/2015CPE 731, ILP5 9 Example Latencies: –Integer 1 cycle –Load 2 cycles –FP add 3 cycles –Branch 1 cycle (single issue) Loop: l.d f0, 0(r1) add.d f4, f0, f2 s.d f4, 0(r1) adddiu r1, r1, -8 bne r1, r2, loop

4/15/2015CPE 731, ILP5 10 Example (3 iterations) Loop: l.d f0, 0(r1) I A M W add.d f4, f0, f2 I E.. W s.d f4, 0(r1) I A M adddiu r1, r1, -8 I E W bne r1, r2, loop I E Loop: l.d f0, 0(r1) I A M W add.d f4, f0, f2 I E.. W s.d f4, 0(r1) I A M adddiu r1, r1, -8 I E W bne r1, r2, loop I E Loop: l.d f0, 0(r1) I A M W add.d f4, f0, f2 I E.. W s.d f4, 0(r1) I A M adddiu r1, r1, -8 I E W bne r1, r2, loop I E

4/15/2015CPE 731, ILP5 11 Example IPC = 15 / 16 = 0.94 The performance is limited by: 1.The FP unit is not fully utilized 2.High loop overhead (2 out of 5) 3.Control hazards; the l.d instruction is delayed one cycle.

4/15/2015CPE 731, ILP5 12 Example (with speculative execution) Loop: l.d f0, 0(r1) I A M W C add.d f4, f0, f2 I E.. W C s.d f4, 0(r1) I A C adddiu r1, r1, -8 I E W C bne r1, r2, loop I E C Loop: l.d f0, 0(r1) I A M W C add.d f4, f0, f2 I E.. W C s.d f4, 0(r1) I A C adddiu r1, r1, -8 I E W C bne r1, r2, loop I E C Loop: l.d f0, 0(r1) I A M W C add.d f4, f0, f2 I E.. W C s.d f4, 0(r1) I A C adddiu r1, r1, -8 I E W C bne r1, r2, loop I E C

4/15/2015CPE 731, ILP5 13 Outline Multiple Issue Statically-Scheduled Superscalar Processors Multiple Instruction Issue with Dynamic Scheduling VLIW Perspective

4/15/2015CPE 731, ILP5 14 VLIW: Very Large Instruction Word Each “instruction” has explicit coding for multiple operations –In IA-64, grouping called a “packet” –In Transmeta, grouping called a “molecule” (with “atoms” as ops) Tradeoff instruction space for simple decoding –The long instruction word has room for many operations –By definition, all the operations the compiler puts in the long instruction word are independent => execute in parallel –E.g., 2 integer operations, 2 FP ops, 2 Memory refs, 1 branch »16 to 24 bits per field => 7*16 or 112 bits to 7*24 or 168 bits wide –Need compiling technique that schedules across several branches

4/15/2015CPE 731, ILP5 15 Recall: Unrolled Loop that Minimizes Stalls for Scalar 1 Loop:L.DF0,0(R1) 2 L.DF6,-8(R1) 3 L.DF10,-16(R1) 4 L.DF14,-24(R1) 5 ADD.DF4,F0,F2 6 ADD.DF8,F6,F2 7 ADD.DF12,F10,F2 8 ADD.DF16,F14,F2 9 S.D0(R1),F4 10 S.D-8(R1),F8 11 S.D-16(R1),F12 12 DSUBUIR1,R1,#32 13 BNEZR1,LOOP 14 S.D8(R1),F16; 8-32 = clock cycles, or 3.5 per iteration L.D to ADD.D: 1 Cycle ADD.D to S.D: 2 Cycles

4/15/2015CPE 731, ILP5 16 Loop Unrolling in VLIW Memory MemoryFPFPInt. op/Clock reference 1reference 2operation 1 op. 2 branch L.D F0,0(R1)L.D F6,-8(R1)1 L.D F10,-16(R1)L.D F14,-24(R1)2 L.D F18,-32(R1)L.D F22,-40(R1)ADD.D F4,F0,F2ADD.D F8,F6,F23 L.D F26,-48(R1)ADD.D F12,F10,F2ADD.D F16,F14,F24 ADD.D F20,F18,F2ADD.D F24,F22,F25 S.D 0(R1),F4S.D -8(R1),F8ADD.D F28,F26,F26 S.D -16(R1),F12S.D -24(R1),F167 S.D -32(R1),F20S.D -40(R1),F24DSUBUI R1,R1,#488 S.D -0(R1),F28BNEZ R1,LOOP9 Unrolled 7 times to avoid delays 7 results in 9 clocks, or 1.3 clocks per iteration (1.8X) Average: 2.5 ops per clock, 50% efficiency Note: Need more registers in VLIW (15 vs. 6 in SS)

4/15/2015CPE 731, ILP5 17 Problems with 1st Generation VLIW Increase in code size –generating enough operations in a straight-line code fragment requires ambitiously unrolling loops –whenever VLIW instructions are not full, unused functional units translate to wasted bits in instruction encoding Operated in lock-step; no hazard detection HW –a stall in any functional unit pipeline caused entire processor to stall, since all functional units must be kept synchronized –Compiler might predict function units, but caches hard to predict Binary code compatibility –Pure VLIW => different numbers of functional units and unit latencies require different versions of the code

4/15/2015CPE 731, ILP5 18 Intel/HP IA-64 “Explicitly Parallel Instruction Computer (EPIC)” IA-64: instruction set architecture bit integer regs bit floating point regs –Not separate register files per functional unit as in old VLIW Hardware checks dependencies (interlocks => binary compatibility over time) Predicated execution (select 1 out of 64 1-bit flags) => 40% fewer mispredictions? Itanium™ was first implementation (2001) –Highly parallel and deeply pipelined hardware at 800Mhz –6-wide, 10-stage pipeline at 800Mhz on 0.18 µ process Itanium 2™ is name of 2nd implementation (2005) –6-wide, 8-stage pipeline at 1666Mhz on 0.13 µ process –Caches: 32 KB I, 32 KB D, 128 KB L2I, 128 KB L2D, 9216 KB L3

4/15/2015CPE 731, ILP5 19 EPIC Bundle (128 bits): Instruction group is a sequence of consecutive instrs with no register dependencies among them. There is a stop mark at the end of each group Op1Op2Op3

4/15/2015CPE 731, ILP5 20 Perspective Interest in multiple-issue because wanted to improve performance without affecting uniprocessor programming model Taking advantage of ILP is conceptually simple, but design problems are amazingly complex in practice Conservative in ideas, just faster clock and bigger Processors of last 5 years (Pentium 4, IBM Power 5, AMD Opteron) have the same basic structure and similar sustained issue rates (3 to 4 instructions per clock) as the 1st dynamically scheduled, multiple- issue processors announced in 1995 –Clocks 10 to 20X faster, caches 4 to 8X bigger, 2 to 4X as many renaming registers, and 2X as many load-store units  performance 8 to 16X Peak v. delivered performance gap increasing