1  1998 Morgan Kaufmann Publishers Arithmetic Where we've been: –Performance (seconds, cycles, instructions) –Abstractions: Instruction Set Architecture.

Slides:



Advertisements
Similar presentations
Chapter Three.
Advertisements

1 Chapter Three Last revision: 4/17/ Arithmetic Where we've been: –Performance (seconds, cycles, instructions) –Abstractions: Instruction Set Architecture.
©UCB CPSC 161 Lecture 6 Prof. L.N. Bhuyan
1 CONSTRUCTING AN ARITHMETIC LOGIC UNIT CHAPTER 4: PART II.
Princess Sumaya Univ. Computer Engineering Dept. Chapter 3:
Princess Sumaya Univ. Computer Engineering Dept. Chapter 3: IT Students.
Chapter 3 Arithmetic for Computers. Multiplication More complicated than addition accomplished via shifting and addition More time and more area Let's.
Arithmetic CPSC 321 Computer Architecture Andreas Klappenecker.
1 Representing Numbers Using Bases Numbers in base 10 are called decimal numbers, they are composed of 10 numerals ( ספרות ) = 9* * *10.
1  2004 Morgan Kaufmann Publishers Chapter Three.
Computer Organization Multiplication and Division Feb 2005 Reading: Portions of these slides are derived from: Textbook figures © 1998 Morgan Kaufmann.
1 Chapter 4: Arithmetic Where we've been: –Performance (seconds, cycles, instructions) –Abstractions: Instruction Set Architecture Assembly Language and.
Chapter Four Arithmetic and Logic Unit
Arithmetic I CPSC 321 Andreas Klappenecker. Administrative Issues Office hours of TA Praveen Bhojwani: M 1:00pm-3:00pm.
Arithmetic-Logic Units CPSC 321 Computer Architecture Andreas Klappenecker.
1  2004 Morgan Kaufmann Publishers Lectures for 3rd Edition Note: these lectures are often supplemented with other materials and also problems from the.
1 ECE369 Chapter 3. 2 ECE369 Multiplication More complicated than addition –Accomplished via shifting and addition More time and more area.
1  1998 Morgan Kaufmann Publishers Chapter Four Arithmetic for Computers.
Chapter 3 Arithmetic for Computers. Arithmetic Where we've been: Abstractions: Instruction Set Architecture Assembly Language and Machine Language What's.
Arithmetic for Computers
(original notes from Prof. J. Kelly Flanagan)
1 Bits are just bits (no inherent meaning) — conventions define relationship between bits and numbers Binary numbers (base 2)
Computer Arithmetic Nizamettin AYDIN
1 CS/COE0447 Computer Organization & Assembly Language Chapter 3.
1 Modified from  Modified from 1998 Morgan Kaufmann Publishers Chapter Three: Arithmetic for Computers citation and following credit line is included:
Computer Arithmetic.
1  2004 Morgan Kaufmann Publishers Instructions: bne $t4,$t5,Label Next instruction is at Label if $t4≠$t5 beq $t4,$t5,Label Next instruction is at Label.
EGRE 426 Fall 09 Chapter Three
Computing Systems Basic arithmetic for computers.
ECE232: Hardware Organization and Design
1 ECE369 Sections 3.5, 3.6 and ECE369 Number Systems Fixed Point: Binary point of a real number in a certain position –Can treat real numbers as.
07/19/2005 Arithmetic / Logic Unit – ALU Design Presentation F CSE : Introduction to Computer Architecture Slides by Gojko Babić.
1 EGRE 426 Fall 08 Chapter Three. 2 Arithmetic What's up ahead: –Implementing the Architecture 32 operation result a b ALU.
Csci 136 Computer Architecture II – Constructing An Arithmetic Logic Unit Xiuzhen Cheng
Computing Systems Designing a basic ALU.
Princess Sumaya Univ. Computer Engineering Dept. Chapter 3:
1 Lecture 6 BOOLEAN ALGEBRA and GATES Building a 32 bit processor PH 3: B.1-B.5.
CDA 3101 Fall 2013 Introduction to Computer Organization The Arithmetic Logic Unit (ALU) and MIPS ALU Support 20 September 2013.
1  2004 Morgan Kaufmann Publishers Performance is specific to a particular program/s –Total execution time is a consistent summary of performance For.
1 ELEN 033 Lecture 4 Chapter 4 of Text (COD2E) Chapters 3 and 4 of Goodman and Miller book.
Orange Coast College Business Division Computer Science Department CS 116- Computer Architecture Arithmetic: Part II.
1  2004 Morgan Kaufmann Publishers Lets Build a Processor Almost ready to move into chapter 5 and start building a processor First, let’s review Boolean.
Computer Architecture Lecture Notes Spring 2005 Dr. Michael P. Frank Competency Area 4: Computer Arithmetic.
Addition, Subtraction, Logic Operations and ALU Design
CML CML CS 230: Computer Organization and Assembly Language Aviral Shrivastava Department of Computer Science and Engineering School of Computing and Informatics.
순천향대학교 정보기술공학부 이 상 정 1 3. Arithmetic for Computers.
Prof. Hsien-Hsin Sean Lee
Lec 11Systems Architecture1 Systems Architecture Lecture 11: Arithmetic for Computers Jeremy R. Johnson Anatole D. Ruslanov William M. Mongan Some or all.
10/7/2004Comp 120 Fall October 7 Read 5.1 through 5.3 Register! Questions? Chapter 4 – Floating Point.
1 Arithmetic Where we've been: –Abstractions: Instruction Set Architecture Assembly Language and Machine Language What's up ahead: –Implementing the Architecture.
Computer Arthmetic Chapter Four P&H. Data Representation Why do we not encode numbers as strings of ASCII digits inside computers? What is overflow when.
9/23/2004Comp 120 Fall September Chapter 4 – Arithmetic and its implementation Assignments 5,6 and 7 posted to the class web page.
EE204 L03-ALUHina Anwar Khan EE204 Computer Architecture Lecture 03- ALU.
May 2, 2001System Architecture I1 Systems Architecture I (CS ) Lecture 11: Arithmetic for Computers * Jeremy R. Johnson May 2, 2001 *This lecture.
By Wannarat Computer System Design Lecture 3 Wannarat Suntiamorntut.
1 CPTR 220 Computer Organization Computer Architecture Assembly Programming.
Based on slides from D. Patterson and www-inst.eecs.berkeley.edu/~cs152/ COM 249 – Computer Organization and Assembly Language Chapter 3 Arithmetic For.
1 Chapter 3 Arithmetic for Computers Lecture Slides are from Prof. Jose Delgado-Frias, Mr. Paul Wettin, and Prof. Valeriu Beiu (Washington State University.
1 (Based on text: David A. Patterson & John L. Hennessy, Computer Organization and Design: The Hardware/Software Interface, 3 rd Ed., Morgan Kaufmann,
Computer System Design Lecture 3
Computer Arthmetic Chapter Four P&H.
CS/COE0447 Computer Organization & Assembly Language
Arithmetic Where we've been:
Arithmetic Logical Unit
Computer Arithmetic Multiplication, Floating Point
CS/COE0447 Computer Organization & Assembly Language
CS/COE0447 Computer Organization & Assembly Language
October 8 Rules for Programming Assignments ASK QUESTIONS!
Chapter 3 Arithmetic for Computers
Presentation transcript:

1  1998 Morgan Kaufmann Publishers Arithmetic Where we've been: –Performance (seconds, cycles, instructions) –Abstractions: Instruction Set Architecture Assembly Language and Machine Language What's up ahead: –Implementing the Architecture

2  1998 Morgan Kaufmann Publishers Arithmetic We start with the Arithmetic and Logic Unit 32 operation result a b ALU

3  1998 Morgan Kaufmann Publishers Bits are just bits (no inherent meaning) — conventions define relationship between bits and numbers Binary numbers (base 2) decimal: n -1 Of course it gets more complicated: numbers are finite (overflow) fractions and real numbers negative numbers How do we represent negative numbers? i.e., which bit patterns will represent which numbers? Octal and hexadecimal numbers Floating-point numbers Numbers

4  1998 Morgan Kaufmann Publishers Sign Magnitude: One's Complement Two's Complement 000 = = = = = = = = = = = = = = = = = = = = = = = = -1 Issues: balance, number of zeros, ease of operations. Two’s complement is best. Possible Representations of Signed Numbers

5  1998 Morgan Kaufmann Publishers 32 bit signed numbers: = = = = +2,147,483, = +2,147,483, = –2,147,483, = –2,147,483, = –2,147,483, = – = – = –1 10 maxint minint MIPS

6  1998 Morgan Kaufmann Publishers Negating a two's complement number: invert all bits and add 1 –Remember: “Negate” and “invert” are different operations. You negate a number but invert a bit. Converting n bit numbers into numbers with more than n bits: –MIPS 16 bit immediate gets converted to 32 bits for arithmetic –copy the most significant bit (the sign bit) into the other bits > > "sign extension" –MIPS load byte instructions lbu: no sign extension lb: sign extension Two's Complement Operations

7  1998 Morgan Kaufmann Publishers Unsigned numbers just like in grade school (carry/borrow 1s) Two's complement operations easy –subtraction using addition of opposite number (result is 0001, carry bit is set) Two's complement overflow (result not within number range) –e.g., adding two 4-bit numbers does not yield an 4-bit number (result is - 6, overflow bit is set) Addition & Subtraction

8  1998 Morgan Kaufmann Publishers No overflow when adding a positive and a negative number No overflow when signs are the same for subtraction Overflow occurs when the value affects the sign: –overflow when adding two positives yields a negative –or, adding two negatives gives a positive –or, subtract a negative from a positive and get a negative –or, subtract a positive from a negative and get a positive Consider the operations A + B, and A – B –Can overflow occur if B is 0 ? No. –Can overflow occur if A is 0 ? Yes. Detecting Overflow

9  1998 Morgan Kaufmann Publishers An exception (interrupt) occurs –Control jumps to predefined address for exception –Interrupted address is saved for possible resumption Details based on software system / language –example: flight control vs. homework assignment Don't always want to detect overflow — new MIPS instructions: addu, addiu, subu note: addiu still sign-extends! note: sltu, sltiu for unsigned comparisons Effects of Overflow

10  1998 Morgan Kaufmann Publishers and, andi:bit-by-bit AND or, ori:bit-by-bit OR sll:shift left logical slr:shift right logical shifting left two steps gives shifting right three bits gives Logical Operations

11  1998 Morgan Kaufmann Publishers Let's build a logical unit to support the and and or instructions –we'll just build a 1 bit unit, and use 32 of them –op=0: and; op=1: or Possible Implementation (sum-of-products): result = a b + a op + b op b a operation result Logical unit opabres

12  1998 Morgan Kaufmann Publishers Selects one of the inputs to be the output, based on a control input IEC symbol of a 4-input MUX: Lets build our logical unit using a MUX: S C A B 0 1 Review: The Multiplexor 0101 b a Operation Result MUX G 0 3 _ EN

13  1998 Morgan Kaufmann Publishers Not easy to decide the “best” way to build something –Don't want too many inputs to a single gate –Don’t want to have to go through too many gates –For our purposes, ease of comprehension is important –We use multiplexors Let's look at a 1-bit ALU for addition: How could we build a 32-bit ALU for AND, OR and ADD? Different Implementations c out = a b + a c in + b c in sum = a  b  c in Sum CarryIn CarryOut a b a b c in c out sum

14  1998 Morgan Kaufmann Publishers Building a 32 bit ALU for AND, OR and ADD We need a 4-input MUX. 1-bit ALU:

15  1998 Morgan Kaufmann Publishers Two's complement approach: just negate b and add. A clever solution: In a multiple bit ALU the least significant CarryIn has to be equal to 1 for subtraction. What about subtraction (a – b) ?

16  1998 Morgan Kaufmann Publishers Need to support the set-on-less-than instruction (slt) –remember: slt is an arithmetic instruction –produces a 1 if rs < rt and 0 otherwise –use subtraction: (a-b) < 0 implies a < b Need to support test for equality (beq $t5, $t6, $t7) –use subtraction: (a-b) = 0 implies a = b Tailoring the ALU to the MIPS

Supporting slt Other ALUs: Most significant ALU:

18  1998 Morgan Kaufmann Publishers 32 bit ALU supporting slt a<b  a-b<0, thus Set is the sign bit of the result.

19  1998 Morgan Kaufmann Publishers Final ALU including test for equality Notice control lines: 000 = and 001 = or 010 = add 110 = subtract 111 = slt Note: Zero is a 1 when the result is zero!

20  1998 Morgan Kaufmann Publishers Conclusion We can build an ALU to support the MIPS instruction set –key idea: use a multiplexor to select the output we want –we can efficiently perform subtraction using two’s complement –we can replicate a 1-bit ALU to produce a 32-bit ALU Important points about hardware –all of the gates are always working –the speed of a gate is affected by the number of inputs to the gate –the speed of a circuit is affected by the number of gates in series (on the “critical path” or the “deepest level of logic”)

21  1998 Morgan Kaufmann Publishers Conclusion Our primary focus: comprehension, however, –clever changes to organization can improve performance (similar to using better algorithms in software) –we’ll look at examples for addition, multiplication and division

22  1998 Morgan Kaufmann Publishers A 32-bit ALU is much slower than a 1-bit ALU. There are more than one way to do addition. –the two extremes: ripple carry and sum-of-products Can you see the ripple? How could you get rid of it? c 1 = b 0 c 0 + a 0 c 0 + a 0 b 0 c 2 = b 1 c 1 + a 1 c 1 + a 1 b 1 c 2 = c 2 (a 0,b 0,c 0,a 1,b 1 ) c 3 = b 2 c 2 + a 2 c 2 + a 2 b 2 c 3 = c 3 (a 0,b 0,c 0,a 1,b 1,a 2,b 2 ) c 4 = b 3 c 3 + a 3 c 3 + a 3 b 3 c 4 = c 4 (a 0,b 0,c 0,a 1,b 1,a 2,b 2,a 3,b 3 ) Not feasible! Too many inputs to the gates. Problem: ripple carry adder is slow

23  1998 Morgan Kaufmann Publishers An approach in-between the two extremes Motivation: –If we didn't know the value of carry-in, what could we do? –When would we always generate a carry? g i = a i b i –When would we propagate the carry? p i = a i + b i –Look at the truth table! Did we get rid of the ripple? c 1 = g 0 + p 0 c 0 c 2 = g 1 + p 1 c 1 c 2 = g 1 +p 1 g 0 +p 1 p 0 c 0 c 3 = g 2 + p 2 c 2 c 3 = g 2 +p 2 g 1 +p 2 p 1 g 0 +p 2 p 1 p 0 c 0 c 4 = g 3 + p 3 c 3 c 4 =... Feasible! A smaller number of inputs to the gates. Carry-lookahead adder

24  1998 Morgan Kaufmann Publishers Can’t build a 16 bit CLA adder (too big) Could use ripple carry of 4-bit CLA adders Better: use the CLA principle again! Principle shown in the figure. See textbook for details. Use principle to build bigger adders

25  1998 Morgan Kaufmann Publishers More complicated than addition –can be accomplished via shifting and addition More time and more area Let's look at 2 versions based on grammar school algorithm 0010 (multiplicand) x_1011 (multiplier) ___ Negative numbers: –easy way: convert to positive and multiply –there are better techniques Multiplication

26  1998 Morgan Kaufmann Publishers Multiplication, First Version

27  1998 Morgan Kaufmann Publishers Multiplication, Final Version

28  1998 Morgan Kaufmann Publishers Booth’s Algorithm The grammar school method was implemented using addition and shifting Booth’s algorithm also uses subtraction Based on two bits of the multiplier either add, subtract or do nothing; always shift Handles two’s complement numbers

29  1998 Morgan Kaufmann Publishers Fast multipliers Combinational implementations –Conventional multiplier algorithm partial products with AND gates adders –Lots of modifications Sequential implementations –Pipelined multiplier registers between levels of logic result delayed effective speed of multiple multiplications increased

30  1998 Morgan Kaufmann Publishers Four-Bit Binary Multiplication Multiplicand B3 B2 B1 B0 Multiplier  A3 A2 A1 A0 1st partial product A0B3 A0B2 A0B1 A0B0 2nd partial product A1B3 A1B2 A1B1 A1B0 3rd partial product A2B3 A2B2 A2B1 A2B0 4th partial product + A3B3 A3B2 A3B1 A3B0 Final product P7 P6 P5 P4 P3 P2 P1 P0

31  1998 Morgan Kaufmann Publishers Classical Implementation P  7:0  PP1 PP2 PP3 PP4 & & & & A0 B0 A0 B1 A0 B2 A0 B3 4 / 4 / 4 / 4 / 6 / 6 /

32  1998 Morgan Kaufmann Publishers Pipelined Multiplier Clk // / / / / / / / /

33  1998 Morgan Kaufmann Publishers Division Simple method: –Initialise the remainder with the dividend –Start from most significant end –Subtract divisor from the remainder if possible (quotient bit 1) –Shift divisor to the right and repeat

34  1998 Morgan Kaufmann Publishers Division, First Version

35  1998 Morgan Kaufmann Publishers Division, Final Version Same hardware for multiply and divide.

36  1998 Morgan Kaufmann Publishers Floating Point (a brief look) We need a way to represent –numbers with fractions, e.g., –very small numbers, e.g., –very large numbers, e.g.,  10 9 Representation: –sign, exponent, significand: (–1) sign  significand  2 exponent –more bits for significand gives more accuracy –more bits for exponent increases range IEEE 754 floating point standard: –single precision: 8 bit exponent, 23 bit significand –double precision: 11 bit exponent, 52 bit significand

37  1998 Morgan Kaufmann Publishers IEEE 754 floating-point standard Leading “1” bit of significand is implicit Exponent is “biased” to make sorting easier –all 0s is smallest exponent all 1s is largest –bias of 127 for single precision and 1023 for double precision –summary: (–1) sign  significand)  2 exponent – bias Example: –decimal: -.75 = -3/4 = -3/2 2 –binary: -.11 = -1.1 x 2 -1 –floating point: exponent = 126 = –IEEE single precision:

38  1998 Morgan Kaufmann Publishers Floating-point addition 1.Shift the significand of the number with the lesser exponent right until the exponents match 2.Add the significands 3.Normalise the sum, checking for overflow or underflow 4.Round the sum

39  1998 Morgan Kaufmann Publishers Floating-point addition

40  1998 Morgan Kaufmann Publishers Floating-point multiplication 1.Add the exponents 2.Multiply the significands 3.Normalise the product, checking for overflow or underflow 4.Round the product 5.Find out the sign of the product

41  1998 Morgan Kaufmann Publishers Floating Point Complexities Operations are somewhat more complicated (see text) In addition to overflow we can have “underflow” Accuracy can be a big problem –IEEE 754 keeps two extra bits during intermediate calculations, guard and round –four rounding modes –positive divided by zero yields “infinity” –zero divide by zero yields “not a number” –other complexities Implementing the standard can be tricky

42  1998 Morgan Kaufmann Publishers Chapter Four Summary Computer arithmetic is constrained by limited precision Bit patterns have no inherent meaning but standards do exist –two’s complement –IEEE 754 floating point Computer instructions determine “meaning” of the bit patterns Performance and accuracy are important so there are many complexities in real machines (i.e., algorithms and implementation). We are ready to move on (and implement the processor)