Memory Interleaving.

Slides:



Advertisements
Similar presentations
Chapter Three: Interconnection Structure
Advertisements

Lecture 10 Instruction Set Architecture
Computer Organization, Bus Structure
MEMORY popo.
Main MemoryCS510 Computer ArchitecturesLecture Lecture 15 Main Memory.
Miss Penalty Reduction Techniques (Sec. 5.4) Multilevel Caches: A second level cache (L2) is added between the original Level-1 cache and main memory.
1/1/ / faculty of Electrical Engineering eindhoven university of technology Processor support devices Part 1:Interrupts and shared memory dr.ir. A.C. Verschueren.
I/O Unit.
Parallell Processing Systems1 Chapter 4 Vector Processors.
Cache Memory Locality of reference: It is observed that when a program refers to memory, the access to memory for data as well as code are confined to.
CSC 4250 Computer Architectures December 8, 2006 Chapter 5. Memory Hierarchy.
Chapter 9 Memory Basics Henry Hexmoor1. 2 Memory Definitions  Memory ─ A collection of storage cells together with the necessary circuits to transfer.
1 Lecture 2: Review of Computer Organization Operating System Spring 2007.
Memory Hierarchy.1 Review: Major Components of a Computer Processor Control Datapath Memory Devices Input Output.
ENEE350 Ankur Srivastava University of Maryland, College Park Based on Slides from Mary Jane Irwin ( )
Registers  Flip-flops are available in a variety of configurations. A simple one with two independent D flip-flops with clear and preset signals is illustrated.
Chapter 7 Interupts DMA Channels Context Switching.
Midterm Tuesday October 23 Covers Chapters 3 through 6 - Buses, Clocks, Timing, Edge Triggering, Level Triggering - Cache Memory Systems - Internal Memory.
1 Lecture 13: Cache Innovations Today: cache access basics and innovations, DRAM (Sections )
Chapter 7 Input/Output. Input/Output Problems Wide variety of peripherals —Delivering different amounts of data —At different speeds —In different formats.
Lecture 41: Review Session #3 Reminders –Office hours during final week TA as usual (Tuesday & Thursday 12:50pm-2:50pm) Hassan: Wednesday 1pm to 4pm or.
Computer Architecture Part III-C: Memory Access and Management.
5.1 Chaper 4 Central Processing Unit Foundations of Computer Science  Cengage Learning.
What’s on the Motherboard? The two main parts of the CPU are the control unit and the arithmetic logic unit. The control unit retrieves instructions from.
The Chip Set. At one time, most of the functions of the chipset were performed by multiple, smaller controller chips Integrated to form a single set of.
Input/Output. Input/Output Problems Wide variety of peripherals —Delivering different amounts of data —At different speeds —In different formats All slower.
Chapter 7 Input/Output Luisa Botero Santiago Del Portillo Ivan Vega.
C.S. Choy95 COMPUTER ORGANIZATION Logic Design Skill to design digital components JAVA Language Skill to program a computer Computer Organization Skill.
Comp-TIA Standards.  AMD- (Advanced Micro Devices) An American multinational semiconductor company that develops computer processors and related technologies.
Higher Computing Computer Systems S. McCrossan 1 Higher Grade Computing Studies 2. Computer Structure Computer Structure The traditional diagram of a computer...
Advanced Higher Computing  Computer Architecture  Chapter 2.
Computer Processing of Data
Computing hardware CPU.
1 Computer System Overview Chapter 1. 2 n An Operating System makes the computing power available to users by controlling the hardware n Let us review.
Survey of Existing Memory Devices Renee Gayle M. Chua.
MICROPROCESSOR INPUT/OUTPUT
Spring EE 437 Lillevik 437s06-l8 University of Portland School of Engineering Advanced Computer Architecture Lecture 8 Project 3: memory agent Programmed.
Memory Systems Embedded Systems Design and Implementation Witawas Srisa-an.
DSP Lecture Series DSP Memory Architecture Dr. E.W. Hu Nov. 28, 2000.
Memory System Unit-IV 4/24/2017 Unit-4 : Memory System.
Main Memory CS448.
CS 312 Computer Architecture Memory Basics Department of Computer Science Southern Illinois University Edwardsville Summer, 2015 Dr. Hiroshi Fujinoki
Parallel architecture Technique. Pipelining Processor Pipelining is a technique of decomposing a sequential process into sub-processes, with each sub-process.
MEMORY ORGANIZTION & ADDRESSING Presented by: Bshara Choufany.
Memory Cell Operation.
B. Ramamurthy.  12 stage pipeline  At peak speed, the processor can request both an instruction and a data word on every clock.  We cannot afford pipeline.
Computer Architecture Lecture 32 Fasih ur Rehman.
Lecture#15. Cache Function The data that is stored within a cache might be values that have been computed earlier or duplicates of original values that.
1 CSCI 2510 Computer Organization Memory System II Cache In Action.
DECStation 3100 Block Instruction Data Effective Program Size Miss Rate Miss Rate Miss Rate 1 6.1% 2.1% 5.4% 4 2.0% 1.7% 1.9% 1 1.2% 1.3% 1.2% 4 0.3%
Lecture 1: Review of Computer Organization
Playstation2 Architecture Architecture Hardware Design.
Topics covered: Memory subsystem CSE243: Introduction to Computer Architecture and Hardware/Software Interface.
The Memory Hierarchy (Lectures #17 - #20) ECE 445 – Computer Organization The slides included herein were taken from the materials accompanying Computer.
Lecture Overview Shift Register Buffering Direct Memory Access.
1 Device Controller I/O units typically consist of A mechanical component: the device itself An electronic component: the device controller or adapter.
Part IVI/O Systems Chapter 13: I/O Systems. I/O Hardware a typical PCI bus structure 2.
Chapter 11 System Performance Enhancement. Basic Operation of a Computer l Program is loaded into memory l Instruction is fetched from memory l Operands.
RAM RAM - random access memory RAM (pronounced ramm) random access memory, a type of computer memory that can be accessed randomly;
نظام المحاضرات الالكترونينظام المحاضرات الالكتروني Cache Memory.
1 load [2], [9] Transfer contents of memory location 9 to memory location 2. Illegal instruction.
Computer Organization
Address – 32 bits WRITE Write Cache Write Main Byte Offset Tag Index Valid Tag Data 16K entries 16.
The Memory System (Chapter 5)
Improving Memory Access 1/3 The Cache and Virtual Memory
Outline Module 1 and 2 dealt with processes, scheduling and synchronization Next two modules will deal with memory and storage Processes require data to.
M. Usha Professor/CSE Sona College of Technology
Miss Rate versus Block Size
Presentation transcript:

Memory Interleaving

interleaved memory Main memory divided into two or more sections. The CPU can access alternate sections immediately, without waiting for memory to catch up (through wait states). Interleaved memory is one technique for compensating for the relatively slow speed of dynamic RAM (DRAM).

interleaved memory Memory interleaving increases bandwidth by allowing simultaneous access to more than one chunk of memory. This improves performance because the processor can transfer more information to/from memory in the same amount of time.

interleaved memory Interleaving works by dividing the system memory into multiple blocks. The most common numbers are two or four, called two-way or four-way interleaving, respectively. Each block of memory is accessed using different sets of control lines, which are merged together on the memory bus. When a read or write is begun to one block, a read or write to other blocks can be overlapped with the first one. The more blocks, the more that overlapping can be done. As an analogy, consider eating a plate of food with a fork. Two-way interleaving would mean dividing the food onto two plates and eating with both hands, using two forks. (Four-way interleaving would require two more hands. :^) ) Remember that here the processor is doing the "eating" and it is much faster than the forks (memory) "feeding" it (unlike a person, whose hands are generally faster.)

interleaved memory each with its own address buffer register (ABR) and data buffer register (DBR), memory access operations may proceed in more than one module at the same time. Two methods of address layout shown

interleaved memory

interleaved memory In the first case, the memory address generated by the CPU is decoded as shown in part a of the figure The high order k bits name one of n modules, and the low-order m bits name a particular word in that module. When consecutive locations are accessed, as happens when a block of data is transferred to a cache, only one module is involved

interleaved memory At the same time, however devices with direct memory access (DMA) ability may be accessing information in other memory modules.

interleaved memory

interleaved memory The second and more effective way to address the modules is shown in Figure 5.246. It is called memory interleaving the low-order kbits of the memory address select a module, and the high-order m bits name a location within that module In this way, consecutive addresses are located in successive modules

interleaved memory Thus, any component of the system that generates requests for access to consecutive memory locations can keep several modules busy at any one time. This results in both faster access to a block of data and higher average utilization of the memory system