1 Lecture 15: DRAM Design Today: DRAM basics, DRAM innovations (Section 5.3)

Slides:



Advertisements
Similar presentations
Main MemoryCS510 Computer ArchitecturesLecture Lecture 15 Main Memory.
Advertisements

Outline Memory characteristics SRAM Content-addressable memory details DRAM © Derek Chiou & Mattan Erez 1.
Lecture 19: Cache Basics Today’s topics: Out-of-order execution
Chapter 5 Internal Memory
5-1 Memory System. Logical Memory Map. Each location size is one byte (Byte Addressable) Logical Memory Map. Each location size is one byte (Byte Addressable)
COEN 180 DRAM. Dynamic Random Access Memory Dynamic: Periodically refresh information in a bit cell. Else it is lost. Small footprint: transistor + capacitor.
1 Lecture 6: Chipkill, PCM Topics: error correction, PCM basics, PCM writes and errors.
Main Mem.. CSE 471 Autumn 011 Main Memory The last level in the cache – main memory hierarchy is the main memory made of DRAM chips DRAM parameters (memory.
CS.305 Computer Architecture Memory: Structures Adapted from Computer Organization and Design, Patterson & Hennessy, © 2005, and from slides kindly made.
1 Lecture 13: DRAM Innovations Today: energy efficiency, row buffer management, scheduling.
Chapter 9 Memory Basics Henry Hexmoor1. 2 Memory Definitions  Memory ─ A collection of storage cells together with the necessary circuits to transfer.
ENGIN112 L30: Random Access Memory November 14, 2003 ENGIN 112 Intro to Electrical and Computer Engineering Lecture 30 Random Access Memory (RAM)
Lecture 12: DRAM Basics Today: DRAM terminology and basics, energy innovations.
1 Lecture 14: Cache Innovations and DRAM Today: cache access basics and innovations, DRAM (Sections )
1 Lecture 16B Memories. 2 Memories in General Computers have mostly RAM ROM (or equivalent) needed to boot ROM is in same class as Programmable Logic.
1 Lecture 16: Virtual Memory Today: DRAM innovations, virtual memory (Sections )
Memory Hierarchy.1 Review: Major Components of a Computer Processor Control Datapath Memory Devices Input Output.
ENEE350 Ankur Srivastava University of Maryland, College Park Based on Slides from Mary Jane Irwin ( )
1 COMP 206: Computer Architecture and Implementation Montek Singh Wed., Nov. 13, 2002 Topic: Main Memory (DRAM) Organization.
Registers  Flip-flops are available in a variety of configurations. A simple one with two independent D flip-flops with clear and preset signals is illustrated.
CS 151 Digital Systems Design Lecture 30 Random Access Memory (RAM)
1 Lecture 13: Cache Innovations Today: cache access basics and innovations, DRAM (Sections )
Chapter 5 Internal Memory
Memory Devices Wen-Hung Liao, Ph.D..
1 Lecture 14: Virtual Memory Today: DRAM and Virtual memory basics (Sections )
1 Lecture 16B Memories. 2 Memories in General RAM - the predominant memory ROM (or equivalent) needed to boot ROM is in same class as Programmable Logic.
1 Lecture 14: DRAM, PCM Today: DRAM scheduling, reliability, PCM Class projects.
1 Lecture 1: Introduction and Memory Systems CS 7810 Course organization:  5 lectures on memory systems  5 lectures on cache coherence and consistency.
1 Towards Scalable and Energy-Efficient Memory System Architectures Rajeev Balasubramonian School of Computing University of Utah.
1 Efficient Data Access in Future Memory Hierarchies Rajeev Balasubramonian School of Computing Research Buffet, Fall 2010.
CPE232 Memory Hierarchy1 CPE 232 Computer Organization Spring 2006 Memory Hierarchy Dr. Gheith Abandah [Adapted from the slides of Professor Mary Irwin.
CSIE30300 Computer Architecture Unit 07: Main Memory Hsin-Chou Chi [Adapted from material by and
Lecture 7: PCM, Cache coherence
1 Lecture: Virtual Memory, DRAM Main Memory Topics: virtual memory, TLB/cache access, DRAM intro (Sections 2.2)
1 CSCI 2510 Computer Organization Memory System I Organization.
Main Memory CS448.
1 Lecture 14: DRAM Main Memory Systems Today: cache/TLB wrap-up, DRAM basics (Section 2.3)
Computer Memory Storage Decoding Addressing 1. Memories We've Seen SIMM = Single Inline Memory Module DIMM = Dual IMM SODIMM = Small Outline DIMM RAM.
The Memory Hierarchy Lecture # 30 15/05/2009Lecture 30_CA&O_Engr Umbreen Sabir.
Computer Architecture Lecture 24 Fasih ur Rehman.
CS/EE 5810 CS/EE 6810 F00: 1 Main Memory. CS/EE 5810 CS/EE 6810 F00: 2 Main Memory Bottom Rung of the Memory Hierarchy 3 important issues –capacity »BellÕs.
1 Lecture 2: Memory Energy Topics: energy breakdowns, handling overfetch, LPDRAM, row buffer management, channel energy, refresh energy.
1 Adapted from UC Berkeley CS252 S01 Lecture 18: Reducing Cache Hit Time and Main Memory Design Virtucal Cache, pipelined cache, cache summary, main memory.
1 KU College of Engineering Elec 204: Digital Systems Design Lecture 22 Memory Definitions Memory ─ A collection of storage cells together with the necessary.
1 Lecture: Memory Technology Innovations Topics: state-of-the-art and upcoming changes: buffer chips, 3D stacking, non-volatile cells, photonics Multiprocessor.
1 Lecture: Memory Technology Innovations Topics: memory schedulers, refresh, state-of-the-art and upcoming changes: buffer chips, 3D stacking, non-volatile.
1 Memory Hierarchy (I). 2 Outline Random-Access Memory (RAM) Nonvolatile Memory Disk Storage Suggested Reading: 6.1.
1 Lecture: DRAM Main Memory Topics: DRAM intro and basics (Section 2.3)
Chapter 11 System Performance Enhancement. Basic Operation of a Computer l Program is loaded into memory l Instruction is fetched from memory l Operands.
CS35101 Computer Architecture Spring 2006 Lecture 18: Memory Hierarchy Paul Durand ( ) [Adapted from M Irwin (
ECE/CS 552: Cache Concepts © Prof. Mikko Lipasti Lecture notes based in part on slides created by Mark Hill, David Wood, Guri Sohi, John Shen and Jim Smith.
1 Lecture 20: OOO, Memory Hierarchy Today’s topics:  Out-of-order execution  Cache basics.
1 Lecture 16: Main Memory Innovations Today: DRAM basics, innovations, trends HW5 due on Thursday; simulations can take a few hours Midterm: 32 scores.
1 Lecture: Memory Basics and Innovations Topics: memory organization basics, schedulers, refresh,
Reducing Hit Time Small and simple caches Way prediction Trace caches
Lecture 23: Cache, Memory, Security
Lecture 15: DRAM Main Memory Systems
Lecture: Memory, Multiprocessors
Lecture 21: Memory Hierarchy
Lecture: DRAM Main Memory
Lecture 23: Cache, Memory, Virtual Memory
Lecture 22: Cache Hierarchies, Memory
Lecture: DRAM Main Memory
Lecture: DRAM Main Memory
Lecture: Memory Technology Innovations
Lecture 6: Reliability, PCM
Lecture 24: Memory, VM, Multiproc
Lecture 15: Memory Design
Lecture 22: Cache Hierarchies, Memory
Presentation transcript:

1 Lecture 15: DRAM Design Today: DRAM basics, DRAM innovations (Section 5.3)

2 DRAM Main Memory Main memory is stored in DRAM cells that have much higher storage density DRAM cells lose their state over time – must be refreshed periodically, hence the name Dynamic DRAM access suffers from long access time and high energy overhead Since the pins on a processor chip are expected to not increase much, we will hit a memory bandwidth wall

3 DRAM Organization 3 … Memory bus or channel Rank DRAM chip or device Bank Array 1/8 th of the row buffer One word of data output DIMM On-chip Memory Controller

4 DRAM Array Access 16Mb DRAM array = 4096 x 4096 array of bits 12 row address bits arrive first Column decoder 12 column address bits arrive next Eight bits returned to CPU, one per cycle 4096 bits are read out Row Access Strobe (RAS) Column Access Strobe (CAS)Row Buffer

5 Salient Points I DIMM, rank, bank, array  form a hierarchy in the storage organization Because of electrical constraints, only a few DIMMs can be attached to a bus Ranks help increase the capacity on a DIMM Multiple DRAM chips are used for every access to improve data transfer bandwidth Multiple banks are provided so we can be simultaneously working on different requests

6 Salient Points II To maximize density, arrays within a bank are made large  rows are wide  row buffers are wide (8KB read for a 64B request) Each array provides a single bit to the output pin in a cycle (for high density and because there are few pins) DRAM chips are described as xN, where N refers to the number of output pins; one rank may be composed of eight x8 DRAM chips (the data bus is 64 bits) The memory controller schedules memory accesses to maximize row buffer hit rates and bank/rank parallelism

7 Salient Points III Banks and ranks offer memory parallelism Row buffers act as a cache within DRAM  Row buffer hit: ~20 ns access time (must only move data from row buffer to pins)  Empty row buffer access: ~40 ns (must first read arrays, then move data from row buffer to pins)  Row buffer conflict: ~60 ns (must first writeback the existing row, then read new row, then move data to pins) In addition, must wait in the queue (tens of nano-seconds) and incur address/cmd/data transfer delays (~10 ns)

8 Technology Trends Improvements in technology (smaller devices)  DRAM capacities double every two years, but latency does not change much Power wall: 25-40% of datacenter power can be attributed to the DRAM system Will soon hit a density wall; may have to be replaced by other technologies (phase change memory, STT-RAM) Interconnects may have to be photonic to overcome the bandwidth limitation imposed by pins on the chip

9 Latency and Power Wall Latency and power can be both improved by employing smaller arrays; incurs a penalty in density and cost Latency and power can be both improved by increasing the row buffer hit rate; requires intelligent mapping of data to rows, clever scheduling of requests, etc. Power can be reduced by minimizing overfetch – either read fewer chips or read parts of a row; incur penalties in area or bandwidth

10 Density Wall New emerging non-volatile memories that have better scalability; instead of storing data in the form of charge, data encoded in cell resistance (phase change memory) or in electron spin (spin torque transfer STT-RAM) Phase change memory: the cell can be either amorphous (high resistance, represents zero) or crystalline (low resistance, represents one) Data is written by heating the material and cooling it at different rates (with electrical pulses); short intense pulse  amorphous; long medium pulse  crystalline

11 Phase Change Memory Can also have multi-level cells; each resistance value represents a different encoding; enables scalability Each cell can only be written about 10 times; need many tricks to improve endurance: write on change, shift bits within a row, re-map data to rows, etc. Reads are relatively quick (~50 ns), writes are very slow (~1000 ns) Has potential to replace DRAM, disk, or both in at least some classes of computers 8

12 Photonics A single waveguide carries light that was generated off-chip to multiple “nodes” The nodes can act as transmitters or receivers; transmitters can vary the amplitude of a light signal based on the input electrical signal Since multiple light wavelengths can be multiplexed on a waveguide and because each wavelength can carry a different signal, photonic interconnects have high bandwidth The E  O and O  E conversion overhead means that the photonic signal must travel far enough to out-do an electrical interconnect in terms of latency and power

13 Title Bullet