Lecture 15: DRAM Main Memory Systems

Slides:



Advertisements
Similar presentations
COEN 180 DRAM. Dynamic Random Access Memory Dynamic: Periodically refresh information in a bit cell. Else it is lost. Small footprint: transistor + capacitor.
Advertisements

Main Mem.. CSE 471 Autumn 011 Main Memory The last level in the cache – main memory hierarchy is the main memory made of DRAM chips DRAM parameters (memory.
1 Lecture: Memory, Coherence Protocols Topics: wrap-up of memory systems, intro to multi-thread programming models.
1 Lecture 13: DRAM Innovations Today: energy efficiency, row buffer management, scheduling.
Lecture 12: DRAM Basics Today: DRAM terminology and basics, energy innovations.
1 Lecture 14: Cache Innovations and DRAM Today: cache access basics and innovations, DRAM (Sections )
1 Lecture 15: DRAM Design Today: DRAM basics, DRAM innovations (Section 5.3)
1 Lecture 16: Virtual Memory Today: DRAM innovations, virtual memory (Sections )
Memory Hierarchy.1 Review: Major Components of a Computer Processor Control Datapath Memory Devices Input Output.
1 Lecture 13: Cache Innovations Today: cache access basics and innovations, DRAM (Sections )
1 Lecture 14: Virtual Memory Today: DRAM and Virtual memory basics (Sections )
1 Lecture 1: Introduction and Memory Systems CS 7810 Course organization:  5 lectures on memory systems  5 lectures on cache coherence and consistency.
1 Towards Scalable and Energy-Efficient Memory System Architectures Rajeev Balasubramonian School of Computing University of Utah.
1 Lecture 4: Memory: HMC, Scheduling Topics: BOOM, memory blades, HMC, scheduling policies.
1 Lecture: Virtual Memory, DRAM Main Memory Topics: virtual memory, TLB/cache access, DRAM intro (Sections 2.2)
1 Lecture 14: DRAM Main Memory Systems Today: cache/TLB wrap-up, DRAM basics (Section 2.3)
Modern DRAM Memory Architectures Sam Miller Tam Chantem Jon Lucas CprE 585 Fall 2003.
1 Lecture 2: Memory Energy Topics: energy breakdowns, handling overfetch, LPDRAM, row buffer management, channel energy, refresh energy.
COMP541 Memories II: DRAMs
1 Adapted from UC Berkeley CS252 S01 Lecture 18: Reducing Cache Hit Time and Main Memory Design Virtucal Cache, pipelined cache, cache summary, main memory.
1 Lecture 5: Refresh, Chipkill Topics: refresh basics and innovations, error correction.
1 Lecture 5: Scheduling and Reliability Topics: scheduling policies, handling DRAM errors.
1 Lecture 3: Memory Buffers and Scheduling Topics: buffers (FB-DIMM, RDIMM, LRDIMM, BoB, BOOM), memory blades, scheduling policies.
1 Lecture: Memory Technology Innovations Topics: memory schedulers, refresh, state-of-the-art and upcoming changes: buffer chips, 3D stacking, non-volatile.
1 Lecture 3: Memory Energy and Buffers Topics: Refresh, floorplan, buffers (SMB, FB-DIMM, BOOM), memory blades, HMC.
1 Lecture: DRAM Main Memory Topics: DRAM intro and basics (Section 2.3)
CSE431 L18 Memory Hierarchy.1Irwin, PSU, 2005 CSE 431 Computer Architecture Fall 2005 Lecture 18: Memory Hierarchy Review Mary Jane Irwin (
1 Lecture 4: Memory Scheduling, Refresh Topics: scheduling policies, refresh basics.
1 Lecture 16: Main Memory Innovations Today: DRAM basics, innovations, trends HW5 due on Thursday; simulations can take a few hours Midterm: 32 scores.
1 Lecture: Memory Basics and Innovations Topics: memory organization basics, schedulers, refresh,
CS161 – Design and Architecture of Computer Main Memory Slides adapted from Onur Mutlu (CMU)
COMP541 Memories II: DRAMs
William Stallings Computer Organization and Architecture 7th Edition
ECE 4100/6100 Advanced Computer Architecture Lecture 11 DRAM
CSE 502: Computer Architecture
Reducing Hit Time Small and simple caches Way prediction Trace caches
SRAM Memory External Interface
COMP541 Memories II: DRAMs
Morgan Kaufmann Publishers Memory & Cache
William Stallings Computer Organization and Architecture 7th Edition
Lecture 23: Cache, Memory, Security
Samira Khan University of Virginia Oct 9, 2017
William Stallings Computer Organization and Architecture 8th Edition
Lecture: Memory, Multiprocessors
Staged-Reads : Mitigating the Impact of DRAM Writes on DRAM Reads
The Main Memory system: DRAM organization
Lecture: DRAM Main Memory
Lecture 23: Cache, Memory, Virtual Memory
Computer Architecture
Lecture 22: Cache Hierarchies, Memory
William Stallings Computer Organization and Architecture 7th Edition
Lecture: DRAM Main Memory
Lecture: DRAM Main Memory
William Stallings Computer Organization and Architecture 8th Edition
Lecture: Memory Technology Innovations
Lecture 24: Memory, VM, Multiproc
Lecture 20: OOO, Memory Hierarchy
Lecture 20: OOO, Memory Hierarchy
Lecture 15: Memory Design
If a DRAM has 512 rows and its refresh time is 9ms, what should be the frequency of row refresh operation on the average?
Chapter 4: MEMORY.
Lecture 22: Cache Hierarchies, Memory
15-740/ Computer Architecture Lecture 19: Main Memory
DRAM Hwansoo Han.
William Stallings Computer Organization and Architecture 8th Edition
Bob Reese Micro II ECE, MSU
18-447: Computer Architecture Lecture 19: Main Memory
Presentation transcript:

Lecture 15: DRAM Main Memory Systems Today: DRAM basics and innovations (Section 2.3)

Memory Architecture Processor Bank Memory Controller DIMM Row Buffer Memory Controller Address/Cmd DIMM Data DIMM: a PCB with DRAM chips on the back and front Rank: a collection of DRAM chips that work together to respond to a request and keep the data bus full A 64-bit data bus will need 8 x8 DRAM chips or 4 x16 DRAM chips or.. Bank: a subset of a rank that is busy during one request Row buffer: the last row (say, 8 KB) read from a bank, acts like a cache

DRAM Array Access 16Mb DRAM array = 4096 x 4096 array of bits 12 row address bits arrive first Row Access Strobe (RAS) 4096 bits are read out Some bits returned to CPU 12 column address bits arrive next Column decoder Column Access Strobe (CAS) Row Buffer

Organizing a Rank DIMM, rank, bank, array  form a hierarchy in the storage organization Because of electrical constraints, only a few DIMMs can be attached to a bus One DIMM can have 1-4 ranks For energy efficiency, use wide-output DRAM chips – better to activate only 4 x16 chips per request than 16 x4 chips For high capacity, use narrow-output DRAM chips – since the ranks on a channel are limited, capacity per rank is boosted by having 16 x4 2Gb chips than 4 x16 2Gb chips

Organizing Banks and Arrays A rank is split into many banks (4-16) to boost parallelism within a rank Ranks and banks offer memory-level parallelism A bank is made up of multiple arrays (subarrays, tiles, mats) To maximize density, arrays within a bank are made large  rows are wide  row buffers are wide (8KB read for a 64B request, called overfetch) Each array provides a single bit to the output pin in a cycle (for high density)

Row Buffers Each bank has a single row buffer Row buffers act as a cache within DRAM Row buffer hit: ~20 ns access time (must only move data from row buffer to pins) Empty row buffer access: ~40 ns (must first read arrays, then move data from row buffer to pins) Row buffer conflict: ~60 ns (must first precharge the bitlines, then read new row, then move data to pins) In addition, must wait in the queue (tens of nano-seconds) and incur address/cmd/data transfer delays (~10 ns)

Reads and Writes A single bus is used for reads and writes The bus direction must be reversed when switching between reads and writes; this takes time and leads to bus idling Hence, writes are performed in bursts; a write buffer stores pending writes until a high water mark is reached Writes are drained until a low water mark is reached

Open/Closed Page Policies If an access stream has locality, a row buffer is kept open Row buffer hits are cheap (open-page policy) Row buffer miss is a bank conflict and expensive because precharge is on the critical path If an access stream has little locality, bitlines are precharged immediately after access (close-page policy) Nearly every access is a row buffer miss The precharge is usually not on the critical path Modern memory controller policies lie somewhere between these two extremes (usually proprietary)

Address Mapping Policies Consecutive cache lines can be placed in the same row to boost row buffer hit rates Consecutive cache lines can be placed in different ranks to boost parallelism Example address mapping policies: row:rank:bank:channel:column:blkoffset row:column:rank:bank:channel:blkoffset

Scheduling Policies FCFS: Issue the first read or write in the queue that is ready for issue First Ready - FCFS: First issue row buffer hits if you can Stall Time Fair: First issue row buffer hits, unless other threads are being neglected

Refresh Every DRAM cell must be refreshed within a 64 ms window A row read/write automatically refreshes the row Every refresh command performs refresh on a number of rows, the memory system is unavailable during that time A refresh command is issued by the memory controller once every 7.8us on average

Error Correction For every 64-bit word, can add an 8-bit code that can detect two errors and correct one error; referred to as SECDED – single error correct double error detect A rank is now made up of 9 x8 chips, instead of 8 x8 chips Stronger forms of error protection exist: a system is chipkill correct if it can handle an entire DRAM chip failure

.. .. .. .. .. .. .. .. Modern Memory System PROC 4 DDR3 channels 64-bit data channels 800 MHz channels 1-2 DIMMs/channel 1-4 ranks/channel .. ..

.. .. .. .. Cutting-Edge Systems SMB PROC The link into the processor is narrow and high frequency The Scalable Memory Buffer chip is a “router” that connects to multiple DDR3 channels (wide and slow) Boosts processor pin bandwidth and memory capacity More expensive, high power

Title Bullet