1 Lecture 16: Main Memory Innovations Today: DRAM basics, innovations, trends HW5 due on Thursday; simulations can take a few hours Midterm: 32 scores.

Slides:



Advertisements
Similar presentations
COEN 180 DRAM. Dynamic Random Access Memory Dynamic: Periodically refresh information in a bit cell. Else it is lost. Small footprint: transistor + capacitor.
Advertisements

1 Lecture 6: Chipkill, PCM Topics: error correction, PCM basics, PCM writes and errors.
1 Lecture: Memory, Coherence Protocols Topics: wrap-up of memory systems, intro to multi-thread programming models.
Lecture 12: DRAM Basics Today: DRAM terminology and basics, energy innovations.
1 Lecture 14: Cache Innovations and DRAM Today: cache access basics and innovations, DRAM (Sections )
1 Lecture 15: DRAM Design Today: DRAM basics, DRAM innovations (Section 5.3)
1 Lecture 16: Virtual Memory Today: DRAM innovations, virtual memory (Sections )
1 Lecture 13: Cache Innovations Today: cache access basics and innovations, DRAM (Sections )
1 Lecture 14: Virtual Memory Today: DRAM and Virtual memory basics (Sections )
1 Lecture 14: DRAM, PCM Today: DRAM scheduling, reliability, PCM Class projects.
1 Lecture 1: Introduction and Memory Systems CS 7810 Course organization:  5 lectures on memory systems  5 lectures on cache coherence and consistency.
1 Towards Scalable and Energy-Efficient Memory System Architectures Rajeev Balasubramonian School of Computing University of Utah.
1 Lecture: Memory, Coherence Protocols Topics: wrap-up of memory systems, multi-thread programming models, snooping-based protocols.
1 Lecture 4: Memory: HMC, Scheduling Topics: BOOM, memory blades, HMC, scheduling policies.
Survey of Existing Memory Devices Renee Gayle M. Chua.
1 Lecture: Virtual Memory, DRAM Main Memory Topics: virtual memory, TLB/cache access, DRAM intro (Sections 2.2)
Memory Cell Operation.
1 Lecture 14: DRAM Main Memory Systems Today: cache/TLB wrap-up, DRAM basics (Section 2.3)
1 Lecture 2: Memory Energy Topics: energy breakdowns, handling overfetch, LPDRAM, row buffer management, channel energy, refresh energy.
COMP541 Memories II: DRAMs
1 Adapted from UC Berkeley CS252 S01 Lecture 18: Reducing Cache Hit Time and Main Memory Design Virtucal Cache, pipelined cache, cache summary, main memory.
1 Lecture 5: Scheduling and Reliability Topics: scheduling policies, handling DRAM errors.
1 Lecture 3: Memory Buffers and Scheduling Topics: buffers (FB-DIMM, RDIMM, LRDIMM, BoB, BOOM), memory blades, scheduling policies.
1 Lecture: Memory Technology Innovations Topics: state-of-the-art and upcoming changes: buffer chips, 3D stacking, non-volatile cells, photonics Multiprocessor.
1 Lecture: Memory Technology Innovations Topics: memory schedulers, refresh, state-of-the-art and upcoming changes: buffer chips, 3D stacking, non-volatile.
1 Lecture 3: Memory Energy and Buffers Topics: Refresh, floorplan, buffers (SMB, FB-DIMM, BOOM), memory blades, HMC.
1 Lecture: DRAM Main Memory Topics: DRAM intro and basics (Section 2.3)
CS203 – Advanced Computer Architecture Main Memory Slides adapted from Onur Mutlu (CMU)
15-740/ Computer Architecture Lecture 25: Main Memory
1 Lecture 4: Memory Scheduling, Refresh Topics: scheduling policies, refresh basics.
1 Lecture: Memory Basics and Innovations Topics: memory organization basics, schedulers, refresh,
CS161 – Design and Architecture of Computer Main Memory Slides adapted from Onur Mutlu (CMU)
COMP541 Memories II: DRAMs
Chapter 5 Internal Memory
William Stallings Computer Organization and Architecture 7th Edition
COMP541 Memories II: DRAMs
William Stallings Computer Organization and Architecture 7th Edition
Lecture 23: Cache, Memory, Security
Samira Khan University of Virginia Oct 9, 2017
Lecture 15: DRAM Main Memory Systems
William Stallings Computer Organization and Architecture 8th Edition
Lecture: Memory, Multiprocessors
Staged-Reads : Mitigating the Impact of DRAM Writes on DRAM Reads
The Main Memory system: DRAM organization
Lecture: DRAM Main Memory
Lecture 23: Cache, Memory, Virtual Memory
Computer Architecture
Lecture 22: Cache Hierarchies, Memory
William Stallings Computer Organization and Architecture 7th Edition
Lecture: DRAM Main Memory
Lecture: DRAM Main Memory
William Stallings Computer Organization and Architecture 8th Edition
Lecture: Memory Technology Innovations
Lecture: Coherence Protocols
Lecture 6: Reliability, PCM
Lecture 24: Memory, VM, Multiproc
Lecture: Coherence Protocols
Lecture 20: OOO, Memory Hierarchy
Lecture 15: Memory Design
Chapter 4: MEMORY.
Lecture 22: Cache Hierarchies, Memory
Lecture 17: Multi-threaded Applications
15-740/ Computer Architecture Lecture 19: Main Memory
DRAM Hwansoo Han.
William Stallings Computer Organization and Architecture 8th Edition
Bob Reese Micro II ECE, MSU
Presentation transcript:

1 Lecture 16: Main Memory Innovations Today: DRAM basics, innovations, trends HW5 due on Thursday; simulations can take a few hours Midterm: 32 scores of 70+ (headed towards an A or A-) Highest: 85; 6 scores of 80+

2 Memory Architecture Processor Memory Controller Address/Cmd Data DIMM Bank Row Buffer DIMM: a PCB with DRAM chips on the back and front Rank: a collection of DRAM chips that work together to respond to a request and keep the data bus full A 64-bit data bus will need 8 x8 DRAM chips or 4 x16 DRAM chips or.. Bank: a subset of a rank that is busy during one request Row buffer: the last row (say, 8 KB) read from a bank, acts like a cache

3 DRAM Array Access 16Mb DRAM array = 4096 x 4096 array of bits 12 row address bits arrive first Column decoder 12 column address bits arrive next Some bits returned to CPU 4096 bits are read out Row Access Strobe (RAS) Column Access Strobe (CAS)Row Buffer

4 Recap A rank can be organized as 16 x4 2Gb chips (high capacity) or as 8 x8 2Gb chips, or 4 x16 2Gb chips (energy-efficient) High density  large arrays  wide row buffers  overfetch Address mapping: consecutive cache lines can be placed in the same row to boost row buffer hits or in different banks or channels to boost parallelism Three types of accesses: row buffer hit, empty row access, row buffer conflict; these are influenced by when we decide to close a row (by precharging bitlines)

5 Open/Closed Page Policies If an access stream has locality, a row buffer is kept open  Row buffer hits are cheap (open-page policy)  Row buffer miss is a bank conflict and expensive because precharge is on the critical path If an access stream has little locality, bitlines are precharged immediately after access (close-page policy)  Nearly every access is a row buffer miss  The precharge is usually not on the critical path Modern memory controller policies lie somewhere between these two extremes (usually proprietary)

6 Reads and Writes A single bus is used for reads and writes The bus direction must be reversed when switching between reads and writes; this takes time and leads to bus idling Hence, writes are performed in bursts; a write buffer stores pending writes until a high water mark is reached Writes are drained until a low water mark is reached

7 Scheduling Policies FCFS: Issue the first read or write in the queue that is ready for issue First Ready - FCFS: First issue row buffer hits if you can Stall Time Fair: First issue row buffer hits, unless other threads are being neglected

8 Refresh Every DRAM cell must be refreshed within a 64 ms window A row read/write automatically refreshes the row Every refresh command performs refresh on a number of rows, the memory system is unavailable during that time A refresh command is issued by the memory controller once every 7.8us on average

9 Error Correction For every 64-bit word, can add an 8-bit code that can detect two errors and correct one error; referred to as SECDED – single error correct double error detect A rank is now made up of 9 x8 chips, instead of 8 x8 chips Stronger forms of error protection exist: a system is chipkill correct if it can handle an entire DRAM chip failure

10 Modern Memory System PROC.. 4 DDR3 channels 64-bit data channels 800 MHz channels 1-2 DIMMs/channel 1-4 ranks/channel

11 Cutting-Edge Systems PROC SMB.. The link into the processor is narrow and high frequency The Scalable Memory Buffer chip is a “router” that connects to multiple DDR3 channels (wide and slow) Boosts processor pin bandwidth and memory capacity More expensive, high power

12 Future Memory Trends Processor pin count is not increasing High memory bandwidth requires high pin frequency High memory capacity requires narrow channels per “DIMM” 3D stacking can enable high memory capacity and high channel frequency (e.g., Micron HMC)

13 Future Memory Cells DRAM cell scaling is expected to slow down Emerging memory cells are expected to have better scaling properties and eventually higher density: phase change memory (PCM), spin torque transfer (STT-RAM), etc. PCM: heat and cool a material with elec pulses – the rate of heat/cool determines if the material is crystalline/amorphous; amorphous has higher resistance (i.e., no longer using capacitive charge to store a bit) Advantages: non-volatile, high density, faster than Flash/disk Disadvantages: poor write latency/energy, low endurance

14 Silicon Photonics Game-changing technology that uses light waves for communication; not mature yet and high cost likely No longer relies on pins; a few waveguides can emerge from a processor Each waveguide carries (say) 64 wavelengths of light (dense wave division multiplexing – DWDM) The signal on a wavelength can be modulated at high frequency – gives very high bandwidth per waveguide

15 Title Bullet