Virtual Memory 3 Hakim Weatherspoon CS 3410, Spring 2011

Slides:



Advertisements
Similar presentations
Computer Organization CS224 Fall 2012 Lesson 44. Virtual Memory  Use main memory as a “cache” for secondary (disk) storage l Managed jointly by CPU hardware.
Advertisements

Lecture 34: Chapter 5 Today’s topic –Virtual Memories 1.
Virtual Memory 3 Hakim Weatherspoon CS 3410, Spring 2011 Computer Science Cornell University P & H Chapter
Virtual Memory 2 Hakim Weatherspoon CS 3410, Spring 2011 Computer Science Cornell University P & H Chapter
Kevin Walsh CS 3410, Spring 2010 Computer Science Cornell University Virtual Memory 2 P & H Chapter
CS 153 Design of Operating Systems Spring 2015
CSCE 212 Chapter 7 Memory Hierarchy Instructor: Jason D. Bakos.
Computer ArchitectureFall 2008 © November 10, 2007 Nael Abu-Ghazaleh Lecture 23 Virtual.
Memory Management and Paging CSCI 3753 Operating Systems Spring 2005 Prof. Rick Han.
Computer ArchitectureFall 2007 © November 21, 2007 Karem A. Sakallah Lecture 23 Virtual Memory (2) CS : Computer Architecture.
Translation Buffers (TLB’s)
©UCB CS 162 Ch 7: Virtual Memory LECTURE 13 Instructor: L.N. Bhuyan
©UCB CS 161 Ch 7: Memory Hierarchy LECTURE 24 Instructor: L.N. Bhuyan
Virtual Memory 2 Hakim Weatherspoon CS 3410, Spring 2013 Computer Science Cornell University P & H Chapter 5.4.
Prof. Kavita Bala and Prof. Hakim Weatherspoon CS 3410, Spring 2014 Computer Science Cornell University P & H Chapter 5.7 Handouts in front/back of the.
Lecture 19: Virtual Memory
Lecture 15: Virtual Memory EEN 312: Processors: Hardware, Software, and Interfacing Department of Electrical and Computer Engineering Spring 2014, Dr.
Multilevel Caches Microprocessors are getting faster and including a small high speed cache on the same chip.
Virtual Memory 3 Hakim Weatherspoon CS 3410, Spring 2012 Computer Science Cornell University P & H Chapter 5.4.
CS2100 Computer Organisation Virtual Memory – Own reading only (AY2015/6) Semester 1.
Lab4: Virtual Memory CS 3410 : Computer System Organization & Programming Spring 2015.
Virtual Memory Ch. 8 & 9 Silberschatz Operating Systems Book.
Virtual Memory Review Goal: give illusion of a large memory Allow many processes to share single memory Strategy Break physical memory up into blocks (pages)
Prof. Hakim Weatherspoon CS 3410, Spring 2015 Computer Science Cornell University P & H Chapter 5.7.
CS203 – Advanced Computer Architecture Virtual Memory.
CS161 – Design and Architecture of Computer
Translation Lookaside Buffer
Memory Hierarchy Ideal memory is fast, large, and inexpensive
Virtual Memory Samira Khan Apr 27, 2017.
Virtual Memory Acknowledgment
Virtual Memory Chapter 7.4.
ECE232: Hardware Organization and Design
Memory COMPUTER ARCHITECTURE
CS161 – Design and Architecture of Computer
Memory and cache CPU Memory I/O.
Section 9: Virtual Memory (VM)
From Address Translation to Demand Paging
From Address Translation to Demand Paging
Today How was the midterm review? Lab4 due today.
CS 704 Advanced Computer Architecture
Virtual Memory Use main memory as a “cache” for secondary (disk) storage Managed jointly by CPU hardware and the operating system (OS) Programs share main.
Memory Hierarchy Virtual Memory, Address Translation
Cache Memory Presentation I
Morgan Kaufmann Publishers
CSE 153 Design of Operating Systems Winter 2018
Virtual Memory 2 Hakim Weatherspoon CS 3410, Spring 2012
Andy Wang Operating Systems COP 4610 / CGS 5765
Memory and cache CPU Memory I/O.
Andy Wang Operating Systems COP 4610 / CGS 5765
Translation Lookaside Buffer
Morgan Kaufmann Publishers Memory Hierarchy: Virtual Memory
Translation Buffers (TLB’s)
CSE 451: Operating Systems Autumn 2003 Lecture 10 Paging & TLBs
CS 3410, Spring 2014 Computer Science Cornell University
© 2004 Ed Lazowska & Hank Levy
Virtual Memory 2 Hakim Weatherspoon CS 3410, Spring 2012
Translation Buffers (TLB’s)
CSC3050 – Computer Architecture
CSE 451: Operating Systems Autumn 2003 Lecture 10 Paging & TLBs
CS 3410 Computer System Organization & Programming
CSE 153 Design of Operating Systems Winter 2019
Virtual Memory Lecture notes from MKP and S. Yalamanchili.
Translation Buffers (TLBs)
Virtual Memory Use main memory as a “cache” for secondary (disk) storage Managed jointly by CPU hardware and the operating system (OS) Programs share main.
CS161 – Design and Architecture of Computer
Sarah Diesburg Operating Systems CS 3430
Andy Wang Operating Systems COP 4610 / CGS 5765
Review What are the advantages/disadvantages of pages versus segments?
Sarah Diesburg Operating Systems COP 4610
CS 444/544 Operating Systems II Virtual Memory Translation
Presentation transcript:

Virtual Memory 3 Hakim Weatherspoon CS 3410, Spring 2011 Computer Science Cornell University P & H Chapter 5.4-5

Announcements PA3 available. Due Tuesday, April 19th Next four weeks Work with pairs Be responsible with new knowledge Scheduling a games night, possibly Friday, April 22nd Next four weeks Two projects and one homeworks Prelim2 will be Thursday, April 28th PA4 will be final project (no final exam) Will not be able to use slip days

Goals for Today Virtual Memory Address Translation Paging Pages, page tables, and memory mgmt unit Paging Role of Operating System Context switches, working set, shared memory Performance How slow is it Making virtual memory fast Translation lookaside buffer (TLB) Virtual Memory Meets Caching

Making Virtual Memory Fast The Translation Lookaside Buffer (TLB)

Translation Lookaside Buffer (TLB) Hardware Translation Lookaside Buffer (TLB) A small, very fast cache of recent address mappings TLB hit: avoids PageTable lookup TLB miss: do PageTable lookup, cache result for later

TLB Diagram V R W X D tag ppn V R W X D invalid 1 V invalid 1

A TLB in the Memory Hierarchy Lookup Cache CPU Mem Disk PageTable Lookup (1) Check TLB for vaddr (~ 1 cycle) (2) TLB Miss: traverse PageTables for vaddr (3a) PageTable has valid entry for in-memory page Load PageTable entry into TLB; try again (tens of cycles) (3b) PageTable has entry for swapped-out (on-disk) page Page Fault: load from disk, fix PageTable, try again (millions of cycles) (3c) PageTable has invalid entry Page Fault: kill process (2) TLB Hit compute paddr, send to cache TLB miss in hardware usually, but not always PageTable stuff in software

TLB Coherency TLB Coherency: What can go wrong? A: PageTable or PageDir contents change swapping/paging activity, new shared pages, … A: Page Table Base Register changes context switch between processes Fully transparent

Translation Lookaside Buffers (TLBs) When PTE changes, PDE changes, PTBR changes…. Full Transparency: TLB coherency in hardware Flush TLB whenever PTBR register changes [easy – why?] Invalidate entries whenever PTE or PDE changes [hard – why?] TLB coherency in software If TLB has a no-write policy… OS invalidates entry after OS modifies page tables OS flushes TLB whenever OS does context switch process ID could just be the PTBR for the process

TLB Parameters TLB parameters (typical) Intel Nehalem TLB (example) very small (64 – 256 entries), so very fast fully associative, or at least set associative tiny block size: why? Intel Nehalem TLB (example) 128-entry L1 Instruction TLB, 4-way LRU 64-entry L1 Data TLB, 4-way LRU 512-entry L2 Unified TLB, 4-way LRU

Virtual Memory meets Caching Virtually vs. physically addressed caches Virtually vs. physically tagged caches

Virtually Addressed Caching Q: Can we remove the TLB from the critical path? A: Virtually-Addressed Caches TLB Lookup CPU Mem Disk Virtually Addressed Cache PageTable Lookup A: have to flush entire cache on context switch A: Doing synonym updates requires significant hardware – essentially an associative lookup on the physical address tags to see if you have multiple hits

Virtual vs. Physical Caches addr Memory DRAM CPU Cache SRAM MMU data Cache works on physical addresses addr Memory DRAM CPU Cache SRAM MMU data Cache works on virtual addresses A1: Physically-addressed: nothing; Virtually-addressed: need to flush cache A2: Physically-addressed: nothing; Virtually-addressed: problems Q: What happens on context switch? Q: What about virtual memory aliasing? Q: So what’s wrong with physically addressed caches?

Indexing vs. Tagging Physically-Addressed Cache slow: requires TLB (and maybe PageTable) lookup first Virtually-Indexed, Virtually Tagged Cache fast: start TLB lookup before cache lookup finishes PageTable changes (paging, context switch, etc.)  need to purge stale cache lines (how?) Synonyms (two virtual mappings for one physical page)  could end up in cache twice (very bad!) Virtually-Indexed, Physically Tagged Cache ~fast: TLB lookup in parallel with cache lookup PageTable changes  no problem: phys. tag mismatch Synonyms  search and evict lines with same phys. tag Virtually-Addressed Cache

Typical Cache Setup Memory DRAM CPU addr L2 Cache SRAM L1 Cache SRAM MMU data TLB SRAM Typical L1: On-chip virtually addressed, physically tagged Typical L2: On-chip physically addressed Typical L3: On-chip …

Caches/TLBs/VM Caches, Virtual Memory, & TLBs Where can block be placed? Direct, n-way, fully associative What block is replaced on miss? LRU, Random, LFU, … How are writes handled? No-write (w/ or w/o automatic invalidation) Write-back (fast, block at time) Write-through (simple, reason about consistency) Where? Caches: direct/n-way/fa VM: fa, but with a table of contents to eliminate searches TLB: fa Replacement? varied Writes? Caches: usually write-back, or maybe write-through, or maybe no-write w/ invalidation VM: write-back TLB: usually no-write

Summary of Cache Design Parameters L1 Paged Memory TLB Size (blocks) 1/4k to 4k 16k to 1M 64 to 4k Size (kB) 16 to 64 1M to 4G 2 to 16 Block size (B) 16-64 4k to 64k 4-32 Miss rates 2%-5% 10-4 to 10-5% 0.01% to 2% Miss penalty 10-25 10M-100M 100-1000