Review °Apply Principle of Locality Recursively °Manage memory to disk? Treat as cache Included protection as bonus, now critical Use Page Table of mappings.

Slides:



Advertisements
Similar presentations
16.317: Microprocessor System Design I
Advertisements

Lecture 34: Chapter 5 Today’s topic –Virtual Memories 1.
COMP 3221: Microprocessors and Embedded Systems Lectures 27: Virtual Memory - III Lecturer: Hui Wu Session 2, 2005 Modified.
Virtual Memory Adapted from lecture notes of Dr. Patterson and Dr. Kubiatowicz of UC Berkeley.
CS61C L35 VM I (1) Garcia © UCB Lecturer PSOE Dan Garcia inst.eecs.berkeley.edu/~cs61c CS61C : Machine Structures Lecture.
CS61C L34 Virtual Memory I (1) Garcia, Spring 2007 © UCB 3D chips from IBM  IBM claims to have found a way to build 3D chips by stacking them together.
Chap. 7.4: Virtual Memory. CS61C L35 VM I (2) Garcia © UCB Review: Caches Cache design choices: size of cache: speed v. capacity direct-mapped v. associative.
CS 61C L24 VM II (1) Garcia, Spring 2004 © UCB Lecturer PSOE Dan Garcia inst.eecs.berkeley.edu/~cs61c CS61C : Machine Structures.
Virtual Memory Adapted from lecture notes of Dr. Patterson and Dr. Kubiatowicz of UC Berkeley and Rabi Mahapatra & Hank Walker.
Inst.eecs.berkeley.edu/~cs61c UCB CS61C : Machine Structures Lecture 34 – Virtual Memory II Jim Cicconi of AT&T says that the surge of video.
Virtual Memory.
CS 61C L23 VM I (1) A Carle, Summer 2006 © UCB inst.eecs.berkeley.edu/~cs61c/su06 CS61C : Machine Structures Lecture #23: VM I Andy Carle.
COMP3221 lec36-vm-I.1 Saeid Nooshabadi COMP 3221 Microprocessors and Embedded Systems Lectures 13: Virtual Memory - I
Memory Management and Paging CSCI 3753 Operating Systems Spring 2005 Prof. Rick Han.
CS 61C L24 VM II (1) A Carle, Summer 2005 © UCB inst.eecs.berkeley.edu/~cs61c/su05 CS61C : Machine Structures Lecture #24: VM II Andy Carle.
CS61C L25 Virtual Memory I (1) Beamer, Summer 2007 © UCB Scott Beamer, Instructor inst.eecs.berkeley.edu/~cs61c CS61C : Machine Structures Lecture #25.
COMP 3221: Microprocessors and Embedded Systems Lectures 27: Virtual Memory - II Lecturer: Hui Wu Session 2, 2005 Modified.
CS61C L19 VM © UC Regents 1 CS61C - Machine Structures Lecture 19 - Virtual Memory November 3, 2000 David Patterson
CS 61C L25 VM III (1) Garcia, Spring 2004 © UCB TA Chema González www-inst.eecs/~cs61c-td inst.eecs.berkeley.edu/~cs61c CS61C : Machine Structures Lecture.
CS61C L35 Virtual Memory II (1) Garcia, Spring 2007 © UCB Hardware repair?!  This technology allows you to “patch” your hardware after it has been installed.
Inst.eecs.berkeley.edu/~cs61c UCB CS61C : Machine Structures Lecture 35 – Virtual Memory III Researchers at three locations have recently added.
Chapter 3.2 : Virtual Memory
Translation Buffers (TLB’s)
CS61C L34 Virtual Memory II (1) Garcia, Fall 2006 © UCB Lecturer SOE Dan Garcia inst.eecs.berkeley.edu/~cs61c UC Berkeley.
CS61C L36 VM II (1) Garcia, Fall 2004 © UCB Lecturer PSOE Dan Garcia inst.eecs.berkeley.edu/~cs61c CS61C : Machine Structures.
CS61C L33 Virtual Memory I (1) Garcia, Fall 2006 © UCB Lecturer SOE Dan Garcia inst.eecs.berkeley.edu/~cs61c UC Berkeley.
©UCB CS 162 Ch 7: Virtual Memory LECTURE 13 Instructor: L.N. Bhuyan
CS 430 – Computer Architecture 1 CS 430 – Computer Architecture Virtual Memory William J. Taffe using slides of David Patterson.
Cs 61C L18 Cache2.1 Patterson Spring 99 ©UCB CS61C Improving Cache Memory and Virtual Memory Introduction Lecture 18 April 2, 1999 Dave Patterson (http.cs.berkeley.edu/~patterson)
Inst.eecs.berkeley.edu/~cs61c UCB CS61C : Machine Structures Lecture 33 – Virtual Memory I OCZ has released a 1 TB solid state drive (the biggest.
Inst.eecs.berkeley.edu/~cs61c UCB CS61C : Machine Structures Lecture 34 – Virtual Memory II Researchers at Stanford have developed “nanoscale.
©UCB CS 161 Ch 7: Memory Hierarchy LECTURE 24 Instructor: L.N. Bhuyan
COMP3221 lec37-vm-II.1 Saeid Nooshabadi COMP 3221 Microprocessors and Embedded Systems Lectures 13: Virtual Memory - II
Review of Memory Management, Virtual Memory CS448.
CS 61C L7.1.2 VM II (1) K. Meinz, Summer 2004 © UCB CS61C : Machine Structures Lecture VM II Kurt Meinz inst.eecs.berkeley.edu/~cs61c.
1 Chapter 3.2 : Virtual Memory What is virtual memory? What is virtual memory? Virtual memory management schemes Virtual memory management schemes Paging.
Lecture Topics: 11/17 Page tables TLBs Virtual memory flat page tables
Inst.eecs.berkeley.edu/~cs61c UCB CS61C : Machine Structures Lecture 35 – Virtual Memory III ”The severity of the decline in the market is further evidence.
Virtual Memory. Virtual Memory: Topics Why virtual memory? Virtual to physical address translation Page Table Translation Lookaside Buffer (TLB)
Review (1/2) °Caches are NOT mandatory: Processor performs arithmetic Memory stores data Caches simply make data transfers go faster °Each level of memory.
Virtual Memory. Virtual Memory: Topics Why virtual memory? Virtual to physical address translation Page Table Translation Lookaside Buffer (TLB)
Multilevel Caches Microprocessors are getting faster and including a small high speed cache on the same chip.
Virtual Memory.  Next in memory hierarchy  Motivations:  to remove programming burdens of a small, limited amount of main memory  to allow efficient.
Csci 136 Computer Architecture II – Virtual Memory
Virtual Memory 1 Computer Organization II © McQuain Virtual Memory Use main memory as a “cache” for secondary (disk) storage – Managed jointly.
CS203 – Advanced Computer Architecture Virtual Memory.
CS161 – Design and Architecture of Computer
Lecturer PSOE Dan Garcia
ECE232: Hardware Organization and Design
CS161 – Design and Architecture of Computer
Virtual Memory Use main memory as a “cache” for secondary (disk) storage Managed jointly by CPU hardware and the operating system (OS) Programs share main.
IT 251 Computer Organization and Architecture
CS510 Operating System Foundations
Review (1/2) Caches are NOT mandatory:
Csci 211 Computer System Architecture – Review on Virtual Memory
Some of the slides are adopted from David Patterson (UCB)
Morgan Kaufmann Publishers Memory Hierarchy: Virtual Memory
Translation Buffers (TLB’s)
CSE 451: Operating Systems Autumn 2003 Lecture 10 Paging & TLBs
CSE451 Virtual Memory Paging Autumn 2002
Translation Buffers (TLB’s)
CSE 451: Operating Systems Autumn 2003 Lecture 10 Paging & TLBs
Albert Chae, Instructor
COMP3221: Microprocessors and Embedded Systems
Paging and Segmentation
Translation Buffers (TLBs)
Virtual Memory Use main memory as a “cache” for secondary (disk) storage Managed jointly by CPU hardware and the operating system (OS) Programs share main.
CS61C Virtual Memory, 2nd try + Advice on Courses Lecture 19
Review What are the advantages/disadvantages of pages versus segments?
Virtual Memory 1 1.
Presentation transcript:

Review °Apply Principle of Locality Recursively °Manage memory to disk? Treat as cache Included protection as bonus, now critical Use Page Table of mappings vs. tag/data in cache °Virtual Memory allows protected sharing of memory between processes with less swapping to disk, less fragmentation than always swap or base/bound

Overview °Review Virtual Memory °TLBs °Multilevel Page Tables

Why Virtual Memory? °Want to give each running program its own private address space °Want programs to be protected from each other (bug in one program can’t corrupt memory in another program) °Want programs running simultaneously to share underlying physical memory °Want to use disk as another level in the memory hierarchy Treat main memory as a cache for disk

Review: Address Translation °Each program operates in its own virtual address space; ~only program running °Each is protected from the other °OS can decide where each goes in memory °Hardware (HW) provides virtual -> physical mapping virtual address (inst. fetch load, store) Program operates in its virtual address space HW mapping physical address (inst. fetch load, store) Physical memory (incl. caches)

Review: Paging 0 Physical Memory  Virtual Memory CodeStatic Heap Stack 64 MB °Divide into equal sized chunks (about 4KB) 0 °Any chunk of Virtual Memory assigned to any chuck of Physical Memory (“page”)

Address Mapping: Page Table Virtual Address: page no.offset Page Table Base Reg Page Table located in physical memory index into page table + Physical Memory Address Page Table Val -id Access Rights Physical Page Address. V A.R. P. P. A....

Notes on Page Table °Solves Fragmentation problem: all chunks same size, so all holes can be used °OS must reserve “Swap Space” on disk for each process °To grow a process, ask Operating System If unused pages, OS uses them first If not, OS swaps some old pages to disk (Least Recently Used to pick pages to swap) °Each process has own Page Table °Will add details, but Page Table is essence of Virtual Memory

Virtual Memory Problem #1 °Not enough physical memory! Only, say, 64 MB of physical memory N processes, each 4GB of virtual memory! Could have 1K virtual pages/physical page! °Spatial Locality to the rescue Each page is 4 KB, lots of nearby references No matter how big program is, at any time only accessing a few pages “Working Set”: recently used pages

Virtual Address and a Cache Processor Trans- lation Cache Main Memory VA PA miss hit data Cache typically operates on physical addresses Page Table access is another memory access for each program memory access! Need to fix this!

Virtual Memory Problem #2 °Map every address  1 extra memory accesses for every memory access °Observation: since locality in pages of data, must be locality in virtual addresses of those pages °Why not use a cache of virtual to physical address translations to make translation fast? (small is fast) °For historical reasons, cache is called a Translation Lookaside Buffer, or TLB

Typical TLB Format VirtualPhysicalDirtyRef Valid Access Address AddressRights TLB just a cache on the page table mappings TLB access time comparable to cache (much less than main memory access time) Ref: Used to help calculate LRU on replacement Dirty: since use write back, need to know whether or not to write page to disk when replaced

What if not in TLB? °Option 1: Hardware checks page table and loads new Page Table Entry into TLB °Option 2: Hardware traps to OS, up to OS to decide what to do °MIPS follows Option 2: Hardware knows nothing about page table format

TLB Miss (simplified format) °If the address is not in the TLB, MIPS traps to the operating system When in the operating system, we don't do translation (turn off virtual memory) °The operating system knows which program caused the TLB fault, page fault, and knows what the virtual address desired was requested So we look the data up in the page table 291 valid virtual physical

If the data is in memory °We simply add the entry to the TLB, evicting an old entry from the TLB valid virtual physical

What if the data is on disk? °We load the page off the disk into a free block of memory, using a DMA transfer Meantime we switch to some other process waiting to be run °When the DMA is complete, we get an interrupt and update the process's page table So when we switch back to the task, the desired data will be in memory

What if we don't have enough memory? °We chose some other page belonging to a program and transfer it onto the disk if it is dirty If clean (other copy is up-to-date), just overwrite that data in memory We chose the page to evict based on replacement policy (e.g., LRU) °And update that program's page table to reflect the fact that its memory moved somewhere else

Translation Look-Aside Buffers TLBs usually small, typically entries Like any other cache, the TLB can be fully associative, set associative, or direct mapped Processor TLB Lookup Cache Main Memory VA PA miss hit data Trans- lation hit miss

Virtual Memory Problem #3 °Page Table too big! 4GB Virtual Memory ÷ 4 KB page  ~ 1 million Page Table Entries  4 MB just for Page Table for 1 process, 25 processes  100 MB for Page Tables! °Variety of solutions to tradeoff memory size of mapping function for slower when miss TLB Make TLB large enough, highly associative so rarely miss on address translation

2-level Page Table 0 Physical Memory 64 MB Virtual Memory  Code Static Heap Stack nd Level Page Tables Super Page Table

Page Table Shrink : °Single Page Table Page Number Offset 20 bits12 bits °Multilevel Page Table Page Number Super Page No. Offset 10 bits 12 bits °Only have second level page table for valid entries of super level page table

Space Savings for Multi-Level Page Table °If only 10% of entries of Super Page Table have valid enties, then total mapping size is roughly 1/10-th of single level page table Exercise 7.35 explores exact size

Note: Actual MIPS Process Memory Allocation 0  ( ) Address Code Static User code/data space Heap Stack I/O device registers $sp $gp  2  ( ) I/O Regs Except.Exception Handlers OS code/data space  2  (2 31 ) OS restricts I/O Registers, Exception Handlers to OS

Things to Remember 1/2 °Apply Principle of Locality Recursively °Manage memory to disk? Treat as cache Included protection as bonus, now critical Use Page Table of mappings vs. tag/data in cache °Virtual memory to Physical Memory Translation too slow? Add a cache of Virtual to Physical Address Translations, called a TLB

Things to Remember 2/2 °Virtual Memory allows protected sharing of memory between processes with less swapping to disk, less fragmentation than always swap or base/bound °Spatial Locality means Working Set of Pages is all that must be in memory for process to run fairly well °TLB to reduce performance cost of VM °Need more compact representation to reduce memory size cost of simple 1-level page table (especially 32-  64-bit address)