• Non ci sono risultati.

Address Translation

N/A
N/A
Protected

Academic year: 2022

Condividi "Address Translation"

Copied!
64
0
0

Testo completo

(1)

Address Translation

(2)

Main Points

• Address Translation Concept

– How do we convert a virtual address to a physical address?

• Flexible Address Translation

– Base and bound – Segmentation – Paging

– Multilevel translation

• Efficient Address Translation

– Translation Lookaside Buffers

– Virtually and Physically Addressed Caches

(3)

Address Translation Concept

(4)

Address Translation Goals

• Memory protection

• Memory sharing

• Flexible memory placement

• Sparse addresses

• Runtime lookup efficiency

• Compact translation tables

• Portability

(5)

Address Translation

• What can you do if you can (selectively) gain control whenever a program reads or writes a particular memory location?

– With hardware support

– With compiler-level support

• Memory management is one of the most complex parts of the OS

– Serves many different purposes

(6)

Address Translation Uses

• Process isolation

– Keep a process from touching anyone else’s memory, or the kernel’s

• Efficient interprocess communication

– Shared regions of memory between processes

• Shared code segments

– E.g., common libraries used by many different programs

• Program initialization

– Start running a program before it is entirely in memory

• Dynamic memory allocation

– Allocate and initialize stack/heap pages on demand

(7)

Address Translation (more)

• Cache management

– Page coloring

• Program debugging

– Data breakpoints when address is accessed

• Zero-copy I/O

– Directly from I/O device into/out of user memory

• Memory mapped files

– Access file data using load/store instructions

• Demand-paged virtual memory

– Illusion of near-infinite memory, backed by disk or memory on other machines

(8)

Address Translation (even more)

• Checkpointing/restart

– Transparently save a copy of a process, without stopping the program while the save happens

• Persistent data structures

– Implement data structures that can survive system reboots

• Process migration

– Transparently move processes between machines

• Information flow control

– Track what data is being shared externally

• Distributed shared memory

– Illusion of memory that is shared between machines

(9)

Address Translation (summary)

Implement a

virtual memory

(10)

Immagine in memoria di un processo

Indirizzi virtuali

entry point del programma

base dello stack codice

dati

stack

0 160

4096

6140 5632

(11)

Preparazione di un programma per l’esecuzione

O1 O2 O3

MC S1

S2 S3

traduzione (compilatore)

libreria

collegamento (linker)

caricamento (loader)

S1, S2, S3: moduli sorgente; O1, O2, O3: moduli oggetto;

MC: modulo di caricamento (file eseguibile);

IP: immagine del processo

IP

codice

dati

stack

0 160

4096

6140 5632

(12)

Rilocazione degli indirizzi: rilocazione statica

caricatore rilocante 0

160

LOAD 4112 B:………

JUMP 168 168

192 256

A:………

modulo di caricamento (indirizzi virtuali) 4112

Stack Pointer Program Counter

16380 10240

10400 10432 10408

14352 14336

16380 10496

B: ………

JUMP 10408 LOAD 14352

A: ………

memoria fisica

10240

immagine processo

(13)

Rilocazione degli indirizzi: rilocazione dinamica

caricatore

Program Counter

Stack Pointer

6140 10240

10400 10432 10408

14352 14336

16380 10496

B: ………

JUMP 168 LOAD 4112

A: ………

memoria fisica

160

immagine processo 0

160

LOAD 4112 B:………

JUMP 168 168

192 256

A:………

modulo di caricamento (indirizzi virtuali) 4112

6140

(14)

Virtual Base and Bounds

(15)

Rilocazione dinamica: Memory Management Unit

10240 10400 10432 10408

14352 14336

16380 10496

B: ………

JUMP 168 LOAD 4112

A: ………

memoria fisica

immagine processo

registro limite: 6140

+

indirizzo fisico y = 14352 indirizzo

virtuale x = 4112

eccezione

<

si CPU no

MMU

registro base:10240

(16380- 10240= 6140)

(16)

Virtual Base and Bounds

• Pros?

– Simple

– Fast (2 registers, adder, comparator)

– Can relocate in physical memory without changing process

• Cons?

– Can’t keep program from accidentally overwriting its own code

– Can’t share code/data with other processes – Can’t grow stack/heap as needed

(17)

Segmentation

• Segment is a contiguous region of memory

– Virtual or (for now) physical memory

• Each process has a segment table (in hardware)

– Entry in table = segment

• Segment can be located anywhere in physical memory

– Start – Length

– Access permission

• Processes can share segments

– Same start, length, same/different access permissions

(18)

Segmentazione

0 160

0

LOAD (1,16) B: ……

JUMP(0,168) 168

16 192 256

A: ……

10240 10400 10432 10408

34016 34000 10496

B: ……

JUMP (0,168) LOAD (1,16)

Spazio virtuale

segmentato Immagine del processo in memoria

memoria fisica caricatore

non rilocante

20000 0

A: ……

segmento 0: codice

segmento 1: dati

segmento 2: stack

segmento codice

segmento stack

segmento dati

Spazio virtuale segmentato

(19)
(20)

main: 240 store #1108, r2

244 store pc+8, r31

248 jump 360

24c

strlen: 360 loadbyte (r2), r3

420 jump (r31)

x: 1108 a b c \0

x: 108 a b c \0

main: 4240 store #1108, r2

4244 store pc+8, r31

4248 jump 360

424c

strlen: 4360 loadbyte (r2),r3

4420 jump (r31)

Segment start length

code 0x4000 0x700

data 0 0x500

heap - -

stack 0x2000 0x1000

Virtual Memory Physical Memory

2 bit segment # 12 bit offset

(21)

UNIX fork and Copy on Write

• UNIX fork

– Makes a complete copy of a process

• Segments allow a more efficient implementation

– Copy segment table into child

– Mark parent and child segments read-only – Start child process; return to parent

– If child or parent writes to a segment, will trap into kernel

• make a copy of the segment and resume

(22)
(23)

Zero-on-Reference

• How much physical memory do we need to allocate for the stack or heap?

– Zero bytes!

• When program touches the heap

– Segmentation fault into OS kernel – Kernel allocates some memory

• How much?

– Zeros the memory

• avoid accidentally leaking information!

– Restart process

(24)

Segmentation

• Pros?

– Can share code/data segments between processes – Can protect code segment from being overwritten – Can transparently grow stack/heap as needed

– Can detect if need to copy-on-write

• Cons?

– Complex memory management

• Need to find chunk of a particular size

– May need to rearrange memory from time to time to make room for new segment or growing segment

• External fragmentation: wasted space between chunks

(25)

Paged Translation

• Manage memory in fixed size units, or pages

• Finding a free page is easy

– Bitmap allocation: 0011111100000001100 – Each bit represents one physical page frame

• Each process has its own page table

– Stored in physical memory – Hardware registers

• pointer to page table start

• page table length

(26)
(27)

A B C D E F G H I J K L

I J K L

E F G H A B C D 4

3 1

Page Table

Process View Physical Memory

(28)

Paging Questions

• What must be saved/restored on a process context switch?

– Pointer to page table/size of page table – Page table itself is in main memory

• What if page size is very small?

• What if page size is very large?

– Internal fragmentation: if we don’t need all of the

space inside a fixed size chunk

(29)

Paging and Copy on Write

• Can we share memory between processes?

– Set entries in both page tables to point to same page frames

– Need core map of page frames to track which processes are pointing to which page frames

• UNIX fork with copy on write at page granularity

– Copy page table entries to new process – Mark all pages as read-only

– Trap into kernel on write (in child or parent) – Copy page and resume execution

(30)

Paging and Fast Program Start

• Can I start running a program before its code is in physical memory?

– Set all page table entries to invalid

– When a page is referenced for first time

• Trap to OS kernel

• OS kernel brings in page

• Resumes execution

– Remaining pages can be transferred in the

background while program is running

(31)

Sparse Address Spaces

• Might want many separate segments

– Per-processor heaps – Per-thread stacks

– Memory-mapped files

– Dynamically linked libraries

• What if virtual address space is sparse?

– On 32-bit UNIX, code starts at 0 – Stack starts at 2^31

– 4KB pages => 500K page table entries

– 64-bits => 4 quadrillion page table entries

(32)

Multi-level Translation

• Tree of translation tables

– Paged segmentation – Multi-level page tables

– Multi-level paged segmentation

• All 3: Fixed size page as lowest level unit

– Efficient memory allocation – Efficient disk transfers

– Easier to build translation lookaside buffers

– Efficient reverse lookup (from physical -> virtual) – Page granularity for protection/sharing

(33)

Paged Segmentation

• Process memory is segmented

• Segment table entry:

– Pointer to page table

– Page table length (# of pages in segment) – Access permissions

• Page table entry:

– Page frame

– Access permissions

• Share/protection at either page or segment-level

(34)
(35)

Multilevel Paging

(36)

Paginazione a due livelli

x =

20

offset pagina virtuale

10 10 RPTP

tabella di 1° livello tabella di 2° livello ptp2

pf

y = pf of

offset pagina fisica

pagina fisica of

L1 L2

CPU pg

12 of

idx2 idx

pf

Caricamento dinamico delle tabelle delle pagine di secondo livello

==> minore occupazione di memoria

(37)

Confronto tra tabelle delle pagine a 1 o 2 livelli

IPOTESI:

• Indirizzi logici di 32 bit; pagine logiche e fisiche di 4 kByte.

==> lunghezza del campo offset : 12 bit; indice di pagina codificato con 20 bit

• Descrittori di pagina (elementi della tabella delle pagine) codificati con 4 byte, di cui:

- 3 byte (24 bit) per la codifica dell’indice di blocco;

- 1 byte riservato agli indicatori

TABELLA DELLE PAGINE A 1 LIVELLO:

• Numero di elementi della tabella delle pagine: 220

• Spazio occupato dalla tabella delle pagine: 220 * 4= 222 byte= 4 Mbyte

• Massima dimensione della memoria fisica: 224 blocchi

==> 224 * 212 = 236 byte = 64 Gbyte

(38)

Confronto tra tabelle delle pagine a 1 o 2 livelli

IPOTESI (come nel caso precedente)

Indirizzi logici di 32 bit; pagine logiche e fisiche di 4 kByte.

==> lunghezza del campo offset : 12 bit; indice di pagina codificato con 20 bit

Elementi di ogni tabella delle pagine (di primo o secondo livello) codificati con 4 byte, di cui:

- 3 byte (24 bit) per individuare un indice di blocco;

- 1 byte riservato agli indicatori

TABELLA DELLE PAGINE A 2 LIVELLI:

Ipotesi: 210 tabelle delle pagine di secondo livello;

==> La tabella di primo livello ha 210 elementi

==> ripartizione dell’indirizzo logico:

- 12 bit per offset;

- 10 bit per indirizzare la tabella di primo livello

- 10 bit per indirizzare la tabella di secondo livello selezionata;

(39)

Confronto tra tabelle delle pagine a 1 o 2 livelli

TABELLA DELLE PAGINE A 2 LIVELLI:

==> ogni elemento di tabella di primo livello corrisponde a una tabella di secondo livello - 3 byte: indice di blocco nel quale risiede la tabella di secondo livello (se presente) - 1 byte: indicatori (tra cui indicatore di presenza).

==> ogni elemento di tabella di secondo livello corrisponde a una pagina - 3 byte: indice di blocco nel quale risiede la pagina (se presente) - 1 byte: indicatori (tra cui indicatore di presenza).

lunghezza di ogni tabella di primo o secondo livello: 210 elementi ==> 210*4 = 4 Kbyte

massima dimensione della memoria fisica : 224 blocchi ==> 224 * 212 = 236 byte

= 64 Gbyte.

(40)

x86 Multilevel Paged Segmentation

• Global Descriptor Table (segment table)

– Pointer to page table for each segment – Segment length

– Segment access permissions

– Context switch: change global descriptor table register (GDTR, pointer to global descriptor table)

• Multilevel page table

– 4KB pages; each level of page table fits in one page

• Only fill page table if needed

– 32-bit: two level page table (per segment) – 64-bit: four level page table (per segment)

(41)

Multilevel Translation

• Pros:

– Allocate/fill only as many page tables as used – Simple memory allocation

– Share at segment or page level

• Cons:

– Space overhead: at least one pointer per virtual page

– Two or more lookups per memory reference

(42)

Portability

• Many operating systems keep their own memory translation data structures

– List of memory objects (segments) – Virtual -> physical

– Physical -> virtual

– Simplifies porting from x86 to ARM, 32 bit to 64 bit

• Inverted page table

– Hash from virtual page -> physical page – Space proportional to # of physical pages

(43)

Do we need multi-level page tables?

• Use inverted page table in hardware instead of multilevel tree

– IBM PowerPC

– Hash virtual page # to inverted page table bucket – Location in IPT => physical page frame

• Pros/cons?

(44)

Efficient Address Translation

• Translation lookaside buffer (TLB)

– Cache of recent virtual page -> physical page translations

– If cache hit, use translation

– If cache miss, walk multi-level page table

• Cost of translation =

Cost of TLB lookup +

Prob(TLB miss) * cost of page table lookup

(45)
(46)

Software Loaded TLB

• Do we need a page table at all?

– MIPS processor architecture – If translation is in TLB, ok

– If translation is not in TLB, trap to kernel – Kernel computes translation and loads TLB

– Kernel can use whatever data structures it wants

• Pros/cons?

(47)

When Do TLBs Work/Not Work?

(48)

When Do TLBs Work/Not Work?

• Video Frame Buffer: 32 bits x 1K x 1K =

4MB

(49)

Superpages

• TLB entry can be

– A page

– A superpage: a set of contiguous pages

– x86: superpage is set of pages in one page table – x86 TLB entries

• 4KB

• 2MB

• 1GB

(50)

When Do TLBs Work/Not Work, part 2

• What happens on a context switch?

– Reuse TLB?

– Discard TLB?

• Motivates hardware tagged TLB

– Each TLB entry has process ID

– TLB hit only if process ID matches current process

(51)
(52)

When Do TLBs Work/Not Work, part 3

• What happens when the OS changes the permissions on a page?

– For demand paging, copy on write, zero on reference, …

• TLB may contain old translation

– OS must ask hardware to purge TLB entry

• On a multicore: TLB shootdown

– OS must ask each CPU to purge TLB entry

(53)

TLB Shootdown

(54)

Address Translation with TLB

(55)

Virtually Addressed Caches

(56)

Memory Hierarchy

i7 has 8MB as shared 3rd level cache; 2nd level cache is per-core

(57)

Hardware Design Principle

The bigger the memory, the slower the memory

(58)

Translation on a Modern Processor

(59)

MI FERMEREI QUI

(60)

Question

• What is the cost of a first level TLB miss?

– Second level TLB lookup

• What is the cost of a second level TLB miss?

– x86: 2-4 level page table walk

• How expensive is a 4-level page table walk on

a modern processor?

(61)

Questions

• With a virtual cache, what do we need to do on a context switch?

• What if the virtual cache > page size?

– Page size: 4KB (x86)

– First level cache size: 64KB (i7)

– Cache block size: 32 bytes

(62)

Aliasing

• Alias: two (or more) virtual cache entries that refer to the same physical memory

– What if we modify one alias and then context switch?

• Typical solution

– On a write, lookup virtual cache and TLB in parallel – Physical address from TLB used to check for

aliases

(63)
(64)

Multicore and Hyperthreading

• Modern CPU has several functional units

– Instruction decode – Arithmetic/branch – Floating point

– Instruction/data cache – TLB

• Multicore: replicate functional units (i7: 4)

– Share second/third level cache, second level TLB

• Hyperthreading: logical processors that share functional units (i7: 2)

– Better functional unit utilization during memory stalls

• No difference from the OS/programmer perspective

– Except for performance, affinity, …

Riferimenti

Documenti correlati

(b) If (i) on or prior to the publication of the Reference Index to be used to calculate any Rate of Interest in relation to the index-linked interest provisions of these Notes, the

Il Vettori sembra quindi ap- plicare criteri propri della correzione (emendatio) a quelli dell’interpreta- zione (explicatio), per cui il recupero della lezione più

Its partners were represented by a core group of regional public innovation agencies and a Reflection Group in which there was Veneto

This document provides health professionals with a summary of up-to-date evidence based on an extensive literature review on the relations between air pollution

NAT overload sometimes called PAT (Port Address Translation) maps multiple unregistered or private IP addresses to a single registered or public IP address by using different

The role of the primary treating physical therapist, especially for the young child between the ages of 1 and 5 years, will incorporate the typical role that the grandmother and

Autologous and allogeneic hematopoietic stem cell transplantation (HSCT) is the treatment of choice for patients with some malignant and non-malignant hematological diseases..

С другой стороны, проведя анализ, мы можем утверждать, что именно Сибирский текст русской культуры и литературы является той подоплекой,