The role of wrapper ﹣ PG ﹣ dir

During the initialization of the memory system, there are the following codes: ...

During the initialization of the memory system, there are the following codes:

1: static void __init pagetable_init(void) 2: { 3: pgd_t *pgd_base = swapper_pg_dir; 4: 5: permanent_kmaps_init(pgd_base); 6: }

Here, we see the mysterious swapper ﹣ PG ﹣ dir. After a global search, we find

1: /* 2: * Build a proper pagetable for the kernel mappings. Up until this 3: * point, we've been running on some set of pagetables constructed by 4: * the boot process. 5: * 6: * If we're booting on native hardware, this will be a pagetable 7: * constructed in arch/x86/kernel/head_32.S. The root of the 8: * pagetable will be swapper_pg_dir. 9: * 10: * If we're booting paravirtualized under a hypervisor, then there are 11: * more options: we may already be running PAE, and the pagetable may 12: * or may not be based in swapper_pg_dir. In any case, 13: * paravirt_pagetable_setup_start() will set up swapper_pg_dir 14: * appropriately for the rest of the initialization to work. 15: * 16: * In general, pagetable_init() assumes that the pagetable may already 17: * be partially populated, and so it avoids stomping on any existing 18: * mappings. 19: */ 20: void __init early_ioremap_page_table_range_init(void) 21: { 22: pgd_t *pgd_base = swapper_pg_dir; 23: unsigned long vaddr, end; 24: 25: /* 26: * Fixed mappings, only the page table structure has to be 27: * created - mappings will be set by set_fixmap(): 28: */ 29: vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK; 30: end = (FIXADDR_TOP + PMD_SIZE - 1) & PMD_MASK; 31: page_table_range_init(vaddr, end, pgd_base); 32: early_ioremap_reset(); 33: }

In head_32.S, the following BSS segments are defined. The BSS segment does not occupy space in the kernel image file, but when the kernel is loaded into memory, the corresponding space will be reserved.

In BSS segment, four pages are reserved, and their addresses are marked with initial page table, initial PG fixmap, empty zero page and wrapper PG dir respectively.

1: /* 2: * BSS section 3: */ 4: __PAGE_ALIGNED_BSS 5: .align PAGE_SIZE 6: #ifdef CONFIG_X86_PAE 7: initial_pg_pmd: 8: .fill 1024*KPMDS,4,0 9: #else 10: ENTRY(initial_page_table) 11: .fill 1024,4,0 12: #endif 13: initial_pg_fixmap: 14: .fill 1024,4,0 15: ENTRY(empty_zero_page) 16: .fill 4096,1,0 17: ENTRY(swapper_pg_dir) 18: .fill 1024,4,0

Set the initial page table to the initial page directory with the following code

1: /* 2: * Enable paging 3: */ 4: movl $pa(initial_page_table), %eax 5: movl %eax,%cr3 /* set the page table pointer.. */ 6: movl %cr0,%eax 7: orl $X86_CR0_PG,%eax 8: movl %eax,%cr0 /* ..and set paging (PG) bit */ 9: ljmp $__BOOT_CS,$1f /* Clear prefetch and normalize %eip */

In the kernel initialization phase, setup arch calls the following functions:

1: void __init setup_arch(char **cmdline_p) 2: { 3: 4: ...... 5: /* max_pfn_mapped is updated here */ 6: max_low_pfn_mapped = init_memory_mapping(0, max_low_pfn<<PAGE_SHIFT); 7: max_pfn_mapped = max_low_pfn_mapped; 8: ...... 9: x86_init.paging.pagetable_setup_start(swapper_pg_dir); 10: paging_init(); 11: x86_init.paging.pagetable_setup_done(swapper_pg_dir); 12: 13: ...... 14: }

Init? Memory? Mapping calls kernel? Physical? Mapping? Init, initializes the wrapper? PG? Dir

1: /* 2: * This maps the physical memory to kernel virtual address space, a total 3: * of max_low_pfn pages, by creating page tables starting from address 4: * PAGE_OFFSET: 5: */ 6: unsigned long __init 7: kernel_physical_mapping_init(unsigned long start, 8: unsigned long end, 9: unsigned long page_size_mask) 10: { 11: int use_pse = page_size_mask == (1<<PG_LEVEL_2M); 12: unsigned long last_map_addr = end; 13: unsigned long start_pfn, end_pfn; 14: pgd_t *pgd_base = swapper_pg_dir; 15: int pgd_idx, pmd_idx, pte_ofs; 16: unsigned long pfn; 17: pgd_t *pgd; 18: pmd_t *pmd; 19: pte_t *pte; 20: unsigned pages_2m, pages_4k; 21: int mapping_iter; 22: 23: start_pfn = start >> PAGE_SHIFT; 24: end_pfn = end >> PAGE_SHIFT; 25: 26: /* 27: * First iteration will setup identity mapping using large/small pages 28: * based on use_pse, with other attributes same as set by 29: * the early code in head_32.S 30: * 31: * Second iteration will setup the appropriate attributes (NX, GLOBAL..) 32: * as desired for the kernel identity mapping. 33: * 34: * This two pass mechanism conforms to the TLB app note which says: 35: * 36: * "Software should not write to a paging-structure entry in a way 37: * that would change, for any linear address, both the page size 38: * and either the page frame or attributes." 39: */ 40: mapping_iter = 1; 41: 42: if (!cpu_has_pse) 43: use_pse = 0; 44: 45: at: 46: pages_2m = pages_4k = 0; 47: pfn = start_pfn; 48: pgd_idx = pgd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET); 49: pgd = pgd_base + pgd_idx; 50: for (; pgd_idx < PTRS_PER_PGD; pgd++, pgd_idx++) { 51: pmd = one_md_table_init(pgd); 52: 53: if (pfn >= end_pfn) 54: continue; 55: ef CONFIG_X86_PAE 56: pmd_idx = pmd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET); 57: pmd += pmd_idx; 58: e 59: pmd_idx = 0; 60: if 61: for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn; 62: pmd++, pmd_idx++) { 63: unsigned int addr = pfn * PAGE_SIZE + PAGE_OFFSET; 64: 65: /* 66: * Map with big pages if possible, otherwise 67: * create normal page tables: 68: */ 69: if (use_pse) { 70: unsigned int addr2; 71: pgprot_t prot = PAGE_KERNEL_LARGE; 72: /* 73: * first pass will use the same initial 74: * identity mapping attribute + _PAGE_PSE. 75: */ 76: pgprot_t init_prot = 77: __pgprot(PTE_IDENT_ATTR | 78: _PAGE_PSE); 79: 80: addr2 = (pfn + PTRS_PER_PTE-1) * PAGE_SIZE + 81: PAGE_OFFSET + PAGE_SIZE-1; 82: 83: if (is_kernel_text(addr) || 84: is_kernel_text(addr2)) 85: prot = PAGE_KERNEL_LARGE_EXEC; 86: 87: pages_2m++; 88: if (mapping_iter == 1) 89: set_pmd(pmd, pfn_pmd(pfn, init_prot)); 90: else 91: set_pmd(pmd, pfn_pmd(pfn, prot)); 92: 93: pfn += PTRS_PER_PTE; 94: continue; 95: } 96: pte = one_page_table_init(pmd); 97: 98: pte_ofs = pte_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET); 99: pte += pte_ofs; 100: for (; pte_ofs < PTRS_PER_PTE && pfn < end_pfn; 101: pte++, pfn++, pte_ofs++, addr += PAGE_SIZE) { 102: pgprot_t prot = PAGE_KERNEL; 103: /* 104: * first pass will use the same initial 105: * identity mapping attribute. 106: */ 107: pgprot_t init_prot = __pgprot(PTE_IDENT_ATTR); 108: 109: if (is_kernel_text(addr)) 110: prot = PAGE_KERNEL_EXEC; 111: 112: pages_4k++; 113: if (mapping_iter == 1) { 114: set_pte(pte, pfn_pte(pfn, init_prot)); 115: last_map_addr = (pfn << PAGE_SHIFT) + PAGE_SIZE; 116: } else 117: set_pte(pte, pfn_pte(pfn, prot)); 118: } 119: } 120: } 121: if (mapping_iter == 1) { 122: /* 123: * update direct mapping page count only in the first 124: * iteration. 125: */ 126: update_page_count(PG_LEVEL_2M, pages_2m); 127: update_page_count(PG_LEVEL_4K, pages_4k); 128: 129: /* 130: * local global flush tlb, which will flush the previous 131: * mappings present in both small and large page TLB's. 132: */ 133: __flush_tlb_all(); 134: 135: /* 136: * Second iteration will set the actual desired PTE attributes. 137: */ 138: mapping_iter = 2; 139: goto repeat; 140: } 141: return last_map_addr; 142: }

pgd_t *pgd_base = swapper_pg_dir;

Assign the wrapper? PG? Dir as the page directory address to PGD? Base

start_pfn = start >> PAGE_SHIFT;

end_pfn = end >> PAGE_SHIFT;

pgd_idx = pgd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);

pgd = pgd_base + pgd_idx;

PGD? IDX, PGD represents the index in the page directory and the corresponding page directory entry

1: pgprot_t prot = PAGE_KERNEL; 2: /* 3: * first pass will use the same initial 4: * identity mapping attribute. 5: */ 6: pgprot_t init_prot = __pgprot(PTE_IDENT_ATTR); 7: 8: if (is_kernel_text(addr)) 9: prot = PAGE_KERNEL_EXEC; 10: 11: pages_4k++; 12: if (mapping_iter == 1) { 13: set_pte(pte, pfn_pte(pfn, init_prot)); 14: last_map_addr = (pfn << PAGE_SHIFT) + PAGE_SIZE; 15: } else 16: set_pte(pte, pfn_pte(pfn, prot));

Finally, through two rounds of traversal, the property is set to the corresponding page table item.

5 May 2020, 19:31 | Views: 1530

Add new comment

For adding a comment, please log in
or create account

0 comments