2 * Copyright 2003 PathScale Inc
3 * Derived from include/asm-i386/pgtable.h
4 * Licensed under the GPL
7 #ifndef __UM_PGTABLE_3LEVEL_H
8 #define __UM_PGTABLE_3LEVEL_H
10 #include <asm-generic/pgtable-nopud.h>
12 /* PGDIR_SHIFT determines what a third-level page table entry can map */
14 #define PGDIR_SHIFT 30
15 #define PGDIR_SIZE (1UL << PGDIR_SHIFT)
16 #define PGDIR_MASK (~(PGDIR_SIZE-1))
18 /* PMD_SHIFT determines the size of the area a second-level page table can
23 #define PMD_SIZE (1UL << PMD_SHIFT)
24 #define PMD_MASK (~(PMD_SIZE-1))
27 * entries per page directory level
30 #define PTRS_PER_PTE 512
31 #define PTRS_PER_PMD 512
32 #define USER_PTRS_PER_PGD ((TASK_SIZE + (PGDIR_SIZE - 1)) / PGDIR_SIZE)
33 #define PTRS_PER_PGD 512
34 #define FIRST_USER_PGD_NR 0
36 #define pte_ERROR(e) \
37 printk("%s:%d: bad pte %p(%016lx).\n", __FILE__, __LINE__, &(e), \
39 #define pmd_ERROR(e) \
40 printk("%s:%d: bad pmd %p(%016lx).\n", __FILE__, __LINE__, &(e), \
42 #define pgd_ERROR(e) \
43 printk("%s:%d: bad pgd %p(%016lx).\n", __FILE__, __LINE__, &(e), \
46 #define pud_none(x) (!(pud_val(x) & ~_PAGE_NEWPAGE))
47 #define pud_bad(x) ((pud_val(x) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE)
48 #define pud_present(x) (pud_val(x) & _PAGE_PRESENT)
49 #define pud_populate(mm, pud, pmd) \
50 set_pud(pud, __pud(_PAGE_TABLE + __pa(pmd)))
52 #define set_pud(pudptr, pudval) set_64bit((phys_t *) (pudptr), pud_val(pudval))
53 static inline int pgd_newpage(pgd_t pgd)
55 return(pgd_val(pgd) & _PAGE_NEWPAGE);
58 static inline void pgd_mkuptodate(pgd_t pgd) { pgd_val(pgd) &= ~_PAGE_NEWPAGE; }
61 #define pte_present(x) pte_get_bits(x, (_PAGE_PRESENT | _PAGE_PROTNONE))
63 static inline pte_t pte_mknewprot(pte_t pte)
65 pte_set_bits(pte, _PAGE_NEWPROT);
69 static inline pte_t pte_mknewpage(pte_t pte)
71 pte_set_bits(pte, _PAGE_NEWPAGE);
75 static inline void set_pte(pte_t *pteptr, pte_t pteval)
77 pte_copy(*pteptr, pteval);
79 /* If it's a swap entry, it needs to be marked _PAGE_NEWPAGE so
80 * fix_range knows to unmap it. _PAGE_NEWPROT is specific to
84 *pteptr = pte_mknewpage(*pteptr);
85 if(pte_present(*pteptr)) *pteptr = pte_mknewprot(*pteptr);
88 #define set_pmd(pmdptr, pmdval) set_64bit((phys_t *) (pmdptr), pmd_val(pmdval))
90 static inline pmd_t *pmd_alloc_one(struct mm_struct *mm, unsigned long address)
92 pmd_t *pmd = (pmd_t *) __get_free_page(GFP_KERNEL);
95 memset(pmd, 0, PAGE_SIZE);
100 static inline void pmd_free(pmd_t *pmd){
101 free_page((unsigned long) pmd);
104 #define __pmd_free_tlb(tlb,x) do { } while (0)
106 static inline void pud_clear (pud_t * pud) { }
108 #define pud_page(pud) \
109 ((struct page *) __va(pud_val(pud) & PAGE_MASK))
111 /* Find an entry in the second-level page table.. */
112 #define pmd_offset(pud, address) ((pmd_t *) pud_page(*(pud)) + \
115 #define pte_page(x) pfn_to_page(pte_pfn(x))
117 static inline int pte_none(pte_t pte)
119 return pte_is_zero(pte);
122 static inline unsigned long pte_pfn(pte_t pte)
124 return phys_to_pfn(pte_val(pte));
127 static inline pte_t pfn_pte(pfn_t page_nr, pgprot_t pgprot)
130 phys_t phys = pfn_to_phys(page_nr);
132 pte_set_val(pte, phys, pgprot);
136 static inline pmd_t pfn_pmd(pfn_t page_nr, pgprot_t pgprot)
138 return __pmd((page_nr << PAGE_SHIFT) | pgprot_val(pgprot));
142 * Bits 0 through 3 are taken in the low part of the pte,
143 * put the 32 bits of offset into the high part.
145 #define PTE_FILE_MAX_BITS 32
149 #define pte_to_pgoff(p) ((p).pte >> 32)
151 #define pgoff_to_pte(off) ((pte_t) { ((off) < 32) | _PAGE_FILE })
155 #define pte_to_pgoff(pte) ((pte).pte_high)
157 #define pgoff_to_pte(off) ((pte_t) { _PAGE_FILE, (off) })
164 * Overrides for Emacs so that we follow Linus's tabbing style.
165 * Emacs will notice this stuff at the end of the file and automatically
166 * adjust the settings for this buffer only. This must remain at the end
168 * ---------------------------------------------------------------------------
170 * c-file-style: "linux"