1 #ifndef _X86_64_PGALLOC_H
2 #define _X86_64_PGALLOC_H
4 #include <asm/fixmap.h>
6 #include <linux/threads.h>
9 #define pmd_populate_kernel(mm, pmd, pte) \
10 set_pmd(pmd, __pmd(_PAGE_TABLE | __pa(pte)))
11 #define pud_populate(mm, pud, pmd) \
12 set_pud(pud, __pud(_PAGE_TABLE | __pa(pmd)))
13 #define pgd_populate(mm, pgd, pud) \
14 set_pgd(pgd, __pgd(_PAGE_TABLE | __pa(pud)))
16 static inline void pmd_populate(struct mm_struct
*mm
, pmd_t
*pmd
, struct page
*pte
)
18 set_pmd(pmd
, __pmd(_PAGE_TABLE
| (page_to_pfn(pte
) << PAGE_SHIFT
)));
21 static inline pmd_t
*get_pmd(void)
23 return (pmd_t
*)get_zeroed_page(GFP_KERNEL
);
26 static inline void pmd_free(pmd_t
*pmd
)
28 BUG_ON((unsigned long)pmd
& (PAGE_SIZE
-1));
29 free_page((unsigned long)pmd
);
32 static inline pmd_t
*pmd_alloc_one (struct mm_struct
*mm
, unsigned long addr
)
34 return (pmd_t
*)get_zeroed_page(GFP_KERNEL
|__GFP_REPEAT
);
37 static inline pud_t
*pud_alloc_one(struct mm_struct
*mm
, unsigned long addr
)
39 return (pud_t
*)get_zeroed_page(GFP_KERNEL
|__GFP_REPEAT
);
42 static inline void pud_free (pud_t
*pud
)
44 BUG_ON((unsigned long)pud
& (PAGE_SIZE
-1));
45 free_page((unsigned long)pud
);
48 static inline void pgd_list_add(pgd_t
*pgd
)
50 struct page
*page
= virt_to_page(pgd
);
53 page
->index
= (pgoff_t
)pgd_list
;
55 pgd_list
->private = (unsigned long)&page
->index
;
57 page
->private = (unsigned long)&pgd_list
;
58 spin_unlock(&pgd_lock
);
61 static inline void pgd_list_del(pgd_t
*pgd
)
63 struct page
*next
, **pprev
, *page
= virt_to_page(pgd
);
66 next
= (struct page
*)page
->index
;
67 pprev
= (struct page
**)page
->private;
70 next
->private = (unsigned long)pprev
;
71 spin_unlock(&pgd_lock
);
74 static inline pgd_t
*pgd_alloc(struct mm_struct
*mm
)
77 pgd_t
*pgd
= (pgd_t
*)__get_free_page(GFP_KERNEL
|__GFP_REPEAT
);
82 * Copy kernel pointers in from init.
83 * Could keep a freelist or slab cache of those because the kernel
86 boundary
= pgd_index(__PAGE_OFFSET
);
87 memset(pgd
, 0, boundary
* sizeof(pgd_t
));
88 memcpy(pgd
+ boundary
,
89 init_level4_pgt
+ boundary
,
90 (PTRS_PER_PGD
- boundary
) * sizeof(pgd_t
));
94 static inline void pgd_free(pgd_t
*pgd
)
96 BUG_ON((unsigned long)pgd
& (PAGE_SIZE
-1));
98 free_page((unsigned long)pgd
);
101 static inline pte_t
*pte_alloc_one_kernel(struct mm_struct
*mm
, unsigned long address
)
103 return (pte_t
*)get_zeroed_page(GFP_KERNEL
|__GFP_REPEAT
);
106 static inline struct page
*pte_alloc_one(struct mm_struct
*mm
, unsigned long address
)
108 void *p
= (void *)get_zeroed_page(GFP_KERNEL
|__GFP_REPEAT
);
111 return virt_to_page(p
);
114 /* Should really implement gc for free page table pages. This could be
115 done with a reference count in struct page. */
117 static inline void pte_free_kernel(pte_t
*pte
)
119 BUG_ON((unsigned long)pte
& (PAGE_SIZE
-1));
120 free_page((unsigned long)pte
);
123 static inline void pte_free(struct page
*pte
)
128 #define __pte_free_tlb(tlb,pte) tlb_remove_page((tlb),(pte))
130 #define __pmd_free_tlb(tlb,x) tlb_remove_page((tlb),virt_to_page(x))
131 #define __pud_free_tlb(tlb,x) tlb_remove_page((tlb),virt_to_page(x))
133 #endif /* _X86_64_PGALLOC_H */