5e5419734c
(with Martin Schwidefsky <schwidefsky@de.ibm.com>) The pgd/pud/pmd/pte page table allocation functions get a mm_struct pointer as first argument. The free functions do not get the mm_struct argument. This is 1) asymmetrical and 2) to do mm related page table allocations the mm argument is needed on the free function as well. [kamalesh@linux.vnet.ibm.com: i386 fix] [akpm@linux-foundation.org: coding-syle fixes] Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org> Signed-off-by: Martin Schwidefsky <schwidefsky@de.ibm.com> Cc: <linux-arch@vger.kernel.org> Signed-off-by: Kamalesh Babulal <kamalesh@linux.vnet.ibm.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
67 lines
1.6 KiB
C
67 lines
1.6 KiB
C
/*
|
|
* include/asm-xtensa/pgalloc.h
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* Copyright (C) 2001-2007 Tensilica Inc.
|
|
*/
|
|
|
|
#ifndef _XTENSA_PGALLOC_H
|
|
#define _XTENSA_PGALLOC_H
|
|
|
|
#ifdef __KERNEL__
|
|
|
|
#include <linux/highmem.h>
|
|
|
|
/*
|
|
* Allocating and freeing a pmd is trivial: the 1-entry pmd is
|
|
* inside the pgd, so has no extra memory associated with it.
|
|
*/
|
|
|
|
#define pmd_populate_kernel(mm, pmdp, ptep) \
|
|
(pmd_val(*(pmdp)) = ((unsigned long)ptep))
|
|
#define pmd_populate(mm, pmdp, page) \
|
|
(pmd_val(*(pmdp)) = ((unsigned long)page_to_virt(page)))
|
|
|
|
static inline pgd_t*
|
|
pgd_alloc(struct mm_struct *mm)
|
|
{
|
|
return (pgd_t*) __get_free_pages(GFP_KERNEL | __GFP_ZERO, PGD_ORDER);
|
|
}
|
|
|
|
static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd)
|
|
{
|
|
free_page((unsigned long)pgd);
|
|
}
|
|
|
|
/* Use a slab cache for the pte pages (see also sparc64 implementation) */
|
|
|
|
extern struct kmem_cache *pgtable_cache;
|
|
|
|
static inline pte_t *pte_alloc_one_kernel(struct mm_struct *mm,
|
|
unsigned long address)
|
|
{
|
|
return kmem_cache_alloc(pgtable_cache, GFP_KERNEL|__GFP_REPEAT);
|
|
}
|
|
|
|
static inline struct page *pte_alloc_one(struct mm_struct *mm,
|
|
unsigned long addr)
|
|
{
|
|
return virt_to_page(pte_alloc_one_kernel(mm, addr));
|
|
}
|
|
|
|
static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
|
|
{
|
|
kmem_cache_free(pgtable_cache, pte);
|
|
}
|
|
|
|
static inline void pte_free(struct mm_struct *mm, struct page *page)
|
|
{
|
|
kmem_cache_free(pgtable_cache, page_address(page));
|
|
}
|
|
|
|
#endif /* __KERNEL__ */
|
|
#endif /* _XTENSA_PGALLOC_H */
|