forked from luck/tmp_suning_uos_patched
f84c914b98
Signed-off-by: Kirill A. Shutemov <kirill.shutemov@linux.intel.com> Cc: Geert Uytterhoeven <geert@linux-m68k.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
107 lines
2.3 KiB
C
107 lines
2.3 KiB
C
#ifndef M68K_MCF_PGALLOC_H
|
|
#define M68K_MCF_PGALLOC_H
|
|
|
|
#include <asm/tlb.h>
|
|
#include <asm/tlbflush.h>
|
|
|
|
extern inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
|
|
{
|
|
free_page((unsigned long) pte);
|
|
}
|
|
|
|
extern const char bad_pmd_string[];
|
|
|
|
extern inline pte_t *pte_alloc_one_kernel(struct mm_struct *mm,
|
|
unsigned long address)
|
|
{
|
|
unsigned long page = __get_free_page(GFP_DMA|__GFP_REPEAT);
|
|
|
|
if (!page)
|
|
return NULL;
|
|
|
|
memset((void *)page, 0, PAGE_SIZE);
|
|
return (pte_t *) (page);
|
|
}
|
|
|
|
extern inline pmd_t *pmd_alloc_kernel(pgd_t *pgd, unsigned long address)
|
|
{
|
|
return (pmd_t *) pgd;
|
|
}
|
|
|
|
#define pmd_alloc_one_fast(mm, address) ({ BUG(); ((pmd_t *)1); })
|
|
#define pmd_alloc_one(mm, address) ({ BUG(); ((pmd_t *)2); })
|
|
|
|
#define pte_alloc_one_fast(mm, addr) pte_alloc_one(mm, addr)
|
|
|
|
#define pmd_populate(mm, pmd, page) (pmd_val(*pmd) = \
|
|
(unsigned long)(page_address(page)))
|
|
|
|
#define pmd_populate_kernel(mm, pmd, pte) (pmd_val(*pmd) = (unsigned long)(pte))
|
|
|
|
#define pmd_pgtable(pmd) pmd_page(pmd)
|
|
|
|
static inline void __pte_free_tlb(struct mmu_gather *tlb, pgtable_t page,
|
|
unsigned long address)
|
|
{
|
|
__free_page(page);
|
|
}
|
|
|
|
#define __pmd_free_tlb(tlb, pmd, address) do { } while (0)
|
|
|
|
static inline struct page *pte_alloc_one(struct mm_struct *mm,
|
|
unsigned long address)
|
|
{
|
|
struct page *page = alloc_pages(GFP_DMA|__GFP_REPEAT, 0);
|
|
pte_t *pte;
|
|
|
|
if (!page)
|
|
return NULL;
|
|
if (!pgtable_page_ctor(page)) {
|
|
__free_page(page);
|
|
return NULL;
|
|
}
|
|
|
|
pte = kmap(page);
|
|
if (pte) {
|
|
clear_page(pte);
|
|
__flush_page_to_ram(pte);
|
|
flush_tlb_kernel_page(pte);
|
|
nocache_page(pte);
|
|
}
|
|
kunmap(page);
|
|
|
|
return page;
|
|
}
|
|
|
|
extern inline void pte_free(struct mm_struct *mm, struct page *page)
|
|
{
|
|
__free_page(page);
|
|
}
|
|
|
|
/*
|
|
* In our implementation, each pgd entry contains 1 pmd that is never allocated
|
|
* or freed. pgd_present is always 1, so this should never be called. -NL
|
|
*/
|
|
#define pmd_free(mm, pmd) BUG()
|
|
|
|
static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd)
|
|
{
|
|
free_page((unsigned long) pgd);
|
|
}
|
|
|
|
static inline pgd_t *pgd_alloc(struct mm_struct *mm)
|
|
{
|
|
pgd_t *new_pgd;
|
|
|
|
new_pgd = (pgd_t *)__get_free_page(GFP_DMA | __GFP_NOWARN);
|
|
if (!new_pgd)
|
|
return NULL;
|
|
memcpy(new_pgd, swapper_pg_dir, PAGE_SIZE);
|
|
memset(new_pgd, 0, PAGE_OFFSET >> PGDIR_SHIFT);
|
|
return new_pgd;
|
|
}
|
|
|
|
#define pgd_populate(mm, pmd, pte) BUG()
|
|
|
|
#endif /* M68K_MCF_PGALLOC_H */
|