2005-04-17 00:20:36 +02:00
|
|
|
#ifndef _PGTABLE_NOPMD_H
|
|
|
|
#define _PGTABLE_NOPMD_H
|
|
|
|
|
|
|
|
#ifndef __ASSEMBLY__
|
|
|
|
|
|
|
|
#include <asm-generic/pgtable-nopud.h>
|
|
|
|
|
2008-07-29 00:46:40 +02:00
|
|
|
struct mm_struct;
|
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
#define __PAGETABLE_PMD_FOLDED
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Having the pmd type consist of a pud gets the size right, and allows
|
|
|
|
* us to conceptually access the pud entry that this pmd is folded into
|
|
|
|
* without casting.
|
|
|
|
*/
|
|
|
|
typedef struct { pud_t pud; } pmd_t;
|
|
|
|
|
|
|
|
#define PMD_SHIFT PUD_SHIFT
|
|
|
|
#define PTRS_PER_PMD 1
|
|
|
|
#define PMD_SIZE (1UL << PMD_SHIFT)
|
|
|
|
#define PMD_MASK (~(PMD_SIZE-1))
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The "pud_xxx()" functions here are trivial for a folded two-level
|
|
|
|
* setup: the pmd is never bad, and a pmd always exists (as it's folded
|
|
|
|
* into the pud entry)
|
|
|
|
*/
|
|
|
|
static inline int pud_none(pud_t pud) { return 0; }
|
|
|
|
static inline int pud_bad(pud_t pud) { return 0; }
|
|
|
|
static inline int pud_present(pud_t pud) { return 1; }
|
|
|
|
static inline void pud_clear(pud_t *pud) { }
|
|
|
|
#define pmd_ERROR(pmd) (pud_ERROR((pmd).pud))
|
|
|
|
|
|
|
|
#define pud_populate(mm, pmd, pte) do { } while (0)
|
|
|
|
|
|
|
|
/*
|
|
|
|
* (pmds are folded into puds so this doesn't get actually called,
|
|
|
|
* but the define is needed for a generic inline function.)
|
|
|
|
*/
|
|
|
|
#define set_pud(pudptr, pudval) set_pmd((pmd_t *)(pudptr), (pmd_t) { pudval })
|
|
|
|
|
|
|
|
static inline pmd_t * pmd_offset(pud_t * pud, unsigned long address)
|
|
|
|
{
|
|
|
|
return (pmd_t *)pud;
|
|
|
|
}
|
|
|
|
|
|
|
|
#define pmd_val(x) (pud_val((x).pud))
|
|
|
|
#define __pmd(x) ((pmd_t) { __pud(x) } )
|
|
|
|
|
|
|
|
#define pud_page(pud) (pmd_page((pmd_t){ pud }))
|
2006-09-26 08:31:48 +02:00
|
|
|
#define pud_page_vaddr(pud) (pmd_page_vaddr((pmd_t){ pud }))
|
2005-04-17 00:20:36 +02:00
|
|
|
|
|
|
|
/*
|
|
|
|
* allocating and freeing a pmd is trivial: the 1-entry pmd is
|
|
|
|
* inside the pud, so has no extra memory associated with it.
|
|
|
|
*/
|
|
|
|
#define pmd_alloc_one(mm, address) NULL
|
2008-07-29 00:46:40 +02:00
|
|
|
static inline void pmd_free(struct mm_struct *mm, pmd_t *pmd)
|
|
|
|
{
|
|
|
|
}
|
mm: Pass virtual address to [__]p{te,ud,md}_free_tlb()
mm: Pass virtual address to [__]p{te,ud,md}_free_tlb()
Upcoming paches to support the new 64-bit "BookE" powerpc architecture
will need to have the virtual address corresponding to PTE page when
freeing it, due to the way the HW table walker works.
Basically, the TLB can be loaded with "large" pages that cover the whole
virtual space (well, sort-of, half of it actually) represented by a PTE
page, and which contain an "indirect" bit indicating that this TLB entry
RPN points to an array of PTEs from which the TLB can then create direct
entries. Thus, in order to invalidate those when PTE pages are deleted,
we need the virtual address to pass to tlbilx or tlbivax instructions.
The old trick of sticking it somewhere in the PTE page struct page sucks
too much, the address is almost readily available in all call sites and
almost everybody implemets these as macros, so we may as well add the
argument everywhere. I added it to the pmd and pud variants for consistency.
Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Acked-by: David Howells <dhowells@redhat.com> [MN10300 & FRV]
Acked-by: Nick Piggin <npiggin@suse.de>
Acked-by: Martin Schwidefsky <schwidefsky@de.ibm.com> [s390]
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-07-22 07:44:28 +02:00
|
|
|
#define __pmd_free_tlb(tlb, x, a) do { } while (0)
|
2005-04-17 00:20:36 +02:00
|
|
|
|
|
|
|
#undef pmd_addr_end
|
|
|
|
#define pmd_addr_end(addr, end) (end)
|
|
|
|
|
|
|
|
#endif /* __ASSEMBLY__ */
|
|
|
|
|
|
|
|
#endif /* _PGTABLE_NOPMD_H */
|