forked from luck/tmp_suning_uos_patched
a7d2dac802
This updates the 32-bit headers to use the same definitions for the RPN shift inside the PTE as 64-bit, and thus updates _PAGE_CHG_MASK to become identical. This does introduce a runtime visible difference, which is that now, _PAGE_HASHPTE will be part of _PAGE_CHG_MASK and thus preserved. However this should have no practical effect as it should have been preserved in the first place and we got away with not having it there due to our PTE access functions preserving it anyway. Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org>
49 lines
1.6 KiB
C
49 lines
1.6 KiB
C
#ifndef _ASM_POWERPC_PTE_FSL_BOOKE_H
|
|
#define _ASM_POWERPC_PTE_FSL_BOOKE_H
|
|
#ifdef __KERNEL__
|
|
|
|
/* PTE bit definitions for Freescale BookE SW loaded TLB MMU based
|
|
* processors
|
|
*
|
|
MMU Assist Register 3:
|
|
|
|
32 33 34 35 36 ... 50 51 52 53 54 55 56 57 58 59 60 61 62 63
|
|
RPN...................... 0 0 U0 U1 U2 U3 UX SX UW SW UR SR
|
|
|
|
- PRESENT *must* be in the bottom three bits because swap cache
|
|
entries use the top 29 bits.
|
|
|
|
- FILE *must* be in the bottom three bits because swap cache
|
|
entries use the top 29 bits.
|
|
*/
|
|
|
|
/* Definitions for FSL Book-E Cores */
|
|
#define _PAGE_PRESENT 0x00001 /* S: PTE contains a translation */
|
|
#define _PAGE_USER 0x00002 /* S: User page (maps to UR) */
|
|
#define _PAGE_FILE 0x00002 /* S: when !present: nonlinear file mapping */
|
|
#define _PAGE_RW 0x00004 /* S: Write permission (SW) */
|
|
#define _PAGE_DIRTY 0x00008 /* S: Page dirty */
|
|
#define _PAGE_HWEXEC 0x00010 /* H: SX permission */
|
|
#define _PAGE_ACCESSED 0x00020 /* S: Page referenced */
|
|
|
|
#define _PAGE_ENDIAN 0x00040 /* H: E bit */
|
|
#define _PAGE_GUARDED 0x00080 /* H: G bit */
|
|
#define _PAGE_COHERENT 0x00100 /* H: M bit */
|
|
#define _PAGE_NO_CACHE 0x00200 /* H: I bit */
|
|
#define _PAGE_WRITETHRU 0x00400 /* H: W bit */
|
|
#define _PAGE_SPECIAL 0x00800 /* S: Special page */
|
|
|
|
#ifdef CONFIG_PTE_64BIT
|
|
/* ERPN in a PTE never gets cleared, ignore it */
|
|
#define _PTE_NONE_MASK 0xffffffffffff0000ULL
|
|
/* We extend the size of the PTE flags area when using 64-bit PTEs */
|
|
#define PTE_RPN_SHIFT (PAGE_SHIFT + 8)
|
|
#endif
|
|
|
|
#define _PMD_PRESENT 0
|
|
#define _PMD_PRESENT_MASK (PAGE_MASK)
|
|
#define _PMD_BAD (~PAGE_MASK)
|
|
|
|
#endif /* __KERNEL__ */
|
|
#endif /* _ASM_POWERPC_PTE_FSL_BOOKE_H */
|