提交 849f86a6 编写于 作者: P Paul Mackerras 提交者: Michael Ellerman

powerpc/mm/book3s-64: Move _PAGE_PRESENT to the most significant bit

This changes _PAGE_PRESENT for 64-bit Book 3S processors from 0x2 to
0x8000_0000_0000_0000, because that is where PowerISA v3.0 CPUs in
radix mode will expect to find it.
Signed-off-by: NPaul Mackerras <paulus@samba.org>
Signed-off-by: NMichael Ellerman <mpe@ellerman.id.au>
上级 c61a8843
...@@ -210,30 +210,30 @@ static inline char *get_hpte_slot_array(pmd_t *pmdp) ...@@ -210,30 +210,30 @@ static inline char *get_hpte_slot_array(pmd_t *pmdp)
/* /*
* The linux hugepage PMD now include the pmd entries followed by the address * The linux hugepage PMD now include the pmd entries followed by the address
* to the stashed pgtable_t. The stashed pgtable_t contains the hpte bits. * to the stashed pgtable_t. The stashed pgtable_t contains the hpte bits.
* [ 1 bit secondary | 3 bit hidx | 1 bit valid | 000]. We use one byte per * [ 000 | 1 bit secondary | 3 bit hidx | 1 bit valid]. We use one byte per
* each HPTE entry. With 16MB hugepage and 64K HPTE we need 256 entries and * each HPTE entry. With 16MB hugepage and 64K HPTE we need 256 entries and
* with 4K HPTE we need 4096 entries. Both will fit in a 4K pgtable_t. * with 4K HPTE we need 4096 entries. Both will fit in a 4K pgtable_t.
* *
* The last three bits are intentionally left to zero. This memory location * The top three bits are intentionally left as zero. This memory location
* are also used as normal page PTE pointers. So if we have any pointers * are also used as normal page PTE pointers. So if we have any pointers
* left around while we collapse a hugepage, we need to make sure * left around while we collapse a hugepage, we need to make sure
* _PAGE_PRESENT bit of that is zero when we look at them * _PAGE_PRESENT bit of that is zero when we look at them
*/ */
static inline unsigned int hpte_valid(unsigned char *hpte_slot_array, int index) static inline unsigned int hpte_valid(unsigned char *hpte_slot_array, int index)
{ {
return (hpte_slot_array[index] >> 3) & 0x1; return hpte_slot_array[index] & 0x1;
} }
static inline unsigned int hpte_hash_index(unsigned char *hpte_slot_array, static inline unsigned int hpte_hash_index(unsigned char *hpte_slot_array,
int index) int index)
{ {
return hpte_slot_array[index] >> 4; return hpte_slot_array[index] >> 1;
} }
static inline void mark_hpte_slot_valid(unsigned char *hpte_slot_array, static inline void mark_hpte_slot_valid(unsigned char *hpte_slot_array,
unsigned int index, unsigned int hidx) unsigned int index, unsigned int hidx)
{ {
hpte_slot_array[index] = hidx << 4 | 0x1 << 3; hpte_slot_array[index] = (hidx << 1) | 0x1;
} }
/* /*
......
...@@ -14,7 +14,6 @@ ...@@ -14,7 +14,6 @@
* combinations that newer processors provide but we currently don't. * combinations that newer processors provide but we currently don't.
*/ */
#define _PAGE_PTE 0x00001 /* distinguishes PTEs from pointers */ #define _PAGE_PTE 0x00001 /* distinguishes PTEs from pointers */
#define _PAGE_PRESENT 0x00002 /* software: pte contains a translation */
#define _PAGE_BIT_SWAP_TYPE 2 #define _PAGE_BIT_SWAP_TYPE 2
#define _PAGE_USER 0x00004 /* page may be accessed by userspace */ #define _PAGE_USER 0x00004 /* page may be accessed by userspace */
#define _PAGE_EXEC 0x00008 /* execute permission */ #define _PAGE_EXEC 0x00008 /* execute permission */
...@@ -39,6 +38,8 @@ ...@@ -39,6 +38,8 @@
#define _PAGE_SOFT_DIRTY 0x00000 #define _PAGE_SOFT_DIRTY 0x00000
#endif #endif
#define _PAGE_PRESENT (1ul << 63) /* pte contains a translation */
/* /*
* We need to differentiate between explicit huge page and THP huge * We need to differentiate between explicit huge page and THP huge
* page, since THP huge page also need to track real subpage details * page, since THP huge page also need to track real subpage details
...@@ -402,7 +403,7 @@ static inline int pte_protnone(pte_t pte) ...@@ -402,7 +403,7 @@ static inline int pte_protnone(pte_t pte)
static inline int pte_present(pte_t pte) static inline int pte_present(pte_t pte)
{ {
return pte_val(pte) & _PAGE_PRESENT; return !!(pte_val(pte) & _PAGE_PRESENT);
} }
/* Conversion functions: convert a page and protection to a page entry, /* Conversion functions: convert a page and protection to a page entry,
......
...@@ -110,7 +110,8 @@ extern unsigned long Hash_size, Hash_mask; ...@@ -110,7 +110,8 @@ extern unsigned long Hash_size, Hash_mask;
#endif /* CONFIG_PPC32 */ #endif /* CONFIG_PPC32 */
#ifdef CONFIG_PPC64 #ifdef CONFIG_PPC64
extern int map_kernel_page(unsigned long ea, unsigned long pa, int flags); extern int map_kernel_page(unsigned long ea, unsigned long pa,
unsigned long flags);
#endif /* CONFIG_PPC64 */ #endif /* CONFIG_PPC64 */
extern unsigned long ioremap_bot; extern unsigned long ioremap_bot;
......
...@@ -88,7 +88,7 @@ static __ref void *early_alloc_pgtable(unsigned long size) ...@@ -88,7 +88,7 @@ static __ref void *early_alloc_pgtable(unsigned long size)
* map_kernel_page adds an entry to the ioremap page table * map_kernel_page adds an entry to the ioremap page table
* and adds an entry to the HPT, possibly bolting it * and adds an entry to the HPT, possibly bolting it
*/ */
int map_kernel_page(unsigned long ea, unsigned long pa, int flags) int map_kernel_page(unsigned long ea, unsigned long pa, unsigned long flags)
{ {
pgd_t *pgdp; pgd_t *pgdp;
pud_t *pudp; pud_t *pudp;
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册