提交 f8639939 编写于 作者: E Eduardo Habkost 提交者: Ingo Molnar

x86, paravirt_ops: use unsigned long instead of u32 for alloc_p*() pfn args

This patch changes the pfn args from 'u32' to 'unsigned long'
on alloc_p*() functions on paravirt_ops, and the corresponding
implementations for Xen and VMI. The prototypes for CONFIG_PARAVIRT=n
are already using unsigned long, so paravirt.h now matches the prototypes
on asm-x86/pgalloc.h.

It shouldn't result in any changes on generated code on 32-bit, with
or without CONFIG_PARAVIRT. On both cases, 'codiff -f' didn't show any
change after applying this patch.

On 64-bit, there are (expected) binary changes only when CONFIG_PARAVIRT
is enabled, as the patch is really supposed to change the size of the
pfn args.

[ v2: KVM_GUEST: use the right parameter type on kvm_release_pt() ]
Signed-off-by: NEduardo Habkost <ehabkost@redhat.com>
Acked-by: NJeremy Fitzhardinge <jeremy@goop.org>
Acked-by: NZachary Amsden <zach@vmware.com>
Signed-off-by: NIngo Molnar <mingo@elte.hu>
上级 6e86841d
...@@ -178,7 +178,7 @@ static void kvm_flush_tlb(void) ...@@ -178,7 +178,7 @@ static void kvm_flush_tlb(void)
kvm_deferred_mmu_op(&ftlb, sizeof ftlb); kvm_deferred_mmu_op(&ftlb, sizeof ftlb);
} }
static void kvm_release_pt(u32 pfn) static void kvm_release_pt(unsigned long pfn)
{ {
struct kvm_mmu_op_release_pt rpt = { struct kvm_mmu_op_release_pt rpt = {
.header.op = KVM_MMU_OP_RELEASE_PT, .header.op = KVM_MMU_OP_RELEASE_PT,
......
...@@ -392,13 +392,13 @@ static void *vmi_kmap_atomic_pte(struct page *page, enum km_type type) ...@@ -392,13 +392,13 @@ static void *vmi_kmap_atomic_pte(struct page *page, enum km_type type)
} }
#endif #endif
static void vmi_allocate_pte(struct mm_struct *mm, u32 pfn) static void vmi_allocate_pte(struct mm_struct *mm, unsigned long pfn)
{ {
vmi_set_page_type(pfn, VMI_PAGE_L1); vmi_set_page_type(pfn, VMI_PAGE_L1);
vmi_ops.allocate_page(pfn, VMI_PAGE_L1, 0, 0, 0); vmi_ops.allocate_page(pfn, VMI_PAGE_L1, 0, 0, 0);
} }
static void vmi_allocate_pmd(struct mm_struct *mm, u32 pfn) static void vmi_allocate_pmd(struct mm_struct *mm, unsigned long pfn)
{ {
/* /*
* This call comes in very early, before mem_map is setup. * This call comes in very early, before mem_map is setup.
...@@ -409,20 +409,20 @@ static void vmi_allocate_pmd(struct mm_struct *mm, u32 pfn) ...@@ -409,20 +409,20 @@ static void vmi_allocate_pmd(struct mm_struct *mm, u32 pfn)
vmi_ops.allocate_page(pfn, VMI_PAGE_L2, 0, 0, 0); vmi_ops.allocate_page(pfn, VMI_PAGE_L2, 0, 0, 0);
} }
static void vmi_allocate_pmd_clone(u32 pfn, u32 clonepfn, u32 start, u32 count) static void vmi_allocate_pmd_clone(unsigned long pfn, unsigned long clonepfn, unsigned long start, unsigned long count)
{ {
vmi_set_page_type(pfn, VMI_PAGE_L2 | VMI_PAGE_CLONE); vmi_set_page_type(pfn, VMI_PAGE_L2 | VMI_PAGE_CLONE);
vmi_check_page_type(clonepfn, VMI_PAGE_L2); vmi_check_page_type(clonepfn, VMI_PAGE_L2);
vmi_ops.allocate_page(pfn, VMI_PAGE_L2 | VMI_PAGE_CLONE, clonepfn, start, count); vmi_ops.allocate_page(pfn, VMI_PAGE_L2 | VMI_PAGE_CLONE, clonepfn, start, count);
} }
static void vmi_release_pte(u32 pfn) static void vmi_release_pte(unsigned long pfn)
{ {
vmi_ops.release_page(pfn, VMI_PAGE_L1); vmi_ops.release_page(pfn, VMI_PAGE_L1);
vmi_set_page_type(pfn, VMI_PAGE_NORMAL); vmi_set_page_type(pfn, VMI_PAGE_NORMAL);
} }
static void vmi_release_pmd(u32 pfn) static void vmi_release_pmd(unsigned long pfn)
{ {
vmi_ops.release_page(pfn, VMI_PAGE_L2); vmi_ops.release_page(pfn, VMI_PAGE_L2);
vmi_set_page_type(pfn, VMI_PAGE_NORMAL); vmi_set_page_type(pfn, VMI_PAGE_NORMAL);
......
...@@ -812,7 +812,7 @@ static int xen_write_msr_safe(unsigned int msr, unsigned low, unsigned high) ...@@ -812,7 +812,7 @@ static int xen_write_msr_safe(unsigned int msr, unsigned low, unsigned high)
/* Early in boot, while setting up the initial pagetable, assume /* Early in boot, while setting up the initial pagetable, assume
everything is pinned. */ everything is pinned. */
static __init void xen_alloc_pte_init(struct mm_struct *mm, u32 pfn) static __init void xen_alloc_pte_init(struct mm_struct *mm, unsigned long pfn)
{ {
#ifdef CONFIG_FLATMEM #ifdef CONFIG_FLATMEM
BUG_ON(mem_map); /* should only be used early */ BUG_ON(mem_map); /* should only be used early */
...@@ -822,7 +822,7 @@ static __init void xen_alloc_pte_init(struct mm_struct *mm, u32 pfn) ...@@ -822,7 +822,7 @@ static __init void xen_alloc_pte_init(struct mm_struct *mm, u32 pfn)
/* Early release_pte assumes that all pts are pinned, since there's /* Early release_pte assumes that all pts are pinned, since there's
only init_mm and anything attached to that is pinned. */ only init_mm and anything attached to that is pinned. */
static void xen_release_pte_init(u32 pfn) static void xen_release_pte_init(unsigned long pfn)
{ {
make_lowmem_page_readwrite(__va(PFN_PHYS(pfn))); make_lowmem_page_readwrite(__va(PFN_PHYS(pfn)));
} }
...@@ -838,7 +838,7 @@ static void pin_pagetable_pfn(unsigned cmd, unsigned long pfn) ...@@ -838,7 +838,7 @@ static void pin_pagetable_pfn(unsigned cmd, unsigned long pfn)
/* This needs to make sure the new pte page is pinned iff its being /* This needs to make sure the new pte page is pinned iff its being
attached to a pinned pagetable. */ attached to a pinned pagetable. */
static void xen_alloc_ptpage(struct mm_struct *mm, u32 pfn, unsigned level) static void xen_alloc_ptpage(struct mm_struct *mm, unsigned long pfn, unsigned level)
{ {
struct page *page = pfn_to_page(pfn); struct page *page = pfn_to_page(pfn);
...@@ -856,12 +856,12 @@ static void xen_alloc_ptpage(struct mm_struct *mm, u32 pfn, unsigned level) ...@@ -856,12 +856,12 @@ static void xen_alloc_ptpage(struct mm_struct *mm, u32 pfn, unsigned level)
} }
} }
static void xen_alloc_pte(struct mm_struct *mm, u32 pfn) static void xen_alloc_pte(struct mm_struct *mm, unsigned long pfn)
{ {
xen_alloc_ptpage(mm, pfn, PT_PTE); xen_alloc_ptpage(mm, pfn, PT_PTE);
} }
static void xen_alloc_pmd(struct mm_struct *mm, u32 pfn) static void xen_alloc_pmd(struct mm_struct *mm, unsigned long pfn)
{ {
xen_alloc_ptpage(mm, pfn, PT_PMD); xen_alloc_ptpage(mm, pfn, PT_PMD);
} }
...@@ -909,7 +909,7 @@ static void xen_pgd_free(struct mm_struct *mm, pgd_t *pgd) ...@@ -909,7 +909,7 @@ static void xen_pgd_free(struct mm_struct *mm, pgd_t *pgd)
} }
/* This should never happen until we're OK to use struct page */ /* This should never happen until we're OK to use struct page */
static void xen_release_ptpage(u32 pfn, unsigned level) static void xen_release_ptpage(unsigned long pfn, unsigned level)
{ {
struct page *page = pfn_to_page(pfn); struct page *page = pfn_to_page(pfn);
...@@ -923,23 +923,23 @@ static void xen_release_ptpage(u32 pfn, unsigned level) ...@@ -923,23 +923,23 @@ static void xen_release_ptpage(u32 pfn, unsigned level)
} }
} }
static void xen_release_pte(u32 pfn) static void xen_release_pte(unsigned long pfn)
{ {
xen_release_ptpage(pfn, PT_PTE); xen_release_ptpage(pfn, PT_PTE);
} }
static void xen_release_pmd(u32 pfn) static void xen_release_pmd(unsigned long pfn)
{ {
xen_release_ptpage(pfn, PT_PMD); xen_release_ptpage(pfn, PT_PMD);
} }
#if PAGETABLE_LEVELS == 4 #if PAGETABLE_LEVELS == 4
static void xen_alloc_pud(struct mm_struct *mm, u32 pfn) static void xen_alloc_pud(struct mm_struct *mm, unsigned long pfn)
{ {
xen_alloc_ptpage(mm, pfn, PT_PUD); xen_alloc_ptpage(mm, pfn, PT_PUD);
} }
static void xen_release_pud(u32 pfn) static void xen_release_pud(unsigned long pfn)
{ {
xen_release_ptpage(pfn, PT_PUD); xen_release_ptpage(pfn, PT_PUD);
} }
......
...@@ -257,13 +257,13 @@ struct pv_mmu_ops { ...@@ -257,13 +257,13 @@ struct pv_mmu_ops {
* Hooks for allocating/releasing pagetable pages when they're * Hooks for allocating/releasing pagetable pages when they're
* attached to a pagetable * attached to a pagetable
*/ */
void (*alloc_pte)(struct mm_struct *mm, u32 pfn); void (*alloc_pte)(struct mm_struct *mm, unsigned long pfn);
void (*alloc_pmd)(struct mm_struct *mm, u32 pfn); void (*alloc_pmd)(struct mm_struct *mm, unsigned long pfn);
void (*alloc_pmd_clone)(u32 pfn, u32 clonepfn, u32 start, u32 count); void (*alloc_pmd_clone)(unsigned long pfn, unsigned long clonepfn, unsigned long start, unsigned long count);
void (*alloc_pud)(struct mm_struct *mm, u32 pfn); void (*alloc_pud)(struct mm_struct *mm, unsigned long pfn);
void (*release_pte)(u32 pfn); void (*release_pte)(unsigned long pfn);
void (*release_pmd)(u32 pfn); void (*release_pmd)(unsigned long pfn);
void (*release_pud)(u32 pfn); void (*release_pud)(unsigned long pfn);
/* Pagetable manipulation functions */ /* Pagetable manipulation functions */
void (*set_pte)(pte_t *ptep, pte_t pteval); void (*set_pte)(pte_t *ptep, pte_t pteval);
...@@ -993,35 +993,35 @@ static inline void paravirt_pgd_free(struct mm_struct *mm, pgd_t *pgd) ...@@ -993,35 +993,35 @@ static inline void paravirt_pgd_free(struct mm_struct *mm, pgd_t *pgd)
PVOP_VCALL2(pv_mmu_ops.pgd_free, mm, pgd); PVOP_VCALL2(pv_mmu_ops.pgd_free, mm, pgd);
} }
static inline void paravirt_alloc_pte(struct mm_struct *mm, unsigned pfn) static inline void paravirt_alloc_pte(struct mm_struct *mm, unsigned long pfn)
{ {
PVOP_VCALL2(pv_mmu_ops.alloc_pte, mm, pfn); PVOP_VCALL2(pv_mmu_ops.alloc_pte, mm, pfn);
} }
static inline void paravirt_release_pte(unsigned pfn) static inline void paravirt_release_pte(unsigned long pfn)
{ {
PVOP_VCALL1(pv_mmu_ops.release_pte, pfn); PVOP_VCALL1(pv_mmu_ops.release_pte, pfn);
} }
static inline void paravirt_alloc_pmd(struct mm_struct *mm, unsigned pfn) static inline void paravirt_alloc_pmd(struct mm_struct *mm, unsigned long pfn)
{ {
PVOP_VCALL2(pv_mmu_ops.alloc_pmd, mm, pfn); PVOP_VCALL2(pv_mmu_ops.alloc_pmd, mm, pfn);
} }
static inline void paravirt_alloc_pmd_clone(unsigned pfn, unsigned clonepfn, static inline void paravirt_alloc_pmd_clone(unsigned long pfn, unsigned long clonepfn,
unsigned start, unsigned count) unsigned long start, unsigned long count)
{ {
PVOP_VCALL4(pv_mmu_ops.alloc_pmd_clone, pfn, clonepfn, start, count); PVOP_VCALL4(pv_mmu_ops.alloc_pmd_clone, pfn, clonepfn, start, count);
} }
static inline void paravirt_release_pmd(unsigned pfn) static inline void paravirt_release_pmd(unsigned long pfn)
{ {
PVOP_VCALL1(pv_mmu_ops.release_pmd, pfn); PVOP_VCALL1(pv_mmu_ops.release_pmd, pfn);
} }
static inline void paravirt_alloc_pud(struct mm_struct *mm, unsigned pfn) static inline void paravirt_alloc_pud(struct mm_struct *mm, unsigned long pfn)
{ {
PVOP_VCALL2(pv_mmu_ops.alloc_pud, mm, pfn); PVOP_VCALL2(pv_mmu_ops.alloc_pud, mm, pfn);
} }
static inline void paravirt_release_pud(unsigned pfn) static inline void paravirt_release_pud(unsigned long pfn)
{ {
PVOP_VCALL1(pv_mmu_ops.release_pud, pfn); PVOP_VCALL1(pv_mmu_ops.release_pud, pfn);
} }
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册