mirror of
https://github.com/torvalds/linux.git
synced 2024-11-17 17:41:44 +00:00
arch/x86/xen/mmu: Cleanup code/data sections definitions
Cleanup code/data sections definitions accordingly to include/linux/init.h. Signed-off-by: Daniel Kiper <dkiper@net-space.pl> [v1: Rebased on top of latest linus's to include fixes in mmu.c] Signed-off-by: Konrad Rzeszutek Wilk <konrad.wilk@oracle.com>
This commit is contained in:
parent
ae15a3b4d1
commit
3f508953dd
@ -1054,7 +1054,7 @@ void xen_mm_pin_all(void)
|
|||||||
* that's before we have page structures to store the bits. So do all
|
* that's before we have page structures to store the bits. So do all
|
||||||
* the book-keeping now.
|
* the book-keeping now.
|
||||||
*/
|
*/
|
||||||
static __init int xen_mark_pinned(struct mm_struct *mm, struct page *page,
|
static int __init xen_mark_pinned(struct mm_struct *mm, struct page *page,
|
||||||
enum pt_level level)
|
enum pt_level level)
|
||||||
{
|
{
|
||||||
SetPagePinned(page);
|
SetPagePinned(page);
|
||||||
@ -1271,7 +1271,7 @@ void xen_exit_mmap(struct mm_struct *mm)
|
|||||||
spin_unlock(&mm->page_table_lock);
|
spin_unlock(&mm->page_table_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
static __init void xen_pagetable_setup_start(pgd_t *base)
|
static void __init xen_pagetable_setup_start(pgd_t *base)
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1291,7 +1291,7 @@ static __init void xen_mapping_pagetable_reserve(u64 start, u64 end)
|
|||||||
|
|
||||||
static void xen_post_allocator_init(void);
|
static void xen_post_allocator_init(void);
|
||||||
|
|
||||||
static __init void xen_pagetable_setup_done(pgd_t *base)
|
static void __init xen_pagetable_setup_done(pgd_t *base)
|
||||||
{
|
{
|
||||||
xen_setup_shared_info();
|
xen_setup_shared_info();
|
||||||
xen_post_allocator_init();
|
xen_post_allocator_init();
|
||||||
@ -1488,7 +1488,7 @@ static void xen_pgd_free(struct mm_struct *mm, pgd_t *pgd)
|
|||||||
}
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_X86_32
|
#ifdef CONFIG_X86_32
|
||||||
static __init pte_t mask_rw_pte(pte_t *ptep, pte_t pte)
|
static pte_t __init mask_rw_pte(pte_t *ptep, pte_t pte)
|
||||||
{
|
{
|
||||||
/* If there's an existing pte, then don't allow _PAGE_RW to be set */
|
/* If there's an existing pte, then don't allow _PAGE_RW to be set */
|
||||||
if (pte_val_ma(*ptep) & _PAGE_PRESENT)
|
if (pte_val_ma(*ptep) & _PAGE_PRESENT)
|
||||||
@ -1498,7 +1498,7 @@ static __init pte_t mask_rw_pte(pte_t *ptep, pte_t pte)
|
|||||||
return pte;
|
return pte;
|
||||||
}
|
}
|
||||||
#else /* CONFIG_X86_64 */
|
#else /* CONFIG_X86_64 */
|
||||||
static __init pte_t mask_rw_pte(pte_t *ptep, pte_t pte)
|
static pte_t __init mask_rw_pte(pte_t *ptep, pte_t pte)
|
||||||
{
|
{
|
||||||
unsigned long pfn = pte_pfn(pte);
|
unsigned long pfn = pte_pfn(pte);
|
||||||
|
|
||||||
@ -1519,7 +1519,7 @@ static __init pte_t mask_rw_pte(pte_t *ptep, pte_t pte)
|
|||||||
|
|
||||||
/* Init-time set_pte while constructing initial pagetables, which
|
/* Init-time set_pte while constructing initial pagetables, which
|
||||||
doesn't allow RO pagetable pages to be remapped RW */
|
doesn't allow RO pagetable pages to be remapped RW */
|
||||||
static __init void xen_set_pte_init(pte_t *ptep, pte_t pte)
|
static void __init xen_set_pte_init(pte_t *ptep, pte_t pte)
|
||||||
{
|
{
|
||||||
pte = mask_rw_pte(ptep, pte);
|
pte = mask_rw_pte(ptep, pte);
|
||||||
|
|
||||||
@ -1537,7 +1537,7 @@ static void pin_pagetable_pfn(unsigned cmd, unsigned long pfn)
|
|||||||
|
|
||||||
/* Early in boot, while setting up the initial pagetable, assume
|
/* Early in boot, while setting up the initial pagetable, assume
|
||||||
everything is pinned. */
|
everything is pinned. */
|
||||||
static __init void xen_alloc_pte_init(struct mm_struct *mm, unsigned long pfn)
|
static void __init xen_alloc_pte_init(struct mm_struct *mm, unsigned long pfn)
|
||||||
{
|
{
|
||||||
#ifdef CONFIG_FLATMEM
|
#ifdef CONFIG_FLATMEM
|
||||||
BUG_ON(mem_map); /* should only be used early */
|
BUG_ON(mem_map); /* should only be used early */
|
||||||
@ -1547,7 +1547,7 @@ static __init void xen_alloc_pte_init(struct mm_struct *mm, unsigned long pfn)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/* Used for pmd and pud */
|
/* Used for pmd and pud */
|
||||||
static __init void xen_alloc_pmd_init(struct mm_struct *mm, unsigned long pfn)
|
static void __init xen_alloc_pmd_init(struct mm_struct *mm, unsigned long pfn)
|
||||||
{
|
{
|
||||||
#ifdef CONFIG_FLATMEM
|
#ifdef CONFIG_FLATMEM
|
||||||
BUG_ON(mem_map); /* should only be used early */
|
BUG_ON(mem_map); /* should only be used early */
|
||||||
@ -1557,13 +1557,13 @@ static __init void xen_alloc_pmd_init(struct mm_struct *mm, unsigned long pfn)
|
|||||||
|
|
||||||
/* Early release_pte assumes that all pts are pinned, since there's
|
/* Early release_pte assumes that all pts are pinned, since there's
|
||||||
only init_mm and anything attached to that is pinned. */
|
only init_mm and anything attached to that is pinned. */
|
||||||
static __init void xen_release_pte_init(unsigned long pfn)
|
static void __init xen_release_pte_init(unsigned long pfn)
|
||||||
{
|
{
|
||||||
pin_pagetable_pfn(MMUEXT_UNPIN_TABLE, pfn);
|
pin_pagetable_pfn(MMUEXT_UNPIN_TABLE, pfn);
|
||||||
make_lowmem_page_readwrite(__va(PFN_PHYS(pfn)));
|
make_lowmem_page_readwrite(__va(PFN_PHYS(pfn)));
|
||||||
}
|
}
|
||||||
|
|
||||||
static __init void xen_release_pmd_init(unsigned long pfn)
|
static void __init xen_release_pmd_init(unsigned long pfn)
|
||||||
{
|
{
|
||||||
make_lowmem_page_readwrite(__va(PFN_PHYS(pfn)));
|
make_lowmem_page_readwrite(__va(PFN_PHYS(pfn)));
|
||||||
}
|
}
|
||||||
@ -1689,7 +1689,7 @@ static void set_page_prot(void *addr, pgprot_t prot)
|
|||||||
BUG();
|
BUG();
|
||||||
}
|
}
|
||||||
|
|
||||||
static __init void xen_map_identity_early(pmd_t *pmd, unsigned long max_pfn)
|
static void __init xen_map_identity_early(pmd_t *pmd, unsigned long max_pfn)
|
||||||
{
|
{
|
||||||
unsigned pmdidx, pteidx;
|
unsigned pmdidx, pteidx;
|
||||||
unsigned ident_pte;
|
unsigned ident_pte;
|
||||||
@ -1772,7 +1772,7 @@ static void convert_pfn_mfn(void *v)
|
|||||||
* of the physical mapping once some sort of allocator has been set
|
* of the physical mapping once some sort of allocator has been set
|
||||||
* up.
|
* up.
|
||||||
*/
|
*/
|
||||||
__init pgd_t *xen_setup_kernel_pagetable(pgd_t *pgd,
|
pgd_t * __init xen_setup_kernel_pagetable(pgd_t *pgd,
|
||||||
unsigned long max_pfn)
|
unsigned long max_pfn)
|
||||||
{
|
{
|
||||||
pud_t *l3;
|
pud_t *l3;
|
||||||
@ -1843,7 +1843,7 @@ __init pgd_t *xen_setup_kernel_pagetable(pgd_t *pgd,
|
|||||||
static RESERVE_BRK_ARRAY(pmd_t, initial_kernel_pmd, PTRS_PER_PMD);
|
static RESERVE_BRK_ARRAY(pmd_t, initial_kernel_pmd, PTRS_PER_PMD);
|
||||||
static RESERVE_BRK_ARRAY(pmd_t, swapper_kernel_pmd, PTRS_PER_PMD);
|
static RESERVE_BRK_ARRAY(pmd_t, swapper_kernel_pmd, PTRS_PER_PMD);
|
||||||
|
|
||||||
static __init void xen_write_cr3_init(unsigned long cr3)
|
static void __init xen_write_cr3_init(unsigned long cr3)
|
||||||
{
|
{
|
||||||
unsigned long pfn = PFN_DOWN(__pa(swapper_pg_dir));
|
unsigned long pfn = PFN_DOWN(__pa(swapper_pg_dir));
|
||||||
|
|
||||||
@ -1880,7 +1880,7 @@ static __init void xen_write_cr3_init(unsigned long cr3)
|
|||||||
pv_mmu_ops.write_cr3 = &xen_write_cr3;
|
pv_mmu_ops.write_cr3 = &xen_write_cr3;
|
||||||
}
|
}
|
||||||
|
|
||||||
__init pgd_t *xen_setup_kernel_pagetable(pgd_t *pgd,
|
pgd_t * __init xen_setup_kernel_pagetable(pgd_t *pgd,
|
||||||
unsigned long max_pfn)
|
unsigned long max_pfn)
|
||||||
{
|
{
|
||||||
pmd_t *kernel_pmd;
|
pmd_t *kernel_pmd;
|
||||||
@ -1986,7 +1986,7 @@ static void xen_set_fixmap(unsigned idx, phys_addr_t phys, pgprot_t prot)
|
|||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
__init void xen_ident_map_ISA(void)
|
void __init xen_ident_map_ISA(void)
|
||||||
{
|
{
|
||||||
unsigned long pa;
|
unsigned long pa;
|
||||||
|
|
||||||
@ -2009,7 +2009,7 @@ __init void xen_ident_map_ISA(void)
|
|||||||
xen_flush_tlb();
|
xen_flush_tlb();
|
||||||
}
|
}
|
||||||
|
|
||||||
static __init void xen_post_allocator_init(void)
|
static void __init xen_post_allocator_init(void)
|
||||||
{
|
{
|
||||||
#ifdef CONFIG_XEN_DEBUG
|
#ifdef CONFIG_XEN_DEBUG
|
||||||
pv_mmu_ops.make_pte = PV_CALLEE_SAVE(xen_make_pte_debug);
|
pv_mmu_ops.make_pte = PV_CALLEE_SAVE(xen_make_pte_debug);
|
||||||
@ -2046,7 +2046,7 @@ static void xen_leave_lazy_mmu(void)
|
|||||||
preempt_enable();
|
preempt_enable();
|
||||||
}
|
}
|
||||||
|
|
||||||
static const struct pv_mmu_ops xen_mmu_ops __initdata = {
|
static const struct pv_mmu_ops xen_mmu_ops __initconst = {
|
||||||
.read_cr2 = xen_read_cr2,
|
.read_cr2 = xen_read_cr2,
|
||||||
.write_cr2 = xen_write_cr2,
|
.write_cr2 = xen_write_cr2,
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user