Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 7708ad64 authored by Jeremy Fitzhardinge's avatar Jeremy Fitzhardinge Committed by Ingo Molnar
Browse files

xen: add xen_ prefixes to make tracing with ftrace easier



It's easier to pattern match on Xen function if they all start with xen_.

Signed-off-by: default avatarJeremy Fitzhardinge <jeremy.fitzhardinge@citrix.com>
Signed-off-by: default avatarIngo Molnar <mingo@elte.hu>
parent 11ad93e5
Loading
Loading
Loading
Loading
+34 −32
Original line number Original line Diff line number Diff line
@@ -229,14 +229,14 @@ void make_lowmem_page_readwrite(void *vaddr)
}
}




static bool page_pinned(void *ptr)
static bool xen_page_pinned(void *ptr)
{
{
	struct page *page = virt_to_page(ptr);
	struct page *page = virt_to_page(ptr);


	return PagePinned(page);
	return PagePinned(page);
}
}


static void extend_mmu_update(const struct mmu_update *update)
static void xen_extend_mmu_update(const struct mmu_update *update)
{
{
	struct multicall_space mcs;
	struct multicall_space mcs;
	struct mmu_update *u;
	struct mmu_update *u;
@@ -265,7 +265,7 @@ void xen_set_pmd_hyper(pmd_t *ptr, pmd_t val)
	/* ptr may be ioremapped for 64-bit pagetable setup */
	/* ptr may be ioremapped for 64-bit pagetable setup */
	u.ptr = arbitrary_virt_to_machine(ptr).maddr;
	u.ptr = arbitrary_virt_to_machine(ptr).maddr;
	u.val = pmd_val_ma(val);
	u.val = pmd_val_ma(val);
	extend_mmu_update(&u);
	xen_extend_mmu_update(&u);


	xen_mc_issue(PARAVIRT_LAZY_MMU);
	xen_mc_issue(PARAVIRT_LAZY_MMU);


@@ -276,7 +276,7 @@ void xen_set_pmd(pmd_t *ptr, pmd_t val)
{
{
	/* If page is not pinned, we can just update the entry
	/* If page is not pinned, we can just update the entry
	   directly */
	   directly */
	if (!page_pinned(ptr)) {
	if (!xen_page_pinned(ptr)) {
		*ptr = val;
		*ptr = val;
		return;
		return;
	}
	}
@@ -334,7 +334,7 @@ void xen_ptep_modify_prot_commit(struct mm_struct *mm, unsigned long addr,


	u.ptr = virt_to_machine(ptep).maddr | MMU_PT_UPDATE_PRESERVE_AD;
	u.ptr = virt_to_machine(ptep).maddr | MMU_PT_UPDATE_PRESERVE_AD;
	u.val = pte_val_ma(pte);
	u.val = pte_val_ma(pte);
	extend_mmu_update(&u);
	xen_extend_mmu_update(&u);


	xen_mc_issue(PARAVIRT_LAZY_MMU);
	xen_mc_issue(PARAVIRT_LAZY_MMU);
}
}
@@ -400,7 +400,7 @@ void xen_set_pud_hyper(pud_t *ptr, pud_t val)
	/* ptr may be ioremapped for 64-bit pagetable setup */
	/* ptr may be ioremapped for 64-bit pagetable setup */
	u.ptr = arbitrary_virt_to_machine(ptr).maddr;
	u.ptr = arbitrary_virt_to_machine(ptr).maddr;
	u.val = pud_val_ma(val);
	u.val = pud_val_ma(val);
	extend_mmu_update(&u);
	xen_extend_mmu_update(&u);


	xen_mc_issue(PARAVIRT_LAZY_MMU);
	xen_mc_issue(PARAVIRT_LAZY_MMU);


@@ -411,7 +411,7 @@ void xen_set_pud(pud_t *ptr, pud_t val)
{
{
	/* If page is not pinned, we can just update the entry
	/* If page is not pinned, we can just update the entry
	   directly */
	   directly */
	if (!page_pinned(ptr)) {
	if (!xen_page_pinned(ptr)) {
		*ptr = val;
		*ptr = val;
		return;
		return;
	}
	}
@@ -490,7 +490,7 @@ static void __xen_set_pgd_hyper(pgd_t *ptr, pgd_t val)


	u.ptr = virt_to_machine(ptr).maddr;
	u.ptr = virt_to_machine(ptr).maddr;
	u.val = pgd_val_ma(val);
	u.val = pgd_val_ma(val);
	extend_mmu_update(&u);
	xen_extend_mmu_update(&u);
}
}


/*
/*
@@ -519,10 +519,10 @@ void xen_set_pgd(pgd_t *ptr, pgd_t val)


	/* If page is not pinned, we can just update the entry
	/* If page is not pinned, we can just update the entry
	   directly */
	   directly */
	if (!page_pinned(ptr)) {
	if (!xen_page_pinned(ptr)) {
		*ptr = val;
		*ptr = val;
		if (user_ptr) {
		if (user_ptr) {
			WARN_ON(page_pinned(user_ptr));
			WARN_ON(xen_page_pinned(user_ptr));
			*user_ptr = val;
			*user_ptr = val;
		}
		}
		return;
		return;
@@ -555,7 +555,7 @@ void xen_set_pgd(pgd_t *ptr, pgd_t val)
 * For 64-bit, we must skip the Xen hole in the middle of the address
 * For 64-bit, we must skip the Xen hole in the middle of the address
 * space, just after the big x86-64 virtual hole.
 * space, just after the big x86-64 virtual hole.
 */
 */
static int pgd_walk(pgd_t *pgd, int (*func)(struct page *, enum pt_level),
static int xen_pgd_walk(pgd_t *pgd, int (*func)(struct page *, enum pt_level),
			unsigned long limit)
			unsigned long limit)
{
{
	int flush = 0;
	int flush = 0;
@@ -644,7 +644,9 @@ out:
	return flush;
	return flush;
}
}


static spinlock_t *lock_pte(struct page *page)
/* If we're using split pte locks, then take the page's lock and
   return a pointer to it.  Otherwise return NULL. */
static spinlock_t *xen_pte_lock(struct page *page)
{
{
	spinlock_t *ptl = NULL;
	spinlock_t *ptl = NULL;


@@ -656,7 +658,7 @@ static spinlock_t *lock_pte(struct page *page)
	return ptl;
	return ptl;
}
}


static void do_unlock(void *v)
static void xen_pte_unlock(void *v)
{
{
	spinlock_t *ptl = v;
	spinlock_t *ptl = v;
	spin_unlock(ptl);
	spin_unlock(ptl);
@@ -674,7 +676,7 @@ static void xen_do_pin(unsigned level, unsigned long pfn)
	MULTI_mmuext_op(mcs.mc, op, 1, NULL, DOMID_SELF);
	MULTI_mmuext_op(mcs.mc, op, 1, NULL, DOMID_SELF);
}
}


static int pin_page(struct page *page, enum pt_level level)
static int xen_pin_page(struct page *page, enum pt_level level)
{
{
	unsigned pgfl = TestSetPagePinned(page);
	unsigned pgfl = TestSetPagePinned(page);
	int flush;
	int flush;
@@ -715,7 +717,7 @@ static int pin_page(struct page *page, enum pt_level level)
		 */
		 */
		ptl = NULL;
		ptl = NULL;
		if (level == PT_PTE)
		if (level == PT_PTE)
			ptl = lock_pte(page);
			ptl = xen_pte_lock(page);


		MULTI_update_va_mapping(mcs.mc, (unsigned long)pt,
		MULTI_update_va_mapping(mcs.mc, (unsigned long)pt,
					pfn_pte(pfn, PAGE_KERNEL_RO),
					pfn_pte(pfn, PAGE_KERNEL_RO),
@@ -726,7 +728,7 @@ static int pin_page(struct page *page, enum pt_level level)


			/* Queue a deferred unlock for when this batch
			/* Queue a deferred unlock for when this batch
			   is completed. */
			   is completed. */
			xen_mc_callback(do_unlock, ptl);
			xen_mc_callback(xen_pte_unlock, ptl);
		}
		}
	}
	}


@@ -740,7 +742,7 @@ void xen_pgd_pin(pgd_t *pgd)
{
{
	xen_mc_batch();
	xen_mc_batch();


	if (pgd_walk(pgd, pin_page, USER_LIMIT)) {
	if (xen_pgd_walk(pgd, xen_pin_page, USER_LIMIT)) {
		/* re-enable interrupts for kmap_flush_unused */
		/* re-enable interrupts for kmap_flush_unused */
		xen_mc_issue(0);
		xen_mc_issue(0);
		kmap_flush_unused();
		kmap_flush_unused();
@@ -754,14 +756,14 @@ void xen_pgd_pin(pgd_t *pgd)
		xen_do_pin(MMUEXT_PIN_L4_TABLE, PFN_DOWN(__pa(pgd)));
		xen_do_pin(MMUEXT_PIN_L4_TABLE, PFN_DOWN(__pa(pgd)));


		if (user_pgd) {
		if (user_pgd) {
			pin_page(virt_to_page(user_pgd), PT_PGD);
			xen_pin_page(virt_to_page(user_pgd), PT_PGD);
			xen_do_pin(MMUEXT_PIN_L4_TABLE, PFN_DOWN(__pa(user_pgd)));
			xen_do_pin(MMUEXT_PIN_L4_TABLE, PFN_DOWN(__pa(user_pgd)));
		}
		}
	}
	}
#else /* CONFIG_X86_32 */
#else /* CONFIG_X86_32 */
#ifdef CONFIG_X86_PAE
#ifdef CONFIG_X86_PAE
	/* Need to make sure unshared kernel PMD is pinnable */
	/* Need to make sure unshared kernel PMD is pinnable */
	pin_page(virt_to_page(pgd_page(pgd[pgd_index(TASK_SIZE)])), PT_PMD);
	xen_pin_page(virt_to_page(pgd_page(pgd[pgd_index(TASK_SIZE)])), PT_PMD);
#endif
#endif
	xen_do_pin(MMUEXT_PIN_L3_TABLE, PFN_DOWN(__pa(pgd)));
	xen_do_pin(MMUEXT_PIN_L3_TABLE, PFN_DOWN(__pa(pgd)));
#endif /* CONFIG_X86_64 */
#endif /* CONFIG_X86_64 */
@@ -796,7 +798,7 @@ void xen_mm_pin_all(void)
 * that's before we have page structures to store the bits.  So do all
 * that's before we have page structures to store the bits.  So do all
 * the book-keeping now.
 * the book-keeping now.
 */
 */
static __init int mark_pinned(struct page *page, enum pt_level level)
static __init int xen_mark_pinned(struct page *page, enum pt_level level)
{
{
	SetPagePinned(page);
	SetPagePinned(page);
	return 0;
	return 0;
@@ -804,10 +806,10 @@ static __init int mark_pinned(struct page *page, enum pt_level level)


void __init xen_mark_init_mm_pinned(void)
void __init xen_mark_init_mm_pinned(void)
{
{
	pgd_walk(init_mm.pgd, mark_pinned, FIXADDR_TOP);
	xen_pgd_walk(init_mm.pgd, xen_mark_pinned, FIXADDR_TOP);
}
}


static int unpin_page(struct page *page, enum pt_level level)
static int xen_unpin_page(struct page *page, enum pt_level level)
{
{
	unsigned pgfl = TestClearPagePinned(page);
	unsigned pgfl = TestClearPagePinned(page);


@@ -825,7 +827,7 @@ static int unpin_page(struct page *page, enum pt_level level)
		 * partially-pinned state.
		 * partially-pinned state.
		 */
		 */
		if (level == PT_PTE) {
		if (level == PT_PTE) {
			ptl = lock_pte(page);
			ptl = xen_pte_lock(page);


			if (ptl)
			if (ptl)
				xen_do_pin(MMUEXT_UNPIN_TABLE, pfn);
				xen_do_pin(MMUEXT_UNPIN_TABLE, pfn);
@@ -839,7 +841,7 @@ static int unpin_page(struct page *page, enum pt_level level)


		if (ptl) {
		if (ptl) {
			/* unlock when batch completed */
			/* unlock when batch completed */
			xen_mc_callback(do_unlock, ptl);
			xen_mc_callback(xen_pte_unlock, ptl);
		}
		}
	}
	}


@@ -859,17 +861,17 @@ static void xen_pgd_unpin(pgd_t *pgd)


		if (user_pgd) {
		if (user_pgd) {
			xen_do_pin(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(user_pgd)));
			xen_do_pin(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(user_pgd)));
			unpin_page(virt_to_page(user_pgd), PT_PGD);
			xen_unpin_page(virt_to_page(user_pgd), PT_PGD);
		}
		}
	}
	}
#endif
#endif


#ifdef CONFIG_X86_PAE
#ifdef CONFIG_X86_PAE
	/* Need to make sure unshared kernel PMD is unpinned */
	/* Need to make sure unshared kernel PMD is unpinned */
	unpin_page(virt_to_page(pgd_page(pgd[pgd_index(TASK_SIZE)])), PT_PMD);
	xen_unpin_page(virt_to_page(pgd_page(pgd[pgd_index(TASK_SIZE)])), PT_PMD);
#endif
#endif


	pgd_walk(pgd, unpin_page, USER_LIMIT);
	xen_pgd_walk(pgd, xen_unpin_page, USER_LIMIT);


	xen_mc_issue(0);
	xen_mc_issue(0);
}
}
@@ -936,7 +938,7 @@ static void drop_other_mm_ref(void *info)
	}
	}
}
}


static void drop_mm_ref(struct mm_struct *mm)
static void xen_drop_mm_ref(struct mm_struct *mm)
{
{
	cpumask_t mask;
	cpumask_t mask;
	unsigned cpu;
	unsigned cpu;
@@ -966,7 +968,7 @@ static void drop_mm_ref(struct mm_struct *mm)
		smp_call_function_mask(mask, drop_other_mm_ref, mm, 1);
		smp_call_function_mask(mask, drop_other_mm_ref, mm, 1);
}
}
#else
#else
static void drop_mm_ref(struct mm_struct *mm)
static void xen_drop_mm_ref(struct mm_struct *mm)
{
{
	if (current->active_mm == mm)
	if (current->active_mm == mm)
		load_cr3(swapper_pg_dir);
		load_cr3(swapper_pg_dir);
@@ -990,13 +992,13 @@ static void drop_mm_ref(struct mm_struct *mm)
void xen_exit_mmap(struct mm_struct *mm)
void xen_exit_mmap(struct mm_struct *mm)
{
{
	get_cpu();		/* make sure we don't move around */
	get_cpu();		/* make sure we don't move around */
	drop_mm_ref(mm);
	xen_drop_mm_ref(mm);
	put_cpu();
	put_cpu();


	spin_lock(&mm->page_table_lock);
	spin_lock(&mm->page_table_lock);


	/* pgd may not be pinned in the error exit path of execve */
	/* pgd may not be pinned in the error exit path of execve */
	if (page_pinned(mm->pgd))
	if (xen_page_pinned(mm->pgd))
		xen_pgd_unpin(mm->pgd);
		xen_pgd_unpin(mm->pgd);


	spin_unlock(&mm->page_table_lock);
	spin_unlock(&mm->page_table_lock);