This is a follow-up of commit:
cfd8983f03
("x86, locking/spinlocks: Remove ticket (spin)lock implementation")
The static_key structure 'paravirt_ticketlocks_enabled' is now removed as it is
no longer used.
As a result, the init functions kvm_spinlock_init_jump() and
xen_init_spinlocks_jump() are also removed.
A simple build and boot test was done to verify it.
Signed-off-by: Waiman Long <longman@redhat.com>
Reviewed-by: Boris Ostrovsky <boris.ostrovsky@oracle.com>
Cc: Alok Kataria <akataria@vmware.com>
Cc: Chris Wright <chrisw@sous-sol.org>
Cc: Jeremy Fitzhardinge <jeremy@goop.org>
Cc: Juergen Gross <jgross@suse.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Paolo Bonzini <pbonzini@redhat.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Radim Krčmář <rkrcmar@redhat.com>
Cc: Rusty Russell <rusty@rustcorp.com.au>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: kvm@vger.kernel.org
Cc: linux-arch@vger.kernel.org
Cc: virtualization@lists.linux-foundation.org
Cc: xen-devel@lists.xenproject.org
Link: http://lkml.kernel.org/r/1484252878-1962-1-git-send-email-longman@redhat.com
Signed-off-by: Ingo Molnar <mingo@kernel.org>
45 lines
1.1 KiB
C
45 lines
1.1 KiB
C
/*
|
|
* Split spinlock implementation out into its own file, so it can be
|
|
* compiled in a FTRACE-compatible way.
|
|
*/
|
|
#include <linux/spinlock.h>
|
|
#include <linux/export.h>
|
|
#include <linux/jump_label.h>
|
|
|
|
#include <asm/paravirt.h>
|
|
|
|
__visible void __native_queued_spin_unlock(struct qspinlock *lock)
|
|
{
|
|
native_queued_spin_unlock(lock);
|
|
}
|
|
PV_CALLEE_SAVE_REGS_THUNK(__native_queued_spin_unlock);
|
|
|
|
bool pv_is_native_spin_unlock(void)
|
|
{
|
|
return pv_lock_ops.queued_spin_unlock.func ==
|
|
__raw_callee_save___native_queued_spin_unlock;
|
|
}
|
|
|
|
__visible bool __native_vcpu_is_preempted(int cpu)
|
|
{
|
|
return false;
|
|
}
|
|
PV_CALLEE_SAVE_REGS_THUNK(__native_vcpu_is_preempted);
|
|
|
|
bool pv_is_native_vcpu_is_preempted(void)
|
|
{
|
|
return pv_lock_ops.vcpu_is_preempted.func ==
|
|
__raw_callee_save___native_vcpu_is_preempted;
|
|
}
|
|
|
|
struct pv_lock_ops pv_lock_ops = {
|
|
#ifdef CONFIG_SMP
|
|
.queued_spin_lock_slowpath = native_queued_spin_lock_slowpath,
|
|
.queued_spin_unlock = PV_CALLEE_SAVE(__native_queued_spin_unlock),
|
|
.wait = paravirt_nop,
|
|
.kick = paravirt_nop,
|
|
.vcpu_is_preempted = PV_CALLEE_SAVE(__native_vcpu_is_preempted),
|
|
#endif /* SMP */
|
|
};
|
|
EXPORT_SYMBOL(pv_lock_ops);
|