forked from Minki/linux
2d9e1e2f58
The standard ticket spinlocks are very expensive in a virtual environment, because their performance depends on Xen's scheduler giving vcpus time in the order that they're supposed to take the spinlock. This implements a Xen-specific spinlock, which should be much more efficient. The fast-path is essentially the old Linux-x86 locks, using a single lock byte. The locker decrements the byte; if the result is 0, then they have the lock. If the lock is negative, then locker must spin until the lock is positive again. When there's contention, the locker spin for 2^16[*] iterations waiting to get the lock. If it fails to get the lock in that time, it adds itself to the contention count in the lock and blocks on a per-cpu event channel. When unlocking the spinlock, the locker looks to see if there's anyone blocked waiting for the lock by checking for a non-zero waiter count. If there's a waiter, it traverses the per-cpu "lock_spinners" variable, which contains which lock each CPU is waiting on. It picks one CPU waiting on the lock and sends it an event to wake it up. This allows efficient fast-path spinlock operation, while allowing spinning vcpus to give up their processor time while waiting for a contended lock. [*] 2^16 iterations is threshold at which 98% locks have been taken according to Thomas Friebel's Xen Summit talk "Preventing Guests from Spinning Around". Therefore, we'd expect the lock and unlock slow paths will only be entered 2% of the time. Signed-off-by: Jeremy Fitzhardinge <jeremy.fitzhardinge@citrix.com> Cc: Jens Axboe <axboe@kernel.dk> Cc: Peter Zijlstra <a.p.zijlstra@chello.nl> Cc: Christoph Lameter <clameter@linux-foundation.org> Cc: Petr Tesarik <ptesarik@suse.cz> Cc: Virtualization <virtualization@lists.linux-foundation.org> Cc: Xen devel <xen-devel@lists.xensource.com> Cc: Thomas Friebel <thomas.friebel@amd.com> Cc: Nick Piggin <nickpiggin@yahoo.com.au> Signed-off-by: Ingo Molnar <mingo@elte.hu>
55 lines
1.6 KiB
C
55 lines
1.6 KiB
C
#ifndef _XEN_EVENTS_H
|
|
#define _XEN_EVENTS_H
|
|
|
|
#include <linux/interrupt.h>
|
|
|
|
#include <xen/interface/event_channel.h>
|
|
#include <asm/xen/hypercall.h>
|
|
#include <asm/xen/events.h>
|
|
|
|
int bind_evtchn_to_irq(unsigned int evtchn);
|
|
int bind_evtchn_to_irqhandler(unsigned int evtchn,
|
|
irq_handler_t handler,
|
|
unsigned long irqflags, const char *devname,
|
|
void *dev_id);
|
|
int bind_virq_to_irqhandler(unsigned int virq, unsigned int cpu,
|
|
irq_handler_t handler,
|
|
unsigned long irqflags, const char *devname,
|
|
void *dev_id);
|
|
int bind_ipi_to_irqhandler(enum ipi_vector ipi,
|
|
unsigned int cpu,
|
|
irq_handler_t handler,
|
|
unsigned long irqflags,
|
|
const char *devname,
|
|
void *dev_id);
|
|
|
|
/*
|
|
* Common unbind function for all event sources. Takes IRQ to unbind from.
|
|
* Automatically closes the underlying event channel (even for bindings
|
|
* made with bind_evtchn_to_irqhandler()).
|
|
*/
|
|
void unbind_from_irqhandler(unsigned int irq, void *dev_id);
|
|
|
|
void xen_send_IPI_one(unsigned int cpu, enum ipi_vector vector);
|
|
int resend_irq_on_evtchn(unsigned int irq);
|
|
void rebind_evtchn_irq(int evtchn, int irq);
|
|
|
|
static inline void notify_remote_via_evtchn(int port)
|
|
{
|
|
struct evtchn_send send = { .port = port };
|
|
(void)HYPERVISOR_event_channel_op(EVTCHNOP_send, &send);
|
|
}
|
|
|
|
extern void notify_remote_via_irq(int irq);
|
|
|
|
extern void xen_irq_resume(void);
|
|
|
|
/* Clear an irq's pending state, in preparation for polling on it */
|
|
void xen_clear_irq_pending(int irq);
|
|
|
|
/* Poll waiting for an irq to become pending. In the usual case, the
|
|
irq will be disabled so it won't deliver an interrupt. */
|
|
void xen_poll_irq(int irq);
|
|
|
|
#endif /* _XEN_EVENTS_H */
|