forked from Minki/linux
Merge git://git.kernel.org/pub/scm/linux/kernel/git/x86/linux-2.6-x86
* git://git.kernel.org/pub/scm/linux/kernel/git/x86/linux-2.6-x86: x86: fix "Kernel panic - not syncing: IO-APIC + timer doesn't work!" genirq: revert lazy irq disable for simple irqs x86: also define AT_VECTOR_SIZE_ARCH x86: kprobes bugfix x86: jprobe bugfix timer: kernel/timer.c section fixes genirq: add unlocked version of set_irq_handler() clockevents: fix reprogramming decision in oneshot broadcast oprofile: op_model_athlon.c support for AMD family 10h barcelona performance counters
This commit is contained in:
commit
3e3b3916a9
@ -1882,13 +1882,16 @@ __setup("no_timer_check", notimercheck);
|
||||
static int __init timer_irq_works(void)
|
||||
{
|
||||
unsigned long t1 = jiffies;
|
||||
unsigned long flags;
|
||||
|
||||
if (no_timer_check)
|
||||
return 1;
|
||||
|
||||
local_save_flags(flags);
|
||||
local_irq_enable();
|
||||
/* Let ten ticks pass... */
|
||||
mdelay((10 * 1000) / HZ);
|
||||
local_irq_restore(flags);
|
||||
|
||||
/*
|
||||
* Expect a few ticks at least, to be sure some possible
|
||||
@ -2167,6 +2170,9 @@ static inline void __init check_timer(void)
|
||||
int apic1, pin1, apic2, pin2;
|
||||
int vector;
|
||||
unsigned int ver;
|
||||
unsigned long flags;
|
||||
|
||||
local_irq_save(flags);
|
||||
|
||||
ver = apic_read(APIC_LVR);
|
||||
ver = GET_APIC_VERSION(ver);
|
||||
@ -2219,7 +2225,7 @@ static inline void __init check_timer(void)
|
||||
}
|
||||
if (disable_timer_pin_1 > 0)
|
||||
clear_IO_APIC_pin(0, pin1);
|
||||
return;
|
||||
goto out;
|
||||
}
|
||||
clear_IO_APIC_pin(apic1, pin1);
|
||||
printk(KERN_ERR "..MP-BIOS bug: 8254 timer not connected to "
|
||||
@ -2242,7 +2248,7 @@ static inline void __init check_timer(void)
|
||||
if (nmi_watchdog == NMI_IO_APIC) {
|
||||
setup_nmi();
|
||||
}
|
||||
return;
|
||||
goto out;
|
||||
}
|
||||
/*
|
||||
* Cleanup, just in case ...
|
||||
@ -2266,7 +2272,7 @@ static inline void __init check_timer(void)
|
||||
|
||||
if (timer_irq_works()) {
|
||||
printk(" works.\n");
|
||||
return;
|
||||
goto out;
|
||||
}
|
||||
apic_write_around(APIC_LVT0, APIC_LVT_MASKED | APIC_DM_FIXED | vector);
|
||||
printk(" failed.\n");
|
||||
@ -2282,11 +2288,13 @@ static inline void __init check_timer(void)
|
||||
|
||||
if (timer_irq_works()) {
|
||||
printk(" works.\n");
|
||||
return;
|
||||
goto out;
|
||||
}
|
||||
printk(" failed :(.\n");
|
||||
panic("IO-APIC + timer doesn't work! Boot with apic=debug and send a "
|
||||
"report. Then try booting with the 'noapic' option");
|
||||
out:
|
||||
local_irq_restore(flags);
|
||||
}
|
||||
|
||||
/*
|
||||
|
@ -1281,10 +1281,13 @@ void disable_IO_APIC(void)
|
||||
static int __init timer_irq_works(void)
|
||||
{
|
||||
unsigned long t1 = jiffies;
|
||||
unsigned long flags;
|
||||
|
||||
local_save_flags(flags);
|
||||
local_irq_enable();
|
||||
/* Let ten ticks pass... */
|
||||
mdelay((10 * 1000) / HZ);
|
||||
local_irq_restore(flags);
|
||||
|
||||
/*
|
||||
* Expect a few ticks at least, to be sure some possible
|
||||
@ -1655,6 +1658,9 @@ static inline void check_timer(void)
|
||||
{
|
||||
struct irq_cfg *cfg = irq_cfg + 0;
|
||||
int apic1, pin1, apic2, pin2;
|
||||
unsigned long flags;
|
||||
|
||||
local_irq_save(flags);
|
||||
|
||||
/*
|
||||
* get/set the timer IRQ vector:
|
||||
@ -1696,7 +1702,7 @@ static inline void check_timer(void)
|
||||
}
|
||||
if (disable_timer_pin_1 > 0)
|
||||
clear_IO_APIC_pin(0, pin1);
|
||||
return;
|
||||
goto out;
|
||||
}
|
||||
clear_IO_APIC_pin(apic1, pin1);
|
||||
apic_printk(APIC_QUIET,KERN_ERR "..MP-BIOS bug: 8254 timer not "
|
||||
@ -1718,7 +1724,7 @@ static inline void check_timer(void)
|
||||
if (nmi_watchdog == NMI_IO_APIC) {
|
||||
setup_nmi();
|
||||
}
|
||||
return;
|
||||
goto out;
|
||||
}
|
||||
/*
|
||||
* Cleanup, just in case ...
|
||||
@ -1741,7 +1747,7 @@ static inline void check_timer(void)
|
||||
|
||||
if (timer_irq_works()) {
|
||||
apic_printk(APIC_VERBOSE," works.\n");
|
||||
return;
|
||||
goto out;
|
||||
}
|
||||
apic_write(APIC_LVT0, APIC_LVT_MASKED | APIC_DM_FIXED | cfg->vector);
|
||||
apic_printk(APIC_VERBOSE," failed.\n");
|
||||
@ -1756,10 +1762,12 @@ static inline void check_timer(void)
|
||||
|
||||
if (timer_irq_works()) {
|
||||
apic_printk(APIC_VERBOSE," works.\n");
|
||||
return;
|
||||
goto out;
|
||||
}
|
||||
apic_printk(APIC_VERBOSE," failed :(.\n");
|
||||
panic("IO-APIC + timer doesn't work! Try using the 'noapic' kernel parameter\n");
|
||||
out:
|
||||
local_irq_restore(flags);
|
||||
}
|
||||
|
||||
static int __init notimercheck(char *s)
|
||||
|
@ -727,9 +727,7 @@ int __kprobes longjmp_break_handler(struct kprobe *p, struct pt_regs *regs)
|
||||
|
||||
if ((addr > (u8 *) jprobe_return) && (addr < (u8 *) jprobe_return_end)) {
|
||||
if (®s->esp != kcb->jprobe_saved_esp) {
|
||||
struct pt_regs *saved_regs =
|
||||
container_of(kcb->jprobe_saved_esp,
|
||||
struct pt_regs, esp);
|
||||
struct pt_regs *saved_regs = &kcb->jprobe_saved_regs;
|
||||
printk("current esp %p does not match saved esp %p\n",
|
||||
®s->esp, kcb->jprobe_saved_esp);
|
||||
printk("Saved registers for jprobe %p\n", jp);
|
||||
|
@ -485,7 +485,6 @@ static void __kprobes resume_execution(struct kprobe *p,
|
||||
struct pt_regs *regs, struct kprobe_ctlblk *kcb)
|
||||
{
|
||||
unsigned long *tos = (unsigned long *)regs->rsp;
|
||||
unsigned long next_rip = 0;
|
||||
unsigned long copy_rip = (unsigned long)p->ainsn.insn;
|
||||
unsigned long orig_rip = (unsigned long)p->addr;
|
||||
kprobe_opcode_t *insn = p->ainsn.insn;
|
||||
@ -494,46 +493,42 @@ static void __kprobes resume_execution(struct kprobe *p,
|
||||
if (*insn >= 0x40 && *insn <= 0x4f)
|
||||
insn++;
|
||||
|
||||
regs->eflags &= ~TF_MASK;
|
||||
switch (*insn) {
|
||||
case 0x9c: /* pushfl */
|
||||
case 0x9c: /* pushfl */
|
||||
*tos &= ~(TF_MASK | IF_MASK);
|
||||
*tos |= kcb->kprobe_old_rflags;
|
||||
break;
|
||||
case 0xc3: /* ret/lret */
|
||||
case 0xcb:
|
||||
case 0xc2:
|
||||
case 0xc2: /* iret/ret/lret */
|
||||
case 0xc3:
|
||||
case 0xca:
|
||||
regs->eflags &= ~TF_MASK;
|
||||
/* rip is already adjusted, no more changes required*/
|
||||
return;
|
||||
case 0xe8: /* call relative - Fix return addr */
|
||||
case 0xcb:
|
||||
case 0xcf:
|
||||
case 0xea: /* jmp absolute -- ip is correct */
|
||||
/* ip is already adjusted, no more changes required */
|
||||
goto no_change;
|
||||
case 0xe8: /* call relative - Fix return addr */
|
||||
*tos = orig_rip + (*tos - copy_rip);
|
||||
break;
|
||||
case 0xff:
|
||||
if ((insn[1] & 0x30) == 0x10) {
|
||||
/* call absolute, indirect */
|
||||
/* Fix return addr; rip is correct. */
|
||||
next_rip = regs->rip;
|
||||
/* Fix return addr; ip is correct. */
|
||||
*tos = orig_rip + (*tos - copy_rip);
|
||||
goto no_change;
|
||||
} else if (((insn[1] & 0x31) == 0x20) || /* jmp near, absolute indirect */
|
||||
((insn[1] & 0x31) == 0x21)) { /* jmp far, absolute indirect */
|
||||
/* rip is correct. */
|
||||
next_rip = regs->rip;
|
||||
/* ip is correct. */
|
||||
goto no_change;
|
||||
}
|
||||
break;
|
||||
case 0xea: /* jmp absolute -- rip is correct */
|
||||
next_rip = regs->rip;
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
}
|
||||
|
||||
regs->eflags &= ~TF_MASK;
|
||||
if (next_rip) {
|
||||
regs->rip = next_rip;
|
||||
} else {
|
||||
regs->rip = orig_rip + (regs->rip - copy_rip);
|
||||
}
|
||||
regs->rip = orig_rip + (regs->rip - copy_rip);
|
||||
no_change:
|
||||
|
||||
return;
|
||||
}
|
||||
|
||||
int __kprobes post_kprobe_handler(struct pt_regs *regs)
|
||||
@ -716,10 +711,8 @@ int __kprobes longjmp_break_handler(struct kprobe *p, struct pt_regs *regs)
|
||||
struct jprobe *jp = container_of(p, struct jprobe, kp);
|
||||
|
||||
if ((addr > (u8 *) jprobe_return) && (addr < (u8 *) jprobe_return_end)) {
|
||||
if ((long *)regs->rsp != kcb->jprobe_saved_rsp) {
|
||||
struct pt_regs *saved_regs =
|
||||
container_of(kcb->jprobe_saved_rsp,
|
||||
struct pt_regs, rsp);
|
||||
if ((unsigned long *)regs->rsp != kcb->jprobe_saved_rsp) {
|
||||
struct pt_regs *saved_regs = &kcb->jprobe_saved_regs;
|
||||
printk("current rsp %p does not match saved rsp %p\n",
|
||||
(long *)regs->rsp, kcb->jprobe_saved_rsp);
|
||||
printk("Saved registers for jprobe %p\n", jp);
|
||||
|
@ -1,6 +1,6 @@
|
||||
/**
|
||||
* @file op_model_athlon.h
|
||||
* athlon / K7 model-specific MSR operations
|
||||
* athlon / K7 / K8 / Family 10h model-specific MSR operations
|
||||
*
|
||||
* @remark Copyright 2002 OProfile authors
|
||||
* @remark Read the file COPYING
|
||||
@ -31,12 +31,16 @@
|
||||
#define CTRL_WRITE(l,h,msrs,c) do {wrmsr(msrs->controls[(c)].addr, (l), (h));} while (0)
|
||||
#define CTRL_SET_ACTIVE(n) (n |= (1<<22))
|
||||
#define CTRL_SET_INACTIVE(n) (n &= ~(1<<22))
|
||||
#define CTRL_CLEAR(x) (x &= (1<<21))
|
||||
#define CTRL_CLEAR_LO(x) (x &= (1<<21))
|
||||
#define CTRL_CLEAR_HI(x) (x &= 0xfffffcf0)
|
||||
#define CTRL_SET_ENABLE(val) (val |= 1<<20)
|
||||
#define CTRL_SET_USR(val,u) (val |= ((u & 1) << 16))
|
||||
#define CTRL_SET_KERN(val,k) (val |= ((k & 1) << 17))
|
||||
#define CTRL_SET_UM(val, m) (val |= (m << 8))
|
||||
#define CTRL_SET_EVENT(val, e) (val |= e)
|
||||
#define CTRL_SET_EVENT_LOW(val, e) (val |= (e & 0xff))
|
||||
#define CTRL_SET_EVENT_HIGH(val, e) (val |= ((e >> 8) & 0xf))
|
||||
#define CTRL_SET_HOST_ONLY(val, h) (val |= ((h & 1) << 9))
|
||||
#define CTRL_SET_GUEST_ONLY(val, h) (val |= ((h & 1) << 8))
|
||||
|
||||
static unsigned long reset_value[NUM_COUNTERS];
|
||||
|
||||
@ -70,7 +74,8 @@ static void athlon_setup_ctrs(struct op_msrs const * const msrs)
|
||||
if (unlikely(!CTRL_IS_RESERVED(msrs,i)))
|
||||
continue;
|
||||
CTRL_READ(low, high, msrs, i);
|
||||
CTRL_CLEAR(low);
|
||||
CTRL_CLEAR_LO(low);
|
||||
CTRL_CLEAR_HI(high);
|
||||
CTRL_WRITE(low, high, msrs, i);
|
||||
}
|
||||
|
||||
@ -89,12 +94,17 @@ static void athlon_setup_ctrs(struct op_msrs const * const msrs)
|
||||
CTR_WRITE(counter_config[i].count, msrs, i);
|
||||
|
||||
CTRL_READ(low, high, msrs, i);
|
||||
CTRL_CLEAR(low);
|
||||
CTRL_CLEAR_LO(low);
|
||||
CTRL_CLEAR_HI(high);
|
||||
CTRL_SET_ENABLE(low);
|
||||
CTRL_SET_USR(low, counter_config[i].user);
|
||||
CTRL_SET_KERN(low, counter_config[i].kernel);
|
||||
CTRL_SET_UM(low, counter_config[i].unit_mask);
|
||||
CTRL_SET_EVENT(low, counter_config[i].event);
|
||||
CTRL_SET_EVENT_LOW(low, counter_config[i].event);
|
||||
CTRL_SET_EVENT_HIGH(high, counter_config[i].event);
|
||||
CTRL_SET_HOST_ONLY(high, 0);
|
||||
CTRL_SET_GUEST_ONLY(high, 0);
|
||||
|
||||
CTRL_WRITE(low, high, msrs, i);
|
||||
} else {
|
||||
reset_value[i] = 0;
|
||||
|
@ -73,7 +73,7 @@ struct kprobe_ctlblk {
|
||||
unsigned long kprobe_status;
|
||||
unsigned long kprobe_old_eflags;
|
||||
unsigned long kprobe_saved_eflags;
|
||||
long *jprobe_saved_esp;
|
||||
unsigned long *jprobe_saved_esp;
|
||||
struct pt_regs jprobe_saved_regs;
|
||||
kprobe_opcode_t jprobes_stack[MAX_STACK_SIZE];
|
||||
struct prev_kprobe prev_kprobe;
|
||||
|
@ -66,7 +66,7 @@ struct kprobe_ctlblk {
|
||||
unsigned long kprobe_status;
|
||||
unsigned long kprobe_old_rflags;
|
||||
unsigned long kprobe_saved_rflags;
|
||||
long *jprobe_saved_rsp;
|
||||
unsigned long *jprobe_saved_rsp;
|
||||
struct pt_regs jprobe_saved_regs;
|
||||
kprobe_opcode_t jprobes_stack[MAX_STACK_SIZE];
|
||||
struct prev_kprobe prev_kprobe;
|
||||
|
@ -7,6 +7,13 @@
|
||||
|
||||
#ifdef __KERNEL__
|
||||
|
||||
/* entries in ARCH_DLINFO: */
|
||||
#ifdef CONFIG_IA32_EMULATION
|
||||
# define AT_VECTOR_SIZE_ARCH 2
|
||||
#else
|
||||
# define AT_VECTOR_SIZE_ARCH 1
|
||||
#endif
|
||||
|
||||
#define __SAVE(reg,offset) "movq %%" #reg ",(14-" #offset ")*8(%%rsp)\n\t"
|
||||
#define __RESTORE(reg,offset) "movq (14-" #offset ")*8(%%rsp),%%" #reg "\n\t"
|
||||
|
||||
|
@ -339,6 +339,13 @@ extern void
|
||||
__set_irq_handler(unsigned int irq, irq_flow_handler_t handle, int is_chained,
|
||||
const char *name);
|
||||
|
||||
/* caller has locked the irq_desc and both params are valid */
|
||||
static inline void __set_irq_handler_unlocked(int irq,
|
||||
irq_flow_handler_t handler)
|
||||
{
|
||||
irq_desc[irq].handle_irq = handler;
|
||||
}
|
||||
|
||||
/*
|
||||
* Set a highlevel flow handler for a given IRQ:
|
||||
*/
|
||||
|
@ -297,18 +297,13 @@ handle_simple_irq(unsigned int irq, struct irq_desc *desc)
|
||||
|
||||
if (unlikely(desc->status & IRQ_INPROGRESS))
|
||||
goto out_unlock;
|
||||
desc->status &= ~(IRQ_REPLAY | IRQ_WAITING);
|
||||
kstat_cpu(cpu).irqs[irq]++;
|
||||
|
||||
action = desc->action;
|
||||
if (unlikely(!action || (desc->status & IRQ_DISABLED))) {
|
||||
if (desc->chip->mask)
|
||||
desc->chip->mask(irq);
|
||||
desc->status &= ~(IRQ_REPLAY | IRQ_WAITING);
|
||||
desc->status |= IRQ_PENDING;
|
||||
if (unlikely(!action || (desc->status & IRQ_DISABLED)))
|
||||
goto out_unlock;
|
||||
}
|
||||
|
||||
desc->status &= ~(IRQ_REPLAY | IRQ_WAITING | IRQ_PENDING);
|
||||
desc->status |= IRQ_INPROGRESS;
|
||||
spin_unlock(&desc->lock);
|
||||
|
||||
|
@ -383,33 +383,6 @@ int tick_resume_broadcast_oneshot(struct clock_event_device *bc)
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* Reprogram the broadcast device:
|
||||
*
|
||||
* Called with tick_broadcast_lock held and interrupts disabled.
|
||||
*/
|
||||
static int tick_broadcast_reprogram(void)
|
||||
{
|
||||
ktime_t expires = { .tv64 = KTIME_MAX };
|
||||
struct tick_device *td;
|
||||
int cpu;
|
||||
|
||||
/*
|
||||
* Find the event which expires next:
|
||||
*/
|
||||
for (cpu = first_cpu(tick_broadcast_oneshot_mask); cpu != NR_CPUS;
|
||||
cpu = next_cpu(cpu, tick_broadcast_oneshot_mask)) {
|
||||
td = &per_cpu(tick_cpu_device, cpu);
|
||||
if (td->evtdev->next_event.tv64 < expires.tv64)
|
||||
expires = td->evtdev->next_event;
|
||||
}
|
||||
|
||||
if (expires.tv64 == KTIME_MAX)
|
||||
return 0;
|
||||
|
||||
return tick_broadcast_set_event(expires, 0);
|
||||
}
|
||||
|
||||
/*
|
||||
* Handle oneshot mode broadcasting
|
||||
*/
|
||||
@ -417,12 +390,13 @@ static void tick_handle_oneshot_broadcast(struct clock_event_device *dev)
|
||||
{
|
||||
struct tick_device *td;
|
||||
cpumask_t mask;
|
||||
ktime_t now;
|
||||
ktime_t now, next_event;
|
||||
int cpu;
|
||||
|
||||
spin_lock(&tick_broadcast_lock);
|
||||
again:
|
||||
dev->next_event.tv64 = KTIME_MAX;
|
||||
next_event.tv64 = KTIME_MAX;
|
||||
mask = CPU_MASK_NONE;
|
||||
now = ktime_get();
|
||||
/* Find all expired events */
|
||||
@ -431,19 +405,31 @@ again:
|
||||
td = &per_cpu(tick_cpu_device, cpu);
|
||||
if (td->evtdev->next_event.tv64 <= now.tv64)
|
||||
cpu_set(cpu, mask);
|
||||
else if (td->evtdev->next_event.tv64 < next_event.tv64)
|
||||
next_event.tv64 = td->evtdev->next_event.tv64;
|
||||
}
|
||||
|
||||
/*
|
||||
* Wakeup the cpus which have an expired event. The broadcast
|
||||
* device is reprogrammed in the return from idle code.
|
||||
* Wakeup the cpus which have an expired event.
|
||||
*/
|
||||
if (!tick_do_broadcast(mask)) {
|
||||
tick_do_broadcast(mask);
|
||||
|
||||
/*
|
||||
* Two reasons for reprogram:
|
||||
*
|
||||
* - The global event did not expire any CPU local
|
||||
* events. This happens in dyntick mode, as the maximum PIT
|
||||
* delta is quite small.
|
||||
*
|
||||
* - There are pending events on sleeping CPUs which were not
|
||||
* in the event mask
|
||||
*/
|
||||
if (next_event.tv64 != KTIME_MAX) {
|
||||
/*
|
||||
* The global event did not expire any CPU local
|
||||
* events. This happens in dyntick mode, as the
|
||||
* maximum PIT delta is quite small.
|
||||
* Rearm the broadcast device. If event expired,
|
||||
* repeat the above
|
||||
*/
|
||||
if (tick_broadcast_reprogram())
|
||||
if (tick_broadcast_set_event(next_event, 0))
|
||||
goto again;
|
||||
}
|
||||
spin_unlock(&tick_broadcast_lock);
|
||||
|
@ -1219,11 +1219,11 @@ asmlinkage long sys_sysinfo(struct sysinfo __user *info)
|
||||
*/
|
||||
static struct lock_class_key base_lock_keys[NR_CPUS];
|
||||
|
||||
static int __devinit init_timers_cpu(int cpu)
|
||||
static int __cpuinit init_timers_cpu(int cpu)
|
||||
{
|
||||
int j;
|
||||
tvec_base_t *base;
|
||||
static char __devinitdata tvec_base_done[NR_CPUS];
|
||||
static char __cpuinitdata tvec_base_done[NR_CPUS];
|
||||
|
||||
if (!tvec_base_done[cpu]) {
|
||||
static char boot_done;
|
||||
|
Loading…
Reference in New Issue
Block a user