forked from Minki/linux
39c715b717
This patch implements a number of smp_processor_id() cleanup ideas that Arjan van de Ven and I came up with. The previous __smp_processor_id/_smp_processor_id/smp_processor_id API spaghetti was hard to follow both on the implementational and on the usage side. Some of the complexity arose from picking wrong names, some of the complexity comes from the fact that not all architectures defined __smp_processor_id. In the new code, there are two externally visible symbols: - smp_processor_id(): debug variant. - raw_smp_processor_id(): nondebug variant. Replaces all existing uses of _smp_processor_id() and __smp_processor_id(). Defined by every SMP architecture in include/asm-*/smp.h. There is one new internal symbol, dependent on DEBUG_PREEMPT: - debug_smp_processor_id(): internal debug variant, mapped to smp_processor_id(). Also, i moved debug_smp_processor_id() from lib/kernel_lock.c into a new lib/smp_processor_id.c file. All related comments got updated and/or clarified. I have build/boot tested the following 8 .config combinations on x86: {SMP,UP} x {PREEMPT,!PREEMPT} x {DEBUG_PREEMPT,!DEBUG_PREEMPT} I have also build/boot tested x64 on UP/PREEMPT/DEBUG_PREEMPT. (Other architectures are untested, but should work just fine.) Signed-off-by: Ingo Molnar <mingo@elte.hu> Signed-off-by: Arjan van de Ven <arjan@infradead.org> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
119 lines
3.5 KiB
C
119 lines
3.5 KiB
C
#ifndef _ASM_M32R_SMP_H
|
|
#define _ASM_M32R_SMP_H
|
|
|
|
/* $Id$ */
|
|
|
|
#include <linux/config.h>
|
|
|
|
#ifdef CONFIG_SMP
|
|
#ifndef __ASSEMBLY__
|
|
|
|
#include <linux/cpumask.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/threads.h>
|
|
#include <asm/m32r.h>
|
|
|
|
#define PHYSID_ARRAY_SIZE 1
|
|
|
|
struct physid_mask
|
|
{
|
|
unsigned long mask[PHYSID_ARRAY_SIZE];
|
|
};
|
|
|
|
typedef struct physid_mask physid_mask_t;
|
|
|
|
#define physid_set(physid, map) set_bit(physid, (map).mask)
|
|
#define physid_clear(physid, map) clear_bit(physid, (map).mask)
|
|
#define physid_isset(physid, map) test_bit(physid, (map).mask)
|
|
#define physid_test_and_set(physid, map) test_and_set_bit(physid, (map).mask)
|
|
|
|
#define physids_and(dst, src1, src2) bitmap_and((dst).mask, (src1).mask, (src2).mask, MAX_APICS)
|
|
#define physids_or(dst, src1, src2) bitmap_or((dst).mask, (src1).mask, (src2).mask, MAX_APICS)
|
|
#define physids_clear(map) bitmap_zero((map).mask, MAX_APICS)
|
|
#define physids_complement(dst, src) bitmap_complement((dst).mask,(src).mask, MAX_APICS)
|
|
#define physids_empty(map) bitmap_empty((map).mask, MAX_APICS)
|
|
#define physids_equal(map1, map2) bitmap_equal((map1).mask, (map2).mask, MAX_APICS)
|
|
#define physids_weight(map) bitmap_weight((map).mask, MAX_APICS)
|
|
#define physids_shift_right(d, s, n) bitmap_shift_right((d).mask, (s).mask, n, MAX_APICS)
|
|
#define physids_shift_left(d, s, n) bitmap_shift_left((d).mask, (s).mask, n, MAX_APICS)
|
|
#define physids_coerce(map) ((map).mask[0])
|
|
|
|
#define physids_promote(physids) \
|
|
({ \
|
|
physid_mask_t __physid_mask = PHYSID_MASK_NONE; \
|
|
__physid_mask.mask[0] = physids; \
|
|
__physid_mask; \
|
|
})
|
|
|
|
#define physid_mask_of_physid(physid) \
|
|
({ \
|
|
physid_mask_t __physid_mask = PHYSID_MASK_NONE; \
|
|
physid_set(physid, __physid_mask); \
|
|
__physid_mask; \
|
|
})
|
|
|
|
#define PHYSID_MASK_ALL { {[0 ... PHYSID_ARRAY_SIZE-1] = ~0UL} }
|
|
#define PHYSID_MASK_NONE { {[0 ... PHYSID_ARRAY_SIZE-1] = 0UL} }
|
|
|
|
extern physid_mask_t phys_cpu_present_map;
|
|
|
|
/*
|
|
* Some lowlevel functions might want to know about
|
|
* the real CPU ID <-> CPU # mapping.
|
|
*/
|
|
extern volatile int physid_2_cpu[NR_CPUS];
|
|
extern volatile int cpu_2_physid[NR_CPUS];
|
|
#define physid_to_cpu(physid) physid_2_cpu[physid]
|
|
#define cpu_to_physid(cpu_id) cpu_2_physid[cpu_id]
|
|
|
|
#define raw_smp_processor_id() (current_thread_info()->cpu)
|
|
|
|
extern cpumask_t cpu_callout_map;
|
|
#define cpu_possible_map cpu_callout_map
|
|
|
|
static __inline__ int hard_smp_processor_id(void)
|
|
{
|
|
return (int)*(volatile long *)M32R_CPUID_PORTL;
|
|
}
|
|
|
|
static __inline__ int cpu_logical_map(int cpu)
|
|
{
|
|
return cpu;
|
|
}
|
|
|
|
static __inline__ int cpu_number_map(int cpu)
|
|
{
|
|
return cpu;
|
|
}
|
|
|
|
static __inline__ unsigned int num_booting_cpus(void)
|
|
{
|
|
return cpus_weight(cpu_callout_map);
|
|
}
|
|
|
|
extern void smp_send_timer(void);
|
|
extern unsigned long send_IPI_mask_phys(cpumask_t, int, int);
|
|
|
|
#endif /* not __ASSEMBLY__ */
|
|
|
|
#define NO_PROC_ID (0xff) /* No processor magic marker */
|
|
|
|
#define PROC_CHANGE_PENALTY (15) /* Schedule penalty */
|
|
|
|
/*
|
|
* M32R-mp IPI
|
|
*/
|
|
#define RESCHEDULE_IPI (M32R_IRQ_IPI0-M32R_IRQ_IPI0)
|
|
#define INVALIDATE_TLB_IPI (M32R_IRQ_IPI1-M32R_IRQ_IPI0)
|
|
#define CALL_FUNCTION_IPI (M32R_IRQ_IPI2-M32R_IRQ_IPI0)
|
|
#define LOCAL_TIMER_IPI (M32R_IRQ_IPI3-M32R_IRQ_IPI0)
|
|
#define INVALIDATE_CACHE_IPI (M32R_IRQ_IPI4-M32R_IRQ_IPI0)
|
|
#define CPU_BOOT_IPI (M32R_IRQ_IPI5-M32R_IRQ_IPI0)
|
|
|
|
#define IPI_SHIFT (0)
|
|
#define NR_IPIS (8)
|
|
|
|
#endif /* CONFIG_SMP */
|
|
|
|
#endif /* _ASM_M32R_SMP_H */
|