|  | #ifndef _ASM_M32R_SMP_H | 
|  | #define _ASM_M32R_SMP_H | 
|  |  | 
|  | #ifdef CONFIG_SMP | 
|  | #ifndef __ASSEMBLY__ | 
|  |  | 
|  | #include <linux/cpumask.h> | 
|  | #include <linux/spinlock.h> | 
|  | #include <linux/threads.h> | 
|  | #include <asm/m32r.h> | 
|  |  | 
|  | #define PHYSID_ARRAY_SIZE       1 | 
|  |  | 
|  | struct physid_mask | 
|  | { | 
|  | unsigned long mask[PHYSID_ARRAY_SIZE]; | 
|  | }; | 
|  |  | 
|  | typedef struct physid_mask physid_mask_t; | 
|  |  | 
|  | #define physid_set(physid, map)                 set_bit(physid, (map).mask) | 
|  | #define physid_clear(physid, map)               clear_bit(physid, (map).mask) | 
|  | #define physid_isset(physid, map)               test_bit(physid, (map).mask) | 
|  | #define physid_test_and_set(physid, map)        test_and_set_bit(physid, (map).mask) | 
|  |  | 
|  | #define physids_and(dst, src1, src2)            bitmap_and((dst).mask, (src1).mask, (src2).mask, MAX_APICS) | 
|  | #define physids_or(dst, src1, src2)             bitmap_or((dst).mask, (src1).mask, (src2).mask, MAX_APICS) | 
|  | #define physids_clear(map)                      bitmap_zero((map).mask, MAX_APICS) | 
|  | #define physids_complement(dst, src)            bitmap_complement((dst).mask,(src).mask, MAX_APICS) | 
|  | #define physids_empty(map)                      bitmap_empty((map).mask, MAX_APICS) | 
|  | #define physids_equal(map1, map2)               bitmap_equal((map1).mask, (map2).mask, MAX_APICS) | 
|  | #define physids_weight(map)                     bitmap_weight((map).mask, MAX_APICS) | 
|  | #define physids_shift_right(d, s, n)            bitmap_shift_right((d).mask, (s).mask, n, MAX_APICS) | 
|  | #define physids_shift_left(d, s, n)             bitmap_shift_left((d).mask, (s).mask, n, MAX_APICS) | 
|  | #define physids_coerce(map)                     ((map).mask[0]) | 
|  |  | 
|  | #define physids_promote(physids)					\ | 
|  | ({								\ | 
|  | physid_mask_t __physid_mask = PHYSID_MASK_NONE;		\ | 
|  | __physid_mask.mask[0] = physids;			\ | 
|  | __physid_mask;						\ | 
|  | }) | 
|  |  | 
|  | #define physid_mask_of_physid(physid)					\ | 
|  | ({								\ | 
|  | physid_mask_t __physid_mask = PHYSID_MASK_NONE;		\ | 
|  | physid_set(physid, __physid_mask);			\ | 
|  | __physid_mask;						\ | 
|  | }) | 
|  |  | 
|  | #define PHYSID_MASK_ALL         { {[0 ... PHYSID_ARRAY_SIZE-1] = ~0UL} } | 
|  | #define PHYSID_MASK_NONE        { {[0 ... PHYSID_ARRAY_SIZE-1] = 0UL} } | 
|  |  | 
|  | extern physid_mask_t phys_cpu_present_map; | 
|  |  | 
|  | /* | 
|  | * Some lowlevel functions might want to know about | 
|  | * the real CPU ID <-> CPU # mapping. | 
|  | */ | 
|  | extern volatile int cpu_2_physid[NR_CPUS]; | 
|  | #define cpu_to_physid(cpu_id)	cpu_2_physid[cpu_id] | 
|  |  | 
|  | #define raw_smp_processor_id()	(current_thread_info()->cpu) | 
|  |  | 
|  | extern cpumask_t cpu_callout_map; | 
|  | extern cpumask_t cpu_possible_map; | 
|  | extern cpumask_t cpu_present_map; | 
|  |  | 
|  | static __inline__ int hard_smp_processor_id(void) | 
|  | { | 
|  | return (int)*(volatile long *)M32R_CPUID_PORTL; | 
|  | } | 
|  |  | 
|  | static __inline__ int cpu_logical_map(int cpu) | 
|  | { | 
|  | return cpu; | 
|  | } | 
|  |  | 
|  | static __inline__ int cpu_number_map(int cpu) | 
|  | { | 
|  | return cpu; | 
|  | } | 
|  |  | 
|  | static __inline__ unsigned int num_booting_cpus(void) | 
|  | { | 
|  | return cpus_weight(cpu_callout_map); | 
|  | } | 
|  |  | 
|  | extern void smp_send_timer(void); | 
|  | extern unsigned long send_IPI_mask_phys(cpumask_t, int, int); | 
|  |  | 
|  | extern void arch_send_call_function_single_ipi(int cpu); | 
|  | extern void arch_send_call_function_ipi(cpumask_t mask); | 
|  |  | 
|  | #endif	/* not __ASSEMBLY__ */ | 
|  |  | 
|  | #define NO_PROC_ID (0xff)	/* No processor magic marker */ | 
|  |  | 
|  | #define PROC_CHANGE_PENALTY	(15)	/* Schedule penalty */ | 
|  |  | 
|  | /* | 
|  | * M32R-mp IPI | 
|  | */ | 
|  | #define RESCHEDULE_IPI		(M32R_IRQ_IPI0-M32R_IRQ_IPI0) | 
|  | #define INVALIDATE_TLB_IPI	(M32R_IRQ_IPI1-M32R_IRQ_IPI0) | 
|  | #define CALL_FUNCTION_IPI	(M32R_IRQ_IPI2-M32R_IRQ_IPI0) | 
|  | #define LOCAL_TIMER_IPI		(M32R_IRQ_IPI3-M32R_IRQ_IPI0) | 
|  | #define INVALIDATE_CACHE_IPI	(M32R_IRQ_IPI4-M32R_IRQ_IPI0) | 
|  | #define CPU_BOOT_IPI		(M32R_IRQ_IPI5-M32R_IRQ_IPI0) | 
|  | #define CALL_FUNC_SINGLE_IPI	(M32R_IRQ_IPI6-M32R_IRQ_IPI0) | 
|  |  | 
|  | #define IPI_SHIFT	(0) | 
|  | #define NR_IPIS		(8) | 
|  |  | 
|  | #else	/* CONFIG_SMP */ | 
|  |  | 
|  | #define hard_smp_processor_id()		0 | 
|  |  | 
|  | #endif /* CONFIG_SMP */ | 
|  |  | 
|  | #endif	/* _ASM_M32R_SMP_H */ |