| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 1 | /* | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 2 |  *  This program is free software; you can distribute it and/or modify it | 
 | 3 |  *  under the terms of the GNU General Public License (Version 2) as | 
 | 4 |  *  published by the Free Software Foundation. | 
 | 5 |  * | 
 | 6 |  *  This program is distributed in the hope it will be useful, but WITHOUT | 
 | 7 |  *  ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | 
 | 8 |  *  FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License | 
 | 9 |  *  for more details. | 
 | 10 |  * | 
 | 11 |  *  You should have received a copy of the GNU General Public License along | 
 | 12 |  *  with this program; if not, write to the Free Software Foundation, Inc., | 
 | 13 |  *  59 Temple Place - Suite 330, Boston MA 02111-1307, USA. | 
 | 14 |  * | 
| Ralf Baechle | 41c594a | 2006-04-05 09:45:45 +0100 | [diff] [blame] | 15 |  * Copyright (C) 2004, 05, 06 MIPS Technologies, Inc. | 
 | 16 |  *    Elizabeth Clarke (beth@mips.com) | 
 | 17 |  *    Ralf Baechle (ralf@linux-mips.org) | 
 | 18 |  * Copyright (C) 2006 Ralf Baechle (ralf@linux-mips.org) | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 19 |  */ | 
 | 20 | #include <linux/kernel.h> | 
 | 21 | #include <linux/sched.h> | 
 | 22 | #include <linux/cpumask.h> | 
 | 23 | #include <linux/interrupt.h> | 
 | 24 | #include <linux/compiler.h> | 
| Ralf Baechle | 0ab7aef | 2007-03-02 20:42:04 +0000 | [diff] [blame] | 25 | #include <linux/smp.h> | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 26 |  | 
 | 27 | #include <asm/atomic.h> | 
| Ralf Baechle | 41c594a | 2006-04-05 09:45:45 +0100 | [diff] [blame] | 28 | #include <asm/cacheflush.h> | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 29 | #include <asm/cpu.h> | 
 | 30 | #include <asm/processor.h> | 
 | 31 | #include <asm/system.h> | 
 | 32 | #include <asm/hardirq.h> | 
 | 33 | #include <asm/mmu_context.h> | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 34 | #include <asm/time.h> | 
 | 35 | #include <asm/mipsregs.h> | 
 | 36 | #include <asm/mipsmtregs.h> | 
| Ralf Baechle | 41c594a | 2006-04-05 09:45:45 +0100 | [diff] [blame] | 37 | #include <asm/mips_mt.h> | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 38 |  | 
 | 39 | #define MIPS_CPU_IPI_RESCHED_IRQ 0 | 
 | 40 | #define MIPS_CPU_IPI_CALL_IRQ 1 | 
 | 41 |  | 
 | 42 | static int cpu_ipi_resched_irq, cpu_ipi_call_irq; | 
 | 43 |  | 
 | 44 | #if 0 | 
 | 45 | static void dump_mtregisters(int vpe, int tc) | 
 | 46 | { | 
 | 47 | 	printk("vpe %d tc %d\n", vpe, tc); | 
 | 48 |  | 
 | 49 | 	settc(tc); | 
 | 50 |  | 
 | 51 | 	printk("  c0 status  0x%lx\n", read_vpe_c0_status()); | 
 | 52 | 	printk("  vpecontrol 0x%lx\n", read_vpe_c0_vpecontrol()); | 
 | 53 | 	printk("  vpeconf0    0x%lx\n", read_vpe_c0_vpeconf0()); | 
 | 54 | 	printk("  tcstatus 0x%lx\n", read_tc_c0_tcstatus()); | 
 | 55 | 	printk("  tcrestart 0x%lx\n", read_tc_c0_tcrestart()); | 
 | 56 | 	printk("  tcbind 0x%lx\n", read_tc_c0_tcbind()); | 
 | 57 | 	printk("  tchalt 0x%lx\n", read_tc_c0_tchalt()); | 
 | 58 | } | 
 | 59 | #endif | 
 | 60 |  | 
 | 61 | void __init sanitize_tlb_entries(void) | 
 | 62 | { | 
 | 63 | 	int i, tlbsiz; | 
 | 64 | 	unsigned long mvpconf0, ncpu; | 
 | 65 |  | 
 | 66 | 	if (!cpu_has_mipsmt) | 
 | 67 | 		return; | 
 | 68 |  | 
| Ralf Baechle | 41c594a | 2006-04-05 09:45:45 +0100 | [diff] [blame] | 69 | 	/* Enable VPC */ | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 70 | 	set_c0_mvpcontrol(MVPCONTROL_VPC); | 
 | 71 |  | 
| Ralf Baechle | fdc9bb1 | 2006-02-10 14:25:16 +0000 | [diff] [blame] | 72 | 	back_to_back_c0_hazard(); | 
 | 73 |  | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 74 | 	/* Disable TLB sharing */ | 
 | 75 | 	clear_c0_mvpcontrol(MVPCONTROL_STLB); | 
 | 76 |  | 
 | 77 | 	mvpconf0 = read_c0_mvpconf0(); | 
 | 78 |  | 
 | 79 | 	printk(KERN_INFO "MVPConf0 0x%lx TLBS %lx PTLBE %ld\n", mvpconf0, | 
 | 80 | 		   (mvpconf0 & MVPCONF0_TLBS) >> MVPCONF0_TLBS_SHIFT, | 
 | 81 | 			   (mvpconf0 & MVPCONF0_PTLBE) >> MVPCONF0_PTLBE_SHIFT); | 
 | 82 |  | 
 | 83 | 	tlbsiz = (mvpconf0 & MVPCONF0_PTLBE) >> MVPCONF0_PTLBE_SHIFT; | 
 | 84 | 	ncpu = ((mvpconf0 & MVPCONF0_PVPE) >> MVPCONF0_PVPE_SHIFT) + 1; | 
 | 85 |  | 
 | 86 | 	printk(" tlbsiz %d ncpu %ld\n", tlbsiz, ncpu); | 
 | 87 |  | 
 | 88 | 	if (tlbsiz > 0) { | 
 | 89 | 		/* share them out across the vpe's */ | 
 | 90 | 		tlbsiz /= ncpu; | 
 | 91 |  | 
 | 92 | 		printk(KERN_INFO "setting Config1.MMU_size to %d\n", tlbsiz); | 
 | 93 |  | 
 | 94 | 		for (i = 0; i < ncpu; i++) { | 
 | 95 | 			settc(i); | 
 | 96 |  | 
 | 97 | 			if (i == 0) | 
 | 98 | 				write_c0_config1((read_c0_config1() & ~(0x3f << 25)) | (tlbsiz << 25)); | 
 | 99 | 			else | 
 | 100 | 				write_vpe_c0_config1((read_vpe_c0_config1() & ~(0x3f << 25)) | | 
 | 101 | 						   (tlbsiz << 25)); | 
 | 102 | 		} | 
 | 103 | 	} | 
 | 104 |  | 
 | 105 | 	clear_c0_mvpcontrol(MVPCONTROL_VPC); | 
 | 106 | } | 
 | 107 |  | 
| Ralf Baechle | 937a801 | 2006-10-07 19:44:33 +0100 | [diff] [blame] | 108 | static void ipi_resched_dispatch(void) | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 109 | { | 
| Atsushi Nemoto | 97dcb82 | 2007-01-08 02:14:29 +0900 | [diff] [blame] | 110 | 	do_IRQ(MIPS_CPU_IRQ_BASE + MIPS_CPU_IPI_RESCHED_IRQ); | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 111 | } | 
 | 112 |  | 
| Ralf Baechle | 937a801 | 2006-10-07 19:44:33 +0100 | [diff] [blame] | 113 | static void ipi_call_dispatch(void) | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 114 | { | 
| Atsushi Nemoto | 97dcb82 | 2007-01-08 02:14:29 +0900 | [diff] [blame] | 115 | 	do_IRQ(MIPS_CPU_IRQ_BASE + MIPS_CPU_IPI_CALL_IRQ); | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 116 | } | 
 | 117 |  | 
| Ralf Baechle | 937a801 | 2006-10-07 19:44:33 +0100 | [diff] [blame] | 118 | static irqreturn_t ipi_resched_interrupt(int irq, void *dev_id) | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 119 | { | 
 | 120 | 	return IRQ_HANDLED; | 
 | 121 | } | 
 | 122 |  | 
| Ralf Baechle | 937a801 | 2006-10-07 19:44:33 +0100 | [diff] [blame] | 123 | static irqreturn_t ipi_call_interrupt(int irq, void *dev_id) | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 124 | { | 
 | 125 | 	smp_call_function_interrupt(); | 
 | 126 |  | 
 | 127 | 	return IRQ_HANDLED; | 
 | 128 | } | 
 | 129 |  | 
 | 130 | static struct irqaction irq_resched = { | 
 | 131 | 	.handler	= ipi_resched_interrupt, | 
| Chris Dearman | ffe9ee4 | 2007-05-24 22:24:20 +0100 | [diff] [blame] | 132 | 	.flags		= IRQF_DISABLED|IRQF_PERCPU, | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 133 | 	.name		= "IPI_resched" | 
 | 134 | }; | 
 | 135 |  | 
 | 136 | static struct irqaction irq_call = { | 
 | 137 | 	.handler	= ipi_call_interrupt, | 
| Chris Dearman | ffe9ee4 | 2007-05-24 22:24:20 +0100 | [diff] [blame] | 138 | 	.flags		= IRQF_DISABLED|IRQF_PERCPU, | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 139 | 	.name		= "IPI_call" | 
 | 140 | }; | 
 | 141 |  | 
| Ralf Baechle | 781b0f8 | 2006-10-31 18:25:10 +0000 | [diff] [blame] | 142 | static void __init smp_copy_vpe_config(void) | 
 | 143 | { | 
 | 144 | 	write_vpe_c0_status( | 
 | 145 | 		(read_c0_status() & ~(ST0_IM | ST0_IE | ST0_KSU)) | ST0_CU0); | 
 | 146 |  | 
 | 147 | 	/* set config to be the same as vpe0, particularly kseg0 coherency alg */ | 
 | 148 | 	write_vpe_c0_config( read_c0_config()); | 
 | 149 |  | 
 | 150 | 	/* make sure there are no software interrupts pending */ | 
 | 151 | 	write_vpe_c0_cause(0); | 
 | 152 |  | 
 | 153 | 	/* Propagate Config7 */ | 
 | 154 | 	write_vpe_c0_config7(read_c0_config7()); | 
| Ralf Baechle | 70e46f4 | 2006-10-31 18:33:09 +0000 | [diff] [blame] | 155 |  | 
 | 156 | 	write_vpe_c0_count(read_c0_count()); | 
| Ralf Baechle | 781b0f8 | 2006-10-31 18:25:10 +0000 | [diff] [blame] | 157 | } | 
 | 158 |  | 
 | 159 | static unsigned int __init smp_vpe_init(unsigned int tc, unsigned int mvpconf0, | 
 | 160 | 	unsigned int ncpu) | 
 | 161 | { | 
 | 162 | 	if (tc > ((mvpconf0 & MVPCONF0_PVPE) >> MVPCONF0_PVPE_SHIFT)) | 
 | 163 | 		return ncpu; | 
 | 164 |  | 
 | 165 | 	/* Deactivate all but VPE 0 */ | 
 | 166 | 	if (tc != 0) { | 
 | 167 | 		unsigned long tmp = read_vpe_c0_vpeconf0(); | 
 | 168 |  | 
 | 169 | 		tmp &= ~VPECONF0_VPA; | 
 | 170 |  | 
 | 171 | 		/* master VPE */ | 
 | 172 | 		tmp |= VPECONF0_MVP; | 
 | 173 | 		write_vpe_c0_vpeconf0(tmp); | 
 | 174 |  | 
 | 175 | 		/* Record this as available CPU */ | 
 | 176 | 		cpu_set(tc, phys_cpu_present_map); | 
 | 177 | 		__cpu_number_map[tc]	= ++ncpu; | 
 | 178 | 		__cpu_logical_map[ncpu]	= tc; | 
 | 179 | 	} | 
 | 180 |  | 
 | 181 | 	/* Disable multi-threading with TC's */ | 
 | 182 | 	write_vpe_c0_vpecontrol(read_vpe_c0_vpecontrol() & ~VPECONTROL_TE); | 
 | 183 |  | 
 | 184 | 	if (tc != 0) | 
 | 185 | 		smp_copy_vpe_config(); | 
 | 186 |  | 
 | 187 | 	return ncpu; | 
 | 188 | } | 
 | 189 |  | 
 | 190 | static void __init smp_tc_init(unsigned int tc, unsigned int mvpconf0) | 
 | 191 | { | 
 | 192 | 	unsigned long tmp; | 
 | 193 |  | 
 | 194 | 	if (!tc) | 
 | 195 | 		return; | 
 | 196 |  | 
 | 197 | 	/* bind a TC to each VPE, May as well put all excess TC's | 
 | 198 | 	   on the last VPE */ | 
 | 199 | 	if (tc >= (((mvpconf0 & MVPCONF0_PVPE) >> MVPCONF0_PVPE_SHIFT)+1)) | 
 | 200 | 		write_tc_c0_tcbind(read_tc_c0_tcbind() | ((mvpconf0 & MVPCONF0_PVPE) >> MVPCONF0_PVPE_SHIFT)); | 
 | 201 | 	else { | 
 | 202 | 		write_tc_c0_tcbind(read_tc_c0_tcbind() | tc); | 
 | 203 |  | 
 | 204 | 		/* and set XTC */ | 
 | 205 | 		write_vpe_c0_vpeconf0(read_vpe_c0_vpeconf0() | (tc << VPECONF0_XTC_SHIFT)); | 
 | 206 | 	} | 
 | 207 |  | 
 | 208 | 	tmp = read_tc_c0_tcstatus(); | 
 | 209 |  | 
 | 210 | 	/* mark not allocated and not dynamically allocatable */ | 
 | 211 | 	tmp &= ~(TCSTATUS_A | TCSTATUS_DA); | 
 | 212 | 	tmp |= TCSTATUS_IXMT;		/* interrupt exempt */ | 
 | 213 | 	write_tc_c0_tcstatus(tmp); | 
 | 214 |  | 
 | 215 | 	write_tc_c0_tchalt(TCHALT_H); | 
 | 216 | } | 
 | 217 |  | 
| Ralf Baechle | 87353d8 | 2007-11-19 12:23:51 +0000 | [diff] [blame] | 218 | static void vsmp_send_ipi_single(int cpu, unsigned int action) | 
 | 219 | { | 
 | 220 | 	int i; | 
 | 221 | 	unsigned long flags; | 
 | 222 | 	int vpflags; | 
 | 223 |  | 
 | 224 | 	local_irq_save(flags); | 
 | 225 |  | 
 | 226 | 	vpflags = dvpe();	/* cant access the other CPU's registers whilst MVPE enabled */ | 
 | 227 |  | 
 | 228 | 	switch (action) { | 
 | 229 | 	case SMP_CALL_FUNCTION: | 
 | 230 | 		i = C_SW1; | 
 | 231 | 		break; | 
 | 232 |  | 
 | 233 | 	case SMP_RESCHEDULE_YOURSELF: | 
 | 234 | 	default: | 
 | 235 | 		i = C_SW0; | 
 | 236 | 		break; | 
 | 237 | 	} | 
 | 238 |  | 
 | 239 | 	/* 1:1 mapping of vpe and tc... */ | 
 | 240 | 	settc(cpu); | 
 | 241 | 	write_vpe_c0_cause(read_vpe_c0_cause() | i); | 
 | 242 | 	evpe(vpflags); | 
 | 243 |  | 
 | 244 | 	local_irq_restore(flags); | 
 | 245 | } | 
 | 246 |  | 
 | 247 | static void vsmp_send_ipi_mask(cpumask_t mask, unsigned int action) | 
 | 248 | { | 
 | 249 | 	unsigned int i; | 
 | 250 |  | 
 | 251 | 	for_each_cpu_mask(i, mask) | 
 | 252 | 		vsmp_send_ipi_single(i, action); | 
 | 253 | } | 
 | 254 |  | 
 | 255 | static void __cpuinit vsmp_init_secondary(void) | 
 | 256 | { | 
 | 257 | 	/* Enable per-cpu interrupts */ | 
 | 258 |  | 
 | 259 | 	/* This is Malta specific: IPI,performance and timer inetrrupts */ | 
 | 260 | 	write_c0_status((read_c0_status() & ~ST0_IM ) | | 
 | 261 | 	                (STATUSF_IP0 | STATUSF_IP1 | STATUSF_IP6 | STATUSF_IP7)); | 
 | 262 | } | 
 | 263 |  | 
 | 264 | static void __cpuinit vsmp_smp_finish(void) | 
 | 265 | { | 
 | 266 | 	write_c0_compare(read_c0_count() + (8* mips_hpt_frequency/HZ)); | 
 | 267 |  | 
 | 268 | #ifdef CONFIG_MIPS_MT_FPAFF | 
 | 269 | 	/* If we have an FPU, enroll ourselves in the FPU-full mask */ | 
 | 270 | 	if (cpu_has_fpu) | 
 | 271 | 		cpu_set(smp_processor_id(), mt_fpu_cpumask); | 
 | 272 | #endif /* CONFIG_MIPS_MT_FPAFF */ | 
 | 273 |  | 
 | 274 | 	local_irq_enable(); | 
 | 275 | } | 
 | 276 |  | 
 | 277 | static void vsmp_cpus_done(void) | 
 | 278 | { | 
 | 279 | } | 
 | 280 |  | 
 | 281 | /* | 
 | 282 |  * Setup the PC, SP, and GP of a secondary processor and start it | 
 | 283 |  * running! | 
 | 284 |  * smp_bootstrap is the place to resume from | 
 | 285 |  * __KSTK_TOS(idle) is apparently the stack pointer | 
 | 286 |  * (unsigned long)idle->thread_info the gp | 
 | 287 |  * assumes a 1:1 mapping of TC => VPE | 
 | 288 |  */ | 
 | 289 | static void __cpuinit vsmp_boot_secondary(int cpu, struct task_struct *idle) | 
 | 290 | { | 
 | 291 | 	struct thread_info *gp = task_thread_info(idle); | 
 | 292 | 	dvpe(); | 
 | 293 | 	set_c0_mvpcontrol(MVPCONTROL_VPC); | 
 | 294 |  | 
 | 295 | 	settc(cpu); | 
 | 296 |  | 
 | 297 | 	/* restart */ | 
 | 298 | 	write_tc_c0_tcrestart((unsigned long)&smp_bootstrap); | 
 | 299 |  | 
 | 300 | 	/* enable the tc this vpe/cpu will be running */ | 
 | 301 | 	write_tc_c0_tcstatus((read_tc_c0_tcstatus() & ~TCSTATUS_IXMT) | TCSTATUS_A); | 
 | 302 |  | 
 | 303 | 	write_tc_c0_tchalt(0); | 
 | 304 |  | 
 | 305 | 	/* enable the VPE */ | 
 | 306 | 	write_vpe_c0_vpeconf0(read_vpe_c0_vpeconf0() | VPECONF0_VPA); | 
 | 307 |  | 
 | 308 | 	/* stack pointer */ | 
 | 309 | 	write_tc_gpr_sp( __KSTK_TOS(idle)); | 
 | 310 |  | 
 | 311 | 	/* global pointer */ | 
 | 312 | 	write_tc_gpr_gp((unsigned long)gp); | 
 | 313 |  | 
 | 314 | 	flush_icache_range((unsigned long)gp, | 
 | 315 | 	                   (unsigned long)(gp + sizeof(struct thread_info))); | 
 | 316 |  | 
 | 317 | 	/* finally out of configuration and into chaos */ | 
 | 318 | 	clear_c0_mvpcontrol(MVPCONTROL_VPC); | 
 | 319 |  | 
 | 320 | 	evpe(EVPE_ENABLE); | 
 | 321 | } | 
 | 322 |  | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 323 | /* | 
 | 324 |  * Common setup before any secondaries are started | 
 | 325 |  * Make sure all CPU's are in a sensible state before we boot any of the | 
 | 326 |  * secondarys | 
 | 327 |  */ | 
| Ralf Baechle | 87353d8 | 2007-11-19 12:23:51 +0000 | [diff] [blame] | 328 | static void __init vsmp_smp_setup(void) | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 329 | { | 
| Ralf Baechle | 781b0f8 | 2006-10-31 18:25:10 +0000 | [diff] [blame] | 330 | 	unsigned int mvpconf0, ntc, tc, ncpu = 0; | 
| Ralf Baechle | 0ab7aef | 2007-03-02 20:42:04 +0000 | [diff] [blame] | 331 | 	unsigned int nvpe; | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 332 |  | 
| Ralf Baechle | f088fc8 | 2006-04-05 09:45:47 +0100 | [diff] [blame] | 333 | #ifdef CONFIG_MIPS_MT_FPAFF | 
 | 334 | 	/* If we have an FPU, enroll ourselves in the FPU-full mask */ | 
 | 335 | 	if (cpu_has_fpu) | 
 | 336 | 		cpu_set(0, mt_fpu_cpumask); | 
 | 337 | #endif /* CONFIG_MIPS_MT_FPAFF */ | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 338 | 	if (!cpu_has_mipsmt) | 
 | 339 | 		return; | 
 | 340 |  | 
 | 341 | 	/* disable MT so we can configure */ | 
 | 342 | 	dvpe(); | 
 | 343 | 	dmt(); | 
 | 344 |  | 
 | 345 | 	/* Put MVPE's into 'configuration state' */ | 
 | 346 | 	set_c0_mvpcontrol(MVPCONTROL_VPC); | 
 | 347 |  | 
| Ralf Baechle | 781b0f8 | 2006-10-31 18:25:10 +0000 | [diff] [blame] | 348 | 	mvpconf0 = read_c0_mvpconf0(); | 
 | 349 | 	ntc = (mvpconf0 & MVPCONF0_PTC) >> MVPCONF0_PTC_SHIFT; | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 350 |  | 
| Ralf Baechle | 0ab7aef | 2007-03-02 20:42:04 +0000 | [diff] [blame] | 351 | 	nvpe = ((mvpconf0 & MVPCONF0_PVPE) >> MVPCONF0_PVPE_SHIFT) + 1; | 
 | 352 | 	smp_num_siblings = nvpe; | 
 | 353 |  | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 354 | 	/* we'll always have more TC's than VPE's, so loop setting everything | 
 | 355 | 	   to a sensible state */ | 
| Ralf Baechle | 781b0f8 | 2006-10-31 18:25:10 +0000 | [diff] [blame] | 356 | 	for (tc = 0; tc <= ntc; tc++) { | 
 | 357 | 		settc(tc); | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 358 |  | 
| Ralf Baechle | 781b0f8 | 2006-10-31 18:25:10 +0000 | [diff] [blame] | 359 | 		smp_tc_init(tc, mvpconf0); | 
 | 360 | 		ncpu = smp_vpe_init(tc, mvpconf0, ncpu); | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 361 | 	} | 
 | 362 |  | 
 | 363 | 	/* Release config state */ | 
 | 364 | 	clear_c0_mvpcontrol(MVPCONTROL_VPC); | 
 | 365 |  | 
 | 366 | 	/* We'll wait until starting the secondaries before starting MVPE */ | 
 | 367 |  | 
| Ralf Baechle | 781b0f8 | 2006-10-31 18:25:10 +0000 | [diff] [blame] | 368 | 	printk(KERN_INFO "Detected %i available secondary CPU(s)\n", ncpu); | 
| Ralf Baechle | 41c594a | 2006-04-05 09:45:45 +0100 | [diff] [blame] | 369 | } | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 370 |  | 
| Ralf Baechle | 87353d8 | 2007-11-19 12:23:51 +0000 | [diff] [blame] | 371 | static void __init vsmp_prepare_cpus(unsigned int max_cpus) | 
| Ralf Baechle | 41c594a | 2006-04-05 09:45:45 +0100 | [diff] [blame] | 372 | { | 
| Ralf Baechle | 8c976e3 | 2007-07-03 18:25:58 +0200 | [diff] [blame] | 373 | 	mips_mt_set_cpuoptions(); | 
 | 374 |  | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 375 | 	/* set up ipi interrupts */ | 
 | 376 | 	if (cpu_has_vint) { | 
| Ralf Baechle | 937a801 | 2006-10-07 19:44:33 +0100 | [diff] [blame] | 377 | 		set_vi_handler(MIPS_CPU_IPI_RESCHED_IRQ, ipi_resched_dispatch); | 
 | 378 | 		set_vi_handler(MIPS_CPU_IPI_CALL_IRQ, ipi_call_dispatch); | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 379 | 	} | 
 | 380 |  | 
| Atsushi Nemoto | 97dcb82 | 2007-01-08 02:14:29 +0900 | [diff] [blame] | 381 | 	cpu_ipi_resched_irq = MIPS_CPU_IRQ_BASE + MIPS_CPU_IPI_RESCHED_IRQ; | 
 | 382 | 	cpu_ipi_call_irq = MIPS_CPU_IRQ_BASE + MIPS_CPU_IPI_CALL_IRQ; | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 383 |  | 
 | 384 | 	setup_irq(cpu_ipi_resched_irq, &irq_resched); | 
 | 385 | 	setup_irq(cpu_ipi_call_irq, &irq_call); | 
 | 386 |  | 
| Atsushi Nemoto | 1417836 | 2006-11-14 01:13:18 +0900 | [diff] [blame] | 387 | 	set_irq_handler(cpu_ipi_resched_irq, handle_percpu_irq); | 
| Atsushi Nemoto | 1417836 | 2006-11-14 01:13:18 +0900 | [diff] [blame] | 388 | 	set_irq_handler(cpu_ipi_call_irq, handle_percpu_irq); | 
| Ralf Baechle | 340ee4b | 2005-08-17 17:44:08 +0000 | [diff] [blame] | 389 | } | 
 | 390 |  | 
| Ralf Baechle | 87353d8 | 2007-11-19 12:23:51 +0000 | [diff] [blame] | 391 | struct plat_smp_ops vsmp_smp_ops = { | 
 | 392 | 	.send_ipi_single	= vsmp_send_ipi_single, | 
 | 393 | 	.send_ipi_mask		= vsmp_send_ipi_mask, | 
 | 394 | 	.init_secondary		= vsmp_init_secondary, | 
 | 395 | 	.smp_finish		= vsmp_smp_finish, | 
 | 396 | 	.cpus_done		= vsmp_cpus_done, | 
 | 397 | 	.boot_secondary		= vsmp_boot_secondary, | 
 | 398 | 	.smp_setup		= vsmp_smp_setup, | 
 | 399 | 	.prepare_cpus		= vsmp_prepare_cpus, | 
 | 400 | }; |