xref: /titanic_41/usr/src/uts/i86pc/os/x_call.c (revision 49d3bc91e27cd871b950d56c01398fa2f2e12ab4)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2009 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 /*
26  * Copyright (c) 2010, Intel Corporation.
27  * All rights reserved.
28  */
29 
30 #include <sys/types.h>
31 #include <sys/param.h>
32 #include <sys/t_lock.h>
33 #include <sys/thread.h>
34 #include <sys/cpuvar.h>
35 #include <sys/x_call.h>
36 #include <sys/xc_levels.h>
37 #include <sys/cpu.h>
38 #include <sys/psw.h>
39 #include <sys/sunddi.h>
40 #include <sys/debug.h>
41 #include <sys/systm.h>
42 #include <sys/archsystm.h>
43 #include <sys/machsystm.h>
44 #include <sys/mutex_impl.h>
45 #include <sys/stack.h>
46 #include <sys/promif.h>
47 #include <sys/x86_archext.h>
48 
49 /*
50  * Implementation for cross-processor calls via interprocessor interrupts
51  *
52  * This implementation uses a message passing architecture to allow multiple
53  * concurrent cross calls to be in flight at any given time. We use the cmpxchg
54  * instruction, aka casptr(), to implement simple efficient work queues for
55  * message passing between CPUs with almost no need for regular locking.
56  * See xc_extract() and xc_insert() below.
57  *
58  * The general idea is that initiating a cross call means putting a message
59  * on a target(s) CPU's work queue. Any synchronization is handled by passing
60  * the message back and forth between initiator and target(s).
61  *
62  * Every CPU has xc_work_cnt, which indicates it has messages to process.
63  * This value is incremented as message traffic is initiated and decremented
64  * with every message that finishes all processing.
65  *
66  * The code needs no mfence or other membar_*() calls. The uses of
67  * casptr(), cas32() and atomic_dec_32() for the message passing are
68  * implemented with LOCK prefix instructions which are equivalent to mfence.
69  *
70  * One interesting aspect of this implmentation is that it allows 2 or more
71  * CPUs to initiate cross calls to intersecting sets of CPUs at the same time.
72  * The cross call processing by the CPUs will happen in any order with only
73  * a guarantee, for xc_call() and xc_sync(), that an initiator won't return
74  * from cross calls before all slaves have invoked the function.
75  *
76  * The reason for this asynchronous approach is to allow for fast global
77  * TLB shootdowns. If all CPUs, say N, tried to do a global TLB invalidation
78  * on a different Virtual Address at the same time. The old code required
79  * N squared IPIs. With this method, depending on timing, it could happen
80  * with just N IPIs.
81  */
82 
83 /*
84  * The default is to not enable collecting counts of IPI information, since
85  * the updating of shared cachelines could cause excess bus traffic.
86  */
87 uint_t xc_collect_enable = 0;
88 uint64_t xc_total_cnt = 0;	/* total #IPIs sent for cross calls */
89 uint64_t xc_multi_cnt = 0;	/* # times we piggy backed on another IPI */
90 
91 /*
92  * Values for message states. Here are the normal transitions. A transition
93  * of "->" happens in the slave cpu and "=>" happens in the master cpu as
94  * the messages are passed back and forth.
95  *
96  * FREE => ASYNC ->                       DONE => FREE
97  * FREE => CALL ->                        DONE => FREE
98  * FREE => SYNC -> WAITING => RELEASED -> DONE => FREE
99  *
100  * The interesing one above is ASYNC. You might ask, why not go directly
101  * to FREE, instead of DONE. If it did that, it might be possible to exhaust
102  * the master's xc_free list if a master can generate ASYNC messages faster
103  * then the slave can process them. That could be handled with more complicated
104  * handling. However since nothing important uses ASYNC, I've not bothered.
105  */
106 #define	XC_MSG_FREE	(0)	/* msg in xc_free queue */
107 #define	XC_MSG_ASYNC	(1)	/* msg in slave xc_msgbox */
108 #define	XC_MSG_CALL	(2)	/* msg in slave xc_msgbox */
109 #define	XC_MSG_SYNC	(3)	/* msg in slave xc_msgbox */
110 #define	XC_MSG_WAITING	(4)	/* msg in master xc_msgbox or xc_waiters */
111 #define	XC_MSG_RELEASED	(5)	/* msg in slave xc_msgbox */
112 #define	XC_MSG_DONE	(6)	/* msg in master xc_msgbox */
113 
114 /*
115  * We allow for one high priority message at a time to happen in the system.
116  * This is used for panic, kmdb, etc., so no locking is done.
117  */
118 static volatile cpuset_t xc_priority_set_store;
119 static volatile ulong_t *xc_priority_set = CPUSET2BV(xc_priority_set_store);
120 static xc_data_t xc_priority_data;
121 
122 /*
123  * Wrappers to avoid C compiler warnings due to volatile. The atomic bit
124  * operations don't accept volatile bit vectors - which is a bit silly.
125  */
126 #define	XC_BT_SET(vector, b)	BT_ATOMIC_SET((ulong_t *)(vector), (b))
127 #define	XC_BT_CLEAR(vector, b)	BT_ATOMIC_CLEAR((ulong_t *)(vector), (b))
128 
129 /*
130  * Decrement a CPU's work count
131  */
132 static void
133 xc_decrement(struct machcpu *mcpu)
134 {
135 	atomic_dec_32(&mcpu->xc_work_cnt);
136 }
137 
138 /*
139  * Increment a CPU's work count and return the old value
140  */
141 static int
142 xc_increment(struct machcpu *mcpu)
143 {
144 	int old;
145 	do {
146 		old = mcpu->xc_work_cnt;
147 	} while (cas32((uint32_t *)&mcpu->xc_work_cnt, old, old + 1) != old);
148 	return (old);
149 }
150 
151 /*
152  * Put a message into a queue. The insertion is atomic no matter
153  * how many different inserts/extracts to the same queue happen.
154  */
155 static void
156 xc_insert(void *queue, xc_msg_t *msg)
157 {
158 	xc_msg_t *old_head;
159 
160 	/*
161 	 * FREE messages should only ever be getting inserted into
162 	 * the xc_master CPUs xc_free queue.
163 	 */
164 	ASSERT(msg->xc_command != XC_MSG_FREE ||
165 	    cpu[msg->xc_master] == NULL || /* possible only during init */
166 	    queue == &cpu[msg->xc_master]->cpu_m.xc_free);
167 
168 	do {
169 		old_head = (xc_msg_t *)*(volatile xc_msg_t **)queue;
170 		msg->xc_next = old_head;
171 	} while (casptr(queue, old_head, msg) != old_head);
172 }
173 
174 /*
175  * Extract a message from a queue. The extraction is atomic only
176  * when just one thread does extractions from the queue.
177  * If the queue is empty, NULL is returned.
178  */
179 static xc_msg_t *
180 xc_extract(xc_msg_t **queue)
181 {
182 	xc_msg_t *old_head;
183 
184 	do {
185 		old_head = (xc_msg_t *)*(volatile xc_msg_t **)queue;
186 		if (old_head == NULL)
187 			return (old_head);
188 	} while (casptr(queue, old_head, old_head->xc_next) != old_head);
189 	old_head->xc_next = NULL;
190 	return (old_head);
191 }
192 
193 /*
194  * Initialize the machcpu fields used for cross calls
195  */
196 static uint_t xc_initialized = 0;
197 
198 void
199 xc_init_cpu(struct cpu *cpup)
200 {
201 	xc_msg_t *msg;
202 	int c;
203 
204 	/*
205 	 * Allocate message buffers for the new CPU.
206 	 */
207 	for (c = 0; c < max_ncpus; ++c) {
208 		if (plat_dr_support_cpu()) {
209 			/*
210 			 * Allocate a message buffer for every CPU possible
211 			 * in system, including our own, and add them to our xc
212 			 * message queue.
213 			 */
214 			msg = kmem_zalloc(sizeof (*msg), KM_SLEEP);
215 			msg->xc_command = XC_MSG_FREE;
216 			msg->xc_master = cpup->cpu_id;
217 			xc_insert(&cpup->cpu_m.xc_free, msg);
218 		} else if (cpu[c] != NULL && cpu[c] != cpup) {
219 			/*
220 			 * Add a new message buffer to each existing CPU's free
221 			 * list, as well as one for my list for each of them.
222 			 * Note: cpu0 is statically inserted into cpu[] array,
223 			 * so need to check cpu[c] isn't cpup itself to avoid
224 			 * allocating extra message buffers for cpu0.
225 			 */
226 			msg = kmem_zalloc(sizeof (*msg), KM_SLEEP);
227 			msg->xc_command = XC_MSG_FREE;
228 			msg->xc_master = c;
229 			xc_insert(&cpu[c]->cpu_m.xc_free, msg);
230 
231 			msg = kmem_zalloc(sizeof (*msg), KM_SLEEP);
232 			msg->xc_command = XC_MSG_FREE;
233 			msg->xc_master = cpup->cpu_id;
234 			xc_insert(&cpup->cpu_m.xc_free, msg);
235 		}
236 	}
237 
238 	if (!plat_dr_support_cpu()) {
239 		/*
240 		 * Add one for self messages if CPU hotplug is disabled.
241 		 */
242 		msg = kmem_zalloc(sizeof (*msg), KM_SLEEP);
243 		msg->xc_command = XC_MSG_FREE;
244 		msg->xc_master = cpup->cpu_id;
245 		xc_insert(&cpup->cpu_m.xc_free, msg);
246 	}
247 
248 	if (!xc_initialized)
249 		xc_initialized = 1;
250 }
251 
252 void
253 xc_fini_cpu(struct cpu *cpup)
254 {
255 	xc_msg_t *msg;
256 
257 	ASSERT((cpup->cpu_flags & CPU_READY) == 0);
258 	ASSERT(cpup->cpu_m.xc_msgbox == NULL);
259 	ASSERT(cpup->cpu_m.xc_work_cnt == 0);
260 
261 	while ((msg = xc_extract(&cpup->cpu_m.xc_free)) != NULL) {
262 		kmem_free(msg, sizeof (*msg));
263 	}
264 }
265 
266 #define	XC_FLUSH_MAX_WAITS		1000
267 
268 /* Flush inflight message buffers. */
269 int
270 xc_flush_cpu(struct cpu *cpup)
271 {
272 	int i;
273 
274 	ASSERT((cpup->cpu_flags & CPU_READY) == 0);
275 
276 	/*
277 	 * Pause all working CPUs, which ensures that there's no CPU in
278 	 * function xc_common().
279 	 * This is used to work around a race condition window in xc_common()
280 	 * between checking CPU_READY flag and increasing working item count.
281 	 */
282 	pause_cpus(cpup);
283 	start_cpus();
284 
285 	for (i = 0; i < XC_FLUSH_MAX_WAITS; i++) {
286 		if (cpup->cpu_m.xc_work_cnt == 0) {
287 			break;
288 		}
289 		DELAY(1);
290 	}
291 	for (; i < XC_FLUSH_MAX_WAITS; i++) {
292 		if (!BT_TEST(xc_priority_set, cpup->cpu_id)) {
293 			break;
294 		}
295 		DELAY(1);
296 	}
297 
298 	return (i >= XC_FLUSH_MAX_WAITS ? ETIME : 0);
299 }
300 
301 /*
302  * X-call message processing routine. Note that this is used by both
303  * senders and recipients of messages.
304  *
305  * We're protected against changing CPUs by either being in a high-priority
306  * interrupt, having preemption disabled or by having a raised SPL.
307  */
308 /*ARGSUSED*/
309 uint_t
310 xc_serv(caddr_t arg1, caddr_t arg2)
311 {
312 	struct machcpu *mcpup = &(CPU->cpu_m);
313 	xc_msg_t *msg;
314 	xc_data_t *data;
315 	xc_msg_t *xc_waiters = NULL;
316 	uint32_t num_waiting = 0;
317 	xc_func_t func;
318 	xc_arg_t a1;
319 	xc_arg_t a2;
320 	xc_arg_t a3;
321 	uint_t rc = DDI_INTR_UNCLAIMED;
322 
323 	while (mcpup->xc_work_cnt != 0) {
324 		rc = DDI_INTR_CLAIMED;
325 
326 		/*
327 		 * We may have to wait for a message to arrive.
328 		 */
329 		for (msg = NULL; msg == NULL;
330 		    msg = xc_extract(&mcpup->xc_msgbox)) {
331 
332 			/*
333 			 * Alway check for and handle a priority message.
334 			 */
335 			if (BT_TEST(xc_priority_set, CPU->cpu_id)) {
336 				func = xc_priority_data.xc_func;
337 				a1 = xc_priority_data.xc_a1;
338 				a2 = xc_priority_data.xc_a2;
339 				a3 = xc_priority_data.xc_a3;
340 				XC_BT_CLEAR(xc_priority_set, CPU->cpu_id);
341 				xc_decrement(mcpup);
342 				func(a1, a2, a3);
343 				if (mcpup->xc_work_cnt == 0)
344 					return (rc);
345 			}
346 
347 			/*
348 			 * wait for a message to arrive
349 			 */
350 			SMT_PAUSE();
351 		}
352 
353 
354 		/*
355 		 * process the message
356 		 */
357 		switch (msg->xc_command) {
358 
359 		/*
360 		 * ASYNC gives back the message immediately, then we do the
361 		 * function and return with no more waiting.
362 		 */
363 		case XC_MSG_ASYNC:
364 			data = &cpu[msg->xc_master]->cpu_m.xc_data;
365 			func = data->xc_func;
366 			a1 = data->xc_a1;
367 			a2 = data->xc_a2;
368 			a3 = data->xc_a3;
369 			msg->xc_command = XC_MSG_DONE;
370 			xc_insert(&cpu[msg->xc_master]->cpu_m.xc_msgbox, msg);
371 			if (func != NULL)
372 				(void) (*func)(a1, a2, a3);
373 			xc_decrement(mcpup);
374 			break;
375 
376 		/*
377 		 * SYNC messages do the call, then send it back to the master
378 		 * in WAITING mode
379 		 */
380 		case XC_MSG_SYNC:
381 			data = &cpu[msg->xc_master]->cpu_m.xc_data;
382 			if (data->xc_func != NULL)
383 				(void) (*data->xc_func)(data->xc_a1,
384 				    data->xc_a2, data->xc_a3);
385 			msg->xc_command = XC_MSG_WAITING;
386 			xc_insert(&cpu[msg->xc_master]->cpu_m.xc_msgbox, msg);
387 			break;
388 
389 		/*
390 		 * WAITING messsages are collected by the master until all
391 		 * have arrived. Once all arrive, we release them back to
392 		 * the slaves
393 		 */
394 		case XC_MSG_WAITING:
395 			xc_insert(&xc_waiters, msg);
396 			if (++num_waiting < mcpup->xc_wait_cnt)
397 				break;
398 			while ((msg = xc_extract(&xc_waiters)) != NULL) {
399 				msg->xc_command = XC_MSG_RELEASED;
400 				xc_insert(&cpu[msg->xc_slave]->cpu_m.xc_msgbox,
401 				    msg);
402 				--num_waiting;
403 			}
404 			if (num_waiting != 0)
405 				panic("wrong number waiting");
406 			mcpup->xc_wait_cnt = 0;
407 			break;
408 
409 		/*
410 		 * CALL messages do the function and then, like RELEASE,
411 		 * send the message is back to master as DONE.
412 		 */
413 		case XC_MSG_CALL:
414 			data = &cpu[msg->xc_master]->cpu_m.xc_data;
415 			if (data->xc_func != NULL)
416 				(void) (*data->xc_func)(data->xc_a1,
417 				    data->xc_a2, data->xc_a3);
418 			/*FALLTHROUGH*/
419 		case XC_MSG_RELEASED:
420 			msg->xc_command = XC_MSG_DONE;
421 			xc_insert(&cpu[msg->xc_master]->cpu_m.xc_msgbox, msg);
422 			xc_decrement(mcpup);
423 			break;
424 
425 		/*
426 		 * DONE means a slave has completely finished up.
427 		 * Once we collect all the DONE messages, we'll exit
428 		 * processing too.
429 		 */
430 		case XC_MSG_DONE:
431 			msg->xc_command = XC_MSG_FREE;
432 			xc_insert(&mcpup->xc_free, msg);
433 			xc_decrement(mcpup);
434 			break;
435 
436 		case XC_MSG_FREE:
437 			panic("free message 0x%p in msgbox", (void *)msg);
438 			break;
439 
440 		default:
441 			panic("bad message 0x%p in msgbox", (void *)msg);
442 			break;
443 		}
444 	}
445 	return (rc);
446 }
447 
448 /*
449  * Initiate cross call processing.
450  */
451 static void
452 xc_common(
453 	xc_func_t func,
454 	xc_arg_t arg1,
455 	xc_arg_t arg2,
456 	xc_arg_t arg3,
457 	ulong_t *set,
458 	uint_t command)
459 {
460 	int c;
461 	struct cpu *cpup;
462 	xc_msg_t *msg;
463 	xc_data_t *data;
464 	int cnt;
465 	int save_spl;
466 
467 	if (!xc_initialized) {
468 		if (BT_TEST(set, CPU->cpu_id) && (CPU->cpu_flags & CPU_READY) &&
469 		    func != NULL)
470 			(void) (*func)(arg1, arg2, arg3);
471 		return;
472 	}
473 
474 	save_spl = splr(ipltospl(XC_HI_PIL));
475 
476 	/*
477 	 * fill in cross call data
478 	 */
479 	data = &CPU->cpu_m.xc_data;
480 	data->xc_func = func;
481 	data->xc_a1 = arg1;
482 	data->xc_a2 = arg2;
483 	data->xc_a3 = arg3;
484 
485 	/*
486 	 * Post messages to all CPUs involved that are CPU_READY
487 	 */
488 	CPU->cpu_m.xc_wait_cnt = 0;
489 	for (c = 0; c < max_ncpus; ++c) {
490 		if (!BT_TEST(set, c))
491 			continue;
492 		cpup = cpu[c];
493 		if (cpup == NULL || !(cpup->cpu_flags & CPU_READY))
494 			continue;
495 
496 		/*
497 		 * Fill out a new message.
498 		 */
499 		msg = xc_extract(&CPU->cpu_m.xc_free);
500 		if (msg == NULL)
501 			panic("Ran out of free xc_msg_t's");
502 		msg->xc_command = command;
503 		if (msg->xc_master != CPU->cpu_id)
504 			panic("msg %p has wrong xc_master", (void *)msg);
505 		msg->xc_slave = c;
506 
507 		/*
508 		 * Increment my work count for all messages that I'll
509 		 * transition from DONE to FREE.
510 		 * Also remember how many XC_MSG_WAITINGs to look for
511 		 */
512 		(void) xc_increment(&CPU->cpu_m);
513 		if (command == XC_MSG_SYNC)
514 			++CPU->cpu_m.xc_wait_cnt;
515 
516 		/*
517 		 * Increment the target CPU work count then insert the message
518 		 * in the target msgbox. If I post the first bit of work
519 		 * for the target to do, send an IPI to the target CPU.
520 		 */
521 		cnt = xc_increment(&cpup->cpu_m);
522 		xc_insert(&cpup->cpu_m.xc_msgbox, msg);
523 		if (cpup != CPU) {
524 			if (cnt == 0) {
525 				CPU_STATS_ADDQ(CPU, sys, xcalls, 1);
526 				send_dirint(c, XC_HI_PIL);
527 				if (xc_collect_enable)
528 					++xc_total_cnt;
529 			} else if (xc_collect_enable) {
530 				++xc_multi_cnt;
531 			}
532 		}
533 	}
534 
535 	/*
536 	 * Now drop into the message handler until all work is done
537 	 */
538 	(void) xc_serv(NULL, NULL);
539 	splx(save_spl);
540 }
541 
542 /*
543  * Push out a priority cross call.
544  */
545 static void
546 xc_priority_common(
547 	xc_func_t func,
548 	xc_arg_t arg1,
549 	xc_arg_t arg2,
550 	xc_arg_t arg3,
551 	ulong_t *set)
552 {
553 	int i;
554 	int c;
555 	struct cpu *cpup;
556 
557 	/*
558 	 * Wait briefly for any previous xc_priority to have finished.
559 	 */
560 	for (c = 0; c < max_ncpus; ++c) {
561 		cpup = cpu[c];
562 		if (cpup == NULL || !(cpup->cpu_flags & CPU_READY))
563 			continue;
564 
565 		/*
566 		 * The value of 40000 here is from old kernel code. It
567 		 * really should be changed to some time based value, since
568 		 * under a hypervisor, there's no guarantee a remote CPU
569 		 * is even scheduled.
570 		 */
571 		for (i = 0; BT_TEST(xc_priority_set, c) && i < 40000; ++i)
572 			SMT_PAUSE();
573 
574 		/*
575 		 * Some CPU did not respond to a previous priority request. It's
576 		 * probably deadlocked with interrupts blocked or some such
577 		 * problem. We'll just erase the previous request - which was
578 		 * most likely a kmdb_enter that has already expired - and plow
579 		 * ahead.
580 		 */
581 		if (BT_TEST(xc_priority_set, c)) {
582 			XC_BT_CLEAR(xc_priority_set, c);
583 			if (cpup->cpu_m.xc_work_cnt > 0)
584 				xc_decrement(&cpup->cpu_m);
585 		}
586 	}
587 
588 	/*
589 	 * fill in cross call data
590 	 */
591 	xc_priority_data.xc_func = func;
592 	xc_priority_data.xc_a1 = arg1;
593 	xc_priority_data.xc_a2 = arg2;
594 	xc_priority_data.xc_a3 = arg3;
595 
596 	/*
597 	 * Post messages to all CPUs involved that are CPU_READY
598 	 * We'll always IPI, plus bang on the xc_msgbox for i86_mwait()
599 	 */
600 	for (c = 0; c < max_ncpus; ++c) {
601 		if (!BT_TEST(set, c))
602 			continue;
603 		cpup = cpu[c];
604 		if (cpup == NULL || !(cpup->cpu_flags & CPU_READY) ||
605 		    cpup == CPU)
606 			continue;
607 		(void) xc_increment(&cpup->cpu_m);
608 		XC_BT_SET(xc_priority_set, c);
609 		send_dirint(c, XC_HI_PIL);
610 		for (i = 0; i < 10; ++i) {
611 			(void) casptr(&cpup->cpu_m.xc_msgbox,
612 			    cpup->cpu_m.xc_msgbox, cpup->cpu_m.xc_msgbox);
613 		}
614 	}
615 }
616 
617 /*
618  * Do cross call to all other CPUs with absolutely no waiting or handshaking.
619  * This should only be used for extraordinary operations, like panic(), which
620  * need to work, in some fashion, in a not completely functional system.
621  * All other uses that want minimal waiting should use xc_call_nowait().
622  */
623 void
624 xc_priority(
625 	xc_arg_t arg1,
626 	xc_arg_t arg2,
627 	xc_arg_t arg3,
628 	ulong_t *set,
629 	xc_func_t func)
630 {
631 	extern int IGNORE_KERNEL_PREEMPTION;
632 	int save_spl = splr(ipltospl(XC_HI_PIL));
633 	int save_kernel_preemption = IGNORE_KERNEL_PREEMPTION;
634 
635 	IGNORE_KERNEL_PREEMPTION = 1;
636 	xc_priority_common((xc_func_t)func, arg1, arg2, arg3, set);
637 	IGNORE_KERNEL_PREEMPTION = save_kernel_preemption;
638 	splx(save_spl);
639 }
640 
641 /*
642  * Wrapper for kmdb to capture other CPUs, causing them to enter the debugger.
643  */
644 void
645 kdi_xc_others(int this_cpu, void (*func)(void))
646 {
647 	extern int IGNORE_KERNEL_PREEMPTION;
648 	int save_kernel_preemption;
649 	cpuset_t set;
650 
651 	if (!xc_initialized)
652 		return;
653 
654 	save_kernel_preemption = IGNORE_KERNEL_PREEMPTION;
655 	IGNORE_KERNEL_PREEMPTION = 1;
656 	CPUSET_ALL_BUT(set, this_cpu);
657 	xc_priority_common((xc_func_t)func, 0, 0, 0, CPUSET2BV(set));
658 	IGNORE_KERNEL_PREEMPTION = save_kernel_preemption;
659 }
660 
661 
662 
663 /*
664  * Invoke function on specified processors. Remotes may continue after
665  * service with no waiting. xc_call_nowait() may return immediately too.
666  */
667 void
668 xc_call_nowait(
669 	xc_arg_t arg1,
670 	xc_arg_t arg2,
671 	xc_arg_t arg3,
672 	ulong_t *set,
673 	xc_func_t func)
674 {
675 	xc_common(func, arg1, arg2, arg3, set, XC_MSG_ASYNC);
676 }
677 
678 /*
679  * Invoke function on specified processors. Remotes may continue after
680  * service with no waiting. xc_call() returns only after remotes have finished.
681  */
682 void
683 xc_call(
684 	xc_arg_t arg1,
685 	xc_arg_t arg2,
686 	xc_arg_t arg3,
687 	ulong_t *set,
688 	xc_func_t func)
689 {
690 	xc_common(func, arg1, arg2, arg3, set, XC_MSG_CALL);
691 }
692 
693 /*
694  * Invoke function on specified processors. Remotes wait until all have
695  * finished. xc_sync() also waits until all remotes have finished.
696  */
697 void
698 xc_sync(
699 	xc_arg_t arg1,
700 	xc_arg_t arg2,
701 	xc_arg_t arg3,
702 	ulong_t *set,
703 	xc_func_t func)
704 {
705 	xc_common(func, arg1, arg2, arg3, set, XC_MSG_SYNC);
706 }
707