| /linux/Documentation/translations/zh_CN/core-api/ |
| H A D | workqueue.rst | 577 pool[00] ref= 1 nice= 0 idle/workers= 4/ 4 cpu= 0 578 pool[01] ref= 1 nice=-20 idle/workers= 2/ 2 cpu= 0 579 pool[02] ref= 1 nice= 0 idle/workers= 4/ 4 cpu= 1 580 pool[03] ref= 1 nice=-20 idle/workers= 2/ 2 cpu= 1 581 pool[04] ref= 1 nice= 0 idle/workers= 4/ 4 cpu= 2 582 pool[05] ref= 1 nice=-20 idle/workers= 2/ 2 cpu= 2 583 pool[06] ref= 1 nice= 0 idle/workers= 3/ 3 cpu= 3 584 pool[07] ref= 1 nice=-20 idle/workers= 2/ 2 cpu= 3 585 pool[08] ref=42 nice= 0 idle/workers= 6/ 6 cpus=0000000f 586 pool[09] ref=28 nice= 0 idle/workers= 3/ 3 cpus=00000003 [all …]
|
| /linux/tools/testing/selftests/powerpc/math/ |
| H A D | fpu_preempt.c | 24 /* Time to wait for workers to get preempted (seconds) */ 69 printf("\tWaiting for all workers to start..."); in test_preempt_fpu() 74 printf("\tWaiting for %d seconds to let some workers get preempted...", PREEMPT_TIME); in test_preempt_fpu() 78 printf("\tStopping workers..."); in test_preempt_fpu()
|
| H A D | vmx_preempt.c | 24 /* Time to wait for workers to get preempted (seconds) */ 78 printf("\tWaiting for all workers to start..."); in test_preempt_vmx() 83 printf("\tWaiting for %d seconds to let some workers get preempted...", PREEMPT_TIME); in test_preempt_vmx() 87 printf("\tStopping workers..."); in test_preempt_vmx()
|
| H A D | vsx_preempt.c | 23 /* Time to wait for workers to get preempted (seconds) */ 110 printf("\tWaiting for %d workers to start...", threads_starting); in test_preempt_vsx() 115 printf("\tWaiting for %d seconds to let some workers get preempted...", PREEMPT_TIME); in test_preempt_vsx() 119 printf("\tStopping workers..."); in test_preempt_vsx()
|
| H A D | fpu_signal.c | 89 printf("\tWaiting for all workers to start..."); in test_signal_fpu() 103 printf("\tStopping workers..."); in test_signal_fpu()
|
| H A D | vmx_signal.c | 114 printf("\tWaiting for %d workers to start... %d", threads, threads_starting); in test_signal_vmx() 131 printf("\tKilling workers..."); in test_signal_vmx()
|
| /linux/tools/testing/selftests/bpf/ |
| H A D | test_progs.c | 536 if (verbose() && !env.workers) in test__end_subtest() 888 { "workers", ARG_NUM_WORKERS, "WORKERS", OPTION_ARG_OPTIONAL, 889 "Number of workers to run in parallel, default to number of cpus." }, 1088 env->workers = atoi(arg); in parse_arg() 1089 if (!env->workers) { in parse_arg() 1094 env->workers = get_nprocs(); in parse_arg() 1305 for (i = 0; i < env.workers; i++) in sigint_handler() 1683 dispatcher_threads = calloc(sizeof(pthread_t), env.workers); in server_main() 1684 data = calloc(sizeof(struct dispatch_data), env.workers); in server_main() [all...] |
| /linux/drivers/gpio/ |
| H A D | gpio-mpsse.c | 21 struct list_head workers; /* polling work threads */ member 364 * Stops all workers except `my_worker`. 375 &priv->workers, list) { in gpio_mpsse_stop_all_except() 410 * We only want one worker. Workers race to acquire irq_race and tear in gpio_mpsse_poll() 411 * down all other workers. This is a cond guard so that we don't deadlock in gpio_mpsse_poll() 501 * As a result, these workers will stick around until irq is reenabled in gpio_mpsse_irq_disable() 505 list_for_each_entry(worker, &priv->workers, list) in gpio_mpsse_irq_disable() 531 list_add(&worker->list, &priv->workers); in gpio_mpsse_irq_enable() 599 INIT_LIST_HEAD(&priv->workers); in gpio_mpsse_probe()
|
| /linux/tools/workqueue/ |
| H A D | wq_dump.py | 28 idle number of idle workers 29 workers number of all workers 31 cpus CPUs the workers in the pool can run on (unbound pool) 160 print(f'idle/workers={pool.nr_idle.value_():3}/{pool.nr_workers.value_():3} ', end='')
|
| H A D | wq_monitor.py | 64 PWQ_STAT_REPATRIATED = prog['PWQ_STAT_REPATRIATED'] # unbound workers brought back into scope
|
| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | hyperv_tlb_flush.c | 57 * Pass the following info to 'workers' and 'sender' 156 * Prepare to test: 'disable' workers by setting the expectation to '0', 164 /* 'Disable' workers */ in prepare_to_test() 168 /* Make sure workers are 'disabled' before we swap PTEs. */ in prepare_to_test() 171 /* Make sure workers have enough time to notice */ in prepare_to_test() 187 /* Set the expectation for workers, '0' means don't test */ in post_test() 191 /* Make sure workers have enough time to test */ in post_test() 630 * for 'workers' and issues TLB flush hypercalls. in main()
|
| /linux/tools/testing/selftests/mm/ |
| H A D | test_vmalloc.sh | 86 echo "available test cases are run by NUM_CPUS workers simultaneously." 121 echo "# Runs 1 test(id_1), repeats it 5 times by NUM_CPUS workers" 129 echo -n "# Runs all tests by NUM_CPUS workers, shuffled order, repeats "
|
| /linux/kernel/ |
| H A D | workqueue.c | 66 * While associated (!DISASSOCIATED), all workers are bound to the 70 * While DISASSOCIATED, the cpu may be offline and all workers have 83 POOL_DISASSOCIATED = 1 << 2, /* cpu can't serve workers */ 123 * Rescue workers are used only on emergencies and shared by 206 int nr_workers; /* L: total number of workers */ 207 int nr_idle; /* L: currently idle workers */ 209 struct list_head idle_list; /* L: list of idle workers */ 213 struct timer_list mayday_timer; /* L: SOS timer for workers */ 215 /* a workers is either on busy_hash or idle_list, or the manager */ 217 /* L: hash of busy workers */ 220 struct list_head workers; /* A: attached workers */ global() member [all...] |
| H A D | workqueue_internal.h | 18 * The poor guys doing the actual heavy lifting. All on-duty workers are 47 struct list_head node; /* A: anchored at pool->workers */
|
| H A D | padata.c | 396 /* Restrict parallel_wq workers to pd->cpumask.pcpu. */ in padata_setup_cpumasks() 509 /* Initialize all percpu queues used by serial workers */ 894 * serial_cpumask [RW] - cpumask for serial workers 895 * parallel_cpumask [RW] - cpumask for parallel workers
|
| /linux/lib/ |
| H A D | test_vmalloc.c | 27 "Number of workers to perform tests(min: 1 max: USHRT_MAX)"); 538 * A maximum number of workers is defined as hard-coded in init_test_configuration() 570 * Put on hold all workers. in do_concurrent_test() 587 * Now let the workers do their job. in do_concurrent_test() 592 * Sleep quiet until all workers are done with 1 second in do_concurrent_test()
|
| H A D | test_objpool.c | 396 /* tell workers threads to quit */ in ot_start_sync() 399 /* wait all workers threads finish and quit */ in ot_start_sync() 580 /* tell workers threads to quit */ in ot_start_async() 586 /* wait all workers threads finish and quit */ in ot_start_async()
|
| /linux/io_uring/ |
| H A D | io-wq.c | 38 IO_WQ_BIT_EXIT_ON_IDLE = 1, /* allow all workers to exit on idle */ 91 * The list of free workers. Protected by #workers_lock 97 * The list of all workers. Protected by #workers_lock 321 * below the max number of workers, create one. 327 * wasn't setup with any unbounded workers. in io_wq_create_worker() 330 pr_warn_once("io-wq is not configured for unbound workers"); in io_wq_create_worker() 713 * keeping io-wq workers around for tasks that no longer have in io_wq_worker() 768 * Called when worker is going to sleep. If there are no workers currently 1473 * Set max number of unbounded workers, returns old value. If new_count is 0,
|
| /linux/include/uapi/linux/ |
| H A D | vhost.h | 63 * virtqueue. If userspace is not able to call this for workers its created, 64 * the kernel will free all the device's workers when the device is closed. 261 * - Vhost will create vhost workers as kernel threads.
|
| /linux/tools/perf/trace/beauty/include/uapi/linux/ |
| H A D | vhost.h | 63 * virtqueue. If userspace is not able to call this for workers its created, 64 * the kernel will free all the device's workers when the device is closed. 261 * - Vhost will create vhost workers as kernel threads.
|
| /linux/tools/include/uapi/linux/ |
| H A D | vhost.h | |
| /linux/fs/btrfs/ |
| H A D | async-thread.c | 38 /* Up limit of concurrency workers */ 41 /* Current number of concurrency workers */
|
| /linux/drivers/vhost/ |
| H A D | Kconfig | 107 to configure the default mode for vhost workers.
|
| /linux/net/l2tp/ |
| H A D | Kconfig | 23 with home workers to connect to their offices.
|
| /linux/tools/docs/ |
| H A D | get_abi.py | 154 parser.add_argument("-j", "--jobs", "--max-workers", type=int, default=1,
|