Home
last modified time | relevance | path

Searched full:workers (Results 1 – 25 of 95) sorted by relevance

1234

/linux/Documentation/translations/zh_CN/core-api/
H A Dworkqueue.rst577 pool[00] ref= 1 nice= 0 idle/workers= 4/ 4 cpu= 0
578 pool[01] ref= 1 nice=-20 idle/workers= 2/ 2 cpu= 0
579 pool[02] ref= 1 nice= 0 idle/workers= 4/ 4 cpu= 1
580 pool[03] ref= 1 nice=-20 idle/workers= 2/ 2 cpu= 1
581 pool[04] ref= 1 nice= 0 idle/workers= 4/ 4 cpu= 2
582 pool[05] ref= 1 nice=-20 idle/workers= 2/ 2 cpu= 2
583 pool[06] ref= 1 nice= 0 idle/workers= 3/ 3 cpu= 3
584 pool[07] ref= 1 nice=-20 idle/workers= 2/ 2 cpu= 3
585 pool[08] ref=42 nice= 0 idle/workers= 6/ 6 cpus=0000000f
586 pool[09] ref=28 nice= 0 idle/workers= 3/ 3 cpus=00000003
[all …]
/linux/tools/testing/selftests/sched_ext/
H A Dcyclic_kick_wait.c102 struct worker_ctx workers[NR_WORKERS] = {}; in run() local
141 workers[i].cpu = test_cpus[i / WORKERS_PER_CPU]; in run()
144 ret = pthread_create(&workers[i].tid, NULL, worker_fn, &workers[i]); in run()
150 workers[i].started = true; in run()
164 workers[i].stop = true; in run()
167 ret = join_worker(&workers[i]); in run()
/linux/tools/testing/selftests/powerpc/math/
H A Dfpu_preempt.c24 /* Time to wait for workers to get preempted (seconds) */
69 printf("\tWaiting for all workers to start..."); in test_preempt_fpu()
74 printf("\tWaiting for %d seconds to let some workers get preempted...", PREEMPT_TIME); in test_preempt_fpu()
78 printf("\tStopping workers..."); in test_preempt_fpu()
H A Dvmx_preempt.c24 /* Time to wait for workers to get preempted (seconds) */
78 printf("\tWaiting for all workers to start..."); in test_preempt_vmx()
83 printf("\tWaiting for %d seconds to let some workers get preempted...", PREEMPT_TIME); in test_preempt_vmx()
87 printf("\tStopping workers..."); in test_preempt_vmx()
H A Dvsx_preempt.c23 /* Time to wait for workers to get preempted (seconds) */
110 printf("\tWaiting for %d workers to start...", threads_starting); in test_preempt_vsx()
115 printf("\tWaiting for %d seconds to let some workers get preempted...", PREEMPT_TIME); in test_preempt_vsx()
119 printf("\tStopping workers..."); in test_preempt_vsx()
H A Dfpu_signal.c89 printf("\tWaiting for all workers to start..."); in test_signal_fpu()
103 printf("\tStopping workers..."); in test_signal_fpu()
H A Dvmx_signal.c114 printf("\tWaiting for %d workers to start... %d", threads, threads_starting); in test_signal_vmx()
131 printf("\tKilling workers..."); in test_signal_vmx()
/linux/tools/testing/selftests/bpf/
H A Dtest_progs.c554 if (verbose() && !env.workers) in test__start_subtest()
915 { "workers", ARG_NUM_WORKERS, "WORKERS", OPTION_ARG_OPTIONAL, in start_libbpf_log_capture()
916 "Number of workers to run in parallel, default to number of cpus." }, in start_libbpf_log_capture()
1115 env->workers = atoi(arg); in parse_arg()
1116 if (!env->workers) { in parse_arg()
1121 env->workers = get_nprocs(); in parse_arg()
1346 for (i = 0; i < env.workers; i++) in str_msg()
1724 dispatcher_threads = calloc(sizeof(pthread_t), env.workers); in server_main()
1725 data = calloc(sizeof(struct dispatch_data), env.workers); in server_main()
[all...]
/linux/fs/erofs/
H A DKconfig190 bool "EROFS per-cpu decompression kthread workers"
193 Saying Y here enables per-CPU kthread workers pool to carry out
199 bool "EROFS high priority per-CPU kthread workers"
203 This permits EROFS to configure per-CPU kthread workers to run
/linux/drivers/gpio/
H A Dgpio-mpsse.c21 struct list_head workers; /* polling work threads */ member
364 * Stops all workers except `my_worker`.
375 &priv->workers, list) { in gpio_mpsse_stop_all_except()
410 * We only want one worker. Workers race to acquire irq_race and tear in gpio_mpsse_poll()
411 * down all other workers. This is a cond guard so that we don't deadlock in gpio_mpsse_poll()
501 * As a result, these workers will stick around until irq is reenabled in gpio_mpsse_irq_disable()
505 list_for_each_entry(worker, &priv->workers, list) in gpio_mpsse_irq_disable()
531 list_add(&worker->list, &priv->workers); in gpio_mpsse_irq_enable()
592 INIT_LIST_HEAD(&priv->workers); in gpio_mpsse_probe()
/linux/tools/workqueue/
H A Dwq_dump.py28 idle number of idle workers
29 workers number of all workers
31 cpus CPUs the workers in the pool can run on (unbound pool)
161 print(f'idle/workers={pool.nr_idle.value_():3}/{pool.nr_workers.value_():3} ', end='')
/linux/tools/testing/selftests/kvm/x86/
H A Dhyperv_tlb_flush.c57 * Pass the following info to 'workers' and 'sender'
156 * Prepare to test: 'disable' workers by setting the expectation to '0',
164 /* 'Disable' workers */ in prepare_to_test()
168 /* Make sure workers are 'disabled' before we swap PTEs. */ in prepare_to_test()
171 /* Make sure workers have enough time to notice */ in prepare_to_test()
187 /* Set the expectation for workers, '0' means don't test */ in post_test()
191 /* Make sure workers have enough time to test */ in post_test()
630 * for 'workers' and issues TLB flush hypercalls. in main()
/linux/tools/testing/selftests/mm/
H A Dtest_vmalloc.sh86 echo "available test cases are run by NUM_CPUS workers simultaneously."
121 echo "# Runs 1 test(id_1), repeats it 5 times by NUM_CPUS workers"
129 echo -n "# Runs all tests by NUM_CPUS workers, shuffled order, repeats "
/linux/kernel/
H A Dworkqueue.c67 * While associated (!DISASSOCIATED), all workers are bound to the
71 * While DISASSOCIATED, the cpu may be offline and all workers have
84 POOL_DISASSOCIATED = 1 << 2, /* cpu can't serve workers */
124 * Rescue workers are used only on emergencies and shared by
215 int nr_workers; /* L: total number of workers */
216 int nr_idle; /* L: currently idle workers */
218 struct list_head idle_list; /* L: list of idle workers */
222 struct timer_list mayday_timer; /* L: SOS timer for workers */
224 /* a workers is either on busy_hash or idle_list, or the manager */
226 /* L: hash of busy workers */
220 struct list_head workers; /* A: attached workers */ global() member
[all...]
H A Dworkqueue_internal.h18 * The poor guys doing the actual heavy lifting. All on-duty workers are
48 struct list_head node; /* A: anchored at pool->workers */
H A Dpadata.c396 /* Restrict parallel_wq workers to pd->cpumask.pcpu. */ in padata_setup_cpumasks()
509 /* Initialize all percpu queues used by serial workers */
892 * serial_cpumask [RW] - cpumask for serial workers
893 * parallel_cpumask [RW] - cpumask for parallel workers
/linux/lib/
H A Dtest_vmalloc.c27 "Number of workers to perform tests(min: 1 max: USHRT_MAX)");
538 * A maximum number of workers is defined as hard-coded in init_test_configuration()
570 * Put on hold all workers. in do_concurrent_test()
587 * Now let the workers do their job. in do_concurrent_test()
592 * Sleep quiet until all workers are done with 1 second in do_concurrent_test()
H A Dtest_objpool.c396 /* tell workers threads to quit */ in ot_start_sync()
399 /* wait all workers threads finish and quit */ in ot_start_sync()
580 /* tell workers threads to quit */ in ot_start_async()
586 /* wait all workers threads finish and quit */ in ot_start_async()
/linux/io_uring/
H A Dio-wq.c38 IO_WQ_BIT_EXIT_ON_IDLE = 1, /* allow all workers to exit on idle */
91 * The list of free workers. Protected by #workers_lock
97 * The list of all workers. Protected by #workers_lock
321 * below the max number of workers, create one.
327 * wasn't setup with any unbounded workers. in io_wq_create_worker()
330 pr_warn_once("io-wq is not configured for unbound workers"); in io_wq_create_worker()
713 * keeping io-wq workers around for tasks that no longer have in io_wq_worker()
768 * Called when worker is going to sleep. If there are no workers currently
1473 * Set max number of unbounded workers, returns old value. If new_count is 0,
/linux/include/uapi/linux/
H A Dvhost.h63 * virtqueue. If userspace is not able to call this for workers its created,
64 * the kernel will free all the device's workers when the device is closed.
261 * - Vhost will create vhost workers as kernel threads.
/linux/tools/perf/trace/beauty/include/uapi/linux/
H A Dvhost.h63 * virtqueue. If userspace is not able to call this for workers its created,
64 * the kernel will free all the device's workers when the device is closed.
261 * - Vhost will create vhost workers as kernel threads.
/linux/tools/include/uapi/linux/
H A Dvhost.h
/linux/fs/btrfs/
H A Dasync-thread.c38 /* Up limit of concurrency workers */
41 /* Current number of concurrency workers */
/linux/drivers/vhost/
H A DKconfig107 to configure the default mode for vhost workers.
/linux/samples/workqueue/stall_detector/
H A Dwq_stall.c7 * detector correctly identifies stuck workers and produces useful

1234