| /linux/tools/lib/python/ |
| H A D | jobserver.py | 52 - run(): executes a command setting PARALLELISM=<available slots jobs + 1>. 58 self.jobs = b"" 132 # Something went wrong. Clear self.jobs to avoid writing 134 self.jobs = b"" 147 # the jobs and give up. 149 if self.jobs: 150 os.write(self.writer, self.jobs) 154 self.jobs += slot 159 self.claim = len(self.jobs) + 1 168 if len(self.jobs) [all...] |
| /linux/scripts/ |
| H A D | generate_initcall_order.pl | 18 my $jobs = {}; # child process pid -> file handle 169 if (!exists($jobs->{$pid})) { 173 my $fh = $jobs->{$pid}; 181 delete($jobs->{$pid}); 202 $jobs->{$pid} = $fh; 213 if (scalar(keys(%{$jobs})) >= $njobs) { 219 while (scalar(keys(%{$jobs})) > 0) {
|
| /linux/drivers/gpu/drm/panfrost/ |
| H A D | panfrost_job.c | 164 struct panfrost_job *job = pfdev->jobs[slot][0]; in panfrost_dequeue_job() 178 pfdev->jobs[slot][0] = pfdev->jobs[slot][1]; in panfrost_dequeue_job() 179 pfdev->jobs[slot][1] = NULL; in panfrost_dequeue_job() 191 if (!pfdev->jobs[slot][0]) { in panfrost_enqueue_job() 192 pfdev->jobs[slot][0] = job; in panfrost_enqueue_job() 196 WARN_ON(pfdev->jobs[slot][1]); in panfrost_enqueue_job() 197 pfdev->jobs[slot][1] = job; in panfrost_enqueue_job() 199 panfrost_get_job_chain_flag(pfdev->jobs[slot][0])); in panfrost_enqueue_job() 578 } else if (pfdev->jobs[j][0] && !(js_state & MK_JS_MASK(j))) { in panfrost_jm_handle_irq() 605 if (!failed[j] || !pfdev->jobs[j][0]) in panfrost_jm_handle_irq() [all …]
|
| H A D | TODO | 11 - Compute job support. So called 'compute only' jobs need to be plumbed up to
|
| /linux/drivers/md/ |
| H A D | dm-kcopyd.c | 417 static struct kcopyd_job *pop_io_job(struct list_head *jobs, in pop_io_job() argument 426 list_for_each_entry(job, jobs, list) { in pop_io_job() 443 static struct kcopyd_job *pop(struct list_head *jobs, in pop() argument 450 if (!list_empty(jobs)) { in pop() 451 if (jobs == &kc->io_jobs) in pop() 452 job = pop_io_job(jobs, kc); in pop() 454 job = list_entry(jobs->next, struct kcopyd_job, list); in pop() 463 static void push(struct list_head *jobs, struct kcopyd_job *job) in push() argument 469 list_add_tail(&job->list, jobs); in push() 474 static void push_head(struct list_head *jobs, struct kcopyd_job *job) in push_head() argument [all …]
|
| /linux/Documentation/core-api/ |
| H A D | padata.rst | 9 Padata is a mechanism by which the kernel can farm jobs out to be done in 16 Padata also supports multithreaded jobs, splitting up the job evenly while load 25 The first step in using padata to run serialized jobs is to set up a 26 padata_instance structure for overall control of how jobs are to be run:: 39 jobs to be serialized independently. A padata_instance may have one or more 40 padata_shells associated with it, each allowing a separate series of jobs. 45 The CPUs used to run jobs can be changed in two ways, programmatically with 52 parallel cpumask describes which processors will be used to execute jobs 116 true parallelism is achieved by submitting multiple jobs. parallel() runs with 141 pains to ensure that jobs are completed in the order in which they were [all …]
|
| /linux/drivers/gpu/drm/imagination/ |
| H A D | pvr_job.c | 714 if (!args->jobs.count) in pvr_submit_jobs() 717 err = PVR_UOBJ_GET_ARRAY(job_args, &args->jobs); in pvr_submit_jobs() 721 job_data = kvmalloc_objs(*job_data, args->jobs.count, in pvr_submit_jobs() 728 err = pvr_job_data_init(pvr_dev, pvr_file, job_args, &args->jobs.count, in pvr_submit_jobs() 733 jobs_alloced = args->jobs.count; in pvr_submit_jobs() 747 err = prepare_job_syncs_for_each(pvr_file, job_data, &args->jobs.count, in pvr_submit_jobs() 752 err = prepare_job_resvs_for_each(&exec, job_data, args->jobs.count); in pvr_submit_jobs() 756 err = pvr_jobs_link_geom_frag(job_data, &args->jobs.count); in pvr_submit_jobs() 763 update_job_resvs_for_each(job_data, args->jobs.count); in pvr_submit_jobs() 764 push_jobs(job_data, args->jobs.count); in pvr_submit_jobs()
|
| /linux/drivers/accel/rocket/ |
| H A D | rocket_job.c | 595 struct drm_rocket_job *jobs; in rocket_ioctl_submit() local 612 jobs = kvmalloc_objs(*jobs, args->job_count); in rocket_ioctl_submit() 613 if (!jobs) { in rocket_ioctl_submit() 619 if (copy_from_user(&jobs[i], in rocket_ioctl_submit() 620 u64_to_user_ptr(args->jobs) + i * args->job_struct_size, in rocket_ioctl_submit() 621 sizeof(*jobs))) { in rocket_ioctl_submit() 630 rocket_ioctl_submit_job(dev, file, &jobs[i]); in rocket_ioctl_submit() 633 kvfree(jobs); in rocket_ioctl_submit()
|
| /linux/drivers/accel/ethosu/ |
| H A D | ethosu_job.c | 488 struct drm_ethosu_job __free(kvfree) *jobs = in ethosu_ioctl_submit() 489 kvmalloc_objs(*jobs, args->job_count); in ethosu_ioctl_submit() 490 if (!jobs) in ethosu_ioctl_submit() 493 if (copy_from_user(jobs, in ethosu_ioctl_submit() 494 (void __user *)(uintptr_t)args->jobs, in ethosu_ioctl_submit() 495 args->job_count * sizeof(*jobs))) { in ethosu_ioctl_submit() 501 ret = ethosu_ioctl_submit_job(dev, file, &jobs[i]); in ethosu_ioctl_submit()
|
| /linux/Documentation/gpu/ |
| H A D | drm-compute.rst | 10 Some hardware may schedule compute jobs, and have no way to pre-empt them, or 16 As with normal compute jobs, dma-fence may not be used at all. In this case, 34 older compute jobs to start a new one. 43 into cgroups, that will allow jobs to run next to each other without
|
| /linux/tools/docs/ |
| H A D | sphinx-build-wrapper | 150 Get the number of jobs to be used for docs build passed via command 153 The number of jobs can be on different places: 157 3) if called via GNU make, -j specifies the desired number of jobs. 166 # and desired number of parallel jobs. 169 parser.add_argument('-j', '--jobs', type=int) 198 elif sphinx_args.jobs: 199 self.n_jobs = sphinx_args.jobs 298 many jobs are still available from a job pool. claim all remaining 299 jobs, as we don't want sphinx-build to run in parallel with other 300 jobs [all...] |
| H A D | get_abi.py | 178 max_workers=args.jobs,
|
| /linux/Documentation/admin-guide/device-mapper/ |
| H A D | kcopyd.rst | 10 to set aside for their copy jobs. This is done with a call to 43 When a user is done with all their copy jobs, they should call
|
| /linux/tools/testing/kunit/ |
| H A D | kunit_kernel.py | 75 def make(self, jobs: int, build_dir: str, make_options: Optional[List[str]]) -> None: 77 'ARCH=' + self._linux_arch, 'O=' + build_dir, '--jobs=' + str(jobs)] 340 def build_kernel(self, jobs: int, build_dir: str, make_options: Optional[List[str]]) -> bool: 343 self._ops.make(jobs, build_dir, make_options)
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_cs.h | 62 struct amdgpu_job *jobs[AMDGPU_CS_GANG_SIZE]; member
|
| /linux/tools/perf/scripts/python/ |
| H A D | parallel-perf.py | 724 if self.jobs < 0 or self.nr < 0 or self.interval < 0: 728 if self.jobs == 0: 729 self.jobs = NumberOfCPUs() 734 self.nr = self.jobs 835 result = RunWork(self.worklist, self.jobs, verbosity=self.verbosity)
|
| /linux/Documentation/dev-tools/kunit/ |
| H A D | run_wrapper.rst | 33 ./tools/testing/kunit/kunit.py run --timeout=30 --jobs=`nproc --all` 36 - ``--jobs`` sets the number of threads to build the kernel. 233 --jobs=12 \ 306 - ``--jobs``: Specifies the number of jobs (commands) to run simultaneously.
|
| /linux/Documentation/admin-guide/cgroup-v1/ |
| H A D | cpusets.rst | 87 can benefit from explicitly placing jobs on properly sized subsets of 100 executing jobs. The location of the running jobs pages may also be moved 252 jobs can share common kernel data, such as file system pages, while 254 construct a large mem_exclusive cpuset to hold all the jobs, and 268 This enables batch managers monitoring jobs running in dedicated 273 submitted jobs, which may choose to terminate or re-prioritize jobs that 276 computing jobs that will dramatically fail to meet required performance 381 This policy can provide substantial improvements for jobs that need 384 the several nodes in the jobs cpuset in order to fit. Without this 385 policy, especially for jobs that might have one thread reading in the [all …]
|
| /linux/Documentation/translations/zh_CN/mm/ |
| H A D | hwpoison.rst | 127 echo `jobs -p` > /sys/fs/cgroup/mem/hwpoison/tasks
|
| /linux/tools/testing/selftests/kho/ |
| H A D | vmtest.sh | 175 local make_cmd="make ARCH=$arch CROSS_COMPILE=$CROSS_COMPILE -j$jobs"
|
| /linux/tools/testing/selftests/net/ |
| H A D | udpgro.sh | 22 [ -n "${jobs}" ] && kill -1 ${jobs} 2>/dev/null
|
| /linux/Documentation/admin-guide/mm/ |
| H A D | multigen_lru.rst | 103 scheduler needs to estimate the working sets of the existing jobs. 163 existing jobs.
|
| /linux/Documentation/driver-api/tty/ |
| H A D | index.rst | 25 implementing echoes, signal handling, jobs control, special characters
|
| /linux/Documentation/accounting/ |
| H A D | psi.rst | 27 dynamically using techniques such as load shedding, migrating jobs to 29 priority or restartable batch jobs.
|
| /linux/tools/cgroup/ |
| H A D | iocost_coef_gen.py | 89 def run_fio(testfile, duration, iotype, iodepth, blocksize, jobs): argument
|