Home
last modified time | relevance | path

Searched refs:workload (Results 1 – 25 of 116) sorted by relevance

12345

/linux/drivers/gpu/drm/i915/gvt/
H A Dexeclist.c370 static int prepare_execlist_workload(struct intel_vgpu_workload *workload) in prepare_execlist_workload() argument
372 struct intel_vgpu *vgpu = workload->vgpu; in prepare_execlist_workload()
377 if (!workload->emulate_schedule_in) in prepare_execlist_workload()
380 ctx[0] = *get_desc_from_elsp_dwords(&workload->elsp_dwords, 0); in prepare_execlist_workload()
381 ctx[1] = *get_desc_from_elsp_dwords(&workload->elsp_dwords, 1); in prepare_execlist_workload()
383 ret = emulate_execlist_schedule_in(&s->execlist[workload->engine->id], in prepare_execlist_workload()
392 static int complete_execlist_workload(struct intel_vgpu_workload *workload) in complete_execlist_workload() argument
394 struct intel_vgpu *vgpu = workload->vgpu; in complete_execlist_workload()
397 &s->execlist[workload->engine->id]; in complete_execlist_workload()
399 struct list_head *next = workload_q_head(vgpu, workload->engine)->next; in complete_execlist_workload()
[all …]
H A Dcmd_parser.h50 int intel_gvt_scan_and_shadow_ringbuffer(struct intel_vgpu_workload *workload);
56 int intel_gvt_scan_engine_context(struct intel_vgpu_workload *workload);
H A Dtrace.h231 void *workload, const char *cmd_name),
234 buf_addr_type, workload, cmd_name),
243 __field(void*, workload)
255 __entry->workload = workload;
271 __entry->workload)
H A Dscheduler.h137 void intel_vgpu_queue_workload(struct intel_vgpu_workload *workload);
164 void intel_vgpu_destroy_workload(struct intel_vgpu_workload *workload);
/linux/Documentation/admin-guide/
H A Dworkload-tracing.rst4 Discovering Linux kernel subsystems used by a workload
14 * Understanding system resources necessary to build and run a workload
17 in use by a workload. The completeness of the system usage information
18 depends on the completeness of coverage of a workload.
24 * Once we discover and understand the workload needs, we can focus on them
32 the system resources in use by a workload. Once we discover and understand
33 the workload needs, we can focus on them to avoid regressions and use it
37 the workload and doesn't include all the system calls that can be invoked
39 these system calls that are invoked. As an example, if a workload opens a
42 is a workload that provides full coverage of a workload then the method
[all …]
/linux/tools/perf/tests/shell/
H A Dstat_bpf_counters.sh7 workload="perf test -w sqrtloop"
44 …base_instructions=$(perf stat --no-big-num -e instructions -- $workload 2>&1 | awk '/instructions/…
45 …bpf_instructions=$(perf stat --no-big-num --bpf-counters -e instructions -- $workload 2>&1 | awk …
54 …num -e instructions/name=base_instructions/,instructions/name=bpf_instructions/b -- $workload 2>&1)
H A Dtrace_summary.sh17 workload="true"
20 echo "testing: perf trace ${args} -- ${workload}"
21 perf trace ${args} -- ${workload} >${OUTPUT} 2>&1
H A Dstat_metrics_values.sh14 workload="perf bench futex hash -r 2 -s"
22 $PYTHON $pythonvalidator -rule $rulefile -output_dir $tmpdir -wl "${workload}" \
H A Dtest_intel_pt.sh24 workload="${temp_dir}/workload"
64 cat << _end_of_file_ | /usr/bin/cc -o "${workload}" -xc - -pthread && have_workload=true
255 $workload &
257 $workload &
/linux/tools/perf/tests/
H A Dperf-record.c129 err = sched__get_first_possible_cpu(evlist->workload.pid, cpu_mask); in test__PERF_RECORD()
142 if (sched_setaffinity(evlist->workload.pid, cpu_mask_size, cpu_mask) < 0) { in test__PERF_RECORD()
231 if ((pid_t)sample.pid != evlist->workload.pid) { in test__PERF_RECORD()
233 name, evlist->workload.pid, sample.pid); in test__PERF_RECORD()
237 if ((pid_t)sample.tid != evlist->workload.pid) { in test__PERF_RECORD()
239 name, evlist->workload.pid, sample.tid); in test__PERF_RECORD()
248 (pid_t)event->comm.pid != evlist->workload.pid) { in test__PERF_RECORD()
H A Dbuiltin-test.c158 #define workloads__for_each(workload) \ argument
159 for (unsigned i = 0; i < ARRAY_SIZE(workloads) && ({ workload = workloads[i]; 1; }); i++)
783 const char *workload = NULL; in cmd_test() local
795 …OPT_STRING('w', "workload", &workload, "work", "workload to run for testing, use '--list-workloads… in cmd_test()
823 if (workload) in cmd_test()
824 return run_workload(workload, argc, argv); in cmd_test()
/linux/tools/perf/tests/shell/lib/
H A Dperf_metric_validation.py22 \tis {1} in workload(s): {2} \n\
31 \tis {1} in workload(s): {2}\n\
39 workload='true', metrics='', cputype='cpu'): argument
49 self.workloads = [x for x in workload.split(",") if x]
380 def _run_perf(self, metric, workload: str):
383 wl = workload.split()
392 def collect_perf(self, workload: str):
416 wl = workload
425 workload = self.workloads[self.wlidx]
427 data = self._run_perf(metric, workload)
[all …]
/linux/Documentation/admin-guide/mm/
H A Didle_page_tracking.rst9 accessed by a workload and which are idle. This information can be useful for
10 estimating the workload's working set size, which, in turn, can be taken into
11 account when configuring the workload parameters, setting memory cgroup limits,
12 or deciding where to place the workload within a compute cluster.
51 workload one should:
53 1. Mark all the workload's pages as idle by setting corresponding bits in
55 ``/proc/pid/pagemap`` if the workload is represented by a process, or by
56 filtering out alien pages using ``/proc/kpagecgroup`` in case the workload
59 2. Wait until the workload accesses its working set.
/linux/rust/pin-init/examples/
H A Dstatic_init.rs93 let workload = 1_000; in main() localVariable
100 for _ in 0..workload { in main()
109 for _ in 0..workload { in main()
122 assert_eq!(*mtx.lock(), workload * thread_count * 2); in main()
H A Dpthread_mutex.rs158 let workload = 1_000_000; in main() localVariable
165 for _ in 0..workload { in main()
170 for _ in 0..workload { in main()
182 assert_eq!(*mtx.lock(), workload * thread_count * 2); in main()
H A Dmutex.rs199 let workload = if cfg!(miri) { 100 } else { 1_000 }; in main() localVariable
206 for _ in 0..workload { in main()
211 for _ in 0..workload { in main()
223 assert_eq!(*mtx.lock(), workload * thread_count * 2); in main()
/linux/tools/perf/bench/
H A Dfind-bit-bench.c34 static noinline void workload(int val) in workload() function
82 workload(bit); in do_for_each_set_bit()
97 workload(bit); in do_for_each_set_bit()
/linux/Documentation/filesystems/nfs/
H A Dknfsd-stats.rst54 Depending on the NFS workload patterns and various network stack
58 However this is a more accurate and less workload-dependent measure
74 pool for the NFS workload (the workload is thread-limited), in which
76 performance of the NFS workload.
93 threads configured than can be used by the NFS workload. This is
99 slow; the idle timeout is 60 minutes. Unless the NFS workload
/linux/drivers/gpu/drm/amd/pm/powerplay/hwmgr/
H A Dpp_psm.c275 long workload[1]; in psm_adjust_power_state_dynamic() local
300 workload[0] = hwmgr->workload_setting[index]; in psm_adjust_power_state_dynamic()
302 if (hwmgr->power_profile_mode != workload[0] && hwmgr->hwmgr_func->set_power_profile_mode) in psm_adjust_power_state_dynamic()
303 hwmgr->hwmgr_func->set_power_profile_mode(hwmgr, workload, 0); in psm_adjust_power_state_dynamic()
/linux/Documentation/gpu/
H A Ddrm-compute.rst11 have their memory swapped out from them. Or they simply want their workload
19 waiting for the workload to complete. Effectively this terminates the workload
22 Since this is undesirable, there need to be mitigations to prevent a workload
/linux/Documentation/scheduler/
H A Dsched-capacity.rst73 With a workload that periodically does a fixed amount of work, you will get an
104 Executing the same workload as described in 1.3.1, which each CPU running at its
112 workload on CPU1
152 One issue that needs to be taken into account is that a workload's duty cycle is
154 periodic workload at a given frequency F::
163 Now, consider running the *same* workload at frequency F/2::
185 identical workload on CPUs of different capacity values will yield different
193 Executing a given periodic workload on each CPU at their maximum frequency would
384 workload on CPU0
391 workload on CPU1
[all …]
/linux/Documentation/tools/rtla/
H A Dcommon_timerlat_options.txt58 Set timerlat to run without a workload, and then dispatches user-space workloads
59 to wait on the timerlat_fd. Once the workload is awakened, it goes to sleep again
65 Set timerlat to run without workload, waiting for the user to dispatch a per-cpu
/linux/Documentation/translations/zh_CN/scheduler/
H A Dsched-capacity.rst108 workload on CPU1
339 workload on CPU0
346 workload on CPU1
/linux/Documentation/admin-guide/pm/
H A Dintel-speed-select.rst10 variety of diverse workload requirements.
82 This feature allows configuration of a server dynamically based on workload
216 workload, disable turbo::
220 Then runs a busy workload on all CPUs, for example::
262 level 0 to 2800 MHz at performance level 4. As a result, any workload, which can
540 the user control base frequency. If some critical workload threads demand
568 Before enabling Intel(R) SST-BF and measuring its impact on a workload
569 performance, execute some workload and measure performance and get a baseline
584 To compare, pick a multi-threaded workload where each thread can be scheduled on
588 Below, the workload is measuring average scheduler wakeup latency, so a lower
[all …]
/linux/tools/perf/util/
H A Devlist.c81 evlist->workload.pid = -1; in evlist__init()
1437 evlist->workload.cork_fd = -1; in evlist__prepare_workload()
1449 evlist->workload.pid = fork(); in evlist__prepare_workload()
1450 if (evlist->workload.pid < 0) { in evlist__prepare_workload()
1455 if (!evlist->workload.pid) { in evlist__prepare_workload()
1526 perf_thread_map__set_pid(evlist->core.threads, 0, evlist->workload.pid); in evlist__prepare_workload()
1540 evlist->workload.cork_fd = go_pipe[1]; in evlist__prepare_workload()
1555 if (evlist->workload.cork_fd >= 0) { in evlist__start_workload()
1561 ret = write(evlist->workload.cork_fd, &bf, 1); in evlist__start_workload()
1565 close(evlist->workload.cork_fd); in evlist__start_workload()
[all …]

12345