| /linux/drivers/gpu/drm/i915/gvt/ |
| H A D | execlist.c | 370 static int prepare_execlist_workload(struct intel_vgpu_workload *workload) in prepare_execlist_workload() argument 372 struct intel_vgpu *vgpu = workload->vgpu; in prepare_execlist_workload() 377 if (!workload->emulate_schedule_in) in prepare_execlist_workload() 380 ctx[0] = *get_desc_from_elsp_dwords(&workload->elsp_dwords, 0); in prepare_execlist_workload() 381 ctx[1] = *get_desc_from_elsp_dwords(&workload->elsp_dwords, 1); in prepare_execlist_workload() 383 ret = emulate_execlist_schedule_in(&s->execlist[workload->engine->id], in prepare_execlist_workload() 392 static int complete_execlist_workload(struct intel_vgpu_workload *workload) in complete_execlist_workload() argument 394 struct intel_vgpu *vgpu = workload->vgpu; in complete_execlist_workload() 397 &s->execlist[workload->engine->id]; in complete_execlist_workload() 399 struct list_head *next = workload_q_head(vgpu, workload->engine)->next; in complete_execlist_workload() [all …]
|
| H A D | cmd_parser.h | 50 int intel_gvt_scan_and_shadow_ringbuffer(struct intel_vgpu_workload *workload); 56 int intel_gvt_scan_engine_context(struct intel_vgpu_workload *workload);
|
| H A D | trace.h | 231 void *workload, const char *cmd_name), 234 buf_addr_type, workload, cmd_name), 243 __field(void*, workload) 255 __entry->workload = workload; 271 __entry->workload)
|
| H A D | scheduler.h | 137 void intel_vgpu_queue_workload(struct intel_vgpu_workload *workload); 164 void intel_vgpu_destroy_workload(struct intel_vgpu_workload *workload);
|
| /linux/Documentation/admin-guide/ |
| H A D | workload-tracing.rst | 4 Discovering Linux kernel subsystems used by a workload 14 * Understanding system resources necessary to build and run a workload 17 in use by a workload. The completeness of the system usage information 18 depends on the completeness of coverage of a workload. 24 * Once we discover and understand the workload needs, we can focus on them 32 the system resources in use by a workload. Once we discover and understand 33 the workload needs, we can focus on them to avoid regressions and use it 37 the workload and doesn't include all the system calls that can be invoked 39 these system calls that are invoked. As an example, if a workload opens a 42 is a workload that provides full coverage of a workload then the method [all …]
|
| /linux/tools/perf/tests/shell/ |
| H A D | stat_bpf_counters.sh | 7 workload="perf test -w sqrtloop" 44 …base_instructions=$(perf stat --no-big-num -e instructions -- $workload 2>&1 | awk '/instructions/… 45 …bpf_instructions=$(perf stat --no-big-num --bpf-counters -e instructions -- $workload 2>&1 | awk … 54 …num -e instructions/name=base_instructions/,instructions/name=bpf_instructions/b -- $workload 2>&1)
|
| H A D | trace_summary.sh | 17 workload="true" 20 echo "testing: perf trace ${args} -- ${workload}" 21 perf trace ${args} -- ${workload} >${OUTPUT} 2>&1
|
| H A D | stat_metrics_values.sh | 14 workload="perf bench futex hash -r 2 -s" 22 $PYTHON $pythonvalidator -rule $rulefile -output_dir $tmpdir -wl "${workload}" \
|
| H A D | test_intel_pt.sh | 24 workload="${temp_dir}/workload" 64 cat << _end_of_file_ | /usr/bin/cc -o "${workload}" -xc - -pthread && have_workload=true 255 $workload & 257 $workload &
|
| /linux/tools/perf/tests/ |
| H A D | perf-record.c | 129 err = sched__get_first_possible_cpu(evlist->workload.pid, cpu_mask); in test__PERF_RECORD() 142 if (sched_setaffinity(evlist->workload.pid, cpu_mask_size, cpu_mask) < 0) { in test__PERF_RECORD() 231 if ((pid_t)sample.pid != evlist->workload.pid) { in test__PERF_RECORD() 233 name, evlist->workload.pid, sample.pid); in test__PERF_RECORD() 237 if ((pid_t)sample.tid != evlist->workload.pid) { in test__PERF_RECORD() 239 name, evlist->workload.pid, sample.tid); in test__PERF_RECORD() 248 (pid_t)event->comm.pid != evlist->workload.pid) { in test__PERF_RECORD()
|
| H A D | builtin-test.c | 158 #define workloads__for_each(workload) \ argument 159 for (unsigned i = 0; i < ARRAY_SIZE(workloads) && ({ workload = workloads[i]; 1; }); i++) 783 const char *workload = NULL; in cmd_test() local 795 …OPT_STRING('w', "workload", &workload, "work", "workload to run for testing, use '--list-workloads… in cmd_test() 823 if (workload) in cmd_test() 824 return run_workload(workload, argc, argv); in cmd_test()
|
| /linux/tools/perf/tests/shell/lib/ |
| H A D | perf_metric_validation.py | 22 \tis {1} in workload(s): {2} \n\ 31 \tis {1} in workload(s): {2}\n\ 39 workload='true', metrics='', cputype='cpu'): argument 49 self.workloads = [x for x in workload.split(",") if x] 380 def _run_perf(self, metric, workload: str): 383 wl = workload.split() 392 def collect_perf(self, workload: str): 416 wl = workload 425 workload = self.workloads[self.wlidx] 427 data = self._run_perf(metric, workload) [all …]
|
| /linux/Documentation/admin-guide/mm/ |
| H A D | idle_page_tracking.rst | 9 accessed by a workload and which are idle. This information can be useful for 10 estimating the workload's working set size, which, in turn, can be taken into 11 account when configuring the workload parameters, setting memory cgroup limits, 12 or deciding where to place the workload within a compute cluster. 51 workload one should: 53 1. Mark all the workload's pages as idle by setting corresponding bits in 55 ``/proc/pid/pagemap`` if the workload is represented by a process, or by 56 filtering out alien pages using ``/proc/kpagecgroup`` in case the workload 59 2. Wait until the workload accesses its working set.
|
| /linux/rust/pin-init/examples/ |
| H A D | static_init.rs | 93 let workload = 1_000; in main() localVariable 100 for _ in 0..workload { in main() 109 for _ in 0..workload { in main() 122 assert_eq!(*mtx.lock(), workload * thread_count * 2); in main()
|
| H A D | pthread_mutex.rs | 158 let workload = 1_000_000; in main() localVariable 165 for _ in 0..workload { in main() 170 for _ in 0..workload { in main() 182 assert_eq!(*mtx.lock(), workload * thread_count * 2); in main()
|
| H A D | mutex.rs | 199 let workload = if cfg!(miri) { 100 } else { 1_000 }; in main() localVariable 206 for _ in 0..workload { in main() 211 for _ in 0..workload { in main() 223 assert_eq!(*mtx.lock(), workload * thread_count * 2); in main()
|
| /linux/tools/perf/bench/ |
| H A D | find-bit-bench.c | 34 static noinline void workload(int val) in workload() function 82 workload(bit); in do_for_each_set_bit() 97 workload(bit); in do_for_each_set_bit()
|
| /linux/Documentation/filesystems/nfs/ |
| H A D | knfsd-stats.rst | 54 Depending on the NFS workload patterns and various network stack 58 However this is a more accurate and less workload-dependent measure 74 pool for the NFS workload (the workload is thread-limited), in which 76 performance of the NFS workload. 93 threads configured than can be used by the NFS workload. This is 99 slow; the idle timeout is 60 minutes. Unless the NFS workload
|
| /linux/drivers/gpu/drm/amd/pm/powerplay/hwmgr/ |
| H A D | pp_psm.c | 275 long workload[1]; in psm_adjust_power_state_dynamic() local 300 workload[0] = hwmgr->workload_setting[index]; in psm_adjust_power_state_dynamic() 302 if (hwmgr->power_profile_mode != workload[0] && hwmgr->hwmgr_func->set_power_profile_mode) in psm_adjust_power_state_dynamic() 303 hwmgr->hwmgr_func->set_power_profile_mode(hwmgr, workload, 0); in psm_adjust_power_state_dynamic()
|
| /linux/Documentation/gpu/ |
| H A D | drm-compute.rst | 11 have their memory swapped out from them. Or they simply want their workload 19 waiting for the workload to complete. Effectively this terminates the workload 22 Since this is undesirable, there need to be mitigations to prevent a workload
|
| /linux/Documentation/scheduler/ |
| H A D | sched-capacity.rst | 73 With a workload that periodically does a fixed amount of work, you will get an 104 Executing the same workload as described in 1.3.1, which each CPU running at its 112 workload on CPU1 152 One issue that needs to be taken into account is that a workload's duty cycle is 154 periodic workload at a given frequency F:: 163 Now, consider running the *same* workload at frequency F/2:: 185 identical workload on CPUs of different capacity values will yield different 193 Executing a given periodic workload on each CPU at their maximum frequency would 384 workload on CPU0 391 workload on CPU1 [all …]
|
| /linux/Documentation/tools/rtla/ |
| H A D | common_timerlat_options.txt | 58 Set timerlat to run without a workload, and then dispatches user-space workloads 59 to wait on the timerlat_fd. Once the workload is awakened, it goes to sleep again 65 Set timerlat to run without workload, waiting for the user to dispatch a per-cpu
|
| /linux/Documentation/translations/zh_CN/scheduler/ |
| H A D | sched-capacity.rst | 108 workload on CPU1 339 workload on CPU0 346 workload on CPU1
|
| /linux/Documentation/admin-guide/pm/ |
| H A D | intel-speed-select.rst | 10 variety of diverse workload requirements. 82 This feature allows configuration of a server dynamically based on workload 216 workload, disable turbo:: 220 Then runs a busy workload on all CPUs, for example:: 262 level 0 to 2800 MHz at performance level 4. As a result, any workload, which can 540 the user control base frequency. If some critical workload threads demand 568 Before enabling Intel(R) SST-BF and measuring its impact on a workload 569 performance, execute some workload and measure performance and get a baseline 584 To compare, pick a multi-threaded workload where each thread can be scheduled on 588 Below, the workload is measuring average scheduler wakeup latency, so a lower [all …]
|
| /linux/tools/perf/util/ |
| H A D | evlist.c | 81 evlist->workload.pid = -1; in evlist__init() 1437 evlist->workload.cork_fd = -1; in evlist__prepare_workload() 1449 evlist->workload.pid = fork(); in evlist__prepare_workload() 1450 if (evlist->workload.pid < 0) { in evlist__prepare_workload() 1455 if (!evlist->workload.pid) { in evlist__prepare_workload() 1526 perf_thread_map__set_pid(evlist->core.threads, 0, evlist->workload.pid); in evlist__prepare_workload() 1540 evlist->workload.cork_fd = go_pipe[1]; in evlist__prepare_workload() 1555 if (evlist->workload.cork_fd >= 0) { in evlist__start_workload() 1561 ret = write(evlist->workload.cork_fd, &bf, 1); in evlist__start_workload() 1565 close(evlist->workload.cork_fd); in evlist__start_workload() [all …]
|