| /linux/tools/perf/tests/shell/lib/ |
| H A D | perf_metric_validation.py | 13 self.workloads = [wl] # multiple workloads possible 24 … \tRelationship rule description: \'{5}\'".format(self.metric, self.collectedValue, self.workloads, 28 \tworkload(s): {1}".format(self.metric, self.workloads) 33 .format(self.metric, self.collectedValue, self.workloads, 49 self.workloads = [x for x in workload.split(",") if x] 50 self.wlidx = 0 # idx of current workloads 204 … [TestError([m], self.workloads[self.wlidx], negmetric[m], 0) for m in negmetric.keys()]) 279 …self.errlist.append(TestError([m['Name'] for m in rule['Metrics']], self.workloads[self.wlidx], [], 282 …self.errlist.append(TestError([m['Name'] for m in rule['Metrics']], self.workloads[self.wlidx], [v… 334 self.errlist.extend([TestError([name], self.workloads[self.wlidx], val, [all …]
|
| /linux/Documentation/timers/ |
| H A D | no_hz.rst | 26 workloads, you will normally -not- want this option. 39 right approach, for example, in heavy workloads with lots of tasks 42 hundreds of microseconds). For these types of workloads, scheduling 56 are running light workloads, you should therefore read the following 118 computationally intensive short-iteration workloads: If any CPU is 228 aggressive real-time workloads, which have the option of disabling 230 some workloads will no doubt want to use adaptive ticks to 232 options for these workloads: 252 workloads, which have few such transitions. Careful benchmarking 253 will be required to determine whether or not other workloads
|
| /linux/Documentation/gpu/ |
| H A D | drm-compute.rst | 2 Long running workloads and compute 5 Long running workloads (compute) are workloads that will not complete in 10 7 This means that other techniques need to be used to manage those workloads,
|
| H A D | drm-vm-bind-async.rst | 103 exec functions. For long-running workloads, such pipelining of a bind 109 operations for long-running workloads will not allow for pipelining 110 anyway since long-running workloads don't allow for dma-fences as 121 deeply pipelined behind other VM_BIND operations and workloads
|
| /linux/drivers/cpuidle/ |
| H A D | Kconfig | 33 Some workloads benefit from using it and it generally should be safe 45 Some virtualized workloads benefit from using it.
|
| /linux/Documentation/networking/device_drivers/ethernet/intel/ |
| H A D | idpf.rst | 81 Driver defaults are meant to fit a wide variety of workloads, but if further 89 is tuned for general workloads. The user can customize the interrupt rate 90 control for specific workloads, via ethtool, adjusting the number of
|
| /linux/Documentation/accounting/ |
| H A D | psi.rst | 10 When CPU, memory or IO devices are contended, workloads experience 19 such resource crunches and the time impact it has on complex workloads 23 scarcity aids users in sizing workloads to hardware--or provisioning
|
| /linux/Documentation/admin-guide/pm/ |
| H A D | intel_uncore_frequency_scaling.rst | 23 Users may have some latency sensitive workloads where they do not want any 24 change to uncore frequency. Also, users may have workloads which require 133 latency sensitive workloads further tuning can be done by SW to
|
| /linux/fs/squashfs/ |
| H A D | Kconfig | 106 poor performance on parallel I/O workloads when using multiple CPU 121 poor performance on parallel I/O workloads when using multiple CPU 159 reducinng performance in workloads like fio-based benchmarks.
|
| /linux/tools/perf/tests/ |
| H A D | builtin-test.c | 147 static struct test_workload *workloads[] = { variable 164 for (unsigned i = 0; i < ARRAY_SIZE(workloads) && ({ workload = workloads[i]; 1; }); i++) 800 …OPT_STRING('w', "workload", &workload, "work", "workload to run for testing, use '--list-workloads… in cmd_test() 801 …OPT_BOOLEAN(0, "list-workloads", &list_workloads, "List the available builtin workloads to use wit… in cmd_test()
|
| H A D | tests.h | 216 * Define test workloads to be used in test suites. 234 /* The list of test workloads */
|
| /linux/Documentation/driver-api/md/ |
| H A D | raid5-cache.rst | 58 completely avoid the overhead, so it's very helpful for some workloads. A 74 mode depending on the workloads. It's recommended to use a cache disk with at
|
| /linux/Documentation/admin-guide/hw-vuln/ |
| H A D | core-scheduling.rst | 9 workloads may benefit from running on the same core as they don't need the same 24 world workloads. In theory, core scheduling aims to perform at least as good as 30 total number of CPUs. Please measure the performance of your workloads always.
|
| /linux/drivers/crypto/cavium/nitrox/ |
| H A D | Kconfig | 18 for accelerating crypto workloads.
|
| /linux/drivers/infiniband/hw/mana/ |
| H A D | Kconfig | 8 for workloads (e.g. DPDK, MPI etc) that uses RDMA verbs to directly
|
| /linux/Documentation/tools/rtla/ |
| H A D | common_options.txt | 55 …ior differs between workload types. User workloads created by rtla will inherit rtla's cgroup. Ker…
|
| /linux/Documentation/arch/x86/ |
| H A D | sgx.rst | 249 SGX workloads, (or just any new workloads), and migrate all valuable 250 workloads. Although a machine reboot can recover all EPC memory, the bug
|
| H A D | orc-unwinder.rst | 34 Gorman [1]_ have shown a slowdown of 5-10% for some workloads. 46 footprint. That can transform to even higher speedups for workloads
|
| /linux/tools/perf/pmu-events/arch/x86/jaketown/ |
| H A D | uncore-cache.json | 1310 …eads. This count only tracks the regular credits Common high banwidth workloads should be able t… 1320 …eads. This count only tracks the regular credits Common high banwidth workloads should be able t… 1330 …eads. This count only tracks the regular credits Common high banwidth workloads should be able t… 1340 …eads. This count only tracks the regular credits Common high banwidth workloads should be able t… 1350 …l' credits. This statistic is generally not interesting for general IA workloads, but may be of i… 1360 …l' credits. This statistic is generally not interesting for general IA workloads, but may be of i… 1370 …l' credits. This statistic is generally not interesting for general IA workloads, but may be of i… 1380 …l' credits. This statistic is generally not interesting for general IA workloads, but may be of i… 1903 …ites. This count only tracks the regular credits Common high banwidth workloads should be able t… 1913 …ites. This count only tracks the regular credits Common high banwidth workloads should be able t… [all …]
|
| /linux/Documentation/block/ |
| H A D | bfq-iosched.rst | 87 background workloads are being executed: 122 sequential workloads considered in our tests. With random workloads, 123 and with all the workloads on flash-based devices, BFQ achieves, 142 possibly heavy workloads are being served, BFQ guarantees:
|
| /linux/Documentation/RCU/ |
| H A D | checklist.rst | 222 IPI-sensitive real-time workloads can use the rcupdate.rcu_normal 232 of the system, especially to real-time workloads running on the 398 real-time workloads, this is the whole point of using the 453 real-time workloads than is synchronize_rcu_expedited().
|
| /linux/Documentation/admin-guide/ |
| H A D | workload-tracing.rst | 34 to evaluate safety considerations. We use strace tool to trace workloads. 67 We used strace to trace the perf, stress-ng, paxtest workloads to illustrate 69 be applied to trace other workloads. 101 paxtest workloads to show how to analyze a workload and identify Linux 102 subsystems used by these workloads. Let's start with an overview of these 103 three workloads to get a better understanding of what they do and how to 173 by three workloads we have chose for this analysis. 312 Tracing workloads 315 Now that we understand the workloads, let's start tracing them. 595 information on the resources in use by workloads using strace.
|
| /linux/samples/coresight/ |
| H A D | coresight-cfg-sample.c | 23 * experimentation with mark / space ratios for various workloads
|
| /linux/tools/perf/tests/shell/ |
| H A D | sched.sh | 32 # Start two noploop workloads on CPU0 to trigger scheduling.
|
| /linux/Documentation/filesystems/ext4/ |
| H A D | orphan.rst | 18 global single linked list is a scalability bottleneck for workloads that result
|