| /linux/tools/testing/selftests/cgroup/ |
| H A D | memcg_protection.m | 6 % hierarchy to illustrate how overcommitted protection spreads among siblings 9 % Simulation assumes siblings consumed the initial amount of memory (w/out 16 % n vector nominal protection of siblings set at the given level (memory.low) 48 siblings = sum(u); variable 52 e = protected * min(1, E / siblings); % normalize overcommit 55 unclaimed = max(0, E - siblings); 56 parent_overuse = sum(c) - siblings; 79 % XXX here I do parallel reclaim of all siblings
|
| /linux/Documentation/admin-guide/hw-vuln/ |
| H A D | core-scheduling.rst | 100 siblings of a core such that all the selected tasks running on a core are 107 the sibling has the task enqueued. For rest of the siblings in the core, 112 Once a task has been selected for all the siblings in the core, an IPI is sent to 113 siblings for whom a new task was selected. Siblings on receiving the IPI will 125 siblings could be forced to select a lower priority task if the highest 157 and are considered system-wide trusted. The forced-idling of siblings running 174 the siblings to switch to the new task. But there could be hardware delays in 176 cause an attacker task to start running on a CPU before its siblings receive the 177 IPI. Even though cache is flushed on entry to user mode, victim tasks on siblings 185 Core scheduling cannot protect against MDS attacks between the siblings [all …]
|
| /linux/drivers/infiniband/hw/irdma/ |
| H A D | ws.c | 133 list_for_each_entry(node, &parent->child_list_head, siblings) { in ws_find_node() 139 list_for_each_entry(node, &parent->child_list_head, siblings) { in ws_find_node() 214 list_del(&tc_node->siblings); in irdma_remove_leaf() 219 list_del(&vsi_node->siblings); in irdma_remove_leaf() 295 list_add(&vsi_node->siblings, &ws_tree_root->child_list_head); in irdma_ws_add() 322 list_add(&tc_node->siblings, &vsi_node->child_list_head); in irdma_ws_add() 356 list_del(&tc_node->siblings); in irdma_ws_add() 362 list_del(&vsi_node->siblings); in irdma_ws_add()
|
| H A D | ws.h | 19 struct list_head siblings; member
|
| /linux/drivers/gpu/drm/i915/gt/uc/ |
| H A D | selftest_guc_multi_lrc.c | 34 struct intel_engine_cs *siblings[MAX_ENGINE_INSTANCE + 1]; in multi_lrc_create_parent() local 43 siblings[i++] = engine; in multi_lrc_create_parent() 49 logical_sort(siblings, i); in multi_lrc_create_parent() 51 return intel_engine_create_parallel(siblings, 1, i); in multi_lrc_create_parent()
|
| /linux/arch/x86/kernel/cpu/microcode/ |
| H A D | core.c | 295 * - HT siblings must be idle and not execute other code while the other sibling 450 * If the update was successful, let the siblings run the apply() in __load_primary() 499 /* Kick soft-offlined SMT siblings if required */ in load_primary() 507 /* Unconditionally release soft-offlined SMT siblings if required */ in load_primary() 582 unsigned int cpu, updated = 0, failed = 0, timedout = 0, siblings = 0; in load_late_stop_cpus() local 626 case UCODE_OK: siblings++; break; in load_late_stop_cpus() 642 pr_warn("%u offline siblings did not respond.\n", in load_late_stop_cpus() 654 pr_info("load: updated on %u primary CPUs with %u siblings\n", updated, siblings); in load_late_stop_cpus() 657 num_online_cpus() - (updated + siblings)); in load_late_stop_cpus() 662 return updated + siblings == num_online_cpus() ? 0 : -EIO; in load_late_stop_cpus()
|
| /linux/include/uapi/linux/ |
| H A D | membarrier.h | 70 * threads siblings have passed through a state 94 * call, that all its running threads siblings 120 * siblings have any currently running rseq
|
| /linux/tools/testing/selftests/drivers/platform/x86/intel/ifs/ |
| H A D | test_ifs.sh | 32 readonly SIBLINGS="siblings" 381 "$SIBLINGS") 446 ifs_test_cpus "$SIBLINGS" "$IFS_SCAN_MODE" "$image_id" 485 ifs_test_cpus "$SIBLINGS" "$IFS_ARRAY_BIST_SCAN_MODE"
|
| /linux/Documentation/devicetree/bindings/net/dsa/ |
| H A D | marvell,mv88e6xxx.yaml | 28 This switch and its siblings will be autodetected from 34 This switch and its siblings will be autodetected from 39 This switch and its siblings will be autodetected from
|
| /linux/tools/testing/selftests/landlock/ |
| H A D | scoped_base_variants.h | 74 * Parent + child domain (siblings) 144 * Inherited + parent and child domain (siblings)
|
| /linux/kernel/ |
| H A D | Kconfig.preempt | 153 selection across SMT siblings. When enabled -- see 154 prctl(PR_SCHED_CORE) -- task selection ensures that all SMT siblings
|
| /linux/drivers/iommu/ |
| H A D | fsl_pamu_domain.h | 22 struct list_head link; /* link to domain siblings */
|
| /linux/arch/x86/power/ |
| H A D | hibernate.c | 199 * that SMT siblings are sleeping in hlt, as mwait is not safe in arch_resume_nosmt() 204 * SMT siblings out of hlt, and offline them again so that they in arch_resume_nosmt()
|
| /linux/arch/sparc/include/asm/ |
| H A D | oplib_32.h | 118 * siblings exist. 145 /* Search all siblings starting at the passed node for "name" matching
|
| H A D | oplib_64.h | 186 * siblings exist. 217 /* Search all siblings starting at the passed node for "name" matching
|
| /linux/tools/perf/util/ |
| H A D | dwarf-aux.h | 74 DIE_FIND_CB_SIBLING = 2, /* Search only siblings */ 75 DIE_FIND_CB_CONTINUE = 3, /* Search children and siblings */
|
| H A D | cputopo.h | 68 /* Are the sets of SMT siblings all enabled or all disabled in user_requested_cpus. */
|
| /linux/drivers/nvme/host/ |
| H A D | multipath.c | 282 list_for_each_entry_srcu(ns, &head->list, siblings, in nvme_mpath_revalidate_paths() 316 list_for_each_entry_srcu(ns, &head->list, siblings, in __nvme_find_path() 355 ns = list_next_or_null_rcu(&head->list, &ns->siblings, struct nvme_ns, in nvme_next_ns() 356 siblings); in nvme_next_ns() 359 return list_first_or_null_rcu(&head->list, struct nvme_ns, siblings); in nvme_next_ns() 416 list_for_each_entry_srcu(ns, &head->list, siblings, in nvme_queue_depth_path() 485 list_for_each_entry_srcu(ns, &head->list, siblings, in nvme_available_path() 1205 list_for_each_entry_srcu(ns, &head->list, siblings, in nvme_mpath_add_sysfs_link()
|
| /linux/arch/alpha/kernel/ |
| H A D | gct.c | 39 /* Now walk the tree, siblings first. */ in gct6_find_nodes()
|
| /linux/arch/mips/mm/ |
| H A D | context.c | 239 * If this CPU shares FTLB entries with its siblings and one or more of in check_switch_mmu_context() 240 * those siblings hasn't yet invalidated its TLB following a version in check_switch_mmu_context()
|
| /linux/Documentation/admin-guide/pm/ |
| H A D | intel_epb.rst | 40 example, SMT siblings or cores in one package). For this reason, updating the
|
| /linux/include/linux/ |
| H A D | damon.h | 55 * @list: List head for siblings. 93 * @list: List head for siblings. 180 * @list: List head for siblings. 406 * @list: List head for siblings. 499 * @list: List head for siblings. 675 /* List head for siblings. */
|
| /linux/arch/sparc/prom/ |
| H A D | tree_32.c | 69 /* Return the next sibling of node 'node' or zero if no more siblings 181 /* Search siblings at 'node_start' for a node with name
|
| /linux/Documentation/networking/devlink/ |
| H A D | devlink-port.rst | 402 Allows for usage of strict priority arbiter among siblings. This 410 siblings. This arbitration scheme can be used simultaneously with the 412 BW relative to its siblings. Values are relative like a percentage 414 its siblings.
|
| /linux/Documentation/devicetree/bindings/display/mediatek/ |
| H A D | mediatek,wdma.yaml | 16 WDMA device node must be siblings to the central MMSYS_CONFIG node.
|