Home
last modified time | relevance | path

Searched full:siblings (Results 1 – 25 of 174) sorted by relevance

1234567

/linux/tools/testing/selftests/cgroup/
H A Dmemcg_protection.m6 % hierarchy to illustrate how overcommitted protection spreads among siblings
9 % Simulation assumes siblings consumed the initial amount of memory (w/out
16 % n vector nominal protection of siblings set at the given level (memory.low)
48 siblings = sum(u); variable
52 e = protected * min(1, E / siblings); % normalize overcommit
55 unclaimed = max(0, E - siblings);
56 parent_overuse = sum(c) - siblings;
79 % XXX here I do parallel reclaim of all siblings
/linux/Documentation/admin-guide/hw-vuln/
H A Dcore-scheduling.rst100 siblings of a core such that all the selected tasks running on a core are
107 the sibling has the task enqueued. For rest of the siblings in the core,
112 Once a task has been selected for all the siblings in the core, an IPI is sent to
113 siblings for whom a new task was selected. Siblings on receiving the IPI will
125 siblings could be forced to select a lower priority task if the highest
157 and are considered system-wide trusted. The forced-idling of siblings running
174 the siblings to switch to the new task. But there could be hardware delays in
176 cause an attacker task to start running on a CPU before its siblings receive the
177 IPI. Even though cache is flushed on entry to user mode, victim tasks on siblings
185 Core scheduling cannot protect against MDS attacks between the siblings
[all …]
/linux/drivers/infiniband/hw/irdma/
H A Dws.c133 list_for_each_entry(node, &parent->child_list_head, siblings) { in ws_find_node()
139 list_for_each_entry(node, &parent->child_list_head, siblings) { in ws_find_node()
214 list_del(&tc_node->siblings); in irdma_remove_leaf()
219 list_del(&vsi_node->siblings); in irdma_remove_leaf()
295 list_add(&vsi_node->siblings, &ws_tree_root->child_list_head); in irdma_ws_add()
322 list_add(&tc_node->siblings, &vsi_node->child_list_head); in irdma_ws_add()
356 list_del(&tc_node->siblings); in irdma_ws_add()
362 list_del(&vsi_node->siblings); in irdma_ws_add()
H A Dws.h19 struct list_head siblings; member
/linux/drivers/gpu/drm/i915/gt/uc/
H A Dselftest_guc_multi_lrc.c34 struct intel_engine_cs *siblings[MAX_ENGINE_INSTANCE + 1]; in multi_lrc_create_parent() local
43 siblings[i++] = engine; in multi_lrc_create_parent()
49 logical_sort(siblings, i); in multi_lrc_create_parent()
51 return intel_engine_create_parallel(siblings, 1, i); in multi_lrc_create_parent()
/linux/drivers/gpio/
H A Dgpio-sim.c593 struct list_head siblings; member
627 struct list_head siblings; member
724 list_for_each_entry(line, &bank->line_list, siblings) { in gpio_sim_get_line_names_size()
739 list_for_each_entry(line, &bank->line_list, siblings) { in gpio_sim_set_line_names()
752 list_for_each_entry(line, &bank->line_list, siblings) { in gpio_sim_get_reserved_ranges_size()
768 list_for_each_entry(line, &bank->line_list, siblings) { in gpio_sim_set_reserved_ranges()
802 list_for_each_entry(bank, &dev->bank_list, siblings) { in gpio_sim_add_hogs()
803 list_for_each_entry(line, &bank->line_list, siblings) { in gpio_sim_add_hogs()
820 list_for_each_entry(bank, &dev->bank_list, siblings) { in gpio_sim_add_hogs()
821 list_for_each_entry(line, &bank->line_list, siblings) { in gpio_sim_add_hogs()
[all …]
H A Dgpio-virtuser.c1012 struct list_head siblings; member
1031 struct list_head siblings; member
1374 list_for_each_entry(lookup, &dev->lookup_list, siblings) in gpio_virtuser_get_lookup_count()
1399 list_for_each_entry(lookup, &dev->lookup_list, siblings) { in gpio_virtuser_make_lookup_table()
1401 list_for_each_entry(entry, &lookup->entry_list, siblings) { in gpio_virtuser_make_lookup_table()
1440 list_for_each_entry(lookup, &dev->lookup_list, siblings) in gpio_virtuser_make_device_swnode()
1513 list_for_each_entry(lookup, &dev->lookup_list, siblings) { in gpio_virtuser_device_lockup_configfs()
1514 list_for_each_entry(entry, &lookup->entry_list, siblings) { in gpio_virtuser_device_lockup_configfs()
1576 list_del(&entry->siblings); in gpio_virtuser_lookup_entry_config_group_release()
1616 list_add_tail(&entry->siblings, &lookup->entry_list); in gpio_virtuser_make_lookup_entry_group()
[all …]
/linux/drivers/gpu/drm/i915/gt/
H A Dselftest_execlists.c3713 struct intel_engine_cs **siblings, in nop_virtual_engine() argument
3730 ve[n] = intel_engine_create_virtual(siblings, nsibling, 0); in nop_virtual_engine()
3841 struct intel_engine_cs **siblings, in __select_siblings() argument
3854 siblings[n++] = gt->engine_class[class][inst]; in __select_siblings()
3863 struct intel_engine_cs **siblings) in select_siblings() argument
3865 return __select_siblings(gt, class, siblings, NULL); in select_siblings()
3871 struct intel_engine_cs *siblings[MAX_ENGINE_INSTANCE + 1]; in live_virtual_engine() local
3892 nsibling = select_siblings(gt, class, siblings); in live_virtual_engine()
3897 err = nop_virtual_engine(gt, siblings, nsibling, in live_virtual_engine()
3903 err = nop_virtual_engine(gt, siblings, nsibling, n, CHAIN); in live_virtual_engine()
[all …]
H A Dintel_execlists_submission.c204 struct intel_engine_cs *siblings[]; member
214 execlists_create_virtual(struct intel_engine_cs **siblings, unsigned int count,
635 * We have to kick all the siblings again in case we need to in __execlists_schedule_out()
1051 if (likely(engine == ve->siblings[0])) in virtual_xfer_context()
1065 if (ve->siblings[n] == engine) { in virtual_xfer_context()
1066 swap(ve->siblings[n], ve->siblings[0]); in virtual_xfer_context()
1440 str_yes_no(engine != ve->siblings[0])); in execlists_dequeue()
1462 * ve->siblings[] on an idle context, where in execlists_dequeue()
1463 * we may be using ve->siblings[] in in execlists_dequeue()
1467 GEM_BUG_ON(ve->siblings[0] != engine); in execlists_dequeue()
[all …]
/linux/drivers/gpu/drm/i915/gem/
H A Di915_gem_context.c206 kfree(pc->user_engines[i].siblings); in proto_context_close()
407 struct intel_engine_cs **siblings; in set_proto_ctx_engines_balance() local
445 siblings = kmalloc_objs(*siblings, num_siblings); in set_proto_ctx_engines_balance()
446 if (!siblings) in set_proto_ctx_engines_balance()
457 siblings[n] = intel_engine_lookup_user(i915, in set_proto_ctx_engines_balance()
460 if (!siblings[n]) { in set_proto_ctx_engines_balance()
471 set->engines[idx].engine = siblings[0]; in set_proto_ctx_engines_balance()
472 kfree(siblings); in set_proto_ctx_engines_balance()
476 set->engines[idx].siblings = siblings; in set_proto_ctx_engines_balance()
482 kfree(siblings); in set_proto_ctx_engines_balance()
[all …]
/linux/include/uapi/linux/
H A Dmembarrier.h70 * threads siblings have passed through a state
94 * call, that all its running threads siblings
120 * siblings have any currently running rseq
/linux/tools/testing/selftests/drivers/platform/x86/intel/ifs/
H A Dtest_ifs.sh32 readonly SIBLINGS="siblings"
381 "$SIBLINGS")
446 ifs_test_cpus "$SIBLINGS" "$IFS_SCAN_MODE" "$image_id"
485 ifs_test_cpus "$SIBLINGS" "$IFS_ARRAY_BIST_SCAN_MODE"
/linux/tools/testing/selftests/landlock/
H A Dscoped_base_variants.h74 * Parent + child domain (siblings)
144 * Inherited + parent and child domain (siblings)
/linux/kernel/
H A DKconfig.preempt156 selection across SMT siblings. When enabled -- see
157 prctl(PR_SCHED_CORE) -- task selection ensures that all SMT siblings
/linux/arch/x86/power/
H A Dhibernate.c199 * that SMT siblings are sleeping in hlt, as mwait is not safe in arch_resume_nosmt()
204 * SMT siblings out of hlt, and offline them again so that they in arch_resume_nosmt()
/linux/drivers/nvme/host/
H A Dmultipath.c282 list_for_each_entry_srcu(ns, &head->list, siblings, in nvme_mpath_revalidate_paths()
316 list_for_each_entry_srcu(ns, &head->list, siblings, in __nvme_find_path()
355 ns = list_next_or_null_rcu(&head->list, &ns->siblings, struct nvme_ns, in nvme_next_ns()
356 siblings); in nvme_next_ns()
359 return list_first_or_null_rcu(&head->list, struct nvme_ns, siblings); in nvme_next_ns()
416 list_for_each_entry_srcu(ns, &head->list, siblings, in nvme_queue_depth_path()
485 list_for_each_entry_srcu(ns, &head->list, siblings, in nvme_available_path()
1205 list_for_each_entry_srcu(ns, &head->list, siblings, in nvme_mpath_add_sysfs_link()
/linux/arch/sparc/include/asm/
H A Doplib_32.h118 * siblings exist.
145 /* Search all siblings starting at the passed node for "name" matching
H A Doplib_64.h186 * siblings exist.
217 /* Search all siblings starting at the passed node for "name" matching
/linux/tools/perf/util/
H A Ddwarf-aux.h74 DIE_FIND_CB_SIBLING = 2, /* Search only siblings */
75 DIE_FIND_CB_CONTINUE = 3, /* Search children and siblings */
/linux/drivers/iommu/
H A Dfsl_pamu_domain.h22 struct list_head link; /* link to domain siblings */
/linux/arch/alpha/kernel/
H A Dgct.c39 /* Now walk the tree, siblings first. */ in gct6_find_nodes()
/linux/arch/mips/mm/
H A Dcontext.c239 * If this CPU shares FTLB entries with its siblings and one or more of in check_switch_mmu_context()
240 * those siblings hasn't yet invalidated its TLB following a version in check_switch_mmu_context()
/linux/Documentation/networking/devlink/
H A Ddevlink-port.rst402 Allows for usage of strict priority arbiter among siblings. This
410 siblings. This arbitration scheme can be used simultaneously with the
412 BW relative to its siblings. Values are relative like a percentage
414 its siblings.
/linux/arch/sparc/prom/
H A Dtree_32.c69 /* Return the next sibling of node 'node' or zero if no more siblings
181 /* Search siblings at 'node_start' for a node with name
/linux/arch/powerpc/kernel/
H A Dsmp.c118 * the set its siblings that share the L1-cache.
124 * corresponds to the set its siblings within the core that share the
1452 /* Verify that L1-cache siblings are a subset of L2 cache-siblings */ in update_mask_by_l2()
1455 pr_warn_once("CPU %d : Inconsistent L1 and L2 cache siblings\n", in update_mask_by_l2()
1464 /* Assume only core siblings share cache with this CPU */ in update_mask_by_l2()
1554 /* Assume only siblings are part of this CPU's coregroup */ in update_coregroup_mask()

1234567