| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | test_core_reloc_kernel.c | 31 struct task_struct *group_leader; member 73 group_leader, in test_core_kernel() 76 group_leader, group_leader, in test_core_kernel() 79 group_leader, group_leader, group_leader, in test_core_kernel() 82 group_leader, group_leader, group_leader, in test_core_kernel() 83 group_leader, in test_core_kernel() 86 group_leader, group_leader, group_leader, in test_core_kernel() 87 group_leader, group_leader, in test_core_kernel() 90 group_leader, group_leader, group_leader, in test_core_kernel() 91 group_leader, group_leader, group_leader, in test_core_kernel() [all …]
|
| H A D | rcu_read_lock.c | 334 struct task_struct *task, *group_leader; 338 /* the pointer group_leader is explicitly marked as trusted */ in global_subprog() 339 group_leader = task->real_parent->group_leader; in global_subprog() 340 (void)bpf_task_storage_get(&map_a, group_leader, 0, 0); in global_subprog() 284 struct task_struct *task, *group_leader; task_trusted_non_rcuptr() local
|
| H A D | task_kfunc_success.c | 360 acquired = bpf_task_acquire(task->group_leader); in BPF_PROG()
|
| /linux/security/yama/ |
| H A D | yama_lsm.c | 235 myself = current->group_leader; in yama_task_prctl() 278 parent = rcu_dereference(parent->group_leader); in task_is_descendant() 281 walker = rcu_dereference(walker->group_leader); in task_is_descendant() 322 tracee = rcu_dereference(tracee->group_leader); in ptracer_exception_found()
|
| /linux/kernel/events/ |
| H A D | core.c | 735 struct perf_event *leader = event->group_leader; in __perf_effective_state() 1129 struct perf_event *group_leader) in perf_cgroup_connect() argument 1156 if (group_leader && group_leader->cgrp != cgrp) { in perf_cgroup_connect() 1234 struct perf_event *group_leader) in perf_cgroup_connect() argument 1712 if (event->group_leader != event) in get_event_type() 1713 event = event->group_leader; in get_event_type() 1990 if (event->group_leader == event) { in list_add_event() 2099 event->group_leader->nr_siblings); in perf_event__header_size() 2143 struct perf_event *sibling, *group_leader = event->group_leader; in perf_event_validate_size() local 2146 group_leader->nr_siblings + 1) > 16*1024) in perf_event_validate_size() [all …]
|
| /linux/kernel/ |
| H A D | sys.c | 1117 struct task_struct *group_leader = current->group_leader; in SYSCALL_DEFINE2() local 1123 pid = task_pid_vnr(group_leader); in SYSCALL_DEFINE2() 1144 if (same_thread_group(p->real_parent, group_leader)) { in SYSCALL_DEFINE2() 1146 if (task_session(p) != task_session(group_leader)) in SYSCALL_DEFINE2() 1153 if (p != group_leader) in SYSCALL_DEFINE2() 1167 if (!g || task_session(g) != task_session(group_leader)) in SYSCALL_DEFINE2() 1259 struct task_struct *curr = current->group_leader; in set_special_pids() 1270 struct task_struct *group_leader = current->group_leader; in ksys_setsid() local [all...] |
| H A D | fork.c | 1727 task_lock(current->group_leader); in copy_signal() 1729 task_unlock(current->group_leader); in copy_signal() 2294 p->group_leader = current->group_leader; in copy_process() 2297 p->group_leader = p; in copy_process() 2378 p->exit_signal = current->group_leader->exit_signal; in copy_process() 2964 leader = top = top->group_leader; in walk_process_tree() 2983 leader = parent->group_leader; in walk_process_tree()
|
| H A D | tsacct.c | 34 delta = now_ns - tsk->group_leader->start_time; in bacct_add_tsk()
|
| /linux/drivers/gpu/drm/scheduler/ |
| H A D | sched_entity.c | 73 entity->last_user = current->group_leader; in drm_sched_entity_init() 309 last_user = cmpxchg(&entity->last_user, current->group_leader, NULL); in drm_sched_entity_flush() 310 if (last_user == current->group_leader && in drm_sched_entity_flush() 592 WRITE_ONCE(entity->last_user, current->group_leader); in drm_sched_entity_push_job()
|
| /linux/lib/ |
| H A D | is_single_threaded.c | 33 if (unlikely(p == task->group_leader)) in current_is_single_threaded()
|
| /linux/fs/ |
| H A D | exec.c | 267 task_lock(current->group_leader); in bprm_mm_init() 269 task_unlock(current->group_leader); in bprm_mm_init() 943 struct task_struct *leader = tsk->group_leader; in de_thread() 995 tsk->group_leader = tsk; in de_thread() 996 leader->group_leader = tsk; in de_thread() 1338 task_lock(current->group_leader); in finalize_exec() 1340 task_unlock(current->group_leader); in finalize_exec()
|
| /linux/tools/bpf/bpftool/skeleton/ |
| H A D | pid_iter.bpf.c | 118 task->group_leader->comm); in iter()
|
| /linux/drivers/perf/hisilicon/ |
| H A D | hisi_pcie_pmu.c | 336 struct perf_event *sibling, *leader = event->group_leader; in hisi_pcie_pmu_validate_event_group() 350 for_each_sibling_event(sibling, event->group_leader) { in hisi_pcie_pmu_validate_event_group() 439 sibling->group_leader == event->group_leader) in hisi_pcie_pmu_get_event_idx()
|
| H A D | hisi_uncore_pmu.c | 99 struct perf_event *sibling, *leader = event->group_leader; in hisi_validate_event_group() 117 for_each_sibling_event(sibling, event->group_leader) { in hisi_validate_event_group()
|
| /linux/drivers/perf/ |
| H A D | arm_smmuv3_pmu.c | 426 if (!is_software_event(event->group_leader)) { in smmu_pmu_event_init() 427 if (!smmu_pmu_events_compatible(event->group_leader, event)) in smmu_pmu_event_init() 442 if (event->group_leader == event) in smmu_pmu_event_init() 445 for_each_sibling_event(sibling, event->group_leader) { in smmu_pmu_event_init()
|
| H A D | arm_dmc620_pmu.c | 547 if (event->group_leader == event) in dmc620_pmu_event_init() 554 if (!is_software_event(event->group_leader)) in dmc620_pmu_event_init() 557 for_each_sibling_event(sibling, event->group_leader) { in dmc620_pmu_event_init()
|
| H A D | alibaba_uncore_drw_pmu.c | 551 if (event->group_leader != event && in ali_drw_pmu_event_init() 552 !is_software_event(event->group_leader)) { in ali_drw_pmu_event_init() 557 for_each_sibling_event(sibling, event->group_leader) { in ali_drw_pmu_event_init()
|
| H A D | fsl_imx9_ddr_perf.c | 579 if (event->group_leader->pmu != event->pmu && in ddr_perf_event_init() 580 !is_software_event(event->group_leader)) in ddr_perf_event_init() 583 for_each_sibling_event(sibling, event->group_leader) { in ddr_perf_event_init()
|
| /linux/security/apparmor/ |
| H A D | resource.c | 202 (void) update_rlimit_cpu(current->group_leader, in __aa_transition_rlimits()
|
| /linux/arch/x86/kernel/fpu/ |
| H A D | core.c | 233 fpuperm = &x86_task_fpu(current->group_leader)->guest_perm; in fpu_lock_guest_permissions() 623 struct fpu *src_fpu = x86_task_fpu(current->group_leader); in fpu_inherit_perms()
|
| H A D | xstate.c | 1653 struct task_struct *thread, *leader = current->group_leader; in validate_sigaltstack() 1677 struct fpu *fpu = x86_task_fpu(current->group_leader); in __xstate_request_perm() 1784 fpu = x86_task_fpu(current->group_leader); in __xfd_enable_feature()
|
| /linux/drivers/dma/idxd/ |
| H A D | perfmon.c | 127 struct perf_event *leader = event->group_leader; in perfmon_validate_group() 192 if (event->group_leader != event) in perfmon_pmu_event_init()
|
| /linux/drivers/iommu/iommufd/ |
| H A D | ioas.c | 462 .source_task = current->group_leader, in charge_current() 503 pages->source_task = current->group_leader; in change_mm()
|
| /linux/arch/loongarch/kernel/ |
| H A D | perf_event.c | 644 struct perf_event *sibling, *leader = event->group_leader; in validate_group() 816 if (event->group_leader != event) in __hw_perf_event_init()
|
| /linux/arch/x86/events/ |
| H A D | perf_event.h | 127 return check_leader_group(event->group_leader, PERF_X86_EVENT_BRANCH_COUNTERS); in is_branch_counters_group() 132 return check_leader_group(event->group_leader, PERF_X86_EVENT_PEBS_CNTR); in is_pebs_counter_event_group() 137 return check_leader_group(event->group_leader, PERF_X86_EVENT_ACR); in is_acr_event_group()
|