| /linux/net/sunrpc/ |
| H A D | sched.c | 45 static void rpc_release_task(struct rpc_task *task); 68 bool rpc_task_set_rpc_status(struct rpc_task *task, int rpc_status) in rpc_task_set_rpc_status() argument 70 if (cmpxchg(&task->tk_rpc_status, 0, rpc_status) == 0) in rpc_task_set_rpc_status() 76 rpc_task_timeout(const struct rpc_task *task) in rpc_task_timeout() argument 78 unsigned long timeout = READ_ONCE(task->tk_timeout); in rpc_task_timeout() 90 * Disable the timer for a given RPC task. Should be called with 95 __rpc_disable_timer(struct rpc_wait_queue *queue, struct rpc_task *task) in __rpc_disable_timer() argument 97 if (list_empty(&task->u.tk_wait.timer_list)) in __rpc_disable_timer() 99 task->tk_timeout = 0; in __rpc_disable_timer() 100 list_del(&task->u.tk_wait.timer_list); in __rpc_disable_timer() [all …]
|
| H A D | clnt.c | 53 static void call_start(struct rpc_task *task); 54 static void call_reserve(struct rpc_task *task); 55 static void call_reserveresult(struct rpc_task *task); 56 static void call_allocate(struct rpc_task *task); 57 static void call_encode(struct rpc_task *task); 58 static void call_decode(struct rpc_task *task); 59 static void call_bind(struct rpc_task *task); 60 static void call_bind_status(struct rpc_task *task); 61 static void call_transmit(struct rpc_task *task); 62 static void call_status(struct rpc_task *task); [all …]
|
| H A D | xprt.c | 75 static void xprt_request_init(struct rpc_task *task); 258 * @task: task that is requesting access to the transport 265 int xprt_reserve_xprt(struct rpc_xprt *xprt, struct rpc_task *task) in xprt_reserve_xprt() argument 267 struct rpc_rqst *req = task->tk_rqstp; in xprt_reserve_xprt() 270 if (task == xprt->snd_task) in xprt_reserve_xprt() 276 xprt->snd_task = task; in xprt_reserve_xprt() 279 trace_xprt_reserve_xprt(xprt, task); in xprt_reserve_xprt() 285 task->tk_status = -EAGAIN; in xprt_reserve_xprt() 286 if (RPC_IS_SOFT(task) || RPC_IS_SOFTCONN(task)) in xprt_reserve_xprt() 287 rpc_sleep_on_timeout(&xprt->sending, task, NULL, in xprt_reserve_xprt() [all …]
|
| /linux/drivers/gpu/drm/exynos/ |
| H A D | exynos_drm_ipp.c | 87 WARN_ON(ipp->task); in exynos_drm_ipp_unregister() 261 struct exynos_drm_ipp_task *task; in exynos_drm_ipp_task_alloc() local 263 task = kzalloc_obj(*task); in exynos_drm_ipp_task_alloc() 264 if (!task) in exynos_drm_ipp_task_alloc() 267 task->dev = ipp->dev; in exynos_drm_ipp_task_alloc() 268 task->ipp = ipp; in exynos_drm_ipp_task_alloc() 271 task->src.rect.w = task->dst.rect.w = UINT_MAX; in exynos_drm_ipp_task_alloc() 272 task->src.rect.h = task->dst.rect.h = UINT_MAX; in exynos_drm_ipp_task_alloc() 273 task->transform.rotation = DRM_MODE_ROTATE_0; in exynos_drm_ipp_task_alloc() 275 DRM_DEV_DEBUG_DRIVER(task->dev, "Allocated task %p\n", task); in exynos_drm_ipp_task_alloc() [all …]
|
| /linux/include/asm-generic/ |
| H A D | syscall.h | 12 * and only when the caller is sure that the task of interest 23 * syscall_get_nr - find what system call a task is executing 24 * @task: task of interest, must be blocked 25 * @regs: task_pt_regs() of @task 27 * If @task is executing a system call or is at system call 29 * If @task is not executing a system call, i.e. it's blocked 36 * It's only valid to call this when @task is known to be blocked. 38 int syscall_get_nr(struct task_struct *task, struct pt_regs *regs); 41 * syscall_set_nr - change the system call a task is executing 42 * @task: task of interest, must be blocked [all …]
|
| /linux/kernel/bpf/ |
| H A D | task_iter.c | 39 struct task_struct *task; in task_group_seq_get_next() local 46 task = get_pid_task(pid, PIDTYPE_TGID); in task_group_seq_get_next() 47 if (!task) in task_group_seq_get_next() 53 return task; in task_group_seq_get_next() 58 * same for task_seq_start() to pick up the correct task. in task_group_seq_get_next() 62 task = get_pid_task(pid, PIDTYPE_PID); in task_group_seq_get_next() 64 return task; in task_group_seq_get_next() 67 task = find_task_by_pid_ns(common->pid_visiting, common->ns); in task_group_seq_get_next() 68 if (!task) in task_group_seq_get_next() 72 task = __next_thread(task); in task_group_seq_get_next() [all …]
|
| /linux/drivers/scsi/aic94xx/ |
| H A D | aic94xx_tmf.c | 3 * Aic94xx Task Management Functions 214 static int asd_clear_nexus_tag(struct sas_task *task) in asd_clear_nexus_tag() argument 216 struct asd_ha_struct *asd_ha = task->dev->port->ha->lldd_ha; in asd_clear_nexus_tag() 217 struct asd_ascb *tascb = task->lldd_task; in asd_clear_nexus_tag() 221 memcpy(scb->clear_nexus.ssp_task.lun, task->ssp_task.LUN, 8); in asd_clear_nexus_tag() 223 if (task->dev->tproto) in asd_clear_nexus_tag() 225 task->dev->lldd_dev); in asd_clear_nexus_tag() 229 static int asd_clear_nexus_index(struct sas_task *task) in asd_clear_nexus_index() argument 231 struct asd_ha_struct *asd_ha = task->dev->port->ha->lldd_ha; in asd_clear_nexus_index() 232 struct asd_ascb *tascb = task->lldd_task; in asd_clear_nexus_index() [all …]
|
| /linux/drivers/media/i2c/ |
| H A D | saa711x_regs.h | 77 /* Task independent global settings */ 87 /* Task A definition */ 134 /* Task B definition */ 361 /* Task independent global settings: R_80_GLOBAL_CNTL_1 to R_8F_STATUS_INFO_SCALER */ 383 /* Task A definition: R_90_A_TASK_HANDLING_CNTL to R_BF_A_VERT_LUMA_PHASE_OFF_11 */ 384 /* Task A: Basic settings and acquisition window definition */ 386 "Task A: Task handling control"}, 388 "Task A: X port formats and configuration"}, 390 "Task A: X port input reference signal definition"}, 392 "Task A: I port output formats and configuration"}, [all …]
|
| /linux/drivers/scsi/pm8001/ |
| H A D | pm8001_sas.c | 46 * pm8001_find_tag - from sas task to find out tag that belongs to this task 47 * @task: the task sent to the LLDD 48 * @tag: the found tag associated with the task 50 static int pm8001_find_tag(struct sas_task *task, u32 *tag) in pm8001_find_tag() argument 52 if (task->lldd_task) { in pm8001_find_tag() 54 ccb = task->lldd_task; in pm8001_find_tag() 64 * @tag: the found tag associated with the task 80 * pm8001_tag_alloc - allocate a empty tag for task used. 104 static void pm80xx_get_tag_opcodes(struct sas_task *task, int *ata_op, in pm80xx_get_tag_opcodes() argument 114 if (!task) in pm80xx_get_tag_opcodes() [all …]
|
| /linux/drivers/scsi/isci/ |
| H A D | task.c | 64 #include "task.h" 71 * @task: request to complete 72 * @response: response code for the completed task. 73 * @status: status code for the completed task. 76 static void isci_task_refuse(struct isci_host *ihost, struct sas_task *task, in isci_task_refuse() argument 84 dev_dbg(&ihost->pdev->dev, "%s: task = %p, response=%d, status=%d\n", in isci_task_refuse() 85 __func__, task, response, status); in isci_task_refuse() 87 spin_lock_irqsave(&task->task_state_lock, flags); in isci_task_refuse() 89 task->task_status.resp = response; in isci_task_refuse() 90 task->task_status.stat = status; in isci_task_refuse() [all …]
|
| /linux/arch/powerpc/kernel/ |
| H A D | signal.h | 36 struct task_struct *task); 38 struct task_struct *task); 39 extern unsigned long copy_vsx_from_user(struct task_struct *task, 41 extern unsigned long copy_ckvsx_from_user(struct task_struct *task, 43 unsigned long copy_fpr_to_user(void __user *to, struct task_struct *task); 44 unsigned long copy_ckfpr_to_user(void __user *to, struct task_struct *task); 45 unsigned long copy_fpr_from_user(struct task_struct *task, void __user *from); 46 unsigned long copy_ckfpr_from_user(struct task_struct *task, void __user *from); 48 #define unsafe_copy_fpr_to_user(to, task, label) do { \ argument 49 struct task_struct *__t = task; \ [all …]
|
| /linux/scripts/gdb/linux/ |
| H A D | tasks.py | 4 # task & thread tools 39 for task in task_lists(): 40 if int(task['pid']) == pid: 41 return task 46 """Find Linux task by PID and return the task_struct variable. 55 task = get_task_by_pid(pid) 56 if task: 57 return task.dereference() 59 raise gdb.GdbError("No task of PID " + str(pid)) 72 gdb.write("{:>10} {:>12} {:>7}\n".format("TASK", "PI 86 get_thread_info(task) global() argument 103 invoke(self, task) global() argument [all...] |
| /linux/kernel/livepatch/ |
| H A D | transition.c | 85 struct task_struct *g, *task; 122 for_each_process_thread(g, task) { in klp_complete_transition() 123 WARN_ON_ONCE(test_tsk_thread_flag(task, TIF_PATCH_PENDING)); in klp_complete_transition() 124 task->patch_state = KLP_TRANSITION_IDLE; in klp_complete_transition() 129 task = idle_task(cpu); in klp_complete_transition() 130 WARN_ON_ONCE(test_tsk_thread_flag(task, TIF_PATCH_PENDING)); in klp_complete_transition() 131 task->patch_state = KLP_TRANSITION_IDLE; in klp_complete_transition() 169 * Switch the patched state of the task to the set of functions in the target in klp_cancel_transition() 172 * NOTE: If task is not 'current', the caller must ensure the task i in klp_cancel_transition() 94 struct task_struct *g, *task; klp_complete_transition() local 184 klp_update_patch_state(struct task_struct * task) klp_update_patch_state() argument 263 klp_check_stack(struct task_struct * task,const char ** oldname) klp_check_stack() argument 293 klp_check_and_switch_task(struct task_struct * task,void * arg) klp_check_and_switch_task() argument 314 klp_try_switch_task(struct task_struct * task) klp_try_switch_task() argument 410 struct task_struct *g, *task; klp_send_signals() local 454 struct task_struct *g, *task; klp_try_complete_transition() local 532 struct task_struct *g, *task; klp_start_transition() local 575 struct task_struct *g, *task; klp_init_transition() local 651 struct task_struct *g, *task; klp_reverse_transition() local 730 struct task_struct *g, *task; klp_force_transition() local [all...] |
| /linux/fs/proc/ |
| H A D | base.c | 114 * in /proc for a task before it execs a suid executable. 208 static int get_task_root(struct task_struct *task, struct path *root) in get_task_root() argument 212 task_lock(task); in get_task_root() 213 if (task->fs) { in get_task_root() 214 get_fs_root(task->fs, root); in get_task_root() 217 task_unlock(task); in get_task_root() 223 struct task_struct *task = get_proc_task(d_inode(dentry)); in proc_cwd_link() local 226 if (task) { in proc_cwd_link() 227 task_lock(task); in proc_cwd_link() 228 if (task->fs) { in proc_cwd_link() [all …]
|
| H A D | fd.c | 28 struct task_struct *task; in seq_show() local 30 task = get_proc_task(m->private); in seq_show() 31 if (!task) in seq_show() 34 task_lock(task); in seq_show() 35 files = task->files; in seq_show() 51 task_unlock(task); in seq_show() 52 put_task_struct(task); in seq_show() 82 * that the current task has PTRACE_MODE_READ in addition to the normal 89 struct task_struct *task = get_proc_task(inode); in proc_fdinfo_permission() local 91 if (!task) in proc_fdinfo_permission() [all …]
|
| /linux/drivers/md/dm-vdo/indexer/ |
| H A D | radix-sort.c | 45 struct task { struct 60 struct task *end_of_stack; argument 61 struct task insertion_list[256]; 62 struct task stack[]; 72 static inline void insert_key(const struct task task, sort_key_t *next) in insert_key() argument 78 while ((--next >= task.first_key) && in insert_key() 79 (compare(unsorted, next[0], task.offset, task.length) < 0)) in insert_key() 90 static inline void insertion_sort(const struct task task) in insertion_sort() argument 94 for (next = task.first_key + 1; next <= task.last_key; next++) in insertion_sort() 95 insert_key(task, next); in insertion_sort() [all …]
|
| /linux/include/rv/ |
| H A D | ltl_monitor.h | 12 #include <trace/events/task.h> 24 static void ltl_atoms_fetch(struct task_struct *task, struct ltl_monitor *mon); 25 static void ltl_atoms_init(struct task_struct *task, struct ltl_monitor *mon, bool task_creation); 27 static struct ltl_monitor *ltl_get_monitor(struct task_struct *task) in ltl_get_monitor() argument 29 return &task->rv[ltl_monitor_slot].ltl_mon; in ltl_get_monitor() 32 static void ltl_task_init(struct task_struct *task, bool task_creation) in ltl_task_init() argument 34 struct ltl_monitor *mon = ltl_get_monitor(task); in ltl_task_init() 41 ltl_atoms_init(task, mon, task_creation); in ltl_task_init() 42 ltl_atoms_fetch(task, mon); in ltl_task_init() 45 static void handle_task_newtask(void *data, struct task_struct *task, u64 flags) in handle_task_newtask() argument [all …]
|
| /linux/drivers/video/fbdev/ |
| H A D | uvesafb.c | 69 * find the kernel part of the task struct, copy the registers and 70 * the buffer contents and then complete the task. 75 struct uvesafb_ktask *task; in uvesafb_cn_callback() local 84 task = uvfb_tasks[msg->seq]; in uvesafb_cn_callback() 86 if (!task || msg->ack != task->ack) { in uvesafb_cn_callback() 94 if (task->t.buf_len < utask->buf_len || in uvesafb_cn_callback() 103 memcpy(&task->t, utask, sizeof(*utask)); in uvesafb_cn_callback() 105 if (task->t.buf_len && task->buf) in uvesafb_cn_callback() 106 memcpy(task->buf, utask + 1, task->t.buf_len); in uvesafb_cn_callback() 108 complete(task->done); in uvesafb_cn_callback() [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | rcu_read_lock.c | 43 struct task_struct *task; in get_cgroup_id() 46 task = bpf_get_current_task_btf(); in get_cgroup_id() 47 if (task->pid != target_pid) in get_cgroup_id() 52 cgroups = task->cgroups; in task_succ() 64 struct task_struct *task, *real_parent; in task_succ() 68 task = bpf_get_current_task_btf(); in task_succ() 69 if (task->pid != target_pid) in task_succ() 74 real_parent = task->real_parent; in task_succ() 93 struct task_struct *task, *real_parent; in two_regions() 96 task in two_regions() 33 struct task_struct *task; get_cgroup_id() local 54 struct task_struct *task, *real_parent; task_succ() local 83 struct task_struct *task, *real_parent; no_lock() local 95 struct task_struct *task, *real_parent; two_regions() local 114 struct task_struct *task, *real_parent; non_sleepable_1() local 130 struct task_struct *task, *real_parent; non_sleepable_2() local 149 struct task_struct *task, *real_parent, *gparent; task_acquire() local 177 struct task_struct *task; miss_lock() local 191 struct task_struct *task; miss_unlock() local 203 struct task_struct *task, *real_parent; non_sleepable_rcu_mismatch() local 221 struct task_struct *task, *real_parent; inproper_sleepable_helper() local 265 struct task_struct *task, *real_parent; nested_rcu_region() local 284 struct task_struct *task, *group_leader; task_trusted_non_rcuptr() local 298 struct task_struct *task, *real_parent; task_untrusted_rcuptr() local 312 struct task_struct *task, *real_parent; cross_rcu_region() local [all...] |
| /linux/Documentation/scheduler/ |
| H A D | sched-ext.rst | 20 a runnable task stalls, or on invoking the SysRq key sequence 49 If a task explicitly sets its scheduling policy to ``SCHED_EXT``, it will be 111 Whether a given task is on sched_ext can be determined as follows: 130 * Decide which CPU a task should be migrated to before being 133 * then insert the task directly into SCX_DSQ_LOCAL and skip the 157 * Do a direct insertion of a task to the global DSQ. This ops.enqueue() 162 * default ops.enqueue implementation, which just dispatches the task 205 A CPU always executes a task from its local DSQ. A task is "inserted" into a 206 DSQ. A task in a non-local DSQ is "move"d into the target CPU's local DSQ. 208 When a CPU is looking for the next task to run, if the local DSQ is not [all …]
|
| /linux/Documentation/admin-guide/mm/ |
| H A D | numa_memory_policy.rst | 20 both cpusets and policies are applied to a task, the restrictions of the cpuset 44 Task/Process Policy 45 this is an optional, per-task policy. When defined for a 46 specific task, this policy controls all page allocations made 47 by or on behalf of the task that aren't controlled by a more 48 specific scope. If a task does not define a task policy, then 50 task policy "fall back" to the System Default Policy. 52 The task policy applies to the entire address space of a task. Thus, 54 [clone() w/o the CLONE_VM flag] and exec*(). This allows a parent task 55 to establish the task policy for a child task exec()'d from an [all …]
|
| /linux/drivers/scsi/qedi/ |
| H A D | qedi_fw_iscsi.h | 31 * task context. 33 * @param task_params - Pointer to task parameters struct 37 * @param sgl_task_params - Pointer to SGL task params 49 * Request task context. 51 * @param task_params - Pointer to task parameters struct 53 * @param tx_sgl_task_params - Pointer to SGL task params 54 * @param rx_sgl_task_params - Pointer to SGL task params 62 * task context. 64 * @param task_params - Pointer to task parameters struct 66 * @param tx_sgl_task_params - Pointer to SGL task params [all …]
|
| /linux/arch/arm64/kernel/ |
| H A D | fpsimd.c | 60 * (a) for each task, we need to remember which CPU was the last one to have 61 * the task's FPSIMD state loaded into its FPSIMD registers; 62 * (b) for each CPU, we need to remember which task's userland FPSIMD state has 69 * address of the userland FPSIMD state of the task that was loaded onto the CPU 75 * task's fpsimd_cpu are still mutually in sync. If this is the case, we 79 * indicate whether or not the userland FPSIMD state of the current task is 82 * task. If the task is behaving as a VMM, then this is will be managed by 89 * called from softirq context, which will save the task's FPSIMD context back 91 * task's FPSIMD state from task context and thereby corrupting the state, it 92 * is necessary to protect any manipulation of a task's fpsimd_state or [all …]
|
| /linux/tools/perf/util/bpf_skel/ |
| H A D | kwork_top.bpf.c | 105 static __always_inline void update_task_info(struct task_struct *task, __u32 cpu) in update_task_info() argument 108 .pid = task->pid, in update_task_info() 114 .tgid = task->tgid, in update_task_info() 115 .is_kthread = task->flags & PF_KTHREAD ? 1 : 0, in update_task_info() 117 BPF_CORE_READ_STR_INTO(&data.comm, task, comm); in update_task_info() 139 static void on_sched_out(struct task_struct *task, __u64 ts, __u32 cpu) in on_sched_out() argument 144 pelem = bpf_task_storage_get(&kwork_top_task_time, task, NULL, 0); in on_sched_out() 152 .pid = task->pid, in on_sched_out() 153 .task_p = (__u64)task, in on_sched_out() 157 update_task_info(task, cpu); in on_sched_out() [all …]
|
| /linux/kernel/sched/ |
| H A D | ext_internal.h | 116 * By default, if there are no other task to run on the CPU, ext core 117 * keeps running the current task even after its slice expires. If this 124 * An exiting task may schedule after PF_EXITING is set. In such cases, 125 * bpf_task_from_pid() may not be able to find the task and if the BPF 126 * scheduler depends on pid lookup for dispatching, the task will be 143 * A migration disabled task can only execute on its current CPU. By 148 * A migration disabled task never invokes ops.select_cpu() as it can 205 /* the cgroup the task is joining */ 212 /* Whether the task exited before running on sched_ext. */ 228 /* next task is being scheduled by &sched_class_rt */ [all …]
|