| /linux/net/sunrpc/ |
| H A D | sched.c | 45 static void rpc_release_task(struct rpc_task *task); 68 bool rpc_task_set_rpc_status(struct rpc_task *task, int rpc_status) in rpc_task_set_rpc_status() argument 70 if (cmpxchg(&task->tk_rpc_status, 0, rpc_status) == 0) in rpc_task_set_rpc_status() 76 rpc_task_timeout(const struct rpc_task *task) in rpc_task_timeout() argument 78 unsigned long timeout = READ_ONCE(task->tk_timeout); in rpc_task_timeout() 90 * Disable the timer for a given RPC task. Should be called with 95 __rpc_disable_timer(struct rpc_wait_queue *queue, struct rpc_task *task) in __rpc_disable_timer() argument 97 if (list_empty(&task->u.tk_wait.timer_list)) in __rpc_disable_timer() 99 task->tk_timeout = 0; in __rpc_disable_timer() 100 list_del(&task->u.tk_wait.timer_list); in __rpc_disable_timer() [all …]
|
| H A D | clnt.c | 53 static void call_start(struct rpc_task *task); 54 static void call_reserve(struct rpc_task *task); 55 static void call_reserveresult(struct rpc_task *task); 56 static void call_allocate(struct rpc_task *task); 57 static void call_encode(struct rpc_task *task); 58 static void call_decode(struct rpc_task *task); 59 static void call_bind(struct rpc_task *task); 60 static void call_bind_status(struct rpc_task *task); 61 static void call_transmit(struct rpc_task *task); 62 static void call_status(struct rpc_task *task); [all …]
|
| H A D | xprt.c | 75 static void xprt_request_init(struct rpc_task *task); 258 * @task: task that is requesting access to the transport 265 int xprt_reserve_xprt(struct rpc_xprt *xprt, struct rpc_task *task) in xprt_reserve_xprt() argument 267 struct rpc_rqst *req = task->tk_rqstp; in xprt_reserve_xprt() 270 if (task == xprt->snd_task) in xprt_reserve_xprt() 276 xprt->snd_task = task; in xprt_reserve_xprt() 279 trace_xprt_reserve_xprt(xprt, task); in xprt_reserve_xprt() 285 task->tk_status = -EAGAIN; in xprt_reserve_xprt() 286 if (RPC_IS_SOFT(task) || RPC_IS_SOFTCONN(task)) in xprt_reserve_xprt() 287 rpc_sleep_on_timeout(&xprt->sending, task, NULL, in xprt_reserve_xprt() [all …]
|
| /linux/include/asm-generic/ |
| H A D | syscall.h | 12 * and only when the caller is sure that the task of interest 23 * syscall_get_nr - find what system call a task is executing 24 * @task: task of interest, must be blocked 25 * @regs: task_pt_regs() of @task 27 * If @task is executing a system call or is at system call 29 * If @task is not executing a system call, i.e. it's blocked 36 * It's only valid to call this when @task is known to be blocked. 38 int syscall_get_nr(struct task_struct *task, struct pt_regs *regs); 41 * syscall_set_nr - change the system call a task is executing 42 * @task: task of interest, must be blocked [all …]
|
| /linux/drivers/gpu/drm/exynos/ |
| H A D | exynos_drm_ipp.c | 87 WARN_ON(ipp->task); in exynos_drm_ipp_unregister() 261 struct exynos_drm_ipp_task *task; in exynos_drm_ipp_task_alloc() local 263 task = kzalloc_obj(*task); in exynos_drm_ipp_task_alloc() 264 if (!task) in exynos_drm_ipp_task_alloc() 267 task->dev = ipp->dev; in exynos_drm_ipp_task_alloc() 268 task->ipp = ipp; in exynos_drm_ipp_task_alloc() 271 task->src.rect.w = task->dst.rect.w = UINT_MAX; in exynos_drm_ipp_task_alloc() 272 task->src.rect.h = task->dst.rect.h = UINT_MAX; in exynos_drm_ipp_task_alloc() 273 task->transform.rotation = DRM_MODE_ROTATE_0; in exynos_drm_ipp_task_alloc() 275 DRM_DEV_DEBUG_DRIVER(task->dev, "Allocated task %p\n", task); in exynos_drm_ipp_task_alloc() [all …]
|
| /linux/drivers/scsi/aic94xx/ |
| H A D | aic94xx_tmf.c | 3 * Aic94xx Task Management Functions 214 static int asd_clear_nexus_tag(struct sas_task *task) in asd_clear_nexus_tag() argument 216 struct asd_ha_struct *asd_ha = task->dev->port->ha->lldd_ha; in asd_clear_nexus_tag() 217 struct asd_ascb *tascb = task->lldd_task; in asd_clear_nexus_tag() 221 memcpy(scb->clear_nexus.ssp_task.lun, task->ssp_task.LUN, 8); in asd_clear_nexus_tag() 223 if (task->dev->tproto) in asd_clear_nexus_tag() 225 task->dev->lldd_dev); in asd_clear_nexus_tag() 229 static int asd_clear_nexus_index(struct sas_task *task) in asd_clear_nexus_index() argument 231 struct asd_ha_struct *asd_ha = task->dev->port->ha->lldd_ha; in asd_clear_nexus_index() 232 struct asd_ascb *tascb = task->lldd_task; in asd_clear_nexus_index() [all …]
|
| /linux/drivers/media/i2c/ |
| H A D | saa711x_regs.h | 77 /* Task independent global settings */ 87 /* Task A definition */ 134 /* Task B definition */ 361 /* Task independent global settings: R_80_GLOBAL_CNTL_1 to R_8F_STATUS_INFO_SCALER */ 383 /* Task A definition: R_90_A_TASK_HANDLING_CNTL to R_BF_A_VERT_LUMA_PHASE_OFF_11 */ 384 /* Task A: Basic settings and acquisition window definition */ 386 "Task A: Task handling control"}, 388 "Task A: X port formats and configuration"}, 390 "Task A: X port input reference signal definition"}, 392 "Task A: I port output formats and configuration"}, [all …]
|
| /linux/Documentation/scheduler/ |
| H A D | sched-ext.rst | 20 a runnable task stalls, or on invoking the SysRq key sequence 49 If a task explicitly sets its scheduling policy to ``SCHED_EXT``, it will be 120 the task and the core scheduler silently picked a fallback CPU. 123 * ``SCX_EV_DISPATCH_KEEP_LAST``: a task continued running because no other 124 task was available (only when ``SCX_OPS_ENQ_LAST`` is not set). 125 * ``SCX_EV_ENQ_SKIP_EXITING``: an exiting task was dispatched to the local DSQ 127 * ``SCX_EV_ENQ_SKIP_MIGRATION_DISABLED``: a migration-disabled task was 130 * ``SCX_EV_REENQ_IMMED``: a task dispatched with ``SCX_ENQ_IMMED`` was 135 * ``SCX_EV_REFILL_SLICE_DFL``: a task's time slice was refilled with the 140 * ``SCX_EV_INSERT_NOT_OWNED``: attempted to insert a task not owned by this [all …]
|
| /linux/scripts/gdb/linux/ |
| H A D | tasks.py | 4 # task & thread tools 39 for task in task_lists(): 40 if int(task['pid']) == pid: 41 return task 46 """Find Linux task by PID and return the task_struct variable. 55 task = get_task_by_pid(pid) 56 if task: 57 return task.dereference() 59 raise gdb.GdbError("No task of PID " + str(pid)) 72 gdb.write("{:>10} {:>12} {:>7}\n".format("TASK", "PI 86 get_thread_info(task) global() argument 103 invoke(self, task) global() argument [all...] |
| /linux/arch/powerpc/kernel/ |
| H A D | signal.h | 36 struct task_struct *task); 38 struct task_struct *task); 39 extern unsigned long copy_vsx_from_user(struct task_struct *task, 41 extern unsigned long copy_ckvsx_from_user(struct task_struct *task, 43 unsigned long copy_fpr_to_user(void __user *to, struct task_struct *task); 44 unsigned long copy_ckfpr_to_user(void __user *to, struct task_struct *task); 45 unsigned long copy_fpr_from_user(struct task_struct *task, void __user *from); 46 unsigned long copy_ckfpr_from_user(struct task_struct *task, void __user *from); 48 #define unsafe_copy_fpr_to_user(to, task, label) do { \ argument 49 struct task_struct *__t = task; \ [all …]
|
| /linux/drivers/scsi/pm8001/ |
| H A D | pm8001_sas.c | 46 * pm8001_find_tag - from sas task to find out tag that belongs to this task 47 * @task: the task sent to the LLDD 48 * @tag: the found tag associated with the task 50 static int pm8001_find_tag(struct sas_task *task, u32 *tag) in pm8001_find_tag() argument 52 if (task->lldd_task) { in pm8001_find_tag() 54 ccb = task->lldd_task; in pm8001_find_tag() 64 * @tag: the found tag associated with the task 80 * pm8001_tag_alloc - allocate a empty tag for task used. 104 static void pm80xx_get_tag_opcodes(struct sas_task *task, int *ata_op, in pm80xx_get_tag_opcodes() argument 114 if (!task) in pm80xx_get_tag_opcodes() [all …]
|
| /linux/drivers/scsi/isci/ |
| H A D | task.c | 64 #include "task.h" 71 * @task: request to complete 72 * @response: response code for the completed task. 73 * @status: status code for the completed task. 76 static void isci_task_refuse(struct isci_host *ihost, struct sas_task *task, in isci_task_refuse() argument 84 dev_dbg(&ihost->pdev->dev, "%s: task = %p, response=%d, status=%d\n", in isci_task_refuse() 85 __func__, task, response, status); in isci_task_refuse() 87 spin_lock_irqsave(&task->task_state_lock, flags); in isci_task_refuse() 89 task->task_status.resp = response; in isci_task_refuse() 90 task->task_status.stat = status; in isci_task_refuse() [all …]
|
| /linux/kernel/livepatch/ |
| H A D | transition.c | 85 struct task_struct *g, *task; 122 for_each_process_thread(g, task) { in klp_complete_transition() 123 WARN_ON_ONCE(test_tsk_thread_flag(task, TIF_PATCH_PENDING)); in klp_complete_transition() 124 task->patch_state = KLP_TRANSITION_IDLE; in klp_complete_transition() 129 task = idle_task(cpu); in klp_complete_transition() 130 WARN_ON_ONCE(test_tsk_thread_flag(task, TIF_PATCH_PENDING)); in klp_complete_transition() 131 task->patch_state = KLP_TRANSITION_IDLE; in klp_complete_transition() 169 * Switch the patched state of the task to the set of functions in the target in klp_cancel_transition() 172 * NOTE: If task is not 'current', the caller must ensure the task i in klp_cancel_transition() 94 struct task_struct *g, *task; klp_complete_transition() local 184 klp_update_patch_state(struct task_struct * task) klp_update_patch_state() argument 263 klp_check_stack(struct task_struct * task,const char ** oldname) klp_check_stack() argument 293 klp_check_and_switch_task(struct task_struct * task,void * arg) klp_check_and_switch_task() argument 314 klp_try_switch_task(struct task_struct * task) klp_try_switch_task() argument 410 struct task_struct *g, *task; klp_send_signals() local 454 struct task_struct *g, *task; klp_try_complete_transition() local 532 struct task_struct *g, *task; klp_start_transition() local 575 struct task_struct *g, *task; klp_init_transition() local 651 struct task_struct *g, *task; klp_reverse_transition() local 730 struct task_struct *g, *task; klp_force_transition() local [all...] |
| /linux/arch/riscv/include/asm/ |
| H A D | usercfi.h | 36 void set_shstk_base(struct task_struct *task, unsigned long shstk_addr, unsigned long size); 37 unsigned long get_shstk_base(struct task_struct *task, unsigned long *size); 38 void set_active_shstk(struct task_struct *task, unsigned long shstk_addr); 39 bool is_shstk_enabled(struct task_struct *task); 40 bool is_shstk_locked(struct task_struct *task); 41 bool is_shstk_allocated(struct task_struct *task); 42 void set_shstk_lock(struct task_struct *task, bool lock); 43 void set_shstk_status(struct task_struct *task, bool enable); 44 unsigned long get_active_shstk(struct task_struct *task); 47 bool is_indir_lp_enabled(struct task_struct *task); [all …]
|
| /linux/rust/kernel/ |
| H A D | task.rs | 33 /// Returns the currently running task. 39 // leave current task context: 47 unsafe { &*$crate::task::Task::current() } 75 /// Getting the current task and storing it in some struct. The reference count is automatically 79 /// use kernel::{task::Task, sync::aref::ARef}; 82 /// creator: ARef<Task>, 96 pub struct Task(pub(crate) Opaque<bindings::task_struct>); struct 98 // SAFETY: By design, the only way to access a `Task` is via the `current` function or via an 99 // `ARef<Task>` obtained through the `AlwaysRefCounted` impl. This means that the only situation in 100 // which a `Task` can be accessed mutably is when the refcount drops to zero and the destructor [all …]
|
| /linux/fs/proc/ |
| H A D | base.c | 114 * in /proc for a task before it execs a suid executable. 208 static int get_task_root(struct task_struct *task, struct path *root) in get_task_root() argument 212 task_lock(task); in get_task_root() 213 if (task->fs) { in get_task_root() 214 get_fs_root(task->fs, root); in get_task_root() 217 task_unlock(task); in get_task_root() 223 struct task_struct *task = get_proc_task(d_inode(dentry)); in proc_cwd_link() local 226 if (task) { in proc_cwd_link() 227 task_lock(task); in proc_cwd_link() 228 if (task->fs) { in proc_cwd_link() [all …]
|
| /linux/include/rv/ |
| H A D | ltl_monitor.h | 12 #include <trace/events/task.h> 24 static void ltl_atoms_fetch(struct task_struct *task, struct ltl_monitor *mon); 25 static void ltl_atoms_init(struct task_struct *task, struct ltl_monitor *mon, bool task_creation); 27 static struct ltl_monitor *ltl_get_monitor(struct task_struct *task) in ltl_get_monitor() argument 29 return &task->rv[ltl_monitor_slot].ltl_mon; in ltl_get_monitor() 32 static void ltl_task_init(struct task_struct *task, bool task_creation) in ltl_task_init() argument 34 struct ltl_monitor *mon = ltl_get_monitor(task); in ltl_task_init() 41 ltl_atoms_init(task, mon, task_creation); in ltl_task_init() 42 ltl_atoms_fetch(task, mon); in ltl_task_init() 45 static void handle_task_newtask(void *data, struct task_struct *task, u64 flags) in handle_task_newtask() argument [all …]
|
| /linux/drivers/video/fbdev/ |
| H A D | uvesafb.c | 69 * find the kernel part of the task struct, copy the registers and 70 * the buffer contents and then complete the task. 75 struct uvesafb_ktask *task; in uvesafb_cn_callback() local 84 task = uvfb_tasks[msg->seq]; in uvesafb_cn_callback() 86 if (!task || msg->ack != task->ack) { in uvesafb_cn_callback() 94 if (task->t.buf_len < utask->buf_len || in uvesafb_cn_callback() 103 memcpy(&task->t, utask, sizeof(*utask)); in uvesafb_cn_callback() 105 if (task->t.buf_len && task->buf) in uvesafb_cn_callback() 106 memcpy(task->buf, utask + 1, task->t.buf_len); in uvesafb_cn_callback() 108 complete(task->done); in uvesafb_cn_callback() [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | rcu_read_lock.c | 43 struct task_struct *task; in get_cgroup_id() 46 task = bpf_get_current_task_btf(); in get_cgroup_id() 47 if (task->pid != target_pid) in get_cgroup_id() 52 cgroups = task->cgroups; in task_succ() 64 struct task_struct *task, *real_parent; in task_succ() 68 task = bpf_get_current_task_btf(); in task_succ() 69 if (task->pid != target_pid) in task_succ() 74 real_parent = task->real_parent; in task_succ() 93 struct task_struct *task, *real_parent; in two_regions() 96 task in two_regions() 33 struct task_struct *task; get_cgroup_id() local 54 struct task_struct *task, *real_parent; task_succ() local 83 struct task_struct *task, *real_parent; no_lock() local 95 struct task_struct *task, *real_parent; two_regions() local 114 struct task_struct *task, *real_parent; non_sleepable_1() local 130 struct task_struct *task, *real_parent; non_sleepable_2() local 149 struct task_struct *task, *real_parent, *gparent; task_acquire() local 177 struct task_struct *task; miss_lock() local 191 struct task_struct *task; miss_unlock() local 203 struct task_struct *task, *real_parent; non_sleepable_rcu_mismatch() local 221 struct task_struct *task, *real_parent; inproper_sleepable_helper() local 265 struct task_struct *task, *real_parent; nested_rcu_region() local 284 struct task_struct *task, *group_leader; task_trusted_non_rcuptr() local 298 struct task_struct *task, *real_parent; task_untrusted_rcuptr() local 312 struct task_struct *task, *real_parent; cross_rcu_region() local [all...] |
| /linux/drivers/scsi/qedi/ |
| H A D | qedi_fw_iscsi.h | 31 * task context. 33 * @param task_params - Pointer to task parameters struct 37 * @param sgl_task_params - Pointer to SGL task params 49 * Request task context. 51 * @param task_params - Pointer to task parameters struct 53 * @param tx_sgl_task_params - Pointer to SGL task params 54 * @param rx_sgl_task_params - Pointer to SGL task params 62 * task context. 64 * @param task_params - Pointer to task parameters struct 66 * @param tx_sgl_task_params - Pointer to SGL task params [all …]
|
| /linux/tools/perf/util/bpf_skel/ |
| H A D | kwork_top.bpf.c | 105 static __always_inline void update_task_info(struct task_struct *task, __u32 cpu) in update_task_info() argument 108 .pid = task->pid, in update_task_info() 114 .tgid = task->tgid, in update_task_info() 115 .is_kthread = task->flags & PF_KTHREAD ? 1 : 0, in update_task_info() 117 BPF_CORE_READ_STR_INTO(&data.comm, task, comm); in update_task_info() 139 static void on_sched_out(struct task_struct *task, __u64 ts, __u32 cpu) in on_sched_out() argument 144 pelem = bpf_task_storage_get(&kwork_top_task_time, task, NULL, 0); in on_sched_out() 152 .pid = task->pid, in on_sched_out() 153 .task_p = (__u64)task, in on_sched_out() 157 update_task_info(task, cpu); in on_sched_out() [all …]
|
| /linux/kernel/sched/ |
| H A D | ext_internal.h | 125 * By default, if there are no other task to run on the CPU, ext core 126 * keeps running the current task even after its slice expires. If this 133 * An exiting task may schedule after PF_EXITING is set. In such cases, 134 * bpf_task_from_pid() may not be able to find the task and if the BPF 135 * scheduler depends on pid lookup for dispatching, the task will be 152 * A migration disabled task can only execute on its current CPU. By 157 * A migration disabled task never invokes ops.select_cpu() as it can 215 /* the cgroup the task is joining */ 222 /* Whether the task exited before running on sched_ext. */ 238 /* next task is being scheduled by &sched_class_rt */ [all …]
|
| /linux/arch/riscv/kernel/ |
| H A D | stacktrace.c | 20 * This disables KASAN checking when reading a value from another task's stack, 21 * since the other task could be running on another CPU and could have poisoned 24 #define READ_ONCE_TASK_STACK(task, x) \ argument 28 if ((task) == current) \ 48 void notrace walk_stackframe(struct task_struct *task, struct pt_regs *regs, in walk_stackframe() argument 59 } else if (task == NULL || task == current) { in walk_stackframe() 65 /* task blocked in __switch_to */ in walk_stackframe() 66 fp = task->thread.s[0]; in walk_stackframe() 67 sp = task->thread.sp; in walk_stackframe() 68 pc = task->thread.ra; in walk_stackframe() [all …]
|
| /linux/arch/s390/include/asm/ |
| H A D | unwind.h | 14 * There four combinations for task and regs: 15 * 1) task==NULL, regs==NULL: the unwind starts for the task that is currently 17 * 2) task==NULL, regs!=NULL: the unwind starts from the sp/ip found in 18 * the struct pt_regs of an interrupt frame for the current task 19 * 3) task!=NULL, regs==NULL: the unwind starts for an inactive task with 20 * the sp picked up from task->thread.ksp and the ip picked up from the 22 * 4) task!=NULL, regs!=NULL: the sp/ip are picked up from the interrupt 23 * frame 'regs' of a inactive task 37 struct task_struct *task; member 50 ip = ftrace_graph_ret_addr(state->task, &state->graph_idx, ip, (void *)state->sp); in unwind_recover_ret_addr() [all …]
|
| /linux/drivers/scsi/qedf/ |
| H A D | drv_fcoe_fw_funcs.h | 29 * @brief init_initiator_rw_fcoe_task - Initializes FCoE task context for 30 * read/write task types and init fcoe_sqe 32 * @param task_params - Pointer to task parameters struct 33 * @param sgl_task_params - Pointer to SGL task params 45 * @brief init_initiator_midpath_fcoe_task - Initializes FCoE task context for 46 * midpath/unsolicited task types and init fcoe_sqe 48 * @param task_params - Pointer to task parameters struct 50 * @param tx_sgl_task_params - Pointer to Tx SGL task params 51 * @param rx_sgl_task_params - Pointer to Rx SGL task params 63 * @brief init_initiator_abort_fcoe_task - Initializes FCoE task context for [all …]
|