Home
last modified time | relevance | path

Searched full:task (Results 1 – 25 of 1933) sorted by relevance

12345678910>>...78

/linux/net/sunrpc/
H A Dsched.c45 static void rpc_release_task(struct rpc_task *task);
68 bool rpc_task_set_rpc_status(struct rpc_task *task, int rpc_status) in rpc_task_set_rpc_status() argument
70 if (cmpxchg(&task->tk_rpc_status, 0, rpc_status) == 0) in rpc_task_set_rpc_status()
76 rpc_task_timeout(const struct rpc_task *task) in rpc_task_timeout() argument
78 unsigned long timeout = READ_ONCE(task->tk_timeout); in rpc_task_timeout()
90 * Disable the timer for a given RPC task. Should be called with
95 __rpc_disable_timer(struct rpc_wait_queue *queue, struct rpc_task *task) in __rpc_disable_timer() argument
97 if (list_empty(&task->u.tk_wait.timer_list)) in __rpc_disable_timer()
99 task->tk_timeout = 0; in __rpc_disable_timer()
100 list_del(&task->u.tk_wait.timer_list); in __rpc_disable_timer()
[all …]
H A Dclnt.c53 static void call_start(struct rpc_task *task);
54 static void call_reserve(struct rpc_task *task);
55 static void call_reserveresult(struct rpc_task *task);
56 static void call_allocate(struct rpc_task *task);
57 static void call_encode(struct rpc_task *task);
58 static void call_decode(struct rpc_task *task);
59 static void call_bind(struct rpc_task *task);
60 static void call_bind_status(struct rpc_task *task);
61 static void call_transmit(struct rpc_task *task);
62 static void call_status(struct rpc_task *task);
[all …]
H A Dxprt.c75 static void xprt_request_init(struct rpc_task *task);
258 * @task: task that is requesting access to the transport
265 int xprt_reserve_xprt(struct rpc_xprt *xprt, struct rpc_task *task) in xprt_reserve_xprt() argument
267 struct rpc_rqst *req = task->tk_rqstp; in xprt_reserve_xprt()
270 if (task == xprt->snd_task) in xprt_reserve_xprt()
276 xprt->snd_task = task; in xprt_reserve_xprt()
279 trace_xprt_reserve_xprt(xprt, task); in xprt_reserve_xprt()
285 task->tk_status = -EAGAIN; in xprt_reserve_xprt()
286 if (RPC_IS_SOFT(task) || RPC_IS_SOFTCONN(task)) in xprt_reserve_xprt()
287 rpc_sleep_on_timeout(&xprt->sending, task, NULL, in xprt_reserve_xprt()
[all …]
/linux/include/asm-generic/
H A Dsyscall.h12 * and only when the caller is sure that the task of interest
23 * syscall_get_nr - find what system call a task is executing
24 * @task: task of interest, must be blocked
25 * @regs: task_pt_regs() of @task
27 * If @task is executing a system call or is at system call
29 * If @task is not executing a system call, i.e. it's blocked
36 * It's only valid to call this when @task is known to be blocked.
38 int syscall_get_nr(struct task_struct *task, struct pt_regs *regs);
41 * syscall_set_nr - change the system call a task is executing
42 * @task: task of interest, must be blocked
[all …]
/linux/drivers/gpu/drm/exynos/
H A Dexynos_drm_ipp.c87 WARN_ON(ipp->task); in exynos_drm_ipp_unregister()
261 struct exynos_drm_ipp_task *task; in exynos_drm_ipp_task_alloc() local
263 task = kzalloc_obj(*task); in exynos_drm_ipp_task_alloc()
264 if (!task) in exynos_drm_ipp_task_alloc()
267 task->dev = ipp->dev; in exynos_drm_ipp_task_alloc()
268 task->ipp = ipp; in exynos_drm_ipp_task_alloc()
271 task->src.rect.w = task->dst.rect.w = UINT_MAX; in exynos_drm_ipp_task_alloc()
272 task->src.rect.h = task->dst.rect.h = UINT_MAX; in exynos_drm_ipp_task_alloc()
273 task->transform.rotation = DRM_MODE_ROTATE_0; in exynos_drm_ipp_task_alloc()
275 DRM_DEV_DEBUG_DRIVER(task->dev, "Allocated task %p\n", task); in exynos_drm_ipp_task_alloc()
[all …]
/linux/drivers/scsi/aic94xx/
H A Daic94xx_tmf.c3 * Aic94xx Task Management Functions
214 static int asd_clear_nexus_tag(struct sas_task *task) in asd_clear_nexus_tag() argument
216 struct asd_ha_struct *asd_ha = task->dev->port->ha->lldd_ha; in asd_clear_nexus_tag()
217 struct asd_ascb *tascb = task->lldd_task; in asd_clear_nexus_tag()
221 memcpy(scb->clear_nexus.ssp_task.lun, task->ssp_task.LUN, 8); in asd_clear_nexus_tag()
223 if (task->dev->tproto) in asd_clear_nexus_tag()
225 task->dev->lldd_dev); in asd_clear_nexus_tag()
229 static int asd_clear_nexus_index(struct sas_task *task) in asd_clear_nexus_index() argument
231 struct asd_ha_struct *asd_ha = task->dev->port->ha->lldd_ha; in asd_clear_nexus_index()
232 struct asd_ascb *tascb = task->lldd_task; in asd_clear_nexus_index()
[all …]
/linux/drivers/media/i2c/
H A Dsaa711x_regs.h77 /* Task independent global settings */
87 /* Task A definition */
134 /* Task B definition */
361 /* Task independent global settings: R_80_GLOBAL_CNTL_1 to R_8F_STATUS_INFO_SCALER */
383 /* Task A definition: R_90_A_TASK_HANDLING_CNTL to R_BF_A_VERT_LUMA_PHASE_OFF_11 */
384 /* Task A: Basic settings and acquisition window definition */
386 "Task A: Task handling control"},
388 "Task A: X port formats and configuration"},
390 "Task A: X port input reference signal definition"},
392 "Task A: I port output formats and configuration"},
[all …]
/linux/Documentation/scheduler/
H A Dsched-ext.rst20 a runnable task stalls, or on invoking the SysRq key sequence
49 If a task explicitly sets its scheduling policy to ``SCHED_EXT``, it will be
120 the task and the core scheduler silently picked a fallback CPU.
123 * ``SCX_EV_DISPATCH_KEEP_LAST``: a task continued running because no other
124 task was available (only when ``SCX_OPS_ENQ_LAST`` is not set).
125 * ``SCX_EV_ENQ_SKIP_EXITING``: an exiting task was dispatched to the local DSQ
127 * ``SCX_EV_ENQ_SKIP_MIGRATION_DISABLED``: a migration-disabled task was
130 * ``SCX_EV_REENQ_IMMED``: a task dispatched with ``SCX_ENQ_IMMED`` was
135 * ``SCX_EV_REFILL_SLICE_DFL``: a task's time slice was refilled with the
140 * ``SCX_EV_INSERT_NOT_OWNED``: attempted to insert a task not owned by this
[all …]
/linux/scripts/gdb/linux/
H A Dtasks.py4 # task & thread tools
39 for task in task_lists():
40 if int(task['pid']) == pid:
41 return task
46 """Find Linux task by PID and return the task_struct variable.
55 task = get_task_by_pid(pid)
56 if task:
57 return task.dereference()
59 raise gdb.GdbError("No task of PID " + str(pid))
72 gdb.write("{:>10} {:>12} {:>7}\n".format("TASK", "PI
86 get_thread_info(task) global() argument
103 invoke(self, task) global() argument
[all...]
/linux/arch/powerpc/kernel/
H A Dsignal.h36 struct task_struct *task);
38 struct task_struct *task);
39 extern unsigned long copy_vsx_from_user(struct task_struct *task,
41 extern unsigned long copy_ckvsx_from_user(struct task_struct *task,
43 unsigned long copy_fpr_to_user(void __user *to, struct task_struct *task);
44 unsigned long copy_ckfpr_to_user(void __user *to, struct task_struct *task);
45 unsigned long copy_fpr_from_user(struct task_struct *task, void __user *from);
46 unsigned long copy_ckfpr_from_user(struct task_struct *task, void __user *from);
48 #define unsafe_copy_fpr_to_user(to, task, label) do { \ argument
49 struct task_struct *__t = task; \
[all …]
/linux/drivers/scsi/pm8001/
H A Dpm8001_sas.c46 * pm8001_find_tag - from sas task to find out tag that belongs to this task
47 * @task: the task sent to the LLDD
48 * @tag: the found tag associated with the task
50 static int pm8001_find_tag(struct sas_task *task, u32 *tag) in pm8001_find_tag() argument
52 if (task->lldd_task) { in pm8001_find_tag()
54 ccb = task->lldd_task; in pm8001_find_tag()
64 * @tag: the found tag associated with the task
80 * pm8001_tag_alloc - allocate a empty tag for task used.
104 static void pm80xx_get_tag_opcodes(struct sas_task *task, int *ata_op, in pm80xx_get_tag_opcodes() argument
114 if (!task) in pm80xx_get_tag_opcodes()
[all …]
/linux/drivers/scsi/isci/
H A Dtask.c64 #include "task.h"
71 * @task: request to complete
72 * @response: response code for the completed task.
73 * @status: status code for the completed task.
76 static void isci_task_refuse(struct isci_host *ihost, struct sas_task *task, in isci_task_refuse() argument
84 dev_dbg(&ihost->pdev->dev, "%s: task = %p, response=%d, status=%d\n", in isci_task_refuse()
85 __func__, task, response, status); in isci_task_refuse()
87 spin_lock_irqsave(&task->task_state_lock, flags); in isci_task_refuse()
89 task->task_status.resp = response; in isci_task_refuse()
90 task->task_status.stat = status; in isci_task_refuse()
[all …]
/linux/kernel/livepatch/
H A Dtransition.c85 struct task_struct *g, *task;
122 for_each_process_thread(g, task) { in klp_complete_transition()
123 WARN_ON_ONCE(test_tsk_thread_flag(task, TIF_PATCH_PENDING)); in klp_complete_transition()
124 task->patch_state = KLP_TRANSITION_IDLE; in klp_complete_transition()
129 task = idle_task(cpu); in klp_complete_transition()
130 WARN_ON_ONCE(test_tsk_thread_flag(task, TIF_PATCH_PENDING)); in klp_complete_transition()
131 task->patch_state = KLP_TRANSITION_IDLE; in klp_complete_transition()
169 * Switch the patched state of the task to the set of functions in the target in klp_cancel_transition()
172 * NOTE: If task is not 'current', the caller must ensure the task i in klp_cancel_transition()
94 struct task_struct *g, *task; klp_complete_transition() local
184 klp_update_patch_state(struct task_struct * task) klp_update_patch_state() argument
263 klp_check_stack(struct task_struct * task,const char ** oldname) klp_check_stack() argument
293 klp_check_and_switch_task(struct task_struct * task,void * arg) klp_check_and_switch_task() argument
314 klp_try_switch_task(struct task_struct * task) klp_try_switch_task() argument
410 struct task_struct *g, *task; klp_send_signals() local
454 struct task_struct *g, *task; klp_try_complete_transition() local
532 struct task_struct *g, *task; klp_start_transition() local
575 struct task_struct *g, *task; klp_init_transition() local
651 struct task_struct *g, *task; klp_reverse_transition() local
730 struct task_struct *g, *task; klp_force_transition() local
[all...]
/linux/arch/riscv/include/asm/
H A Dusercfi.h36 void set_shstk_base(struct task_struct *task, unsigned long shstk_addr, unsigned long size);
37 unsigned long get_shstk_base(struct task_struct *task, unsigned long *size);
38 void set_active_shstk(struct task_struct *task, unsigned long shstk_addr);
39 bool is_shstk_enabled(struct task_struct *task);
40 bool is_shstk_locked(struct task_struct *task);
41 bool is_shstk_allocated(struct task_struct *task);
42 void set_shstk_lock(struct task_struct *task, bool lock);
43 void set_shstk_status(struct task_struct *task, bool enable);
44 unsigned long get_active_shstk(struct task_struct *task);
47 bool is_indir_lp_enabled(struct task_struct *task);
[all …]
/linux/rust/kernel/
H A Dtask.rs33 /// Returns the currently running task.
39 // leave current task context:
47 unsafe { &*$crate::task::Task::current() }
75 /// Getting the current task and storing it in some struct. The reference count is automatically
79 /// use kernel::{task::Task, sync::aref::ARef};
82 /// creator: ARef<Task>,
96 pub struct Task(pub(crate) Opaque<bindings::task_struct>); struct
98 // SAFETY: By design, the only way to access a `Task` is via the `current` function or via an
99 // `ARef<Task>` obtained through the `AlwaysRefCounted` impl. This means that the only situation in
100 // which a `Task` can be accessed mutably is when the refcount drops to zero and the destructor
[all …]
/linux/fs/proc/
H A Dbase.c114 * in /proc for a task before it execs a suid executable.
208 static int get_task_root(struct task_struct *task, struct path *root) in get_task_root() argument
212 task_lock(task); in get_task_root()
213 if (task->fs) { in get_task_root()
214 get_fs_root(task->fs, root); in get_task_root()
217 task_unlock(task); in get_task_root()
223 struct task_struct *task = get_proc_task(d_inode(dentry)); in proc_cwd_link() local
226 if (task) { in proc_cwd_link()
227 task_lock(task); in proc_cwd_link()
228 if (task->fs) { in proc_cwd_link()
[all …]
/linux/include/rv/
H A Dltl_monitor.h12 #include <trace/events/task.h>
24 static void ltl_atoms_fetch(struct task_struct *task, struct ltl_monitor *mon);
25 static void ltl_atoms_init(struct task_struct *task, struct ltl_monitor *mon, bool task_creation);
27 static struct ltl_monitor *ltl_get_monitor(struct task_struct *task) in ltl_get_monitor() argument
29 return &task->rv[ltl_monitor_slot].ltl_mon; in ltl_get_monitor()
32 static void ltl_task_init(struct task_struct *task, bool task_creation) in ltl_task_init() argument
34 struct ltl_monitor *mon = ltl_get_monitor(task); in ltl_task_init()
41 ltl_atoms_init(task, mon, task_creation); in ltl_task_init()
42 ltl_atoms_fetch(task, mon); in ltl_task_init()
45 static void handle_task_newtask(void *data, struct task_struct *task, u64 flags) in handle_task_newtask() argument
[all …]
/linux/drivers/video/fbdev/
H A Duvesafb.c69 * find the kernel part of the task struct, copy the registers and
70 * the buffer contents and then complete the task.
75 struct uvesafb_ktask *task; in uvesafb_cn_callback() local
84 task = uvfb_tasks[msg->seq]; in uvesafb_cn_callback()
86 if (!task || msg->ack != task->ack) { in uvesafb_cn_callback()
94 if (task->t.buf_len < utask->buf_len || in uvesafb_cn_callback()
103 memcpy(&task->t, utask, sizeof(*utask)); in uvesafb_cn_callback()
105 if (task->t.buf_len && task->buf) in uvesafb_cn_callback()
106 memcpy(task->buf, utask + 1, task->t.buf_len); in uvesafb_cn_callback()
108 complete(task->done); in uvesafb_cn_callback()
[all …]
/linux/tools/testing/selftests/bpf/progs/
H A Drcu_read_lock.c43 struct task_struct *task; in get_cgroup_id()
46 task = bpf_get_current_task_btf(); in get_cgroup_id()
47 if (task->pid != target_pid) in get_cgroup_id()
52 cgroups = task->cgroups; in task_succ()
64 struct task_struct *task, *real_parent; in task_succ()
68 task = bpf_get_current_task_btf(); in task_succ()
69 if (task->pid != target_pid) in task_succ()
74 real_parent = task->real_parent; in task_succ()
93 struct task_struct *task, *real_parent; in two_regions()
96 task in two_regions()
33 struct task_struct *task; get_cgroup_id() local
54 struct task_struct *task, *real_parent; task_succ() local
83 struct task_struct *task, *real_parent; no_lock() local
95 struct task_struct *task, *real_parent; two_regions() local
114 struct task_struct *task, *real_parent; non_sleepable_1() local
130 struct task_struct *task, *real_parent; non_sleepable_2() local
149 struct task_struct *task, *real_parent, *gparent; task_acquire() local
177 struct task_struct *task; miss_lock() local
191 struct task_struct *task; miss_unlock() local
203 struct task_struct *task, *real_parent; non_sleepable_rcu_mismatch() local
221 struct task_struct *task, *real_parent; inproper_sleepable_helper() local
265 struct task_struct *task, *real_parent; nested_rcu_region() local
284 struct task_struct *task, *group_leader; task_trusted_non_rcuptr() local
298 struct task_struct *task, *real_parent; task_untrusted_rcuptr() local
312 struct task_struct *task, *real_parent; cross_rcu_region() local
[all...]
/linux/drivers/scsi/qedi/
H A Dqedi_fw_iscsi.h31 * task context.
33 * @param task_params - Pointer to task parameters struct
37 * @param sgl_task_params - Pointer to SGL task params
49 * Request task context.
51 * @param task_params - Pointer to task parameters struct
53 * @param tx_sgl_task_params - Pointer to SGL task params
54 * @param rx_sgl_task_params - Pointer to SGL task params
62 * task context.
64 * @param task_params - Pointer to task parameters struct
66 * @param tx_sgl_task_params - Pointer to SGL task params
[all …]
/linux/tools/perf/util/bpf_skel/
H A Dkwork_top.bpf.c105 static __always_inline void update_task_info(struct task_struct *task, __u32 cpu) in update_task_info() argument
108 .pid = task->pid, in update_task_info()
114 .tgid = task->tgid, in update_task_info()
115 .is_kthread = task->flags & PF_KTHREAD ? 1 : 0, in update_task_info()
117 BPF_CORE_READ_STR_INTO(&data.comm, task, comm); in update_task_info()
139 static void on_sched_out(struct task_struct *task, __u64 ts, __u32 cpu) in on_sched_out() argument
144 pelem = bpf_task_storage_get(&kwork_top_task_time, task, NULL, 0); in on_sched_out()
152 .pid = task->pid, in on_sched_out()
153 .task_p = (__u64)task, in on_sched_out()
157 update_task_info(task, cpu); in on_sched_out()
[all …]
/linux/kernel/sched/
H A Dext_internal.h125 * By default, if there are no other task to run on the CPU, ext core
126 * keeps running the current task even after its slice expires. If this
133 * An exiting task may schedule after PF_EXITING is set. In such cases,
134 * bpf_task_from_pid() may not be able to find the task and if the BPF
135 * scheduler depends on pid lookup for dispatching, the task will be
152 * A migration disabled task can only execute on its current CPU. By
157 * A migration disabled task never invokes ops.select_cpu() as it can
215 /* the cgroup the task is joining */
222 /* Whether the task exited before running on sched_ext. */
238 /* next task is being scheduled by &sched_class_rt */
[all …]
/linux/arch/riscv/kernel/
H A Dstacktrace.c20 * This disables KASAN checking when reading a value from another task's stack,
21 * since the other task could be running on another CPU and could have poisoned
24 #define READ_ONCE_TASK_STACK(task, x) \ argument
28 if ((task) == current) \
48 void notrace walk_stackframe(struct task_struct *task, struct pt_regs *regs, in walk_stackframe() argument
59 } else if (task == NULL || task == current) { in walk_stackframe()
65 /* task blocked in __switch_to */ in walk_stackframe()
66 fp = task->thread.s[0]; in walk_stackframe()
67 sp = task->thread.sp; in walk_stackframe()
68 pc = task->thread.ra; in walk_stackframe()
[all …]
/linux/arch/s390/include/asm/
H A Dunwind.h14 * There four combinations for task and regs:
15 * 1) task==NULL, regs==NULL: the unwind starts for the task that is currently
17 * 2) task==NULL, regs!=NULL: the unwind starts from the sp/ip found in
18 * the struct pt_regs of an interrupt frame for the current task
19 * 3) task!=NULL, regs==NULL: the unwind starts for an inactive task with
20 * the sp picked up from task->thread.ksp and the ip picked up from the
22 * 4) task!=NULL, regs!=NULL: the sp/ip are picked up from the interrupt
23 * frame 'regs' of a inactive task
37 struct task_struct *task; member
50 ip = ftrace_graph_ret_addr(state->task, &state->graph_idx, ip, (void *)state->sp); in unwind_recover_ret_addr()
[all …]
/linux/drivers/scsi/qedf/
H A Ddrv_fcoe_fw_funcs.h29 * @brief init_initiator_rw_fcoe_task - Initializes FCoE task context for
30 * read/write task types and init fcoe_sqe
32 * @param task_params - Pointer to task parameters struct
33 * @param sgl_task_params - Pointer to SGL task params
45 * @brief init_initiator_midpath_fcoe_task - Initializes FCoE task context for
46 * midpath/unsolicited task types and init fcoe_sqe
48 * @param task_params - Pointer to task parameters struct
50 * @param tx_sgl_task_params - Pointer to Tx SGL task params
51 * @param rx_sgl_task_params - Pointer to Rx SGL task params
63 * @brief init_initiator_abort_fcoe_task - Initializes FCoE task context for
[all …]

12345678910>>...78