Home
last modified time | relevance | path

Searched full:task (Results 1 – 25 of 2127) sorted by relevance

12345678910>>...86

/linux/net/sunrpc/
H A Dsched.c45 static void rpc_release_task(struct rpc_task *task);
68 bool rpc_task_set_rpc_status(struct rpc_task *task, int rpc_status) in rpc_task_set_rpc_status() argument
70 if (cmpxchg(&task->tk_rpc_status, 0, rpc_status) == 0) in rpc_task_set_rpc_status()
76 rpc_task_timeout(const struct rpc_task *task) in rpc_task_timeout() argument
78 unsigned long timeout = READ_ONCE(task->tk_timeout); in rpc_task_timeout()
90 * Disable the timer for a given RPC task. Should be called with
95 __rpc_disable_timer(struct rpc_wait_queue *queue, struct rpc_task *task) in __rpc_disable_timer() argument
97 if (list_empty(&task->u.tk_wait.timer_list)) in __rpc_disable_timer()
99 task->tk_timeout = 0; in __rpc_disable_timer()
100 list_del(&task->u.tk_wait.timer_list); in __rpc_disable_timer()
[all …]
H A Dclnt.c53 static void call_start(struct rpc_task *task);
54 static void call_reserve(struct rpc_task *task);
55 static void call_reserveresult(struct rpc_task *task);
56 static void call_allocate(struct rpc_task *task);
57 static void call_encode(struct rpc_task *task);
58 static void call_decode(struct rpc_task *task);
59 static void call_bind(struct rpc_task *task);
60 static void call_bind_status(struct rpc_task *task);
61 static void call_transmit(struct rpc_task *task);
62 static void call_status(struct rpc_task *task);
[all …]
H A Dxprt.c75 static void xprt_request_init(struct rpc_task *task);
258 * @task: task that is requesting access to the transport
265 int xprt_reserve_xprt(struct rpc_xprt *xprt, struct rpc_task *task) in xprt_reserve_xprt() argument
267 struct rpc_rqst *req = task->tk_rqstp; in xprt_reserve_xprt()
270 if (task == xprt->snd_task) in xprt_reserve_xprt()
276 xprt->snd_task = task; in xprt_reserve_xprt()
279 trace_xprt_reserve_xprt(xprt, task); in xprt_reserve_xprt()
285 task->tk_status = -EAGAIN; in xprt_reserve_xprt()
286 if (RPC_IS_SOFT(task) || RPC_IS_SOFTCONN(task)) in xprt_reserve_xprt()
287 rpc_sleep_on_timeout(&xprt->sending, task, NULL, in xprt_reserve_xprt()
[all …]
/linux/drivers/gpu/drm/exynos/
H A Dexynos_drm_ipp.c87 WARN_ON(ipp->task); in exynos_drm_ipp_unregister()
261 struct exynos_drm_ipp_task *task; in exynos_drm_ipp_task_alloc() local
263 task = kzalloc_obj(*task); in exynos_drm_ipp_task_alloc()
264 if (!task) in exynos_drm_ipp_task_alloc()
267 task->dev = ipp->dev; in exynos_drm_ipp_task_alloc()
268 task->ipp = ipp; in exynos_drm_ipp_task_alloc()
271 task->src.rect.w = task->dst.rect.w = UINT_MAX; in exynos_drm_ipp_task_alloc()
272 task->src.rect.h = task->dst.rect.h = UINT_MAX; in exynos_drm_ipp_task_alloc()
273 task->transform.rotation = DRM_MODE_ROTATE_0; in exynos_drm_ipp_task_alloc()
275 DRM_DEV_DEBUG_DRIVER(task->dev, "Allocated task %p\n", task); in exynos_drm_ipp_task_alloc()
[all …]
/linux/include/asm-generic/
H A Dsyscall.h12 * and only when the caller is sure that the task of interest
23 * syscall_get_nr - find what system call a task is executing
24 * @task: task of interest, must be blocked
25 * @regs: task_pt_regs() of @task
27 * If @task is executing a system call or is at system call
29 * If @task is not executing a system call, i.e. it's blocked
36 * It's only valid to call this when @task is known to be blocked.
38 int syscall_get_nr(struct task_struct *task, struct pt_regs *regs);
41 * syscall_set_nr - change the system call a task is executing
42 * @task: task of interest, must be blocked
[all …]
/linux/kernel/bpf/
H A Dtask_iter.c39 struct task_struct *task; in task_group_seq_get_next() local
46 task = get_pid_task(pid, PIDTYPE_TGID); in task_group_seq_get_next()
47 if (!task) in task_group_seq_get_next()
53 return task; in task_group_seq_get_next()
58 * same for task_seq_start() to pick up the correct task. in task_group_seq_get_next()
62 task = get_pid_task(pid, PIDTYPE_PID); in task_group_seq_get_next()
64 return task; in task_group_seq_get_next()
67 task = find_task_by_pid_ns(common->pid_visiting, common->ns); in task_group_seq_get_next()
68 if (!task) in task_group_seq_get_next()
72 task = __next_thread(task); in task_group_seq_get_next()
[all …]
/linux/drivers/scsi/aic94xx/
H A Daic94xx_tmf.c3 * Aic94xx Task Management Functions
214 static int asd_clear_nexus_tag(struct sas_task *task) in asd_clear_nexus_tag() argument
216 struct asd_ha_struct *asd_ha = task->dev->port->ha->lldd_ha; in asd_clear_nexus_tag()
217 struct asd_ascb *tascb = task->lldd_task; in asd_clear_nexus_tag()
221 memcpy(scb->clear_nexus.ssp_task.lun, task->ssp_task.LUN, 8); in asd_clear_nexus_tag()
223 if (task->dev->tproto) in asd_clear_nexus_tag()
225 task->dev->lldd_dev); in asd_clear_nexus_tag()
229 static int asd_clear_nexus_index(struct sas_task *task) in asd_clear_nexus_index() argument
231 struct asd_ha_struct *asd_ha = task->dev->port->ha->lldd_ha; in asd_clear_nexus_index()
232 struct asd_ascb *tascb = task->lldd_task; in asd_clear_nexus_index()
[all …]
/linux/drivers/media/i2c/
H A Dsaa711x_regs.h77 /* Task independent global settings */
87 /* Task A definition */
134 /* Task B definition */
361 /* Task independent global settings: R_80_GLOBAL_CNTL_1 to R_8F_STATUS_INFO_SCALER */
383 /* Task A definition: R_90_A_TASK_HANDLING_CNTL to R_BF_A_VERT_LUMA_PHASE_OFF_11 */
384 /* Task A: Basic settings and acquisition window definition */
386 "Task A: Task handling control"},
388 "Task A: X port formats and configuration"},
390 "Task A: X port input reference signal definition"},
392 "Task A: I port output formats and configuration"},
[all …]
/linux/drivers/scsi/pm8001/
H A Dpm8001_sas.c46 * pm8001_find_tag - from sas task to find out tag that belongs to this task
47 * @task: the task sent to the LLDD
48 * @tag: the found tag associated with the task
50 static int pm8001_find_tag(struct sas_task *task, u32 *tag) in pm8001_find_tag() argument
52 if (task->lldd_task) { in pm8001_find_tag()
54 ccb = task->lldd_task; in pm8001_find_tag()
64 * @tag: the found tag associated with the task
80 * pm8001_tag_alloc - allocate a empty tag for task used.
104 static void pm80xx_get_tag_opcodes(struct sas_task *task, int *ata_op, in pm80xx_get_tag_opcodes() argument
114 if (!task) in pm80xx_get_tag_opcodes()
[all …]
/linux/drivers/scsi/isci/
H A Dtask.c64 #include "task.h"
71 * @task: request to complete
72 * @response: response code for the completed task.
73 * @status: status code for the completed task.
76 static void isci_task_refuse(struct isci_host *ihost, struct sas_task *task, in isci_task_refuse() argument
84 dev_dbg(&ihost->pdev->dev, "%s: task = %p, response=%d, status=%d\n", in isci_task_refuse()
85 __func__, task, response, status); in isci_task_refuse()
87 spin_lock_irqsave(&task->task_state_lock, flags); in isci_task_refuse()
89 task->task_status.resp = response; in isci_task_refuse()
90 task->task_status.stat = status; in isci_task_refuse()
[all …]
/linux/arch/powerpc/kernel/
H A Dsignal.h36 struct task_struct *task);
38 struct task_struct *task);
39 extern unsigned long copy_vsx_from_user(struct task_struct *task,
41 extern unsigned long copy_ckvsx_from_user(struct task_struct *task,
43 unsigned long copy_fpr_to_user(void __user *to, struct task_struct *task);
44 unsigned long copy_ckfpr_to_user(void __user *to, struct task_struct *task);
45 unsigned long copy_fpr_from_user(struct task_struct *task, void __user *from);
46 unsigned long copy_ckfpr_from_user(struct task_struct *task, void __user *from);
48 #define unsafe_copy_fpr_to_user(to, task, label) do { \ argument
49 struct task_struct *__t = task; \
[all …]
/linux/scripts/gdb/linux/
H A Dtasks.py4 # task & thread tools
39 for task in task_lists():
40 if int(task['pid']) == pid:
41 return task
46 """Find Linux task by PID and return the task_struct variable.
55 task = get_task_by_pid(pid)
56 if task:
57 return task.dereference()
59 raise gdb.GdbError("No task of PID " + str(pid))
72 gdb.write("{:>10} {:>12} {:>7}\n".format("TASK", "PI
86 get_thread_info(task) global() argument
103 invoke(self, task) global() argument
[all...]
/linux/kernel/livepatch/
H A Dtransition.c85 struct task_struct *g, *task;
122 for_each_process_thread(g, task) { in klp_complete_transition()
123 WARN_ON_ONCE(test_tsk_thread_flag(task, TIF_PATCH_PENDING)); in klp_complete_transition()
124 task->patch_state = KLP_TRANSITION_IDLE; in klp_complete_transition()
129 task = idle_task(cpu); in klp_complete_transition()
130 WARN_ON_ONCE(test_tsk_thread_flag(task, TIF_PATCH_PENDING)); in klp_complete_transition()
131 task->patch_state = KLP_TRANSITION_IDLE; in klp_complete_transition()
169 * Switch the patched state of the task to the set of functions in the target in klp_cancel_transition()
172 * NOTE: If task is not 'current', the caller must ensure the task i in klp_cancel_transition()
94 struct task_struct *g, *task; klp_complete_transition() local
184 klp_update_patch_state(struct task_struct * task) klp_update_patch_state() argument
263 klp_check_stack(struct task_struct * task,const char ** oldname) klp_check_stack() argument
293 klp_check_and_switch_task(struct task_struct * task,void * arg) klp_check_and_switch_task() argument
314 klp_try_switch_task(struct task_struct * task) klp_try_switch_task() argument
410 struct task_struct *g, *task; klp_send_signals() local
454 struct task_struct *g, *task; klp_try_complete_transition() local
532 struct task_struct *g, *task; klp_start_transition() local
575 struct task_struct *g, *task; klp_init_transition() local
651 struct task_struct *g, *task; klp_reverse_transition() local
730 struct task_struct *g, *task; klp_force_transition() local
[all...]
/linux/fs/proc/
H A Dbase.c114 * in /proc for a task before it execs a suid executable.
208 static int get_task_root(struct task_struct *task, struct path *root) in get_task_root() argument
212 task_lock(task); in get_task_root()
213 if (task->fs) { in get_task_root()
214 get_fs_root(task->fs, root); in get_task_root()
217 task_unlock(task); in get_task_root()
223 struct task_struct *task = get_proc_task(d_inode(dentry)); in proc_cwd_link() local
226 if (task) { in proc_cwd_link()
227 task_lock(task); in proc_cwd_link()
228 if (task->fs) { in proc_cwd_link()
[all …]
H A Dfd.c28 struct task_struct *task; in seq_show() local
30 task = get_proc_task(m->private); in seq_show()
31 if (!task) in seq_show()
34 task_lock(task); in seq_show()
35 files = task->files; in seq_show()
51 task_unlock(task); in seq_show()
52 put_task_struct(task); in seq_show()
82 * that the current task has PTRACE_MODE_READ in addition to the normal
89 struct task_struct *task = get_proc_task(inode); in proc_fdinfo_permission() local
91 if (!task) in proc_fdinfo_permission()
[all …]
/linux/drivers/md/dm-vdo/indexer/
H A Dradix-sort.c45 struct task { struct
60 struct task *end_of_stack; argument
61 struct task insertion_list[256];
62 struct task stack[];
72 static inline void insert_key(const struct task task, sort_key_t *next) in insert_key() argument
78 while ((--next >= task.first_key) && in insert_key()
79 (compare(unsorted, next[0], task.offset, task.length) < 0)) in insert_key()
90 static inline void insertion_sort(const struct task task) in insertion_sort() argument
94 for (next = task.first_key + 1; next <= task.last_key; next++) in insertion_sort()
95 insert_key(task, next); in insertion_sort()
[all …]
/linux/include/rv/
H A Dltl_monitor.h12 #include <trace/events/task.h>
24 static void ltl_atoms_fetch(struct task_struct *task, struct ltl_monitor *mon);
25 static void ltl_atoms_init(struct task_struct *task, struct ltl_monitor *mon, bool task_creation);
27 static struct ltl_monitor *ltl_get_monitor(struct task_struct *task) in ltl_get_monitor() argument
29 return &task->rv[ltl_monitor_slot].ltl_mon; in ltl_get_monitor()
32 static void ltl_task_init(struct task_struct *task, bool task_creation) in ltl_task_init() argument
34 struct ltl_monitor *mon = ltl_get_monitor(task); in ltl_task_init()
41 ltl_atoms_init(task, mon, task_creation); in ltl_task_init()
42 ltl_atoms_fetch(task, mon); in ltl_task_init()
45 static void handle_task_newtask(void *data, struct task_struct *task, u64 flags) in handle_task_newtask() argument
[all …]
/linux/drivers/video/fbdev/
H A Duvesafb.c69 * find the kernel part of the task struct, copy the registers and
70 * the buffer contents and then complete the task.
75 struct uvesafb_ktask *task; in uvesafb_cn_callback() local
84 task = uvfb_tasks[msg->seq]; in uvesafb_cn_callback()
86 if (!task || msg->ack != task->ack) { in uvesafb_cn_callback()
94 if (task->t.buf_len < utask->buf_len || in uvesafb_cn_callback()
103 memcpy(&task->t, utask, sizeof(*utask)); in uvesafb_cn_callback()
105 if (task->t.buf_len && task->buf) in uvesafb_cn_callback()
106 memcpy(task->buf, utask + 1, task->t.buf_len); in uvesafb_cn_callback()
108 complete(task->done); in uvesafb_cn_callback()
[all …]
/linux/tools/testing/selftests/bpf/progs/
H A Drcu_read_lock.c43 struct task_struct *task; in get_cgroup_id()
46 task = bpf_get_current_task_btf(); in get_cgroup_id()
47 if (task->pid != target_pid) in get_cgroup_id()
52 cgroups = task->cgroups; in task_succ()
64 struct task_struct *task, *real_parent; in task_succ()
68 task = bpf_get_current_task_btf(); in task_succ()
69 if (task->pid != target_pid) in task_succ()
74 real_parent = task->real_parent; in task_succ()
93 struct task_struct *task, *real_parent; in two_regions()
96 task in two_regions()
33 struct task_struct *task; get_cgroup_id() local
54 struct task_struct *task, *real_parent; task_succ() local
83 struct task_struct *task, *real_parent; no_lock() local
95 struct task_struct *task, *real_parent; two_regions() local
114 struct task_struct *task, *real_parent; non_sleepable_1() local
130 struct task_struct *task, *real_parent; non_sleepable_2() local
149 struct task_struct *task, *real_parent, *gparent; task_acquire() local
177 struct task_struct *task; miss_lock() local
191 struct task_struct *task; miss_unlock() local
203 struct task_struct *task, *real_parent; non_sleepable_rcu_mismatch() local
221 struct task_struct *task, *real_parent; inproper_sleepable_helper() local
265 struct task_struct *task, *real_parent; nested_rcu_region() local
284 struct task_struct *task, *group_leader; task_trusted_non_rcuptr() local
298 struct task_struct *task, *real_parent; task_untrusted_rcuptr() local
312 struct task_struct *task, *real_parent; cross_rcu_region() local
[all...]
/linux/Documentation/scheduler/
H A Dsched-ext.rst20 a runnable task stalls, or on invoking the SysRq key sequence
49 If a task explicitly sets its scheduling policy to ``SCHED_EXT``, it will be
111 Whether a given task is on sched_ext can be determined as follows:
130 * Decide which CPU a task should be migrated to before being
133 * then insert the task directly into SCX_DSQ_LOCAL and skip the
157 * Do a direct insertion of a task to the global DSQ. This ops.enqueue()
162 * default ops.enqueue implementation, which just dispatches the task
205 A CPU always executes a task from its local DSQ. A task is "inserted" into a
206 DSQ. A task in a non-local DSQ is "move"d into the target CPU's local DSQ.
208 When a CPU is looking for the next task to run, if the local DSQ is not
[all …]
/linux/Documentation/admin-guide/mm/
H A Dnuma_memory_policy.rst20 both cpusets and policies are applied to a task, the restrictions of the cpuset
44 Task/Process Policy
45 this is an optional, per-task policy. When defined for a
46 specific task, this policy controls all page allocations made
47 by or on behalf of the task that aren't controlled by a more
48 specific scope. If a task does not define a task policy, then
50 task policy "fall back" to the System Default Policy.
52 The task policy applies to the entire address space of a task. Thus,
54 [clone() w/o the CLONE_VM flag] and exec*(). This allows a parent task
55 to establish the task policy for a child task exec()'d from an
[all …]
/linux/drivers/scsi/qedi/
H A Dqedi_fw_iscsi.h31 * task context.
33 * @param task_params - Pointer to task parameters struct
37 * @param sgl_task_params - Pointer to SGL task params
49 * Request task context.
51 * @param task_params - Pointer to task parameters struct
53 * @param tx_sgl_task_params - Pointer to SGL task params
54 * @param rx_sgl_task_params - Pointer to SGL task params
62 * task context.
64 * @param task_params - Pointer to task parameters struct
66 * @param tx_sgl_task_params - Pointer to SGL task params
[all …]
/linux/arch/arm64/kernel/
H A Dfpsimd.c60 * (a) for each task, we need to remember which CPU was the last one to have
61 * the task's FPSIMD state loaded into its FPSIMD registers;
62 * (b) for each CPU, we need to remember which task's userland FPSIMD state has
69 * address of the userland FPSIMD state of the task that was loaded onto the CPU
75 * task's fpsimd_cpu are still mutually in sync. If this is the case, we
79 * indicate whether or not the userland FPSIMD state of the current task is
82 * task. If the task is behaving as a VMM, then this is will be managed by
89 * called from softirq context, which will save the task's FPSIMD context back
91 * task's FPSIMD state from task context and thereby corrupting the state, it
92 * is necessary to protect any manipulation of a task's fpsimd_state or
[all …]
/linux/tools/perf/util/bpf_skel/
H A Dkwork_top.bpf.c105 static __always_inline void update_task_info(struct task_struct *task, __u32 cpu) in update_task_info() argument
108 .pid = task->pid, in update_task_info()
114 .tgid = task->tgid, in update_task_info()
115 .is_kthread = task->flags & PF_KTHREAD ? 1 : 0, in update_task_info()
117 BPF_CORE_READ_STR_INTO(&data.comm, task, comm); in update_task_info()
139 static void on_sched_out(struct task_struct *task, __u64 ts, __u32 cpu) in on_sched_out() argument
144 pelem = bpf_task_storage_get(&kwork_top_task_time, task, NULL, 0); in on_sched_out()
152 .pid = task->pid, in on_sched_out()
153 .task_p = (__u64)task, in on_sched_out()
157 update_task_info(task, cpu); in on_sched_out()
[all …]
/linux/kernel/sched/
H A Dext_internal.h116 * By default, if there are no other task to run on the CPU, ext core
117 * keeps running the current task even after its slice expires. If this
124 * An exiting task may schedule after PF_EXITING is set. In such cases,
125 * bpf_task_from_pid() may not be able to find the task and if the BPF
126 * scheduler depends on pid lookup for dispatching, the task will be
143 * A migration disabled task can only execute on its current CPU. By
148 * A migration disabled task never invokes ops.select_cpu() as it can
205 /* the cgroup the task is joining */
212 /* Whether the task exited before running on sched_ext. */
228 /* next task is being scheduled by &sched_class_rt */
[all …]

12345678910>>...86