| /linux/fs/smb/server/ |
| H A D | ksmbd_work.c | 21 struct ksmbd_work *work = kmem_cache_zalloc(work_cache, KSMBD_DEFAULT_GFP); in ksmbd_alloc_work_struct() local 23 if (work) { in ksmbd_alloc_work_struct() 24 work->compound_fid = KSMBD_NO_FID; in ksmbd_alloc_work_struct() 25 work->compound_pfid = KSMBD_NO_FID; in ksmbd_alloc_work_struct() 26 INIT_LIST_HEAD(&work->request_entry); in ksmbd_alloc_work_struct() 27 INIT_LIST_HEAD(&work->async_request_entry); in ksmbd_alloc_work_struct() 28 INIT_LIST_HEAD(&work->fp_entry); in ksmbd_alloc_work_struct() 29 INIT_LIST_HEAD(&work->aux_read_list); in ksmbd_alloc_work_struct() 30 work->iov_alloc_cnt = 4; in ksmbd_alloc_work_struct() 31 work->iov = kzalloc_objs(struct kvec, work->iov_alloc_cnt, in ksmbd_alloc_work_struct() [all …]
|
| H A D | server.c | 89 * @work: smb work containing server thread information 93 static inline int check_conn_state(struct ksmbd_work *work) in check_conn_state() argument 97 if (ksmbd_conn_exiting(work->conn) || in check_conn_state() 98 ksmbd_conn_need_reconnect(work->conn)) { in check_conn_state() 99 rsp_hdr = smb_get_msg(work->response_buf); in check_conn_state() 109 static int __process_request(struct ksmbd_work *work, struct ksmbd_conn *conn, in __process_request() argument 116 if (check_conn_state(work)) in __process_request() 119 if (ksmbd_verify_smb_message(work)) { in __process_request() 120 conn->ops->set_rsp_status(work, STATUS_INVALID_PARAMETER); in __process_request() 124 command = conn->ops->get_cmd_val(work); in __process_request() [all …]
|
| H A D | smb2pdu.h | 327 bool is_smb2_neg_cmd(struct ksmbd_work *work); 328 bool is_smb2_rsp(struct ksmbd_work *work); 330 u16 get_smb2_cmd_val(struct ksmbd_work *work); 331 void set_smb2_rsp_status(struct ksmbd_work *work, __le32 err); 332 int init_smb2_rsp_hdr(struct ksmbd_work *work); 333 int smb2_allocate_rsp_buf(struct ksmbd_work *work); 334 bool is_chained_smb2_message(struct ksmbd_work *work); 335 int init_smb2_neg_rsp(struct ksmbd_work *work); 336 void smb2_set_err_rsp(struct ksmbd_work *work); 337 int smb2_check_user_session(struct ksmbd_work *work); [all …]
|
| H A D | smb2pdu.c | 45 static void __wbuf(struct ksmbd_work *work, void **req, void **rsp) in __wbuf() argument 47 if (work->next_smb2_rcv_hdr_off) { in __wbuf() 48 *req = ksmbd_req_buf_next(work); in __wbuf() 49 *rsp = ksmbd_resp_buf_next(work); in __wbuf() 51 *req = smb_get_msg(work->request_buf); in __wbuf() 52 *rsp = smb_get_msg(work->response_buf); in __wbuf() 94 * @work: smb work 99 int smb2_get_ksmbd_tcon(struct ksmbd_work *work) in smb2_get_ksmbd_tcon() argument 101 struct smb2_hdr *req_hdr = ksmbd_req_buf_next(work); in smb2_get_ksmbd_tcon() 112 if (xa_empty(&work->sess->tree_conns)) { in smb2_get_ksmbd_tcon() [all …]
|
| H A D | ksmbd_work.h | 86 struct work_struct work; member 96 * @work: smb work containing response buffer 98 static inline void *ksmbd_resp_buf_next(struct ksmbd_work *work) in ksmbd_resp_buf_next() argument 100 return work->response_buf + work->next_smb2_rsp_hdr_off + 4; in ksmbd_resp_buf_next() 105 * @work: smb work containing response buffer 107 static inline void *ksmbd_resp_buf_curr(struct ksmbd_work *work) in ksmbd_resp_buf_curr() argument 109 return work->response_buf + work->curr_smb2_rsp_hdr_off + 4; in ksmbd_resp_buf_curr() 114 * @work: smb work containing response buffer 116 static inline void *ksmbd_req_buf_next(struct ksmbd_work *work) in ksmbd_req_buf_next() argument 118 return work->request_buf + work->next_smb2_rcv_hdr_off + 4; in ksmbd_req_buf_next() [all …]
|
| H A D | smb_common.c | 153 * @work: smb work 159 int ksmbd_verify_smb_message(struct ksmbd_work *work) in ksmbd_verify_smb_message() argument 161 struct smb2_hdr *smb2_hdr = ksmbd_req_buf_next(work); in ksmbd_verify_smb_message() 165 return ksmbd_smb2_check_message(work); in ksmbd_verify_smb_message() 167 hdr = smb_get_msg(work->request_buf); in ksmbd_verify_smb_message() 170 work->conn->outstanding_credits++; in ksmbd_verify_smb_message() 329 * @work: smb work containing smb header 333 static u16 get_smb1_cmd_val(struct ksmbd_work *work) in get_smb1_cmd_val() argument 340 * @work: smb work containing smb request 344 static int init_smb1_rsp_hdr(struct ksmbd_work *work) in init_smb1_rsp_hdr() argument [all …]
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_flip_work.c | 47 static void drm_flip_work_queue_task(struct drm_flip_work *work, struct drm_flip_task *task) in drm_flip_work_queue_task() argument 51 spin_lock_irqsave(&work->lock, flags); in drm_flip_work_queue_task() 52 list_add_tail(&task->node, &work->queued); in drm_flip_work_queue_task() 53 spin_unlock_irqrestore(&work->lock, flags); in drm_flip_work_queue_task() 57 * drm_flip_work_queue - queue work 58 * @work: the flip-work 61 * Queues work, that will later be run (passed back to drm_flip_func_t 62 * func) on a work queue after drm_flip_work_commit() is called. 64 void drm_flip_work_queue(struct drm_flip_work *work, void *val) in drm_flip_work_queue() argument 71 drm_flip_work_queue_task(work, task); in drm_flip_work_queue() [all …]
|
| /linux/include/trace/events/ |
| H A D | workqueue.h | 14 * workqueue_queue_work - called when a work gets queued 17 * @work: pointer to struct work_struct 19 * This event occurs when a work is queued immediately or once a 20 * delayed work is actually queued on a workqueue (ie: once the delay 26 struct work_struct *work), 28 TP_ARGS(req_cpu, pwq, work), 31 __field( void *, work ) 39 __entry->work = work; 40 __entry->function = work [all...] |
| /linux/virt/kvm/ |
| H A D | async_pf.c | 45 static void async_pf_execute(struct work_struct *work) in async_pf_execute() argument 48 container_of(work, struct kvm_async_pf, work); in async_pf_execute() 63 * work item is fully processed. in async_pf_execute() 99 static void kvm_flush_and_free_async_pf_work(struct kvm_async_pf *work) in kvm_flush_and_free_async_pf_work() argument 102 * The async #PF is "done", but KVM must wait for the work item itself, in kvm_flush_and_free_async_pf_work() 105 * after the last call to module_put(). Note, flushing the work item in kvm_flush_and_free_async_pf_work() 111 * need to be flushed (but sanity check that the work wasn't queued). in kvm_flush_and_free_async_pf_work() 113 if (work->wakeup_all) in kvm_flush_and_free_async_pf_work() 114 WARN_ON_ONCE(work in kvm_flush_and_free_async_pf_work() 124 struct kvm_async_pf *work = kvm_clear_async_pf_completion_queue() local 139 struct kvm_async_pf *work = kvm_clear_async_pf_completion_queue() local 155 struct kvm_async_pf *work; kvm_check_async_pf_completion() local 182 struct kvm_async_pf *work; kvm_setup_async_pf() local 218 struct kvm_async_pf *work; kvm_async_pf_wakeup_all() local [all...] |
| /linux/fs/btrfs/ |
| H A D | async-thread.c | 29 /* List head pointing to ordered work list */ 55 struct btrfs_fs_info * __pure btrfs_work_owner(const struct btrfs_work *work) in btrfs_work_owner() argument 57 return work->wq->fs_info; in btrfs_work_owner() 163 * Hook for threshold which will be called before executing the work, 213 struct btrfs_work *work; in run_ordered_work() local 222 work = list_first_entry(list, struct btrfs_work, ordered_list); in run_ordered_work() 223 if (!test_bit(WORK_DONE_BIT, &work->flags)) in run_ordered_work() 229 * updates from ordinary work function. in run_ordered_work() 235 * we leave the work item on the list as a barrier so in run_ordered_work() 236 * that later work item in run_ordered_work() 294 struct btrfs_work *work = container_of(normal_work, struct btrfs_work, btrfs_work_helper() local 329 btrfs_init_work(struct btrfs_work * work,btrfs_func_t func,btrfs_ordered_func_t ordered_func) btrfs_init_work() argument 339 btrfs_queue_work(struct btrfs_workqueue * wq,struct btrfs_work * work) btrfs_queue_work() argument [all...] |
| /linux/LICENSES/dual/ |
| H A D | copyleft-next-0.3.1 | 26 of, publicly perform and publicly display My Work. 40 Legal Notices contained in My Work (to the extent they remain 47 If You Distribute a Derived Work, You must license the entire Derived 48 Work as a whole under this License, with prominent notice of such 50 separate Distribution of portions of the Derived Work. 52 If the Derived Work includes material licensed under the GPL, You may 53 instead license the Derived Work under the GPL. 57 When Distributing a Covered Work, You may not impose further 58 restrictions on the exercise of rights in the Covered Work granted under 64 However, You may Distribute a Covered Work incorporating material [all …]
|
| H A D | Apache-2.0 | 49 "Work" shall mean the work of authorship, whether in Source or Object form, 51 is included in or attached to the work (an example is provided in the 54 "Derivative Works" shall mean any work, whether in Source or Object form, 55 that is based on (or derived from) the Work and for which the editorial 57 a whole, an original work of authorship. For the purposes of this License, 59 merely link (or bind by name) to the interfaces of, the Work and Derivative 62 "Contribution" shall mean any work of authorship, including the original 63 version of the Work and any modifications or additions to that Work or 65 inclusion in the Work by the copyright owner or by an individual or Legal 72 and improving the Work, but excluding communication that is conspicuously [all …]
|
| /linux/include/linux/ |
| H A D | workqueue.h | 3 * workqueue.h --- work queue handling for Linux. 21 * The first word is the work queue pointer and the flags rolled into 24 #define work_data_bits(work) ((unsigned long *)(&(work)->data)) argument 27 WORK_STRUCT_PENDING_BIT = 0, /* work item is pending execution */ 28 WORK_STRUCT_INACTIVE_BIT, /* work item is inactive */ 30 WORK_STRUCT_LINKED_BIT, /* next work is linked to this one */ 67 * When a work item is off queue, the high bits encode off-queue flags 115 struct work_struct work; member 118 /* target workqueue and CPU ->timer uses to queue ->work */ 124 struct work_struct work; global() member 212 to_delayed_work(struct work_struct * work) to_delayed_work() argument 217 to_rcu_work(struct work_struct * work) to_rcu_work() argument 223 struct work_struct work; global() member 264 work_static(struct work_struct * work) work_static() argument 269 __init_work(struct work_struct * work,int onstack) __init_work() argument 270 destroy_work_on_stack(struct work_struct * work) destroy_work_on_stack() argument 271 destroy_delayed_work_on_stack(struct delayed_work * work) destroy_delayed_work_on_stack() argument 272 work_static(struct work_struct * work) work_static() argument 355 work_pending(work) global() argument 667 queue_work(struct workqueue_struct * wq,struct work_struct * work) queue_work() argument 709 schedule_work_on(int cpu,struct work_struct * work) schedule_work_on() argument 728 schedule_work(struct work_struct * work) schedule_work() argument 750 enable_and_queue_work(struct workqueue_struct * wq,struct work_struct * work) enable_and_queue_work() argument [all...] |
| H A D | completion.h | 35 #define COMPLETION_INITIALIZER(work) \ argument 36 { 0, __SWAIT_QUEUE_HEAD_INITIALIZER((work).wait) } 38 #define COMPLETION_INITIALIZER_ONSTACK_MAP(work, map) \ argument 39 (*({ init_completion_map(&(work), &(map)); &(work); })) 41 #define COMPLETION_INITIALIZER_ONSTACK(work) \ argument 42 (*({ init_completion(&work); &work; })) 46 * @work: identifier for the completion structure 52 #define DECLARE_COMPLETION(work) \ argument 53 struct completion work = COMPLETION_INITIALIZER(work) 62 * @work: identifier for the completion structure [all …]
|
| /linux/tools/perf/ |
| H A D | builtin-kwork.c | 315 struct kwork_work *work; in work_search() local 319 work = container_of(node, struct kwork_work, node); in work_search() 320 cmp = work_cmp(sort_list, key, work); in work_search() 326 if (work->name == NULL) in work_search() 327 work->name = key->name; in work_search() 328 return work; in work_search() 362 struct kwork_work *work = zalloc(sizeof(*work)); in work_new() local 364 if (work == NULL) { in work_new() 365 pr_err("Failed to zalloc kwork work\n"); in work_new() 370 INIT_LIST_HEAD(&work->atom_list[i]); in work_new() [all …]
|
| /linux/drivers/staging/octeon/ |
| H A D | ethernet-rx.c | 59 * @work: Work queue entry pointing to the packet. 63 static inline int cvm_oct_check_rcv_error(struct cvmx_wqe *work) in cvm_oct_check_rcv_error() argument 68 port = work->word0.pip.cn68xx.pknd; in cvm_oct_check_rcv_error() 70 port = work->word1.cn38xx.ipprt; in cvm_oct_check_rcv_error() 72 if ((work->word2.snoip.err_code == 10) && (work->word1.len <= 64)) in cvm_oct_check_rcv_error() 81 if (work->word2.snoip.err_code == 5 || in cvm_oct_check_rcv_error() 82 work->word2.snoip.err_code == 7) { in cvm_oct_check_rcv_error() 99 cvmx_phys_to_ptr(work->packet_ptr.s.addr); in cvm_oct_check_rcv_error() 102 while (i < work->word1.len - 1) { in cvm_oct_check_rcv_error() 111 work->packet_ptr.s.addr += i + 1; in cvm_oct_check_rcv_error() [all …]
|
| /linux/kernel/ |
| H A D | kthread.c | 317 * functions which do some additional work in non-modular code such as 1010 * when they finish. There is defined a safe point for freezing when one work 1019 struct kthread_work *work; in kthread_worker_fn() local 1042 work = NULL; in kthread_worker_fn() 1045 work = list_first_entry(&worker->work_list, in kthread_worker_fn() 1047 list_del_init(&work->node); in kthread_worker_fn() 1049 worker->current_work = work; in kthread_worker_fn() 1052 if (work) { in kthread_worker_fn() 1053 kthread_work_func_t func = work->func; in kthread_worker_fn() 1055 trace_sched_kthread_work_execute_start(work); in kthread_worker_fn() 1187 queuing_blocked(struct kthread_worker * worker,struct kthread_work * work) queuing_blocked() argument 1195 kthread_insert_work_sanity_check(struct kthread_worker * worker,struct kthread_work * work) kthread_insert_work_sanity_check() argument 1205 kthread_insert_work(struct kthread_worker * worker,struct kthread_work * work,struct list_head * pos) kthread_insert_work() argument 1231 kthread_queue_work(struct kthread_worker * worker,struct kthread_work * work) kthread_queue_work() argument 1258 struct kthread_work *work = &dwork->work; kthread_delayed_work_timer_fn() local 1288 struct kthread_work *work = &dwork->work; __kthread_queue_delayed_work() local 1331 struct kthread_work *work = &dwork->work; kthread_queue_delayed_work() local 1348 struct kthread_work work; global() member 1352 kthread_flush_work_fn(struct kthread_work * work) kthread_flush_work_fn() argument 1365 kthread_flush_work(struct kthread_work * work) kthread_flush_work() argument 1404 kthread_cancel_delayed_work_timer(struct kthread_work * work,unsigned long * flags) kthread_cancel_delayed_work_timer() argument 1437 __kthread_cancel_work(struct kthread_work * work) __kthread_cancel_work() argument 1478 struct kthread_work *work = &dwork->work; kthread_mod_delayed_work() local 1521 __kthread_cancel_work_sync(struct kthread_work * work,bool is_dwork) __kthread_cancel_work_sync() argument 1574 kthread_cancel_work_sync(struct kthread_work * work) kthread_cancel_work_sync() argument [all...] |
| H A D | workqueue.c | 18 * This is the generic async execution mechanism. Work items as are 21 * normal work items and the other for high priority ones) and some extra 242 PWQ_STAT_STARTED, /* work items started execution */ 243 PWQ_STAT_COMPLETED, /* work items completed execution */ 249 PWQ_STAT_RESCUED, /* linked work items executed by rescuer */ 273 * When pwq->nr_active >= max_active, new work item is queued to 277 * All work items marked with WORK_STRUCT_INACTIVE do not participate in 278 * nr_active and all work items in pwq->inactive_works are marked with 279 * WORK_STRUCT_INACTIVE. But not all WORK_STRUCT_INACTIVE work items are 281 * pool->worklist or worker->scheduled. Those work itme 616 struct work_struct *work = addr; work_is_static_object() local 627 struct work_struct *work = addr; work_fixup_init() local 645 struct work_struct *work = addr; work_fixup_free() local 665 debug_work_activate(struct work_struct * work) debug_work_activate() argument 670 debug_work_deactivate(struct work_struct * work) debug_work_deactivate() argument 675 __init_work(struct work_struct * work,int onstack) __init_work() argument 684 destroy_work_on_stack(struct work_struct * work) destroy_work_on_stack() argument 690 destroy_delayed_work_on_stack(struct delayed_work * work) destroy_delayed_work_on_stack() argument 698 debug_work_activate(struct work_struct * work) debug_work_activate() argument 699 debug_work_deactivate(struct work_struct * work) debug_work_deactivate() argument 789 set_work_data(struct work_struct * work,unsigned long data) set_work_data() argument 795 set_work_pwq(struct work_struct * work,struct pool_workqueue * pwq,unsigned long flags) set_work_pwq() argument 802 set_work_pool_and_keep_pending(struct work_struct * work,int pool_id,unsigned long flags) set_work_pool_and_keep_pending() argument 809 set_work_pool_and_clear_pending(struct work_struct * work,int pool_id,unsigned long flags) set_work_pool_and_clear_pending() argument 857 get_work_pwq(struct work_struct * work) get_work_pwq() argument 882 get_work_pool(struct work_struct * work) get_work_pool() argument 1113 find_worker_executing_work(struct worker_pool * pool,struct work_struct * work) find_worker_executing_work() argument 1126 mayday_cursor_func(struct work_struct * work) mayday_cursor_func() argument 1146 move_linked_works(struct work_struct * work,struct list_head * head,struct work_struct ** nextp) move_linked_works() argument 1186 assign_work(struct work_struct * work,struct worker * worker,struct work_struct ** nextp) assign_work() argument 1288 struct work_struct *work = list_first_entry(&pool->worklist, kick_pool() local 1693 __pwq_activate_work(struct pool_workqueue * pwq,struct work_struct * work) __pwq_activate_work() argument 1806 struct work_struct *work = pwq_activate_first_inactive() local 1871 struct work_struct *work; node_activate_pending_pwq() local 2063 try_to_grab_pending(struct work_struct * work,u32 cflags,unsigned long * irq_flags) try_to_grab_pending() argument 2167 work_grab_pending(struct work_struct * work,u32 cflags,unsigned long * irq_flags) work_grab_pending() argument 2193 insert_work(struct pool_workqueue * pwq,struct work_struct * work,struct list_head * head,unsigned int extra_flags) insert_work() argument 2249 __queue_work(int cpu,struct workqueue_struct * wq,struct work_struct * work) __queue_work() argument 2366 clear_pending_if_disabled(struct work_struct * work) clear_pending_if_disabled() argument 2396 queue_work_on(int cpu,struct workqueue_struct * wq,struct work_struct * work) queue_work_on() argument 2464 queue_work_node(int node,struct workqueue_struct * wq,struct work_struct * work) queue_work_node() argument 2508 struct work_struct *work = &dwork->work; __queue_delayed_work() local 2565 struct work_struct *work = &dwork->work; queue_delayed_work_on() local 2639 struct work_struct *work = &rwork->work; queue_rcu_work() local 2962 idle_cull_fn(struct work_struct * work) idle_cull_fn() argument 3024 struct work_struct *work; pool_mayday_timeout() local 3174 process_one_work(struct worker * worker,struct work_struct * work) process_one_work() argument 3349 struct work_struct *work; process_scheduled_works() local 3434 struct work_struct *work = worker_thread() local 3462 struct work_struct *work, *n; assign_rescuer_work() local 3656 struct work_struct *work = bh_worker() local 3693 struct work_struct work; global() member 3698 drain_dead_softirq_workfn(struct work_struct * work) drain_dead_softirq_workfn() argument 3808 struct work_struct work; global() member 3813 wq_barrier_func(struct work_struct * work) wq_barrier_func() argument 3998 touch_work_lockdep_map(struct work_struct * work,struct workqueue_struct * wq) touch_work_lockdep_map() argument 4225 start_flush_work(struct work_struct * work,struct wq_barrier * barr,bool from_cancel) start_flush_work() argument 4281 __flush_work(struct work_struct * work,bool from_cancel) __flush_work() argument 4345 flush_work(struct work_struct * work) flush_work() argument 4412 __cancel_work(struct work_struct * work,u32 cflags) __cancel_work() argument 4431 __cancel_work_sync(struct work_struct * work,u32 cflags) __cancel_work_sync() argument 4458 cancel_work(struct work_struct * work) cancel_work() argument 4482 cancel_work_sync(struct work_struct * work) cancel_work_sync() argument 4537 disable_work(struct work_struct * work) disable_work() argument 4556 disable_work_sync(struct work_struct * work) disable_work_sync() argument 4572 enable_work(struct work_struct * work) enable_work() argument 4650 struct work_struct *work = per_cpu_ptr(works, cpu); schedule_on_each_cpu() local 5150 pwq_release_workfn(struct kthread_work * work) pwq_release_workfn() argument 6165 work_busy(struct work_struct * work) work_busy() argument 6312 pr_cont_work(bool comma,struct work_struct * work,struct pr_cont_work_struct * pcwsp) pr_cont_work() argument 6333 struct work_struct *work; show_pwq() local 6823 struct work_struct work; global() member 6829 work_for_cpu_fn(struct work_struct * work) work_for_cpu_fn() argument [all...] |
| /linux/net/wireless/ |
| H A D | debugfs.c | 148 struct wiphy_work work; in wiphy_locked_debugfs_read() 164 struct wiphy_work *work) in wiphy_locked_debugfs_read() 166 struct debugfs_read_work *w = container_of(work, typeof(*w), work); in wiphy_locked_debugfs_read() 177 wiphy_work_cancel(w->wiphy, &w->work); in wiphy_locked_debugfs_read() 192 struct debugfs_read_work work = { 200 .completion = COMPLETION_INITIALIZER_ONSTACK(work.completion), 204 .cancel_data = &work, 210 wiphy_work_init(&work.work, wiphy_locked_debugfs_read_wor 115 struct wiphy_work work; global() member 131 wiphy_locked_debugfs_read_work(struct wiphy * wiphy,struct wiphy_work * work) wiphy_locked_debugfs_read_work() argument 159 struct debugfs_read_work work = { wiphy_locked_debugfs_read() local 195 struct wiphy_work work; global() member 211 wiphy_locked_debugfs_write_work(struct wiphy * wiphy,struct wiphy_work * work) wiphy_locked_debugfs_write_work() argument 238 struct debugfs_write_work work = { wiphy_locked_debugfs_write() local [all...] |
| /linux/LICENSES/deprecated/ |
| H A D | CC0-1.0 | 26 and subsequent owner(s) (each and all, an "owner") of an original work of 27 authorship and/or a database (each, a "Work"). 29 Certain owners wish to permanently relinquish those rights to a Work for 37 works, or to gain reputation or greater distribution for their Work in 42 associating CC0 with a Work (the "Affirmer"), to the extent that he or she 43 is an owner of Copyright and Related Rights in the Work, voluntarily 44 elects to apply CC0 to the Work and publicly distribute the Work under its 46 Work and the meaning and intended legal effect of CC0 on those rights. 48 1. Copyright and Related Rights. A Work made available under CC0 may be 54 communicate, and translate a Work; [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | task_work.c | 45 struct elem *work = value; in process_work() local 47 bpf_copy_from_user_str(work->data, sizeof(work->data), (const void *)user_ptr, 0); in process_work() 57 struct elem *work; in oncpu_hash_map() local 65 work = bpf_map_lookup_elem(&hmap, &key); in oncpu_hash_map() 66 if (!work) in oncpu_hash_map() 68 bpf_task_work_schedule_resume(task, &work->tw, &hmap, process_work); in oncpu_hash_map() 75 struct elem *work; in oncpu_array_map() 79 work = bpf_map_lookup_elem(&arrmap, &key); in oncpu_array_map() 80 if (!work) in oncpu_array_map() 76 struct elem *work; oncpu_array_map() local 91 struct elem *work; oncpu_lru_map() local [all...] |
| /linux/io_uring/ |
| H A D | io-wq.c | 68 struct delayed_work work; member 160 static inline unsigned int io_get_work_hash(struct io_wq_work *work) in io_get_work_hash() argument 162 return __io_get_work_hash(atomic_read(&work->flags)); in io_get_work_hash() 274 * If there's work to do, returns true with acct->lock acquired. If not, 308 * starting work or finishing work. In either case, if it does in io_acct_activate_free_worker() 309 * to go sleep, we'll kick off a new task for this work anyway. in io_acct_activate_free_worker() 326 * Most likely an attempt to queue unbounded work on an io_wq that in io_wq_create_worker() 415 * work item after we canceled in io_wq_exit_workers(). in io_queue_worker_create() 432 /* Defer if current and next work ar 433 io_wq_hash_defer(struct io_wq_work * work,struct io_wq_acct * acct) io_wq_hash_defer() argument 529 struct io_wq_work *work, *tail; io_get_next_work() local 583 io_assign_current_work(struct io_worker * worker,struct io_wq_work * work) io_assign_current_work() argument 606 struct io_wq_work *work; io_worker_handle_work() local 801 io_wq_work_match_all(struct io_wq_work * work,void * data) io_wq_work_match_all() argument 883 io_workqueue_create(struct work_struct * work) io_workqueue_create() argument 990 io_run_cancel(struct io_wq_work * work,struct io_wq * wq) io_run_cancel() argument 1000 io_wq_insert_work(struct io_wq * wq,struct io_wq_acct * acct,struct io_wq_work * work,unsigned int work_flags) io_wq_insert_work() argument 1020 io_wq_work_match_item(struct io_wq_work * work,void * data) io_wq_work_match_item() argument 1025 io_wq_enqueue(struct io_wq * wq,struct io_wq_work * work) io_wq_enqueue() argument 1079 io_wq_hash_work(struct io_wq_work * work,void * val) io_wq_hash_work() argument 1089 __io_wq_worker_cancel(struct io_worker * worker,struct io_cb_cancel_data * match,struct io_wq_work * work) __io_wq_worker_cancel() argument 1118 io_wq_remove_pending(struct io_wq * wq,struct io_wq_acct * acct,struct io_wq_work * work,struct io_wq_work_node * prev) io_wq_remove_pending() argument 1140 struct io_wq_work *work; io_acct_cancel_pending_work() local [all...] |
| /linux/drivers/infiniband/core/ |
| H A D | cm.c | 95 struct cm_work *work); 187 struct delayed_work work; member 198 struct cm_work work; member 269 static void cm_work_handler(struct work_struct *work); 710 __be32 remote_id = timewait_info->work.remote_id; in cm_insert_remote_id() 716 if (be32_lt(remote_id, cur_timewait_info->work.remote_id)) in cm_insert_remote_id() 718 else if (be32_gt(remote_id, cur_timewait_info->work.remote_id)) in cm_insert_remote_id() 744 if (be32_lt(remote_id, timewait_info->work.remote_id)) in cm_find_remote_id() 746 else if (be32_gt(remote_id, timewait_info->work.remote_id)) in cm_find_remote_id() 753 res = cm_acquire_id(timewait_info->work in cm_find_remote_id() 888 struct cm_work *work; cm_dequeue_work() local 898 cm_free_work(struct cm_work * work) cm_free_work() argument 906 cm_queue_work_unlock(struct cm_id_private * cm_id_priv,struct cm_work * work) cm_queue_work_unlock() argument 1060 struct cm_work *work; cm_destroy_id() local 1771 cm_get_bth_pkey(struct cm_work * work) cm_get_bth_pkey() argument 1798 cm_opa_to_ib_sgid(struct cm_work * work,struct sa_path_rec * path) cm_opa_to_ib_sgid() argument 1818 cm_format_req_event(struct cm_work * work,struct cm_id_private * cm_id_priv,struct ib_cm_id * listen_id) cm_format_req_event() argument 1863 cm_process_work(struct cm_id_private * cm_id_priv,struct cm_work * work) cm_process_work() argument 1952 cm_dup_req_handler(struct cm_work * work,struct cm_id_private * cm_id_priv) cm_dup_req_handler() argument 2001 cm_match_req(struct cm_work * work,struct cm_id_private * cm_id_priv) cm_match_req() argument 2094 cm_req_handler(struct cm_work * work) cm_req_handler() argument 2416 cm_format_rep_event(struct cm_work * work,enum ib_qp_type qp_type) cm_format_rep_event() argument 2444 cm_dup_rep_handler(struct cm_work * work) cm_dup_rep_handler() argument 2489 cm_rep_handler(struct cm_work * work) cm_rep_handler() argument 2594 cm_establish_handler(struct cm_work * work) cm_establish_handler() argument 2617 cm_rtu_handler(struct cm_work * work) cm_rtu_handler() argument 2839 cm_dreq_handler(struct cm_work * work) cm_dreq_handler() argument 2913 cm_drep_handler(struct cm_work * work) cm_drep_handler() argument 3017 cm_format_rej_event(struct cm_work * work) cm_format_rej_event() argument 3055 cm_rej_handler(struct cm_work * work) cm_rej_handler() argument 3171 cm_mra_handler(struct cm_work * work) cm_mra_handler() argument 3284 cm_lap_handler(struct cm_work * work) cm_lap_handler() argument 3382 cm_apr_handler(struct cm_work * work) cm_apr_handler() argument 3426 cm_timewait_handler(struct cm_work * work) cm_timewait_handler() argument 3525 cm_format_sidr_req_event(struct cm_work * work,const struct cm_id_private * rx_cm_id,struct ib_cm_id * listen_id) cm_format_sidr_req_event() argument 3546 cm_sidr_req_handler(struct cm_work * work) cm_sidr_req_handler() argument 3702 cm_format_sidr_rep_event(struct cm_work * work,const struct cm_id_private * cm_id_priv) cm_format_sidr_rep_event() argument 3723 cm_sidr_rep_handler(struct cm_work * work) cm_sidr_rep_handler() argument 3841 struct cm_work *work = container_of(_work, struct cm_work, work.work); cm_work_handler() local 3896 struct cm_work *work; cm_establish() local 4000 struct cm_work *work; cm_recv_handler() local [all...] |
| /linux/LICENSES/preferred/ |
| H A D | LGPL-2.1 | 94 work, a derivative of the original library. The ordinary General Public 127 follow. Pay close attention to the difference between a "work based on the 128 library" and a "work that uses the library". The former contains code 144 The "Library", below, refers to any such software library or work which 145 has been distributed under these terms. A "work based on the Library" 146 means either the Library or any derivative work under copyright law: 147 that is to say, a work containing the Library or a portion of it, either 152 "Source code" for a work means the preferred form of the work for making 161 program is covered only if its contents constitute a work based on the 177 thus forming a work based on the Library, and copy and distribute such [all …]
|
| H A D | LGPL-2.0 | 88 a textual and legal sense, the linked executable is a combined work, a 108 follow. Pay close attention to the difference between a "work based on the 109 library" and a "work that uses the library". The former contains code 128 The "Library", below, refers to any such software library or work which 129 has been distributed under these terms. A "work based on the Library" 130 means either the Library or any derivative work under copyright law: 131 that is to say, a work containing the Library or a portion of it, either 136 "Source code" for a work means the preferred form of the work for making 145 program is covered only if its contents constitute a work based on the 161 thus forming a work based on the Library, and copy and distribute such [all …]
|