| /linux/kernel/cgroup/ |
| H A D | rstat.c | 100 if (llist_on_list(&rstatc->lnode)) in css_rstat_updated() 105 * and may try to insert the same per-cpu lnode into the llist. Note in css_rstat_updated() 107 * this same per-cpu lnode can be modified through init_llist_node() in css_rstat_updated() 111 * fact that lnode points to itself when not on a list and then use in css_rstat_updated() 114 * successful and the winner will eventually add the per-cpu lnode to in css_rstat_updated() 120 self = &rstatc->lnode; in css_rstat_updated() 121 if (!try_cmpxchg(&rstatc->lnode.next, &self, NULL)) in css_rstat_updated() 125 llist_add(&rstatc->lnode, lhead); in __css_process_update_tree() 160 struct llist_node *lnode; in css_process_update_tree() 162 while ((lnode in css_process_update_tree() 157 struct llist_node *lnode; css_process_update_tree() local [all...] |
| H A D | cgroup.c | 7033 struct llist_node *lnode; in cgrp_dead_tasks_iwork_fn() local 7036 lnode = llist_del_all(this_cpu_ptr(&cgrp_dead_tasks)); in cgrp_dead_tasks_iwork_fn() 7037 llist_for_each_entry_safe(task, next, lnode, cg_dead_lnode) { in cgrp_dead_tasks_iwork_fn()
|
| /linux/drivers/scsi/csiostor/ |
| H A D | csio_scsi.c | 93 return ((ioreq->lnode == sld->lnode) && in csio_scsi_match_io() 98 return ((ioreq->lnode == sld->lnode) && in csio_scsi_match_io() 101 return (ioreq->lnode == sld->lnode); in csio_scsi_match_io() 203 struct csio_hw *hw = req->lnode->hwp; in csio_scsi_init_cmd_wr() 258 struct csio_hw *hw = req->lnode->hwp; in csio_scsi_cmd() 362 struct csio_hw *hw = req->lnode->hwp; in csio_scsi_init_read_wr() 415 struct csio_hw *hw = req->lnode->hwp; in csio_scsi_init_write_wr() 483 struct csio_hw *hw = req->lnode->hwp; in csio_scsi_read() 520 struct csio_hw *hw = req->lnode->hwp; in csio_scsi_write() 556 struct csio_hw *hw = req->lnode->hwp; in csio_setup_ddp() [all …]
|
| H A D | csio_lnode.c | 270 struct csio_lnode *ln = fdmi_req->lnode; in csio_ln_fdmi_done() 299 struct csio_lnode *ln = fdmi_req->lnode; in csio_ln_fdmi_rhba_cbfn() 412 struct csio_lnode *ln = fdmi_req->lnode; in csio_ln_fdmi_dprt_cbfn() 513 struct csio_lnode *ln = fdmi_req->lnode; in csio_ln_fdmi_dhba_cbfn() 584 fdmi_req->lnode = ln; in csio_ln_fdmi_start() 1699 io_req->lnode->hwp, io_req->iq_idx)); in csio_ln_prep_ecwr() 1729 struct csio_lnode *ln = io_req->lnode; in csio_ln_mgmt_submit_wr() 1810 struct csio_hw *hw = csio_lnode_to_hw(io_req->lnode); in csio_ln_mgmt_submit_req()
|
| H A D | csio_scsi.h | 187 struct csio_lnode *lnode; member
|
| H A D | csio_wr.h | 251 struct csio_lnode *lnode; /* Owner lnode */ member
|
| /linux/net/netfilter/ |
| H A D | nf_conntrack_expect.c | 62 hlist_del_rcu(&exp->lnode); in nf_ct_unlink_expect_report() 239 hlist_for_each_entry_safe(exp, next, &help->expectations, lnode) { in nf_ct_remove_expectations() 405 hlist_add_head_rcu(&exp->lnode, &master_help->expectations); in nf_ct_expect_insert() 422 hlist_for_each_entry(exp, &master_help->expectations, lnode) { in evict_oldest_expect()
|
| H A D | nf_nat_sip.c | 340 hlist_for_each_entry(pair_exp, &help->expectations, lnode) { in nf_nat_sip_expected()
|
| /linux/include/net/netfilter/ |
| H A D | nf_conntrack_expect.h | 20 struct hlist_node lnode; member
|
| /linux/block/ |
| H A D | blk-cgroup.c | 1046 struct llist_node *lnode; in __blkcg_rstat_flush() local 1052 lnode = llist_del_all(lhead); in __blkcg_rstat_flush() 1053 if (!lnode) in __blkcg_rstat_flush() 1067 llist_for_each_entry_safe(bisc, next_bisc, lnode, lnode) { in __blkcg_rstat_flush() 1109 llist_add(&parent->iostat.lnode, plhead); in __blkcg_rstat_flush() 2223 llist_add(&bis->lnode, lhead); in blk_cgroup_bio_start()
|
| /linux/include/linux/ |
| H A D | cgroup-defs.h | 399 struct llist_node lnode; /* lockless list for update */ member
|
| H A D | filter.h | 1288 return list_empty(&fp->aux->ksym.lnode) || in bpf_prog_kallsyms_verify_off() 1289 fp->aux->ksym.lnode.prev == LIST_POISON2; in bpf_prog_kallsyms_verify_off()
|
| /linux/kernel/bpf/ |
| H A D | core.c | 132 INIT_LIST_HEAD_RCU(&fp->aux->ksym.lnode); in bpf_prog_alloc_no_stats() 134 INIT_LIST_HEAD_RCU(&fp->aux->ksym_prefix.lnode); in bpf_prog_alloc_no_stats() 638 WARN_ON_ONCE(!list_empty(&ksym->lnode)); in bpf_ksym_add() 639 list_add_tail_rcu(&ksym->lnode, &bpf_kallsyms); in bpf_ksym_add() 646 if (list_empty(&ksym->lnode)) in __bpf_ksym_del() 650 list_del_rcu(&ksym->lnode); in __bpf_ksym_del() 808 list_for_each_entry_rcu(ksym, &bpf_kallsyms, lnode) {
|
| H A D | trampoline.c | 405 INIT_LIST_HEAD_RCU(&ksym->lnode); in bpf_tramp_image_alloc()
|
| /linux/tools/power/x86/turbostat/ |
| H A D | turbostat.c | 5877 int pkg, node, lnode, cpu, cpux; in set_node_data() local 5886 lnode = 0; in set_node_data() 5893 cpus[cpu].logical_node_id = lnode; in set_node_data() 5902 cpus[cpux].logical_node_id = lnode; in set_node_data() 5906 lnode++; in set_node_data() 5907 if (lnode > topo.nodes_per_pkg) in set_node_data() 5908 topo.nodes_per_pkg = lnode; in set_node_data()
|