| /linux/kernel/cgroup/ |
| H A D | rstat.c | 100 if (llist_on_list(&rstatc->lnode)) in css_rstat_updated() 120 self = &rstatc->lnode; in css_rstat_updated() 121 if (!try_cmpxchg(&rstatc->lnode.next, &self, NULL)) in css_rstat_updated() 125 llist_add(&rstatc->lnode, lhead); in css_rstat_updated() 160 struct llist_node *lnode; in css_process_update_tree() local 162 while ((lnode = llist_del_first_init(lhead))) { in css_process_update_tree() 180 rstatc = container_of(lnode, struct css_rstat_cpu, lnode); in css_process_update_tree() 465 init_llist_node(&rstatc->lnode); in css_rstat_init()
|
| H A D | cgroup.c | 7030 struct llist_node *lnode; in cgrp_dead_tasks_iwork_fn() local 7033 lnode = llist_del_all(this_cpu_ptr(&cgrp_dead_tasks)); in cgrp_dead_tasks_iwork_fn() 7034 llist_for_each_entry_safe(task, next, lnode, cg_dead_lnode) { in cgrp_dead_tasks_iwork_fn()
|
| /linux/drivers/scsi/csiostor/ |
| H A D | csio_scsi.c | 93 return ((ioreq->lnode == sld->lnode) && in csio_scsi_match_io() 98 return ((ioreq->lnode == sld->lnode) && in csio_scsi_match_io() 101 return (ioreq->lnode == sld->lnode); in csio_scsi_match_io() 203 struct csio_hw *hw = req->lnode->hwp; in csio_scsi_init_cmd_wr() 258 struct csio_hw *hw = req->lnode->hwp; in csio_scsi_cmd() 362 struct csio_hw *hw = req->lnode->hwp; in csio_scsi_init_read_wr() 415 struct csio_hw *hw = req->lnode->hwp; in csio_scsi_init_write_wr() 483 struct csio_hw *hw = req->lnode->hwp; in csio_scsi_read() 520 struct csio_hw *hw = req->lnode->hwp; in csio_scsi_write() 556 struct csio_hw *hw = req->lnode->hwp; in csio_setup_ddp() [all …]
|
| H A D | csio_lnode.c | 270 struct csio_lnode *ln = fdmi_req->lnode; in csio_ln_fdmi_done() 299 struct csio_lnode *ln = fdmi_req->lnode; in csio_ln_fdmi_rhba_cbfn() 412 struct csio_lnode *ln = fdmi_req->lnode; in csio_ln_fdmi_dprt_cbfn() 513 struct csio_lnode *ln = fdmi_req->lnode; in csio_ln_fdmi_dhba_cbfn() 584 fdmi_req->lnode = ln; in csio_ln_fdmi_start() 1699 io_req->lnode->hwp, io_req->iq_idx)); in csio_ln_prep_ecwr() 1729 struct csio_lnode *ln = io_req->lnode; in csio_ln_mgmt_submit_wr() 1810 struct csio_hw *hw = csio_lnode_to_hw(io_req->lnode); in csio_ln_mgmt_submit_req()
|
| H A D | csio_scsi.h | 187 struct csio_lnode *lnode; member
|
| H A D | csio_wr.h | 251 struct csio_lnode *lnode; /* Owner lnode */ member
|
| /linux/drivers/accel/habanalabs/common/ |
| H A D | memory.c | 1429 struct hl_vm_hw_block_list_node *lnode = in hw_block_vm_close() local 1431 struct hl_ctx *ctx = lnode->ctx; in hw_block_vm_close() 1434 new_mmap_size = lnode->mapped_size - (vma->vm_end - vma->vm_start); in hw_block_vm_close() 1436 lnode->mapped_size = new_mmap_size; in hw_block_vm_close() 1441 list_del(&lnode->node); in hw_block_vm_close() 1444 kfree(lnode); in hw_block_vm_close() 1462 struct hl_vm_hw_block_list_node *lnode; in hl_hw_block_mmap() local 1485 lnode = kzalloc_obj(*lnode); in hl_hw_block_mmap() 1486 if (!lnode) in hl_hw_block_mmap() 1491 kfree(lnode); in hl_hw_block_mmap() [all …]
|
| H A D | debugfs.c | 239 struct hl_vm_hw_block_list_node *lnode; in vm_show() local 290 list_for_each_entry(lnode, &ctx->hw_block_mem_list, node) { in vm_show() 293 lnode->vaddr, lnode->block_size, lnode->mapped_size, in vm_show() 294 lnode->id); in vm_show()
|
| /linux/net/netfilter/ |
| H A D | nf_conntrack_expect.c | 62 hlist_del_rcu(&exp->lnode); in nf_ct_unlink_expect_report() 239 hlist_for_each_entry_safe(exp, next, &help->expectations, lnode) { in nf_ct_remove_expectations() 405 hlist_add_head_rcu(&exp->lnode, &master_help->expectations); in nf_ct_expect_insert() 422 hlist_for_each_entry(exp, &master_help->expectations, lnode) { in evict_oldest_expect()
|
| H A D | nf_nat_sip.c | 340 hlist_for_each_entry(pair_exp, &help->expectations, lnode) { in nf_nat_sip_expected()
|
| H A D | nf_conntrack_netlink.c | 3226 hlist_for_each_entry_rcu(exp, &help->expectations, lnode) { in ctnetlink_exp_ct_dump_table()
|
| /linux/include/net/netfilter/ |
| H A D | nf_conntrack_expect.h | 20 struct hlist_node lnode; member
|
| /linux/block/ |
| H A D | blk-cgroup.c | 1046 struct llist_node *lnode; in __blkcg_rstat_flush() local 1052 lnode = llist_del_all(lhead); in __blkcg_rstat_flush() 1053 if (!lnode) in __blkcg_rstat_flush() 1067 llist_for_each_entry_safe(bisc, next_bisc, lnode, lnode) { in __blkcg_rstat_flush() 1109 llist_add(&parent->iostat.lnode, plhead); in __blkcg_rstat_flush() 2223 llist_add(&bis->lnode, lhead); in blk_cgroup_bio_start()
|
| /linux/drivers/gpu/drm/exynos/ |
| H A D | exynos_drm_g2d.c | 362 struct g2d_cmdlist_node *lnode; in g2d_add_cmdlist_to_inuse() local 368 lnode = list_entry(file_priv->inuse_cmdlist.prev, in g2d_add_cmdlist_to_inuse() 370 lnode->cmdlist->data[lnode->cmdlist->last] = node->dma_addr; in g2d_add_cmdlist_to_inuse()
|
| /linux/include/linux/ |
| H A D | cgroup-defs.h | 399 struct llist_node lnode; /* lockless list for update */ member
|
| H A D | bpf.h | 1303 struct list_head lnode; member 1505 .lnode = LIST_HEAD_INIT(_name.ksym.lnode), \
|
| /linux/kernel/bpf/ |
| H A D | core.c | 134 INIT_LIST_HEAD_RCU(&fp->aux->ksym.lnode); in bpf_prog_alloc_no_stats() 136 INIT_LIST_HEAD_RCU(&fp->aux->ksym_prefix.lnode); in bpf_prog_alloc_no_stats() 642 WARN_ON_ONCE(!list_empty(&ksym->lnode)); in bpf_ksym_add() 643 list_add_tail_rcu(&ksym->lnode, &bpf_kallsyms); in bpf_ksym_add() 650 if (list_empty(&ksym->lnode)) in __bpf_ksym_del() 654 list_del_rcu(&ksym->lnode); in __bpf_ksym_del() 812 list_for_each_entry_rcu(ksym, &bpf_kallsyms, lnode) { in bpf_get_kallsym()
|
| H A D | bpf_struct_ops.c | 648 INIT_LIST_HEAD_RCU(&ksym->lnode); in bpf_struct_ops_ksym_init()
|
| H A D | trampoline.c | 591 INIT_LIST_HEAD_RCU(&ksym->lnode); in bpf_tramp_image_alloc()
|
| /linux/tools/power/x86/turbostat/ |
| H A D | turbostat.c | 6052 int pkg, node, lnode, cpu, cpux; in set_node_data() local 6061 lnode = 0; in set_node_data() 6068 cpus[cpu].logical_node_id = lnode; in set_node_data() 6077 cpus[cpux].logical_node_id = lnode; in set_node_data() 6081 lnode++; in set_node_data() 6082 if (lnode > topo.nodes_per_pkg) in set_node_data() 6083 topo.nodes_per_pkg = lnode; in set_node_data()
|