Home
last modified time | relevance | path

Searched refs:gl (Results 1 – 25 of 26) sorted by relevance

12

/linux/fs/gfs2/
H A Dglock.c57 struct gfs2_glock *gl; /* current glock struct */ member
61 typedef void (*glock_examiner) (struct gfs2_glock * gl);
63 static void do_xmote(struct gfs2_glock *gl, struct gfs2_holder *gh,
65 static void request_demote(struct gfs2_glock *gl, unsigned int state,
120 static void wake_up_glock(struct gfs2_glock *gl) in wake_up_glock() argument
122 wait_queue_head_t *wq = glock_waitqueue(&gl->gl_name); in wake_up_glock()
125 __wake_up(wq, TASK_NORMAL, 1, &gl->gl_name); in wake_up_glock()
130 struct gfs2_glock *gl = container_of(rcu, struct gfs2_glock, gl_rcu); in gfs2_glock_dealloc() local
132 kfree(gl->gl_lksb.sb_lvbptr); in gfs2_glock_dealloc()
133 if (gl->gl_ops->go_flags & GLOF_ASPACE) { in gfs2_glock_dealloc()
[all …]
H A Dmain.c52 struct gfs2_glock *gl = foo; in gfs2_init_glock_once() local
54 INIT_LIST_HEAD(&gl->gl_holders); in gfs2_init_glock_once()
55 INIT_LIST_HEAD(&gl->gl_lru); in gfs2_init_glock_once()
56 INIT_LIST_HEAD(&gl->gl_ail_list); in gfs2_init_glock_once()
57 atomic_set(&gl->gl_ail_count, 0); in gfs2_init_glock_once()
58 atomic_set(&gl->gl_revokes, 0); in gfs2_init_glock_once()
H A Dlops.c73 struct gfs2_glock *gl = bd->bd_gl; in maybe_release_space() local
74 struct gfs2_sbd *sdp = glock_sbd(gl); in maybe_release_space()
75 struct gfs2_rgrpd *rgd = gfs2_glock2rgrp(gl); in maybe_release_space()
76 unsigned int index = bd->bd_bh->b_blocknr - glock_number(gl); in maybe_release_space()
122 struct gfs2_glock *gl = bd->bd_gl; in gfs2_unpin() local
123 list_add(&bd->bd_ail_gl_list, &gl->gl_ail_list); in gfs2_unpin()
124 atomic_inc(&gl->gl_ail_count); in gfs2_unpin()
780 struct gfs2_glock *gl = ip->i_gl; in buf_lo_scan_elements() local
803 bh_ip = gfs2_meta_new(gl, blkno); in buf_lo_scan_elements()
890 struct gfs2_glock *gl; in gfs2_drain_revokes() local
[all …]
H A Dglops.h25 int gfs2_inode_metasync(struct gfs2_glock *gl);
26 void gfs2_ail_flush(struct gfs2_glock *gl, bool fsync);
H A Drgrp.c720 struct gfs2_glock *gl; in gfs2_clear_rgrpd() local
724 gl = rgd->rd_gl; in gfs2_clear_rgrpd()
728 if (gl) { in gfs2_clear_rgrpd()
729 if (gl->gl_state != LM_ST_UNLOCKED) { in gfs2_clear_rgrpd()
730 gfs2_glock_cb(gl, LM_ST_UNLOCKED); in gfs2_clear_rgrpd()
731 flush_delayed_work(&gl->gl_work); in gfs2_clear_rgrpd()
734 glock_clear_object(gl, rgd); in gfs2_clear_rgrpd()
735 gfs2_glock_put(gl); in gfs2_clear_rgrpd()
1037 struct gfs2_glock *gl = ip->i_gl; in gfs2_rindex_update() local
1044 if (!gfs2_glock_is_locked_by_me(gl)) { in gfs2_rindex_update()
[all …]
H A Dsuper.c1012 struct gfs2_glock *gl = ip->i_iopen_gh.gh_gl; in gfs2_drop_inode() local
1013 if (glock_needs_demote(gl)) in gfs2_drop_inode()
1025 struct gfs2_glock *gl = ip->i_iopen_gh.gh_gl; in gfs2_drop_inode() local
1027 gfs2_glock_hold(gl); in gfs2_drop_inode()
1028 if (!gfs2_queue_verify_delete(gl, true)) in gfs2_drop_inode()
1029 gfs2_glock_put_async(gl); in gfs2_drop_inode()
1167 static void gfs2_glock_put_eventually(struct gfs2_glock *gl) in gfs2_glock_put_eventually() argument
1170 gfs2_glock_put_async(gl); in gfs2_glock_put_eventually()
1172 gfs2_glock_put(gl); in gfs2_glock_put_eventually()
1436 struct gfs2_glock *gl = ip->i_iopen_gh.gh_gl; in gfs2_evict_inode() local
[all …]
H A Dfile.c1472 struct gfs2_glock *gl = gfs2_glock_hold(fl_gh->gh_gl); in __flock_holder_uninit() local
1482 gfs2_glock_put(gl); in __flock_holder_uninit()
1490 struct gfs2_glock *gl; in do_flock() local
1515 &gfs2_flock_glops, CREATE, &gl); in do_flock()
1519 gfs2_holder_init(gl, state, flags, fl_gh); in do_flock()
1521 gfs2_glock_put(gl); in do_flock()
H A Drgrp.h34 int gfs2_rgrp_go_instantiate(struct gfs2_glock *gl);
H A Drecovery.c36 struct gfs2_glock *gl = ip->i_gl; in gfs2_replay_read_block() local
50 *bh = gfs2_meta_ra(gl, dblock, extlen); in gfs2_replay_read_block()
H A Dbmap.c294 static void gfs2_metapath_ra(struct gfs2_glock *gl, __be64 *start, __be64 *end) in gfs2_metapath_ra() argument
304 rabh = gfs2_getbuf(gl, be64_to_cpu(*t), CREATE); in gfs2_metapath_ra()
609 struct gfs2_glock *gl, unsigned int i, in gfs2_indirect_init() argument
617 mp->mp_bh[i] = gfs2_meta_new(gl, bn); in gfs2_indirect_init()
618 gfs2_trans_add_meta(gl, mp->mp_bh[i]); in gfs2_indirect_init()
H A Ddir.c1484 struct gfs2_glock *gl = ip->i_gl; in gfs2_dir_readahead() local
1504 bh = gfs2_getbuf(gl, blocknr, 1); in gfs2_dir_readahead()
/linux/drivers/target/iscsi/cxgbit/
H A Dcxgbit_main.c219 cxgbit_copy_frags(struct sk_buff *skb, const struct pkt_gl *gl, in cxgbit_copy_frags() argument
226 __skb_fill_page_desc(skb, skb_frag_idx, gl->frags[0].page, in cxgbit_copy_frags()
227 gl->frags[0].offset + offset, in cxgbit_copy_frags()
228 gl->frags[0].size - offset); in cxgbit_copy_frags()
229 for (i = 1; i < gl->nfrags; i++) in cxgbit_copy_frags()
231 gl->frags[i].page, in cxgbit_copy_frags()
232 gl->frags[i].offset, in cxgbit_copy_frags()
233 gl->frags[i].size); in cxgbit_copy_frags()
235 skb_shinfo(skb)->nr_frags += gl->nfrags; in cxgbit_copy_frags()
238 get_page(gl->frags[gl->nfrags - 1].page); in cxgbit_copy_frags()
[all …]
/linux/drivers/net/ethernet/chelsio/cxgb4vf/
H A Dsge.c1470 const struct pkt_gl *gl, in copy_frags() argument
1476 __skb_fill_page_desc(skb, 0, gl->frags[0].page, in copy_frags()
1477 gl->frags[0].offset + offset, in copy_frags()
1478 gl->frags[0].size - offset); in copy_frags()
1479 skb_shinfo(skb)->nr_frags = gl->nfrags; in copy_frags()
1480 for (i = 1; i < gl->nfrags; i++) in copy_frags()
1481 __skb_fill_page_desc(skb, i, gl->frags[i].page, in copy_frags()
1482 gl->frags[i].offset, in copy_frags()
1483 gl->frags[i].size); in copy_frags()
1486 get_page(gl->frags[gl->nfrags - 1].page); in copy_frags()
[all …]
H A Dcxgb4vf_main.c488 const struct pkt_gl *gl) in fwevtq_handler() argument
/linux/drivers/net/ethernet/chelsio/inline_crypto/chtls/
H A Dchtls_main.c339 static struct sk_buff *copy_gl_to_skb_pkt(const struct pkt_gl *gl, in copy_gl_to_skb_pkt() argument
349 skb = alloc_skb(size_add(gl->tot_len, in copy_gl_to_skb_pkt()
354 __skb_put(skb, gl->tot_len + sizeof(struct cpl_pass_accept_req) in copy_gl_to_skb_pkt()
359 , gl->va + pktshift, in copy_gl_to_skb_pkt()
360 gl->tot_len - pktshift); in copy_gl_to_skb_pkt()
366 const struct pkt_gl *gl, const __be64 *rsp) in chtls_recv_packet() argument
372 skb = copy_gl_to_skb_pkt(gl, rsp, cdev->lldi->sge_pktshift); in chtls_recv_packet()
440 const struct pkt_gl *gl) in chtls_uld_rx_handler() argument
449 if (chtls_recv_packet(cdev, gl, rsp) < 0) in chtls_uld_rx_handler()
454 if (!gl) in chtls_uld_rx_handler()
[all …]
/linux/drivers/infiniband/hw/cxgb4/
H A Ddevice.c1102 static inline struct sk_buff *copy_gl_to_skb_pkt(const struct pkt_gl *gl, in copy_gl_to_skb_pkt() argument
1115 skb = alloc_skb(size_add(gl->tot_len, in copy_gl_to_skb_pkt()
1122 __skb_put(skb, gl->tot_len + sizeof(struct cpl_pass_accept_req) + in copy_gl_to_skb_pkt()
1137 gl->va + pktshift, in copy_gl_to_skb_pkt()
1138 gl->tot_len - pktshift); in copy_gl_to_skb_pkt()
1142 static inline int recv_rx_pkt(struct c4iw_dev *dev, const struct pkt_gl *gl, in recv_rx_pkt() argument
1151 skb = copy_gl_to_skb_pkt(gl , rsp, dev->rdev.lldi.sge_pktshift); in recv_rx_pkt()
1167 const struct pkt_gl *gl) in c4iw_uld_rx_handler() argument
1174 if (gl == NULL) { in c4iw_uld_rx_handler()
1183 } else if (gl == CXGB4_MSG_AN) { in c4iw_uld_rx_handler()
[all …]
/linux/drivers/net/ethernet/chelsio/cxgb4/
H A Dsge.c3306 const struct pkt_gl *gl, unsigned int offset) in copy_frags() argument
3311 __skb_fill_page_desc(skb, 0, gl->frags[0].page, in copy_frags()
3312 gl->frags[0].offset + offset, in copy_frags()
3313 gl->frags[0].size - offset); in copy_frags()
3314 skb_shinfo(skb)->nr_frags = gl->nfrags; in copy_frags()
3315 for (i = 1; i < gl->nfrags; i++) in copy_frags()
3316 __skb_fill_page_desc(skb, i, gl->frags[i].page, in copy_frags()
3317 gl->frags[i].offset, in copy_frags()
3318 gl->frags[i].size); in copy_frags()
3321 get_page(gl->frags[gl->nfrags - 1].page); in copy_frags()
[all …]
H A Dcxgb4_uld.h483 const struct pkt_gl *gl);
487 const struct pkt_gl *gl,
521 struct sk_buff *cxgb4_pktgl_to_skb(const struct pkt_gl *gl,
H A Dcxgb4_uld.c75 const struct pkt_gl *gl) in uldrx_handler() argument
88 rsp, gl, &q->lro_mgr, in uldrx_handler()
92 rsp, gl); in uldrx_handler()
99 if (!gl) in uldrx_handler()
101 else if (gl == CXGB4_MSG_AN) in uldrx_handler()
/linux/drivers/scsi/qedi/
H A Dqedi_main.c1518 struct global_queue **gl = qedi->global_queues; in qedi_free_global_queues() local
1521 if (!gl[i]) in qedi_free_global_queues()
1524 if (gl[i]->cq) in qedi_free_global_queues()
1525 dma_free_coherent(&qedi->pdev->dev, gl[i]->cq_mem_size, in qedi_free_global_queues()
1526 gl[i]->cq, gl[i]->cq_dma); in qedi_free_global_queues()
1527 if (gl[i]->cq_pbl) in qedi_free_global_queues()
1528 dma_free_coherent(&qedi->pdev->dev, gl[i]->cq_pbl_size, in qedi_free_global_queues()
1529 gl[i]->cq_pbl, gl[i]->cq_pbl_dma); in qedi_free_global_queues()
1531 kfree(gl[i]); in qedi_free_global_queues()
/linux/tools/testing/cxl/test/
H A Dmem.c577 struct cxl_mbox_get_log *gl = cmd->payload_in; in mock_get_log() local
578 u32 offset = le32_to_cpu(gl->offset); in mock_get_log()
579 u32 length = le32_to_cpu(gl->length); in mock_get_log()
583 if (cmd->size_in < sizeof(*gl)) in mock_get_log()
589 if (!uuid_equal(&gl->uuid, &uuid)) in mock_get_log()
/linux/drivers/scsi/qedf/
H A Dqedf_main.c2946 struct global_queue **gl = qedf->global_queues; in qedf_free_global_queues() local
2949 if (!gl[i]) in qedf_free_global_queues()
2952 if (gl[i]->cq) in qedf_free_global_queues()
2954 gl[i]->cq_mem_size, gl[i]->cq, gl[i]->cq_dma); in qedf_free_global_queues()
2955 if (gl[i]->cq_pbl) in qedf_free_global_queues()
2956 dma_free_coherent(&qedf->pdev->dev, gl[i]->cq_pbl_size, in qedf_free_global_queues()
2957 gl[i]->cq_pbl, gl[i]->cq_pbl_dma); in qedf_free_global_queues()
2959 kfree(gl[i]); in qedf_free_global_queues()
/linux/tools/perf/util/
H A Dprobe-finder.c1033 static int pubname_search_cb(Dwarf *dbg, Dwarf_Global *gl, void *data) in pubname_search_cb() argument
1038 if (dwarf_offdie(dbg, gl->die_offset, param->sp_die)) { in pubname_search_cb()
1043 if (!dwarf_offdie(dbg, gl->cu_offset, param->cu_die)) in pubname_search_cb()
/linux/Documentation/userspace-api/media/v4l/
H A Dv4l2.rst247 :revision: 0.28 / 2009-08-26 (*gl*)
/linux/drivers/net/ethernet/fungible/funeth/
H A Dfuneth_main.c186 struct fun_dataop_gl gl; in fun_config_rss() member
208 fun_dataop_gl_init(&cmd.gl, 0, 0, len, fp->rss_dma_addr); in fun_config_rss()

12