| /linux/fs/gfs2/ |
| H A D | glock.c | 57 struct gfs2_glock *gl; /* current glock struct */ member 61 typedef void (*glock_examiner) (struct gfs2_glock * gl); 63 static void do_xmote(struct gfs2_glock *gl, struct gfs2_holder *gh, 65 static void request_demote(struct gfs2_glock *gl, unsigned int state, 120 static void wake_up_glock(struct gfs2_glock *gl) in wake_up_glock() argument 122 wait_queue_head_t *wq = glock_waitqueue(&gl->gl_name); in wake_up_glock() 125 __wake_up(wq, TASK_NORMAL, 1, &gl->gl_name); in wake_up_glock() 130 struct gfs2_glock *gl = container_of(rcu, struct gfs2_glock, gl_rcu); in gfs2_glock_dealloc() local 132 kfree(gl->gl_lksb.sb_lvbptr); in gfs2_glock_dealloc() 133 if (gl->gl_ops->go_flags & GLOF_ASPACE) { in gfs2_glock_dealloc() [all …]
|
| H A D | main.c | 52 struct gfs2_glock *gl = foo; in gfs2_init_glock_once() local 54 INIT_LIST_HEAD(&gl->gl_holders); in gfs2_init_glock_once() 55 INIT_LIST_HEAD(&gl->gl_lru); in gfs2_init_glock_once() 56 INIT_LIST_HEAD(&gl->gl_ail_list); in gfs2_init_glock_once() 57 atomic_set(&gl->gl_ail_count, 0); in gfs2_init_glock_once() 58 atomic_set(&gl->gl_revokes, 0); in gfs2_init_glock_once()
|
| H A D | lops.c | 73 struct gfs2_glock *gl = bd->bd_gl; in maybe_release_space() local 74 struct gfs2_sbd *sdp = glock_sbd(gl); in maybe_release_space() 75 struct gfs2_rgrpd *rgd = gfs2_glock2rgrp(gl); in maybe_release_space() 76 unsigned int index = bd->bd_bh->b_blocknr - glock_number(gl); in maybe_release_space() 122 struct gfs2_glock *gl = bd->bd_gl; in gfs2_unpin() local 123 list_add(&bd->bd_ail_gl_list, &gl->gl_ail_list); in gfs2_unpin() 124 atomic_inc(&gl->gl_ail_count); in gfs2_unpin() 780 struct gfs2_glock *gl = ip->i_gl; in buf_lo_scan_elements() local 803 bh_ip = gfs2_meta_new(gl, blkno); in buf_lo_scan_elements() 890 struct gfs2_glock *gl; in gfs2_drain_revokes() local [all …]
|
| H A D | glops.h | 25 int gfs2_inode_metasync(struct gfs2_glock *gl); 26 void gfs2_ail_flush(struct gfs2_glock *gl, bool fsync);
|
| H A D | rgrp.c | 720 struct gfs2_glock *gl; in gfs2_clear_rgrpd() local 724 gl = rgd->rd_gl; in gfs2_clear_rgrpd() 728 if (gl) { in gfs2_clear_rgrpd() 729 if (gl->gl_state != LM_ST_UNLOCKED) { in gfs2_clear_rgrpd() 730 gfs2_glock_cb(gl, LM_ST_UNLOCKED); in gfs2_clear_rgrpd() 731 flush_delayed_work(&gl->gl_work); in gfs2_clear_rgrpd() 734 glock_clear_object(gl, rgd); in gfs2_clear_rgrpd() 735 gfs2_glock_put(gl); in gfs2_clear_rgrpd() 1037 struct gfs2_glock *gl = ip->i_gl; in gfs2_rindex_update() local 1044 if (!gfs2_glock_is_locked_by_me(gl)) { in gfs2_rindex_update() [all …]
|
| H A D | super.c | 1012 struct gfs2_glock *gl = ip->i_iopen_gh.gh_gl; in gfs2_drop_inode() local 1013 if (glock_needs_demote(gl)) in gfs2_drop_inode() 1025 struct gfs2_glock *gl = ip->i_iopen_gh.gh_gl; in gfs2_drop_inode() local 1027 gfs2_glock_hold(gl); in gfs2_drop_inode() 1028 if (!gfs2_queue_verify_delete(gl, true)) in gfs2_drop_inode() 1029 gfs2_glock_put_async(gl); in gfs2_drop_inode() 1167 static void gfs2_glock_put_eventually(struct gfs2_glock *gl) in gfs2_glock_put_eventually() argument 1170 gfs2_glock_put_async(gl); in gfs2_glock_put_eventually() 1172 gfs2_glock_put(gl); in gfs2_glock_put_eventually() 1436 struct gfs2_glock *gl = ip->i_iopen_gh.gh_gl; in gfs2_evict_inode() local [all …]
|
| H A D | file.c | 1472 struct gfs2_glock *gl = gfs2_glock_hold(fl_gh->gh_gl); in __flock_holder_uninit() local 1482 gfs2_glock_put(gl); in __flock_holder_uninit() 1490 struct gfs2_glock *gl; in do_flock() local 1515 &gfs2_flock_glops, CREATE, &gl); in do_flock() 1519 gfs2_holder_init(gl, state, flags, fl_gh); in do_flock() 1521 gfs2_glock_put(gl); in do_flock()
|
| H A D | rgrp.h | 34 int gfs2_rgrp_go_instantiate(struct gfs2_glock *gl);
|
| H A D | recovery.c | 36 struct gfs2_glock *gl = ip->i_gl; in gfs2_replay_read_block() local 50 *bh = gfs2_meta_ra(gl, dblock, extlen); in gfs2_replay_read_block()
|
| H A D | bmap.c | 294 static void gfs2_metapath_ra(struct gfs2_glock *gl, __be64 *start, __be64 *end) in gfs2_metapath_ra() argument 304 rabh = gfs2_getbuf(gl, be64_to_cpu(*t), CREATE); in gfs2_metapath_ra() 609 struct gfs2_glock *gl, unsigned int i, in gfs2_indirect_init() argument 617 mp->mp_bh[i] = gfs2_meta_new(gl, bn); in gfs2_indirect_init() 618 gfs2_trans_add_meta(gl, mp->mp_bh[i]); in gfs2_indirect_init()
|
| H A D | dir.c | 1484 struct gfs2_glock *gl = ip->i_gl; in gfs2_dir_readahead() local 1504 bh = gfs2_getbuf(gl, blocknr, 1); in gfs2_dir_readahead()
|
| /linux/drivers/target/iscsi/cxgbit/ |
| H A D | cxgbit_main.c | 219 cxgbit_copy_frags(struct sk_buff *skb, const struct pkt_gl *gl, in cxgbit_copy_frags() argument 226 __skb_fill_page_desc(skb, skb_frag_idx, gl->frags[0].page, in cxgbit_copy_frags() 227 gl->frags[0].offset + offset, in cxgbit_copy_frags() 228 gl->frags[0].size - offset); in cxgbit_copy_frags() 229 for (i = 1; i < gl->nfrags; i++) in cxgbit_copy_frags() 231 gl->frags[i].page, in cxgbit_copy_frags() 232 gl->frags[i].offset, in cxgbit_copy_frags() 233 gl->frags[i].size); in cxgbit_copy_frags() 235 skb_shinfo(skb)->nr_frags += gl->nfrags; in cxgbit_copy_frags() 238 get_page(gl->frags[gl->nfrags - 1].page); in cxgbit_copy_frags() [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb4vf/ |
| H A D | sge.c | 1470 const struct pkt_gl *gl, in copy_frags() argument 1476 __skb_fill_page_desc(skb, 0, gl->frags[0].page, in copy_frags() 1477 gl->frags[0].offset + offset, in copy_frags() 1478 gl->frags[0].size - offset); in copy_frags() 1479 skb_shinfo(skb)->nr_frags = gl->nfrags; in copy_frags() 1480 for (i = 1; i < gl->nfrags; i++) in copy_frags() 1481 __skb_fill_page_desc(skb, i, gl->frags[i].page, in copy_frags() 1482 gl->frags[i].offset, in copy_frags() 1483 gl->frags[i].size); in copy_frags() 1486 get_page(gl->frags[gl->nfrags - 1].page); in copy_frags() [all …]
|
| H A D | cxgb4vf_main.c | 488 const struct pkt_gl *gl) in fwevtq_handler() argument
|
| /linux/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
| H A D | chtls_main.c | 339 static struct sk_buff *copy_gl_to_skb_pkt(const struct pkt_gl *gl, in copy_gl_to_skb_pkt() argument 349 skb = alloc_skb(size_add(gl->tot_len, in copy_gl_to_skb_pkt() 354 __skb_put(skb, gl->tot_len + sizeof(struct cpl_pass_accept_req) in copy_gl_to_skb_pkt() 359 , gl->va + pktshift, in copy_gl_to_skb_pkt() 360 gl->tot_len - pktshift); in copy_gl_to_skb_pkt() 366 const struct pkt_gl *gl, const __be64 *rsp) in chtls_recv_packet() argument 372 skb = copy_gl_to_skb_pkt(gl, rsp, cdev->lldi->sge_pktshift); in chtls_recv_packet() 440 const struct pkt_gl *gl) in chtls_uld_rx_handler() argument 449 if (chtls_recv_packet(cdev, gl, rsp) < 0) in chtls_uld_rx_handler() 454 if (!gl) in chtls_uld_rx_handler() [all …]
|
| /linux/drivers/infiniband/hw/cxgb4/ |
| H A D | device.c | 1102 static inline struct sk_buff *copy_gl_to_skb_pkt(const struct pkt_gl *gl, in copy_gl_to_skb_pkt() argument 1115 skb = alloc_skb(size_add(gl->tot_len, in copy_gl_to_skb_pkt() 1122 __skb_put(skb, gl->tot_len + sizeof(struct cpl_pass_accept_req) + in copy_gl_to_skb_pkt() 1137 gl->va + pktshift, in copy_gl_to_skb_pkt() 1138 gl->tot_len - pktshift); in copy_gl_to_skb_pkt() 1142 static inline int recv_rx_pkt(struct c4iw_dev *dev, const struct pkt_gl *gl, in recv_rx_pkt() argument 1151 skb = copy_gl_to_skb_pkt(gl , rsp, dev->rdev.lldi.sge_pktshift); in recv_rx_pkt() 1167 const struct pkt_gl *gl) in c4iw_uld_rx_handler() argument 1174 if (gl == NULL) { in c4iw_uld_rx_handler() 1183 } else if (gl == CXGB4_MSG_AN) { in c4iw_uld_rx_handler() [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb4/ |
| H A D | sge.c | 3306 const struct pkt_gl *gl, unsigned int offset) in copy_frags() argument 3311 __skb_fill_page_desc(skb, 0, gl->frags[0].page, in copy_frags() 3312 gl->frags[0].offset + offset, in copy_frags() 3313 gl->frags[0].size - offset); in copy_frags() 3314 skb_shinfo(skb)->nr_frags = gl->nfrags; in copy_frags() 3315 for (i = 1; i < gl->nfrags; i++) in copy_frags() 3316 __skb_fill_page_desc(skb, i, gl->frags[i].page, in copy_frags() 3317 gl->frags[i].offset, in copy_frags() 3318 gl->frags[i].size); in copy_frags() 3321 get_page(gl->frags[gl->nfrags - 1].page); in copy_frags() [all …]
|
| H A D | cxgb4_uld.h | 483 const struct pkt_gl *gl); 487 const struct pkt_gl *gl, 521 struct sk_buff *cxgb4_pktgl_to_skb(const struct pkt_gl *gl,
|
| H A D | cxgb4_uld.c | 75 const struct pkt_gl *gl) in uldrx_handler() argument 88 rsp, gl, &q->lro_mgr, in uldrx_handler() 92 rsp, gl); in uldrx_handler() 99 if (!gl) in uldrx_handler() 101 else if (gl == CXGB4_MSG_AN) in uldrx_handler()
|
| /linux/drivers/scsi/qedi/ |
| H A D | qedi_main.c | 1518 struct global_queue **gl = qedi->global_queues; in qedi_free_global_queues() local 1521 if (!gl[i]) in qedi_free_global_queues() 1524 if (gl[i]->cq) in qedi_free_global_queues() 1525 dma_free_coherent(&qedi->pdev->dev, gl[i]->cq_mem_size, in qedi_free_global_queues() 1526 gl[i]->cq, gl[i]->cq_dma); in qedi_free_global_queues() 1527 if (gl[i]->cq_pbl) in qedi_free_global_queues() 1528 dma_free_coherent(&qedi->pdev->dev, gl[i]->cq_pbl_size, in qedi_free_global_queues() 1529 gl[i]->cq_pbl, gl[i]->cq_pbl_dma); in qedi_free_global_queues() 1531 kfree(gl[i]); in qedi_free_global_queues()
|
| /linux/tools/testing/cxl/test/ |
| H A D | mem.c | 577 struct cxl_mbox_get_log *gl = cmd->payload_in; in mock_get_log() local 578 u32 offset = le32_to_cpu(gl->offset); in mock_get_log() 579 u32 length = le32_to_cpu(gl->length); in mock_get_log() 583 if (cmd->size_in < sizeof(*gl)) in mock_get_log() 589 if (!uuid_equal(&gl->uuid, &uuid)) in mock_get_log()
|
| /linux/drivers/scsi/qedf/ |
| H A D | qedf_main.c | 2946 struct global_queue **gl = qedf->global_queues; in qedf_free_global_queues() local 2949 if (!gl[i]) in qedf_free_global_queues() 2952 if (gl[i]->cq) in qedf_free_global_queues() 2954 gl[i]->cq_mem_size, gl[i]->cq, gl[i]->cq_dma); in qedf_free_global_queues() 2955 if (gl[i]->cq_pbl) in qedf_free_global_queues() 2956 dma_free_coherent(&qedf->pdev->dev, gl[i]->cq_pbl_size, in qedf_free_global_queues() 2957 gl[i]->cq_pbl, gl[i]->cq_pbl_dma); in qedf_free_global_queues() 2959 kfree(gl[i]); in qedf_free_global_queues()
|
| /linux/tools/perf/util/ |
| H A D | probe-finder.c | 1033 static int pubname_search_cb(Dwarf *dbg, Dwarf_Global *gl, void *data) in pubname_search_cb() argument 1038 if (dwarf_offdie(dbg, gl->die_offset, param->sp_die)) { in pubname_search_cb() 1043 if (!dwarf_offdie(dbg, gl->cu_offset, param->cu_die)) in pubname_search_cb()
|
| /linux/Documentation/userspace-api/media/v4l/ |
| H A D | v4l2.rst | 247 :revision: 0.28 / 2009-08-26 (*gl*)
|
| /linux/drivers/net/ethernet/fungible/funeth/ |
| H A D | funeth_main.c | 186 struct fun_dataop_gl gl; in fun_config_rss() member 208 fun_dataop_gl_init(&cmd.gl, 0, 0, len, fp->rss_dma_addr); in fun_config_rss()
|