Home
last modified time | relevance | path

Searched refs:gl (Results 1 – 25 of 27) sorted by relevance

12

/linux/fs/gfs2/
H A Dglock.c57 struct gfs2_glock *gl; /* current glock struct */ member
61 typedef void (*glock_examiner) (struct gfs2_glock * gl);
63 static void do_xmote(struct gfs2_glock *gl, struct gfs2_holder *gh,
65 static void request_demote(struct gfs2_glock *gl, unsigned int state,
118 * @gl: the glock
120 static void wake_up_glock(struct gfs2_glock *gl) in wake_up_glock() argument
122 wait_queue_head_t *wq = glock_waitqueue(&gl->gl_name); in wake_up_glock()
125 __wake_up(wq, TASK_NORMAL, 1, &gl->gl_name); in wake_up_glock()
130 struct gfs2_glock *gl = container_of(rcu, struct gfs2_glock, gl_rcu); in gfs2_glock_dealloc() local
132 kfree(gl in gfs2_glock_dealloc()
141 __gfs2_glock_free(struct gfs2_glock * gl) __gfs2_glock_free() argument
149 gfs2_glock_free(struct gfs2_glock * gl) gfs2_glock_free() argument
157 gfs2_glock_free_later(struct gfs2_glock * gl) gfs2_glock_free_later() argument
172 struct gfs2_glock *gl; gfs2_free_dead_glocks() local
186 gfs2_glock_hold(struct gfs2_glock * gl) gfs2_glock_hold() argument
193 gfs2_glock_add_to_lru(struct gfs2_glock * gl) gfs2_glock_add_to_lru() argument
206 gfs2_glock_remove_from_lru(struct gfs2_glock * gl) gfs2_glock_remove_from_lru() argument
221 gfs2_glock_queue_work(struct gfs2_glock * gl,unsigned long delay) gfs2_glock_queue_work() argument
236 __gfs2_glock_put(struct gfs2_glock * gl) __gfs2_glock_put() argument
254 __gfs2_glock_put_or_lock(struct gfs2_glock * gl) __gfs2_glock_put_or_lock() argument
274 gfs2_glock_put(struct gfs2_glock * gl) gfs2_glock_put() argument
289 gfs2_glock_put_async(struct gfs2_glock * gl) gfs2_glock_put_async() argument
312 may_grant(struct gfs2_glock * gl,struct gfs2_holder * current_gh,struct gfs2_holder * gh) may_grant() argument
372 do_error(struct gfs2_glock * gl,const int ret) do_error() argument
396 find_first_holder(const struct gfs2_glock * gl) find_first_holder() argument
417 struct gfs2_glock *gl = gh->gh_gl; gfs2_instantiate() local
460 do_promote(struct gfs2_glock * gl) do_promote() argument
496 find_first_waiter(const struct gfs2_glock * gl) find_first_waiter() argument
514 find_last_waiter(const struct gfs2_glock * gl) find_last_waiter() argument
530 state_change(struct gfs2_glock * gl,unsigned int new_state) state_change() argument
540 gfs2_set_demote(int nr,struct gfs2_glock * gl) gfs2_set_demote() argument
549 gfs2_demote_wake(struct gfs2_glock * gl) gfs2_demote_wake() argument
564 finish_xmote(struct gfs2_glock * gl,unsigned int ret) finish_xmote() argument
656 do_xmote(struct gfs2_glock * gl,struct gfs2_holder * gh,unsigned int target,bool may_cancel) do_xmote() argument
755 run_queue(struct gfs2_glock * gl,const int nonblock) run_queue() argument
814 glock_set_object(struct gfs2_glock * gl,void * object) glock_set_object() argument
831 glock_clear_object(struct gfs2_glock * gl,void * object) glock_clear_object() argument
843 gfs2_inode_remember_delete(struct gfs2_glock * gl,u64 generation) gfs2_inode_remember_delete() argument
853 gfs2_inode_already_deleted(struct gfs2_glock * gl,u64 generation) gfs2_inode_already_deleted() argument
862 gfs2_glock_poke(struct gfs2_glock * gl) gfs2_glock_poke() argument
875 gfs2_grab_existing_inode(struct gfs2_glock * gl) gfs2_grab_existing_inode() argument
894 gfs2_try_to_evict(struct gfs2_glock * gl) gfs2_try_to_evict() argument
927 gfs2_queue_try_to_evict(struct gfs2_glock * gl) gfs2_queue_try_to_evict() argument
936 gfs2_queue_verify_delete(struct gfs2_glock * gl,bool later) gfs2_queue_verify_delete() argument
950 struct gfs2_glock *gl = container_of(dwork, struct gfs2_glock, gl_delete); delete_work_func() local
986 struct gfs2_glock *gl = container_of(work, struct gfs2_glock, gl_work.work); glock_work_func() local
1036 struct gfs2_glock *gl; find_insert_glock() local
1087 struct gfs2_glock *gl, *tmp; gfs2_glock_get() local
1181 __gfs2_holder_init(struct gfs2_glock * gl,unsigned int state,u16 flags,struct gfs2_holder * gh,unsigned long ip) __gfs2_holder_init() argument
1227 gfs2_glock_update_hold_time(struct gfs2_glock * gl,unsigned long start_time) gfs2_glock_update_hold_time() argument
1367 request_demote(struct gfs2_glock * gl,unsigned int state,unsigned long delay,bool remote) request_demote() argument
1402 gfs2_should_queue_trylock(struct gfs2_glock * gl,struct gfs2_holder * gh) gfs2_should_queue_trylock() argument
1439 struct gfs2_glock *gl = gh->gh_gl; add_to_queue() local
1493 struct gfs2_glock *gl = gh->gh_gl; gfs2_glock_nq() local
1552 struct gfs2_glock *gl = gh->gh_gl; __gfs2_glock_dq() local
1596 struct gfs2_glock *gl = gh->gh_gl; gfs2_glock_dq() local
1643 struct gfs2_glock *gl = gh->gh_gl; gfs2_glock_dq_wait() local
1677 struct gfs2_glock *gl; gfs2_glock_nq_num() local
1793 gfs2_glock_cb(struct gfs2_glock * gl,unsigned int state) gfs2_glock_cb() argument
1827 gfs2_should_freeze(const struct gfs2_glock * gl) gfs2_should_freeze() argument
1855 gfs2_glock_complete(struct gfs2_glock * gl,int ret) gfs2_glock_complete() argument
1893 can_free_glock(struct gfs2_glock * gl) can_free_glock() argument
1921 struct gfs2_glock *gl; gfs2_dispose_glock_lru() local
1962 struct gfs2_glock *gl, *next; gfs2_scan_glock_lru() local
2008 struct gfs2_glock *gl; glock_hash_walk() local
2027 gfs2_cancel_delete_work(struct gfs2_glock * gl) gfs2_cancel_delete_work() argument
2035 flush_delete_work(struct gfs2_glock * gl) flush_delete_work() argument
2059 thaw_glock(struct gfs2_glock * gl) thaw_glock() argument
2079 clear_glock(struct gfs2_glock * gl) clear_glock() argument
2104 dump_glock(struct seq_file * seq,struct gfs2_glock * gl,bool fsid) dump_glock() argument
2111 dump_glock_func(struct gfs2_glock * gl) dump_glock_func() argument
2116 withdraw_glock(struct gfs2_glock * gl) withdraw_glock() argument
2250 gflags2str(char * buf,const struct gfs2_glock * gl) gflags2str() argument
2317 gfs2_dump_glock(struct seq_file * seq,struct gfs2_glock * gl,bool fsid) gfs2_dump_glock() argument
2360 struct gfs2_glock *gl = iter_ptr; gfs2_glstats_seq_show() local
2463 struct gfs2_glock *gl = gi->gl; gfs2_glock_iter_next() local
2770 struct gfs2_glock *gl; gfs2_glockfd_seq_show() local
[all...]
H A Dmain.c52 struct gfs2_glock *gl = foo; in gfs2_init_glock_once() local
54 INIT_LIST_HEAD(&gl->gl_holders); in gfs2_init_glock_once()
55 INIT_LIST_HEAD(&gl->gl_lru); in gfs2_init_glock_once()
56 INIT_LIST_HEAD(&gl->gl_ail_list); in gfs2_init_glock_once()
57 atomic_set(&gl->gl_ail_count, 0); in gfs2_init_glock_once()
58 atomic_set(&gl->gl_revokes, 0); in gfs2_init_glock_once()
H A Dlops.c73 struct gfs2_glock *gl = bd->bd_gl; in maybe_release_space() local
74 struct gfs2_sbd *sdp = glock_sbd(gl); in maybe_release_space()
75 struct gfs2_rgrpd *rgd = gfs2_glock2rgrp(gl); in maybe_release_space()
76 unsigned int index = bd->bd_bh->b_blocknr - glock_number(gl); in maybe_release_space()
122 struct gfs2_glock *gl = bd->bd_gl; in gfs2_unpin() local
123 list_add(&bd->bd_ail_gl_list, &gl->gl_ail_list); in gfs2_unpin()
124 atomic_inc(&gl->gl_ail_count); in gfs2_unpin()
780 struct gfs2_glock *gl = ip->i_gl; in buf_lo_scan_elements() local
803 bh_ip = gfs2_meta_new(gl, blkno); in buf_lo_scan_elements()
890 struct gfs2_glock *gl; in gfs2_drain_revokes() local
[all …]
H A Drgrp.c720 struct gfs2_glock *gl; in gfs2_clear_rgrpd() local
724 gl = rgd->rd_gl; in gfs2_clear_rgrpd()
728 if (gl) { in gfs2_clear_rgrpd()
729 if (gl->gl_state != LM_ST_UNLOCKED) { in gfs2_clear_rgrpd()
730 gfs2_glock_cb(gl, LM_ST_UNLOCKED); in gfs2_clear_rgrpd()
731 flush_delayed_work(&gl->gl_work); in gfs2_clear_rgrpd()
734 glock_clear_object(gl, rgd); in gfs2_clear_rgrpd()
735 gfs2_glock_put(gl); in gfs2_clear_rgrpd()
1037 struct gfs2_glock *gl = ip->i_gl; in gfs2_rindex_update() local
1044 if (!gfs2_glock_is_locked_by_me(gl)) { in gfs2_rindex_update()
[all …]
H A Dglops.h25 int gfs2_inode_metasync(struct gfs2_glock *gl);
26 void gfs2_ail_flush(struct gfs2_glock *gl, bool fsync);
H A Dsuper.c1012 struct gfs2_glock *gl = ip->i_iopen_gh.gh_gl; in gfs2_drop_inode() local
1013 if (glock_needs_demote(gl)) in gfs2_drop_inode()
1025 struct gfs2_glock *gl = ip->i_iopen_gh.gh_gl; in gfs2_drop_inode() local
1027 gfs2_glock_hold(gl); in gfs2_drop_inode()
1028 if (!gfs2_queue_verify_delete(gl, true)) in gfs2_drop_inode()
1029 gfs2_glock_put_async(gl); in gfs2_drop_inode()
1161 * @gl: The glock to put
1167 static void gfs2_glock_put_eventually(struct gfs2_glock *gl) in gfs2_glock_put_eventually() argument
1170 gfs2_glock_put_async(gl); in gfs2_glock_put_eventually()
1172 gfs2_glock_put(gl); in gfs2_glock_put_eventually()
1436 struct gfs2_glock *gl = ip->i_iopen_gh.gh_gl; gfs2_evict_inode() local
[all...]
H A Dfile.c1472 struct gfs2_glock *gl = gfs2_glock_hold(fl_gh->gh_gl); in __flock_holder_uninit() local
1482 gfs2_glock_put(gl); in __flock_holder_uninit()
1490 struct gfs2_glock *gl; in do_flock() local
1515 &gfs2_flock_glops, CREATE, &gl); in do_flock()
1519 gfs2_holder_init(gl, state, flags, fl_gh); in do_flock()
1521 gfs2_glock_put(gl); in do_flock()
H A Drgrp.h34 int gfs2_rgrp_go_instantiate(struct gfs2_glock *gl);
H A Drecovery.c36 struct gfs2_glock *gl = ip->i_gl; in gfs2_replay_read_block() local
50 *bh = gfs2_meta_ra(gl, dblock, extlen); in gfs2_replay_read_block()
H A Dbmap.c294 static void gfs2_metapath_ra(struct gfs2_glock *gl, __be64 *start, __be64 *end) in gfs2_metapath_ra() argument
304 rabh = gfs2_getbuf(gl, be64_to_cpu(*t), CREATE); in gfs2_metapath_ra()
609 struct gfs2_glock *gl, unsigned int i, in gfs2_indirect_init() argument
617 mp->mp_bh[i] = gfs2_meta_new(gl, bn); in gfs2_indirect_init()
618 gfs2_trans_add_meta(gl, mp->mp_bh[i]); in gfs2_indirect_init()
H A Ddir.c1484 struct gfs2_glock *gl = ip->i_gl; in gfs2_dir_readahead() local
1504 bh = gfs2_getbuf(gl, blocknr, 1); in gfs2_dir_readahead()
/linux/drivers/target/iscsi/cxgbit/
H A Dcxgbit_main.c219 cxgbit_copy_frags(struct sk_buff *skb, const struct pkt_gl *gl, in cxgbit_copy_frags() argument
226 __skb_fill_page_desc(skb, skb_frag_idx, gl->frags[0].page, in cxgbit_copy_frags()
227 gl->frags[0].offset + offset, in cxgbit_copy_frags()
228 gl->frags[0].size - offset); in cxgbit_copy_frags()
229 for (i = 1; i < gl->nfrags; i++) in cxgbit_copy_frags()
231 gl->frags[i].page, in cxgbit_copy_frags()
232 gl->frags[i].offset, in cxgbit_copy_frags()
233 gl->frags[i].size); in cxgbit_copy_frags()
235 skb_shinfo(skb)->nr_frags += gl->nfrags; in cxgbit_copy_frags()
238 get_page(gl->frags[gl->nfrags - 1].page); in cxgbit_copy_frags()
[all …]
/linux/drivers/net/ethernet/chelsio/cxgb4vf/
H A Dsge.c1470 const struct pkt_gl *gl, in copy_frags() argument
1476 __skb_fill_page_desc(skb, 0, gl->frags[0].page, in copy_frags()
1477 gl->frags[0].offset + offset, in copy_frags()
1478 gl->frags[0].size - offset); in copy_frags()
1479 skb_shinfo(skb)->nr_frags = gl->nfrags; in copy_frags()
1480 for (i = 1; i < gl->nfrags; i++) in copy_frags()
1481 __skb_fill_page_desc(skb, i, gl->frags[i].page, in copy_frags()
1482 gl->frags[i].offset, in copy_frags()
1483 gl->frags[i].size); in copy_frags()
1486 get_page(gl->frags[gl->nfrags - 1].page); in copy_frags()
[all …]
/linux/drivers/net/ethernet/chelsio/inline_crypto/chtls/
H A Dchtls_main.c339 static struct sk_buff *copy_gl_to_skb_pkt(const struct pkt_gl *gl, in copy_gl_to_skb_pkt() argument
349 skb = alloc_skb(size_add(gl->tot_len, in copy_gl_to_skb_pkt()
354 __skb_put(skb, gl->tot_len + sizeof(struct cpl_pass_accept_req) in copy_gl_to_skb_pkt()
359 , gl->va + pktshift, in copy_gl_to_skb_pkt()
360 gl->tot_len - pktshift); in copy_gl_to_skb_pkt()
366 const struct pkt_gl *gl, const __be64 *rsp) in chtls_recv_packet() argument
372 skb = copy_gl_to_skb_pkt(gl, rsp, cdev->lldi->sge_pktshift); in chtls_recv_packet()
440 const struct pkt_gl *gl) in chtls_uld_rx_handler() argument
449 if (chtls_recv_packet(cdev, gl, rsp) < 0) in chtls_uld_rx_handler()
454 if (!gl) in chtls_uld_rx_handler()
[all …]
/linux/drivers/infiniband/hw/cxgb4/
H A Ddevice.c1102 static inline struct sk_buff *copy_gl_to_skb_pkt(const struct pkt_gl *gl, in copy_gl_to_skb_pkt() argument
1115 skb = alloc_skb(size_add(gl->tot_len, in copy_gl_to_skb_pkt()
1122 __skb_put(skb, gl->tot_len + sizeof(struct cpl_pass_accept_req) + in copy_gl_to_skb_pkt()
1137 gl->va + pktshift, in copy_gl_to_skb_pkt()
1138 gl->tot_len - pktshift); in copy_gl_to_skb_pkt()
1142 static inline int recv_rx_pkt(struct c4iw_dev *dev, const struct pkt_gl *gl, in recv_rx_pkt() argument
1151 skb = copy_gl_to_skb_pkt(gl , rsp, dev->rdev.lldi.sge_pktshift); in recv_rx_pkt()
1167 const struct pkt_gl *gl) in c4iw_uld_rx_handler() argument
1174 if (gl == NULL) { in c4iw_uld_rx_handler()
1183 } else if (gl == CXGB4_MSG_AN) { in c4iw_uld_rx_handler()
[all …]
/linux/drivers/net/ethernet/chelsio/cxgb4/
H A Dsge.c3306 const struct pkt_gl *gl, unsigned int offset) in copy_frags() argument
3311 __skb_fill_page_desc(skb, 0, gl->frags[0].page, in copy_frags()
3312 gl->frags[0].offset + offset, in copy_frags()
3313 gl->frags[0].size - offset); in copy_frags()
3314 skb_shinfo(skb)->nr_frags = gl->nfrags; in copy_frags()
3315 for (i = 1; i < gl->nfrags; i++) in copy_frags()
3316 __skb_fill_page_desc(skb, i, gl->frags[i].page, in copy_frags()
3317 gl->frags[i].offset, in copy_frags()
3318 gl->frags[i].size); in copy_frags()
3321 get_page(gl->frags[gl->nfrags - 1].page); in copy_frags()
[all …]
H A Dcxgb4_uld.h483 const struct pkt_gl *gl);
487 const struct pkt_gl *gl,
521 struct sk_buff *cxgb4_pktgl_to_skb(const struct pkt_gl *gl,
H A Dcxgb4_uld.c75 const struct pkt_gl *gl) in uldrx_handler() argument
88 rsp, gl, &q->lro_mgr, in uldrx_handler()
92 rsp, gl); in uldrx_handler()
99 if (!gl) in uldrx_handler()
101 else if (gl == CXGB4_MSG_AN) in uldrx_handler()
/linux/drivers/scsi/qedi/
H A Dqedi_main.c1518 struct global_queue **gl = qedi->global_queues; in qedi_free_global_queues() local
1521 if (!gl[i]) in qedi_free_global_queues()
1524 if (gl[i]->cq) in qedi_free_global_queues()
1525 dma_free_coherent(&qedi->pdev->dev, gl[i]->cq_mem_size, in qedi_free_global_queues()
1526 gl[i]->cq, gl[i]->cq_dma); in qedi_free_global_queues()
1527 if (gl[i]->cq_pbl) in qedi_free_global_queues()
1528 dma_free_coherent(&qedi->pdev->dev, gl[i]->cq_pbl_size, in qedi_free_global_queues()
1529 gl[i]->cq_pbl, gl[i]->cq_pbl_dma); in qedi_free_global_queues()
1531 kfree(gl[i]); in qedi_free_global_queues()
/linux/tools/testing/cxl/test/
H A Dmem.c577 struct cxl_mbox_get_log *gl = cmd->payload_in; in mock_get_log() local
578 u32 offset = le32_to_cpu(gl->offset); in mock_get_log()
579 u32 length = le32_to_cpu(gl->length); in mock_get_log()
583 if (cmd->size_in < sizeof(*gl)) in mock_get_log()
589 if (!uuid_equal(&gl->uuid, &uuid)) in mock_get_log()
/linux/arch/s390/kvm/
H A Dgaccess.c1509 int flags, i, hl, gl, l, rc; in _gaccess_do_shadow() local
1530 gl = get_level(table, ptep); in _gaccess_do_shadow()
1543 for (i = gl ; i >= w->level; i--) { in _gaccess_do_shadow()
1559 l = min3(gl, hl, w->level); in _gaccess_do_shadow()
1564 if (l < gl) { in _gaccess_do_shadow()
/linux/drivers/scsi/qedf/
H A Dqedf_main.c2946 struct global_queue **gl = qedf->global_queues; in qedf_free_global_queues() local
2949 if (!gl[i]) in qedf_free_global_queues()
2952 if (gl[i]->cq) in qedf_free_global_queues()
2954 gl[i]->cq_mem_size, gl[i]->cq, gl[i]->cq_dma); in qedf_free_global_queues()
2955 if (gl[i]->cq_pbl) in qedf_free_global_queues()
2956 dma_free_coherent(&qedf->pdev->dev, gl[i]->cq_pbl_size, in qedf_free_global_queues()
2957 gl[i]->cq_pbl, gl[i]->cq_pbl_dma); in qedf_free_global_queues()
2959 kfree(gl[i]); in qedf_free_global_queues()
/linux/tools/perf/util/
H A Dprobe-finder.c1033 static int pubname_search_cb(Dwarf *dbg, Dwarf_Global *gl, void *data) in pubname_search_cb() argument
1038 if (dwarf_offdie(dbg, gl->die_offset, param->sp_die)) { in pubname_search_cb()
1043 if (!dwarf_offdie(dbg, gl->cu_offset, param->cu_die)) in pubname_search_cb()
/linux/Documentation/userspace-api/media/v4l/
H A Dv4l2.rst247 :revision: 0.28 / 2009-08-26 (*gl*)
/linux/drivers/net/ethernet/fungible/funeth/
H A Dfuneth_main.c186 struct fun_dataop_gl gl; in fun_config_rss() member
208 fun_dataop_gl_init(&cmd.gl, 0, 0, len, fp->rss_dma_addr); in fun_config_rss()

12