Home
last modified time | relevance | path

Searched refs:dreq (Results 1 – 25 of 25) sorted by relevance

/linux/fs/nfs/
H A Ddirect.c72 static void nfs_direct_write_complete(struct nfs_direct_req *dreq);
75 static inline void get_dreq(struct nfs_direct_req *dreq) in get_dreq() argument
77 atomic_inc(&dreq->io_count); in get_dreq()
80 static inline int put_dreq(struct nfs_direct_req *dreq) in put_dreq() argument
82 return atomic_dec_and_test(&dreq->io_count); in put_dreq()
86 nfs_direct_handle_truncated(struct nfs_direct_req *dreq, in nfs_direct_handle_truncated() argument
93 if (dreq->max_count >= dreq_len) { in nfs_direct_handle_truncated()
94 dreq->max_count = dreq_len; in nfs_direct_handle_truncated()
95 if (dreq->count > dreq_len) in nfs_direct_handle_truncated()
96 dreq->count = dreq_len; in nfs_direct_handle_truncated()
[all …]
H A Dcache_lib.c68 void nfs_cache_defer_req_put(struct nfs_cache_defer_req *dreq) in nfs_cache_defer_req_put() argument
70 if (refcount_dec_and_test(&dreq->count)) in nfs_cache_defer_req_put()
71 kfree(dreq); in nfs_cache_defer_req_put()
76 struct nfs_cache_defer_req *dreq; in nfs_dns_cache_revisit() local
78 dreq = container_of(d, struct nfs_cache_defer_req, deferred_req); in nfs_dns_cache_revisit()
80 complete(&dreq->completion); in nfs_dns_cache_revisit()
81 nfs_cache_defer_req_put(dreq); in nfs_dns_cache_revisit()
86 struct nfs_cache_defer_req *dreq; in nfs_dns_cache_defer() local
88 dreq = container_of(req, struct nfs_cache_defer_req, req); in nfs_dns_cache_defer()
89 dreq->deferred_req.revisit = nfs_dns_cache_revisit; in nfs_dns_cache_defer()
[all …]
H A Ddns_resolve.c283 struct nfs_cache_defer_req *dreq) in do_cache_lookup() argument
289 ret = cache_check(cd, &(*item)->h, &dreq->req); in do_cache_lookup()
325 struct nfs_cache_defer_req *dreq; in do_cache_lookup_wait() local
328 dreq = nfs_cache_defer_req_alloc(); in do_cache_lookup_wait()
329 if (!dreq) in do_cache_lookup_wait()
331 ret = do_cache_lookup(cd, key, item, dreq); in do_cache_lookup_wait()
333 ret = nfs_cache_wait_for_upcall(dreq); in do_cache_lookup_wait()
337 nfs_cache_defer_req_put(dreq); in do_cache_lookup_wait()
H A Dcache_lib.h24 extern void nfs_cache_defer_req_put(struct nfs_cache_defer_req *dreq);
25 extern int nfs_cache_wait_for_upcall(struct nfs_cache_defer_req *dreq);
H A Dwrite.c842 cinfo->dreq = NULL; in nfs_init_cinfo_from_inode()
848 struct nfs_direct_req *dreq) in nfs_init_cinfo() argument
850 if (dreq) in nfs_init_cinfo()
851 nfs_init_cinfo_from_dreq(cinfo, dreq); in nfs_init_cinfo()
972 if ((ret == max) && !cinfo->dreq) in nfs_scan_commit_list()
1713 data->dreq = cinfo->dreq; in nfs_init_commit()
1847 nfs_init_cinfo(&cinfo, data->inode, data->dreq); in nfs_commit_release_pages()
H A Dnfstrace.h1683 const struct nfs_direct_req *dreq
1686 TP_ARGS(dreq),
1699 const struct inode *inode = dreq->inode;
1706 __entry->offset = dreq->io_start;
1707 __entry->count = dreq->count;
1708 __entry->error = dreq->error;
1709 __entry->flags = dreq->flags;
1727 const struct nfs_direct_req *dreq \
1729 TP_ARGS(dreq))
H A Dpagelist.c108 hdr->dreq = desc->pg_dreq; in nfs_pgheader_init()
1382 desc->pg_dreq = hdr->dreq; in nfs_pageio_resend()
H A Dnfs4proc.c5767 if (hdr->ds_clp != NULL || hdr->dreq != NULL) in nfs4_write_need_cache_consistency_data()
/linux/net/dccp/
H A Dminisocks.c
H A Doptions.c
H A Doutput.c
H A Dipv4.c
H A Dipv6.c
H A Dfeat.c
H A Ddccp.h
/linux/net/sunrpc/
H A Dcache.c598 static void __unhash_deferred_req(struct cache_deferred_req *dreq) in __unhash_deferred_req() argument
600 hlist_del_init(&dreq->hash); in __unhash_deferred_req()
601 if (!list_empty(&dreq->recent)) { in __unhash_deferred_req()
602 list_del_init(&dreq->recent); in __unhash_deferred_req()
607 static void __hash_deferred_req(struct cache_deferred_req *dreq, struct cache_head *item) in __hash_deferred_req() argument
611 INIT_LIST_HEAD(&dreq->recent); in __hash_deferred_req()
612 hlist_add_head(&dreq->hash, &cache_defer_hash[hash]); in __hash_deferred_req()
615 static void setup_deferral(struct cache_deferred_req *dreq, in setup_deferral() argument
620 dreq->item = item; in setup_deferral()
624 __hash_deferred_req(dreq, item); in setup_deferral()
[all …]
H A Dsvc_xprt.c1185 static void svc_revisit(struct cache_deferred_req *dreq, int too_many) in svc_defer()
1188 container_of(dreq, struct svc_deferred_req, handle); in svc_defer()
1152 svc_revisit(struct cache_deferred_req * dreq,int too_many) svc_revisit() argument
/linux/drivers/s390/block/
H A Ddasd_diag.c168 struct dasd_diag_req *dreq; in dasd_start_diag() local
179 dreq = cqr->data; in dasd_start_diag()
184 private->iob.block_count = dreq->block_count; in dasd_start_diag()
186 private->iob.bio_list = dreq->bio; in dasd_start_diag()
513 struct dasd_diag_req *dreq; in dasd_diag_build_cp() local
546 cqr = dasd_smalloc_request(DASD_DIAG_MAGIC, 0, struct_size(dreq, bio, count), in dasd_diag_build_cp()
551 dreq = (struct dasd_diag_req *) cqr->data; in dasd_diag_build_cp()
552 dreq->block_count = count; in dasd_diag_build_cp()
553 dbio = dreq->bio; in dasd_diag_build_cp()
/linux/drivers/dma/
H A Dbcm2835-dma.c72 unsigned int dreq; member
642 if (c->dreq != 0) in bcm2835_dma_prep_slave_sg()
643 info |= BCM2835_DMA_PER_MAP(c->dreq); in bcm2835_dma_prep_slave_sg()
716 if (c->dreq != 0) in bcm2835_dma_prep_dma_cyclic()
717 info |= BCM2835_DMA_PER_MAP(c->dreq); in bcm2835_dma_prep_dma_cyclic()
856 to_bcm2835_dma_chan(chan)->dreq = spec->args[0]; in bcm2835_dma_xlate()
/linux/fs/nilfs2/
H A Dbtree.c1743 union nilfs_bmap_ptr_req *dreq, in nilfs_btree_prepare_convert_and_insert() argument
1757 dreq->bpr_ptr = nilfs_btree_find_target_v(btree, NULL, key); in nilfs_btree_prepare_convert_and_insert()
1765 ret = nilfs_bmap_prepare_alloc_ptr(btree, dreq, dat); in nilfs_btree_prepare_convert_and_insert()
1772 nreq->bpr_ptr = dreq->bpr_ptr + 1; in nilfs_btree_prepare_convert_and_insert()
1792 nilfs_bmap_abort_alloc_ptr(btree, dreq, dat); in nilfs_btree_prepare_convert_and_insert()
1803 union nilfs_bmap_ptr_req *dreq, in nilfs_btree_commit_convert_and_insert() argument
1823 nilfs_bmap_commit_alloc_ptr(btree, dreq, dat); in nilfs_btree_commit_convert_and_insert()
1830 nilfs_btree_node_insert(node, n, key, dreq->bpr_ptr, ncblk); in nilfs_btree_commit_convert_and_insert()
1845 nilfs_bmap_commit_alloc_ptr(btree, dreq, dat); in nilfs_btree_commit_convert_and_insert()
1852 nilfs_btree_node_insert(node, n, key, dreq->bpr_ptr, in nilfs_btree_commit_convert_and_insert()
[all …]
/linux/Documentation/devicetree/bindings/dma/
H A Dst_fdma.txt51 -bit 2-0: Holdoff value, dreq will be masked for
/linux/drivers/infiniband/core/
H A Dcm_trace.h183 DEFINE_CM_SEND_EVENT(dreq);
H A Dcm.c4296 CM_COUNTER_ATTR(dreq, _group, CM_DREQ_COUNTER), \
/linux/drivers/net/ethernet/marvell/octeontx2/af/
H A Drvu_nix.c6331 struct nix_mcast_grp_destroy_req dreq = { 0 }; in rvu_nix_mcast_flr_free_entries() local
6357 dreq.hdr.pcifunc = elem->pcifunc; in rvu_nix_mcast_flr_free_entries()
6358 dreq.mcast_grp_idx = elem->mcast_grp_idx; in rvu_nix_mcast_flr_free_entries()
6359 dreq.is_af = 1; in rvu_nix_mcast_flr_free_entries()
6360 rvu_mbox_handler_nix_mcast_grp_destroy(rvu, &dreq, NULL); in rvu_nix_mcast_flr_free_entries()
6511 struct nix_mcast_grp_destroy_req dreq = { 0 }; in rvu_mbox_handler_nix_mcast_grp_update() local
6549 dreq.hdr.pcifunc = elem->pcifunc; in rvu_mbox_handler_nix_mcast_grp_update()
6550 dreq.mcast_grp_idx = elem->mcast_grp_idx; in rvu_mbox_handler_nix_mcast_grp_update()
6551 dreq.is_af = 1; in rvu_mbox_handler_nix_mcast_grp_update()
6552 rvu_mbox_handler_nix_mcast_grp_destroy(rvu, &dreq, NULL); in rvu_mbox_handler_nix_mcast_grp_update()
/linux/fs/nfs/blocklayout/
H A Dblocklayout.c243 const bool is_dio = (header->dreq != NULL); in bl_read_pagelist()