| /linux/fs/nfs/ |
| H A D | direct.c | 72 static void nfs_direct_write_complete(struct nfs_direct_req *dreq); 75 static inline void get_dreq(struct nfs_direct_req *dreq) in get_dreq() argument 77 atomic_inc(&dreq->io_count); in get_dreq() 80 static inline int put_dreq(struct nfs_direct_req *dreq) in put_dreq() argument 82 return atomic_dec_and_test(&dreq->io_count); in put_dreq() 86 nfs_direct_handle_truncated(struct nfs_direct_req *dreq, in nfs_direct_handle_truncated() argument 93 if (dreq->max_count >= dreq_len) { in nfs_direct_handle_truncated() 94 dreq->max_count = dreq_len; in nfs_direct_handle_truncated() 95 if (dreq->count > dreq_len) in nfs_direct_handle_truncated() 96 dreq->count = dreq_len; in nfs_direct_handle_truncated() [all …]
|
| H A D | cache_lib.c | 68 void nfs_cache_defer_req_put(struct nfs_cache_defer_req *dreq) in nfs_cache_defer_req_put() argument 70 if (refcount_dec_and_test(&dreq->count)) in nfs_cache_defer_req_put() 71 kfree(dreq); in nfs_cache_defer_req_put() 76 struct nfs_cache_defer_req *dreq; in nfs_dns_cache_revisit() local 78 dreq = container_of(d, struct nfs_cache_defer_req, deferred_req); in nfs_dns_cache_revisit() 80 complete(&dreq->completion); in nfs_dns_cache_revisit() 81 nfs_cache_defer_req_put(dreq); in nfs_dns_cache_revisit() 86 struct nfs_cache_defer_req *dreq; in nfs_dns_cache_defer() local 88 dreq = container_of(req, struct nfs_cache_defer_req, req); in nfs_dns_cache_defer() 89 dreq->deferred_req.revisit = nfs_dns_cache_revisit; in nfs_dns_cache_defer() [all …]
|
| H A D | dns_resolve.c | 283 struct nfs_cache_defer_req *dreq) in do_cache_lookup() argument 289 ret = cache_check(cd, &(*item)->h, &dreq->req); in do_cache_lookup() 325 struct nfs_cache_defer_req *dreq; in do_cache_lookup_wait() local 328 dreq = nfs_cache_defer_req_alloc(); in do_cache_lookup_wait() 329 if (!dreq) in do_cache_lookup_wait() 331 ret = do_cache_lookup(cd, key, item, dreq); in do_cache_lookup_wait() 333 ret = nfs_cache_wait_for_upcall(dreq); in do_cache_lookup_wait() 337 nfs_cache_defer_req_put(dreq); in do_cache_lookup_wait()
|
| H A D | cache_lib.h | 24 extern void nfs_cache_defer_req_put(struct nfs_cache_defer_req *dreq); 25 extern int nfs_cache_wait_for_upcall(struct nfs_cache_defer_req *dreq);
|
| H A D | write.c | 842 cinfo->dreq = NULL; in nfs_init_cinfo_from_inode() 848 struct nfs_direct_req *dreq) in nfs_init_cinfo() argument 850 if (dreq) in nfs_init_cinfo() 851 nfs_init_cinfo_from_dreq(cinfo, dreq); in nfs_init_cinfo() 972 if ((ret == max) && !cinfo->dreq) in nfs_scan_commit_list() 1713 data->dreq = cinfo->dreq; in nfs_init_commit() 1847 nfs_init_cinfo(&cinfo, data->inode, data->dreq); in nfs_commit_release_pages()
|
| H A D | internal.h | 634 struct nfs_direct_req *dreq); 736 struct nfs_direct_req *dreq); 737 extern ssize_t nfs_dreq_bytes_left(struct nfs_direct_req *dreq, loff_t offset); 861 if (folio && !cinfo->dreq) { in nfs_folio_mark_unstable()
|
| H A D | nfstrace.h | 1680 const struct nfs_direct_req *dreq 1683 TP_ARGS(dreq), 1696 const struct inode *inode = dreq->inode; 1703 __entry->offset = dreq->io_start; 1704 __entry->count = dreq->count; 1705 __entry->error = dreq->error; 1706 __entry->flags = dreq->flags; 1724 const struct nfs_direct_req *dreq \ 1726 TP_ARGS(dreq))
|
| H A D | pagelist.c | 108 hdr->dreq = desc->pg_dreq; in nfs_pgheader_init() 1381 desc->pg_dreq = hdr->dreq; in nfs_pageio_resend()
|
| H A D | nfs4proc.c | 5835 if (hdr->ds_clp != NULL || hdr->dreq != NULL) in nfs4_write_need_cache_consistency_data()
|
| /linux/net/dccp/ |
| H A D | minisocks.c | |
| H A D | options.c | |
| H A D | output.c | |
| H A D | ipv4.c | |
| H A D | ipv6.c | |
| H A D | feat.c | |
| H A D | dccp.h | |
| /linux/net/sunrpc/ |
| H A D | cache.c | 597 static void __unhash_deferred_req(struct cache_deferred_req *dreq) in __unhash_deferred_req() argument 599 hlist_del_init(&dreq->hash); in __unhash_deferred_req() 600 if (!list_empty(&dreq->recent)) { in __unhash_deferred_req() 601 list_del_init(&dreq->recent); in __unhash_deferred_req() 606 static void __hash_deferred_req(struct cache_deferred_req *dreq, struct cache_head *item) in __hash_deferred_req() argument 610 INIT_LIST_HEAD(&dreq->recent); in __hash_deferred_req() 611 hlist_add_head(&dreq->hash, &cache_defer_hash[hash]); in __hash_deferred_req() 614 static void setup_deferral(struct cache_deferred_req *dreq, in setup_deferral() argument 619 dreq->item = item; in setup_deferral() 623 __hash_deferred_req(dreq, item); in setup_deferral() [all …]
|
| H A D | svc_xprt.c | 1152 static void svc_revisit(struct cache_deferred_req *dreq, int too_many) in svc_revisit() argument 1155 container_of(dreq, struct svc_deferred_req, handle); in svc_revisit()
|
| /linux/drivers/dma/ |
| H A D | bcm2835-dma.c | 72 unsigned int dreq; member 659 if (c->dreq != 0) in bcm2835_dma_prep_slave_sg() 660 info |= BCM2835_DMA_PER_MAP(c->dreq); in bcm2835_dma_prep_slave_sg() 733 if (c->dreq != 0) in bcm2835_dma_prep_dma_cyclic() 734 info |= BCM2835_DMA_PER_MAP(c->dreq); in bcm2835_dma_prep_dma_cyclic() 873 to_bcm2835_dma_chan(chan)->dreq = spec->args[0]; in bcm2835_dma_xlate()
|
| /linux/fs/nilfs2/ |
| H A D | btree.c | 1743 union nilfs_bmap_ptr_req *dreq, in nilfs_btree_prepare_convert_and_insert() argument 1757 dreq->bpr_ptr = nilfs_btree_find_target_v(btree, NULL, key); in nilfs_btree_prepare_convert_and_insert() 1765 ret = nilfs_bmap_prepare_alloc_ptr(btree, dreq, dat); in nilfs_btree_prepare_convert_and_insert() 1772 nreq->bpr_ptr = dreq->bpr_ptr + 1; in nilfs_btree_prepare_convert_and_insert() 1792 nilfs_bmap_abort_alloc_ptr(btree, dreq, dat); in nilfs_btree_prepare_convert_and_insert() 1803 union nilfs_bmap_ptr_req *dreq, in nilfs_btree_commit_convert_and_insert() argument 1823 nilfs_bmap_commit_alloc_ptr(btree, dreq, dat); in nilfs_btree_commit_convert_and_insert() 1830 nilfs_btree_node_insert(node, n, key, dreq->bpr_ptr, ncblk); in nilfs_btree_commit_convert_and_insert() 1845 nilfs_bmap_commit_alloc_ptr(btree, dreq, dat); in nilfs_btree_commit_convert_and_insert() 1852 nilfs_btree_node_insert(node, n, key, dreq->bpr_ptr, in nilfs_btree_commit_convert_and_insert() [all …]
|
| /linux/Documentation/devicetree/bindings/dma/ |
| H A D | st_fdma.txt | 51 -bit 2-0: Holdoff value, dreq will be masked for
|
| /linux/drivers/infiniband/core/ |
| H A D | cm_trace.h | 183 DEFINE_CM_SEND_EVENT(dreq);
|
| H A D | cm.c | 4296 CM_COUNTER_ATTR(dreq, _group, CM_DREQ_COUNTER), \
|
| /linux/include/linux/ |
| H A D | nfs_xdr.h | 1664 struct nfs_direct_req *dreq; member 1709 struct nfs_direct_req *dreq; /* O_DIRECT request */ member 1721 struct nfs_direct_req *dreq; /* O_DIRECT request */ member
|
| /linux/fs/nfs/blocklayout/ |
| H A D | blocklayout.c | 243 const bool is_dio = (header->dreq != NULL); in bl_read_pagelist()
|