/linux/fs/nfs/ |
H A D | direct.c | 72 static void nfs_direct_write_complete(struct nfs_direct_req *dreq); 75 static inline void get_dreq(struct nfs_direct_req *dreq) in get_dreq() argument 77 atomic_inc(&dreq->io_count); in get_dreq() 80 static inline int put_dreq(struct nfs_direct_req *dreq) in put_dreq() argument 82 return atomic_dec_and_test(&dreq->io_count); in put_dreq() 86 nfs_direct_handle_truncated(struct nfs_direct_req *dreq, in nfs_direct_handle_truncated() argument 93 if (dreq->max_count >= dreq_len) { in nfs_direct_handle_truncated() 94 dreq->max_count = dreq_len; in nfs_direct_handle_truncated() 95 if (dreq->count > dreq_len) in nfs_direct_handle_truncated() 96 dreq->count = dreq_len; in nfs_direct_handle_truncated() [all …]
|
H A D | cache_lib.c | 68 void nfs_cache_defer_req_put(struct nfs_cache_defer_req *dreq) in nfs_cache_defer_req_put() argument 70 if (refcount_dec_and_test(&dreq->count)) in nfs_cache_defer_req_put() 71 kfree(dreq); in nfs_cache_defer_req_put() 76 struct nfs_cache_defer_req *dreq; in nfs_dns_cache_revisit() local 78 dreq = container_of(d, struct nfs_cache_defer_req, deferred_req); in nfs_dns_cache_revisit() 80 complete(&dreq->completion); in nfs_dns_cache_revisit() 81 nfs_cache_defer_req_put(dreq); in nfs_dns_cache_revisit() 86 struct nfs_cache_defer_req *dreq; in nfs_dns_cache_defer() local 88 dreq = container_of(req, struct nfs_cache_defer_req, req); in nfs_dns_cache_defer() 89 dreq->deferred_req.revisit = nfs_dns_cache_revisit; in nfs_dns_cache_defer() [all …]
|
H A D | dns_resolve.c | 283 struct nfs_cache_defer_req *dreq) in do_cache_lookup() argument 289 ret = cache_check(cd, &(*item)->h, &dreq->req); in do_cache_lookup() 325 struct nfs_cache_defer_req *dreq; in do_cache_lookup_wait() local 328 dreq = nfs_cache_defer_req_alloc(); in do_cache_lookup_wait() 329 if (!dreq) in do_cache_lookup_wait() 331 ret = do_cache_lookup(cd, key, item, dreq); in do_cache_lookup_wait() 333 ret = nfs_cache_wait_for_upcall(dreq); in do_cache_lookup_wait() 337 nfs_cache_defer_req_put(dreq); in do_cache_lookup_wait()
|
H A D | cache_lib.h | 24 extern void nfs_cache_defer_req_put(struct nfs_cache_defer_req *dreq); 25 extern int nfs_cache_wait_for_upcall(struct nfs_cache_defer_req *dreq);
|
H A D | nfstrace.h | 1541 const struct nfs_direct_req *dreq 1544 TP_ARGS(dreq), 1557 const struct inode *inode = dreq->inode; 1564 __entry->offset = dreq->io_start; 1565 __entry->count = dreq->count; 1566 __entry->error = dreq->error; 1567 __entry->flags = dreq->flags; 1585 const struct nfs_direct_req *dreq \ 1587 TP_ARGS(dreq))
|
H A D | pagelist.c | 108 hdr->dreq = desc->pg_dreq; in nfs_pgheader_init() 1380 desc->pg_dreq = hdr->dreq; in nfs_pageio_resend()
|
H A D | nfs4proc.c | 5709 if (hdr->ds_clp != NULL || hdr->dreq != NULL) in nfs4_write_need_cache_consistency_data()
|
/linux/net/dccp/ |
H A D | minisocks.c | 85 struct dccp_request_sock *dreq = dccp_rsk(req); in dccp_create_openreq_child() local 94 newdp->dccps_service = dreq->dreq_service; in dccp_create_openreq_child() 95 newdp->dccps_timestamp_echo = dreq->dreq_timestamp_echo; in dccp_create_openreq_child() 96 newdp->dccps_timestamp_time = dreq->dreq_timestamp_time; in dccp_create_openreq_child() 111 newdp->dccps_iss = dreq->dreq_iss; in dccp_create_openreq_child() 112 newdp->dccps_gss = dreq->dreq_gss; in dccp_create_openreq_child() 114 newdp->dccps_isr = dreq->dreq_isr; in dccp_create_openreq_child() 115 newdp->dccps_gsr = dreq->dreq_gsr; in dccp_create_openreq_child() 120 if (dccp_feat_activate_values(newsk, &dreq->dreq_featneg)) { in dccp_create_openreq_child() 141 struct dccp_request_sock *dreq = dccp_rsk(req); in dccp_check_req() local [all …]
|
H A D | options.c | 48 int dccp_parse_options(struct sock *sk, struct dccp_request_sock *dreq, in dccp_parse_options() argument 99 if (dreq != NULL && (opt >= DCCPO_MIN_RX_CCID_SPECIFIC || in dccp_parse_options() 125 rc = dccp_feat_parse_options(sk, dreq, mandatory, opt, in dccp_parse_options() 144 if (dreq != NULL) { in dccp_parse_options() 145 dreq->dreq_timestamp_echo = ntohl(opt_val); in dccp_parse_options() 146 dreq->dreq_timestamp_time = dccp_timestamp(); in dccp_parse_options() 353 struct dccp_request_sock *dreq, in dccp_insert_option_timestamp_echo() argument 360 if (dreq != NULL) { in dccp_insert_option_timestamp_echo() 361 elapsed_time = dccp_timestamp() - dreq->dreq_timestamp_time; in dccp_insert_option_timestamp_echo() 362 tstamp_echo = htonl(dreq->dreq_timestamp_echo); in dccp_insert_option_timestamp_echo() [all …]
|
H A D | output.c | 403 struct dccp_request_sock *dreq; in dccp_make_response() local 422 dreq = dccp_rsk(req); in dccp_make_response() 424 dccp_inc_seqno(&dreq->dreq_gss); in dccp_make_response() 426 DCCP_SKB_CB(skb)->dccpd_seq = dreq->dreq_gss; in dccp_make_response() 429 if (dccp_feat_server_ccid_dependencies(dreq)) in dccp_make_response() 432 if (dccp_insert_options_rsk(dreq, skb)) in dccp_make_response() 444 dccp_hdr_set_seq(dh, dreq->dreq_gss); in dccp_make_response() 445 dccp_hdr_set_ack(dccp_hdr_ack_bits(skb), dreq->dreq_gsr); in dccp_make_response() 446 dccp_hdr_response(skb)->dccph_resp_service = dreq->dreq_service; in dccp_make_response()
|
H A D | ipv6.c | 323 struct dccp_request_sock *dreq; in dccp_v6_conn_request() local 361 dreq = dccp_rsk(req); in dccp_v6_conn_request() 362 if (dccp_parse_options(sk, dreq, skb)) in dccp_v6_conn_request() 394 dreq->dreq_isr = dcb->dccpd_seq; in dccp_v6_conn_request() 395 dreq->dreq_gsr = dreq->dreq_isr; in dccp_v6_conn_request() 396 dreq->dreq_iss = dccp_v6_init_sequence(skb); in dccp_v6_conn_request() 397 dreq->dreq_gss = dreq->dreq_iss; in dccp_v6_conn_request() 398 dreq->dreq_service = service; in dccp_v6_conn_request()
|
H A D | feat.c | 633 int dccp_feat_insert_opts(struct dccp_sock *dp, struct dccp_request_sock *dreq, in dccp_feat_insert_opts() argument 636 struct list_head *fn = dreq ? &dreq->dreq_featneg : &dp->dccps_featneg; in dccp_feat_insert_opts() 1004 int dccp_feat_server_ccid_dependencies(struct dccp_request_sock *dreq) in dccp_feat_server_ccid_dependencies() argument 1006 struct list_head *fn = &dreq->dreq_featneg; in dccp_feat_server_ccid_dependencies() 1408 int dccp_feat_parse_options(struct sock *sk, struct dccp_request_sock *dreq, in dccp_feat_parse_options() argument 1412 struct list_head *fn = dreq ? &dreq->dreq_featneg : &dp->dccps_featneg; in dccp_feat_parse_options()
|
H A D | dccp.h | 456 int dccp_feat_server_ccid_dependencies(struct dccp_request_sock *dreq);
|
/linux/drivers/crypto/marvell/cesa/ |
H A D | tdma.c | 37 void mv_cesa_dma_step(struct mv_cesa_req *dreq) in mv_cesa_dma_step() argument 39 struct mv_cesa_engine *engine = dreq->engine; in mv_cesa_dma_step() 51 writel_relaxed(dreq->chain.first->cur_dma, in mv_cesa_dma_step() 58 void mv_cesa_dma_cleanup(struct mv_cesa_req *dreq) in mv_cesa_dma_cleanup() argument 62 for (tdma = dreq->chain.first; tdma;) { in mv_cesa_dma_cleanup() 75 dreq->chain.first = NULL; in mv_cesa_dma_cleanup() 76 dreq->chain.last = NULL; in mv_cesa_dma_cleanup() 79 void mv_cesa_dma_prepare(struct mv_cesa_req *dreq, in mv_cesa_dma_prepare() argument 84 for (tdma = dreq->chain.first; tdma; tdma = tdma->next) { in mv_cesa_dma_prepare() 97 struct mv_cesa_req *dreq) in mv_cesa_tdma_chain() argument [all …]
|
H A D | cesa.h | 826 void mv_cesa_dma_step(struct mv_cesa_req *dreq); 828 static inline int mv_cesa_dma_process(struct mv_cesa_req *dreq, in mv_cesa_dma_process() argument 840 void mv_cesa_dma_prepare(struct mv_cesa_req *dreq, 842 void mv_cesa_dma_cleanup(struct mv_cesa_req *dreq); 844 struct mv_cesa_req *dreq);
|
/linux/net/sunrpc/ |
H A D | cache.c | 596 static void __unhash_deferred_req(struct cache_deferred_req *dreq) in __unhash_deferred_req() argument 598 hlist_del_init(&dreq->hash); in __unhash_deferred_req() 599 if (!list_empty(&dreq->recent)) { in __unhash_deferred_req() 600 list_del_init(&dreq->recent); in __unhash_deferred_req() 605 static void __hash_deferred_req(struct cache_deferred_req *dreq, struct cache_head *item) in __hash_deferred_req() argument 609 INIT_LIST_HEAD(&dreq->recent); in __hash_deferred_req() 610 hlist_add_head(&dreq->hash, &cache_defer_hash[hash]); in __hash_deferred_req() 613 static void setup_deferral(struct cache_deferred_req *dreq, in setup_deferral() argument 618 dreq->item = item; in setup_deferral() 622 __hash_deferred_req(dreq, item); in setup_deferral() [all …]
|
H A D | svc_xprt.c | 1141 static void svc_revisit(struct cache_deferred_req *dreq, int too_many) in svc_xprt_destroy_all() 1144 container_of(dreq, struct svc_deferred_req, handle); 1148 svc_revisit(struct cache_deferred_req * dreq,int too_many) svc_revisit() argument
|
/linux/fs/nilfs2/ |
H A D | btree.c | 1743 union nilfs_bmap_ptr_req *dreq, in nilfs_btree_prepare_convert_and_insert() 1757 dreq->bpr_ptr = nilfs_btree_find_target_v(btree, NULL, key); in nilfs_btree_prepare_convert_and_insert() 1765 ret = nilfs_bmap_prepare_alloc_ptr(btree, dreq, dat); in nilfs_btree_prepare_convert_and_insert() 1772 nreq->bpr_ptr = dreq->bpr_ptr + 1; in nilfs_btree_prepare_convert_and_insert() 1792 nilfs_bmap_abort_alloc_ptr(btree, dreq, dat); in nilfs_btree_prepare_convert_and_insert() 1803 union nilfs_bmap_ptr_req *dreq, in nilfs_btree_commit_convert_and_insert() 1823 nilfs_bmap_commit_alloc_ptr(btree, dreq, dat); in nilfs_btree_commit_convert_and_insert() 1830 nilfs_btree_node_insert(node, n, key, dreq->bpr_ptr, ncblk); in nilfs_btree_commit_convert_and_insert() 1845 nilfs_bmap_commit_alloc_ptr(btree, dreq, dat); in nilfs_btree_commit_convert_and_insert() 1852 nilfs_btree_node_insert(node, n, key, dreq in nilfs_btree_commit_convert_and_insert() 1744 nilfs_btree_prepare_convert_and_insert(struct nilfs_bmap * btree,__u64 key,union nilfs_bmap_ptr_req * dreq,union nilfs_bmap_ptr_req * nreq,struct buffer_head ** bhp,struct nilfs_bmap_stats * stats) nilfs_btree_prepare_convert_and_insert() argument 1804 nilfs_btree_commit_convert_and_insert(struct nilfs_bmap * btree,__u64 key,__u64 ptr,const __u64 * keys,const __u64 * ptrs,int n,union nilfs_bmap_ptr_req * dreq,union nilfs_bmap_ptr_req * nreq,struct buffer_head * bh) nilfs_btree_commit_convert_and_insert() argument 1886 union nilfs_bmap_ptr_req dreq, nreq, *di, *ni; nilfs_btree_convert_and_insert() local [all...] |
/linux/Documentation/devicetree/bindings/dma/ |
H A D | st_fdma.txt | 51 -bit 2-0: Holdoff value, dreq will be masked for
|
/linux/include/linux/ |
H A D | dccp.h | 180 extern int dccp_parse_options(struct sock *sk, struct dccp_request_sock *dreq,
|
H A D | nfs_xdr.h | 1653 struct nfs_direct_req *dreq; member 1697 struct nfs_direct_req *dreq; /* O_DIRECT request */ member 1709 struct nfs_direct_req *dreq; /* O_DIRECT request */ member
|
/linux/drivers/infiniband/core/ |
H A D | cm_trace.h | 183 DEFINE_CM_SEND_EVENT(dreq);
|
/linux/drivers/net/ethernet/marvell/octeontx2/af/ |
H A D | rvu_nix.c | 6303 struct nix_mcast_grp_destroy_req dreq = { 0 }; in rvu_nix_mcast_flr_free_entries() local 6329 dreq.hdr.pcifunc = elem->pcifunc; in rvu_nix_mcast_flr_free_entries() 6330 dreq.mcast_grp_idx = elem->mcast_grp_idx; in rvu_nix_mcast_flr_free_entries() 6331 dreq.is_af = 1; in rvu_nix_mcast_flr_free_entries() 6332 rvu_mbox_handler_nix_mcast_grp_destroy(rvu, &dreq, NULL); in rvu_nix_mcast_flr_free_entries() 6483 struct nix_mcast_grp_destroy_req dreq = { 0 }; in rvu_mbox_handler_nix_mcast_grp_update() local 6521 dreq.hdr.pcifunc = elem->pcifunc; in rvu_mbox_handler_nix_mcast_grp_update() 6522 dreq.mcast_grp_idx = elem->mcast_grp_idx; in rvu_mbox_handler_nix_mcast_grp_update() 6523 dreq.is_af = 1; in rvu_mbox_handler_nix_mcast_grp_update() 6524 rvu_mbox_handler_nix_mcast_grp_destroy(rvu, &dreq, NULL); in rvu_mbox_handler_nix_mcast_grp_update()
|
/linux/fs/nfs/blocklayout/ |
H A D | blocklayout.c | 243 const bool is_dio = (header->dreq != NULL); in bl_read_pagelist()
|