| /linux/fs/netfs/ |
| H A D | direct_write.c | 15 static void netfs_unbuffered_write_done(struct netfs_io_request *wreq) in netfs_unbuffered_write_done() argument 17 struct netfs_inode *ictx = netfs_inode(wreq->inode); in netfs_unbuffered_write_done() 19 _enter("R=%x", wreq->debug_id); in netfs_unbuffered_write_done() 22 trace_netfs_rreq(wreq, netfs_rreq_trace_write_done); in netfs_unbuffered_write_done() 24 if (!wreq->error) in netfs_unbuffered_write_done() 25 netfs_update_i_size(ictx, &ictx->inode, wreq->start, wreq->transferred); in netfs_unbuffered_write_done() 27 if (wreq->origin == NETFS_DIO_WRITE && in netfs_unbuffered_write_done() 28 wreq->mapping->nrpages) { in netfs_unbuffered_write_done() 35 pgoff_t first = wreq->start >> PAGE_SHIFT; in netfs_unbuffered_write_done() 36 pgoff_t last = (wreq->start + wreq->transferred - 1) >> PAGE_SHIFT; in netfs_unbuffered_write_done() [all …]
|
| H A D | write_issue.c | 95 struct netfs_io_request *wreq; in netfs_create_write_req() local 102 wreq = netfs_alloc_request(mapping, file, start, 0, origin); in netfs_create_write_req() 103 if (IS_ERR(wreq)) in netfs_create_write_req() 104 return wreq; in netfs_create_write_req() 106 _enter("R=%x", wreq->debug_id); in netfs_create_write_req() 108 ictx = netfs_inode(wreq->inode); in netfs_create_write_req() 110 fscache_begin_write_operation(&wreq->cache_resources, netfs_i_cookie(ictx)); in netfs_create_write_req() 111 if (rolling_buffer_init(&wreq->buffer, wreq->debug_id, ITER_SOURCE) < 0) in netfs_create_write_req() 114 wreq->cleaned_to = wreq->start; in netfs_create_write_req() 116 wreq->io_streams[0].stream_nr = 0; in netfs_create_write_req() [all …]
|
| H A D | write_collect.c | 111 static void netfs_writeback_unlock_folios(struct netfs_io_request *wreq, in netfs_writeback_unlock_folios() argument 114 struct folio_queue *folioq = wreq->buffer.tail; in netfs_writeback_unlock_folios() 115 unsigned long long collected_to = wreq->collected_to; in netfs_writeback_unlock_folios() 116 unsigned int slot = wreq->buffer.first_tail_slot; in netfs_writeback_unlock_folios() 120 netfs_dump_request(wreq); in netfs_writeback_unlock_folios() 124 if (wreq->origin == NETFS_PGPRIV2_COPY_TO_CACHE) { in netfs_writeback_unlock_folios() 125 if (netfs_pgpriv2_unlock_copied_folios(wreq)) in netfs_writeback_unlock_folios() 131 folioq = rolling_buffer_delete_spent(&wreq->buffer); in netfs_writeback_unlock_folios() 146 wreq->debug_id, folio->index)) in netfs_writeback_unlock_folios() 154 fend = min_t(unsigned long long, fpos + flen, wreq->i_size); in netfs_writeback_unlock_folios() [all …]
|
| H A D | fscache_io.c | 197 struct fscache_write_request *wreq = priv; in fscache_wreq_done() local 199 if (wreq->using_pgpriv2) in fscache_wreq_done() 200 fscache_clear_page_bits(wreq->mapping, wreq->start, wreq->len, in fscache_wreq_done() 201 wreq->set_bits); in fscache_wreq_done() 203 if (wreq->term_func) in fscache_wreq_done() 204 wreq->term_func(wreq->term_func_priv, transferred_or_error); in fscache_wreq_done() 205 fscache_end_operation(&wreq->cache_resources); in fscache_wreq_done() 206 kfree(wreq); in fscache_wreq_done() 216 struct fscache_write_request *wreq; in __fscache_write_to_cache() local 226 wreq = kzalloc_obj(struct fscache_write_request, GFP_NOFS); in __fscache_write_to_cache() [all …]
|
| H A D | buffered_write.c | 120 struct netfs_io_request *wreq = NULL; in netfs_perform_write() local 138 wreq = netfs_begin_writethrough(iocb, iter->count); in netfs_perform_write() 139 if (IS_ERR(wreq)) { in netfs_perform_write() 141 ret = PTR_ERR(wreq); in netfs_perform_write() 142 wreq = NULL; in netfs_perform_write() 146 wreq->iocb = iocb; in netfs_perform_write() 361 if (likely(!wreq)) { in netfs_perform_write() 365 netfs_advance_writethrough(wreq, &wbc, folio, copied, in netfs_perform_write() 391 if (unlikely(wreq)) { in netfs_perform_write() 392 ret2 = netfs_end_writethrough(wreq, &wbc, writethrough); in netfs_perform_write()
|
| H A D | internal.h | 118 bool netfs_pgpriv2_unlock_copied_folios(struct netfs_io_request *wreq); 191 bool netfs_write_collection(struct netfs_io_request *wreq); 201 void netfs_prepare_write(struct netfs_io_request *wreq, 207 void netfs_issue_write(struct netfs_io_request *wreq, 209 size_t netfs_advance_write(struct netfs_io_request *wreq, 213 int netfs_advance_writethrough(struct netfs_io_request *wreq, struct writeback_control *wbc, 216 ssize_t netfs_end_writethrough(struct netfs_io_request *wreq, struct writeback_control *wbc, 222 void netfs_retry_writes(struct netfs_io_request *wreq);
|
| /linux/include/trace/events/ |
| H A D | netfs.h | 532 TP_PROTO(const struct netfs_io_request *wreq, 535 TP_ARGS(wreq, what), 538 __field(unsigned int, wreq) 547 struct netfs_inode *__ctx = netfs_inode(wreq->inode); 549 __entry->wreq = wreq->debug_id; 551 __entry->ino = wreq->inode->i_ino; 553 __entry->start = wreq->start; 554 __entry->len = wreq->len; 558 __entry->wreq, 595 TP_PROTO(const struct netfs_io_request *wreq), [all …]
|
| /linux/fs/afs/ |
| H A D | write.c | 36 static void afs_get_writeback_key(struct netfs_io_request *wreq) in afs_get_writeback_key() argument 38 struct afs_wb_key *wbk, *old = wreq->netfs_priv2; in afs_get_writeback_key() 39 struct afs_vnode *vnode = AFS_FS_I(wreq->inode); in afs_get_writeback_key() 41 key_put(wreq->netfs_priv); in afs_get_writeback_key() 42 wreq->netfs_priv = NULL; in afs_get_writeback_key() 43 wreq->netfs_priv2 = NULL; in afs_get_writeback_key() 55 wreq->netfs_priv = key_get(wbk->key); in afs_get_writeback_key() 56 wreq->netfs_priv2 = wbk; in afs_get_writeback_key() 106 struct netfs_io_request *wreq = subreq->rreq; in afs_issue_write_worker() local 108 struct afs_vnode *vnode = AFS_FS_I(wreq->inode); in afs_issue_write_worker() [all …]
|
| H A D | file.c | 442 static void afs_netfs_invalidate_cache(struct netfs_io_request *wreq) in afs_netfs_invalidate_cache() argument 444 struct afs_vnode *vnode = AFS_FS_I(wreq->inode); in afs_netfs_invalidate_cache()
|
| H A D | internal.h | 1685 void afs_begin_writeback(struct netfs_io_request *wreq); 1686 void afs_retry_request(struct netfs_io_request *wreq, struct netfs_io_stream *stream);
|
| /linux/fs/cachefiles/ |
| H A D | io.c | 628 struct netfs_io_request *wreq = subreq->rreq; in cachefiles_prepare_write_subreq() local 629 struct netfs_cache_resources *cres = &wreq->cache_resources; in cachefiles_prepare_write_subreq() 630 struct netfs_io_stream *stream = &wreq->io_streams[subreq->stream_nr]; in cachefiles_prepare_write_subreq() 632 _enter("W=%x[%x] %llx", wreq->debug_id, subreq->debug_index, subreq->start); in cachefiles_prepare_write_subreq() 647 struct netfs_io_request *wreq = subreq->rreq; in cachefiles_issue_write() local 648 struct netfs_cache_resources *cres = &wreq->cache_resources; in cachefiles_issue_write() 651 struct netfs_io_stream *stream = &wreq->io_streams[subreq->stream_nr]; in cachefiles_issue_write() 658 wreq->debug_id, subreq->debug_index, start, start + len - 1); in cachefiles_issue_write() 676 if (start + len == wreq->i_size) { in cachefiles_issue_write()
|
| /linux/net/9p/ |
| H A D | trans_fd.c | 85 struct p9_req_t *wreq; member 442 m->wreq = req; in p9_write_work() 465 p9_req_put(m->client, m->wreq); in p9_write_work() 466 m->wreq = NULL; in p9_write_work() 811 if (m->wreq) { in p9_conn_destroy() 812 p9_req_put(m->client, m->wreq); in p9_conn_destroy() 813 m->wreq = NULL; in p9_conn_destroy()
|
| H A D | client.c | 1563 struct netfs_io_request *wreq = subreq->rreq; in p9_client_write_subreq() local 1564 struct p9_fid *fid = wreq->netfs_priv; in p9_client_write_subreq() 1577 0, wreq->len, P9_ZC_HDR_SZ, "dqd", in p9_client_write_subreq()
|
| /linux/include/linux/ |
| H A D | netfs.h | 309 void (*begin_writeback)(struct netfs_io_request *wreq); 312 void (*retry_request)(struct netfs_io_request *wreq, struct netfs_io_stream *stream); 313 void (*invalidate_cache)(struct netfs_io_request *wreq);
|
| /linux/drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/ |
| H A D | chcr_ipsec.h | 33 struct fw_ulptx_wr wreq; member
|
| H A D | chcr_ipsec.c | 595 wr->wreq.op_to_compl = htonl(FW_WR_OP_V(FW_ULPTX_WR)); in ch_ipsec_crypto_wreq() 605 wr->wreq.flowid_len16 = htonl(wr_mid); in ch_ipsec_crypto_wreq()
|
| /linux/fs/smb/client/ |
| H A D | compress.c | 293 const struct smb2_write_req *wreq = rq->rq_iov->iov_base; in should_compress() local 295 if (le32_to_cpu(wreq->Length) < SMB_COMPRESS_MIN_LEN) in should_compress()
|
| H A D | file.c | 146 static void cifs_netfs_invalidate_cache(struct netfs_io_request *wreq) in cifs_netfs_invalidate_cache() argument 148 cifs_invalidate_cache(wreq->inode, 0); in cifs_netfs_invalidate_cache() 253 static void cifs_begin_writeback(struct netfs_io_request *wreq) in cifs_begin_writeback() argument 255 struct cifs_io_request *req = container_of(wreq, struct cifs_io_request, rreq); in cifs_begin_writeback() 258 ret = cifs_get_writable_file(CIFS_I(wreq->inode), FIND_ANY, &req->cfile); in cifs_begin_writeback() 264 wreq->io_streams[0].avail = true; in cifs_begin_writeback() 2494 struct netfs_io_request *wreq = wdata->rreq; in cifs_write_subrequest_terminated() local 2495 struct netfs_inode *ictx = netfs_inode(wreq->inode); in cifs_write_subrequest_terminated()
|
| /linux/drivers/crypto/chelsio/ |
| H A D | chcr_core.h | 89 struct fw_crypto_lookaside_wr wreq; member
|
| /linux/net/wireless/ |
| H A D | scan.c | 3502 struct iw_scan_req *wreq = NULL; in cfg80211_wext_siwscan() local 3511 wreq = (struct iw_scan_req *)extra; in cfg80211_wext_siwscan() 3524 if (wreq && wreq->num_channels) { in cfg80211_wext_siwscan() 3526 if (unlikely(wreq->num_channels > IW_MAX_FREQUENCIES)) in cfg80211_wext_siwscan() 3528 n_channels = wreq->num_channels; in cfg80211_wext_siwscan() 3569 if (wreq && wreq->num_channels) { in cfg80211_wext_siwscan() 3572 for (k = 0; k < wreq->num_channels; k++) { in cfg80211_wext_siwscan() 3574 &wreq->channel_list[k]; in cfg80211_wext_siwscan() 3601 if (wreq) { in cfg80211_wext_siwscan() 3603 if (wreq->essid_len > IEEE80211_MAX_SSID_LEN) in cfg80211_wext_siwscan() [all …]
|