/linux/fs/netfs/ |
H A D | iterator.c | 113 unsigned int nbv = iter->nr_segs, ix = 0, nsegs = 0; in netfs_limit_bvec() local 134 nsegs++; in netfs_limit_bvec() 136 if (span >= max_size || nsegs >= max_segs) in netfs_limit_bvec() 155 unsigned int nsegs = 0; in netfs_limit_xarray() local 182 nsegs++; in netfs_limit_xarray() 183 if (span >= max_size || nsegs >= max_segs) in netfs_limit_xarray() 200 unsigned int nsegs = 0; in netfs_limit_folioq() local 221 nsegs++; in netfs_limit_folioq() 226 if (span >= max_size || nsegs >= max_segs) in netfs_limit_folioq()
|
/linux/block/ |
H A D | blk-merge.c | 134 unsigned *nsegs) in bio_split_discard() argument 140 *nsegs = 1; in bio_split_discard() 170 const struct queue_limits *lim, unsigned *nsegs) in bio_split_write_zeroes() argument 172 *nsegs = 0; in bio_split_write_zeroes() 274 const struct bio_vec *bv, unsigned *nsegs, unsigned *bytes, in bvec_split_segs() argument 282 while (len && *nsegs < max_segs) { in bvec_split_segs() 285 (*nsegs)++; in bvec_split_segs() 316 unsigned nsegs = 0, bytes = 0; in bio_split_rw_at() local 326 if (nsegs < lim->max_segments && in bio_split_rw_at() 329 nsegs++; in bio_split_rw_at() [all …]
|
H A D | blk-map.c | 569 unsigned int nsegs = 0, bytes = 0; in blk_rq_map_user_bvec() local 600 if (nsegs >= nr_segs || bytes > UINT_MAX - bv->bv_len) in blk_rq_map_user_bvec() 607 nsegs++; in blk_rq_map_user_bvec()
|
H A D | blk.h | 335 unsigned *nsegs); 337 const struct queue_limits *lim, unsigned *nsegs);
|
/linux/net/sunrpc/xprtrdma/ |
H A D | rpc_rdma.c | 298 int nsegs, bool writing, in rpcrdma_mr_prepare() argument 310 return frwr_map(r_xprt, seg, nsegs, writing, req->rl_slot.rq_xid, *mr); in rpcrdma_mr_prepare() 342 int nsegs; in rpcrdma_encode_read_list() local 351 nsegs = rpcrdma_convert_iovs(r_xprt, &rqst->rq_snd_buf, pos, in rpcrdma_encode_read_list() 353 if (nsegs < 0) in rpcrdma_encode_read_list() 354 return nsegs; in rpcrdma_encode_read_list() 357 seg = rpcrdma_mr_prepare(r_xprt, req, seg, nsegs, false, &mr); in rpcrdma_encode_read_list() 364 trace_xprtrdma_chunk_read(rqst->rq_task, pos, mr, nsegs); in rpcrdma_encode_read_list() 366 nsegs -= mr->mr_nents; in rpcrdma_encode_read_list() 367 } while (nsegs); in rpcrdma_encode_read_list() [all …]
|
H A D | frwr_ops.c | 287 int nsegs, bool writing, __be32 xid, in frwr_map() argument 296 if (nsegs > ep->re_max_fr_depth) in frwr_map() 297 nsegs = ep->re_max_fr_depth; in frwr_map() 298 for (i = 0; i < nsegs;) { in frwr_map() 306 if ((i < nsegs && seg->mr_offset) || in frwr_map()
|
H A D | xprt_rdma.h | 540 int nsegs, bool writing, __be32 xid,
|
/linux/fs/nilfs2/ |
H A D | sufile.c | 156 int nilfs_sufile_updatev(struct inode *sufile, __u64 *segnumv, size_t nsegs, in nilfs_sufile_updatev() argument 168 if (unlikely(nsegs == 0)) in nilfs_sufile_updatev() 172 for (seg = segnumv; seg < segnumv + nsegs; seg++) { in nilfs_sufile_updatev() 198 if (++seg >= segnumv + nsegs) in nilfs_sufile_updatev() 269 __u64 nsegs; in nilfs_sufile_set_alloc_range() local 273 nsegs = nilfs_sufile_get_nsegments(sufile); in nilfs_sufile_set_alloc_range() 275 if (start <= end && end < nsegs) { in nilfs_sufile_set_alloc_range() 687 unsigned long nsegs, ncleaned; in nilfs_sufile_truncate_range() local 694 nsegs = nilfs_sufile_get_nsegments(sufile); in nilfs_sufile_truncate_range() 697 if (start > end || start >= nsegs) in nilfs_sufile_truncate_range() [all …]
|
H A D | sufile.h | 85 size_t nsegs, size_t *ndone) in nilfs_sufile_freev() argument 87 return nilfs_sufile_updatev(sufile, segnumv, nsegs, 0, ndone, in nilfs_sufile_freev() 102 __u64 *segnumv, size_t nsegs, in nilfs_sufile_cancel_freev() argument 105 return nilfs_sufile_updatev(sufile, segnumv, nsegs, 0, ndone, in nilfs_sufile_cancel_freev()
|
H A D | the_nilfs.c | 401 unsigned long nilfs_nrsvsegs(struct the_nilfs *nilfs, unsigned long nsegs) in nilfs_nrsvsegs() argument 404 DIV_ROUND_UP(nsegs * nilfs->ns_r_segments_percentage, in nilfs_nrsvsegs() 420 void nilfs_set_nsegments(struct the_nilfs *nilfs, unsigned long nsegs) in nilfs_set_nsegments() argument 422 nilfs->ns_nsegments = nsegs; in nilfs_set_nsegments() 423 nilfs->ns_nrsvsegs = nilfs_nrsvsegs(nilfs, nsegs); in nilfs_set_nsegments() 784 size_t nsegs) in nilfs_discard_segments() argument 794 for (sn = segnump; sn < segnump + nsegs; sn++) { in nilfs_discard_segments()
|
H A D | the_nilfs.h | 277 unsigned long nilfs_nrsvsegs(struct the_nilfs *nilfs, unsigned long nsegs); 278 void nilfs_set_nsegments(struct the_nilfs *nilfs, unsigned long nsegs);
|
H A D | ioctl.c | 857 size_t len, nsegs; in nilfs_ioctl_clean_segments() local 872 nsegs = argv[4].v_nmembs; in nilfs_ioctl_clean_segments() 882 nsegs, sizeof(__u64)); in nilfs_ioctl_clean_segments() 894 if (argv[n].v_nmembs > nsegs * nilfs->ns_blocks_per_segment) in nilfs_ioctl_clean_segments()
|
/linux/include/uapi/linux/ |
H A D | elf-fdpic.h | 29 Elf32_Half nsegs; /* number of segments */ member 44 Elf64_Half nsegs; /* number of segments */ member
|
/linux/drivers/net/ethernet/google/gve/ |
H A D | gve_tx.c | 745 int nsegs; in gve_tx() local 760 nsegs = gve_tx_add_skb_no_copy(priv, tx, skb); in gve_tx() 762 nsegs = gve_tx_add_skb_copy(priv, tx, skb); in gve_tx() 765 if (nsegs) { in gve_tx() 768 tx->req += nsegs; in gve_tx() 869 int nsegs; in gve_xdp_xmit_one() local 874 nsegs = gve_tx_fill_xdp(priv, tx, data, len, frame_p, false); in gve_xdp_xmit_one() 875 tx->req += nsegs; in gve_xdp_xmit_one() 951 int sent = 0, nsegs; in gve_xsk_tx() local 965 nsegs = gve_tx_fill_xdp(priv, tx, data, desc.len, NULL, true); in gve_xsk_tx() [all …]
|
/linux/include/trace/events/ |
H A D | rpcrdma.h | 395 int nsegs 398 TP_ARGS(task, pos, mr, nsegs), 408 __field(int, nsegs) 419 __entry->nsegs = nsegs; 427 __entry->nents < __entry->nsegs ? "more" : "last" 437 int nsegs \ 439 TP_ARGS(task, pos, mr, nsegs)) 445 int nsegs 448 TP_ARGS(task, mr, nsegs), 457 __field(int, nsegs) [all …]
|
/linux/fs/ |
H A D | binfmt_elf_fdpic.c | 561 len += sizeof(struct elf_fdpic_loadseg) * exec_params->loadmap->nsegs; in create_elf_fdpic_tables() 573 interp_params->loadmap->nsegs; in create_elf_fdpic_tables() 762 loadmap->nsegs = nloads; in elf_fdpic_map_file() 784 for (loop = loadmap->nsegs; loop > 0; loop--, seg++) { in elf_fdpic_map_file() 809 for (loop = loadmap->nsegs; loop > 0; loop--, seg++) { in elf_fdpic_map_file() 830 for (loop = loadmap->nsegs; loop > 0; loop--, seg++) { in elf_fdpic_map_file() 864 nloads = loadmap->nsegs; in elf_fdpic_map_file() 877 loadmap->nsegs--; in elf_fdpic_map_file() 894 for (loop = 0; loop < loadmap->nsegs; loop++, seg++) in elf_fdpic_map_file()
|
/linux/crypto/ |
H A D | rsa-pkcs1pad.c | 188 int nsegs = next ? 2 : 1; in pkcs1pad_sg_set_buf() local 190 sg_init_table(sg, nsegs); in pkcs1pad_sg_set_buf() 194 sg_chain(sg, nsegs, next); in pkcs1pad_sg_set_buf()
|
/linux/drivers/media/dvb-frontends/ |
H A D | dib8000.c | 4014 int guard, rate_num, rate_denum = 1, bits_per_symbol, nsegs; in dib8000_get_time_us() local 4056 nsegs = c->layer[i].segment_count; in dib8000_get_time_us() 4057 if (nsegs == 0 || nsegs > 13) in dib8000_get_time_us() 4100 denom += bits_per_symbol * rate_num * fft_div * nsegs * 384; in dib8000_get_time_us() 4221 unsigned nsegs = c->layer[i].segment_count; in dib8000_get_stats() local 4223 if (nsegs == 0 || nsegs > 13) in dib8000_get_stats()
|
/linux/drivers/scsi/bfa/ |
H A D | bfa_svc.c | 438 u16 nsegs, idx, per_seg_fcxp; in bfa_fcxp_meminfo() local 451 nsegs = BFI_MEM_DMA_NSEGS(num_fcxps, per_fcxp_sz); in bfa_fcxp_meminfo() 454 bfa_mem_dma_seg_iter(fcxp_mod, seg_ptr, nsegs, idx) { in bfa_fcxp_meminfo() 5130 u16 nsegs, idx, per_seg_sgpg, num_sgpg; in bfa_sgpg_meminfo() local 5140 nsegs = BFI_MEM_DMA_NSEGS(num_sgpg, sgpg_sz); in bfa_sgpg_meminfo() 5143 bfa_mem_dma_seg_iter(sgpg_mod, seg_ptr, nsegs, idx) { in bfa_sgpg_meminfo() 5168 u16 i, idx, nsegs, per_seg_sgpg, num_sgpg; in bfa_sgpg_attach() local 5183 nsegs = BFI_MEM_DMA_NSEGS(num_sgpg, sgpg_sz); in bfa_sgpg_attach() 5188 bfa_mem_dma_seg_iter(mod, seg_ptr, nsegs, idx) { in bfa_sgpg_attach() 5412 u16 nsegs, idx, per_seg_uf = 0; in bfa_uf_meminfo() local [all …]
|
H A D | bfa_fcpim.c | 3579 u16 nsegs, idx, per_seg_ios, num_io_req; in bfa_fcp_meminfo() local 3613 nsegs = BFI_MEM_DMA_NSEGS(num_io_req, BFI_IOIM_SNSLEN); in bfa_fcp_meminfo() 3616 bfa_mem_dma_seg_iter(fcp, seg_ptr, nsegs, idx) { in bfa_fcp_meminfo() 3636 u16 idx, nsegs, num_io_req; in bfa_fcp_attach() local 3649 nsegs = BFI_MEM_DMA_NSEGS(num_io_req, BFI_IOIM_SNSLEN); in bfa_fcp_attach() 3651 bfa_mem_dma_seg_iter(fcp, seg_ptr, nsegs, idx) { in bfa_fcp_attach()
|
/linux/include/net/ |
H A D | addrconf.h | 109 unsigned char nsegs);
|
/linux/drivers/scsi/lpfc/ |
H A D | lpfc_nvmet.c | 2724 int i, cnt, nsegs; in lpfc_nvmet_prep_fcp_wqe() local 2755 nsegs = rsp->sg_cnt; in lpfc_nvmet_prep_fcp_wqe() 2911 if (nsegs == 1 && phba->cfg_enable_pbde) { in lpfc_nvmet_prep_fcp_wqe() 2992 nsegs = 0; in lpfc_nvmet_prep_fcp_wqe() 3009 for_each_sg(rsp->sg, sgel, nsegs, i) { in lpfc_nvmet_prep_fcp_wqe()
|
/linux/net/ipv6/ |
H A D | addrconf.c | 4566 unsigned char nsegs) in ipv6_chk_rpl_srh_loop() argument 4576 for (i = 0; i < nsegs; i++) { in ipv6_chk_rpl_srh_loop()
|
/linux/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt.c | 6124 u32 nsegs, n, segs = 0, flags; in bnxt_hwrm_vnic_set_tpa() local 6146 nsegs = (MAX_SKB_FRAGS - 1) * n; in bnxt_hwrm_vnic_set_tpa() 6151 nsegs = (MAX_SKB_FRAGS - n) / n; in bnxt_hwrm_vnic_set_tpa() 6158 segs = ilog2(nsegs); in bnxt_hwrm_vnic_set_tpa()
|