Home
last modified time | relevance | path

Searched refs:sector (Results 1 – 25 of 195) sorted by relevance

12345678

/linux/drivers/block/drbd/
H A Ddrbd_interval.c16 #define NODE_END(node) ((node)->sector + ((node)->size >> 9))
28 sector_t this_end = this->sector + (this->size >> 9); in drbd_insert_interval()
39 if (this->sector < here->sector) in drbd_insert_interval()
41 else if (this->sector > here->sector) in drbd_insert_interval()
69 drbd_contains_interval(struct rb_root *root, sector_t sector, in drbd_contains_interval() argument
78 if (sector < here->sector) in drbd_contains_interval()
80 else if (sector > here->sector) in drbd_contains_interval()
118 drbd_find_overlap(struct rb_root *root, sector_t sector, unsigned int size) in drbd_find_overlap() argument
122 sector_t end = sector + (size >> 9); in drbd_find_overlap()
131 sector < interval_end(node->rb_left)) { in drbd_find_overlap()
[all …]
H A Ddrbd_actlog.c127 sector_t sector, enum req_op op) in _drbd_md_sync_page_io() argument
144 bio->bi_iter.bi_sector = sector; in _drbd_md_sync_page_io()
178 sector_t sector, enum req_op op) in drbd_md_sync_page_io() argument
187 (unsigned long long)sector, (op == REQ_OP_WRITE) ? "WRITE" : "READ", in drbd_md_sync_page_io()
190 if (sector < drbd_md_first_sector(bdev) || in drbd_md_sync_page_io()
191 sector + 7 > drbd_md_last_sector(bdev)) in drbd_md_sync_page_io()
194 (unsigned long long)sector, in drbd_md_sync_page_io()
197 err = _drbd_md_sync_page_io(device, bdev, sector, op); in drbd_md_sync_page_io()
200 (unsigned long long)sector, in drbd_md_sync_page_io()
245 unsigned first = i->sector >> (AL_EXTENT_SHIFT-9); in drbd_al_begin_io_fastpath()
[all …]
H A Ddrbd_worker.c127 drbd_set_out_of_sync(peer_device, peer_req->i.sector, peer_req->i.size); in drbd_endio_write_sec_final()
157 drbd_rs_complete_io(device, i.sector); in drbd_endio_write_sec_final()
183 (unsigned long long)peer_req->i.sector); in drbd_peer_request_endio()
356 sector_t sector = peer_req->i.sector; in w_e_send_csum() local
367 err = drbd_send_drequest_csum(peer_device, sector, size, in w_e_send_csum()
387 static int read_for_csum(struct drbd_peer_device *peer_device, sector_t sector, int size) in read_for_csum() argument
397 peer_req = drbd_alloc_peer_req(peer_device, ID_SYNCER /* unused */, sector, in read_for_csum()
591 sector_t sector; in make_resync_request() local
657 sector = BM_BIT_TO_SECT(bit); in make_resync_request()
659 if (drbd_try_rs_begin_io(peer_device, sector)) { in make_resync_request()
[all …]
H A Ddrbd_receiver.c177 drbd_alloc_peer_req(struct drbd_peer_device *peer_device, u64 id, sector_t sector, in drbd_alloc_peer_req() argument
208 peer_req->i.sector = sector; in drbd_alloc_peer_req()
1414 if (drbd_issue_discard_or_zero_out(device, peer_req->i.sector, in drbd_issue_peer_discard_or_zero_out()
1452 sector_t sector = peer_req->i.sector; in drbd_submit_peer_request() local
1506 bio->bi_iter.bi_sector = sector; in drbd_submit_peer_request()
1519 sector += len >> 9; in drbd_submit_peer_request()
1659 read_in_block(struct drbd_peer_device *peer_device, u64 id, sector_t sector, in read_in_block() argument
1709 if (sector + (ds>>9) > capacity) { in read_in_block()
1713 (unsigned long long)sector, ds); in read_in_block()
1720 peer_req = drbd_alloc_peer_req(peer_device, id, sector, ds, data_size, GFP_NOIO); in read_in_block()
[all …]
H A Ddrbd_interval.h10 sector_t sector; /* start sector of the interval */ member
41 #define drbd_for_each_overlap(i, root, sector, size) \ argument
42 for (i = drbd_find_overlap(root, sector, size); \
44 i = drbd_next_overlap(i, sector, size))
/linux/include/trace/events/
H A Dblock.h32 __field( sector_t, sector )
38 __entry->sector = bh->b_blocknr;
44 (unsigned long long)__entry->sector, __entry->size
91 __field( sector_t, sector )
100 __entry->sector = blk_rq_trace_sector(rq);
111 (unsigned long long)__entry->sector, __entry->nr_sector,
126 __field( sector_t, sector )
136 __entry->sector = blk_rq_pos(rq);
148 (unsigned long long)__entry->sector, __entry->nr_sector,
198 __field( sector_t, sector )
[all …]
H A Dbcache.h18 __field(sector_t, sector )
28 __entry->sector = bio->bi_iter.bi_sector;
36 __entry->rwbs, (unsigned long long)__entry->sector,
96 __field(sector_t, sector )
103 __entry->sector = bio->bi_iter.bi_sector;
110 (unsigned long long)__entry->sector, __entry->nr_sector)
129 __field(sector_t, sector )
138 __entry->sector = bio->bi_iter.bi_sector;
147 __entry->rwbs, (unsigned long long)__entry->sector,
159 __field(sector_t, sector )
[all …]
/linux/block/
H A Dblk-zoned.c156 static void disk_zone_set_cond(struct gendisk *disk, sector_t sector, in disk_zone_set_cond() argument
164 unsigned int zno = disk_zone_no(disk, sector); in disk_zone_set_cond()
191 bool bdev_zone_is_seq(struct block_device *bdev, sector_t sector) in bdev_zone_is_seq() argument
194 unsigned int zno = disk_zone_no(disk, sector); in bdev_zone_is_seq()
222 static int blkdev_do_report_zones(struct block_device *bdev, sector_t sector, in blkdev_do_report_zones() argument
231 if (!nr_zones || sector >= get_capacity(disk)) in blkdev_do_report_zones()
234 return disk->fops->report_zones(disk, sector, nr_zones, args); in blkdev_do_report_zones()
256 int blkdev_report_zones(struct block_device *bdev, sector_t sector, in blkdev_report_zones() argument
264 return blkdev_do_report_zones(bdev, sector, nr_zones, &args); in blkdev_report_zones()
293 sector_t sector, sector_t nr_sectors) in blkdev_zone_mgmt() argument
[all …]
H A Dblk-ia-ranges.c18 return sprintf(buf, "%llu\n", iar->sector); in blk_ia_range_sector_show()
176 sector_t sector) in disk_find_ia_range() argument
183 if (sector >= iar->sector && in disk_find_ia_range()
184 sector < iar->sector + iar->nr_sectors) in disk_find_ia_range()
196 sector_t sector = 0; in disk_check_ia_ranges() local
208 tmp = disk_find_ia_range(iars, sector); in disk_check_ia_ranges()
209 if (!tmp || tmp->sector != sector) { in disk_check_ia_ranges()
216 swap(iar->sector, tmp->sector); in disk_check_ia_ranges()
220 sector += iar->nr_sectors; in disk_check_ia_ranges()
223 if (sector != capacity) { in disk_check_ia_ranges()
[all …]
/linux/drivers/block/null_blk/
H A Dzoned.c56 sector_t sector = 0; in null_init_zoned_dev() local
134 zone->start = sector; in null_init_zoned_dev()
141 sector += dev->zone_size_sects; in null_init_zoned_dev()
148 zone->start = sector; in null_init_zoned_dev()
164 sector += dev->zone_size_sects; in null_init_zoned_dev()
193 int null_report_zones(struct gendisk *disk, sector_t sector, in null_report_zones() argument
203 first_zone = null_zone_no(dev, sector); in null_report_zones()
241 sector_t sector, unsigned int len) in null_zone_valid_read_len() argument
244 struct nullb_zone *zone = &dev->zones[null_zone_no(dev, sector)]; in null_zone_valid_read_len()
249 sector + nr_sectors <= zone->wp) in null_zone_valid_read_len()
[all …]
H A Dmain.c900 static void null_free_sector(struct nullb *nullb, sector_t sector, in null_free_sector() argument
909 idx = sector >> PAGE_SECTORS_SHIFT; in null_free_sector()
910 sector_bit = (sector & SECTOR_MASK); in null_free_sector()
973 sector_t sector, bool for_write, bool is_cache) in __null_lookup_page() argument
980 idx = sector >> PAGE_SECTORS_SHIFT; in __null_lookup_page()
981 sector_bit = (sector & SECTOR_MASK); in __null_lookup_page()
994 sector_t sector, bool for_write, bool ignore_cache) in null_lookup_page() argument
999 page = __null_lookup_page(nullb, sector, for_write, true); in null_lookup_page()
1002 return __null_lookup_page(nullb, sector, for_write, false); in null_lookup_page()
1006 sector_t sector, bool ignore_cache) in null_insert_page() argument
[all …]
/linux/drivers/scsi/
H A Dsr_vendor.c174 unsigned long sector; in sr_cd_check() local
186 sector = 0; /* the multisession sector offset goes here */ in sr_cd_check()
212 sector = buffer[11] + (buffer[10] << 8) + in sr_cd_check()
216 sector = 0; in sr_cd_check()
243 sector = min * CD_SECS * CD_FRAMES + sec * CD_FRAMES + frame; in sr_cd_check()
271 sector = min * CD_SECS * CD_FRAMES + sec * CD_FRAMES + frame; in sr_cd_check()
272 if (sector) in sr_cd_check()
273 sector -= CD_MSF_OFFSET; in sr_cd_check()
309 sector = buffer[11] + (buffer[10] << 8) + in sr_cd_check()
318 sector = 0; in sr_cd_check()
[all …]
/linux/drivers/block/
H A Dbrd.c49 static struct page *brd_lookup_page(struct brd_device *brd, sector_t sector) in brd_lookup_page() argument
52 XA_STATE(xas, &brd->brd_pages, sector >> PAGE_SECTORS_SHIFT); in brd_lookup_page()
85 static struct page *brd_insert_page(struct brd_device *brd, sector_t sector, in brd_insert_page() argument
96 ret = __xa_cmpxchg(&brd->brd_pages, sector >> PAGE_SECTORS_SHIFT, NULL, in brd_insert_page()
141 sector_t sector = bio->bi_iter.bi_sector; in brd_rw_bvec() local
142 u32 offset = (sector & (PAGE_SECTORS - 1)) << SECTOR_SHIFT; in brd_rw_bvec()
149 page = brd_lookup_page(brd, sector); in brd_rw_bvec()
151 page = brd_insert_page(brd, sector, opf); in brd_rw_bvec()
180 static void brd_do_discard(struct brd_device *brd, sector_t sector, u32 size) in brd_do_discard() argument
182 sector_t aligned_sector = round_up(sector, PAGE_SECTORS); in brd_do_discard()
[all …]
/linux/drivers/usb/storage/
H A Djumpshot.c155 u32 sector, in jumpshot_read_data() argument
171 if (sector > 0x0FFFFFFF) in jumpshot_read_data()
193 command[2] = sector & 0xFF; in jumpshot_read_data()
194 command[3] = (sector >> 8) & 0xFF; in jumpshot_read_data()
195 command[4] = (sector >> 16) & 0xFF; in jumpshot_read_data()
197 command[5] = 0xE0 | ((sector >> 24) & 0x0F); in jumpshot_read_data()
217 sector += thistime; in jumpshot_read_data()
232 u32 sector, in jumpshot_write_data() argument
248 if (sector > 0x0FFFFFFF) in jumpshot_write_data()
275 command[2] = sector & 0xFF; in jumpshot_write_data()
[all …]
/linux/samples/bpf/
H A Dtracex3.bpf.c15 sector_t sector; member
32 .sector = ctx->sector in bpf_prog1()
63 .sector = ctx->sector in bpf_prog2()
/linux/tools/testing/selftests/ublk/trace/
H A Dseq_io.bt24 $expected = @start_order[args.sector];
30 printf("out_of_order: sector %llu started at seq %llu but completed at seq %llu\n",
31 args.sector, $expected, @complete_seq);
34 delete(@start_order[args.sector]);
42 @start_order[args.sector] = @start_seq;
/linux/fs/zonefs/
H A Dtrace.h30 __field(sector_t, sector)
38 __entry->sector = z->z_sector;
43 blk_op_str(__entry->op), __entry->sector,
54 __field(sector_t, sector)
62 __entry->sector = zonefs_inode_zone(inode)->z_sector;
70 __entry->sector, __entry->size, __entry->wpoffset,
/linux/drivers/mtd/
H A Drfd_ftl.c91 static int rfd_ftl_writesect(struct mtd_blktrans_dev *dev, u_long sector, char *buf);
239 static int rfd_ftl_readsect(struct mtd_blktrans_dev *dev, u_long sector, char *buf) in rfd_ftl_readsect() argument
246 if (sector >= part->sector_count) in rfd_ftl_readsect()
249 addr = part->sector_map[sector]; in rfd_ftl_readsect()
600 static int do_writesect(struct mtd_blktrans_dev *dev, u_long sector, char *buf, ulong *old_addr) in do_writesect() argument
641 part->sector_map[sector] = addr; in do_writesect()
643 entry = cpu_to_le16(sector == 0 ? SECTOR_ZERO : sector); in do_writesect()
666 static int rfd_ftl_writesect(struct mtd_blktrans_dev *dev, u_long sector, char *buf) in rfd_ftl_writesect() argument
673 pr_debug("rfd_ftl_writesect(sector=0x%lx)\n", sector); in rfd_ftl_writesect()
680 if (sector >= part->sector_count) { in rfd_ftl_writesect()
[all …]
/linux/drivers/vdpa/vdpa_sim/
H A Dvdpa_sim_blk.c118 u64 sector; in vdpasim_blk_handle_req() local
156 sector = vdpasim64_to_cpu(vdpasim, hdr.sector); in vdpasim_blk_handle_req()
157 offset = sector << SECTOR_SHIFT; in vdpasim_blk_handle_req()
161 sector != 0) { in vdpasim_blk_handle_req()
164 type, sector); in vdpasim_blk_handle_req()
171 if (!vdpasim_blk_check_range(vdpasim, sector, in vdpasim_blk_handle_req()
194 if (!vdpasim_blk_check_range(vdpasim, sector, in vdpasim_blk_handle_req()
255 sector = le64_to_cpu(range.sector); in vdpasim_blk_handle_req()
256 offset = sector << SECTOR_SHIFT; in vdpasim_blk_handle_req()
277 if (!vdpasim_blk_check_range(vdpasim, sector, num_sectors, in vdpasim_blk_handle_req()
/linux/drivers/md/
H A Ddm-log-writes.c97 __le64 sector; member
126 sector_t sector; member
214 sector_t sector) in write_metadata() argument
223 bio->bi_iter.bi_sector = sector; in write_metadata()
224 bio->bi_end_io = (sector == WRITE_LOG_SUPER_SECTOR) ? in write_metadata()
260 sector_t sector) in write_inline_data() argument
276 bio->bi_iter.bi_sector = sector; in write_inline_data()
308 sector += bio_pages * PAGE_SECTORS; in write_inline_data()
319 struct pending_block *block, sector_t sector) in log_one_block() argument
326 entry.sector = cpu_to_le64(block->sector); in log_one_block()
[all …]
H A Draid0.c309 sector_t sector = *sectorp; in find_zone() local
312 if (sector < z[i].zone_end) { in find_zone()
314 *sectorp = sector - z[i-1].zone_end; in find_zone()
325 sector_t sector, sector_t *sector_offset) in map_sector() argument
336 sect_in_chunk = sector & (chunk_sects - 1); in map_sector()
337 sector >>= chunksect_bits; in map_sector()
343 sect_in_chunk = sector_div(sector, chunk_sects); in map_sector()
354 + sector_div(sector, zone->nb_dev)]; in map_sector()
561 sector_t sector = bio_sector; in raid0_map_submit_bio() local
565 zone = find_zone(mddev->private, &sector); in raid0_map_submit_bio()
[all …]
H A Draid1.c64 sector_t lo = r1_bio->sector; in check_and_add_serial()
86 int idx = sector_to_idx(r1_bio->sector); in wait_for_serialization()
262 sector_t sect = r1_bio->sector; in put_buf()
283 idx = sector_to_idx(r1_bio->sector); in reschedule_retry()
312 sector_t sector = r1_bio->sector; in raid_end_bio_io() local
329 allow_barrier(conf, sector); in raid_end_bio_io()
340 r1_bio->sector + (r1_bio->sectors); in update_head_pos()
407 (unsigned long long)r1_bio->sector); in raid1_end_read_request()
454 sector_t lo = r1_bio->sector; in raid1_end_write_request()
455 sector_t hi = r1_bio->sector + r1_bio->sectors; in raid1_end_write_request()
[all …]
H A Draid5-ppl.c166 pr_debug("%s: stripe %llu\n", __func__, (unsigned long long)sh->sector); in ops_run_partial_parity()
277 pr_debug("%s: stripe: %llu\n", __func__, (unsigned long long)sh->sector); in ppl_log_stripe()
303 if (!data_disks || dev->sector < data_sector) in ppl_log_stripe()
304 data_sector = dev->sector; in ppl_log_stripe()
328 if ((sh->sector == sh_last->sector + RAID5_STRIPE_SECTORS(conf)) && in ppl_log_stripe()
461 log->rdev->ppl.sector + log->rdev->ppl.size - log->next_io_sector < in ppl_submit_iounit()
463 log->next_io_sector = log->rdev->ppl.sector; in ppl_submit_iounit()
857 sector_t sector; in ppl_recover_entry() local
876 sector = raid5_compute_sector(conf, r_sector, 0, in ppl_recover_entry()
881 (unsigned long long)sector); in ppl_recover_entry()
[all …]
H A Draid5.c431 (unsigned long long)sh->sector); in remove_hash()
438 struct hlist_head *hp = stripe_hash(conf, sh->sector); in insert_hash()
441 (unsigned long long)sh->sector); in insert_hash()
580 static void init_stripe(struct stripe_head *sh, sector_t sector, int previous) in init_stripe() argument
591 (unsigned long long)sector); in init_stripe()
596 sh->sector = sector; in init_stripe()
597 stripe_set_idx(sector, conf, previous, sh); in init_stripe()
606 (unsigned long long)sh->sector, i, dev->toread, in init_stripe()
612 dev->sector = raid5_compute_blocknr(sh, i, previous); in init_stripe()
622 static struct stripe_head *__find_stripe(struct r5conf *conf, sector_t sector, in __find_stripe() argument
[all …]
/linux/fs/exfat/
H A Dnls.c648 sector_t sector, unsigned long long num_sectors, in exfat_load_upcase_table() argument
663 num_sectors += sector; in exfat_load_upcase_table()
665 while (sector < num_sectors) { in exfat_load_upcase_table()
668 bh = sb_bread(sb, sector); in exfat_load_upcase_table()
671 (unsigned long long)sector); in exfat_load_upcase_table()
674 sector++; in exfat_load_upcase_table()
742 sector_t sector; in exfat_create_upcase_table() local
773 sector = exfat_cluster_to_sector(sbi, tbl_clu); in exfat_create_upcase_table()
775 ret = exfat_load_upcase_table(sb, sector, num_sectors, in exfat_create_upcase_table()

12345678