Lines Matching refs:limits

502 	struct queue_limits	limits;  member
639 #define blk_queue_nonrot(q) (!((q)->limits.features & BLK_FEAT_ROTATIONAL))
640 #define blk_queue_io_stat(q) ((q)->limits.features & BLK_FEAT_IO_STAT)
642 ((q)->limits.flags & BLK_FLAG_IOSTATS_PASSTHROUGH)
643 #define blk_queue_dax(q) ((q)->limits.features & BLK_FEAT_DAX)
644 #define blk_queue_pci_p2pdma(q) ((q)->limits.features & BLK_FEAT_PCI_P2PDMA)
660 ((q)->limits.features & BLK_FEAT_SKIP_TAGSET_QUIESCE)
691 (q->limits.features & BLK_FEAT_ZONED); in blk_queue_is_zoned()
715 return sector >> ilog2(disk->queue->limits.chunk_sectors); in disk_zone_no()
725 return bdev->bd_disk->queue->limits.max_open_zones; in bdev_max_open_zones()
730 return bdev->bd_disk->queue->limits.max_active_zones; in bdev_max_active_zones()
952 return q->limits; in queue_limits_start_update()
983 q->limits.max_discard_sectors = 0; in blk_queue_disable_discard()
988 q->limits.max_secure_erase_sectors = 0; in blk_queue_disable_secure_erase()
993 q->limits.max_write_zeroes_sectors = 0; in blk_queue_disable_write_zeroes()
1183 return &bdev_get_queue(bdev)->limits; in bdev_limits()
1188 return q->limits.seg_boundary_mask; in queue_segment_boundary()
1193 return q->limits.virt_boundary_mask; in queue_virt_boundary()
1198 return q->limits.max_sectors; in queue_max_sectors()
1208 return q->limits.max_hw_sectors; in queue_max_hw_sectors()
1213 return q->limits.max_segments; in queue_max_segments()
1218 return q->limits.max_discard_segments; in queue_max_discard_segments()
1223 return q->limits.max_segment_size; in queue_max_segment_size()
1228 return blk_queue_is_zoned(q) && !q->limits.max_hw_zone_append_sectors; in queue_emulates_zone_append()
1249 return q->limits.logical_block_size; in queue_logical_block_size()
1259 return q->limits.physical_block_size; in queue_physical_block_size()
1269 return q->limits.io_min; in queue_io_min()
1279 return q->limits.io_opt; in queue_io_opt()
1290 return q->limits.zone_write_granularity; in queue_zone_write_granularity()
1330 return bdev->bd_disk->queue->limits.features & BLK_FEAT_SYNCHRONOUS; in bdev_synchronous()
1338 q->limits.integrity.csum_type != BLK_INTEGRITY_CSUM_NONE) in bdev_stable_writes()
1340 return q->limits.features & BLK_FEAT_STABLE_WRITES; in bdev_stable_writes()
1345 return (q->limits.features & BLK_FEAT_WRITE_CACHE) && in blk_queue_write_cache()
1346 !(q->limits.flags & BLK_FLAG_WRITE_CACHE_DISABLED); in blk_queue_write_cache()
1361 return bdev->bd_disk->queue->limits.features & BLK_FEAT_NOWAIT; in bdev_nowait()
1380 return q->limits.chunk_sectors; in bdev_zone_sectors()
1433 return q->limits.dma_alignment; in queue_dma_alignment()
1439 return q->limits.atomic_write_unit_max; in queue_atomic_write_unit_max_bytes()
1445 return q->limits.atomic_write_unit_min; in queue_atomic_write_unit_min_bytes()
1451 return q->limits.atomic_write_boundary_sectors << SECTOR_SHIFT; in queue_atomic_write_boundary_bytes()
1457 return q->limits.atomic_write_max_sectors << SECTOR_SHIFT; in queue_atomic_write_max_bytes()
1481 unsigned int alignment = blk_lim_dma_alignment_and_pad(&q->limits); in blk_rq_aligned()
1709 struct queue_limits *limits) in blk_atomic_write_start_sect_aligned() argument
1711 unsigned int alignment = max(limits->atomic_write_hw_unit_min, in blk_atomic_write_start_sect_aligned()
1712 limits->atomic_write_hw_boundary); in blk_atomic_write_start_sect_aligned()
1720 struct queue_limits *limits = &bd_queue->limits; in bdev_can_atomic_write() local
1722 if (!limits->atomic_write_unit_min) in bdev_can_atomic_write()
1727 limits); in bdev_can_atomic_write()