Lines Matching refs:request_queue
31 struct request_queue;
162 struct request_queue *queue;
479 struct request_queue {
676 void blk_queue_flag_set(unsigned int flag, struct request_queue *q);
677 void blk_queue_flag_clear(unsigned int flag, struct request_queue *q);
711 extern void blk_set_pm_only(struct request_queue *q);
712 extern void blk_clear_pm_only(struct request_queue *q);
720 static inline bool queue_is_mq(struct request_queue *q)
726 static inline enum rpm_status queue_rpm_status(struct request_queue *q)
731 static inline enum rpm_status queue_rpm_status(struct request_queue *q)
737 static inline bool blk_queue_is_zoned(struct request_queue *q)
760 static inline unsigned int blk_queue_depth(struct request_queue *q)
1024 extern int blk_lld_busy(struct request_queue *q);
1025 extern int blk_queue_enter(struct request_queue *q, blk_mq_req_flags_t flags);
1026 extern void blk_queue_exit(struct request_queue *q);
1027 extern void blk_sync_queue(struct request_queue *q);
1042 static inline struct request_queue *bdev_get_queue(struct block_device *bdev)
1086 queue_limits_start_update(struct request_queue *q)
1091 int queue_limits_commit_update_frozen(struct request_queue *q,
1093 int queue_limits_commit_update(struct request_queue *q,
1095 int queue_limits_set(struct request_queue *q, struct queue_limits *lim);
1106 static inline void queue_limits_cancel_update(struct request_queue *q)
1118 static inline void blk_queue_disable_discard(struct request_queue *q)
1123 static inline void blk_queue_disable_secure_erase(struct request_queue *q)
1128 static inline void blk_queue_disable_write_zeroes(struct request_queue *q)
1137 extern void blk_set_queue_depth(struct request_queue *q, unsigned int depth);
1143 extern void blk_queue_rq_timeout(struct request_queue *, unsigned int);
1150 bool __must_check blk_get_queue(struct request_queue *);
1151 extern void blk_put_queue(struct request_queue *);
1165 * the device's request_queue in a batch, this results in improved scalability
1166 * as the lock contention for request_queue lock is reduced.
1315 static inline unsigned long queue_segment_boundary(const struct request_queue *q)
1320 static inline unsigned long queue_virt_boundary(const struct request_queue *q)
1325 static inline unsigned int queue_max_sectors(const struct request_queue *q)
1330 static inline unsigned int queue_max_bytes(struct request_queue *q)
1335 static inline unsigned int queue_max_hw_sectors(const struct request_queue *q)
1340 static inline unsigned short queue_max_segments(const struct request_queue *q)
1345 static inline unsigned short queue_max_discard_segments(const struct request_queue *q)
1350 static inline unsigned int queue_max_segment_size(const struct request_queue *q)
1355 static inline bool queue_emulates_zone_append(struct request_queue *q)
1383 static inline unsigned queue_logical_block_size(const struct request_queue *q)
1393 static inline unsigned int queue_physical_block_size(const struct request_queue *q)
1403 static inline unsigned int queue_io_min(const struct request_queue *q)
1413 static inline unsigned int queue_io_opt(const struct request_queue *q)
1424 queue_zone_write_granularity(const struct request_queue *q)
1477 struct request_queue *q = bdev_get_queue(bdev);
1485 static inline bool blk_queue_write_cache(struct request_queue *q)
1518 struct request_queue *q = bdev_get_queue(bdev);
1559 static inline unsigned int queue_dma_alignment(const struct request_queue *q)
1565 queue_atomic_write_unit_max_bytes(const struct request_queue *q)
1571 queue_atomic_write_unit_min_bytes(const struct request_queue *q)
1577 queue_atomic_write_boundary_bytes(const struct request_queue *q)
1583 queue_atomic_write_max_bytes(const struct request_queue *q)
1599 static inline bool blk_rq_aligned(struct request_queue *q, unsigned long addr,
1624 struct request_queue *q);
1629 struct request_queue *q)
1838 struct request_queue *bd_queue = bdev->bd_queue;