• Home
  • Raw
  • Download

Lines Matching refs:request_queue

34 struct request_queue;
65 struct request_queue *q; /* the queue this rl belongs to */
142 struct request_queue *q;
285 typedef void (request_fn_proc) (struct request_queue *q);
286 typedef blk_qc_t (make_request_fn) (struct request_queue *q, struct bio *bio);
287 typedef int (prep_rq_fn) (struct request_queue *, struct request *);
288 typedef void (unprep_rq_fn) (struct request_queue *, struct request *);
293 typedef int (lld_busy_fn) (struct request_queue *q);
295 typedef int (init_rq_fn)(struct request_queue *, struct request *, gfp_t);
296 typedef void (exit_rq_fn)(struct request_queue *, struct request *);
404 struct request_queue { struct
670 static inline void queue_lockdep_assert_held(struct request_queue *q) in queue_lockdep_assert_held()
677 struct request_queue *q) in queue_flag_set_unlocked()
683 struct request_queue *q) in queue_flag_test_and_clear()
696 struct request_queue *q) in queue_flag_test_and_set()
708 static inline void queue_flag_set(unsigned int flag, struct request_queue *q) in queue_flag_set()
715 struct request_queue *q) in queue_flag_clear_unlocked()
720 static inline int queue_in_flight(struct request_queue *q) in queue_in_flight()
725 static inline void queue_flag_clear(unsigned int flag, struct request_queue *q) in queue_flag_clear()
775 static inline bool queue_is_rq_based(struct request_queue *q) in queue_is_rq_based()
780 static inline unsigned int blk_queue_cluster(struct request_queue *q) in blk_queue_cluster()
786 blk_queue_zoned_model(struct request_queue *q) in blk_queue_zoned_model()
791 static inline bool blk_queue_is_zoned(struct request_queue *q) in blk_queue_is_zoned()
802 static inline unsigned int blk_queue_zone_sectors(struct request_queue *q) in blk_queue_zone_sectors()
861 static inline unsigned int blk_queue_depth(struct request_queue *q) in blk_queue_depth()
957 extern void blk_rq_init(struct request_queue *q, struct request *rq);
960 extern void __blk_put_request(struct request_queue *, struct request *);
961 extern struct request *blk_get_request(struct request_queue *, unsigned int op,
963 extern void blk_requeue_request(struct request_queue *, struct request *);
964 extern int blk_lld_busy(struct request_queue *q);
970 extern blk_status_t blk_insert_cloned_request(struct request_queue *q,
973 extern void blk_delay_queue(struct request_queue *, unsigned long);
974 extern void blk_queue_split(struct request_queue *, struct bio **);
975 extern void blk_recount_segments(struct request_queue *, struct bio *);
979 extern int scsi_cmd_ioctl(struct request_queue *, struct gendisk *, fmode_t,
981 extern int sg_scsi_ioctl(struct request_queue *, struct gendisk *, fmode_t,
984 extern int blk_queue_enter(struct request_queue *q, bool nowait);
985 extern void blk_queue_exit(struct request_queue *q);
986 extern void blk_start_queue(struct request_queue *q);
987 extern void blk_start_queue_async(struct request_queue *q);
988 extern void blk_stop_queue(struct request_queue *q);
989 extern void blk_sync_queue(struct request_queue *q);
990 extern void __blk_stop_queue(struct request_queue *q);
991 extern void __blk_run_queue(struct request_queue *q);
992 extern void __blk_run_queue_uncond(struct request_queue *q);
993 extern void blk_run_queue(struct request_queue *);
994 extern void blk_run_queue_async(struct request_queue *q);
995 extern int blk_rq_map_user(struct request_queue *, struct request *,
999 extern int blk_rq_map_kern(struct request_queue *, struct request *, void *, unsigned int, gfp_t);
1000 extern int blk_rq_map_user_iov(struct request_queue *, struct request *,
1003 extern void blk_execute_rq(struct request_queue *, struct gendisk *,
1005 extern void blk_execute_rq_nowait(struct request_queue *, struct gendisk *,
1011 bool blk_mq_poll(struct request_queue *q, blk_qc_t cookie);
1013 static inline struct request_queue *bdev_get_queue(struct block_device *bdev) in bdev_get_queue()
1066 static inline unsigned int blk_queue_get_max_sectors(struct request_queue *q, in blk_queue_get_max_sectors()
1085 static inline unsigned int blk_max_size_offset(struct request_queue *q, in blk_max_size_offset()
1098 struct request_queue *q = rq->q; in blk_rq_get_max_sectors()
1126 extern struct request *blk_peek_request(struct request_queue *q);
1128 extern struct request *blk_fetch_request(struct request_queue *q);
1162 extern struct request_queue *blk_init_queue_node(request_fn_proc *rfn,
1164 extern struct request_queue *blk_init_queue(request_fn_proc *, spinlock_t *);
1165 extern int blk_init_allocated_queue(struct request_queue *);
1166 extern void blk_cleanup_queue(struct request_queue *);
1167 extern void blk_queue_make_request(struct request_queue *, make_request_fn *);
1168 extern void blk_queue_bounce_limit(struct request_queue *, u64);
1169 extern void blk_queue_max_hw_sectors(struct request_queue *, unsigned int);
1170 extern void blk_queue_chunk_sectors(struct request_queue *, unsigned int);
1171 extern void blk_queue_max_segments(struct request_queue *, unsigned short);
1172 extern void blk_queue_max_discard_segments(struct request_queue *,
1174 extern void blk_queue_max_segment_size(struct request_queue *, unsigned int);
1175 extern void blk_queue_max_discard_sectors(struct request_queue *q,
1177 extern void blk_queue_max_write_same_sectors(struct request_queue *q,
1179 extern void blk_queue_max_write_zeroes_sectors(struct request_queue *q,
1181 extern void blk_queue_logical_block_size(struct request_queue *, unsigned int);
1182 extern void blk_queue_physical_block_size(struct request_queue *, unsigned int);
1183 extern void blk_queue_alignment_offset(struct request_queue *q,
1186 extern void blk_queue_io_min(struct request_queue *q, unsigned int min);
1188 extern void blk_queue_io_opt(struct request_queue *q, unsigned int opt);
1189 extern void blk_set_queue_depth(struct request_queue *q, unsigned int depth);
1198 extern void blk_queue_stack_limits(struct request_queue *t, struct request_queue *b);
1199 extern void blk_queue_dma_pad(struct request_queue *, unsigned int);
1200 extern void blk_queue_update_dma_pad(struct request_queue *, unsigned int);
1201 extern int blk_queue_dma_drain(struct request_queue *q,
1204 extern void blk_queue_lld_busy(struct request_queue *q, lld_busy_fn *fn);
1205 extern void blk_queue_segment_boundary(struct request_queue *, unsigned long);
1206 extern void blk_queue_virt_boundary(struct request_queue *, unsigned long);
1207 extern void blk_queue_prep_rq(struct request_queue *, prep_rq_fn *pfn);
1208 extern void blk_queue_unprep_rq(struct request_queue *, unprep_rq_fn *ufn);
1209 extern void blk_queue_dma_alignment(struct request_queue *, int);
1210 extern void blk_queue_update_dma_alignment(struct request_queue *, int);
1211 extern void blk_queue_softirq_done(struct request_queue *, softirq_done_fn *);
1212 extern void blk_queue_rq_timed_out(struct request_queue *, rq_timed_out_fn *);
1213 extern void blk_queue_rq_timeout(struct request_queue *, unsigned int);
1214 extern void blk_queue_flush_queueable(struct request_queue *q, bool queueable);
1215 extern void blk_queue_write_cache(struct request_queue *q, bool enabled, bool fua);
1242 extern int blk_rq_map_sg(struct request_queue *, struct request *, struct scatterlist *);
1246 bool __must_check blk_get_queue(struct request_queue *);
1247 struct request_queue *blk_alloc_queue(gfp_t);
1248 struct request_queue *blk_alloc_queue_node(gfp_t, int);
1249 extern void blk_put_queue(struct request_queue *);
1250 extern void blk_set_queue_dying(struct request_queue *);
1256 extern void blk_pm_runtime_init(struct request_queue *q, struct device *dev);
1257 extern int blk_pre_runtime_suspend(struct request_queue *q);
1258 extern void blk_post_runtime_suspend(struct request_queue *q, int err);
1259 extern void blk_pre_runtime_resume(struct request_queue *q);
1260 extern void blk_post_runtime_resume(struct request_queue *q, int err);
1261 extern void blk_set_runtime_active(struct request_queue *q);
1263 static inline void blk_pm_runtime_init(struct request_queue *q, in blk_pm_runtime_init()
1265 static inline int blk_pre_runtime_suspend(struct request_queue *q) in blk_pre_runtime_suspend()
1269 static inline void blk_post_runtime_suspend(struct request_queue *q, int err) {} in blk_post_runtime_suspend()
1270 static inline void blk_pre_runtime_resume(struct request_queue *q) {} in blk_pre_runtime_resume()
1271 static inline void blk_post_runtime_resume(struct request_queue *q, int err) {} in blk_post_runtime_resume()
1272 static inline void blk_set_runtime_active(struct request_queue *q) {} in blk_set_runtime_active()
1337 extern int blk_queue_start_tag(struct request_queue *, struct request *);
1338 extern struct request *blk_queue_find_tag(struct request_queue *, int);
1339 extern void blk_queue_end_tag(struct request_queue *, struct request *);
1340 extern int blk_queue_init_tags(struct request_queue *, int, struct blk_queue_tag *, int);
1341 extern void blk_queue_free_tags(struct request_queue *);
1342 extern int blk_queue_resize_tags(struct request_queue *, int);
1343 extern void blk_queue_invalidate_tags(struct request_queue *);
1404 static inline unsigned long queue_segment_boundary(struct request_queue *q) in queue_segment_boundary()
1409 static inline unsigned long queue_virt_boundary(struct request_queue *q) in queue_virt_boundary()
1414 static inline unsigned int queue_max_sectors(struct request_queue *q) in queue_max_sectors()
1419 static inline unsigned int queue_max_hw_sectors(struct request_queue *q) in queue_max_hw_sectors()
1424 static inline unsigned short queue_max_segments(struct request_queue *q) in queue_max_segments()
1429 static inline unsigned short queue_max_discard_segments(struct request_queue *q) in queue_max_discard_segments()
1434 static inline unsigned int queue_max_segment_size(struct request_queue *q) in queue_max_segment_size()
1439 static inline unsigned queue_logical_block_size(struct request_queue *q) in queue_logical_block_size()
1454 static inline unsigned int queue_physical_block_size(struct request_queue *q) in queue_physical_block_size()
1464 static inline unsigned int queue_io_min(struct request_queue *q) in queue_io_min()
1474 static inline unsigned int queue_io_opt(struct request_queue *q) in queue_io_opt()
1484 static inline int queue_alignment_offset(struct request_queue *q) in queue_alignment_offset()
1502 struct request_queue *q = bdev_get_queue(bdev); in bdev_alignment_offset()
1513 static inline int queue_discard_alignment(struct request_queue *q) in queue_discard_alignment()
1546 struct request_queue *q = bdev_get_queue(bdev); in bdev_discard_alignment()
1556 struct request_queue *q = bdev_get_queue(bdev); in bdev_write_same()
1566 struct request_queue *q = bdev_get_queue(bdev); in bdev_write_zeroes_sectors()
1576 struct request_queue *q = bdev_get_queue(bdev); in bdev_zoned_model()
1586 struct request_queue *q = bdev_get_queue(bdev); in bdev_is_zoned()
1596 struct request_queue *q = bdev_get_queue(bdev); in bdev_zone_sectors()
1604 static inline int queue_dma_alignment(struct request_queue *q) in queue_dma_alignment()
1609 static inline int blk_rq_aligned(struct request_queue *q, unsigned long addr, in blk_rq_aligned()
1632 static inline bool queue_flush_queueable(struct request_queue *q) in queue_flush_queueable()
1646 static inline bool __bvec_gap_to_prev(struct request_queue *q, in __bvec_gap_to_prev()
1657 static inline bool bvec_gap_to_prev(struct request_queue *q, in bvec_gap_to_prev()
1670 static inline bool bios_segs_mergeable(struct request_queue *q, in bios_segs_mergeable()
1684 static inline bool bio_will_gap(struct request_queue *q, in bio_will_gap()
1816 extern int blk_rq_map_integrity_sg(struct request_queue *, struct bio *,
1818 extern int blk_rq_count_integrity_sg(struct request_queue *, struct bio *);
1819 extern bool blk_integrity_merge_rq(struct request_queue *, struct request *,
1821 extern bool blk_integrity_merge_bio(struct request_queue *, struct request *,
1845 static inline void blk_queue_max_integrity_segments(struct request_queue *q, in blk_queue_max_integrity_segments()
1852 queue_max_integrity_segments(struct request_queue *q) in queue_max_integrity_segments()
1888 static inline int blk_rq_count_integrity_sg(struct request_queue *q, in blk_rq_count_integrity_sg()
1893 static inline int blk_rq_map_integrity_sg(struct request_queue *q, in blk_rq_map_integrity_sg()
1918 static inline void blk_queue_max_integrity_segments(struct request_queue *q, in blk_queue_max_integrity_segments()
1922 static inline unsigned short queue_max_integrity_segments(struct request_queue *q) in queue_max_integrity_segments()
1926 static inline bool blk_integrity_merge_rq(struct request_queue *rq, in blk_integrity_merge_rq()
1932 static inline bool blk_integrity_merge_bio(struct request_queue *rq, in blk_integrity_merge_bio()