• Home
  • Raw
  • Download

Lines Matching defs:request_queue

404 struct request_queue {  struct
408 struct list_head queue_head;
409 struct request *last_merge;
410 struct elevator_queue *elevator;
411 int nr_rqs[2]; /* # allocated [a]sync rqs */
412 int nr_rqs_elvpriv; /* # allocated rqs w/ elvpriv */
414 atomic_t shared_hctx_restart;
416 struct blk_queue_stats *stats;
417 struct rq_wb *rq_wb;
425 struct request_list root_rl;
427 request_fn_proc *request_fn;
428 make_request_fn *make_request_fn;
429 prep_rq_fn *prep_rq_fn;
430 unprep_rq_fn *unprep_rq_fn;
431 softirq_done_fn *softirq_done_fn;
432 rq_timed_out_fn *rq_timed_out_fn;
433 dma_drain_needed_fn *dma_drain_needed;
434 lld_busy_fn *lld_busy_fn;
436 init_rq_fn *init_rq_fn;
438 exit_rq_fn *exit_rq_fn;
440 void (*initialize_rq_fn)(struct request *rq);
442 const struct blk_mq_ops *mq_ops;
444 unsigned int *mq_map;
447 struct blk_mq_ctx __percpu *queue_ctx;
448 unsigned int nr_queues;
450 unsigned int queue_depth;
453 struct blk_mq_hw_ctx **queue_hw_ctx;
454 unsigned int nr_hw_queues;
459 sector_t end_sector;
460 struct request *boundary_rq;
465 struct delayed_work delay_work;
467 struct backing_dev_info *backing_dev_info;
473 void *queuedata;
478 unsigned long queue_flags;
484 int id;
489 gfp_t bounce_gfp;
496 spinlock_t __queue_lock;
497 spinlock_t *queue_lock;
502 struct kobject kobj;
507 struct kobject mq_kobj;
510 struct blk_integrity integrity;
514 struct device *dev;
515 int rpm_status;
516 unsigned int nr_pending;
522 unsigned long nr_requests; /* Max # of requests */
523 unsigned int nr_congestion_on;
524 unsigned int nr_congestion_off;
525 unsigned int nr_batching;
527 unsigned int dma_drain_size;
528 void *dma_drain_buffer;
529 unsigned int dma_pad_mask;
530 unsigned int dma_alignment;
532 struct blk_queue_tag *queue_tags;
533 struct list_head tag_busy_list;
535 unsigned int nr_sorted;
536 unsigned int in_flight[2];
543 unsigned int request_fn_active;
545 unsigned int rq_timeout;
546 int poll_nsec;
548 struct blk_stat_callback *poll_cb;
549 struct blk_rq_stat poll_stat[BLK_MQ_POLL_STATS_BKTS];
551 struct timer_list timeout;
552 struct work_struct timeout_work;
553 struct list_head timeout_list;
555 struct list_head icq_list;
558 struct blkcg_gq *root_blkg;
559 struct list_head blkg_list;
562 struct queue_limits limits;
567 unsigned int sg_timeout;
568 unsigned int sg_reserved_size;
569 int node;
571 struct blk_trace *blk_trace;
572 struct mutex blk_trace_mutex;
577 struct blk_flush_queue *fq;
579 struct list_head requeue_list;
580 spinlock_t requeue_lock;
581 struct delayed_work requeue_work;
583 struct mutex sysfs_lock;
585 int bypass_depth;
586 atomic_t mq_freeze_depth;
589 bsg_job_fn *bsg_job_fn;
590 struct bsg_class_device bsg_dev;
595 struct throtl_data *td;
597 struct rcu_head rcu_head;
598 wait_queue_head_t mq_freeze_wq;
599 struct percpu_ref q_usage_counter;
600 struct list_head all_q_node;
602 struct blk_mq_tag_set *tag_set;
603 struct list_head tag_set_list;
604 struct bio_set *bio_split;
607 struct dentry *debugfs_dir;
608 struct dentry *sched_debugfs_dir;
611 bool mq_sysfs_init_done;
613 size_t cmd_size;
614 void *rq_alloc_data;
616 struct work_struct release_work;
619 u64 write_hints[BLK_MAX_WRITE_HINTS];