/include/linux/ |
D | blkdev.h | 268 static inline bool blk_rq_is_scsi(struct request *rq) in blk_rq_is_scsi() argument 270 return blk_op_is_scsi(req_op(rq)); in blk_rq_is_scsi() 273 static inline bool blk_rq_is_private(struct request *rq) in blk_rq_is_private() argument 275 return blk_op_is_private(req_op(rq)); in blk_rq_is_private() 278 static inline bool blk_rq_is_passthrough(struct request *rq) in blk_rq_is_passthrough() argument 280 return blk_rq_is_scsi(rq) || blk_rq_is_private(rq); in blk_rq_is_passthrough() 666 #define blk_noretry_request(rq) \ argument 667 ((rq)->cmd_flags & (REQ_FAILFAST_DEV|REQ_FAILFAST_TRANSPORT| \ 677 static inline bool blk_account_rq(struct request *rq) in blk_account_rq() argument 679 return (rq->rq_flags & RQF_STARTED) && !blk_rq_is_passthrough(rq); in blk_account_rq() [all …]
|
D | t10-pi.h | 40 static inline u32 t10_pi_ref_tag(struct request *rq) in t10_pi_ref_tag() argument 42 unsigned int shift = ilog2(queue_logical_block_size(rq->q)); in t10_pi_ref_tag() 45 if (rq->q->integrity.interval_exp) in t10_pi_ref_tag() 46 shift = rq->q->integrity.interval_exp; in t10_pi_ref_tag() 48 return blk_rq_pos(rq) >> (shift - SECTOR_SHIFT) & 0xffffffff; in t10_pi_ref_tag()
|
D | blk-mq.h | 119 struct request *rq; member 202 void (*initialize_rq_fn)(struct request *rq); 222 void (*show_rq)(struct seq_file *m, struct request *rq); 264 void blk_mq_free_request(struct request *rq); 292 u32 blk_mq_unique_tag(struct request *rq); 305 int blk_mq_request_started(struct request *rq); 306 int blk_mq_request_completed(struct request *rq); 307 void blk_mq_start_request(struct request *rq); 308 void blk_mq_end_request(struct request *rq, blk_status_t error); 309 void __blk_mq_end_request(struct request *rq, blk_status_t error); [all …]
|
D | blktrace_api.h | 78 extern void blk_add_driver_data(struct request_queue *q, struct request *rq, 93 # define blk_add_driver_data(q, rq, data, len) do {} while (0) argument 125 static inline sector_t blk_rq_trace_sector(struct request *rq) in blk_rq_trace_sector() argument 131 if (blk_rq_is_passthrough(rq) || blk_rq_pos(rq) == (sector_t)-1) in blk_rq_trace_sector() 133 return blk_rq_pos(rq); in blk_rq_trace_sector() 136 static inline unsigned int blk_rq_trace_nr_sectors(struct request *rq) in blk_rq_trace_nr_sectors() argument 138 return blk_rq_is_passthrough(rq) ? 0 : blk_rq_sectors(rq); in blk_rq_trace_nr_sectors()
|
D | elevator.h | 114 void elv_rqhash_del(struct request_queue *q, struct request *rq); 115 void elv_rqhash_add(struct request_queue *q, struct request *rq); 116 void elv_rqhash_reposition(struct request_queue *q, struct request *rq); 184 #define rq_end_sector(rq) (blk_rq_pos(rq) + blk_rq_sectors(rq)) argument 188 #define rq_fifo_clear(rq) list_del_init(&(rq)->queuelist) argument
|
D | bsg.h | 12 int (*fill_hdr)(struct request *rq, struct sg_io_v4 *hdr, 14 int (*complete_rq)(struct request *rq, struct sg_io_v4 *hdr); 15 void (*free_rq)(struct request *rq);
|
D | ide.h | 56 static inline struct ide_request *ide_req(struct request *rq) in ide_req() argument 58 return blk_mq_rq_to_pdu(rq); in ide_req() 61 static inline bool ata_misc_request(struct request *rq) in ata_misc_request() argument 63 return blk_rq_is_private(rq) && ide_req(rq)->type == ATA_PRIV_MISC; in ata_misc_request() 66 static inline bool ata_taskfile_request(struct request *rq) in ata_taskfile_request() argument 68 return blk_rq_is_private(rq) && ide_req(rq)->type == ATA_PRIV_TASKFILE; in ata_taskfile_request() 71 static inline bool ata_pc_request(struct request *rq) in ata_pc_request() argument 73 return blk_rq_is_private(rq) && ide_req(rq)->type == ATA_PRIV_PC; in ata_pc_request() 76 static inline bool ata_sense_request(struct request *rq) in ata_sense_request() argument 78 return blk_rq_is_private(rq) && ide_req(rq)->type == ATA_PRIV_SENSE; in ata_sense_request() [all …]
|
D | sched.h | 57 struct rq; 2026 const struct sched_avg *sched_trace_rq_avg_rt(struct rq *rq); 2027 const struct sched_avg *sched_trace_rq_avg_dl(struct rq *rq); 2028 const struct sched_avg *sched_trace_rq_avg_irq(struct rq *rq); 2030 int sched_trace_rq_cpu(struct rq *rq);
|
D | blk-cgroup.h | 783 static inline bool blk_cgroup_mergeable(struct request *rq, struct bio *bio) in blk_cgroup_mergeable() argument 785 return rq->bio->bi_blkg == bio->bi_blkg && in blk_cgroup_mergeable() 786 bio_issue_as_root_blkg(rq->bio) == bio_issue_as_root_blkg(bio); in blk_cgroup_mergeable() 889 static inline bool blk_cgroup_mergeable(struct request *rq, struct bio *bio) { return true; } in blk_cgroup_mergeable() argument
|
D | mii.h | 51 static inline struct mii_ioctl_data *if_mii(struct ifreq *rq) in if_mii() argument 53 return (struct mii_ioctl_data *) &rq->ifr_ifru; in if_mii()
|
D | bio-crypt-ctx.h | 143 bool bio_crypt_should_process(struct request *rq);
|
D | device-mapper.h | 62 struct request *rq, 437 union map_info *dm_get_rq_mapinfo(struct request *rq);
|
/include/linux/sched/ |
D | nohz.h | 18 void calc_load_nohz_remote(struct rq *rq); 22 static inline void calc_load_nohz_remote(struct rq *rq) { } in calc_load_nohz_remote() argument
|
/include/trace/hooks/ |
D | sched.h | 26 struct rq; 28 TP_PROTO(struct rq *rq), 29 TP_ARGS(rq)); 32 TP_PROTO(struct rq *rq, struct task_struct *p), 33 TP_ARGS(rq, p), 1); 36 TP_PROTO(struct rq *rq, struct task_struct *p), 37 TP_ARGS(rq, p), 1); 70 TP_PROTO(struct sched_group *busiest, struct rq *dst_rq, int *out_balance),
|
/include/trace/events/ |
D | block.h | 76 TP_PROTO(struct request_queue *q, struct request *rq), 78 TP_ARGS(q, rq), 89 __entry->dev = rq->rq_disk ? disk_devt(rq->rq_disk) : 0; 90 __entry->sector = blk_rq_trace_sector(rq); 91 __entry->nr_sector = blk_rq_trace_nr_sectors(rq); 93 blk_fill_rwbs(__entry->rwbs, rq->cmd_flags, blk_rq_bytes(rq)); 118 TP_PROTO(struct request *rq, int error, unsigned int nr_bytes), 120 TP_ARGS(rq, error, nr_bytes), 132 __entry->dev = rq->rq_disk ? disk_devt(rq->rq_disk) : 0; 133 __entry->sector = blk_rq_pos(rq); [all …]
|
D | nbd.h | 64 struct request *rq), 66 TP_ARGS(nbd_request, index, rq), 77 __entry->request = rq; 97 struct request *rq), 99 TP_ARGS(nbd_request, index, rq),
|
D | sched.h | 633 TP_PROTO(struct rq *rq), 634 TP_ARGS(rq)); 637 TP_PROTO(struct rq *rq), 638 TP_ARGS(rq)); 641 TP_PROTO(struct rq *rq), 642 TP_ARGS(rq));
|
/include/net/ |
D | xdp_sock.h | 125 void xsk_reuseq_free(struct xdp_umem_fq_reuse *rq); 173 struct xdp_umem_fq_reuse *rq = umem->fq_reuse; in xsk_umem_has_addrs_rq() local 175 if (rq->length >= cnt) in xsk_umem_has_addrs_rq() 178 return xsk_umem_has_addrs(umem, cnt - rq->length); in xsk_umem_has_addrs_rq() 183 struct xdp_umem_fq_reuse *rq = umem->fq_reuse; in xsk_umem_peek_addr_rq() local 185 if (!rq->length) in xsk_umem_peek_addr_rq() 188 *addr = rq->handles[rq->length - 1]; in xsk_umem_peek_addr_rq() 194 struct xdp_umem_fq_reuse *rq = umem->fq_reuse; in xsk_umem_discard_addr_rq() local 196 if (!rq->length) in xsk_umem_discard_addr_rq() 199 rq->length--; in xsk_umem_discard_addr_rq() [all …]
|
/include/scsi/ |
D | scsi_request.h | 20 static inline struct scsi_request *scsi_req(struct request *rq) in scsi_req() argument 22 return blk_mq_rq_to_pdu(rq); in scsi_req()
|
/include/linux/sunrpc/ |
D | svcauth.h | 132 int (*accept)(struct svc_rqst *rq, __be32 *authp); 133 int (*release)(struct svc_rqst *rq); 135 int (*set_client)(struct svc_rqst *rq);
|
/include/rdma/ |
D | rdmavt_qp.h | 291 static inline u32 rvt_get_rq_count(struct rvt_rq *rq, u32 head, u32 tail) in rvt_get_rq_count() argument 296 count += rq->size; in rvt_get_rq_count() 468 struct rvt_rq rq; member 559 static inline struct rvt_rwqe *rvt_get_rwqe_ptr(struct rvt_rq *rq, unsigned n) in rvt_get_rwqe_ptr() argument 562 ((char *)rq->kwq->curr_wq + in rvt_get_rwqe_ptr() 564 rq->max_sge * sizeof(struct ib_sge)) * n); in rvt_get_rwqe_ptr() 987 static inline void rvt_free_rq(struct rvt_rq *rq) in rvt_free_rq() argument 989 kvfree(rq->kwq); in rvt_free_rq() 990 rq->kwq = NULL; in rvt_free_rq() 991 vfree(rq->wq); in rvt_free_rq() [all …]
|
/include/drm/ |
D | gpu_scheduler.h | 81 struct drm_sched_rq *rq; member 306 void drm_sched_rq_add_entity(struct drm_sched_rq *rq, 308 void drm_sched_rq_remove_entity(struct drm_sched_rq *rq,
|
/include/linux/mtd/ |
D | blktrans.h | 32 struct request_queue *rq; member
|
/include/linux/fsl/ |
D | ptp_qoriq.h | 194 struct ptp_clock_request *rq, int on);
|
/include/linux/mlx5/ |
D | qp.h | 591 struct mlx5_core_qp *rq); 593 struct mlx5_core_qp *rq);
|