/kernel/linux/linux-5.10/drivers/crypto/ |
D | omap-des.c | 47 #define _calc_walked(inout) (dd->inout##_walk.offset - dd->inout##_sg->offset) 49 #define DES_REG_KEY(dd, x) ((dd)->pdata->key_ofs - \ argument 52 #define DES_REG_IV(dd, x) ((dd)->pdata->iv_ofs + ((x) * 0x04)) argument 54 #define DES_REG_CTRL(dd) ((dd)->pdata->ctrl_ofs) argument 61 #define DES_REG_DATA_N(dd, x) ((dd)->pdata->data_ofs + ((x) * 0x04)) argument 63 #define DES_REG_REV(dd) ((dd)->pdata->rev_ofs) argument 65 #define DES_REG_MASK(dd) ((dd)->pdata->mask_ofs) argument 69 #define DES_REG_IRQ_STATUS(dd) ((dd)->pdata->irq_status_ofs) argument 70 #define DES_REG_IRQ_ENABLE(dd) ((dd)->pdata->irq_enable_ofs) argument 87 struct omap_des_dev *dd; member [all …]
|
D | omap-aes.c | 49 #define omap_aes_read(dd, offset) \ argument 52 _read_ret = __raw_readl(dd->io_base + offset); \ 58 inline u32 omap_aes_read(struct omap_aes_dev *dd, u32 offset) in omap_aes_read() argument 60 return __raw_readl(dd->io_base + offset); in omap_aes_read() 65 #define omap_aes_write(dd, offset, value) \ argument 69 __raw_writel(value, dd->io_base + offset); \ 72 inline void omap_aes_write(struct omap_aes_dev *dd, u32 offset, in omap_aes_write() argument 75 __raw_writel(value, dd->io_base + offset); in omap_aes_write() 79 static inline void omap_aes_write_mask(struct omap_aes_dev *dd, u32 offset, in omap_aes_write_mask() argument 84 val = omap_aes_read(dd, offset); in omap_aes_write_mask() [all …]
|
D | atmel-tdes.c | 75 struct atmel_tdes_dev *dd; member 181 static inline u32 atmel_tdes_read(struct atmel_tdes_dev *dd, u32 offset) in atmel_tdes_read() argument 183 return readl_relaxed(dd->io_base + offset); in atmel_tdes_read() 186 static inline void atmel_tdes_write(struct atmel_tdes_dev *dd, in atmel_tdes_write() argument 189 writel_relaxed(value, dd->io_base + offset); in atmel_tdes_write() 192 static void atmel_tdes_write_n(struct atmel_tdes_dev *dd, u32 offset, in atmel_tdes_write_n() argument 196 atmel_tdes_write(dd, offset, *value); in atmel_tdes_write_n() 205 if (!ctx->dd) { in atmel_tdes_find_dev() 210 ctx->dd = tdes_dd; in atmel_tdes_find_dev() 212 tdes_dd = ctx->dd; in atmel_tdes_find_dev() [all …]
|
D | atmel-aes.c | 104 struct atmel_aes_dev *dd; member 345 static inline u32 atmel_aes_read(struct atmel_aes_dev *dd, u32 offset) in atmel_aes_read() argument 347 u32 value = readl_relaxed(dd->io_base + offset); in atmel_aes_read() 350 if (dd->flags & AES_FLAGS_DUMP_REG) { in atmel_aes_read() 353 dev_vdbg(dd->dev, "read 0x%08x from %s\n", value, in atmel_aes_read() 361 static inline void atmel_aes_write(struct atmel_aes_dev *dd, in atmel_aes_write() argument 365 if (dd->flags & AES_FLAGS_DUMP_REG) { in atmel_aes_write() 368 dev_vdbg(dd->dev, "write 0x%08x into %s\n", value, in atmel_aes_write() 373 writel_relaxed(value, dd->io_base + offset); in atmel_aes_write() 376 static void atmel_aes_read_n(struct atmel_aes_dev *dd, u32 offset, in atmel_aes_read_n() argument [all …]
|
D | omap-aes-gcm.c | 26 static int omap_aes_gcm_handle_queue(struct omap_aes_dev *dd, 29 static void omap_aes_gcm_finish_req(struct omap_aes_dev *dd, int ret) in omap_aes_gcm_finish_req() argument 31 struct aead_request *req = dd->aead_req; in omap_aes_gcm_finish_req() 33 dd->in_sg = NULL; in omap_aes_gcm_finish_req() 34 dd->out_sg = NULL; in omap_aes_gcm_finish_req() 36 crypto_finalize_aead_request(dd->engine, req, ret); in omap_aes_gcm_finish_req() 38 pm_runtime_mark_last_busy(dd->dev); in omap_aes_gcm_finish_req() 39 pm_runtime_put_autosuspend(dd->dev); in omap_aes_gcm_finish_req() 42 static void omap_aes_gcm_done_task(struct omap_aes_dev *dd) in omap_aes_gcm_done_task() argument 48 alen = ALIGN(dd->assoc_len, AES_BLOCK_SIZE); in omap_aes_gcm_done_task() [all …]
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/qib/ |
D | qib_twsi.c | 67 static void i2c_wait_for_writes(struct qib_devdata *dd) in i2c_wait_for_writes() argument 74 dd->f_gpio_mod(dd, 0, 0, 0); in i2c_wait_for_writes() 89 static void scl_out(struct qib_devdata *dd, u8 bit) in scl_out() argument 95 mask = 1UL << dd->gpio_scl_num; in scl_out() 98 dd->f_gpio_mod(dd, 0, bit ? 0 : mask, mask); in scl_out() 110 if (mask & dd->f_gpio_mod(dd, 0, 0, 0)) in scl_out() 115 qib_dev_err(dd, "SCL interface stuck low > %d uSec\n", in scl_out() 118 i2c_wait_for_writes(dd); in scl_out() 121 static void sda_out(struct qib_devdata *dd, u8 bit) in sda_out() argument 125 mask = 1UL << dd->gpio_sda_num; in sda_out() [all …]
|
D | qib_iba6120.c | 306 static inline u32 qib_read_ureg32(const struct qib_devdata *dd, in qib_read_ureg32() argument 309 if (!dd->kregbase || !(dd->flags & QIB_PRESENT)) in qib_read_ureg32() 312 if (dd->userbase) in qib_read_ureg32() 314 ((char __iomem *)dd->userbase + in qib_read_ureg32() 315 dd->ureg_align * ctxt)); in qib_read_ureg32() 318 (dd->uregbase + in qib_read_ureg32() 319 (char __iomem *)dd->kregbase + in qib_read_ureg32() 320 dd->ureg_align * ctxt)); in qib_read_ureg32() 332 static inline void qib_write_ureg(const struct qib_devdata *dd, in qib_write_ureg() argument 337 if (dd->userbase) in qib_write_ureg() [all …]
|
D | qib_init.c | 102 void qib_set_ctxtcnt(struct qib_devdata *dd) in qib_set_ctxtcnt() argument 105 dd->cfgctxts = dd->first_user_ctxt + num_online_cpus(); in qib_set_ctxtcnt() 106 if (dd->cfgctxts > dd->ctxtcnt) in qib_set_ctxtcnt() 107 dd->cfgctxts = dd->ctxtcnt; in qib_set_ctxtcnt() 108 } else if (qib_cfgctxts < dd->num_pports) in qib_set_ctxtcnt() 109 dd->cfgctxts = dd->ctxtcnt; in qib_set_ctxtcnt() 110 else if (qib_cfgctxts <= dd->ctxtcnt) in qib_set_ctxtcnt() 111 dd->cfgctxts = qib_cfgctxts; in qib_set_ctxtcnt() 113 dd->cfgctxts = dd->ctxtcnt; in qib_set_ctxtcnt() 114 dd->freectxts = (dd->first_user_ctxt > dd->cfgctxts) ? 0 : in qib_set_ctxtcnt() [all …]
|
D | qib_tx.c | 61 void qib_disarm_piobufs(struct qib_devdata *dd, unsigned first, unsigned cnt) in qib_disarm_piobufs() argument 68 spin_lock_irqsave(&dd->pioavail_lock, flags); in qib_disarm_piobufs() 70 __clear_bit(i, dd->pio_need_disarm); in qib_disarm_piobufs() 71 dd->f_sendctrl(dd->pport, QIB_SENDCTRL_DISARM_BUF(i)); in qib_disarm_piobufs() 73 spin_unlock_irqrestore(&dd->pioavail_lock, flags); in qib_disarm_piobufs() 82 struct qib_devdata *dd = rcd->dd; in qib_disarm_piobufs_ifneeded() local 103 spin_lock_irq(&dd->pioavail_lock); in qib_disarm_piobufs_ifneeded() 105 if (__test_and_clear_bit(i, dd->pio_need_disarm)) { in qib_disarm_piobufs_ifneeded() 107 dd->f_sendctrl(rcd->ppd, QIB_SENDCTRL_DISARM_BUF(i)); in qib_disarm_piobufs_ifneeded() 110 spin_unlock_irq(&dd->pioavail_lock); in qib_disarm_piobufs_ifneeded() [all …]
|
D | qib_iba7220.c | 230 static inline u32 qib_read_ureg32(const struct qib_devdata *dd, in qib_read_ureg32() argument 233 if (!dd->kregbase || !(dd->flags & QIB_PRESENT)) in qib_read_ureg32() 236 if (dd->userbase) in qib_read_ureg32() 238 ((char __iomem *)dd->userbase + in qib_read_ureg32() 239 dd->ureg_align * ctxt)); in qib_read_ureg32() 242 (dd->uregbase + in qib_read_ureg32() 243 (char __iomem *)dd->kregbase + in qib_read_ureg32() 244 dd->ureg_align * ctxt)); in qib_read_ureg32() 256 static inline void qib_write_ureg(const struct qib_devdata *dd, in qib_write_ureg() argument 261 if (dd->userbase) in qib_write_ureg() [all …]
|
D | qib_pcie.c | 136 int qib_pcie_ddinit(struct qib_devdata *dd, struct pci_dev *pdev, in qib_pcie_ddinit() argument 142 dd->pcidev = pdev; in qib_pcie_ddinit() 143 pci_set_drvdata(pdev, dd); in qib_pcie_ddinit() 148 dd->kregbase = ioremap(addr, len); in qib_pcie_ddinit() 149 if (!dd->kregbase) in qib_pcie_ddinit() 152 dd->kregend = (u64 __iomem *)((void __iomem *) dd->kregbase + len); in qib_pcie_ddinit() 153 dd->physaddr = addr; /* used for io_remap, etc. */ in qib_pcie_ddinit() 159 dd->pcibar0 = addr; in qib_pcie_ddinit() 160 dd->pcibar1 = addr >> 32; in qib_pcie_ddinit() 161 dd->deviceid = ent->device; /* save for later use */ in qib_pcie_ddinit() [all …]
|
D | qib_sd7220.c | 98 static int qib_sd7220_reg_mod(struct qib_devdata *dd, int sdnum, u32 loc, 100 static int ibsd_mod_allchnls(struct qib_devdata *dd, int loc, int val, 102 static int qib_sd_trimdone_poll(struct qib_devdata *dd); 103 static void qib_sd_trimdone_monitor(struct qib_devdata *dd, const char *where); 104 static int qib_sd_setvals(struct qib_devdata *dd); 105 static int qib_sd_early(struct qib_devdata *dd); 106 static int qib_sd_dactrim(struct qib_devdata *dd); 107 static int qib_internal_presets(struct qib_devdata *dd); 109 static int qib_sd_trimself(struct qib_devdata *dd, int val); 110 static int epb_access(struct qib_devdata *dd, int sdnum, int claim); [all …]
|
D | qib_iba7322.c | 162 #define IS_QMH(dd) (SYM_FIELD((dd)->revision, Revision, BoardID) == \ argument 164 #define IS_QME(dd) (SYM_FIELD((dd)->revision, Revision, BoardID) == \ argument 758 static inline void qib_write_kreg(const struct qib_devdata *dd, 767 static void qib_setup_dca(struct qib_devdata *dd); 768 static void setup_dca_notifier(struct qib_devdata *dd, int msixnum); 769 static void reset_dca_notifier(struct qib_devdata *dd, int msixnum); 782 static inline u32 qib_read_ureg32(const struct qib_devdata *dd, in qib_read_ureg32() argument 785 if (!dd->kregbase || !(dd->flags & QIB_PRESENT)) in qib_read_ureg32() 788 (dd->ureg_align * ctxt) + (dd->userbase ? in qib_read_ureg32() 789 (char __iomem *)dd->userbase : in qib_read_ureg32() [all …]
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/ |
D | pcie.c | 66 int hfi1_pcie_init(struct hfi1_devdata *dd) in hfi1_pcie_init() argument 69 struct pci_dev *pdev = dd->pcidev; in hfi1_pcie_init() 85 dd_dev_err(dd, "pci enable failed: error %d\n", -ret); in hfi1_pcie_init() 91 dd_dev_err(dd, "pci_request_regions fails: err %d\n", -ret); in hfi1_pcie_init() 104 dd_dev_err(dd, "Unable to set DMA mask: %d\n", ret); in hfi1_pcie_init() 112 dd_dev_err(dd, "Unable to set DMA consistent mask: %d\n", ret); in hfi1_pcie_init() 143 int hfi1_pcie_ddinit(struct hfi1_devdata *dd, struct pci_dev *pdev) in hfi1_pcie_ddinit() argument 160 dd_dev_err(dd, "chip PIO range does not match\n"); in hfi1_pcie_ddinit() 164 dd->kregbase1 = ioremap(addr, RCV_ARRAY); in hfi1_pcie_ddinit() 165 if (!dd->kregbase1) { in hfi1_pcie_ddinit() [all …]
|
D | firmware.c | 255 static int load_fabric_serdes_firmware(struct hfi1_devdata *dd, 257 static void dump_fw_version(struct hfi1_devdata *dd); 272 static int __read_8051_data(struct hfi1_devdata *dd, u32 addr, u64 *result) in __read_8051_data() argument 280 write_csr(dd, DC_DC8051_CFG_RAM_ACCESS_CTRL, reg); in __read_8051_data() 282 write_csr(dd, DC_DC8051_CFG_RAM_ACCESS_CTRL, in __read_8051_data() 287 while ((read_csr(dd, DC_DC8051_CFG_RAM_ACCESS_STATUS) in __read_8051_data() 292 dd_dev_err(dd, "timeout reading 8051 data\n"); in __read_8051_data() 299 *result = read_csr(dd, DC_DC8051_CFG_RAM_ACCESS_RD_DATA); in __read_8051_data() 308 int read_8051_data(struct hfi1_devdata *dd, u32 addr, u32 len, u64 *result) in read_8051_data() argument 314 spin_lock_irqsave(&dd->dc8051_memlock, flags); in read_8051_data() [all …]
|
D | init.c | 126 static int hfi1_create_kctxt(struct hfi1_devdata *dd, in hfi1_create_kctxt() argument 135 ret = hfi1_create_ctxtdata(ppd, dd->node, &rcd); in hfi1_create_kctxt() 137 dd_dev_err(dd, "Kernel receive context allocation failed\n"); in hfi1_create_kctxt() 160 rcd->sc = sc_alloc(dd, SC_ACK, rcd->rcvhdrqentsize, dd->node); in hfi1_create_kctxt() 162 dd_dev_err(dd, "Kernel send context allocation failed\n"); in hfi1_create_kctxt() 173 int hfi1_create_kctxts(struct hfi1_devdata *dd) in hfi1_create_kctxts() argument 178 dd->rcd = kcalloc_node(dd->num_rcv_contexts, sizeof(*dd->rcd), in hfi1_create_kctxts() 179 GFP_KERNEL, dd->node); in hfi1_create_kctxts() 180 if (!dd->rcd) in hfi1_create_kctxts() 183 for (i = 0; i < dd->first_dyn_alloc_ctxt; ++i) { in hfi1_create_kctxts() [all …]
|
D | chip.c | 142 #define emulator_rev(dd) ((dd)->irev >> 8) argument 144 #define is_emulator_p(dd) ((((dd)->irev) & 0xf) == 3) argument 145 #define is_emulator_s(dd) ((((dd)->irev) & 0xf) == 4) argument 1044 static u32 chip_to_opa_lstate(struct hfi1_devdata *dd, u32 chip_lstate); 1045 static int set_physical_link_state(struct hfi1_devdata *dd, u64 state); 1046 static void read_vc_remote_phy(struct hfi1_devdata *dd, u8 *power_management, 1048 static void read_vc_remote_fabric(struct hfi1_devdata *dd, u8 *vau, u8 *z, 1050 static void read_vc_remote_link_width(struct hfi1_devdata *dd, 1052 static void read_vc_local_link_mode(struct hfi1_devdata *dd, u8 *misc_bits, 1054 static void read_remote_device_id(struct hfi1_devdata *dd, u16 *device_id, [all …]
|
D | msix.c | 59 int msix_initialize(struct hfi1_devdata *dd) in msix_initialize() argument 73 total = 1 + dd->num_sdma + dd->n_krcv_queues + dd->num_netdev_contexts; in msix_initialize() 78 ret = pci_alloc_irq_vectors(dd->pcidev, total, total, PCI_IRQ_MSIX); in msix_initialize() 80 dd_dev_err(dd, "pci_alloc_irq_vectors() failed: %d\n", ret); in msix_initialize() 84 entries = kcalloc(total, sizeof(*dd->msix_info.msix_entries), in msix_initialize() 87 pci_free_irq_vectors(dd->pcidev); in msix_initialize() 91 dd->msix_info.msix_entries = entries; in msix_initialize() 92 spin_lock_init(&dd->msix_info.msix_lock); in msix_initialize() 93 bitmap_zero(dd->msix_info.in_use_msix, total); in msix_initialize() 94 dd->msix_info.max_requested = total; in msix_initialize() [all …]
|
D | aspm.c | 23 static bool aspm_hw_l1_supported(struct hfi1_devdata *dd) in aspm_hw_l1_supported() argument 25 struct pci_dev *parent = dd->pcidev->bus->self; in aspm_hw_l1_supported() 35 pcie_capability_read_dword(dd->pcidev, PCI_EXP_LNKCAP, &dn); in aspm_hw_l1_supported() 42 return (!!dn || is_ax(dd)) && !!up; in aspm_hw_l1_supported() 46 static void aspm_hw_set_l1_ent_latency(struct hfi1_devdata *dd) in aspm_hw_set_l1_ent_latency() argument 51 pci_read_config_dword(dd->pcidev, PCIE_CFG_REG_PL3, ®32); in aspm_hw_set_l1_ent_latency() 54 pci_write_config_dword(dd->pcidev, PCIE_CFG_REG_PL3, reg32); in aspm_hw_set_l1_ent_latency() 57 static void aspm_hw_enable_l1(struct hfi1_devdata *dd) in aspm_hw_enable_l1() argument 59 struct pci_dev *parent = dd->pcidev->bus->self; in aspm_hw_enable_l1() 72 pcie_capability_clear_and_set_word(dd->pcidev, PCI_EXP_LNKCTL, in aspm_hw_enable_l1() [all …]
|
D | netdev_rx.c | 24 struct hfi1_devdata *dd = priv->dd; in hfi1_netdev_setup_ctxt() local 31 ret = hfi1_create_rcvhdrq(dd, uctxt); in hfi1_netdev_setup_ctxt() 53 hfi1_rcvctrl(uctxt->dd, rcvctrl_ops, uctxt); in hfi1_netdev_setup_ctxt() 58 static int hfi1_netdev_allocate_ctxt(struct hfi1_devdata *dd, in hfi1_netdev_allocate_ctxt() argument 64 if (dd->flags & HFI1_FROZEN) in hfi1_netdev_allocate_ctxt() 67 ret = hfi1_create_ctxtdata(dd->pport, dd->node, &uctxt); in hfi1_netdev_allocate_ctxt() 69 dd_dev_err(dd, "Unable to create ctxtdata, failing open\n"); in hfi1_netdev_allocate_ctxt() 85 dd_dev_info(dd, "created netdev context %d\n", uctxt->ctxt); in hfi1_netdev_allocate_ctxt() 91 static void hfi1_netdev_deallocate_ctxt(struct hfi1_devdata *dd, in hfi1_netdev_deallocate_ctxt() argument 100 hfi1_rcvctrl(dd, HFI1_RCVCTRL_CTXT_DIS | in hfi1_netdev_deallocate_ctxt() [all …]
|
D | pio.c | 63 void __cm_reset(struct hfi1_devdata *dd, u64 sendctrl) in __cm_reset() argument 65 write_csr(dd, SEND_CTRL, sendctrl | SEND_CTRL_CM_RESET_SMASK); in __cm_reset() 68 sendctrl = read_csr(dd, SEND_CTRL); in __cm_reset() 75 void pio_send_control(struct hfi1_devdata *dd, int op) in pio_send_control() argument 83 spin_lock_irqsave(&dd->sendctrl_lock, flags); in pio_send_control() 85 reg = read_csr(dd, SEND_CTRL); in pio_send_control() 92 for (i = 0; i < ARRAY_SIZE(dd->vld); i++) in pio_send_control() 93 if (!dd->vld[i].mtu) in pio_send_control() 110 __cm_reset(dd, reg); in pio_send_control() 118 dd_dev_err(dd, "%s: invalid control %d\n", __func__, op); in pio_send_control() [all …]
|
/kernel/linux/linux-5.10/drivers/clk/ti/ |
D | dpll3xxx.c | 49 const struct dpll_data *dd; in _omap3_dpll_write_clken() local 52 dd = clk->dpll_data; in _omap3_dpll_write_clken() 54 v = ti_clk_ll_ops->clk_readl(&dd->control_reg); in _omap3_dpll_write_clken() 55 v &= ~dd->enable_mask; in _omap3_dpll_write_clken() 56 v |= clken_bits << __ffs(dd->enable_mask); in _omap3_dpll_write_clken() 57 ti_clk_ll_ops->clk_writel(v, &dd->control_reg); in _omap3_dpll_write_clken() 63 const struct dpll_data *dd; in _omap3_wait_dpll_status() local 68 dd = clk->dpll_data; in _omap3_wait_dpll_status() 71 state <<= __ffs(dd->idlest_mask); in _omap3_wait_dpll_status() 73 while (((ti_clk_ll_ops->clk_readl(&dd->idlest_reg) & dd->idlest_mask) in _omap3_wait_dpll_status() [all …]
|
/kernel/linux/linux-5.10/drivers/block/mtip32xx/ |
D | mtip32xx.c | 114 static int mtip_block_initialize(struct driver_data *dd); 142 struct driver_data *dd = pci_get_drvdata(pdev); in mtip_check_surprise_removal() local 144 if (dd->sr) in mtip_check_surprise_removal() 150 dd->sr = true; in mtip_check_surprise_removal() 151 if (dd->queue) in mtip_check_surprise_removal() 152 blk_queue_flag_set(QUEUE_FLAG_DEAD, dd->queue); in mtip_check_surprise_removal() 154 dev_warn(&dd->pdev->dev, in mtip_check_surprise_removal() 162 static struct mtip_cmd *mtip_cmd_from_tag(struct driver_data *dd, in mtip_cmd_from_tag() argument 165 struct blk_mq_hw_ctx *hctx = dd->queue->queue_hw_ctx[0]; in mtip_cmd_from_tag() 179 static int mtip_hba_reset(struct driver_data *dd) in mtip_hba_reset() argument [all …]
|
/kernel/linux/linux-5.10/drivers/md/ |
D | dm-dust.c | 78 static int dust_remove_block(struct dust_device *dd, unsigned long long block) in dust_remove_block() argument 83 spin_lock_irqsave(&dd->dust_lock, flags); in dust_remove_block() 84 bblock = dust_rb_search(&dd->badblocklist, block); in dust_remove_block() 87 if (!dd->quiet_mode) { in dust_remove_block() 91 spin_unlock_irqrestore(&dd->dust_lock, flags); in dust_remove_block() 95 rb_erase(&bblock->node, &dd->badblocklist); in dust_remove_block() 96 dd->badblock_count--; in dust_remove_block() 97 if (!dd->quiet_mode) in dust_remove_block() 100 spin_unlock_irqrestore(&dd->dust_lock, flags); in dust_remove_block() 105 static int dust_add_block(struct dust_device *dd, unsigned long long block, in dust_add_block() argument [all …]
|
/kernel/linux/linux-5.10/block/ |
D | mq-deadline.c | 68 deadline_rb_root(struct deadline_data *dd, struct request *rq) in deadline_rb_root() argument 70 return &dd->sort_list[rq_data_dir(rq)]; in deadline_rb_root() 88 deadline_add_rq_rb(struct deadline_data *dd, struct request *rq) in deadline_add_rq_rb() argument 90 struct rb_root *root = deadline_rb_root(dd, rq); in deadline_add_rq_rb() 96 deadline_del_rq_rb(struct deadline_data *dd, struct request *rq) in deadline_del_rq_rb() argument 100 if (dd->next_rq[data_dir] == rq) in deadline_del_rq_rb() 101 dd->next_rq[data_dir] = deadline_latter_request(rq); in deadline_del_rq_rb() 103 elv_rb_del(deadline_rb_root(dd, rq), rq); in deadline_del_rq_rb() 111 struct deadline_data *dd = q->elevator->elevator_data; in deadline_remove_request() local 119 deadline_del_rq_rb(dd, rq); in deadline_remove_request() [all …]
|