Home
last modified time | relevance | path

Searched refs:dd (Results 1 – 25 of 271) sorted by relevance

1234567891011

/kernel/linux/linux-5.10/drivers/crypto/
Domap-des.c47 #define _calc_walked(inout) (dd->inout##_walk.offset - dd->inout##_sg->offset)
49 #define DES_REG_KEY(dd, x) ((dd)->pdata->key_ofs - \ argument
52 #define DES_REG_IV(dd, x) ((dd)->pdata->iv_ofs + ((x) * 0x04)) argument
54 #define DES_REG_CTRL(dd) ((dd)->pdata->ctrl_ofs) argument
61 #define DES_REG_DATA_N(dd, x) ((dd)->pdata->data_ofs + ((x) * 0x04)) argument
63 #define DES_REG_REV(dd) ((dd)->pdata->rev_ofs) argument
65 #define DES_REG_MASK(dd) ((dd)->pdata->mask_ofs) argument
69 #define DES_REG_IRQ_STATUS(dd) ((dd)->pdata->irq_status_ofs) argument
70 #define DES_REG_IRQ_ENABLE(dd) ((dd)->pdata->irq_enable_ofs) argument
87 struct omap_des_dev *dd; member
[all …]
Domap-aes.c49 #define omap_aes_read(dd, offset) \ argument
52 _read_ret = __raw_readl(dd->io_base + offset); \
58 inline u32 omap_aes_read(struct omap_aes_dev *dd, u32 offset) in omap_aes_read() argument
60 return __raw_readl(dd->io_base + offset); in omap_aes_read()
65 #define omap_aes_write(dd, offset, value) \ argument
69 __raw_writel(value, dd->io_base + offset); \
72 inline void omap_aes_write(struct omap_aes_dev *dd, u32 offset, in omap_aes_write() argument
75 __raw_writel(value, dd->io_base + offset); in omap_aes_write()
79 static inline void omap_aes_write_mask(struct omap_aes_dev *dd, u32 offset, in omap_aes_write_mask() argument
84 val = omap_aes_read(dd, offset); in omap_aes_write_mask()
[all …]
Datmel-tdes.c75 struct atmel_tdes_dev *dd; member
181 static inline u32 atmel_tdes_read(struct atmel_tdes_dev *dd, u32 offset) in atmel_tdes_read() argument
183 return readl_relaxed(dd->io_base + offset); in atmel_tdes_read()
186 static inline void atmel_tdes_write(struct atmel_tdes_dev *dd, in atmel_tdes_write() argument
189 writel_relaxed(value, dd->io_base + offset); in atmel_tdes_write()
192 static void atmel_tdes_write_n(struct atmel_tdes_dev *dd, u32 offset, in atmel_tdes_write_n() argument
196 atmel_tdes_write(dd, offset, *value); in atmel_tdes_write_n()
205 if (!ctx->dd) { in atmel_tdes_find_dev()
210 ctx->dd = tdes_dd; in atmel_tdes_find_dev()
212 tdes_dd = ctx->dd; in atmel_tdes_find_dev()
[all …]
Datmel-aes.c104 struct atmel_aes_dev *dd; member
345 static inline u32 atmel_aes_read(struct atmel_aes_dev *dd, u32 offset) in atmel_aes_read() argument
347 u32 value = readl_relaxed(dd->io_base + offset); in atmel_aes_read()
350 if (dd->flags & AES_FLAGS_DUMP_REG) { in atmel_aes_read()
353 dev_vdbg(dd->dev, "read 0x%08x from %s\n", value, in atmel_aes_read()
361 static inline void atmel_aes_write(struct atmel_aes_dev *dd, in atmel_aes_write() argument
365 if (dd->flags & AES_FLAGS_DUMP_REG) { in atmel_aes_write()
368 dev_vdbg(dd->dev, "write 0x%08x into %s\n", value, in atmel_aes_write()
373 writel_relaxed(value, dd->io_base + offset); in atmel_aes_write()
376 static void atmel_aes_read_n(struct atmel_aes_dev *dd, u32 offset, in atmel_aes_read_n() argument
[all …]
Domap-aes-gcm.c26 static int omap_aes_gcm_handle_queue(struct omap_aes_dev *dd,
29 static void omap_aes_gcm_finish_req(struct omap_aes_dev *dd, int ret) in omap_aes_gcm_finish_req() argument
31 struct aead_request *req = dd->aead_req; in omap_aes_gcm_finish_req()
33 dd->in_sg = NULL; in omap_aes_gcm_finish_req()
34 dd->out_sg = NULL; in omap_aes_gcm_finish_req()
36 crypto_finalize_aead_request(dd->engine, req, ret); in omap_aes_gcm_finish_req()
38 pm_runtime_mark_last_busy(dd->dev); in omap_aes_gcm_finish_req()
39 pm_runtime_put_autosuspend(dd->dev); in omap_aes_gcm_finish_req()
42 static void omap_aes_gcm_done_task(struct omap_aes_dev *dd) in omap_aes_gcm_done_task() argument
48 alen = ALIGN(dd->assoc_len, AES_BLOCK_SIZE); in omap_aes_gcm_done_task()
[all …]
/kernel/linux/linux-5.10/drivers/infiniband/hw/qib/
Dqib_twsi.c67 static void i2c_wait_for_writes(struct qib_devdata *dd) in i2c_wait_for_writes() argument
74 dd->f_gpio_mod(dd, 0, 0, 0); in i2c_wait_for_writes()
89 static void scl_out(struct qib_devdata *dd, u8 bit) in scl_out() argument
95 mask = 1UL << dd->gpio_scl_num; in scl_out()
98 dd->f_gpio_mod(dd, 0, bit ? 0 : mask, mask); in scl_out()
110 if (mask & dd->f_gpio_mod(dd, 0, 0, 0)) in scl_out()
115 qib_dev_err(dd, "SCL interface stuck low > %d uSec\n", in scl_out()
118 i2c_wait_for_writes(dd); in scl_out()
121 static void sda_out(struct qib_devdata *dd, u8 bit) in sda_out() argument
125 mask = 1UL << dd->gpio_sda_num; in sda_out()
[all …]
Dqib_iba6120.c306 static inline u32 qib_read_ureg32(const struct qib_devdata *dd, in qib_read_ureg32() argument
309 if (!dd->kregbase || !(dd->flags & QIB_PRESENT)) in qib_read_ureg32()
312 if (dd->userbase) in qib_read_ureg32()
314 ((char __iomem *)dd->userbase + in qib_read_ureg32()
315 dd->ureg_align * ctxt)); in qib_read_ureg32()
318 (dd->uregbase + in qib_read_ureg32()
319 (char __iomem *)dd->kregbase + in qib_read_ureg32()
320 dd->ureg_align * ctxt)); in qib_read_ureg32()
332 static inline void qib_write_ureg(const struct qib_devdata *dd, in qib_write_ureg() argument
337 if (dd->userbase) in qib_write_ureg()
[all …]
Dqib_init.c102 void qib_set_ctxtcnt(struct qib_devdata *dd) in qib_set_ctxtcnt() argument
105 dd->cfgctxts = dd->first_user_ctxt + num_online_cpus(); in qib_set_ctxtcnt()
106 if (dd->cfgctxts > dd->ctxtcnt) in qib_set_ctxtcnt()
107 dd->cfgctxts = dd->ctxtcnt; in qib_set_ctxtcnt()
108 } else if (qib_cfgctxts < dd->num_pports) in qib_set_ctxtcnt()
109 dd->cfgctxts = dd->ctxtcnt; in qib_set_ctxtcnt()
110 else if (qib_cfgctxts <= dd->ctxtcnt) in qib_set_ctxtcnt()
111 dd->cfgctxts = qib_cfgctxts; in qib_set_ctxtcnt()
113 dd->cfgctxts = dd->ctxtcnt; in qib_set_ctxtcnt()
114 dd->freectxts = (dd->first_user_ctxt > dd->cfgctxts) ? 0 : in qib_set_ctxtcnt()
[all …]
Dqib_tx.c61 void qib_disarm_piobufs(struct qib_devdata *dd, unsigned first, unsigned cnt) in qib_disarm_piobufs() argument
68 spin_lock_irqsave(&dd->pioavail_lock, flags); in qib_disarm_piobufs()
70 __clear_bit(i, dd->pio_need_disarm); in qib_disarm_piobufs()
71 dd->f_sendctrl(dd->pport, QIB_SENDCTRL_DISARM_BUF(i)); in qib_disarm_piobufs()
73 spin_unlock_irqrestore(&dd->pioavail_lock, flags); in qib_disarm_piobufs()
82 struct qib_devdata *dd = rcd->dd; in qib_disarm_piobufs_ifneeded() local
103 spin_lock_irq(&dd->pioavail_lock); in qib_disarm_piobufs_ifneeded()
105 if (__test_and_clear_bit(i, dd->pio_need_disarm)) { in qib_disarm_piobufs_ifneeded()
107 dd->f_sendctrl(rcd->ppd, QIB_SENDCTRL_DISARM_BUF(i)); in qib_disarm_piobufs_ifneeded()
110 spin_unlock_irq(&dd->pioavail_lock); in qib_disarm_piobufs_ifneeded()
[all …]
Dqib_iba7220.c230 static inline u32 qib_read_ureg32(const struct qib_devdata *dd, in qib_read_ureg32() argument
233 if (!dd->kregbase || !(dd->flags & QIB_PRESENT)) in qib_read_ureg32()
236 if (dd->userbase) in qib_read_ureg32()
238 ((char __iomem *)dd->userbase + in qib_read_ureg32()
239 dd->ureg_align * ctxt)); in qib_read_ureg32()
242 (dd->uregbase + in qib_read_ureg32()
243 (char __iomem *)dd->kregbase + in qib_read_ureg32()
244 dd->ureg_align * ctxt)); in qib_read_ureg32()
256 static inline void qib_write_ureg(const struct qib_devdata *dd, in qib_write_ureg() argument
261 if (dd->userbase) in qib_write_ureg()
[all …]
Dqib_pcie.c136 int qib_pcie_ddinit(struct qib_devdata *dd, struct pci_dev *pdev, in qib_pcie_ddinit() argument
142 dd->pcidev = pdev; in qib_pcie_ddinit()
143 pci_set_drvdata(pdev, dd); in qib_pcie_ddinit()
148 dd->kregbase = ioremap(addr, len); in qib_pcie_ddinit()
149 if (!dd->kregbase) in qib_pcie_ddinit()
152 dd->kregend = (u64 __iomem *)((void __iomem *) dd->kregbase + len); in qib_pcie_ddinit()
153 dd->physaddr = addr; /* used for io_remap, etc. */ in qib_pcie_ddinit()
159 dd->pcibar0 = addr; in qib_pcie_ddinit()
160 dd->pcibar1 = addr >> 32; in qib_pcie_ddinit()
161 dd->deviceid = ent->device; /* save for later use */ in qib_pcie_ddinit()
[all …]
Dqib_sd7220.c98 static int qib_sd7220_reg_mod(struct qib_devdata *dd, int sdnum, u32 loc,
100 static int ibsd_mod_allchnls(struct qib_devdata *dd, int loc, int val,
102 static int qib_sd_trimdone_poll(struct qib_devdata *dd);
103 static void qib_sd_trimdone_monitor(struct qib_devdata *dd, const char *where);
104 static int qib_sd_setvals(struct qib_devdata *dd);
105 static int qib_sd_early(struct qib_devdata *dd);
106 static int qib_sd_dactrim(struct qib_devdata *dd);
107 static int qib_internal_presets(struct qib_devdata *dd);
109 static int qib_sd_trimself(struct qib_devdata *dd, int val);
110 static int epb_access(struct qib_devdata *dd, int sdnum, int claim);
[all …]
Dqib_iba7322.c162 #define IS_QMH(dd) (SYM_FIELD((dd)->revision, Revision, BoardID) == \ argument
164 #define IS_QME(dd) (SYM_FIELD((dd)->revision, Revision, BoardID) == \ argument
758 static inline void qib_write_kreg(const struct qib_devdata *dd,
767 static void qib_setup_dca(struct qib_devdata *dd);
768 static void setup_dca_notifier(struct qib_devdata *dd, int msixnum);
769 static void reset_dca_notifier(struct qib_devdata *dd, int msixnum);
782 static inline u32 qib_read_ureg32(const struct qib_devdata *dd, in qib_read_ureg32() argument
785 if (!dd->kregbase || !(dd->flags & QIB_PRESENT)) in qib_read_ureg32()
788 (dd->ureg_align * ctxt) + (dd->userbase ? in qib_read_ureg32()
789 (char __iomem *)dd->userbase : in qib_read_ureg32()
[all …]
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/
Dpcie.c66 int hfi1_pcie_init(struct hfi1_devdata *dd) in hfi1_pcie_init() argument
69 struct pci_dev *pdev = dd->pcidev; in hfi1_pcie_init()
85 dd_dev_err(dd, "pci enable failed: error %d\n", -ret); in hfi1_pcie_init()
91 dd_dev_err(dd, "pci_request_regions fails: err %d\n", -ret); in hfi1_pcie_init()
104 dd_dev_err(dd, "Unable to set DMA mask: %d\n", ret); in hfi1_pcie_init()
112 dd_dev_err(dd, "Unable to set DMA consistent mask: %d\n", ret); in hfi1_pcie_init()
143 int hfi1_pcie_ddinit(struct hfi1_devdata *dd, struct pci_dev *pdev) in hfi1_pcie_ddinit() argument
160 dd_dev_err(dd, "chip PIO range does not match\n"); in hfi1_pcie_ddinit()
164 dd->kregbase1 = ioremap(addr, RCV_ARRAY); in hfi1_pcie_ddinit()
165 if (!dd->kregbase1) { in hfi1_pcie_ddinit()
[all …]
Dfirmware.c255 static int load_fabric_serdes_firmware(struct hfi1_devdata *dd,
257 static void dump_fw_version(struct hfi1_devdata *dd);
272 static int __read_8051_data(struct hfi1_devdata *dd, u32 addr, u64 *result) in __read_8051_data() argument
280 write_csr(dd, DC_DC8051_CFG_RAM_ACCESS_CTRL, reg); in __read_8051_data()
282 write_csr(dd, DC_DC8051_CFG_RAM_ACCESS_CTRL, in __read_8051_data()
287 while ((read_csr(dd, DC_DC8051_CFG_RAM_ACCESS_STATUS) in __read_8051_data()
292 dd_dev_err(dd, "timeout reading 8051 data\n"); in __read_8051_data()
299 *result = read_csr(dd, DC_DC8051_CFG_RAM_ACCESS_RD_DATA); in __read_8051_data()
308 int read_8051_data(struct hfi1_devdata *dd, u32 addr, u32 len, u64 *result) in read_8051_data() argument
314 spin_lock_irqsave(&dd->dc8051_memlock, flags); in read_8051_data()
[all …]
Dinit.c126 static int hfi1_create_kctxt(struct hfi1_devdata *dd, in hfi1_create_kctxt() argument
135 ret = hfi1_create_ctxtdata(ppd, dd->node, &rcd); in hfi1_create_kctxt()
137 dd_dev_err(dd, "Kernel receive context allocation failed\n"); in hfi1_create_kctxt()
160 rcd->sc = sc_alloc(dd, SC_ACK, rcd->rcvhdrqentsize, dd->node); in hfi1_create_kctxt()
162 dd_dev_err(dd, "Kernel send context allocation failed\n"); in hfi1_create_kctxt()
173 int hfi1_create_kctxts(struct hfi1_devdata *dd) in hfi1_create_kctxts() argument
178 dd->rcd = kcalloc_node(dd->num_rcv_contexts, sizeof(*dd->rcd), in hfi1_create_kctxts()
179 GFP_KERNEL, dd->node); in hfi1_create_kctxts()
180 if (!dd->rcd) in hfi1_create_kctxts()
183 for (i = 0; i < dd->first_dyn_alloc_ctxt; ++i) { in hfi1_create_kctxts()
[all …]
Dchip.c142 #define emulator_rev(dd) ((dd)->irev >> 8) argument
144 #define is_emulator_p(dd) ((((dd)->irev) & 0xf) == 3) argument
145 #define is_emulator_s(dd) ((((dd)->irev) & 0xf) == 4) argument
1044 static u32 chip_to_opa_lstate(struct hfi1_devdata *dd, u32 chip_lstate);
1045 static int set_physical_link_state(struct hfi1_devdata *dd, u64 state);
1046 static void read_vc_remote_phy(struct hfi1_devdata *dd, u8 *power_management,
1048 static void read_vc_remote_fabric(struct hfi1_devdata *dd, u8 *vau, u8 *z,
1050 static void read_vc_remote_link_width(struct hfi1_devdata *dd,
1052 static void read_vc_local_link_mode(struct hfi1_devdata *dd, u8 *misc_bits,
1054 static void read_remote_device_id(struct hfi1_devdata *dd, u16 *device_id,
[all …]
Dmsix.c59 int msix_initialize(struct hfi1_devdata *dd) in msix_initialize() argument
73 total = 1 + dd->num_sdma + dd->n_krcv_queues + dd->num_netdev_contexts; in msix_initialize()
78 ret = pci_alloc_irq_vectors(dd->pcidev, total, total, PCI_IRQ_MSIX); in msix_initialize()
80 dd_dev_err(dd, "pci_alloc_irq_vectors() failed: %d\n", ret); in msix_initialize()
84 entries = kcalloc(total, sizeof(*dd->msix_info.msix_entries), in msix_initialize()
87 pci_free_irq_vectors(dd->pcidev); in msix_initialize()
91 dd->msix_info.msix_entries = entries; in msix_initialize()
92 spin_lock_init(&dd->msix_info.msix_lock); in msix_initialize()
93 bitmap_zero(dd->msix_info.in_use_msix, total); in msix_initialize()
94 dd->msix_info.max_requested = total; in msix_initialize()
[all …]
Daspm.c23 static bool aspm_hw_l1_supported(struct hfi1_devdata *dd) in aspm_hw_l1_supported() argument
25 struct pci_dev *parent = dd->pcidev->bus->self; in aspm_hw_l1_supported()
35 pcie_capability_read_dword(dd->pcidev, PCI_EXP_LNKCAP, &dn); in aspm_hw_l1_supported()
42 return (!!dn || is_ax(dd)) && !!up; in aspm_hw_l1_supported()
46 static void aspm_hw_set_l1_ent_latency(struct hfi1_devdata *dd) in aspm_hw_set_l1_ent_latency() argument
51 pci_read_config_dword(dd->pcidev, PCIE_CFG_REG_PL3, &reg32); in aspm_hw_set_l1_ent_latency()
54 pci_write_config_dword(dd->pcidev, PCIE_CFG_REG_PL3, reg32); in aspm_hw_set_l1_ent_latency()
57 static void aspm_hw_enable_l1(struct hfi1_devdata *dd) in aspm_hw_enable_l1() argument
59 struct pci_dev *parent = dd->pcidev->bus->self; in aspm_hw_enable_l1()
72 pcie_capability_clear_and_set_word(dd->pcidev, PCI_EXP_LNKCTL, in aspm_hw_enable_l1()
[all …]
Dnetdev_rx.c24 struct hfi1_devdata *dd = priv->dd; in hfi1_netdev_setup_ctxt() local
31 ret = hfi1_create_rcvhdrq(dd, uctxt); in hfi1_netdev_setup_ctxt()
53 hfi1_rcvctrl(uctxt->dd, rcvctrl_ops, uctxt); in hfi1_netdev_setup_ctxt()
58 static int hfi1_netdev_allocate_ctxt(struct hfi1_devdata *dd, in hfi1_netdev_allocate_ctxt() argument
64 if (dd->flags & HFI1_FROZEN) in hfi1_netdev_allocate_ctxt()
67 ret = hfi1_create_ctxtdata(dd->pport, dd->node, &uctxt); in hfi1_netdev_allocate_ctxt()
69 dd_dev_err(dd, "Unable to create ctxtdata, failing open\n"); in hfi1_netdev_allocate_ctxt()
85 dd_dev_info(dd, "created netdev context %d\n", uctxt->ctxt); in hfi1_netdev_allocate_ctxt()
91 static void hfi1_netdev_deallocate_ctxt(struct hfi1_devdata *dd, in hfi1_netdev_deallocate_ctxt() argument
100 hfi1_rcvctrl(dd, HFI1_RCVCTRL_CTXT_DIS | in hfi1_netdev_deallocate_ctxt()
[all …]
Dpio.c63 void __cm_reset(struct hfi1_devdata *dd, u64 sendctrl) in __cm_reset() argument
65 write_csr(dd, SEND_CTRL, sendctrl | SEND_CTRL_CM_RESET_SMASK); in __cm_reset()
68 sendctrl = read_csr(dd, SEND_CTRL); in __cm_reset()
75 void pio_send_control(struct hfi1_devdata *dd, int op) in pio_send_control() argument
83 spin_lock_irqsave(&dd->sendctrl_lock, flags); in pio_send_control()
85 reg = read_csr(dd, SEND_CTRL); in pio_send_control()
92 for (i = 0; i < ARRAY_SIZE(dd->vld); i++) in pio_send_control()
93 if (!dd->vld[i].mtu) in pio_send_control()
110 __cm_reset(dd, reg); in pio_send_control()
118 dd_dev_err(dd, "%s: invalid control %d\n", __func__, op); in pio_send_control()
[all …]
/kernel/linux/linux-5.10/drivers/clk/ti/
Ddpll3xxx.c49 const struct dpll_data *dd; in _omap3_dpll_write_clken() local
52 dd = clk->dpll_data; in _omap3_dpll_write_clken()
54 v = ti_clk_ll_ops->clk_readl(&dd->control_reg); in _omap3_dpll_write_clken()
55 v &= ~dd->enable_mask; in _omap3_dpll_write_clken()
56 v |= clken_bits << __ffs(dd->enable_mask); in _omap3_dpll_write_clken()
57 ti_clk_ll_ops->clk_writel(v, &dd->control_reg); in _omap3_dpll_write_clken()
63 const struct dpll_data *dd; in _omap3_wait_dpll_status() local
68 dd = clk->dpll_data; in _omap3_wait_dpll_status()
71 state <<= __ffs(dd->idlest_mask); in _omap3_wait_dpll_status()
73 while (((ti_clk_ll_ops->clk_readl(&dd->idlest_reg) & dd->idlest_mask) in _omap3_wait_dpll_status()
[all …]
/kernel/linux/linux-5.10/drivers/block/mtip32xx/
Dmtip32xx.c114 static int mtip_block_initialize(struct driver_data *dd);
142 struct driver_data *dd = pci_get_drvdata(pdev); in mtip_check_surprise_removal() local
144 if (dd->sr) in mtip_check_surprise_removal()
150 dd->sr = true; in mtip_check_surprise_removal()
151 if (dd->queue) in mtip_check_surprise_removal()
152 blk_queue_flag_set(QUEUE_FLAG_DEAD, dd->queue); in mtip_check_surprise_removal()
154 dev_warn(&dd->pdev->dev, in mtip_check_surprise_removal()
162 static struct mtip_cmd *mtip_cmd_from_tag(struct driver_data *dd, in mtip_cmd_from_tag() argument
165 struct blk_mq_hw_ctx *hctx = dd->queue->queue_hw_ctx[0]; in mtip_cmd_from_tag()
179 static int mtip_hba_reset(struct driver_data *dd) in mtip_hba_reset() argument
[all …]
/kernel/linux/linux-5.10/drivers/md/
Ddm-dust.c78 static int dust_remove_block(struct dust_device *dd, unsigned long long block) in dust_remove_block() argument
83 spin_lock_irqsave(&dd->dust_lock, flags); in dust_remove_block()
84 bblock = dust_rb_search(&dd->badblocklist, block); in dust_remove_block()
87 if (!dd->quiet_mode) { in dust_remove_block()
91 spin_unlock_irqrestore(&dd->dust_lock, flags); in dust_remove_block()
95 rb_erase(&bblock->node, &dd->badblocklist); in dust_remove_block()
96 dd->badblock_count--; in dust_remove_block()
97 if (!dd->quiet_mode) in dust_remove_block()
100 spin_unlock_irqrestore(&dd->dust_lock, flags); in dust_remove_block()
105 static int dust_add_block(struct dust_device *dd, unsigned long long block, in dust_add_block() argument
[all …]
/kernel/linux/linux-5.10/block/
Dmq-deadline.c68 deadline_rb_root(struct deadline_data *dd, struct request *rq) in deadline_rb_root() argument
70 return &dd->sort_list[rq_data_dir(rq)]; in deadline_rb_root()
88 deadline_add_rq_rb(struct deadline_data *dd, struct request *rq) in deadline_add_rq_rb() argument
90 struct rb_root *root = deadline_rb_root(dd, rq); in deadline_add_rq_rb()
96 deadline_del_rq_rb(struct deadline_data *dd, struct request *rq) in deadline_del_rq_rb() argument
100 if (dd->next_rq[data_dir] == rq) in deadline_del_rq_rb()
101 dd->next_rq[data_dir] = deadline_latter_request(rq); in deadline_del_rq_rb()
103 elv_rb_del(deadline_rb_root(dd, rq), rq); in deadline_del_rq_rb()
111 struct deadline_data *dd = q->elevator->elevator_data; in deadline_remove_request() local
119 deadline_del_rq_rb(dd, rq); in deadline_remove_request()
[all …]

1234567891011