/drivers/net/ethernet/huawei/hinic/ |
D | hinic_hw_eqs.c | 35 #define GET_EQ_NUM_PAGES(eq, pg_size) \ argument 36 (ALIGN((eq)->q_len * (eq)->elem_size, pg_size) / (pg_size)) 38 #define GET_EQ_NUM_ELEMS_IN_PG(eq, pg_size) ((pg_size) / (eq)->elem_size) argument 40 #define EQ_CONS_IDX_REG_ADDR(eq) (((eq)->type == HINIC_AEQ) ? \ argument 41 HINIC_CSR_AEQ_CONS_IDX_ADDR((eq)->q_id) : \ 42 HINIC_CSR_CEQ_CONS_IDX_ADDR((eq)->q_id)) 44 #define EQ_PROD_IDX_REG_ADDR(eq) (((eq)->type == HINIC_AEQ) ? \ argument 45 HINIC_CSR_AEQ_PROD_IDX_ADDR((eq)->q_id) : \ 46 HINIC_CSR_CEQ_PROD_IDX_ADDR((eq)->q_id)) 48 #define EQ_HI_PHYS_ADDR_REG(eq, pg_num) (((eq)->type == HINIC_AEQ) ? \ argument [all …]
|
/drivers/staging/lustre/lnet/lnet/ |
D | lib-eq.c | 70 struct lnet_eq *eq; in LNetEQAlloc() local 92 eq = lnet_eq_alloc(); in LNetEQAlloc() 93 if (!eq) in LNetEQAlloc() 97 LIBCFS_ALLOC(eq->eq_events, count * sizeof(struct lnet_event)); in LNetEQAlloc() 98 if (!eq->eq_events) in LNetEQAlloc() 106 eq->eq_deq_seq = 1; in LNetEQAlloc() 107 eq->eq_enq_seq = 1; in LNetEQAlloc() 108 eq->eq_size = count; in LNetEQAlloc() 109 eq->eq_callback = callback; in LNetEQAlloc() 111 eq->eq_refs = cfs_percpt_alloc(lnet_cpt_table(), in LNetEQAlloc() [all …]
|
/drivers/infiniband/hw/hns/ |
D | hns_roce_eq.c | 39 static void eq_set_cons_index(struct hns_roce_eq *eq, int req_not) in eq_set_cons_index() argument 41 roce_raw_write((eq->cons_index & CONS_INDEX_MASK) | in eq_set_cons_index() 42 (req_not << eq->log_entries), eq->doorbell); in eq_set_cons_index() 47 static struct hns_roce_aeqe *get_aeqe(struct hns_roce_eq *eq, u32 entry) in get_aeqe() argument 49 unsigned long off = (entry & (eq->entries - 1)) * in get_aeqe() 53 (eq->buf_list[off / HNS_ROCE_BA_SIZE].buf) + in get_aeqe() 57 static struct hns_roce_aeqe *next_aeqe_sw(struct hns_roce_eq *eq) in next_aeqe_sw() argument 59 struct hns_roce_aeqe *aeqe = get_aeqe(eq, eq->cons_index); in next_aeqe_sw() 62 !!(eq->cons_index & eq->entries)) ? aeqe : NULL; in next_aeqe_sw() 227 static int hns_roce_aeq_int(struct hns_roce_dev *hr_dev, struct hns_roce_eq *eq) in hns_roce_aeq_int() argument [all …]
|
/drivers/net/ethernet/mellanox/mlx5/core/ |
D | eq.c | 97 static struct mlx5_eqe *get_eqe(struct mlx5_eq *eq, u32 entry) in get_eqe() argument 99 return mlx5_buf_offset(&eq->buf, entry * MLX5_EQE_SIZE); in get_eqe() 102 static struct mlx5_eqe *next_eqe_sw(struct mlx5_eq *eq) in next_eqe_sw() argument 104 struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & (eq->nent - 1)); in next_eqe_sw() 106 return ((eqe->owner & 1) ^ !!(eq->cons_index & eq->nent)) ? NULL : eqe; in next_eqe_sw() 190 static void eq_update_ci(struct mlx5_eq *eq, int arm) in eq_update_ci() argument 192 __be32 __iomem *addr = eq->doorbell + (arm ? 0 : 2); in eq_update_ci() 193 u32 val = (eq->cons_index & 0xffffff) | (eq->eqn << 24); in eq_update_ci() 206 struct mlx5_eq *eq = pfault->eq; in eqe_pf_action() local 208 mlx5_core_page_fault(eq->dev, pfault); in eqe_pf_action() [all …]
|
D | debugfs.c | 361 static u64 eq_read_field(struct mlx5_core_dev *dev, struct mlx5_eq *eq, in eq_read_field() argument 374 err = mlx5_core_eq_query(dev, eq, out, outlen); in eq_read_field() 563 int mlx5_debug_eq_add(struct mlx5_core_dev *dev, struct mlx5_eq *eq) in mlx5_debug_eq_add() argument 571 &eq->dbg, eq->eqn, eq_fields, in mlx5_debug_eq_add() 572 ARRAY_SIZE(eq_fields), eq); in mlx5_debug_eq_add() 574 eq->dbg = NULL; in mlx5_debug_eq_add() 579 void mlx5_debug_eq_remove(struct mlx5_core_dev *dev, struct mlx5_eq *eq) in mlx5_debug_eq_remove() argument 584 if (eq->dbg) in mlx5_debug_eq_remove() 585 rem_res_tree(eq->dbg); in mlx5_debug_eq_remove()
|
D | main.c | 662 struct mlx5_eq *eq, *n; in mlx5_vector2eqn() local 666 list_for_each_entry_safe(eq, n, &table->comp_eqs_list, list) { in mlx5_vector2eqn() 667 if (eq->index == vector) { in mlx5_vector2eqn() 668 *eqn = eq->eqn; in mlx5_vector2eqn() 669 *irqn = eq->irqn; in mlx5_vector2eqn() 683 struct mlx5_eq *eq; in mlx5_eqn2eq() local 686 list_for_each_entry(eq, &table->comp_eqs_list, list) in mlx5_eqn2eq() 687 if (eq->eqn == eqn) { in mlx5_eqn2eq() 689 return eq; in mlx5_eqn2eq() 700 struct mlx5_eq *eq, *n; in free_comp_eqs() local [all …]
|
D | cq.c | 146 struct mlx5_eq *eq; in mlx5_core_create_cq() local 149 eq = mlx5_eqn2eq(dev, eqn); in mlx5_core_create_cq() 150 if (IS_ERR(eq)) in mlx5_core_create_cq() 151 return PTR_ERR(eq); in mlx5_core_create_cq() 167 cq->tasklet_ctx.priv = &eq->tasklet_ctx; in mlx5_core_create_cq()
|
/drivers/infiniband/hw/mthca/ |
D | mthca_eq.c | 173 static inline void tavor_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) in tavor_set_eq_ci() argument 184 mthca_write64(MTHCA_EQ_DB_SET_CI | eq->eqn, ci & (eq->nent - 1), in tavor_set_eq_ci() 189 static inline void arbel_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) in arbel_set_eq_ci() argument 194 dev->eq_regs.arbel.eq_set_ci_base + eq->eqn * 8); in arbel_set_eq_ci() 199 static inline void set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) in set_eq_ci() argument 202 arbel_set_eq_ci(dev, eq, ci); in set_eq_ci() 204 tavor_set_eq_ci(dev, eq, ci); in set_eq_ci() 228 static inline struct mthca_eqe *get_eqe(struct mthca_eq *eq, u32 entry) in get_eqe() argument 230 unsigned long off = (entry & (eq->nent - 1)) * MTHCA_EQ_ENTRY_SIZE; in get_eqe() 231 return eq->page_list[off / PAGE_SIZE].buf + off % PAGE_SIZE; in get_eqe() [all …]
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | eq.c | 97 static void eq_set_ci(struct mlx4_eq *eq, int req_not) in eq_set_ci() argument 99 __raw_writel((__force u32) cpu_to_be32((eq->cons_index & 0xffffff) | in eq_set_ci() 101 eq->doorbell); in eq_set_ci() 106 static struct mlx4_eqe *get_eqe(struct mlx4_eq *eq, u32 entry, u8 eqe_factor, in get_eqe() argument 110 unsigned long offset = (entry & (eq->nent - 1)) * eqe_size; in get_eqe() 118 …return eq->page_list[offset / PAGE_SIZE].buf + (offset + (eqe_factor ? MLX4_EQ_ENTRY_SIZE : 0)) % … in get_eqe() 121 static struct mlx4_eqe *next_eqe_sw(struct mlx4_eq *eq, u8 eqe_factor, u8 size) in next_eqe_sw() argument 123 struct mlx4_eqe *eqe = get_eqe(eq, eq->cons_index, eqe_factor, size); in next_eqe_sw() 124 return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe; in next_eqe_sw() 241 struct mlx4_eq *eq = &priv->eq_table.eq[vec]; in mlx4_set_eq_affinity_hint() local [all …]
|
D | cq.c | 331 cq_context->comp_eqn = priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].eqn; in mlx4_cq_alloc() 351 &priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].tasklet_ctx; in mlx4_cq_alloc() 355 cq->irq = priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].irq; in mlx4_cq_alloc() 384 synchronize_irq(priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq->vector)].irq); in mlx4_cq_free() 385 if (priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq->vector)].irq != in mlx4_cq_free() 386 priv->eq_table.eq[MLX4_EQ_ASYNC].irq) in mlx4_cq_free() 387 synchronize_irq(priv->eq_table.eq[MLX4_EQ_ASYNC].irq); in mlx4_cq_free()
|
/drivers/pci/host/ |
D | pcie-iproc-msi.c | 72 unsigned int eq; member 138 unsigned int eq) in iproc_msi_read_reg() argument 142 return readl_relaxed(pcie->base + msi->reg_offsets[eq][reg]); in iproc_msi_read_reg() 147 int eq, u32 val) in iproc_msi_write_reg() argument 151 writel_relaxed(val, pcie->base + msi->reg_offsets[eq][reg]); in iproc_msi_write_reg() 168 static inline unsigned int iproc_msi_eq_offset(struct iproc_msi *msi, u32 eq) in iproc_msi_eq_offset() argument 171 return eq * EQ_MEM_REGION_SIZE; in iproc_msi_eq_offset() 173 return eq * EQ_LEN * sizeof(u32); in iproc_msi_eq_offset() 298 static inline u32 decode_msi_hwirq(struct iproc_msi *msi, u32 eq, u32 head) in decode_msi_hwirq() argument 303 offs = iproc_msi_eq_offset(msi, eq) + head * sizeof(u32); in decode_msi_hwirq() [all …]
|
/drivers/net/ethernet/ibm/ehea/ |
D | ehea_qmr.c | 256 struct ehea_eq *eq; in ehea_create_eq() local 258 eq = kzalloc(sizeof(*eq), GFP_KERNEL); in ehea_create_eq() 259 if (!eq) in ehea_create_eq() 262 eq->adapter = adapter; in ehea_create_eq() 263 eq->attr.type = type; in ehea_create_eq() 264 eq->attr.max_nr_of_eqes = max_nr_of_eqes; in ehea_create_eq() 265 eq->attr.eqe_gen = eqe_gen; in ehea_create_eq() 266 spin_lock_init(&eq->spinlock); in ehea_create_eq() 269 &eq->attr, &eq->fw_handle); in ehea_create_eq() 275 ret = hw_queue_ctor(&eq->hw_queue, eq->attr.nr_pages, in ehea_create_eq() [all …]
|
/drivers/net/ethernet/tile/ |
D | tilepro.c | 201 lepp_queue_t *eq; member 526 static unsigned int tile_net_lepp_grab_comps(lepp_queue_t *eq, in tile_net_lepp_grab_comps() argument 533 unsigned int comp_head = eq->comp_head; in tile_net_lepp_grab_comps() 534 unsigned int comp_busy = eq->comp_busy; in tile_net_lepp_grab_comps() 537 comps[n++] = eq->comps[comp_head]; in tile_net_lepp_grab_comps() 544 eq->comp_head = comp_head; in tile_net_lepp_grab_comps() 557 lepp_queue_t *eq = priv->eq; in tile_net_lepp_free_comps() local 567 eq->comp_busy = eq->comp_tail; in tile_net_lepp_free_comps() 569 n = tile_net_lepp_grab_comps(eq, olds, wanted, 0); in tile_net_lepp_free_comps() 571 pending = (eq->comp_head != eq->comp_tail); in tile_net_lepp_free_comps() [all …]
|
/drivers/clk/spear/ |
D | spear1340_clock.c | 267 {.xscale = 5, .yscale = 122, .eq = 0}, 269 {.xscale = 10, .yscale = 204, .eq = 0}, 271 {.xscale = 4, .yscale = 25, .eq = 0}, 273 {.xscale = 4, .yscale = 21, .eq = 0}, 275 {.xscale = 5, .yscale = 18, .eq = 0}, 277 {.xscale = 2, .yscale = 6, .eq = 0}, 279 {.xscale = 5, .yscale = 12, .eq = 0}, 281 {.xscale = 2, .yscale = 4, .eq = 0}, 283 {.xscale = 5, .yscale = 18, .eq = 1}, 285 {.xscale = 1, .yscale = 3, .eq = 1}, [all …]
|
D | spear1310_clock.c | 255 {.xscale = 10, .yscale = 204, .eq = 0}, /* 12.29 MHz */ 256 {.xscale = 4, .yscale = 21, .eq = 0}, /* 48 MHz */ 257 {.xscale = 2, .yscale = 6, .eq = 0}, /* 83 MHz */ 258 {.xscale = 2, .yscale = 4, .eq = 0}, /* 125 MHz */ 259 {.xscale = 1, .yscale = 3, .eq = 1}, /* 166 MHz */ 260 {.xscale = 1, .yscale = 2, .eq = 1}, /* 250 MHz */ 266 {.xscale = 2, .yscale = 6, .eq = 0}, /* divided by 6 */ 267 {.xscale = 2, .yscale = 4, .eq = 0}, /* divided by 4 */ 268 {.xscale = 1, .yscale = 3, .eq = 1}, /* divided by 3 */ 269 {.xscale = 1, .yscale = 2, .eq = 1}, /* divided by 2 */ [all …]
|
D | spear3xx_clock.c | 109 {.xscale = 1, .yscale = 81, .eq = 0}, /* 2.049 MHz */ 110 {.xscale = 1, .yscale = 59, .eq = 0}, /* 2.822 MHz */ 111 {.xscale = 2, .yscale = 81, .eq = 0}, /* 4.098 MHz */ 112 {.xscale = 3, .yscale = 89, .eq = 0}, /* 5.644 MHz */ 113 {.xscale = 4, .yscale = 81, .eq = 0}, /* 8.197 MHz */ 114 {.xscale = 4, .yscale = 59, .eq = 0}, /* 11.254 MHz */ 115 {.xscale = 2, .yscale = 27, .eq = 0}, /* 12.296 MHz */ 116 {.xscale = 2, .yscale = 8, .eq = 0}, /* 41.5 MHz */ 117 {.xscale = 2, .yscale = 4, .eq = 0}, /* 83 MHz */ 118 {.xscale = 1, .yscale = 2, .eq = 1}, /* 166 MHz */
|
D | clk-aux-synth.c | 49 u8 eq = rtbl[index].eq ? 1 : 2; in aux_calc_rate() local 52 (rtbl[index].yscale * eq)) * 10000; in aux_calc_rate() 115 val |= (rtbl[i].eq & aux->masks->eq_sel_mask) << in clk_aux_set_rate()
|
D | spear6xx_clock.c | 93 {.xscale = 2, .yscale = 27, .eq = 0}, /* 12.296 MHz */ 94 {.xscale = 2, .yscale = 8, .eq = 0}, /* 41.5 MHz */ 95 {.xscale = 2, .yscale = 4, .eq = 0}, /* 83 MHz */ 96 {.xscale = 1, .yscale = 2, .eq = 1}, /* 166 MHz */
|
/drivers/firmware/broadcom/ |
D | bcm47xx_nvram.c | 171 char *var, *value, *end, *eq; in bcm47xx_nvram_getenv() local 187 eq = strchr(var, '='); in bcm47xx_nvram_getenv() 188 if (!eq) in bcm47xx_nvram_getenv() 190 value = eq + 1; in bcm47xx_nvram_getenv() 191 if (eq - var == strlen(name) && in bcm47xx_nvram_getenv() 192 strncmp(var, name, eq - var) == 0) in bcm47xx_nvram_getenv()
|
/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_hw.c | 111 static inline void *ocrdma_get_eqe(struct ocrdma_eq *eq) in ocrdma_get_eqe() argument 113 return eq->q.va + (eq->q.tail * sizeof(struct ocrdma_eqe)); in ocrdma_get_eqe() 116 static inline void ocrdma_eq_inc_tail(struct ocrdma_eq *eq) in ocrdma_eq_inc_tail() argument 118 eq->q.tail = (eq->q.tail + 1) & (OCRDMA_EQ_LEN - 1); in ocrdma_eq_inc_tail() 434 static int ocrdma_mbx_create_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq) in ocrdma_mbx_create_eq() argument 449 ocrdma_build_q_pages(&cmd->pa[0], cmd->num_pages, eq->q.dma, in ocrdma_mbx_create_eq() 454 eq->q.id = rsp->vector_eqid & 0xffff; in ocrdma_mbx_create_eq() 455 eq->vector = (rsp->vector_eqid >> 16) & 0xffff; in ocrdma_mbx_create_eq() 456 eq->q.created = true; in ocrdma_mbx_create_eq() 462 struct ocrdma_eq *eq, u16 q_len) in ocrdma_create_eq() argument [all …]
|
/drivers/staging/lustre/include/linux/lnet/ |
D | lib-lnet.h | 186 struct lnet_eq *eq; in lnet_eq_alloc() local 188 LIBCFS_ALLOC(eq, sizeof(*eq)); in lnet_eq_alloc() 189 return eq; in lnet_eq_alloc() 193 lnet_eq_free(struct lnet_eq *eq) in lnet_eq_free() argument 195 LIBCFS_FREE(eq, sizeof(*eq)); in lnet_eq_free() 283 lnet_eq2handle(struct lnet_handle_eq *handle, struct lnet_eq *eq) in lnet_eq2handle() argument 285 if (!eq) { in lnet_eq2handle() 290 handle->cookie = eq->eq_lh.lh_cookie; in lnet_eq2handle() 496 void lnet_eq_enqueue_event(struct lnet_eq *eq, struct lnet_event *ev);
|
/drivers/infiniband/hw/hfi1/ |
D | pcie.c | 803 static int load_eq_table(struct hfi1_devdata *dd, const u8 eq[11][3], u8 fs, in load_eq_table() 817 c_minus1 = eq[i][PREC] / div; in load_eq_table() 818 c0 = fs - (eq[i][PREC] / div) - (eq[i][POST] / div); in load_eq_table() 819 c_plus1 = eq[i][POST] / div; in load_eq_table() 839 i, (u32)eq[i][0], (u32)eq[i][1], in load_eq_table() 840 (u32)eq[i][2]); in load_eq_table() 1039 const u8 (*eq)[3]; in do_pcie_gen3_transition() local 1173 eq = discrete_preliminary_eq; in do_pcie_gen3_transition() 1183 eq = integrated_preliminary_eq; in do_pcie_gen3_transition() 1194 ret = load_eq_table(dd, eq, fs, div); in do_pcie_gen3_transition()
|
/drivers/scsi/csiostor/ |
D | csio_wr.h | 411 struct csio_eq eq; member 468 #define csio_q_eqid(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.eq.eqid) 472 ((__hw)->wrm.q_arr[(__idx)]->un.eq.physeqid) 479 #define csio_q_eq_wrap(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.eq.wrap)
|
/drivers/scsi/be2iscsi/ |
D | be_main.c | 678 struct be_queue_info *eq; in be_isr_mcc() local 684 eq = &pbe_eq->q; in be_isr_mcc() 687 eqe = queue_tail_node(eq); in be_isr_mcc() 698 queue_tail_inc(eq); in be_isr_mcc() 699 eqe = queue_tail_node(eq); in be_isr_mcc() 704 hwi_ring_eq_db(phba, eq->id, 1, mcc_events, 1, 1); in be_isr_mcc() 717 struct be_queue_info *eq; in be_isr_msix() local 721 eq = &pbe_eq->q; in be_isr_msix() 725 hwi_ring_eq_db(phba, eq->id, 1, 0, 0, 1); in be_isr_msix() 742 struct be_queue_info *eq; in be_isr() local [all …]
|
/drivers/scsi/lpfc/ |
D | lpfc_debugfs.h | 451 struct lpfc_queue *wq, *cq, *eq; in lpfc_debug_dump_cq() local 489 eq = phba->sli4_hba.hba_eq[eqidx]; in lpfc_debug_dump_cq() 495 eqidx, eq->queue_id); in lpfc_debug_dump_cq() 500 eqidx, eq->queue_id); in lpfc_debug_dump_cq()
|