Home
last modified time | relevance | path

Searched refs:eq (Results 1 – 25 of 61) sorted by relevance

123

/drivers/net/ethernet/huawei/hinic/
Dhinic_hw_eqs.c35 #define GET_EQ_NUM_PAGES(eq, pg_size) \ argument
36 (ALIGN((eq)->q_len * (eq)->elem_size, pg_size) / (pg_size))
38 #define GET_EQ_NUM_ELEMS_IN_PG(eq, pg_size) ((pg_size) / (eq)->elem_size) argument
40 #define EQ_CONS_IDX_REG_ADDR(eq) (((eq)->type == HINIC_AEQ) ? \ argument
41 HINIC_CSR_AEQ_CONS_IDX_ADDR((eq)->q_id) : \
42 HINIC_CSR_CEQ_CONS_IDX_ADDR((eq)->q_id))
44 #define EQ_PROD_IDX_REG_ADDR(eq) (((eq)->type == HINIC_AEQ) ? \ argument
45 HINIC_CSR_AEQ_PROD_IDX_ADDR((eq)->q_id) : \
46 HINIC_CSR_CEQ_PROD_IDX_ADDR((eq)->q_id))
48 #define EQ_HI_PHYS_ADDR_REG(eq, pg_num) (((eq)->type == HINIC_AEQ) ? \ argument
[all …]
/drivers/staging/lustre/lnet/lnet/
Dlib-eq.c70 struct lnet_eq *eq; in LNetEQAlloc() local
92 eq = lnet_eq_alloc(); in LNetEQAlloc()
93 if (!eq) in LNetEQAlloc()
97 LIBCFS_ALLOC(eq->eq_events, count * sizeof(struct lnet_event)); in LNetEQAlloc()
98 if (!eq->eq_events) in LNetEQAlloc()
106 eq->eq_deq_seq = 1; in LNetEQAlloc()
107 eq->eq_enq_seq = 1; in LNetEQAlloc()
108 eq->eq_size = count; in LNetEQAlloc()
109 eq->eq_callback = callback; in LNetEQAlloc()
111 eq->eq_refs = cfs_percpt_alloc(lnet_cpt_table(), in LNetEQAlloc()
[all …]
/drivers/infiniband/hw/hns/
Dhns_roce_eq.c39 static void eq_set_cons_index(struct hns_roce_eq *eq, int req_not) in eq_set_cons_index() argument
41 roce_raw_write((eq->cons_index & CONS_INDEX_MASK) | in eq_set_cons_index()
42 (req_not << eq->log_entries), eq->doorbell); in eq_set_cons_index()
47 static struct hns_roce_aeqe *get_aeqe(struct hns_roce_eq *eq, u32 entry) in get_aeqe() argument
49 unsigned long off = (entry & (eq->entries - 1)) * in get_aeqe()
53 (eq->buf_list[off / HNS_ROCE_BA_SIZE].buf) + in get_aeqe()
57 static struct hns_roce_aeqe *next_aeqe_sw(struct hns_roce_eq *eq) in next_aeqe_sw() argument
59 struct hns_roce_aeqe *aeqe = get_aeqe(eq, eq->cons_index); in next_aeqe_sw()
62 !!(eq->cons_index & eq->entries)) ? aeqe : NULL; in next_aeqe_sw()
227 static int hns_roce_aeq_int(struct hns_roce_dev *hr_dev, struct hns_roce_eq *eq) in hns_roce_aeq_int() argument
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/
Deq.c97 static struct mlx5_eqe *get_eqe(struct mlx5_eq *eq, u32 entry) in get_eqe() argument
99 return mlx5_buf_offset(&eq->buf, entry * MLX5_EQE_SIZE); in get_eqe()
102 static struct mlx5_eqe *next_eqe_sw(struct mlx5_eq *eq) in next_eqe_sw() argument
104 struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & (eq->nent - 1)); in next_eqe_sw()
106 return ((eqe->owner & 1) ^ !!(eq->cons_index & eq->nent)) ? NULL : eqe; in next_eqe_sw()
190 static void eq_update_ci(struct mlx5_eq *eq, int arm) in eq_update_ci() argument
192 __be32 __iomem *addr = eq->doorbell + (arm ? 0 : 2); in eq_update_ci()
193 u32 val = (eq->cons_index & 0xffffff) | (eq->eqn << 24); in eq_update_ci()
206 struct mlx5_eq *eq = pfault->eq; in eqe_pf_action() local
208 mlx5_core_page_fault(eq->dev, pfault); in eqe_pf_action()
[all …]
Ddebugfs.c361 static u64 eq_read_field(struct mlx5_core_dev *dev, struct mlx5_eq *eq, in eq_read_field() argument
374 err = mlx5_core_eq_query(dev, eq, out, outlen); in eq_read_field()
563 int mlx5_debug_eq_add(struct mlx5_core_dev *dev, struct mlx5_eq *eq) in mlx5_debug_eq_add() argument
571 &eq->dbg, eq->eqn, eq_fields, in mlx5_debug_eq_add()
572 ARRAY_SIZE(eq_fields), eq); in mlx5_debug_eq_add()
574 eq->dbg = NULL; in mlx5_debug_eq_add()
579 void mlx5_debug_eq_remove(struct mlx5_core_dev *dev, struct mlx5_eq *eq) in mlx5_debug_eq_remove() argument
584 if (eq->dbg) in mlx5_debug_eq_remove()
585 rem_res_tree(eq->dbg); in mlx5_debug_eq_remove()
Dmain.c662 struct mlx5_eq *eq, *n; in mlx5_vector2eqn() local
666 list_for_each_entry_safe(eq, n, &table->comp_eqs_list, list) { in mlx5_vector2eqn()
667 if (eq->index == vector) { in mlx5_vector2eqn()
668 *eqn = eq->eqn; in mlx5_vector2eqn()
669 *irqn = eq->irqn; in mlx5_vector2eqn()
683 struct mlx5_eq *eq; in mlx5_eqn2eq() local
686 list_for_each_entry(eq, &table->comp_eqs_list, list) in mlx5_eqn2eq()
687 if (eq->eqn == eqn) { in mlx5_eqn2eq()
689 return eq; in mlx5_eqn2eq()
700 struct mlx5_eq *eq, *n; in free_comp_eqs() local
[all …]
Dcq.c146 struct mlx5_eq *eq; in mlx5_core_create_cq() local
149 eq = mlx5_eqn2eq(dev, eqn); in mlx5_core_create_cq()
150 if (IS_ERR(eq)) in mlx5_core_create_cq()
151 return PTR_ERR(eq); in mlx5_core_create_cq()
167 cq->tasklet_ctx.priv = &eq->tasklet_ctx; in mlx5_core_create_cq()
/drivers/infiniband/hw/mthca/
Dmthca_eq.c173 static inline void tavor_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) in tavor_set_eq_ci() argument
184 mthca_write64(MTHCA_EQ_DB_SET_CI | eq->eqn, ci & (eq->nent - 1), in tavor_set_eq_ci()
189 static inline void arbel_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) in arbel_set_eq_ci() argument
194 dev->eq_regs.arbel.eq_set_ci_base + eq->eqn * 8); in arbel_set_eq_ci()
199 static inline void set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) in set_eq_ci() argument
202 arbel_set_eq_ci(dev, eq, ci); in set_eq_ci()
204 tavor_set_eq_ci(dev, eq, ci); in set_eq_ci()
228 static inline struct mthca_eqe *get_eqe(struct mthca_eq *eq, u32 entry) in get_eqe() argument
230 unsigned long off = (entry & (eq->nent - 1)) * MTHCA_EQ_ENTRY_SIZE; in get_eqe()
231 return eq->page_list[off / PAGE_SIZE].buf + off % PAGE_SIZE; in get_eqe()
[all …]
/drivers/net/ethernet/mellanox/mlx4/
Deq.c97 static void eq_set_ci(struct mlx4_eq *eq, int req_not) in eq_set_ci() argument
99 __raw_writel((__force u32) cpu_to_be32((eq->cons_index & 0xffffff) | in eq_set_ci()
101 eq->doorbell); in eq_set_ci()
106 static struct mlx4_eqe *get_eqe(struct mlx4_eq *eq, u32 entry, u8 eqe_factor, in get_eqe() argument
110 unsigned long offset = (entry & (eq->nent - 1)) * eqe_size; in get_eqe()
118 …return eq->page_list[offset / PAGE_SIZE].buf + (offset + (eqe_factor ? MLX4_EQ_ENTRY_SIZE : 0)) % … in get_eqe()
121 static struct mlx4_eqe *next_eqe_sw(struct mlx4_eq *eq, u8 eqe_factor, u8 size) in next_eqe_sw() argument
123 struct mlx4_eqe *eqe = get_eqe(eq, eq->cons_index, eqe_factor, size); in next_eqe_sw()
124 return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe; in next_eqe_sw()
241 struct mlx4_eq *eq = &priv->eq_table.eq[vec]; in mlx4_set_eq_affinity_hint() local
[all …]
Dcq.c331 cq_context->comp_eqn = priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].eqn; in mlx4_cq_alloc()
351 &priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].tasklet_ctx; in mlx4_cq_alloc()
355 cq->irq = priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].irq; in mlx4_cq_alloc()
384 synchronize_irq(priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq->vector)].irq); in mlx4_cq_free()
385 if (priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq->vector)].irq != in mlx4_cq_free()
386 priv->eq_table.eq[MLX4_EQ_ASYNC].irq) in mlx4_cq_free()
387 synchronize_irq(priv->eq_table.eq[MLX4_EQ_ASYNC].irq); in mlx4_cq_free()
/drivers/pci/host/
Dpcie-iproc-msi.c72 unsigned int eq; member
138 unsigned int eq) in iproc_msi_read_reg() argument
142 return readl_relaxed(pcie->base + msi->reg_offsets[eq][reg]); in iproc_msi_read_reg()
147 int eq, u32 val) in iproc_msi_write_reg() argument
151 writel_relaxed(val, pcie->base + msi->reg_offsets[eq][reg]); in iproc_msi_write_reg()
168 static inline unsigned int iproc_msi_eq_offset(struct iproc_msi *msi, u32 eq) in iproc_msi_eq_offset() argument
171 return eq * EQ_MEM_REGION_SIZE; in iproc_msi_eq_offset()
173 return eq * EQ_LEN * sizeof(u32); in iproc_msi_eq_offset()
298 static inline u32 decode_msi_hwirq(struct iproc_msi *msi, u32 eq, u32 head) in decode_msi_hwirq() argument
303 offs = iproc_msi_eq_offset(msi, eq) + head * sizeof(u32); in decode_msi_hwirq()
[all …]
/drivers/net/ethernet/ibm/ehea/
Dehea_qmr.c256 struct ehea_eq *eq; in ehea_create_eq() local
258 eq = kzalloc(sizeof(*eq), GFP_KERNEL); in ehea_create_eq()
259 if (!eq) in ehea_create_eq()
262 eq->adapter = adapter; in ehea_create_eq()
263 eq->attr.type = type; in ehea_create_eq()
264 eq->attr.max_nr_of_eqes = max_nr_of_eqes; in ehea_create_eq()
265 eq->attr.eqe_gen = eqe_gen; in ehea_create_eq()
266 spin_lock_init(&eq->spinlock); in ehea_create_eq()
269 &eq->attr, &eq->fw_handle); in ehea_create_eq()
275 ret = hw_queue_ctor(&eq->hw_queue, eq->attr.nr_pages, in ehea_create_eq()
[all …]
/drivers/net/ethernet/tile/
Dtilepro.c201 lepp_queue_t *eq; member
526 static unsigned int tile_net_lepp_grab_comps(lepp_queue_t *eq, in tile_net_lepp_grab_comps() argument
533 unsigned int comp_head = eq->comp_head; in tile_net_lepp_grab_comps()
534 unsigned int comp_busy = eq->comp_busy; in tile_net_lepp_grab_comps()
537 comps[n++] = eq->comps[comp_head]; in tile_net_lepp_grab_comps()
544 eq->comp_head = comp_head; in tile_net_lepp_grab_comps()
557 lepp_queue_t *eq = priv->eq; in tile_net_lepp_free_comps() local
567 eq->comp_busy = eq->comp_tail; in tile_net_lepp_free_comps()
569 n = tile_net_lepp_grab_comps(eq, olds, wanted, 0); in tile_net_lepp_free_comps()
571 pending = (eq->comp_head != eq->comp_tail); in tile_net_lepp_free_comps()
[all …]
/drivers/clk/spear/
Dspear1340_clock.c267 {.xscale = 5, .yscale = 122, .eq = 0},
269 {.xscale = 10, .yscale = 204, .eq = 0},
271 {.xscale = 4, .yscale = 25, .eq = 0},
273 {.xscale = 4, .yscale = 21, .eq = 0},
275 {.xscale = 5, .yscale = 18, .eq = 0},
277 {.xscale = 2, .yscale = 6, .eq = 0},
279 {.xscale = 5, .yscale = 12, .eq = 0},
281 {.xscale = 2, .yscale = 4, .eq = 0},
283 {.xscale = 5, .yscale = 18, .eq = 1},
285 {.xscale = 1, .yscale = 3, .eq = 1},
[all …]
Dspear1310_clock.c255 {.xscale = 10, .yscale = 204, .eq = 0}, /* 12.29 MHz */
256 {.xscale = 4, .yscale = 21, .eq = 0}, /* 48 MHz */
257 {.xscale = 2, .yscale = 6, .eq = 0}, /* 83 MHz */
258 {.xscale = 2, .yscale = 4, .eq = 0}, /* 125 MHz */
259 {.xscale = 1, .yscale = 3, .eq = 1}, /* 166 MHz */
260 {.xscale = 1, .yscale = 2, .eq = 1}, /* 250 MHz */
266 {.xscale = 2, .yscale = 6, .eq = 0}, /* divided by 6 */
267 {.xscale = 2, .yscale = 4, .eq = 0}, /* divided by 4 */
268 {.xscale = 1, .yscale = 3, .eq = 1}, /* divided by 3 */
269 {.xscale = 1, .yscale = 2, .eq = 1}, /* divided by 2 */
[all …]
Dspear3xx_clock.c109 {.xscale = 1, .yscale = 81, .eq = 0}, /* 2.049 MHz */
110 {.xscale = 1, .yscale = 59, .eq = 0}, /* 2.822 MHz */
111 {.xscale = 2, .yscale = 81, .eq = 0}, /* 4.098 MHz */
112 {.xscale = 3, .yscale = 89, .eq = 0}, /* 5.644 MHz */
113 {.xscale = 4, .yscale = 81, .eq = 0}, /* 8.197 MHz */
114 {.xscale = 4, .yscale = 59, .eq = 0}, /* 11.254 MHz */
115 {.xscale = 2, .yscale = 27, .eq = 0}, /* 12.296 MHz */
116 {.xscale = 2, .yscale = 8, .eq = 0}, /* 41.5 MHz */
117 {.xscale = 2, .yscale = 4, .eq = 0}, /* 83 MHz */
118 {.xscale = 1, .yscale = 2, .eq = 1}, /* 166 MHz */
Dclk-aux-synth.c49 u8 eq = rtbl[index].eq ? 1 : 2; in aux_calc_rate() local
52 (rtbl[index].yscale * eq)) * 10000; in aux_calc_rate()
115 val |= (rtbl[i].eq & aux->masks->eq_sel_mask) << in clk_aux_set_rate()
Dspear6xx_clock.c93 {.xscale = 2, .yscale = 27, .eq = 0}, /* 12.296 MHz */
94 {.xscale = 2, .yscale = 8, .eq = 0}, /* 41.5 MHz */
95 {.xscale = 2, .yscale = 4, .eq = 0}, /* 83 MHz */
96 {.xscale = 1, .yscale = 2, .eq = 1}, /* 166 MHz */
/drivers/firmware/broadcom/
Dbcm47xx_nvram.c171 char *var, *value, *end, *eq; in bcm47xx_nvram_getenv() local
187 eq = strchr(var, '='); in bcm47xx_nvram_getenv()
188 if (!eq) in bcm47xx_nvram_getenv()
190 value = eq + 1; in bcm47xx_nvram_getenv()
191 if (eq - var == strlen(name) && in bcm47xx_nvram_getenv()
192 strncmp(var, name, eq - var) == 0) in bcm47xx_nvram_getenv()
/drivers/infiniband/hw/ocrdma/
Docrdma_hw.c111 static inline void *ocrdma_get_eqe(struct ocrdma_eq *eq) in ocrdma_get_eqe() argument
113 return eq->q.va + (eq->q.tail * sizeof(struct ocrdma_eqe)); in ocrdma_get_eqe()
116 static inline void ocrdma_eq_inc_tail(struct ocrdma_eq *eq) in ocrdma_eq_inc_tail() argument
118 eq->q.tail = (eq->q.tail + 1) & (OCRDMA_EQ_LEN - 1); in ocrdma_eq_inc_tail()
434 static int ocrdma_mbx_create_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq) in ocrdma_mbx_create_eq() argument
449 ocrdma_build_q_pages(&cmd->pa[0], cmd->num_pages, eq->q.dma, in ocrdma_mbx_create_eq()
454 eq->q.id = rsp->vector_eqid & 0xffff; in ocrdma_mbx_create_eq()
455 eq->vector = (rsp->vector_eqid >> 16) & 0xffff; in ocrdma_mbx_create_eq()
456 eq->q.created = true; in ocrdma_mbx_create_eq()
462 struct ocrdma_eq *eq, u16 q_len) in ocrdma_create_eq() argument
[all …]
/drivers/staging/lustre/include/linux/lnet/
Dlib-lnet.h186 struct lnet_eq *eq; in lnet_eq_alloc() local
188 LIBCFS_ALLOC(eq, sizeof(*eq)); in lnet_eq_alloc()
189 return eq; in lnet_eq_alloc()
193 lnet_eq_free(struct lnet_eq *eq) in lnet_eq_free() argument
195 LIBCFS_FREE(eq, sizeof(*eq)); in lnet_eq_free()
283 lnet_eq2handle(struct lnet_handle_eq *handle, struct lnet_eq *eq) in lnet_eq2handle() argument
285 if (!eq) { in lnet_eq2handle()
290 handle->cookie = eq->eq_lh.lh_cookie; in lnet_eq2handle()
496 void lnet_eq_enqueue_event(struct lnet_eq *eq, struct lnet_event *ev);
/drivers/infiniband/hw/hfi1/
Dpcie.c803 static int load_eq_table(struct hfi1_devdata *dd, const u8 eq[11][3], u8 fs, in load_eq_table()
817 c_minus1 = eq[i][PREC] / div; in load_eq_table()
818 c0 = fs - (eq[i][PREC] / div) - (eq[i][POST] / div); in load_eq_table()
819 c_plus1 = eq[i][POST] / div; in load_eq_table()
839 i, (u32)eq[i][0], (u32)eq[i][1], in load_eq_table()
840 (u32)eq[i][2]); in load_eq_table()
1039 const u8 (*eq)[3]; in do_pcie_gen3_transition() local
1173 eq = discrete_preliminary_eq; in do_pcie_gen3_transition()
1183 eq = integrated_preliminary_eq; in do_pcie_gen3_transition()
1194 ret = load_eq_table(dd, eq, fs, div); in do_pcie_gen3_transition()
/drivers/scsi/csiostor/
Dcsio_wr.h411 struct csio_eq eq; member
468 #define csio_q_eqid(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.eq.eqid)
472 ((__hw)->wrm.q_arr[(__idx)]->un.eq.physeqid)
479 #define csio_q_eq_wrap(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.eq.wrap)
/drivers/scsi/be2iscsi/
Dbe_main.c678 struct be_queue_info *eq; in be_isr_mcc() local
684 eq = &pbe_eq->q; in be_isr_mcc()
687 eqe = queue_tail_node(eq); in be_isr_mcc()
698 queue_tail_inc(eq); in be_isr_mcc()
699 eqe = queue_tail_node(eq); in be_isr_mcc()
704 hwi_ring_eq_db(phba, eq->id, 1, mcc_events, 1, 1); in be_isr_mcc()
717 struct be_queue_info *eq; in be_isr_msix() local
721 eq = &pbe_eq->q; in be_isr_msix()
725 hwi_ring_eq_db(phba, eq->id, 1, 0, 0, 1); in be_isr_msix()
742 struct be_queue_info *eq; in be_isr() local
[all …]
/drivers/scsi/lpfc/
Dlpfc_debugfs.h451 struct lpfc_queue *wq, *cq, *eq; in lpfc_debug_dump_cq() local
489 eq = phba->sli4_hba.hba_eq[eqidx]; in lpfc_debug_dump_cq()
495 eqidx, eq->queue_id); in lpfc_debug_dump_cq()
500 eqidx, eq->queue_id); in lpfc_debug_dump_cq()

123