/drivers/staging/lustre/lnet/lnet/ |
D | lib-eq.c | 73 lnet_eq_t *eq; in LNetEQAlloc() local 92 eq = lnet_eq_alloc(); in LNetEQAlloc() 93 if (eq == NULL) in LNetEQAlloc() 97 LIBCFS_ALLOC(eq->eq_events, count * sizeof(lnet_event_t)); in LNetEQAlloc() 98 if (eq->eq_events == NULL) in LNetEQAlloc() 104 eq->eq_deq_seq = 1; in LNetEQAlloc() 105 eq->eq_enq_seq = 1; in LNetEQAlloc() 106 eq->eq_size = count; in LNetEQAlloc() 107 eq->eq_callback = callback; in LNetEQAlloc() 109 eq->eq_refs = cfs_percpt_alloc(lnet_cpt_table(), in LNetEQAlloc() [all …]
|
/drivers/staging/rdma/ehca/ |
D | ehca_eq.c | 54 struct ehca_eq *eq, in ehca_create_eq() argument 64 spin_lock_init(&eq->spinlock); in ehca_create_eq() 65 spin_lock_init(&eq->irq_spinlock); in ehca_create_eq() 66 eq->is_initialized = 0; in ehca_create_eq() 69 ehca_err(ib_dev, "Invalid EQ type %x. eq=%p", type, eq); in ehca_create_eq() 73 ehca_err(ib_dev, "EQ length must not be zero. eq=%p", eq); in ehca_create_eq() 78 &eq->pf, in ehca_create_eq() 81 &eq->ipz_eq_handle, in ehca_create_eq() 82 &eq->length, in ehca_create_eq() 83 &nr_pages, &eq->ist); in ehca_create_eq() [all …]
|
D | ehca_irq.c | 507 tasklet_hi_schedule(&shca->eq.interrupt_task); in ehca_interrupt_eq() 551 struct ehca_eq *eq = &shca->eq; in ehca_process_eq() local 552 struct ehca_eqe_cache_entry *eqe_cache = eq->eqe_cache; in ehca_process_eq() 557 spin_lock(&eq->irq_spinlock); in ehca_process_eq() 564 shca->ipz_hca_handle, eq->ist); in ehca_process_eq() 577 eqe_cache[eqe_cnt].eqe = ehca_poll_eq(shca, eq); in ehca_process_eq() 604 ret = hipz_h_eoi(eq->ist); in ehca_process_eq() 614 if (eq->eqe_cache[i].cq) in ehca_process_eq() 615 reset_eq_pending(eq->eqe_cache[i].cq); in ehca_process_eq() 618 spin_lock(&eq->spinlock); in ehca_process_eq() [all …]
|
D | ehca_main.c | 794 ret = ehca_create_eq(shca, &shca->eq, EHCA_EQ, eq_size); in ehca_probe() 894 ret = ehca_destroy_eq(shca, &shca->eq); in ehca_probe() 936 ret = ehca_destroy_eq(shca, &shca->eq); in ehca_remove() 980 if (shca->eq.is_initialized) { in ehca_poll_eqs() 982 struct ehca_eq *eq = &shca->eq; in ehca_poll_eqs() local 986 spin_lock_irqsave(&eq->spinlock, flags); in ehca_poll_eqs() 987 q_ofs = eq->ipz_queue.current_q_offset; in ehca_poll_eqs() 988 spin_unlock_irqrestore(&eq->spinlock, flags); in ehca_poll_eqs() 990 spin_lock_irqsave(&eq->spinlock, flags); in ehca_poll_eqs() 991 q_ofs2 = eq->ipz_queue.current_q_offset; in ehca_poll_eqs() [all …]
|
D | ehca_iverbs.h | 125 int ehca_create_eq(struct ehca_shca *shca, struct ehca_eq *eq, 128 int ehca_destroy_eq(struct ehca_shca *shca, struct ehca_eq *eq); 130 void *ehca_poll_eq(struct ehca_shca *shca, struct ehca_eq *eq);
|
/drivers/infiniband/hw/mthca/ |
D | mthca_eq.c | 173 static inline void tavor_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) in tavor_set_eq_ci() argument 184 mthca_write64(MTHCA_EQ_DB_SET_CI | eq->eqn, ci & (eq->nent - 1), in tavor_set_eq_ci() 189 static inline void arbel_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) in arbel_set_eq_ci() argument 194 dev->eq_regs.arbel.eq_set_ci_base + eq->eqn * 8); in arbel_set_eq_ci() 199 static inline void set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) in set_eq_ci() argument 202 arbel_set_eq_ci(dev, eq, ci); in set_eq_ci() 204 tavor_set_eq_ci(dev, eq, ci); in set_eq_ci() 228 static inline struct mthca_eqe *get_eqe(struct mthca_eq *eq, u32 entry) in get_eqe() argument 230 unsigned long off = (entry & (eq->nent - 1)) * MTHCA_EQ_ENTRY_SIZE; in get_eqe() 231 return eq->page_list[off / PAGE_SIZE].buf + off % PAGE_SIZE; in get_eqe() [all …]
|
/drivers/net/ethernet/mellanox/mlx5/core/ |
D | eq.c | 105 static struct mlx5_eqe *get_eqe(struct mlx5_eq *eq, u32 entry) in get_eqe() argument 107 return mlx5_buf_offset(&eq->buf, entry * MLX5_EQE_SIZE); in get_eqe() 110 static struct mlx5_eqe *next_eqe_sw(struct mlx5_eq *eq) in next_eqe_sw() argument 112 struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & (eq->nent - 1)); in next_eqe_sw() 114 return ((eqe->owner & 1) ^ !!(eq->cons_index & eq->nent)) ? NULL : eqe; in next_eqe_sw() 188 static void eq_update_ci(struct mlx5_eq *eq, int arm) in eq_update_ci() argument 190 __be32 __iomem *addr = eq->doorbell + (arm ? 0 : 2); in eq_update_ci() 191 u32 val = (eq->cons_index & 0xffffff) | (eq->eqn << 24); in eq_update_ci() 197 static int mlx5_eq_int(struct mlx5_core_dev *dev, struct mlx5_eq *eq) in mlx5_eq_int() argument 206 while ((eqe = next_eqe_sw(eq))) { in mlx5_eq_int() [all …]
|
D | debugfs.c | 358 static u64 eq_read_field(struct mlx5_core_dev *dev, struct mlx5_eq *eq, in eq_read_field() argument 372 err = mlx5_core_eq_query(dev, eq, out, sizeof(*out)); in eq_read_field() 562 int mlx5_debug_eq_add(struct mlx5_core_dev *dev, struct mlx5_eq *eq) in mlx5_debug_eq_add() argument 570 &eq->dbg, eq->eqn, eq_fields, in mlx5_debug_eq_add() 571 ARRAY_SIZE(eq_fields), eq); in mlx5_debug_eq_add() 573 eq->dbg = NULL; in mlx5_debug_eq_add() 578 void mlx5_debug_eq_remove(struct mlx5_core_dev *dev, struct mlx5_eq *eq) in mlx5_debug_eq_remove() argument 583 if (eq->dbg) in mlx5_debug_eq_remove() 584 rem_res_tree(eq->dbg); in mlx5_debug_eq_remove()
|
D | main.c | 574 struct mlx5_eq *eq, *n; in mlx5_vector2eqn() local 578 list_for_each_entry_safe(eq, n, &table->comp_eqs_list, list) { in mlx5_vector2eqn() 579 if (eq->index == vector) { in mlx5_vector2eqn() 580 *eqn = eq->eqn; in mlx5_vector2eqn() 581 *irqn = eq->irqn; in mlx5_vector2eqn() 595 struct mlx5_eq *eq, *n; in free_comp_eqs() local 598 list_for_each_entry_safe(eq, n, &table->comp_eqs_list, list) { in free_comp_eqs() 599 list_del(&eq->list); in free_comp_eqs() 601 if (mlx5_destroy_unmap_eq(dev, eq)) in free_comp_eqs() 603 eq->eqn); in free_comp_eqs() [all …]
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | eq.c | 97 static void eq_set_ci(struct mlx4_eq *eq, int req_not) in eq_set_ci() argument 99 __raw_writel((__force u32) cpu_to_be32((eq->cons_index & 0xffffff) | in eq_set_ci() 101 eq->doorbell); in eq_set_ci() 106 static struct mlx4_eqe *get_eqe(struct mlx4_eq *eq, u32 entry, u8 eqe_factor, in get_eqe() argument 110 unsigned long offset = (entry & (eq->nent - 1)) * eqe_size; in get_eqe() 118 …return eq->page_list[offset / PAGE_SIZE].buf + (offset + (eqe_factor ? MLX4_EQ_ENTRY_SIZE : 0)) % … in get_eqe() 121 static struct mlx4_eqe *next_eqe_sw(struct mlx4_eq *eq, u8 eqe_factor, u8 size) in next_eqe_sw() argument 123 struct mlx4_eqe *eqe = get_eqe(eq, eq->cons_index, eqe_factor, size); in next_eqe_sw() 124 return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe; in next_eqe_sw() 229 struct mlx4_eq *eq = &priv->eq_table.eq[vec]; in mlx4_set_eq_affinity_hint() local [all …]
|
D | cq.c | 324 cq_context->comp_eqn = priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].eqn; in mlx4_cq_alloc() 344 &priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].tasklet_ctx; in mlx4_cq_alloc() 348 cq->irq = priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].irq; in mlx4_cq_alloc() 377 synchronize_irq(priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq->vector)].irq); in mlx4_cq_free() 378 if (priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq->vector)].irq != in mlx4_cq_free() 379 priv->eq_table.eq[MLX4_EQ_ASYNC].irq) in mlx4_cq_free() 380 synchronize_irq(priv->eq_table.eq[MLX4_EQ_ASYNC].irq); in mlx4_cq_free()
|
/drivers/net/ethernet/ibm/ehea/ |
D | ehea_qmr.c | 256 struct ehea_eq *eq; in ehea_create_eq() local 258 eq = kzalloc(sizeof(*eq), GFP_KERNEL); in ehea_create_eq() 259 if (!eq) in ehea_create_eq() 262 eq->adapter = adapter; in ehea_create_eq() 263 eq->attr.type = type; in ehea_create_eq() 264 eq->attr.max_nr_of_eqes = max_nr_of_eqes; in ehea_create_eq() 265 eq->attr.eqe_gen = eqe_gen; in ehea_create_eq() 266 spin_lock_init(&eq->spinlock); in ehea_create_eq() 269 &eq->attr, &eq->fw_handle); in ehea_create_eq() 275 ret = hw_queue_ctor(&eq->hw_queue, eq->attr.nr_pages, in ehea_create_eq() [all …]
|
/drivers/net/ethernet/tile/ |
D | tilepro.c | 201 lepp_queue_t *eq; member 526 static unsigned int tile_net_lepp_grab_comps(lepp_queue_t *eq, in tile_net_lepp_grab_comps() argument 533 unsigned int comp_head = eq->comp_head; in tile_net_lepp_grab_comps() 534 unsigned int comp_busy = eq->comp_busy; in tile_net_lepp_grab_comps() 537 comps[n++] = eq->comps[comp_head]; in tile_net_lepp_grab_comps() 544 eq->comp_head = comp_head; in tile_net_lepp_grab_comps() 557 lepp_queue_t *eq = priv->eq; in tile_net_lepp_free_comps() local 567 eq->comp_busy = eq->comp_tail; in tile_net_lepp_free_comps() 569 n = tile_net_lepp_grab_comps(eq, olds, wanted, 0); in tile_net_lepp_free_comps() 571 pending = (eq->comp_head != eq->comp_tail); in tile_net_lepp_free_comps() [all …]
|
/drivers/clk/spear/ |
D | spear1340_clock.c | 267 {.xscale = 5, .yscale = 122, .eq = 0}, 269 {.xscale = 10, .yscale = 204, .eq = 0}, 271 {.xscale = 4, .yscale = 25, .eq = 0}, 273 {.xscale = 4, .yscale = 21, .eq = 0}, 275 {.xscale = 5, .yscale = 18, .eq = 0}, 277 {.xscale = 2, .yscale = 6, .eq = 0}, 279 {.xscale = 5, .yscale = 12, .eq = 0}, 281 {.xscale = 2, .yscale = 4, .eq = 0}, 283 {.xscale = 5, .yscale = 18, .eq = 1}, 285 {.xscale = 1, .yscale = 3, .eq = 1}, [all …]
|
D | spear1310_clock.c | 255 {.xscale = 10, .yscale = 204, .eq = 0}, /* 12.29 MHz */ 256 {.xscale = 4, .yscale = 21, .eq = 0}, /* 48 MHz */ 257 {.xscale = 2, .yscale = 6, .eq = 0}, /* 83 MHz */ 258 {.xscale = 2, .yscale = 4, .eq = 0}, /* 125 MHz */ 259 {.xscale = 1, .yscale = 3, .eq = 1}, /* 166 MHz */ 260 {.xscale = 1, .yscale = 2, .eq = 1}, /* 250 MHz */ 266 {.xscale = 2, .yscale = 6, .eq = 0}, /* divided by 6 */ 267 {.xscale = 2, .yscale = 4, .eq = 0}, /* divided by 4 */ 268 {.xscale = 1, .yscale = 3, .eq = 1}, /* divided by 3 */ 269 {.xscale = 1, .yscale = 2, .eq = 1}, /* divided by 2 */ [all …]
|
D | spear3xx_clock.c | 109 {.xscale = 1, .yscale = 81, .eq = 0}, /* 2.049 MHz */ 110 {.xscale = 1, .yscale = 59, .eq = 0}, /* 2.822 MHz */ 111 {.xscale = 2, .yscale = 81, .eq = 0}, /* 4.098 MHz */ 112 {.xscale = 3, .yscale = 89, .eq = 0}, /* 5.644 MHz */ 113 {.xscale = 4, .yscale = 81, .eq = 0}, /* 8.197 MHz */ 114 {.xscale = 4, .yscale = 59, .eq = 0}, /* 11.254 MHz */ 115 {.xscale = 2, .yscale = 27, .eq = 0}, /* 12.296 MHz */ 116 {.xscale = 2, .yscale = 8, .eq = 0}, /* 41.5 MHz */ 117 {.xscale = 2, .yscale = 4, .eq = 0}, /* 83 MHz */ 118 {.xscale = 1, .yscale = 2, .eq = 1}, /* 166 MHz */
|
D | clk-aux-synth.c | 49 u8 eq = rtbl[index].eq ? 1 : 2; in aux_calc_rate() local 52 (rtbl[index].yscale * eq)) * 10000; in aux_calc_rate() 115 val |= (rtbl[i].eq & aux->masks->eq_sel_mask) << in clk_aux_set_rate()
|
D | spear6xx_clock.c | 93 {.xscale = 2, .yscale = 27, .eq = 0}, /* 12.296 MHz */ 94 {.xscale = 2, .yscale = 8, .eq = 0}, /* 41.5 MHz */ 95 {.xscale = 2, .yscale = 4, .eq = 0}, /* 83 MHz */ 96 {.xscale = 1, .yscale = 2, .eq = 1}, /* 166 MHz */
|
/drivers/firmware/broadcom/ |
D | bcm47xx_nvram.c | 177 char *var, *value, *end, *eq; in bcm47xx_nvram_getenv() local 193 eq = strchr(var, '='); in bcm47xx_nvram_getenv() 194 if (!eq) in bcm47xx_nvram_getenv() 196 value = eq + 1; in bcm47xx_nvram_getenv() 197 if (eq - var == strlen(name) && in bcm47xx_nvram_getenv() 198 strncmp(var, name, eq - var) == 0) in bcm47xx_nvram_getenv()
|
/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_hw.c | 110 static inline void *ocrdma_get_eqe(struct ocrdma_eq *eq) in ocrdma_get_eqe() argument 112 return eq->q.va + (eq->q.tail * sizeof(struct ocrdma_eqe)); in ocrdma_get_eqe() 115 static inline void ocrdma_eq_inc_tail(struct ocrdma_eq *eq) in ocrdma_eq_inc_tail() argument 117 eq->q.tail = (eq->q.tail + 1) & (OCRDMA_EQ_LEN - 1); in ocrdma_eq_inc_tail() 430 static int ocrdma_mbx_create_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq) in ocrdma_mbx_create_eq() argument 445 ocrdma_build_q_pages(&cmd->pa[0], cmd->num_pages, eq->q.dma, in ocrdma_mbx_create_eq() 450 eq->q.id = rsp->vector_eqid & 0xffff; in ocrdma_mbx_create_eq() 451 eq->vector = (rsp->vector_eqid >> 16) & 0xffff; in ocrdma_mbx_create_eq() 452 eq->q.created = true; in ocrdma_mbx_create_eq() 458 struct ocrdma_eq *eq, u16 q_len) in ocrdma_create_eq() argument [all …]
|
/drivers/staging/lustre/include/linux/lnet/ |
D | lib-lnet.h | 167 lnet_eq_t *eq; in lnet_eq_alloc() local 169 LIBCFS_ALLOC(eq, sizeof(*eq)); in lnet_eq_alloc() 170 return eq; in lnet_eq_alloc() 174 lnet_eq_free(lnet_eq_t *eq) in lnet_eq_free() argument 176 LIBCFS_FREE(eq, sizeof(*eq)); in lnet_eq_free() 265 lnet_eq2handle(lnet_handle_eq_t *handle, lnet_eq_t *eq) in lnet_eq2handle() argument 267 if (eq == NULL) { in lnet_eq2handle() 272 handle->cookie = eq->eq_lh.lh_cookie; in lnet_eq2handle() 463 void lnet_eq_enqueue_event(lnet_eq_t *eq, lnet_event_t *ev);
|
/drivers/staging/rdma/hfi1/ |
D | pcie.c | 733 static int load_eq_table(struct hfi1_devdata *dd, const u8 eq[11][3], u8 fs, in load_eq_table() 746 c_minus1 = eq[i][PREC] / div; in load_eq_table() 747 c0 = fs - (eq[i][PREC] / div) - (eq[i][POST] / div); in load_eq_table() 748 c_plus1 = eq[i][POST] / div; in load_eq_table() 762 i, (u32)eq[i][0], (u32)eq[i][1], (u32)eq[i][2]); in load_eq_table() 885 const u8 (*eq)[3]; in do_pcie_gen3_transition() local 1017 eq = discrete_preliminary_eq; in do_pcie_gen3_transition() 1024 eq = integrated_preliminary_eq; in do_pcie_gen3_transition() 1030 ret = load_eq_table(dd, eq, fs, div); in do_pcie_gen3_transition()
|
/drivers/scsi/csiostor/ |
D | csio_wr.h | 411 struct csio_eq eq; member 468 #define csio_q_eqid(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.eq.eqid) 472 ((__hw)->wrm.q_arr[(__idx)]->un.eq.physeqid) 479 #define csio_q_eq_wrap(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.eq.wrap)
|
/drivers/scsi/be2iscsi/ |
D | be_main.c | 854 struct be_queue_info *eq; in be_isr_mcc() local 861 eq = &pbe_eq->q; in be_isr_mcc() 864 eqe = queue_tail_node(eq); in be_isr_mcc() 878 queue_tail_inc(eq); in be_isr_mcc() 879 eqe = queue_tail_node(eq); in be_isr_mcc() 885 hwi_ring_eq_db(phba, eq->id, 1, num_eq_processed, 1, 1); in be_isr_mcc() 899 struct be_queue_info *eq; in be_isr_msix() local 905 eq = &pbe_eq->q; in be_isr_msix() 907 eqe = queue_tail_node(eq); in be_isr_msix() 917 queue_tail_inc(eq); in be_isr_msix() [all …]
|
D | be_cmds.c | 828 struct be_queue_info *eq, int eq_delay) in beiscsi_cmd_eq_create() argument 833 struct be_dma_mem *q_mem = &eq->dma_mem; in beiscsi_cmd_eq_create() 851 __ilog2_u32(eq->len / 256)); in beiscsi_cmd_eq_create() 860 eq->id = le16_to_cpu(resp->eq_id); in beiscsi_cmd_eq_create() 861 eq->created = true; in beiscsi_cmd_eq_create() 949 struct be_queue_info *cq, struct be_queue_info *eq, in beiscsi_cmd_cq_create() argument 978 AMAP_SET_BITS(struct amap_cq_context, eqid, ctxt, eq->id); in beiscsi_cmd_cq_create() 993 AMAP_SET_BITS(struct amap_cq_context_v2, eqid, ctxt, eq->id); in beiscsi_cmd_cq_create()
|