Home
last modified time | relevance | path

Searched refs:eq_table (Results 1 – 15 of 15) sorted by relevance

/drivers/infiniband/hw/mthca/
Dmthca_eq.c397 if (dev->eq_table.clr_mask) in mthca_tavor_interrupt()
398 writel(dev->eq_table.clr_mask, dev->eq_table.clr_int); in mthca_tavor_interrupt()
408 if (ecr & dev->eq_table.eq[i].eqn_mask) { in mthca_tavor_interrupt()
409 if (mthca_eq_int(dev, &dev->eq_table.eq[i])) in mthca_tavor_interrupt()
410 tavor_set_eq_ci(dev, &dev->eq_table.eq[i], in mthca_tavor_interrupt()
411 dev->eq_table.eq[i].cons_index); in mthca_tavor_interrupt()
412 tavor_eq_req_not(dev, dev->eq_table.eq[i].eqn); in mthca_tavor_interrupt()
437 if (dev->eq_table.clr_mask) in mthca_arbel_interrupt()
438 writel(dev->eq_table.clr_mask, dev->eq_table.clr_int); in mthca_arbel_interrupt()
441 if (mthca_eq_int(dev, &dev->eq_table.eq[i])) { in mthca_arbel_interrupt()
[all …]
Dmthca_main.c683 mdev->eq_table.inta_pin = adapter.inta_pin; in mthca_init_hca()
763 dev->eq_table.eq[MTHCA_EQ_CMD].msi_x_vector); in mthca_setup_hca()
865 mdev->eq_table.eq[MTHCA_EQ_COMP ].msi_x_vector = entries[0].vector; in mthca_enable_msi_x()
866 mdev->eq_table.eq[MTHCA_EQ_ASYNC].msi_x_vector = entries[1].vector; in mthca_enable_msi_x()
867 mdev->eq_table.eq[MTHCA_EQ_CMD ].msi_x_vector = entries[2].vector; in mthca_enable_msi_x()
Dmthca_cq.c841 cq_context->error_eqn = cpu_to_be32(dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn); in mthca_init_cq()
842 cq_context->comp_eqn = cpu_to_be32(dev->eq_table.eq[MTHCA_EQ_COMP].eqn); in mthca_init_cq()
942 synchronize_irq(dev->eq_table.eq[MTHCA_EQ_COMP].msi_x_vector); in mthca_free_cq()
Dmthca_dev.h341 struct mthca_eq_table eq_table; member
/drivers/net/ethernet/mellanox/mlx4/
Deq.c229 struct mlx4_eq *eq = &priv->eq_table.eq[vec]; in mlx4_set_eq_affinity_hint()
845 writel(priv->eq_table.clr_mask, priv->eq_table.clr_int); in mlx4_interrupt()
848 work |= mlx4_eq_int(dev, &priv->eq_table.eq[i]); in mlx4_interrupt()
931 if (!priv->eq_table.uar_map[index]) { in mlx4_get_eq_uar()
932 priv->eq_table.uar_map[index] = in mlx4_get_eq_uar()
936 if (!priv->eq_table.uar_map[index]) { in mlx4_get_eq_uar()
943 return priv->eq_table.uar_map[index] + 0x800 + 8 * (eq->eqn % 4); in mlx4_get_eq_uar()
952 if (priv->eq_table.uar_map[i]) { in mlx4_unmap_uar()
953 iounmap(priv->eq_table.uar_map[i]); in mlx4_unmap_uar()
954 priv->eq_table.uar_map[i] = NULL; in mlx4_unmap_uar()
[all …]
Dcq.c324 cq_context->comp_eqn = priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].eqn; in mlx4_cq_alloc()
344 &priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].tasklet_ctx; in mlx4_cq_alloc()
348 cq->irq = priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].irq; in mlx4_cq_alloc()
377 synchronize_irq(priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq->vector)].irq); in mlx4_cq_free()
378 if (priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq->vector)].irq != in mlx4_cq_free()
379 priv->eq_table.eq[MLX4_EQ_ASYNC].irq) in mlx4_cq_free()
380 synchronize_irq(priv->eq_table.eq[MLX4_EQ_ASYNC].irq); in mlx4_cq_free()
Dmain.c1377 err = mlx4_init_icm_table(dev, &priv->eq_table.cmpt_table, in mlx4_init_cmpt_table()
1439 err = mlx4_init_icm_table(dev, &priv->eq_table.table, in mlx4_init_icm()
1587 mlx4_cleanup_icm_table(dev, &priv->eq_table.table); in mlx4_init_icm()
1590 mlx4_cleanup_icm_table(dev, &priv->eq_table.cmpt_table); in mlx4_init_icm()
1617 mlx4_cleanup_icm_table(dev, &priv->eq_table.table); in mlx4_free_icms()
1618 mlx4_cleanup_icm_table(dev, &priv->eq_table.cmpt_table); in mlx4_free_icms()
2215 priv->eq_table.inta_pin = adapter.inta_pin; in mlx4_init_hca()
2516 priv->eq_table.eq[MLX4_EQ_ASYNC].irq); in mlx4_setup_hca()
2520 priv->eq_table.eq[MLX4_EQ_ASYNC].irq); in mlx4_setup_hca()
2665 eq = &priv->eq_table.eq[eqn]; in mlx4_init_affinity_hint()
[all …]
Dmlx4.h878 struct mlx4_eq_table eq_table; member
/drivers/net/ethernet/mellanox/mlx5/core/
Deq.c440 spin_lock_init(&dev->priv.eq_table.lock); in mlx5_eq_init()
455 struct mlx5_eq_table *table = &dev->priv.eq_table; in mlx5_start_eqs()
503 struct mlx5_eq_table *table = &dev->priv.eq_table; in mlx5_stop_eqs()
Dmain.c258 struct mlx5_eq_table *table = &priv->eq_table; in mlx5_enable_msix()
547 for (i = 0; i < mdev->priv.eq_table.num_comp_vectors; i++) { in mlx5_irq_set_affinity_hints()
566 for (i = 0; i < mdev->priv.eq_table.num_comp_vectors; i++) in mlx5_irq_clear_affinity_hints()
573 struct mlx5_eq_table *table = &dev->priv.eq_table; in mlx5_vector2eqn()
594 struct mlx5_eq_table *table = &dev->priv.eq_table; in free_comp_eqs()
612 struct mlx5_eq_table *table = &dev->priv.eq_table; in alloc_comp_eqs()
Den.h630 return min_t(int, mdev->priv.eq_table.num_comp_vectors, in mlx5e_get_max_num_channels()
/drivers/infiniband/hw/mlx4/
Dmain.c2080 ibdev->eq_table = kcalloc(dev->caps.num_comp_vectors, in mlx4_ib_alloc_eqs()
2081 sizeof(ibdev->eq_table[0]), GFP_KERNEL); in mlx4_ib_alloc_eqs()
2082 if (!ibdev->eq_table) in mlx4_ib_alloc_eqs()
2090 ibdev->eq_table[eq] = total_eqs; in mlx4_ib_alloc_eqs()
2092 &ibdev->eq_table[eq])) in mlx4_ib_alloc_eqs()
2095 ibdev->eq_table[eq] = -1; in mlx4_ib_alloc_eqs()
2100 ibdev->eq_table[i++] = -1) in mlx4_ib_alloc_eqs()
2113 if (!ibdev->eq_table) in mlx4_ib_free_eqs()
2120 mlx4_release_eq(dev, ibdev->eq_table[i]); in mlx4_ib_free_eqs()
2122 kfree(ibdev->eq_table); in mlx4_ib_free_eqs()
[all …]
Dcq.c238 if (dev->eq_table) in mlx4_ib_create_cq()
239 vector = dev->eq_table[vector % ibdev->num_comp_vectors]; in mlx4_ib_create_cq()
Dmlx4_ib.h563 int *eq_table; member
/drivers/infiniband/hw/mlx5/
Dmain.c1361 dev->mdev->priv.eq_table.num_comp_vectors; in mlx5_ib_add()