Searched refs:eq_table (Results 1 – 15 of 15) sorted by relevance
/drivers/infiniband/hw/mthca/ |
D | mthca_eq.c | 397 if (dev->eq_table.clr_mask) in mthca_tavor_interrupt() 398 writel(dev->eq_table.clr_mask, dev->eq_table.clr_int); in mthca_tavor_interrupt() 408 if (ecr & dev->eq_table.eq[i].eqn_mask) { in mthca_tavor_interrupt() 409 if (mthca_eq_int(dev, &dev->eq_table.eq[i])) in mthca_tavor_interrupt() 410 tavor_set_eq_ci(dev, &dev->eq_table.eq[i], in mthca_tavor_interrupt() 411 dev->eq_table.eq[i].cons_index); in mthca_tavor_interrupt() 412 tavor_eq_req_not(dev, dev->eq_table.eq[i].eqn); in mthca_tavor_interrupt() 437 if (dev->eq_table.clr_mask) in mthca_arbel_interrupt() 438 writel(dev->eq_table.clr_mask, dev->eq_table.clr_int); in mthca_arbel_interrupt() 441 if (mthca_eq_int(dev, &dev->eq_table.eq[i])) { in mthca_arbel_interrupt() [all …]
|
D | mthca_main.c | 683 mdev->eq_table.inta_pin = adapter.inta_pin; in mthca_init_hca() 763 dev->eq_table.eq[MTHCA_EQ_CMD].msi_x_vector); in mthca_setup_hca() 865 mdev->eq_table.eq[MTHCA_EQ_COMP ].msi_x_vector = entries[0].vector; in mthca_enable_msi_x() 866 mdev->eq_table.eq[MTHCA_EQ_ASYNC].msi_x_vector = entries[1].vector; in mthca_enable_msi_x() 867 mdev->eq_table.eq[MTHCA_EQ_CMD ].msi_x_vector = entries[2].vector; in mthca_enable_msi_x()
|
D | mthca_cq.c | 841 cq_context->error_eqn = cpu_to_be32(dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn); in mthca_init_cq() 842 cq_context->comp_eqn = cpu_to_be32(dev->eq_table.eq[MTHCA_EQ_COMP].eqn); in mthca_init_cq() 942 synchronize_irq(dev->eq_table.eq[MTHCA_EQ_COMP].msi_x_vector); in mthca_free_cq()
|
D | mthca_dev.h | 341 struct mthca_eq_table eq_table; member
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | eq.c | 229 struct mlx4_eq *eq = &priv->eq_table.eq[vec]; in mlx4_set_eq_affinity_hint() 845 writel(priv->eq_table.clr_mask, priv->eq_table.clr_int); in mlx4_interrupt() 848 work |= mlx4_eq_int(dev, &priv->eq_table.eq[i]); in mlx4_interrupt() 931 if (!priv->eq_table.uar_map[index]) { in mlx4_get_eq_uar() 932 priv->eq_table.uar_map[index] = in mlx4_get_eq_uar() 936 if (!priv->eq_table.uar_map[index]) { in mlx4_get_eq_uar() 943 return priv->eq_table.uar_map[index] + 0x800 + 8 * (eq->eqn % 4); in mlx4_get_eq_uar() 952 if (priv->eq_table.uar_map[i]) { in mlx4_unmap_uar() 953 iounmap(priv->eq_table.uar_map[i]); in mlx4_unmap_uar() 954 priv->eq_table.uar_map[i] = NULL; in mlx4_unmap_uar() [all …]
|
D | cq.c | 324 cq_context->comp_eqn = priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].eqn; in mlx4_cq_alloc() 344 &priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].tasklet_ctx; in mlx4_cq_alloc() 348 cq->irq = priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(vector)].irq; in mlx4_cq_alloc() 377 synchronize_irq(priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq->vector)].irq); in mlx4_cq_free() 378 if (priv->eq_table.eq[MLX4_CQ_TO_EQ_VECTOR(cq->vector)].irq != in mlx4_cq_free() 379 priv->eq_table.eq[MLX4_EQ_ASYNC].irq) in mlx4_cq_free() 380 synchronize_irq(priv->eq_table.eq[MLX4_EQ_ASYNC].irq); in mlx4_cq_free()
|
D | main.c | 1377 err = mlx4_init_icm_table(dev, &priv->eq_table.cmpt_table, in mlx4_init_cmpt_table() 1439 err = mlx4_init_icm_table(dev, &priv->eq_table.table, in mlx4_init_icm() 1587 mlx4_cleanup_icm_table(dev, &priv->eq_table.table); in mlx4_init_icm() 1590 mlx4_cleanup_icm_table(dev, &priv->eq_table.cmpt_table); in mlx4_init_icm() 1617 mlx4_cleanup_icm_table(dev, &priv->eq_table.table); in mlx4_free_icms() 1618 mlx4_cleanup_icm_table(dev, &priv->eq_table.cmpt_table); in mlx4_free_icms() 2215 priv->eq_table.inta_pin = adapter.inta_pin; in mlx4_init_hca() 2516 priv->eq_table.eq[MLX4_EQ_ASYNC].irq); in mlx4_setup_hca() 2520 priv->eq_table.eq[MLX4_EQ_ASYNC].irq); in mlx4_setup_hca() 2665 eq = &priv->eq_table.eq[eqn]; in mlx4_init_affinity_hint() [all …]
|
D | mlx4.h | 878 struct mlx4_eq_table eq_table; member
|
/drivers/net/ethernet/mellanox/mlx5/core/ |
D | eq.c | 440 spin_lock_init(&dev->priv.eq_table.lock); in mlx5_eq_init() 455 struct mlx5_eq_table *table = &dev->priv.eq_table; in mlx5_start_eqs() 503 struct mlx5_eq_table *table = &dev->priv.eq_table; in mlx5_stop_eqs()
|
D | main.c | 258 struct mlx5_eq_table *table = &priv->eq_table; in mlx5_enable_msix() 547 for (i = 0; i < mdev->priv.eq_table.num_comp_vectors; i++) { in mlx5_irq_set_affinity_hints() 566 for (i = 0; i < mdev->priv.eq_table.num_comp_vectors; i++) in mlx5_irq_clear_affinity_hints() 573 struct mlx5_eq_table *table = &dev->priv.eq_table; in mlx5_vector2eqn() 594 struct mlx5_eq_table *table = &dev->priv.eq_table; in free_comp_eqs() 612 struct mlx5_eq_table *table = &dev->priv.eq_table; in alloc_comp_eqs()
|
D | en.h | 630 return min_t(int, mdev->priv.eq_table.num_comp_vectors, in mlx5e_get_max_num_channels()
|
/drivers/infiniband/hw/mlx4/ |
D | main.c | 2080 ibdev->eq_table = kcalloc(dev->caps.num_comp_vectors, in mlx4_ib_alloc_eqs() 2081 sizeof(ibdev->eq_table[0]), GFP_KERNEL); in mlx4_ib_alloc_eqs() 2082 if (!ibdev->eq_table) in mlx4_ib_alloc_eqs() 2090 ibdev->eq_table[eq] = total_eqs; in mlx4_ib_alloc_eqs() 2092 &ibdev->eq_table[eq])) in mlx4_ib_alloc_eqs() 2095 ibdev->eq_table[eq] = -1; in mlx4_ib_alloc_eqs() 2100 ibdev->eq_table[i++] = -1) in mlx4_ib_alloc_eqs() 2113 if (!ibdev->eq_table) in mlx4_ib_free_eqs() 2120 mlx4_release_eq(dev, ibdev->eq_table[i]); in mlx4_ib_free_eqs() 2122 kfree(ibdev->eq_table); in mlx4_ib_free_eqs() [all …]
|
D | cq.c | 238 if (dev->eq_table) in mlx4_ib_create_cq() 239 vector = dev->eq_table[vector % ibdev->num_comp_vectors]; in mlx4_ib_create_cq()
|
D | mlx4_ib.h | 563 int *eq_table; member
|
/drivers/infiniband/hw/mlx5/ |
D | main.c | 1361 dev->mdev->priv.eq_table.num_comp_vectors; in mlx5_ib_add()
|