/drivers/infiniband/hw/mlx5/ |
D | mem.c | 152 __be64 *pas, int access_flags) in __mlx5_ib_populate_pas() argument 173 pas[i] = cpu_to_be64(umem_dma_to_mtt(pa)); in __mlx5_ib_populate_pas() 188 pas[i >> shift] = cpu_to_be64(cur); in __mlx5_ib_populate_pas() 190 i >> shift, be64_to_cpu(pas[i >> shift])); in __mlx5_ib_populate_pas() 200 int page_shift, __be64 *pas, int access_flags) in mlx5_ib_populate_pas() argument 203 ib_umem_num_pages(umem), pas, in mlx5_ib_populate_pas()
|
D | mr.c | 801 __be64 *pas; in reg_umr() local 831 pas = PTR_ALIGN(mr_pas, MLX5_UMR_ALIGN); in reg_umr() 832 mlx5_ib_populate_pas(dev, umem, page_shift, pas, MLX5_IB_MTT_PRESENT); in reg_umr() 834 memset(pas + npages, 0, size - npages * sizeof(u64)); in reg_umr() 836 dma = dma_map_single(ddev, pas, size, DMA_TO_DEVICE); in reg_umr() 892 __be64 *pas; in mlx5_ib_update_mtt() local 922 pas = (__be64 *)get_zeroed_page(GFP_ATOMIC); in mlx5_ib_update_mtt() 923 if (!pas) { in mlx5_ib_update_mtt() 925 pas = mlx5_ib_update_mtt_emergency_buffer; in mlx5_ib_update_mtt() 929 memset(pas, 0, size); in mlx5_ib_update_mtt() [all …]
|
D | srq.c | 122 *inlen = sizeof(**in) + sizeof(*(*in)->pas) * ncont; in create_srq_user() 129 mlx5_ib_populate_pas(dev, srq->umem, page_shift, (*in)->pas, 0); in create_srq_user() 188 *inlen = sizeof(**in) + sizeof(*(*in)->pas) * npages; in create_srq_kernel() 194 mlx5_fill_page_array(&srq->buf, (*in)->pas); in create_srq_kernel()
|
D | cq.c | 658 *inlen = sizeof(**cqb) + sizeof(*(*cqb)->pas) * ncont; in create_cq_user() 664 mlx5_ib_populate_pas(dev, cq->buf.umem, page_shift, (*cqb)->pas, 0); in create_cq_user() 719 *inlen = sizeof(**cqb) + sizeof(*(*cqb)->pas) * cq->buf.buf.npages; in create_cq_kernel() 725 mlx5_fill_page_array(&cq->buf.buf, (*cqb)->pas); in create_cq_kernel() 1133 inlen = sizeof(*in) + npas * sizeof(in->pas[0]); in mlx5_ib_resize_cq() 1142 in->pas, 0); in mlx5_ib_resize_cq() 1144 mlx5_fill_page_array(&cq->resize_buf->buf, in->pas); in mlx5_ib_resize_cq()
|
D | mlx5_ib.h | 596 __be64 *pas, int access_flags); 598 int page_shift, __be64 *pas, int access_flags);
|
D | qp.c | 677 *inlen = sizeof(**in) + sizeof(*(*in)->pas) * ncont; in create_user_qp() 684 mlx5_ib_populate_pas(dev, qp->umem, page_shift, (*in)->pas, 0); in create_user_qp() 778 *inlen = sizeof(**in) + sizeof(*(*in)->pas) * qp->buf.npages; in create_kernel_qp() 791 mlx5_fill_page_array(&qp->buf, (*in)->pas); in create_kernel_qp()
|
/drivers/net/ethernet/mellanox/mlx5/core/ |
D | pagealloc.c | 86 __be64 pas[0]; member 93 __be64 pas[0]; member 323 inlen = sizeof(*in) + npages * sizeof(in->pas[0]); in give_pages() 343 in->pas[i] = cpu_to_be64(addr); in give_pages() 372 free_4k(dev, be64_to_cpu(in->pas[i])); in give_pages() 395 outlen = sizeof(*out) + npages * sizeof(out->pas[0]); in reclaim_pages() 422 addr = be64_to_cpu(out->pas[i]); in reclaim_pages() 495 FIELD_SIZEOF(struct mlx5_manage_pages_outbox, pas[0]); in optimal_reclaimed_pages()
|
D | srq.c | 226 void *pas; in create_xrc_srq_cmd() local 240 pas = MLX5_ADDR_OF(create_xrc_srq_in, create_in, pas); in create_xrc_srq_cmd() 243 memcpy(pas, in->pas, pas_size); in create_xrc_srq_cmd() 349 memcpy(MLX5_ADDR_OF(rmpc, rmpc, wq.pas), in->pas, pas_size); in create_rmp_cmd()
|
D | alloc.c | 200 void mlx5_fill_page_array(struct mlx5_buf *buf, __be64 *pas) in mlx5_fill_page_array() argument 208 pas[i] = cpu_to_be64(addr); in mlx5_fill_page_array()
|
D | eq.c | 356 inlen = sizeof(*in) + sizeof(in->pas[0]) * eq->buf.npages; in mlx5_create_map_eq() 364 mlx5_fill_page_array(&eq->buf, in->pas); in mlx5_create_map_eq()
|
D | en_main.c | 401 (__be64 *)MLX5_ADDR_OF(wq, wq, pas)); in mlx5e_enable_rq() 634 (__be64 *)MLX5_ADDR_OF(wq, wq, pas)); in mlx5e_enable_sq() 818 (__be64 *)MLX5_ADDR_OF(create_cq_in, in, pas)); in mlx5e_enable_cq()
|
/drivers/misc/vmw_vmci/ |
D | vmci_queue_pair.c | 157 dma_addr_t *pas; member 282 queue->kernel_if->u.g.pas[i]); in qp_free_queue() 308 (sizeof(*queue->kernel_if->u.g.pas) + in qp_alloc_queue() 312 pas_size = num_pages * sizeof(*queue->kernel_if->u.g.pas); in qp_alloc_queue() 325 queue->kernel_if->u.g.pas = (dma_addr_t *)(queue->kernel_if + 1); in qp_alloc_queue() 327 (void **)((u8 *)queue->kernel_if->u.g.pas + pas_size); in qp_alloc_queue() 333 &queue->kernel_if->u.g.pas[i], in qp_alloc_queue() 508 produce_q->kernel_if->u.g.pas[i] >> PAGE_SHIFT; in qp_alloc_ppn_set() 521 consume_q->kernel_if->u.g.pas[i] >> PAGE_SHIFT; in qp_alloc_ppn_set()
|