/drivers/infiniband/hw/amso1100/ |
D | c2_mq.c | 46 (struct c2wr_hdr *) (q->msg_pool.host + q->priv * q->msg_size); in c2_mq_alloc() 53 return q->msg_pool.host + q->priv * q->msg_size; in c2_mq_alloc() 81 (q->msg_pool.host + q->priv * q->msg_size); in c2_mq_consume() 87 return q->msg_pool.host + q->priv * q->msg_size; in c2_mq_consume() 102 (q->msg_pool.adapter + q->priv * q->msg_size); in c2_mq_free() 141 void c2_mq_req_init(struct c2_mq *q, u32 index, u32 q_size, u32 msg_size, in c2_mq_req_init() argument 149 q->msg_size = msg_size; in c2_mq_req_init() 158 void c2_mq_rep_init(struct c2_mq *q, u32 index, u32 q_size, u32 msg_size, in c2_mq_rep_init() argument 166 q->msg_size = msg_size; in c2_mq_rep_init()
|
D | c2_cq.c | 99 (q->msg_pool.host + priv * q->msg_size); in c2_cq_clean() 259 dma_free_coherent(&c2dev->pcidev->dev, mq->q_size * mq->msg_size, in c2_free_cq_buf() 264 size_t q_size, size_t msg_size) in c2_alloc_cq_buf() argument 268 if (q_size > SIZE_MAX / msg_size) in c2_alloc_cq_buf() 271 pool_start = dma_alloc_coherent(&c2dev->pcidev->dev, q_size * msg_size, in c2_alloc_cq_buf() 279 msg_size, in c2_alloc_cq_buf() 324 wr.msg_size = cpu_to_be32(cq->mq.msg_size); in c2_init_cq()
|
D | c2_qp.c | 425 u32 q_size, msg_size, mmap_size; in c2_alloc_qp() local 516 msg_size = be32_to_cpu(reply->sq_msg_size); in c2_alloc_qp() 518 mmap_size = PAGE_ALIGN(sizeof(struct c2_mq_shared) + msg_size * q_size); in c2_alloc_qp() 528 msg_size, in c2_alloc_qp() 535 msg_size = be32_to_cpu(reply->rq_msg_size); in c2_alloc_qp() 537 mmap_size = PAGE_ALIGN(sizeof(struct c2_mq_shared) + msg_size * q_size); in c2_alloc_qp() 547 msg_size, in c2_alloc_qp() 802 u32 msg_size; in c2_post_send() local 835 msg_size = sizeof(struct c2wr_send_req) + in c2_post_send() 853 msg_size = sizeof(struct c2wr_rdma_write_req) + in c2_post_send() [all …]
|
D | c2_mq.h | 81 u32 msg_size; member 101 extern void c2_mq_req_init(struct c2_mq *q, u32 index, u32 q_size, u32 msg_size, 103 extern void c2_mq_rep_init(struct c2_mq *q, u32 index, u32 q_size, u32 msg_size,
|
D | c2_mm.c | 72 pbe_count = (c2dev->req_vq.msg_size - in send_pbl_messages() 74 wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL); in send_pbl_messages() 203 wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL); in c2_nsmr_register_phys_kern() 221 pbe_count = (c2dev->req_vq.msg_size - in c2_nsmr_register_phys_kern()
|
D | c2_rnic.c | 225 wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL); in c2_add_addr() 287 wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL); in c2_del_addr() 598 c2dev->aeq.q_size * c2dev->aeq.msg_size, in c2_rnic_init() 602 c2dev->rep_vq.q_size * c2dev->rep_vq.msg_size, in c2_rnic_init() 638 c2dev->aeq.q_size * c2dev->aeq.msg_size, in c2_rnic_term() 644 c2dev->rep_vq.q_size * c2dev->rep_vq.msg_size, in c2_rnic_term()
|
D | c2_vq.c | 87 kmem_cache_create(c2dev->vq_cache_name, c2dev->rep_vq.msg_size, 0, in vq_init() 226 memcpy(msg, wr, c2dev->req_vq.msg_size); in vq_send_wr()
|
D | c2_cm.c | 81 wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL); in c2_llp_connect() 321 wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL); in c2_llp_accept()
|
/drivers/crypto/qat/qat_common/ |
D | qat_crypto.c | 153 int msg_size; in qat_crypto_create_instances() local 198 msg_size = ICP_QAT_FW_REQ_DEFAULT_SZ; in qat_crypto_create_instances() 201 msg_size, key, NULL, 0, &inst->sym_tx)) in qat_crypto_create_instances() 206 msg_size, key, NULL, 0, &inst->rnd_tx)) in qat_crypto_create_instances() 209 msg_size = msg_size >> 1; in qat_crypto_create_instances() 212 msg_size, key, NULL, 0, &inst->pke_tx)) in qat_crypto_create_instances() 215 msg_size = ICP_QAT_FW_RESP_DEFAULT_SZ; in qat_crypto_create_instances() 218 msg_size, key, qat_alg_callback, 0, in qat_crypto_create_instances() 224 msg_size, key, qat_alg_callback, 0, in qat_crypto_create_instances() 230 msg_size, key, qat_alg_callback, 0, in qat_crypto_create_instances()
|
D | adf_transport.c | 69 static int adf_verify_ring_size(uint32_t msg_size, uint32_t msg_num) in adf_verify_ring_size() argument 74 if ((msg_size * msg_num) == ADF_SIZE_TO_RING_SIZE_IN_BYTES(i)) in adf_verify_ring_size() 120 ADF_MAX_INFLIGHTS(ring->ring_size, ring->msg_size)) { in adf_send_message() 126 ADF_MSG_SIZE_TO_BYTES(ring->msg_size)); in adf_send_message() 129 ADF_MSG_SIZE_TO_BYTES(ring->msg_size), in adf_send_message() 146 ADF_MSG_SIZE_TO_BYTES(ring->msg_size), in adf_handle_response() 233 uint32_t msg_size, const char *ring_name, in adf_create_ring() argument 248 if (msg_size > ADF_MSG_SIZE_TO_BYTES(ADF_MAX_MSG_SIZE)) { in adf_create_ring() 252 if (ADF_MAX_INFLIGHTS(adf_verify_ring_size(msg_size, num_msgs), in adf_create_ring() 253 ADF_BYTES_TO_MSG_SIZE(msg_size)) < 2) { in adf_create_ring() [all …]
|
D | adf_transport_debug.c | 66 ADF_MSG_SIZE_TO_BYTES(ring->msg_size))) in adf_ring_start() 70 (ADF_MSG_SIZE_TO_BYTES(ring->msg_size) * (*pos)++); in adf_ring_start() 78 ADF_MSG_SIZE_TO_BYTES(ring->msg_size))) in adf_ring_next() 82 (ADF_MSG_SIZE_TO_BYTES(ring->msg_size) * (*pos)++); in adf_ring_next() 110 ADF_MSG_SIZE_TO_BYTES(ring->msg_size)); in adf_ring_show() 117 for (; i < (ADF_MSG_SIZE_TO_BYTES(ring->msg_size) >> 2); i++) { in adf_ring_show() 119 if ((ADF_MSG_SIZE_TO_BYTES(ring->msg_size) >> 2) != i + 1 && in adf_ring_show()
|
D | adf_transport.h | 57 uint32_t bank_num, uint32_t num_mgs, uint32_t msg_size,
|
D | adf_transport_internal.h | 71 uint8_t msg_size; member
|
/drivers/misc/sgi-xp/ |
D | xpc_uv.c | 334 size_t msg_size) in xpc_send_gru_msg() argument 340 ret = gru_send_message_gpa(gru_mq_desc, msg, msg_size); in xpc_send_gru_msg() 672 xpc_send_activate_IRQ_uv(struct xpc_partition *part, void *msg, size_t msg_size, in xpc_send_activate_IRQ_uv() argument 681 DBUG_ON(msg_size > XPC_ACTIVATE_MSG_SIZE_UV); in xpc_send_activate_IRQ_uv() 716 msg_size); in xpc_send_activate_IRQ_uv() 729 size_t msg_size, int msg_type) in xpc_send_activate_IRQ_part_uv() argument 733 ret = xpc_send_activate_IRQ_uv(part, msg, msg_size, msg_type); in xpc_send_activate_IRQ_part_uv() 740 void *msg, size_t msg_size, int msg_type) in xpc_send_activate_IRQ_ch_uv() argument 745 ret = xpc_send_activate_IRQ_uv(part, msg, msg_size, msg_type); in xpc_send_activate_IRQ_ch_uv() 1538 size_t msg_size; in xpc_send_payload_uv() local [all …]
|
D | xpnet.c | 381 u16 msg_size = sizeof(struct xpnet_message); in xpnet_send() local 391 msg_size += embedded_bytes - 1; in xpnet_send() 410 msg_size, xpnet_send_completed, queued_msg); in xpnet_send()
|
/drivers/isdn/hysdn/ |
D | boardergo.c | 327 int msg_size; in ergo_waitpofready() local 345 msg_size = dpr->ToPcSize - RDY_MAGIC_SIZE; in ergo_waitpofready() 346 if (msg_size > 0) in ergo_waitpofready() 347 if (EvalSysrTokData(card, dpr->ToPcBuf + RDY_MAGIC_SIZE, msg_size)) in ergo_waitpofready()
|
/drivers/net/ethernet/brocade/bna/ |
D | bfa_msgq.h | 45 (_cmd)->msg_size = (_msg_size); \ 57 size_t msg_size; member
|
/drivers/net/wimax/i2400m/ |
D | control.c | 600 size_t msg_size) in i2400m_msg_size_check() argument 606 i2400m, l3l4_hdr, msg_size); in i2400m_msg_size_check() 607 if (msg_size < sizeof(*l3l4_hdr)) { in i2400m_msg_size_check() 610 (size_t) sizeof(*l3l4_hdr), msg_size); in i2400m_msg_size_check() 615 if (msg_size < expected_size) { in i2400m_msg_size_check() 618 expected_size, msg_size); in i2400m_msg_size_check() 625 i2400m, l3l4_hdr, msg_size, result); in i2400m_msg_size_check()
|
/drivers/net/hyperv/ |
D | netvsc.c | 694 u32 msg_size = 0; in netvsc_copy_to_send_buf() local 702 msg_size += len; in netvsc_copy_to_send_buf() 705 return msg_size; in netvsc_copy_to_send_buf() 718 u32 msg_size = 0; in netvsc_send() local 741 msg_size = netvsc_copy_to_send_buf(net_device, in netvsc_send() 754 sendMessage.msg.v1_msg.send_rndis_pkt.send_buf_section_size = msg_size; in netvsc_send()
|
/drivers/net/ethernet/intel/i40evf/ |
D | i40evf_virtchnl.c | 94 event.msg_size = I40EVF_MAX_AQ_BUF_SIZE; in i40evf_verify_api_ver() 95 event.msg_buf = kzalloc(event.msg_size, GFP_KERNEL); in i40evf_verify_api_ver() 161 event.msg_size = len; in i40evf_get_vf_config() 162 event.msg_buf = kzalloc(event.msg_size, GFP_KERNEL); in i40evf_get_vf_config() 191 memcpy(adapter->vf_res, event.msg_buf, min(event.msg_size, len)); in i40evf_get_vf_config()
|
D | i40e_adminq.h | 79 u16 msg_size; member
|
/drivers/block/ |
D | sx8.c | 448 static inline int carm_lookup_bucket(u32 msg_size) in carm_lookup_bucket() argument 453 if (msg_size <= msg_sizes[i]) in carm_lookup_bucket() 643 unsigned int idx, msg_size; in carm_send_special() local 654 msg_size = func(host, idx, mem); in carm_send_special() 659 rc = carm_lookup_bucket(msg_size); in carm_send_special() 847 unsigned int msg_size; in carm_rq_fn() local 913 msg_size = sizeof(struct carm_msg_rw) - sizeof(msg->sg); in carm_rq_fn() 918 msg_size += sizeof(struct carm_msg_sg); in carm_rq_fn() 921 rc = carm_lookup_bucket(msg_size); in carm_rq_fn()
|
/drivers/misc/vmw_vmci/ |
D | vmci_guest.c | 171 u32 msg_size = sizeof(struct vmci_resource_query_hdr) + in vmci_check_host_caps() local 175 check_msg = kmalloc(msg_size, GFP_KERNEL); in vmci_check_host_caps() 184 check_msg->payload_size = msg_size - VMCI_DG_HEADERSIZE; in vmci_check_host_caps()
|
/drivers/usb/host/whci/ |
D | int.c | 77 dn->msg_size); in process_dn_buf()
|
/drivers/net/ethernet/intel/i40e/ |
D | i40e_adminq.h | 79 u16 msg_size; member
|