/drivers/infiniband/hw/hfi1/ |
D | user_sdma.c | 423 if (!info.npkts || req->data_iovs > MAX_VECTORS_PER_REQ) { in hfi1_user_sdma_process_request() 512 if (pcount > req->info.npkts) in hfi1_user_sdma_process_request() 513 pcount = req->info.npkts; in hfi1_user_sdma_process_request() 562 if (req->info.npkts > 1 && HFI1_CAP_IS_USET(SDMA_AHG)) in hfi1_user_sdma_process_request() 578 while (req->seqsubmitted != req->info.npkts) { in hfi1_user_sdma_process_request() 598 if (req->seqsubmitted < req->info.npkts) { in hfi1_user_sdma_process_request() 760 unsigned npkts = 0; in user_sdma_send_pkts() local 777 if (unlikely(req->seqnum == req->info.npkts)) { in user_sdma_send_pkts() 783 if (!maxpkts || maxpkts > req->info.npkts - req->seqnum) in user_sdma_send_pkts() 784 maxpkts = req->info.npkts - req->seqnum; in user_sdma_send_pkts() [all …]
|
D | tid_rdma.c | 1344 flow->npkts += (npages + pmtu_pg - 1) >> ilog2(pmtu_pg); in kern_program_rcv_group() 1405 flow->npkts = 0; in kern_program_rcvarray() 1516 flow->flow_state.lpsn = flow->flow_state.spsn + flow->npkts - 1; in hfi1_kern_exp_rcv_setup() 1519 qpriv->flow_state.psn += flow->npkts; in hfi1_kern_exp_rcv_setup() 1791 u32 npkts = rvt_div_round_up_mtu(qp, *len); in hfi1_build_tid_rdma_read_req() local 1832 if ((qpriv->flow_state.psn + npkts) > MAX_TID_FLOW_PSN - 1) { in hfi1_build_tid_rdma_read_req() 1865 flow->flow_state.ib_spsn + flow->npkts - 1; in hfi1_build_tid_rdma_read_req() 1869 req->s_next_psn += flow->npkts; in hfi1_build_tid_rdma_read_req() 1909 flow->npkts = rvt_div_round_up_mtu(qp, len); in tid_rdma_rcv_read_request() 1943 flow->npkts - 1; in tid_rdma_rcv_read_request() [all …]
|
D | trace_tx.h | 596 __field(u16, npkts) 606 __entry->npkts = i[1]; 616 __entry->npkts,
|
D | tid_rdma.h | 189 u8 npkts; member
|
D | chip.h | 796 u32 intr_adjust, u32 npkts);
|
D | chip.c | 11751 static void adjust_rcv_timeout(struct hfi1_ctxtdata *rcd, u32 npkts) in adjust_rcv_timeout() argument 11765 if (npkts < rcv_intr_count) { in adjust_rcv_timeout() 11794 u32 intr_adjust, u32 npkts) in update_usrhead() argument 11805 adjust_rcv_timeout(rcd, npkts); in update_usrhead()
|
/drivers/input/joystick/ |
D | grip_mp.c | 539 int i, npkts, flags; in grip_poll() local 541 for (npkts = 0; npkts < 4; npkts++) { in grip_poll()
|
/drivers/net/ethernet/sun/ |
D | sunvnet_common.c | 587 u32 start, u32 end, int *npkts, int budget) in vnet_walk_rx() argument 605 (*npkts)++; in vnet_walk_rx() 617 if ((*npkts) >= budget) { in vnet_walk_rx() 630 ack_end, *npkts); in vnet_walk_rx() 636 ack_end, *npkts); in vnet_walk_rx() 643 static int vnet_rx(struct vnet_port *port, void *msgbuf, int *npkts, in vnet_rx() argument 667 npkts, budget); in vnet_rx() 784 int npkts = 0; in vnet_event_napi() local 883 err = vnet_rx(port, &msgbuf, &npkts, budget); in vnet_event_napi() 884 if (npkts >= budget) in vnet_event_napi() [all …]
|
/drivers/infiniband/hw/qib/ |
D | qib_user_sdma.c | 812 int npkts = 0; in qib_user_sdma_queue_pkts() local 821 while (idx < niov && npkts < *maxpkts) { in qib_user_sdma_queue_pkts() 1012 npkts++; in qib_user_sdma_queue_pkts() 1020 *maxpkts = npkts; in qib_user_sdma_queue_pkts() 1382 int npkts = 0; in qib_user_sdma_writev() local 1427 npkts += mxp; in qib_user_sdma_writev() 1438 return (ret < 0) ? ret : npkts; in qib_user_sdma_writev()
|
D | qib_driver.c | 440 u32 qib_kreceive(struct qib_ctxtdata *rcd, u32 *llic, u32 *npkts) in qib_kreceive() argument 574 if (npkts) in qib_kreceive() 575 *npkts = i; in qib_kreceive()
|
D | qib_iba7322.c | 3039 static void adjust_rcv_timeout(struct qib_ctxtdata *rcd, int npkts) in adjust_rcv_timeout() argument 3048 if (npkts < rcv_int_count && timeout > 2) in adjust_rcv_timeout() 3050 else if (npkts >= rcv_int_count && timeout < rcv_int_timeout) in adjust_rcv_timeout() 3075 u32 npkts; in qib_7322intr() local 3134 qib_kreceive(dd->rcd[i], NULL, &npkts); in qib_7322intr() 3163 u32 npkts; in qib_7322pintr() local 3180 qib_kreceive(rcd, NULL, &npkts); in qib_7322pintr() 4432 u32 updegr, u32 egrhd, u32 npkts) in qib_update_7322_usrhead() argument 4439 adjust_rcv_timeout(rcd, npkts); in qib_update_7322_usrhead()
|
D | qib_iba6120.c | 2050 u32 updegr, u32 egrhd, u32 npkts) in qib_update_6120_usrhead() argument
|
D | qib_iba7220.c | 2702 u32 updegr, u32 egrhd, u32 npkts) in qib_update_7220_usrhead() argument
|
/drivers/net/ethernet/mellanox/mlx5/core/ |
D | en_tx.c | 426 u16 npkts; in mlx5e_poll_tx_cq() local 441 npkts = 0; in mlx5e_poll_tx_cq() 507 npkts++; in mlx5e_poll_tx_cq() 525 netdev_tx_completed_queue(sq->txq, npkts, nbytes); in mlx5e_poll_tx_cq()
|
/drivers/net/ethernet/socionext/ |
D | sni_ave.c | 748 int restpkt, npkts; in ave_rx_receive() local 756 for (npkts = 0; npkts < num; npkts++) { in ave_rx_receive() 813 return npkts; in ave_rx_receive()
|