/drivers/infiniband/hw/cxgb4/ |
D | provider.c | 149 if ((addr >= pci_resource_start(rdev->lldi.pdev, 0)) && in c4iw_mmap() 150 (addr < (pci_resource_start(rdev->lldi.pdev, 0) + in c4iw_mmap() 151 pci_resource_len(rdev->lldi.pdev, 0)))) { in c4iw_mmap() 160 } else if ((addr >= pci_resource_start(rdev->lldi.pdev, 2)) && in c4iw_mmap() 161 (addr < (pci_resource_start(rdev->lldi.pdev, 2) + in c4iw_mmap() 162 pci_resource_len(rdev->lldi.pdev, 2)))) { in c4iw_mmap() 170 if (!is_t4(rdev->lldi.adapter_type)) in c4iw_mmap() 251 memcpy(&(gid->raw[0]), dev->rdev.lldi.ports[port-1]->dev_addr, 6); in c4iw_query_gid() 267 memcpy(&props->sys_image_guid, dev->rdev.lldi.ports[0]->dev_addr, 6); in c4iw_query_device() 268 props->hw_ver = CHELSIO_CHIP_RELEASE(dev->rdev.lldi.adapter_type); in c4iw_query_device() [all …]
|
D | device.c | 102 le.poll_sge_ts = cxgb4_read_sge_timestamp(wq->rdev->lldi.ports[0]); in c4iw_log_wr_stats() 130 #define ts2ns(ts) div64_u64((ts) * dev->rdev.lldi.cclk_ps, 1000) in wr_log_show() 383 ret = cxgb4_read_tpte(stagd->devp->rdev.lldi.ports[0], (u32)id<<8, in dump_stag() 386 dev_err(&stagd->devp->rdev.lldi.pdev->dev, in dump_stag() 796 if (rdev->lldi.udb_density != rdev->lldi.ucq_density) { in c4iw_rdev_open() 798 pci_name(rdev->lldi.pdev), rdev->lldi.udb_density, in c4iw_rdev_open() 799 rdev->lldi.ucq_density); in c4iw_rdev_open() 802 if (rdev->lldi.vr->qp.start != rdev->lldi.vr->cq.start || in c4iw_rdev_open() 803 rdev->lldi.vr->qp.size != rdev->lldi.vr->cq.size) { in c4iw_rdev_open() 805 pci_name(rdev->lldi.pdev), rdev->lldi.vr->qp.start, in c4iw_rdev_open() [all …]
|
D | cm.c | 212 error = cxgb4_l2t_send(rdev->lldi.ports[0], skb, l2e); in c4iw_l2t_send() 229 error = cxgb4_ofld_send(rdev->lldi.ports[0], skb); in c4iw_ofld_send() 250 ep->emss = ep->com.dev->rdev.lldi.mtus[TCPOPT_MSS_G(opt)] - in set_emss() 401 ep->com.dev->rdev.lldi.ports[0], in _c4iw_free_ep() 405 cxgb4_remove_tid(ep->com.dev->rdev.lldi.tids, 0, ep->hwtid, in _c4iw_free_ep() 558 cxgb4_clip_release(ep->com.dev->rdev.lldi.ports[0], in act_open_req_arp_failure() 562 cxgb4_free_atid(ep->com.dev->rdev.lldi.tids, ep->atid); in act_open_req_arp_failure() 618 (ep->com.dev->rdev.lldi.pf)); in send_flowc() 736 enum chip_type adapter_type = ep->com.dev->rdev.lldi.adapter_type; in send_connect() 741 netdev = ep->com.dev->rdev.lldi.ports[0]; in send_connect() [all …]
|
D | resource.c | 43 rdev->lldi.vr->qp.start, in c4iw_init_qid_table() 44 rdev->lldi.vr->qp.size, in c4iw_init_qid_table() 45 rdev->lldi.vr->qp.size, 0)) in c4iw_init_qid_table() 48 for (i = rdev->lldi.vr->qp.start; in c4iw_init_qid_table() 49 i < rdev->lldi.vr->qp.start + rdev->lldi.vr->qp.size; i++) in c4iw_init_qid_table() 308 pbl_start = rdev->lldi.vr->pbl.start; in c4iw_pblpool_create() 309 pbl_chunk = rdev->lldi.vr->pbl.size; in c4iw_pblpool_create() 350 pci_name(rdev->lldi.pdev)); in c4iw_rqtpool_alloc() 395 if (rdev->lldi.vr->srq.size) in c4iw_rqtpool_create() 398 rqt_start = rdev->lldi.vr->rq.start + skip; in c4iw_rqtpool_create() [all …]
|
D | mem.c | 56 return (is_t4(dev->rdev.lldi.adapter_type) || in mr_exceeds_hw_limits() 57 is_t5(dev->rdev.lldi.adapter_type)) && in mr_exceeds_hw_limits() 92 T5_ULP_MEMIO_FID_V(rdev->lldi.rxq_ids[0])); in _c4iw_write_mem_dma_aligned() 120 if (is_t4(rdev->lldi.adapter_type)) in _c4iw_write_mem_inline() 200 daddr = dma_map_single(&rdev->lldi.pdev->dev, data, len, DMA_TO_DEVICE); in _c4iw_write_mem_dma() 201 if (dma_mapping_error(&rdev->lldi.pdev->dev, daddr)) in _c4iw_write_mem_dma() 226 dma_unmap_single(&rdev->lldi.pdev->dev, save, len, DMA_TO_DEVICE); in _c4iw_write_mem_dma() 240 if (!rdev->lldi.ulptx_memwrite_dsgl || !use_dsgl) { in write_adapter_mem() 255 pci_name(rdev->lldi.pdev)); in write_adapter_mem() 333 (rdev->lldi.vr->stag.start >> 5), in write_tpt_entry() [all …]
|
D | qp.c | 74 dev_warn(&dev->rdev.lldi.pdev->dev, in alloc_ird() 102 dma_free_coherent(&(rdev->lldi.pdev->dev), sq->memsize, sq->queue, in dealloc_host_sq() 116 if (!ocqp_support || !ocqp_supported(&rdev->lldi)) in alloc_oc_sq() 122 rdev->lldi.vr->ocq.start; in alloc_oc_sq() 124 rdev->lldi.vr->ocq.start); in alloc_oc_sq() 131 sq->queue = dma_alloc_coherent(&(rdev->lldi.pdev->dev), sq->memsize, in alloc_host_sq() 162 dma_free_coherent(&rdev->lldi.pdev->dev, in destroy_qp() 184 ret = cxgb4_bar2_sge_qregs(rdev->lldi.ports[0], qid, qtype, in c4iw_bar2_addrs() 193 if (is_t4(rdev->lldi.adapter_type)) in c4iw_bar2_addrs() 264 wq->rq.queue = dma_alloc_coherent(&rdev->lldi.pdev->dev, in create_qp() [all …]
|
D | ev.c | 43 ret = cxgb4_read_tpte(dev->rdev.lldi.ports[0], stag, in print_tpte() 46 dev_err(&dev->rdev.lldi.pdev->dev, in print_tpte() 66 dev_err(&dev->rdev.lldi.pdev->dev, in dump_err_cqe()
|
D | iw_cxgb4.h | 76 #define PBL_OFF(rdev_p, a) ((a) - (rdev_p)->lldi.vr->pbl.start) 77 #define RQT_OFF(rdev_p, a) ((a) - (rdev_p)->lldi.vr->rq.start) 180 struct cxgb4_lld_info lldi; member 205 return (int)(rdev->lldi.vr->stag.size >> 5); in c4iw_num_stags() 274 func, pci_name(rdev->lldi.pdev), hwtid, qpid); in c4iw_wait_for_reply() 281 pci_name(rdev->lldi.pdev), wr_waitp->ret, hwtid, qpid); in c4iw_wait_for_reply() 334 struct cxgb4_lld_info lldi; member 363 return min(dev->rdev.lldi.max_ordird_qp, c4iw_max_read_depth); in cur_max_read_depth()
|
D | cq.c | 64 dma_free_coherent(&(rdev->lldi.pdev->dev), in destroy_cq() 98 cq->queue = dma_alloc_coherent(&rdev->lldi.pdev->dev, cq->memsize, in create_cq() 142 rdev->lldi.ciq_ids[cq->vector])); in create_cq() 160 cq->gts = rdev->lldi.gts_reg; in create_cq() 168 pci_name(rdev->lldi.pdev), cq->cqid); in create_cq() 174 dma_free_coherent(&rdev->lldi.pdev->dev, cq->memsize, cq->queue, in create_cq() 1014 if (vector >= rhp->rdev.lldi.nciq) in c4iw_create_cq()
|
/drivers/scsi/cxgbi/cxgb4i/ |
D | cxgb4i.c | 210 struct cxgb4_lld_info *lldi = cxgbi_cdev_priv(csk->cdev); in send_act_open_req() local 230 if (is_t4(lldi->adapter_type)) { in send_act_open_req() 253 } else if (is_t5(lldi->adapter_type)) { in send_act_open_req() 319 CHELSIO_CHIP_VERSION(lldi->adapter_type), csk, in send_act_open_req() 329 struct cxgb4_lld_info *lldi = cxgbi_cdev_priv(csk->cdev); in send_act_open_req6() local 349 if (is_t4(lldi->adapter_type)) { in send_act_open_req6() 374 } else if (is_t5(lldi->adapter_type)) { in send_act_open_req6() 430 CHELSIO_CHIP_VERSION(lldi->adapter_type), csk, csk->state, in send_act_open_req6() 717 struct cxgb4_lld_info *lldi = cxgbi_cdev_priv(cdev); in cxgb4i_make_tx_data_wr() local 725 u32 force = is_t5(lldi->adapter_type) ? TX_FORCE_V(!submode) : in cxgb4i_make_tx_data_wr() [all …]
|
/drivers/target/iscsi/cxgbit/ |
D | cxgbit_ddp.c | 298 struct cxgb4_lld_info *lldi = &cdev->lldi; in cxgbit_ddp_init() local 299 struct net_device *ndev = cdev->lldi.ports[0]; in cxgbit_ddp_init() 303 if (!lldi->vr->iscsi.size) { in cxgbit_ddp_init() 310 tformat.pgsz_order[i] = (lldi->iscsi_pgsz_order >> (i << 3)) in cxgbit_ddp_init() 312 cxgbi_tagmask_check(lldi->iscsi_tagmask, &tformat); in cxgbit_ddp_init() 314 ret = cxgbi_ppm_init(lldi->iscsi_ppm, cdev->lldi.ports[0], in cxgbit_ddp_init() 315 cdev->lldi.pdev, &cdev->lldi, &tformat, in cxgbit_ddp_init() 316 lldi->vr->iscsi.size, lldi->iscsi_llimit, in cxgbit_ddp_init() 317 lldi->vr->iscsi.start, 2, in cxgbit_ddp_init() 318 lldi->vr->ppod_edram.start, in cxgbit_ddp_init() [all …]
|
D | cxgbit_cm.c | 63 func, pci_name(cdev->lldi.pdev), tid); in cxgbit_wait_for_reply() 69 pci_name(cdev->lldi.pdev), wr_waitp->ret, tid); in cxgbit_wait_for_reply() 153 __func__, cdev->lldi.ports[0]->name, stid, sin6->sin6_port); in cxgbit_create_server6() 158 ret = cxgb4_clip_get(cdev->lldi.ports[0], in cxgbit_create_server6() 170 ret = cxgb4_create_server6(cdev->lldi.ports[0], in cxgbit_create_server6() 173 cdev->lldi.rxq_ids[0]); in cxgbit_create_server6() 184 cxgb4_clip_release(cdev->lldi.ports[0], in cxgbit_create_server6() 204 __func__, cdev->lldi.ports[0]->name, stid, sin->sin_port); in cxgbit_create_server4() 209 ret = cxgb4_create_server(cdev->lldi.ports[0], in cxgbit_create_server4() 212 cdev->lldi.rxq_ids[0]); in cxgbit_create_server4() [all …]
|
D | cxgbit_main.c | 33 struct cxgb4_lld_info *lldi = &cdev->lldi; in cxgbit_set_mdsl() local 38 mdsl = min_t(u32, lldi->iscsi_iolen - ISCSI_PDU_NONPAYLOAD_LEN, in cxgbit_set_mdsl() 46 static void *cxgbit_uld_add(const struct cxgb4_lld_info *lldi) in cxgbit_uld_add() argument 50 if (is_t4(lldi->adapter_type)) in cxgbit_uld_add() 60 cdev->lldi = *lldi; in cxgbit_uld_add() 71 pci_name(lldi->pdev)); in cxgbit_uld_add() 73 if (lldi->fw_vers >= 0x10d2b00) in cxgbit_uld_add() 84 pci_name(lldi->pdev)); in cxgbit_uld_add() 142 pr_info("cdev %s state UP.\n", pci_name(cdev->lldi.pdev)); in cxgbit_uld_state_change() 147 pr_info("cdev %s state RECOVERY.\n", pci_name(cdev->lldi.pdev)); in cxgbit_uld_state_change() [all …]
|
D | cxgbit.h | 133 struct cxgb4_lld_info lldi; member 349 return (struct cxgbi_ppm *)(*cdev->lldi.iscsi_ppm); in cdev2ppm()
|
D | cxgbit_target.c | 163 const struct cxgb4_lld_info *lldi = &csk->com.cdev->lldi; in cxgbit_tx_data_wr() local 169 u32 force = is_t5(lldi->adapter_type) ? TX_FORCE_V(!submode) : in cxgbit_tx_data_wr() 719 struct cxgbi_ppm *ppm = *csk->com.cdev->lldi.iscsi_ppm; in cxgbit_set_params() 750 if (is_t5(cdev->lldi.adapter_type)) in cxgbit_set_params() 1562 struct cxgb4_lld_info *lldi = &csk->com.cdev->lldi; in cxgbit_rx_skb() local 1566 if (is_t5(lldi->adapter_type)) in cxgbit_rx_skb()
|
/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
D | chtls_main.c | 136 for (i = 0; i < cdev->lldi->nports; i++) { in chtls_inline_feature() 170 kfree(cdev->lldi); in chtls_free_uld() 197 strlcat(tlsdev->name, cdev->lldi->ports[0]->name, in chtls_register_dev() 237 struct cxgb4_lld_info *lldi; in chtls_uld_add() local 245 lldi = kzalloc(sizeof(*lldi), GFP_KERNEL); in chtls_uld_add() 246 if (!lldi) in chtls_uld_add() 252 *lldi = *info; in chtls_uld_add() 253 cdev->lldi = lldi; in chtls_uld_add() 254 cdev->pdev = lldi->pdev; in chtls_uld_add() 255 cdev->tids = lldi->tids; in chtls_uld_add() [all …]
|
D | chtls_cm.c | 209 cxgb4_ofld_send(cdev->lldi->ports[0], skb); in abort_arp_failure() 691 cdev->lldi->rxq_ids[0]); in chtls_listen_start() 707 cdev->lldi->rxq_ids[0]); in chtls_listen_start() 742 cxgb4_remove_server(cdev->lldi->ports[0], stid, in chtls_listen_stop() 743 cdev->lldi->rxq_ids[0], sk->sk_family == PF_INET6); in chtls_listen_stop() 979 tp->advmss = cxgb4_best_aligned_mtu(cdev->lldi->mtus, in chtls_select_mss() 1011 struct cxgb4_lld_info *lldi; in chtls_pass_accept_rpl() local 1024 lldi = csk->cdev->lldi; in chtls_pass_accept_rpl() 1048 if (!is_t5(lldi->adapter_type)) in chtls_pass_accept_rpl() 1180 csk->l2t_entry = cxgb4_l2t_get(cdev->lldi->l2t, n, ndev, 0); in chtls_recv_sock() [all …]
|
D | chtls_hw.c | 158 int chtls_init_kmap(struct chtls_dev *cdev, struct cxgb4_lld_info *lldi) in chtls_init_kmap() argument 163 num_key_ctx = (lldi->vr->key.size / TLS_KEY_CONTEXT_SZ); in chtls_init_kmap() 173 cdev->kmap.start = lldi->vr->key.start; in chtls_init_kmap()
|
D | chtls.h | 212 struct cxgb4_lld_info *lldi; member 559 int chtls_init_kmap(struct chtls_dev *cdev, struct cxgb4_lld_info *lldi);
|
/drivers/net/ethernet/chelsio/libcxgb/ |
D | libcxgb_cm.c | 78 cxgb_our_interface(struct cxgb4_lld_info *lldi, in cxgb_our_interface() argument 85 for (i = 0; i < lldi->nports; i++) in cxgb_our_interface() 86 if (lldi->ports[i] == egress_dev) in cxgb_our_interface() 92 cxgb_find_route(struct cxgb4_lld_info *lldi, in cxgb_find_route() argument 109 if (!cxgb_our_interface(lldi, get_real_dev, n->dev) && in cxgb_find_route() 121 cxgb_find_route6(struct cxgb4_lld_info *lldi, in cxgb_find_route6() argument 138 (!cxgb_our_interface(lldi, get_real_dev, in cxgb_find_route6()
|
/drivers/crypto/chelsio/ |
D | chcr_algo.c | 774 qid = u_ctx->lldi.rxq_ids[rxqidx]; in create_wreq() 775 fid = u_ctx->lldi.rxq_ids[0]; in create_wreq() 778 rx_channel_id = cxgb4_port_e2cchan(u_ctx->lldi.ports[portno]); in create_wreq() 826 rx_channel_id = cxgb4_port_e2cchan(u_ctx->lldi.ports[rx_channel_id]); in create_cipher_wr() 1181 chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, in chcr_handle_cipher_resp() 1204 chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, in chcr_handle_cipher_resp() 1216 wrparam.qid = u_ctx->lldi.rxq_ids[reqctx->rxqidx]; in chcr_handle_cipher_resp() 1225 skb->dev = u_ctx->lldi.ports[0]; in chcr_handle_cipher_resp() 1237 chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, req); in chcr_handle_cipher_resp() 1281 err = chcr_cipher_dma_map(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, req); in process_cipher() [all …]
|
D | chcr_core.h | 108 struct cxgb4_lld_info lldi; member 128 return pci_get_drvdata(u_ctx->lldi.pdev); in padap()
|
/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/ |
D | chcr_ktls.h | 77 struct cxgb4_lld_info lldi; member 99 return u_ctx->lldi.rxq_ids[0]; in chcr_get_first_rx_qid()
|
/drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/ |
D | chcr_ipsec.h | 23 struct cxgb4_lld_info lldi; member
|
D | chcr_ipsec.c | 114 u_ctx->lldi = *infop; in ch_ipsec_uld_add() 126 pr_info("%s: Up\n", pci_name(u_ctx->lldi.pdev)); in ch_ipsec_uld_state_change() 134 pr_info("%s: Down\n", pci_name(u_ctx->lldi.pdev)); in ch_ipsec_uld_state_change() 809 adap = pci_get_drvdata(u_ctx->lldi.pdev); in ch_ipsec_exit()
|