Home
last modified time | relevance | path

Searched refs:lldi (Results 1 – 25 of 29) sorted by relevance

12

/drivers/infiniband/hw/cxgb4/
Dprovider.c149 if ((addr >= pci_resource_start(rdev->lldi.pdev, 0)) && in c4iw_mmap()
150 (addr < (pci_resource_start(rdev->lldi.pdev, 0) + in c4iw_mmap()
151 pci_resource_len(rdev->lldi.pdev, 0)))) { in c4iw_mmap()
160 } else if ((addr >= pci_resource_start(rdev->lldi.pdev, 2)) && in c4iw_mmap()
161 (addr < (pci_resource_start(rdev->lldi.pdev, 2) + in c4iw_mmap()
162 pci_resource_len(rdev->lldi.pdev, 2)))) { in c4iw_mmap()
170 if (!is_t4(rdev->lldi.adapter_type)) in c4iw_mmap()
251 memcpy(&(gid->raw[0]), dev->rdev.lldi.ports[port-1]->dev_addr, 6); in c4iw_query_gid()
267 memcpy(&props->sys_image_guid, dev->rdev.lldi.ports[0]->dev_addr, 6); in c4iw_query_device()
268 props->hw_ver = CHELSIO_CHIP_RELEASE(dev->rdev.lldi.adapter_type); in c4iw_query_device()
[all …]
Ddevice.c102 le.poll_sge_ts = cxgb4_read_sge_timestamp(wq->rdev->lldi.ports[0]); in c4iw_log_wr_stats()
130 #define ts2ns(ts) div64_u64((ts) * dev->rdev.lldi.cclk_ps, 1000) in wr_log_show()
383 ret = cxgb4_read_tpte(stagd->devp->rdev.lldi.ports[0], (u32)id<<8, in dump_stag()
386 dev_err(&stagd->devp->rdev.lldi.pdev->dev, in dump_stag()
796 if (rdev->lldi.udb_density != rdev->lldi.ucq_density) { in c4iw_rdev_open()
798 pci_name(rdev->lldi.pdev), rdev->lldi.udb_density, in c4iw_rdev_open()
799 rdev->lldi.ucq_density); in c4iw_rdev_open()
802 if (rdev->lldi.vr->qp.start != rdev->lldi.vr->cq.start || in c4iw_rdev_open()
803 rdev->lldi.vr->qp.size != rdev->lldi.vr->cq.size) { in c4iw_rdev_open()
805 pci_name(rdev->lldi.pdev), rdev->lldi.vr->qp.start, in c4iw_rdev_open()
[all …]
Dcm.c212 error = cxgb4_l2t_send(rdev->lldi.ports[0], skb, l2e); in c4iw_l2t_send()
229 error = cxgb4_ofld_send(rdev->lldi.ports[0], skb); in c4iw_ofld_send()
250 ep->emss = ep->com.dev->rdev.lldi.mtus[TCPOPT_MSS_G(opt)] - in set_emss()
401 ep->com.dev->rdev.lldi.ports[0], in _c4iw_free_ep()
405 cxgb4_remove_tid(ep->com.dev->rdev.lldi.tids, 0, ep->hwtid, in _c4iw_free_ep()
558 cxgb4_clip_release(ep->com.dev->rdev.lldi.ports[0], in act_open_req_arp_failure()
562 cxgb4_free_atid(ep->com.dev->rdev.lldi.tids, ep->atid); in act_open_req_arp_failure()
618 (ep->com.dev->rdev.lldi.pf)); in send_flowc()
736 enum chip_type adapter_type = ep->com.dev->rdev.lldi.adapter_type; in send_connect()
741 netdev = ep->com.dev->rdev.lldi.ports[0]; in send_connect()
[all …]
Dresource.c43 rdev->lldi.vr->qp.start, in c4iw_init_qid_table()
44 rdev->lldi.vr->qp.size, in c4iw_init_qid_table()
45 rdev->lldi.vr->qp.size, 0)) in c4iw_init_qid_table()
48 for (i = rdev->lldi.vr->qp.start; in c4iw_init_qid_table()
49 i < rdev->lldi.vr->qp.start + rdev->lldi.vr->qp.size; i++) in c4iw_init_qid_table()
308 pbl_start = rdev->lldi.vr->pbl.start; in c4iw_pblpool_create()
309 pbl_chunk = rdev->lldi.vr->pbl.size; in c4iw_pblpool_create()
350 pci_name(rdev->lldi.pdev)); in c4iw_rqtpool_alloc()
395 if (rdev->lldi.vr->srq.size) in c4iw_rqtpool_create()
398 rqt_start = rdev->lldi.vr->rq.start + skip; in c4iw_rqtpool_create()
[all …]
Dmem.c56 return (is_t4(dev->rdev.lldi.adapter_type) || in mr_exceeds_hw_limits()
57 is_t5(dev->rdev.lldi.adapter_type)) && in mr_exceeds_hw_limits()
92 T5_ULP_MEMIO_FID_V(rdev->lldi.rxq_ids[0])); in _c4iw_write_mem_dma_aligned()
120 if (is_t4(rdev->lldi.adapter_type)) in _c4iw_write_mem_inline()
200 daddr = dma_map_single(&rdev->lldi.pdev->dev, data, len, DMA_TO_DEVICE); in _c4iw_write_mem_dma()
201 if (dma_mapping_error(&rdev->lldi.pdev->dev, daddr)) in _c4iw_write_mem_dma()
226 dma_unmap_single(&rdev->lldi.pdev->dev, save, len, DMA_TO_DEVICE); in _c4iw_write_mem_dma()
240 if (!rdev->lldi.ulptx_memwrite_dsgl || !use_dsgl) { in write_adapter_mem()
255 pci_name(rdev->lldi.pdev)); in write_adapter_mem()
333 (rdev->lldi.vr->stag.start >> 5), in write_tpt_entry()
[all …]
Dqp.c74 dev_warn(&dev->rdev.lldi.pdev->dev, in alloc_ird()
102 dma_free_coherent(&(rdev->lldi.pdev->dev), sq->memsize, sq->queue, in dealloc_host_sq()
116 if (!ocqp_support || !ocqp_supported(&rdev->lldi)) in alloc_oc_sq()
122 rdev->lldi.vr->ocq.start; in alloc_oc_sq()
124 rdev->lldi.vr->ocq.start); in alloc_oc_sq()
131 sq->queue = dma_alloc_coherent(&(rdev->lldi.pdev->dev), sq->memsize, in alloc_host_sq()
162 dma_free_coherent(&rdev->lldi.pdev->dev, in destroy_qp()
184 ret = cxgb4_bar2_sge_qregs(rdev->lldi.ports[0], qid, qtype, in c4iw_bar2_addrs()
193 if (is_t4(rdev->lldi.adapter_type)) in c4iw_bar2_addrs()
264 wq->rq.queue = dma_alloc_coherent(&rdev->lldi.pdev->dev, in create_qp()
[all …]
Dev.c43 ret = cxgb4_read_tpte(dev->rdev.lldi.ports[0], stag, in print_tpte()
46 dev_err(&dev->rdev.lldi.pdev->dev, in print_tpte()
66 dev_err(&dev->rdev.lldi.pdev->dev, in dump_err_cqe()
Diw_cxgb4.h76 #define PBL_OFF(rdev_p, a) ((a) - (rdev_p)->lldi.vr->pbl.start)
77 #define RQT_OFF(rdev_p, a) ((a) - (rdev_p)->lldi.vr->rq.start)
180 struct cxgb4_lld_info lldi; member
205 return (int)(rdev->lldi.vr->stag.size >> 5); in c4iw_num_stags()
274 func, pci_name(rdev->lldi.pdev), hwtid, qpid); in c4iw_wait_for_reply()
281 pci_name(rdev->lldi.pdev), wr_waitp->ret, hwtid, qpid); in c4iw_wait_for_reply()
334 struct cxgb4_lld_info lldi; member
363 return min(dev->rdev.lldi.max_ordird_qp, c4iw_max_read_depth); in cur_max_read_depth()
Dcq.c64 dma_free_coherent(&(rdev->lldi.pdev->dev), in destroy_cq()
98 cq->queue = dma_alloc_coherent(&rdev->lldi.pdev->dev, cq->memsize, in create_cq()
142 rdev->lldi.ciq_ids[cq->vector])); in create_cq()
160 cq->gts = rdev->lldi.gts_reg; in create_cq()
168 pci_name(rdev->lldi.pdev), cq->cqid); in create_cq()
174 dma_free_coherent(&rdev->lldi.pdev->dev, cq->memsize, cq->queue, in create_cq()
1014 if (vector >= rhp->rdev.lldi.nciq) in c4iw_create_cq()
/drivers/scsi/cxgbi/cxgb4i/
Dcxgb4i.c210 struct cxgb4_lld_info *lldi = cxgbi_cdev_priv(csk->cdev); in send_act_open_req() local
230 if (is_t4(lldi->adapter_type)) { in send_act_open_req()
253 } else if (is_t5(lldi->adapter_type)) { in send_act_open_req()
319 CHELSIO_CHIP_VERSION(lldi->adapter_type), csk, in send_act_open_req()
329 struct cxgb4_lld_info *lldi = cxgbi_cdev_priv(csk->cdev); in send_act_open_req6() local
349 if (is_t4(lldi->adapter_type)) { in send_act_open_req6()
374 } else if (is_t5(lldi->adapter_type)) { in send_act_open_req6()
430 CHELSIO_CHIP_VERSION(lldi->adapter_type), csk, csk->state, in send_act_open_req6()
717 struct cxgb4_lld_info *lldi = cxgbi_cdev_priv(cdev); in cxgb4i_make_tx_data_wr() local
725 u32 force = is_t5(lldi->adapter_type) ? TX_FORCE_V(!submode) : in cxgb4i_make_tx_data_wr()
[all …]
/drivers/target/iscsi/cxgbit/
Dcxgbit_ddp.c298 struct cxgb4_lld_info *lldi = &cdev->lldi; in cxgbit_ddp_init() local
299 struct net_device *ndev = cdev->lldi.ports[0]; in cxgbit_ddp_init()
303 if (!lldi->vr->iscsi.size) { in cxgbit_ddp_init()
310 tformat.pgsz_order[i] = (lldi->iscsi_pgsz_order >> (i << 3)) in cxgbit_ddp_init()
312 cxgbi_tagmask_check(lldi->iscsi_tagmask, &tformat); in cxgbit_ddp_init()
314 ret = cxgbi_ppm_init(lldi->iscsi_ppm, cdev->lldi.ports[0], in cxgbit_ddp_init()
315 cdev->lldi.pdev, &cdev->lldi, &tformat, in cxgbit_ddp_init()
316 lldi->vr->iscsi.size, lldi->iscsi_llimit, in cxgbit_ddp_init()
317 lldi->vr->iscsi.start, 2, in cxgbit_ddp_init()
318 lldi->vr->ppod_edram.start, in cxgbit_ddp_init()
[all …]
Dcxgbit_cm.c63 func, pci_name(cdev->lldi.pdev), tid); in cxgbit_wait_for_reply()
69 pci_name(cdev->lldi.pdev), wr_waitp->ret, tid); in cxgbit_wait_for_reply()
153 __func__, cdev->lldi.ports[0]->name, stid, sin6->sin6_port); in cxgbit_create_server6()
158 ret = cxgb4_clip_get(cdev->lldi.ports[0], in cxgbit_create_server6()
170 ret = cxgb4_create_server6(cdev->lldi.ports[0], in cxgbit_create_server6()
173 cdev->lldi.rxq_ids[0]); in cxgbit_create_server6()
184 cxgb4_clip_release(cdev->lldi.ports[0], in cxgbit_create_server6()
204 __func__, cdev->lldi.ports[0]->name, stid, sin->sin_port); in cxgbit_create_server4()
209 ret = cxgb4_create_server(cdev->lldi.ports[0], in cxgbit_create_server4()
212 cdev->lldi.rxq_ids[0]); in cxgbit_create_server4()
[all …]
Dcxgbit_main.c33 struct cxgb4_lld_info *lldi = &cdev->lldi; in cxgbit_set_mdsl() local
38 mdsl = min_t(u32, lldi->iscsi_iolen - ISCSI_PDU_NONPAYLOAD_LEN, in cxgbit_set_mdsl()
46 static void *cxgbit_uld_add(const struct cxgb4_lld_info *lldi) in cxgbit_uld_add() argument
50 if (is_t4(lldi->adapter_type)) in cxgbit_uld_add()
60 cdev->lldi = *lldi; in cxgbit_uld_add()
71 pci_name(lldi->pdev)); in cxgbit_uld_add()
73 if (lldi->fw_vers >= 0x10d2b00) in cxgbit_uld_add()
84 pci_name(lldi->pdev)); in cxgbit_uld_add()
142 pr_info("cdev %s state UP.\n", pci_name(cdev->lldi.pdev)); in cxgbit_uld_state_change()
147 pr_info("cdev %s state RECOVERY.\n", pci_name(cdev->lldi.pdev)); in cxgbit_uld_state_change()
[all …]
Dcxgbit.h133 struct cxgb4_lld_info lldi; member
349 return (struct cxgbi_ppm *)(*cdev->lldi.iscsi_ppm); in cdev2ppm()
Dcxgbit_target.c163 const struct cxgb4_lld_info *lldi = &csk->com.cdev->lldi; in cxgbit_tx_data_wr() local
169 u32 force = is_t5(lldi->adapter_type) ? TX_FORCE_V(!submode) : in cxgbit_tx_data_wr()
719 struct cxgbi_ppm *ppm = *csk->com.cdev->lldi.iscsi_ppm; in cxgbit_set_params()
750 if (is_t5(cdev->lldi.adapter_type)) in cxgbit_set_params()
1562 struct cxgb4_lld_info *lldi = &csk->com.cdev->lldi; in cxgbit_rx_skb() local
1566 if (is_t5(lldi->adapter_type)) in cxgbit_rx_skb()
/drivers/net/ethernet/chelsio/inline_crypto/chtls/
Dchtls_main.c136 for (i = 0; i < cdev->lldi->nports; i++) { in chtls_inline_feature()
170 kfree(cdev->lldi); in chtls_free_uld()
197 strlcat(tlsdev->name, cdev->lldi->ports[0]->name, in chtls_register_dev()
237 struct cxgb4_lld_info *lldi; in chtls_uld_add() local
245 lldi = kzalloc(sizeof(*lldi), GFP_KERNEL); in chtls_uld_add()
246 if (!lldi) in chtls_uld_add()
252 *lldi = *info; in chtls_uld_add()
253 cdev->lldi = lldi; in chtls_uld_add()
254 cdev->pdev = lldi->pdev; in chtls_uld_add()
255 cdev->tids = lldi->tids; in chtls_uld_add()
[all …]
Dchtls_cm.c209 cxgb4_ofld_send(cdev->lldi->ports[0], skb); in abort_arp_failure()
691 cdev->lldi->rxq_ids[0]); in chtls_listen_start()
707 cdev->lldi->rxq_ids[0]); in chtls_listen_start()
742 cxgb4_remove_server(cdev->lldi->ports[0], stid, in chtls_listen_stop()
743 cdev->lldi->rxq_ids[0], sk->sk_family == PF_INET6); in chtls_listen_stop()
979 tp->advmss = cxgb4_best_aligned_mtu(cdev->lldi->mtus, in chtls_select_mss()
1011 struct cxgb4_lld_info *lldi; in chtls_pass_accept_rpl() local
1024 lldi = csk->cdev->lldi; in chtls_pass_accept_rpl()
1048 if (!is_t5(lldi->adapter_type)) in chtls_pass_accept_rpl()
1180 csk->l2t_entry = cxgb4_l2t_get(cdev->lldi->l2t, n, ndev, 0); in chtls_recv_sock()
[all …]
Dchtls_hw.c158 int chtls_init_kmap(struct chtls_dev *cdev, struct cxgb4_lld_info *lldi) in chtls_init_kmap() argument
163 num_key_ctx = (lldi->vr->key.size / TLS_KEY_CONTEXT_SZ); in chtls_init_kmap()
173 cdev->kmap.start = lldi->vr->key.start; in chtls_init_kmap()
Dchtls.h212 struct cxgb4_lld_info *lldi; member
559 int chtls_init_kmap(struct chtls_dev *cdev, struct cxgb4_lld_info *lldi);
/drivers/net/ethernet/chelsio/libcxgb/
Dlibcxgb_cm.c78 cxgb_our_interface(struct cxgb4_lld_info *lldi, in cxgb_our_interface() argument
85 for (i = 0; i < lldi->nports; i++) in cxgb_our_interface()
86 if (lldi->ports[i] == egress_dev) in cxgb_our_interface()
92 cxgb_find_route(struct cxgb4_lld_info *lldi, in cxgb_find_route() argument
109 if (!cxgb_our_interface(lldi, get_real_dev, n->dev) && in cxgb_find_route()
121 cxgb_find_route6(struct cxgb4_lld_info *lldi, in cxgb_find_route6() argument
138 (!cxgb_our_interface(lldi, get_real_dev, in cxgb_find_route6()
/drivers/crypto/chelsio/
Dchcr_algo.c774 qid = u_ctx->lldi.rxq_ids[rxqidx]; in create_wreq()
775 fid = u_ctx->lldi.rxq_ids[0]; in create_wreq()
778 rx_channel_id = cxgb4_port_e2cchan(u_ctx->lldi.ports[portno]); in create_wreq()
826 rx_channel_id = cxgb4_port_e2cchan(u_ctx->lldi.ports[rx_channel_id]); in create_cipher_wr()
1181 chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, in chcr_handle_cipher_resp()
1204 chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, in chcr_handle_cipher_resp()
1216 wrparam.qid = u_ctx->lldi.rxq_ids[reqctx->rxqidx]; in chcr_handle_cipher_resp()
1225 skb->dev = u_ctx->lldi.ports[0]; in chcr_handle_cipher_resp()
1237 chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, req); in chcr_handle_cipher_resp()
1281 err = chcr_cipher_dma_map(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, req); in process_cipher()
[all …]
Dchcr_core.h108 struct cxgb4_lld_info lldi; member
128 return pci_get_drvdata(u_ctx->lldi.pdev); in padap()
/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/
Dchcr_ktls.h77 struct cxgb4_lld_info lldi; member
99 return u_ctx->lldi.rxq_ids[0]; in chcr_get_first_rx_qid()
/drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/
Dchcr_ipsec.h23 struct cxgb4_lld_info lldi; member
Dchcr_ipsec.c114 u_ctx->lldi = *infop; in ch_ipsec_uld_add()
126 pr_info("%s: Up\n", pci_name(u_ctx->lldi.pdev)); in ch_ipsec_uld_state_change()
134 pr_info("%s: Down\n", pci_name(u_ctx->lldi.pdev)); in ch_ipsec_uld_state_change()
809 adap = pci_get_drvdata(u_ctx->lldi.pdev); in ch_ipsec_exit()

12