Lines Matching refs:ep
144 static void connect_reply_upcall(struct c4iw_ep *ep, int status);
163 static void deref_qp(struct c4iw_ep *ep) in deref_qp() argument
165 c4iw_qp_rem_ref(&ep->com.qp->ibqp); in deref_qp()
166 clear_bit(QP_REFERENCED, &ep->com.flags); in deref_qp()
167 set_bit(QP_DEREFED, &ep->com.history); in deref_qp()
170 static void ref_qp(struct c4iw_ep *ep) in ref_qp() argument
172 set_bit(QP_REFERENCED, &ep->com.flags); in ref_qp()
173 set_bit(QP_REFED, &ep->com.history); in ref_qp()
174 c4iw_qp_add_ref(&ep->com.qp->ibqp); in ref_qp()
177 static void start_ep_timer(struct c4iw_ep *ep) in start_ep_timer() argument
179 pr_debug("ep %p\n", ep); in start_ep_timer()
180 if (timer_pending(&ep->timer)) { in start_ep_timer()
182 __func__, ep); in start_ep_timer()
185 clear_bit(TIMEOUT, &ep->com.flags); in start_ep_timer()
186 c4iw_get_ep(&ep->com); in start_ep_timer()
187 ep->timer.expires = jiffies + ep_timeout_secs * HZ; in start_ep_timer()
188 add_timer(&ep->timer); in start_ep_timer()
191 static int stop_ep_timer(struct c4iw_ep *ep) in stop_ep_timer() argument
193 pr_debug("ep %p stopping\n", ep); in stop_ep_timer()
194 del_timer_sync(&ep->timer); in stop_ep_timer()
195 if (!test_and_set_bit(TIMEOUT, &ep->com.flags)) { in stop_ep_timer()
196 c4iw_put_ep(&ep->com); in stop_ep_timer()
248 static void set_emss(struct c4iw_ep *ep, u16 opt) in set_emss() argument
250 ep->emss = ep->com.dev->rdev.lldi.mtus[TCPOPT_MSS_G(opt)] - in set_emss()
251 ((AF_INET == ep->com.remote_addr.ss_family) ? in set_emss()
254 ep->mss = ep->emss; in set_emss()
256 ep->emss -= round_up(TCPOLEN_TIMESTAMP, 4); in set_emss()
257 if (ep->emss < 128) in set_emss()
258 ep->emss = 128; in set_emss()
259 if (ep->emss & 7) in set_emss()
261 TCPOPT_MSS_G(opt), ep->mss, ep->emss); in set_emss()
262 pr_debug("mss_idx %u mss %u emss=%u\n", TCPOPT_MSS_G(opt), ep->mss, in set_emss()
263 ep->emss); in set_emss()
330 static void remove_ep_tid(struct c4iw_ep *ep) in remove_ep_tid() argument
334 xa_lock_irqsave(&ep->com.dev->hwtids, flags); in remove_ep_tid()
335 __xa_erase(&ep->com.dev->hwtids, ep->hwtid); in remove_ep_tid()
336 if (xa_empty(&ep->com.dev->hwtids)) in remove_ep_tid()
337 wake_up(&ep->com.dev->wait); in remove_ep_tid()
338 xa_unlock_irqrestore(&ep->com.dev->hwtids, flags); in remove_ep_tid()
341 static int insert_ep_tid(struct c4iw_ep *ep) in insert_ep_tid() argument
346 xa_lock_irqsave(&ep->com.dev->hwtids, flags); in insert_ep_tid()
347 err = __xa_insert(&ep->com.dev->hwtids, ep->hwtid, ep, GFP_KERNEL); in insert_ep_tid()
348 xa_unlock_irqrestore(&ep->com.dev->hwtids, flags); in insert_ep_tid()
358 struct c4iw_ep *ep; in get_ep_from_tid() local
362 ep = xa_load(&dev->hwtids, tid); in get_ep_from_tid()
363 if (ep) in get_ep_from_tid()
364 c4iw_get_ep(&ep->com); in get_ep_from_tid()
366 return ep; in get_ep_from_tid()
375 struct c4iw_listen_ep *ep; in get_ep_from_stid() local
379 ep = xa_load(&dev->stids, stid); in get_ep_from_stid()
380 if (ep) in get_ep_from_stid()
381 c4iw_get_ep(&ep->com); in get_ep_from_stid()
383 return ep; in get_ep_from_stid()
388 struct c4iw_ep *ep; in _c4iw_free_ep() local
390 ep = container_of(kref, struct c4iw_ep, com.kref); in _c4iw_free_ep()
391 pr_debug("ep %p state %s\n", ep, states[ep->com.state]); in _c4iw_free_ep()
392 if (test_bit(QP_REFERENCED, &ep->com.flags)) in _c4iw_free_ep()
393 deref_qp(ep); in _c4iw_free_ep()
394 if (test_bit(RELEASE_RESOURCES, &ep->com.flags)) { in _c4iw_free_ep()
395 if (ep->com.remote_addr.ss_family == AF_INET6) { in _c4iw_free_ep()
398 &ep->com.local_addr; in _c4iw_free_ep()
401 ep->com.dev->rdev.lldi.ports[0], in _c4iw_free_ep()
405 cxgb4_remove_tid(ep->com.dev->rdev.lldi.tids, 0, ep->hwtid, in _c4iw_free_ep()
406 ep->com.local_addr.ss_family); in _c4iw_free_ep()
407 dst_release(ep->dst); in _c4iw_free_ep()
408 cxgb4_l2t_release(ep->l2t); in _c4iw_free_ep()
409 kfree_skb(ep->mpa_skb); in _c4iw_free_ep()
411 if (!skb_queue_empty(&ep->com.ep_skb_list)) in _c4iw_free_ep()
412 skb_queue_purge(&ep->com.ep_skb_list); in _c4iw_free_ep()
413 c4iw_put_wr_wait(ep->com.wr_waitp); in _c4iw_free_ep()
414 kfree(ep); in _c4iw_free_ep()
417 static void release_ep_resources(struct c4iw_ep *ep) in release_ep_resources() argument
419 set_bit(RELEASE_RESOURCES, &ep->com.flags); in release_ep_resources()
427 if (ep->hwtid != -1) in release_ep_resources()
428 remove_ep_tid(ep); in release_ep_resources()
429 c4iw_put_ep(&ep->com); in release_ep_resources()
494 struct c4iw_ep *ep; in _put_ep_safe() local
496 ep = *((struct c4iw_ep **)(skb->cb + 2 * sizeof(void *))); in _put_ep_safe()
497 release_ep_resources(ep); in _put_ep_safe()
503 struct c4iw_ep *ep; in _put_pass_ep_safe() local
505 ep = *((struct c4iw_ep **)(skb->cb + 2 * sizeof(void *))); in _put_pass_ep_safe()
506 c4iw_put_ep(&ep->parent_ep->com); in _put_pass_ep_safe()
507 release_ep_resources(ep); in _put_pass_ep_safe()
517 static void queue_arp_failure_cpl(struct c4iw_ep *ep, struct sk_buff *skb, in queue_arp_failure_cpl() argument
529 *((struct c4iw_ep **)(skb->cb + 2 * sizeof(void *))) = ep; in queue_arp_failure_cpl()
530 sched(ep->com.dev, skb); in queue_arp_failure_cpl()
536 struct c4iw_ep *ep = handle; in pass_accept_rpl_arp_failure() local
539 ep->hwtid); in pass_accept_rpl_arp_failure()
541 __state_set(&ep->com, DEAD); in pass_accept_rpl_arp_failure()
542 queue_arp_failure_cpl(ep, skb, FAKE_CPL_PASS_PUT_EP_SAFE); in pass_accept_rpl_arp_failure()
550 struct c4iw_ep *ep = handle; in act_open_req_arp_failure() local
553 connect_reply_upcall(ep, -EHOSTUNREACH); in act_open_req_arp_failure()
554 __state_set(&ep->com, DEAD); in act_open_req_arp_failure()
555 if (ep->com.remote_addr.ss_family == AF_INET6) { in act_open_req_arp_failure()
557 (struct sockaddr_in6 *)&ep->com.local_addr; in act_open_req_arp_failure()
558 cxgb4_clip_release(ep->com.dev->rdev.lldi.ports[0], in act_open_req_arp_failure()
561 xa_erase_irq(&ep->com.dev->atids, ep->atid); in act_open_req_arp_failure()
562 cxgb4_free_atid(ep->com.dev->rdev.lldi.tids, ep->atid); in act_open_req_arp_failure()
563 queue_arp_failure_cpl(ep, skb, FAKE_CPL_PUT_EP_SAFE); in act_open_req_arp_failure()
573 struct c4iw_ep *ep = handle; in abort_arp_failure() local
574 struct c4iw_rdev *rdev = &ep->com.dev->rdev; in abort_arp_failure()
582 __state_set(&ep->com, DEAD); in abort_arp_failure()
583 queue_arp_failure_cpl(ep, skb, FAKE_CPL_PUT_EP_SAFE); in abort_arp_failure()
588 static int send_flowc(struct c4iw_ep *ep) in send_flowc() argument
591 struct sk_buff *skb = skb_dequeue(&ep->com.ep_skb_list); in send_flowc()
592 u16 vlan = ep->l2t->vlan; in send_flowc()
614 FW_WR_FLOWID_V(ep->hwtid)); in send_flowc()
618 (ep->com.dev->rdev.lldi.pf)); in send_flowc()
620 flowc->mnemval[1].val = cpu_to_be32(ep->tx_chan); in send_flowc()
622 flowc->mnemval[2].val = cpu_to_be32(ep->tx_chan); in send_flowc()
624 flowc->mnemval[3].val = cpu_to_be32(ep->rss_qid); in send_flowc()
626 flowc->mnemval[4].val = cpu_to_be32(ep->snd_seq); in send_flowc()
628 flowc->mnemval[5].val = cpu_to_be32(ep->rcv_seq); in send_flowc()
630 flowc->mnemval[6].val = cpu_to_be32(ep->snd_win); in send_flowc()
632 flowc->mnemval[7].val = cpu_to_be32(ep->emss); in send_flowc()
634 flowc->mnemval[8].val = cpu_to_be32(ep->snd_wscale); in send_flowc()
642 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in send_flowc()
643 return c4iw_ofld_send(&ep->com.dev->rdev, skb); in send_flowc()
646 static int send_halfclose(struct c4iw_ep *ep) in send_halfclose() argument
648 struct sk_buff *skb = skb_dequeue(&ep->com.ep_skb_list); in send_halfclose()
651 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in send_halfclose()
655 cxgb_mk_close_con_req(skb, wrlen, ep->hwtid, ep->txq_idx, in send_halfclose()
658 return c4iw_l2t_send(&ep->com.dev->rdev, skb, ep->l2t); in send_halfclose()
661 static void read_tcb(struct c4iw_ep *ep) in read_tcb() argument
671 set_wr_txq(skb, CPL_PRIORITY_CONTROL, ep->ctrlq_idx); in read_tcb()
674 INIT_TP_WR(req, ep->hwtid); in read_tcb()
675 OPCODE_TID(req) = cpu_to_be32(MK_OPCODE_TID(CPL_GET_TCB, ep->hwtid)); in read_tcb()
676 req->reply_ctrl = htons(REPLY_CHAN_V(0) | QUEUENO_V(ep->rss_qid)); in read_tcb()
682 c4iw_get_ep(&ep->com); in read_tcb()
683 if (WARN_ON(c4iw_ofld_send(&ep->com.dev->rdev, skb))) in read_tcb()
684 c4iw_put_ep(&ep->com); in read_tcb()
687 static int send_abort_req(struct c4iw_ep *ep) in send_abort_req() argument
690 struct sk_buff *req_skb = skb_dequeue(&ep->com.ep_skb_list); in send_abort_req()
692 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in send_abort_req()
696 cxgb_mk_abort_req(req_skb, wrlen, ep->hwtid, ep->txq_idx, in send_abort_req()
697 ep, abort_arp_failure); in send_abort_req()
699 return c4iw_l2t_send(&ep->com.dev->rdev, req_skb, ep->l2t); in send_abort_req()
702 static int send_abort(struct c4iw_ep *ep) in send_abort() argument
704 if (!ep->com.qp || !ep->com.qp->srq) { in send_abort()
705 send_abort_req(ep); in send_abort()
708 set_bit(ABORT_REQ_IN_PROGRESS, &ep->com.flags); in send_abort()
709 read_tcb(ep); in send_abort()
713 static int send_connect(struct c4iw_ep *ep) in send_connect() argument
728 &ep->com.local_addr; in send_connect()
730 &ep->com.remote_addr; in send_connect()
732 &ep->com.local_addr; in send_connect()
734 &ep->com.remote_addr; in send_connect()
736 enum chip_type adapter_type = ep->com.dev->rdev.lldi.adapter_type; in send_connect()
741 netdev = ep->com.dev->rdev.lldi.ports[0]; in send_connect()
762 wrlen = (ep->com.remote_addr.ss_family == AF_INET) ? in send_connect()
766 pr_debug("ep %p atid %u\n", ep, ep->atid); in send_connect()
773 set_wr_txq(skb, CPL_PRIORITY_SETUP, ep->ctrlq_idx); in send_connect()
775 cxgb_best_mtu(ep->com.dev->rdev.lldi.mtus, ep->mtu, &mtu_idx, in send_connect()
777 (ep->com.remote_addr.ss_family == AF_INET) ? 0 : 1); in send_connect()
784 win = ep->rcv_win >> 10; in send_connect()
793 L2T_IDX_V(ep->l2t->idx) | in send_connect()
794 TX_CHAN_V(ep->tx_chan) | in send_connect()
795 SMAC_SEL_V(ep->smac_idx) | in send_connect()
796 DSCP_V(ep->tos >> 2) | in send_connect()
801 RSS_QUEUE_VALID_F | RSS_QUEUE_V(ep->rss_qid); in send_connect()
817 params = cxgb4_select_ntuple(netdev, ep->l2t); in send_connect()
819 if (ep->com.remote_addr.ss_family == AF_INET6) in send_connect()
820 cxgb4_clip_get(ep->com.dev->rdev.lldi.ports[0], in send_connect()
823 t4_set_arp_err_handler(skb, ep, act_open_req_arp_failure); in send_connect()
825 if (ep->com.remote_addr.ss_family == AF_INET) { in send_connect()
850 ((ep->rss_qid<<14) | ep->atid))); in send_connect()
857 if (is_t4(ep->com.dev->rdev.lldi.adapter_type)) { in send_connect()
861 if (is_t5(ep->com.dev->rdev.lldi.adapter_type)) { in send_connect()
900 ((ep->rss_qid<<14)|ep->atid))); in send_connect()
909 if (is_t4(ep->com.dev->rdev.lldi.adapter_type)) { in send_connect()
911 ep->l2t)); in send_connect()
914 if (is_t5(ep->com.dev->rdev.lldi.adapter_type)) { in send_connect()
931 set_bit(ACT_OPEN_REQ, &ep->com.history); in send_connect()
932 ret = c4iw_l2t_send(&ep->com.dev->rdev, skb, ep->l2t); in send_connect()
934 if (ret && ep->com.remote_addr.ss_family == AF_INET6) in send_connect()
935 cxgb4_clip_release(ep->com.dev->rdev.lldi.ports[0], in send_connect()
940 static int send_mpa_req(struct c4iw_ep *ep, struct sk_buff *skb, in send_mpa_req() argument
949 ep, ep->hwtid, ep->plen); in send_mpa_req()
951 mpalen = sizeof(*mpa) + ep->plen; in send_mpa_req()
957 connect_reply_upcall(ep, -ENOMEM); in send_mpa_req()
960 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in send_mpa_req()
968 FW_WR_FLOWID_V(ep->hwtid) | in send_mpa_req()
983 ep->mpa_attr.recv_marker_enabled = 1; in send_mpa_req()
985 ep->mpa_attr.recv_marker_enabled = 0; in send_mpa_req()
990 mpa->private_data_size = htons(ep->plen); in send_mpa_req()
993 ep->tried_with_mpa_v1 = 1; in send_mpa_req()
994 ep->retry_with_mpa_v1 = 0; in send_mpa_req()
1001 pr_debug("initiator ird %u ord %u\n", ep->ird, in send_mpa_req()
1002 ep->ord); in send_mpa_req()
1003 mpa_v2_params.ird = htons((u16)ep->ird); in send_mpa_req()
1004 mpa_v2_params.ord = htons((u16)ep->ord); in send_mpa_req()
1018 if (ep->plen) in send_mpa_req()
1021 ep->mpa_pkt + sizeof(*mpa), ep->plen); in send_mpa_req()
1023 if (ep->plen) in send_mpa_req()
1025 ep->mpa_pkt + sizeof(*mpa), ep->plen); in send_mpa_req()
1034 ep->mpa_skb = skb; in send_mpa_req()
1035 ret = c4iw_l2t_send(&ep->com.dev->rdev, skb, ep->l2t); in send_mpa_req()
1038 start_ep_timer(ep); in send_mpa_req()
1039 __state_set(&ep->com, MPA_REQ_SENT); in send_mpa_req()
1040 ep->mpa_attr.initiator = 1; in send_mpa_req()
1041 ep->snd_seq += mpalen; in send_mpa_req()
1045 static int send_mpa_reject(struct c4iw_ep *ep, const void *pdata, u8 plen) in send_mpa_reject() argument
1054 ep, ep->hwtid, ep->plen); in send_mpa_reject()
1057 if (ep->mpa_attr.version == 2 && ep->mpa_attr.enhanced_rdma_conn) in send_mpa_reject()
1066 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in send_mpa_reject()
1074 FW_WR_FLOWID_V(ep->hwtid) | in send_mpa_reject()
1085 mpa->revision = ep->mpa_attr.version; in send_mpa_reject()
1088 if (ep->mpa_attr.version == 2 && ep->mpa_attr.enhanced_rdma_conn) { in send_mpa_reject()
1093 mpa_v2_params.ird = htons(((u16)ep->ird) | in send_mpa_reject()
1096 mpa_v2_params.ord = htons(((u16)ep->ord) | (peer2peer ? in send_mpa_reject()
1105 if (ep->plen) in send_mpa_reject()
1118 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in send_mpa_reject()
1120 ep->mpa_skb = skb; in send_mpa_reject()
1121 ep->snd_seq += mpalen; in send_mpa_reject()
1122 return c4iw_l2t_send(&ep->com.dev->rdev, skb, ep->l2t); in send_mpa_reject()
1125 static int send_mpa_reply(struct c4iw_ep *ep, const void *pdata, u8 plen) in send_mpa_reply() argument
1134 ep, ep->hwtid, ep->plen); in send_mpa_reply()
1137 if (ep->mpa_attr.version == 2 && ep->mpa_attr.enhanced_rdma_conn) in send_mpa_reply()
1146 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in send_mpa_reply()
1154 FW_WR_FLOWID_V(ep->hwtid) | in send_mpa_reply()
1165 if (ep->mpa_attr.crc_enabled) in send_mpa_reply()
1167 if (ep->mpa_attr.recv_marker_enabled) in send_mpa_reply()
1169 mpa->revision = ep->mpa_attr.version; in send_mpa_reply()
1172 if (ep->mpa_attr.version == 2 && ep->mpa_attr.enhanced_rdma_conn) { in send_mpa_reply()
1177 mpa_v2_params.ird = htons((u16)ep->ird); in send_mpa_reply()
1178 mpa_v2_params.ord = htons((u16)ep->ord); in send_mpa_reply()
1179 if (peer2peer && (ep->mpa_attr.p2p_type != in send_mpa_reply()
1194 if (ep->plen) in send_mpa_reply()
1208 ep->mpa_skb = skb; in send_mpa_reply()
1209 __state_set(&ep->com, MPA_REP_SENT); in send_mpa_reply()
1210 ep->snd_seq += mpalen; in send_mpa_reply()
1211 return c4iw_l2t_send(&ep->com.dev->rdev, skb, ep->l2t); in send_mpa_reply()
1216 struct c4iw_ep *ep; in act_establish() local
1224 ep = lookup_atid(t, atid); in act_establish()
1226 pr_debug("ep %p tid %u snd_isn %u rcv_isn %u\n", ep, tid, in act_establish()
1229 mutex_lock(&ep->com.mutex); in act_establish()
1230 dst_confirm(ep->dst); in act_establish()
1233 ep->hwtid = tid; in act_establish()
1234 cxgb4_insert_tid(t, ep, tid, ep->com.local_addr.ss_family); in act_establish()
1235 insert_ep_tid(ep); in act_establish()
1237 ep->snd_seq = be32_to_cpu(req->snd_isn); in act_establish()
1238 ep->rcv_seq = be32_to_cpu(req->rcv_isn); in act_establish()
1239 ep->snd_wscale = TCPOPT_SND_WSCALE_G(tcp_opt); in act_establish()
1241 set_emss(ep, tcp_opt); in act_establish()
1244 xa_erase_irq(&ep->com.dev->atids, atid); in act_establish()
1246 set_bit(ACT_ESTAB, &ep->com.history); in act_establish()
1249 ret = send_flowc(ep); in act_establish()
1252 if (ep->retry_with_mpa_v1) in act_establish()
1253 ret = send_mpa_req(ep, skb, 1); in act_establish()
1255 ret = send_mpa_req(ep, skb, mpa_rev); in act_establish()
1258 mutex_unlock(&ep->com.mutex); in act_establish()
1261 mutex_unlock(&ep->com.mutex); in act_establish()
1262 connect_reply_upcall(ep, -ENOMEM); in act_establish()
1263 c4iw_ep_disconnect(ep, 0, GFP_KERNEL); in act_establish()
1267 static void close_complete_upcall(struct c4iw_ep *ep, int status) in close_complete_upcall() argument
1271 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in close_complete_upcall()
1275 if (ep->com.cm_id) { in close_complete_upcall()
1277 ep, ep->com.cm_id, ep->hwtid); in close_complete_upcall()
1278 ep->com.cm_id->event_handler(ep->com.cm_id, &event); in close_complete_upcall()
1279 deref_cm_id(&ep->com); in close_complete_upcall()
1280 set_bit(CLOSE_UPCALL, &ep->com.history); in close_complete_upcall()
1284 static void peer_close_upcall(struct c4iw_ep *ep) in peer_close_upcall() argument
1288 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in peer_close_upcall()
1291 if (ep->com.cm_id) { in peer_close_upcall()
1293 ep, ep->com.cm_id, ep->hwtid); in peer_close_upcall()
1294 ep->com.cm_id->event_handler(ep->com.cm_id, &event); in peer_close_upcall()
1295 set_bit(DISCONN_UPCALL, &ep->com.history); in peer_close_upcall()
1299 static void peer_abort_upcall(struct c4iw_ep *ep) in peer_abort_upcall() argument
1303 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in peer_abort_upcall()
1307 if (ep->com.cm_id) { in peer_abort_upcall()
1308 pr_debug("abort delivered ep %p cm_id %p tid %u\n", ep, in peer_abort_upcall()
1309 ep->com.cm_id, ep->hwtid); in peer_abort_upcall()
1310 ep->com.cm_id->event_handler(ep->com.cm_id, &event); in peer_abort_upcall()
1311 deref_cm_id(&ep->com); in peer_abort_upcall()
1312 set_bit(ABORT_UPCALL, &ep->com.history); in peer_abort_upcall()
1316 static void connect_reply_upcall(struct c4iw_ep *ep, int status) in connect_reply_upcall() argument
1321 ep, ep->hwtid, status); in connect_reply_upcall()
1325 memcpy(&event.local_addr, &ep->com.local_addr, in connect_reply_upcall()
1326 sizeof(ep->com.local_addr)); in connect_reply_upcall()
1327 memcpy(&event.remote_addr, &ep->com.remote_addr, in connect_reply_upcall()
1328 sizeof(ep->com.remote_addr)); in connect_reply_upcall()
1331 if (!ep->tried_with_mpa_v1) { in connect_reply_upcall()
1333 event.ord = ep->ird; in connect_reply_upcall()
1334 event.ird = ep->ord; in connect_reply_upcall()
1335 event.private_data_len = ep->plen - in connect_reply_upcall()
1337 event.private_data = ep->mpa_pkt + in connect_reply_upcall()
1342 event.ord = cur_max_read_depth(ep->com.dev); in connect_reply_upcall()
1343 event.ird = cur_max_read_depth(ep->com.dev); in connect_reply_upcall()
1344 event.private_data_len = ep->plen; in connect_reply_upcall()
1345 event.private_data = ep->mpa_pkt + in connect_reply_upcall()
1350 pr_debug("ep %p tid %u status %d\n", ep, in connect_reply_upcall()
1351 ep->hwtid, status); in connect_reply_upcall()
1352 set_bit(CONN_RPL_UPCALL, &ep->com.history); in connect_reply_upcall()
1353 ep->com.cm_id->event_handler(ep->com.cm_id, &event); in connect_reply_upcall()
1356 deref_cm_id(&ep->com); in connect_reply_upcall()
1359 static int connect_request_upcall(struct c4iw_ep *ep) in connect_request_upcall() argument
1364 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in connect_request_upcall()
1367 memcpy(&event.local_addr, &ep->com.local_addr, in connect_request_upcall()
1368 sizeof(ep->com.local_addr)); in connect_request_upcall()
1369 memcpy(&event.remote_addr, &ep->com.remote_addr, in connect_request_upcall()
1370 sizeof(ep->com.remote_addr)); in connect_request_upcall()
1371 event.provider_data = ep; in connect_request_upcall()
1372 if (!ep->tried_with_mpa_v1) { in connect_request_upcall()
1374 event.ord = ep->ord; in connect_request_upcall()
1375 event.ird = ep->ird; in connect_request_upcall()
1376 event.private_data_len = ep->plen - in connect_request_upcall()
1378 event.private_data = ep->mpa_pkt + sizeof(struct mpa_message) + in connect_request_upcall()
1382 event.ord = cur_max_read_depth(ep->com.dev); in connect_request_upcall()
1383 event.ird = cur_max_read_depth(ep->com.dev); in connect_request_upcall()
1384 event.private_data_len = ep->plen; in connect_request_upcall()
1385 event.private_data = ep->mpa_pkt + sizeof(struct mpa_message); in connect_request_upcall()
1387 c4iw_get_ep(&ep->com); in connect_request_upcall()
1388 ret = ep->parent_ep->com.cm_id->event_handler(ep->parent_ep->com.cm_id, in connect_request_upcall()
1391 c4iw_put_ep(&ep->com); in connect_request_upcall()
1392 set_bit(CONNREQ_UPCALL, &ep->com.history); in connect_request_upcall()
1393 c4iw_put_ep(&ep->parent_ep->com); in connect_request_upcall()
1397 static void established_upcall(struct c4iw_ep *ep) in established_upcall() argument
1401 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in established_upcall()
1404 event.ird = ep->ord; in established_upcall()
1405 event.ord = ep->ird; in established_upcall()
1406 if (ep->com.cm_id) { in established_upcall()
1407 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in established_upcall()
1408 ep->com.cm_id->event_handler(ep->com.cm_id, &event); in established_upcall()
1409 set_bit(ESTAB_UPCALL, &ep->com.history); in established_upcall()
1413 static int update_rx_credits(struct c4iw_ep *ep, u32 credits) in update_rx_credits() argument
1420 ep, ep->hwtid, credits); in update_rx_credits()
1432 if (ep->rcv_win > RCV_BUFSIZ_M * 1024) in update_rx_credits()
1433 credits += ep->rcv_win - RCV_BUFSIZ_M * 1024; in update_rx_credits()
1438 cxgb_mk_rx_data_ack(skb, wrlen, ep->hwtid, ep->ctrlq_idx, in update_rx_credits()
1441 c4iw_ofld_send(&ep->com.dev->rdev, skb); in update_rx_credits()
1459 static int process_mpa_reply(struct c4iw_ep *ep, struct sk_buff *skb) in process_mpa_reply() argument
1471 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in process_mpa_reply()
1477 if (ep->mpa_pkt_len + skb->len > sizeof(ep->mpa_pkt)) { in process_mpa_reply()
1485 skb_copy_from_linear_data(skb, &(ep->mpa_pkt[ep->mpa_pkt_len]), in process_mpa_reply()
1487 ep->mpa_pkt_len += skb->len; in process_mpa_reply()
1492 if (ep->mpa_pkt_len < sizeof(*mpa)) in process_mpa_reply()
1494 mpa = (struct mpa_message *) ep->mpa_pkt; in process_mpa_reply()
1521 if (ep->mpa_pkt_len > (sizeof(*mpa) + plen)) { in process_mpa_reply()
1526 ep->plen = (u8) plen; in process_mpa_reply()
1532 if (ep->mpa_pkt_len < (sizeof(*mpa) + plen)) in process_mpa_reply()
1545 if (stop_ep_timer(ep)) in process_mpa_reply()
1553 __state_set(&ep->com, FPDU_MODE); in process_mpa_reply()
1554 ep->mpa_attr.crc_enabled = (mpa->flags & MPA_CRC) | crc_enabled ? 1 : 0; in process_mpa_reply()
1555 ep->mpa_attr.xmit_marker_enabled = mpa->flags & MPA_MARKERS ? 1 : 0; in process_mpa_reply()
1556 ep->mpa_attr.version = mpa->revision; in process_mpa_reply()
1557 ep->mpa_attr.p2p_type = FW_RI_INIT_P2PTYPE_DISABLED; in process_mpa_reply()
1560 ep->mpa_attr.enhanced_rdma_conn = in process_mpa_reply()
1562 if (ep->mpa_attr.enhanced_rdma_conn) { in process_mpa_reply()
1564 (ep->mpa_pkt + sizeof(*mpa)); in process_mpa_reply()
1570 resp_ird, resp_ord, ep->ird, ep->ord); in process_mpa_reply()
1577 if (ep->ird < resp_ord) { in process_mpa_reply()
1579 ep->com.dev->rdev.lldi.max_ordird_qp) in process_mpa_reply()
1580 ep->ird = resp_ord; in process_mpa_reply()
1583 } else if (ep->ird > resp_ord) { in process_mpa_reply()
1584 ep->ird = resp_ord; in process_mpa_reply()
1586 if (ep->ord > resp_ird) { in process_mpa_reply()
1588 ep->ord = resp_ird; in process_mpa_reply()
1594 ep->ird = resp_ord; in process_mpa_reply()
1595 ep->ord = resp_ird; in process_mpa_reply()
1602 ep->mpa_attr.p2p_type = in process_mpa_reply()
1606 ep->mpa_attr.p2p_type = in process_mpa_reply()
1612 ep->mpa_attr.p2p_type = p2p_type; in process_mpa_reply()
1615 ep->mpa_attr.crc_enabled, in process_mpa_reply()
1616 ep->mpa_attr.recv_marker_enabled, in process_mpa_reply()
1617 ep->mpa_attr.xmit_marker_enabled, ep->mpa_attr.version, in process_mpa_reply()
1618 ep->mpa_attr.p2p_type, p2p_type); in process_mpa_reply()
1626 if ((ep->mpa_attr.version == 2) && peer2peer && in process_mpa_reply()
1627 (ep->mpa_attr.p2p_type != p2p_type)) { in process_mpa_reply()
1628 ep->mpa_attr.p2p_type = FW_RI_INIT_P2PTYPE_DISABLED; in process_mpa_reply()
1632 attrs.mpa_attr = ep->mpa_attr; in process_mpa_reply()
1633 attrs.max_ird = ep->ird; in process_mpa_reply()
1634 attrs.max_ord = ep->ord; in process_mpa_reply()
1635 attrs.llp_stream_handle = ep; in process_mpa_reply()
1643 err = c4iw_modify_qp(ep->com.qp->rhp, in process_mpa_reply()
1644 ep->com.qp, mask, &attrs, 1); in process_mpa_reply()
1658 err = c4iw_modify_qp(ep->com.qp->rhp, ep->com.qp, in process_mpa_reply()
1677 err = c4iw_modify_qp(ep->com.qp->rhp, ep->com.qp, in process_mpa_reply()
1685 stop_ep_timer(ep); in process_mpa_reply()
1689 connect_reply_upcall(ep, err); in process_mpa_reply()
1705 static int process_mpa_request(struct c4iw_ep *ep, struct sk_buff *skb) in process_mpa_request() argument
1711 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in process_mpa_request()
1717 if (ep->mpa_pkt_len + skb->len > sizeof(ep->mpa_pkt)) in process_mpa_request()
1725 skb_copy_from_linear_data(skb, &(ep->mpa_pkt[ep->mpa_pkt_len]), in process_mpa_request()
1727 ep->mpa_pkt_len += skb->len; in process_mpa_request()
1733 if (ep->mpa_pkt_len < sizeof(*mpa)) in process_mpa_request()
1737 mpa = (struct mpa_message *) ep->mpa_pkt; in process_mpa_request()
1762 if (ep->mpa_pkt_len > (sizeof(*mpa) + plen)) in process_mpa_request()
1764 ep->plen = (u8) plen; in process_mpa_request()
1769 if (ep->mpa_pkt_len < (sizeof(*mpa) + plen)) in process_mpa_request()
1776 ep->mpa_attr.initiator = 0; in process_mpa_request()
1777 ep->mpa_attr.crc_enabled = (mpa->flags & MPA_CRC) | crc_enabled ? 1 : 0; in process_mpa_request()
1778 ep->mpa_attr.recv_marker_enabled = markers_enabled; in process_mpa_request()
1779 ep->mpa_attr.xmit_marker_enabled = mpa->flags & MPA_MARKERS ? 1 : 0; in process_mpa_request()
1780 ep->mpa_attr.version = mpa->revision; in process_mpa_request()
1782 ep->tried_with_mpa_v1 = 1; in process_mpa_request()
1783 ep->mpa_attr.p2p_type = FW_RI_INIT_P2PTYPE_DISABLED; in process_mpa_request()
1786 ep->mpa_attr.enhanced_rdma_conn = in process_mpa_request()
1788 if (ep->mpa_attr.enhanced_rdma_conn) { in process_mpa_request()
1790 (ep->mpa_pkt + sizeof(*mpa)); in process_mpa_request()
1791 ep->ird = ntohs(mpa_v2_params->ird) & in process_mpa_request()
1793 ep->ird = min_t(u32, ep->ird, in process_mpa_request()
1794 cur_max_read_depth(ep->com.dev)); in process_mpa_request()
1795 ep->ord = ntohs(mpa_v2_params->ord) & in process_mpa_request()
1797 ep->ord = min_t(u32, ep->ord, in process_mpa_request()
1798 cur_max_read_depth(ep->com.dev)); in process_mpa_request()
1800 ep->ird, ep->ord); in process_mpa_request()
1805 ep->mpa_attr.p2p_type = in process_mpa_request()
1809 ep->mpa_attr.p2p_type = in process_mpa_request()
1815 ep->mpa_attr.p2p_type = p2p_type; in process_mpa_request()
1818 ep->mpa_attr.crc_enabled, ep->mpa_attr.recv_marker_enabled, in process_mpa_request()
1819 ep->mpa_attr.xmit_marker_enabled, ep->mpa_attr.version, in process_mpa_request()
1820 ep->mpa_attr.p2p_type); in process_mpa_request()
1822 __state_set(&ep->com, MPA_REQ_RCVD); in process_mpa_request()
1825 mutex_lock_nested(&ep->parent_ep->com.mutex, SINGLE_DEPTH_NESTING); in process_mpa_request()
1826 if (ep->parent_ep->com.state != DEAD) { in process_mpa_request()
1827 if (connect_request_upcall(ep)) in process_mpa_request()
1832 mutex_unlock(&ep->parent_ep->com.mutex); in process_mpa_request()
1836 mutex_unlock(&ep->parent_ep->com.mutex); in process_mpa_request()
1839 (void)stop_ep_timer(ep); in process_mpa_request()
1846 struct c4iw_ep *ep; in rx_data() local
1853 ep = get_ep_from_tid(dev, tid); in rx_data()
1854 if (!ep) in rx_data()
1856 pr_debug("ep %p tid %u dlen %u\n", ep, ep->hwtid, dlen); in rx_data()
1859 mutex_lock(&ep->com.mutex); in rx_data()
1861 switch (ep->com.state) { in rx_data()
1863 update_rx_credits(ep, dlen); in rx_data()
1864 ep->rcv_seq += dlen; in rx_data()
1865 disconnect = process_mpa_reply(ep, skb); in rx_data()
1868 update_rx_credits(ep, dlen); in rx_data()
1869 ep->rcv_seq += dlen; in rx_data()
1870 disconnect = process_mpa_request(ep, skb); in rx_data()
1875 update_rx_credits(ep, dlen); in rx_data()
1879 __func__, ep->com.qp->wq.sq.qid, ep, in rx_data()
1880 ep->com.state, ep->hwtid, status); in rx_data()
1882 c4iw_modify_qp(ep->com.qp->rhp, ep->com.qp, in rx_data()
1890 mutex_unlock(&ep->com.mutex); in rx_data()
1892 c4iw_ep_disconnect(ep, disconnect == 2, GFP_KERNEL); in rx_data()
1893 c4iw_put_ep(&ep->com); in rx_data()
1897 static void complete_cached_srq_buffers(struct c4iw_ep *ep, u32 srqidx) in complete_cached_srq_buffers() argument
1901 adapter_type = ep->com.dev->rdev.lldi.adapter_type; in complete_cached_srq_buffers()
1910 if (ep->com.qp->ibqp.uobject) in complete_cached_srq_buffers()
1911 t4_set_wq_in_error(&ep->com.qp->wq, srqidx); in complete_cached_srq_buffers()
1913 c4iw_flush_srqidx(ep->com.qp, srqidx); in complete_cached_srq_buffers()
1920 struct c4iw_ep *ep; in abort_rpl() local
1925 ep = get_ep_from_tid(dev, tid); in abort_rpl()
1926 if (!ep) { in abort_rpl()
1931 if (ep->com.qp && ep->com.qp->srq) { in abort_rpl()
1933 complete_cached_srq_buffers(ep, srqidx ? srqidx : ep->srqe_idx); in abort_rpl()
1936 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in abort_rpl()
1937 mutex_lock(&ep->com.mutex); in abort_rpl()
1938 switch (ep->com.state) { in abort_rpl()
1940 c4iw_wake_up_noref(ep->com.wr_waitp, -ECONNRESET); in abort_rpl()
1941 __state_set(&ep->com, DEAD); in abort_rpl()
1945 pr_err("%s ep %p state %d\n", __func__, ep, ep->com.state); in abort_rpl()
1948 mutex_unlock(&ep->com.mutex); in abort_rpl()
1951 close_complete_upcall(ep, -ECONNRESET); in abort_rpl()
1952 release_ep_resources(ep); in abort_rpl()
1954 c4iw_put_ep(&ep->com); in abort_rpl()
1958 static int send_fw_act_open_req(struct c4iw_ep *ep, unsigned int atid) in send_fw_act_open_req() argument
1972 ep->com.dev->rdev.lldi.ports[0], in send_fw_act_open_req()
1973 ep->l2t)); in send_fw_act_open_req()
1974 sin = (struct sockaddr_in *)&ep->com.local_addr; in send_fw_act_open_req()
1977 sin = (struct sockaddr_in *)&ep->com.remote_addr; in send_fw_act_open_req()
1987 cxgb_best_mtu(ep->com.dev->rdev.lldi.mtus, ep->mtu, &mtu_idx, in send_fw_act_open_req()
1989 (ep->com.remote_addr.ss_family == AF_INET) ? 0 : 1); in send_fw_act_open_req()
1996 win = ep->rcv_win >> 10; in send_fw_act_open_req()
2006 L2T_IDX_V(ep->l2t->idx) | in send_fw_act_open_req()
2007 TX_CHAN_V(ep->tx_chan) | in send_fw_act_open_req()
2008 SMAC_SEL_V(ep->smac_idx) | in send_fw_act_open_req()
2009 DSCP_V(ep->tos >> 2) | in send_fw_act_open_req()
2013 TX_QUEUE_V(ep->com.dev->rdev.lldi.tx_modq[ep->tx_chan]) | in send_fw_act_open_req()
2016 RSS_QUEUE_VALID_F | RSS_QUEUE_V(ep->rss_qid)); in send_fw_act_open_req()
2025 set_wr_txq(skb, CPL_PRIORITY_CONTROL, ep->ctrlq_idx); in send_fw_act_open_req()
2026 set_bit(ACT_OFLD_CONN, &ep->com.history); in send_fw_act_open_req()
2027 return c4iw_l2t_send(&ep->com.dev->rdev, skb, ep->l2t); in send_fw_act_open_req()
2058 static void set_tcp_window(struct c4iw_ep *ep, struct port_info *pi) in set_tcp_window() argument
2060 ep->snd_win = snd_win; in set_tcp_window()
2061 ep->rcv_win = rcv_win; in set_tcp_window()
2063 ep->snd_win, ep->rcv_win); in set_tcp_window()
2068 static int import_ep(struct c4iw_ep *ep, int iptype, __u8 *peer_ip, in import_ep() argument
2099 ep->l2t = cxgb4_l2t_get(cdev->rdev.lldi.l2t, in import_ep()
2101 if (!ep->l2t) { in import_ep()
2105 ep->mtu = pdev->mtu; in import_ep()
2106 ep->tx_chan = cxgb4_port_chan(pdev); in import_ep()
2107 ep->smac_idx = ((struct port_info *)netdev_priv(pdev))->smt_idx; in import_ep()
2110 ep->txq_idx = cxgb4_port_idx(pdev) * step; in import_ep()
2113 ep->ctrlq_idx = cxgb4_port_idx(pdev); in import_ep()
2114 ep->rss_qid = cdev->rdev.lldi.rxq_ids[ in import_ep()
2116 set_tcp_window(ep, (struct port_info *)netdev_priv(pdev)); in import_ep()
2120 ep->l2t = cxgb4_l2t_get(cdev->rdev.lldi.l2t, in import_ep()
2122 if (!ep->l2t) in import_ep()
2124 ep->mtu = dst_mtu(dst); in import_ep()
2125 ep->tx_chan = cxgb4_port_chan(pdev); in import_ep()
2126 ep->smac_idx = ((struct port_info *)netdev_priv(pdev))->smt_idx; in import_ep()
2129 ep->txq_idx = cxgb4_port_idx(pdev) * step; in import_ep()
2130 ep->ctrlq_idx = cxgb4_port_idx(pdev); in import_ep()
2133 ep->rss_qid = cdev->rdev.lldi.rxq_ids[ in import_ep()
2135 set_tcp_window(ep, (struct port_info *)netdev_priv(pdev)); in import_ep()
2138 ep->retry_with_mpa_v1 = 0; in import_ep()
2139 ep->tried_with_mpa_v1 = 0; in import_ep()
2151 static int c4iw_reconnect(struct c4iw_ep *ep) in c4iw_reconnect() argument
2156 &ep->com.cm_id->m_local_addr; in c4iw_reconnect()
2158 &ep->com.cm_id->m_remote_addr; in c4iw_reconnect()
2160 &ep->com.cm_id->m_local_addr; in c4iw_reconnect()
2162 &ep->com.cm_id->m_remote_addr; in c4iw_reconnect()
2166 pr_debug("qp %p cm_id %p\n", ep->com.qp, ep->com.cm_id); in c4iw_reconnect()
2167 c4iw_init_wr_wait(ep->com.wr_waitp); in c4iw_reconnect()
2178 size = (CN_MAX_CON_BUF - skb_queue_len(&ep->com.ep_skb_list)); in c4iw_reconnect()
2179 if (alloc_ep_skb_list(&ep->com.ep_skb_list, size)) { in c4iw_reconnect()
2187 ep->atid = cxgb4_alloc_atid(ep->com.dev->rdev.lldi.tids, ep); in c4iw_reconnect()
2188 if (ep->atid == -1) { in c4iw_reconnect()
2193 err = xa_insert_irq(&ep->com.dev->atids, ep->atid, ep, GFP_KERNEL); in c4iw_reconnect()
2198 if (ep->com.cm_id->m_local_addr.ss_family == AF_INET) { in c4iw_reconnect()
2199 ep->dst = cxgb_find_route(&ep->com.dev->rdev.lldi, get_real_dev, in c4iw_reconnect()
2203 raddr->sin_port, ep->com.cm_id->tos); in c4iw_reconnect()
2207 ep->dst = cxgb_find_route6(&ep->com.dev->rdev.lldi, in c4iw_reconnect()
2213 ep->com.cm_id->tos, in c4iw_reconnect()
2218 if (!ep->dst) { in c4iw_reconnect()
2223 err = import_ep(ep, iptype, ra, ep->dst, ep->com.dev, false, in c4iw_reconnect()
2224 ep->com.dev->rdev.lldi.adapter_type, in c4iw_reconnect()
2225 ep->com.cm_id->tos); in c4iw_reconnect()
2232 ep->txq_idx, ep->tx_chan, ep->smac_idx, ep->rss_qid, in c4iw_reconnect()
2233 ep->l2t->idx); in c4iw_reconnect()
2235 state_set(&ep->com, CONNECTING); in c4iw_reconnect()
2236 ep->tos = ep->com.cm_id->tos; in c4iw_reconnect()
2239 err = send_connect(ep); in c4iw_reconnect()
2243 cxgb4_l2t_release(ep->l2t); in c4iw_reconnect()
2245 dst_release(ep->dst); in c4iw_reconnect()
2247 xa_erase_irq(&ep->com.dev->atids, ep->atid); in c4iw_reconnect()
2249 cxgb4_free_atid(ep->com.dev->rdev.lldi.tids, ep->atid); in c4iw_reconnect()
2257 connect_reply_upcall(ep, -ECONNRESET); in c4iw_reconnect()
2259 c4iw_put_ep(&ep->com); in c4iw_reconnect()
2266 struct c4iw_ep *ep; in act_open_rpl() local
2278 ep = lookup_atid(t, atid); in act_open_rpl()
2279 la = (struct sockaddr_in *)&ep->com.local_addr; in act_open_rpl()
2280 ra = (struct sockaddr_in *)&ep->com.remote_addr; in act_open_rpl()
2281 la6 = (struct sockaddr_in6 *)&ep->com.local_addr; in act_open_rpl()
2282 ra6 = (struct sockaddr_in6 *)&ep->com.remote_addr; in act_open_rpl()
2284 pr_debug("ep %p atid %u status %u errno %d\n", ep, atid, in act_open_rpl()
2290 ep->stats.connect_neg_adv++; in act_open_rpl()
2297 set_bit(ACT_OPEN_RPL, &ep->com.history); in act_open_rpl()
2310 if (ep->com.local_addr.ss_family == AF_INET && in act_open_rpl()
2312 ret = send_fw_act_open_req(ep, TID_TID_G(AOPEN_ATID_G( in act_open_rpl()
2320 if (ep->retry_count++ < ACT_OPEN_RETRY_COUNT) { in act_open_rpl()
2321 set_bit(ACT_RETRY_INUSE, &ep->com.history); in act_open_rpl()
2322 if (ep->com.remote_addr.ss_family == AF_INET6) { in act_open_rpl()
2325 &ep->com.local_addr; in act_open_rpl()
2327 ep->com.dev->rdev.lldi.ports[0], in act_open_rpl()
2331 xa_erase_irq(&ep->com.dev->atids, atid); in act_open_rpl()
2333 dst_release(ep->dst); in act_open_rpl()
2334 cxgb4_l2t_release(ep->l2t); in act_open_rpl()
2335 c4iw_reconnect(ep); in act_open_rpl()
2340 if (ep->com.local_addr.ss_family == AF_INET) { in act_open_rpl()
2355 connect_reply_upcall(ep, status2errno(status)); in act_open_rpl()
2356 state_set(&ep->com, DEAD); in act_open_rpl()
2358 if (ep->com.remote_addr.ss_family == AF_INET6) { in act_open_rpl()
2360 (struct sockaddr_in6 *)&ep->com.local_addr; in act_open_rpl()
2361 cxgb4_clip_release(ep->com.dev->rdev.lldi.ports[0], in act_open_rpl()
2365 cxgb4_remove_tid(ep->com.dev->rdev.lldi.tids, 0, GET_TID(rpl), in act_open_rpl()
2366 ep->com.local_addr.ss_family); in act_open_rpl()
2368 xa_erase_irq(&ep->com.dev->atids, atid); in act_open_rpl()
2370 dst_release(ep->dst); in act_open_rpl()
2371 cxgb4_l2t_release(ep->l2t); in act_open_rpl()
2372 c4iw_put_ep(&ep->com); in act_open_rpl()
2381 struct c4iw_listen_ep *ep = get_ep_from_stid(dev, stid); in pass_open_rpl() local
2383 if (!ep) { in pass_open_rpl()
2387 pr_debug("ep %p status %d error %d\n", ep, in pass_open_rpl()
2389 c4iw_wake_up_noref(ep->com.wr_waitp, status2errno(rpl->status)); in pass_open_rpl()
2390 c4iw_put_ep(&ep->com); in pass_open_rpl()
2399 struct c4iw_listen_ep *ep = get_ep_from_stid(dev, stid); in close_listsrv_rpl() local
2401 if (!ep) { in close_listsrv_rpl()
2405 pr_debug("ep %p\n", ep); in close_listsrv_rpl()
2406 c4iw_wake_up_noref(ep->com.wr_waitp, status2errno(rpl->status)); in close_listsrv_rpl()
2407 c4iw_put_ep(&ep->com); in close_listsrv_rpl()
2412 static int accept_cr(struct c4iw_ep *ep, struct sk_buff *skb, in accept_cr() argument
2422 enum chip_type adapter_type = ep->com.dev->rdev.lldi.adapter_type; in accept_cr()
2424 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in accept_cr()
2425 cxgb_best_mtu(ep->com.dev->rdev.lldi.mtus, ep->mtu, &mtu_idx, in accept_cr()
2427 (ep->com.remote_addr.ss_family == AF_INET) ? 0 : 1); in accept_cr()
2434 win = ep->rcv_win >> 10; in accept_cr()
2442 L2T_IDX_V(ep->l2t->idx) | in accept_cr()
2443 TX_CHAN_V(ep->tx_chan) | in accept_cr()
2444 SMAC_SEL_V(ep->smac_idx) | in accept_cr()
2445 DSCP_V(ep->tos >> 2) | in accept_cr()
2449 RSS_QUEUE_VALID_F | RSS_QUEUE_V(ep->rss_qid); in accept_cr()
2476 INIT_TP_WR(rpl5, ep->hwtid); in accept_cr()
2479 INIT_TP_WR(rpl, ep->hwtid); in accept_cr()
2482 ep->hwtid)); in accept_cr()
2499 set_wr_txq(skb, CPL_PRIORITY_SETUP, ep->ctrlq_idx); in accept_cr()
2500 t4_set_arp_err_handler(skb, ep, pass_accept_rpl_arp_failure); in accept_cr()
2502 return c4iw_l2t_send(&ep->com.dev->rdev, skb, ep->l2t); in accept_cr()
2678 struct c4iw_ep *ep; in pass_establish() local
2684 ep = get_ep_from_tid(dev, tid); in pass_establish()
2685 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in pass_establish()
2686 ep->snd_seq = be32_to_cpu(req->snd_isn); in pass_establish()
2687 ep->rcv_seq = be32_to_cpu(req->rcv_isn); in pass_establish()
2688 ep->snd_wscale = TCPOPT_SND_WSCALE_G(tcp_opt); in pass_establish()
2690 pr_debug("ep %p hwtid %u tcp_opt 0x%02x\n", ep, tid, tcp_opt); in pass_establish()
2692 set_emss(ep, tcp_opt); in pass_establish()
2694 dst_confirm(ep->dst); in pass_establish()
2695 mutex_lock(&ep->com.mutex); in pass_establish()
2696 ep->com.state = MPA_REQ_WAIT; in pass_establish()
2697 start_ep_timer(ep); in pass_establish()
2698 set_bit(PASS_ESTAB, &ep->com.history); in pass_establish()
2699 ret = send_flowc(ep); in pass_establish()
2700 mutex_unlock(&ep->com.mutex); in pass_establish()
2702 c4iw_ep_disconnect(ep, 1, GFP_KERNEL); in pass_establish()
2703 c4iw_put_ep(&ep->com); in pass_establish()
2711 struct c4iw_ep *ep; in peer_close() local
2718 ep = get_ep_from_tid(dev, tid); in peer_close()
2719 if (!ep) in peer_close()
2722 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in peer_close()
2723 dst_confirm(ep->dst); in peer_close()
2725 set_bit(PEER_CLOSE, &ep->com.history); in peer_close()
2726 mutex_lock(&ep->com.mutex); in peer_close()
2727 switch (ep->com.state) { in peer_close()
2729 __state_set(&ep->com, CLOSING); in peer_close()
2732 __state_set(&ep->com, CLOSING); in peer_close()
2733 connect_reply_upcall(ep, -ECONNRESET); in peer_close()
2743 __state_set(&ep->com, CLOSING); in peer_close()
2744 pr_debug("waking up ep %p tid %u\n", ep, ep->hwtid); in peer_close()
2745 c4iw_wake_up_noref(ep->com.wr_waitp, -ECONNRESET); in peer_close()
2748 __state_set(&ep->com, CLOSING); in peer_close()
2749 pr_debug("waking up ep %p tid %u\n", ep, ep->hwtid); in peer_close()
2750 c4iw_wake_up_noref(ep->com.wr_waitp, -ECONNRESET); in peer_close()
2753 start_ep_timer(ep); in peer_close()
2754 __state_set(&ep->com, CLOSING); in peer_close()
2756 ret = c4iw_modify_qp(ep->com.qp->rhp, ep->com.qp, in peer_close()
2759 peer_close_upcall(ep); in peer_close()
2767 __state_set(&ep->com, MORIBUND); in peer_close()
2771 (void)stop_ep_timer(ep); in peer_close()
2772 if (ep->com.cm_id && ep->com.qp) { in peer_close()
2774 c4iw_modify_qp(ep->com.qp->rhp, ep->com.qp, in peer_close()
2777 close_complete_upcall(ep, 0); in peer_close()
2778 __state_set(&ep->com, DEAD); in peer_close()
2786 WARN_ONCE(1, "Bad endpoint state %u\n", ep->com.state); in peer_close()
2788 mutex_unlock(&ep->com.mutex); in peer_close()
2790 c4iw_ep_disconnect(ep, 0, GFP_KERNEL); in peer_close()
2792 release_ep_resources(ep); in peer_close()
2793 c4iw_put_ep(&ep->com); in peer_close()
2797 static void finish_peer_abort(struct c4iw_dev *dev, struct c4iw_ep *ep) in finish_peer_abort() argument
2799 complete_cached_srq_buffers(ep, ep->srqe_idx); in finish_peer_abort()
2800 if (ep->com.cm_id && ep->com.qp) { in finish_peer_abort()
2804 c4iw_modify_qp(ep->com.qp->rhp, ep->com.qp, in finish_peer_abort()
2807 peer_abort_upcall(ep); in finish_peer_abort()
2808 release_ep_resources(ep); in finish_peer_abort()
2809 c4iw_put_ep(&ep->com); in finish_peer_abort()
2815 struct c4iw_ep *ep; in peer_abort() local
2826 ep = get_ep_from_tid(dev, tid); in peer_abort()
2827 if (!ep) in peer_abort()
2834 ep->hwtid, status, neg_adv_str(status)); in peer_abort()
2835 ep->stats.abort_neg_adv++; in peer_abort()
2842 pr_debug("ep %p tid %u state %u\n", ep, ep->hwtid, in peer_abort()
2843 ep->com.state); in peer_abort()
2844 set_bit(PEER_ABORT, &ep->com.history); in peer_abort()
2851 if (ep->com.state != MPA_REQ_SENT) in peer_abort()
2852 c4iw_wake_up_noref(ep->com.wr_waitp, -ECONNRESET); in peer_abort()
2854 mutex_lock(&ep->com.mutex); in peer_abort()
2855 switch (ep->com.state) { in peer_abort()
2857 c4iw_put_ep(&ep->parent_ep->com); in peer_abort()
2860 (void)stop_ep_timer(ep); in peer_abort()
2863 (void)stop_ep_timer(ep); in peer_abort()
2865 (mpa_rev == 2 && ep->tried_with_mpa_v1)) in peer_abort()
2866 connect_reply_upcall(ep, -ECONNRESET); in peer_abort()
2878 ep->retry_with_mpa_v1 = 1; in peer_abort()
2887 stop_ep_timer(ep); in peer_abort()
2890 if (ep->com.qp && ep->com.qp->srq) { in peer_abort()
2894 complete_cached_srq_buffers(ep, srqidx); in peer_abort()
2897 c4iw_get_ep(&ep->com); in peer_abort()
2898 __state_set(&ep->com, ABORTING); in peer_abort()
2899 set_bit(PEER_ABORT_IN_PROGRESS, &ep->com.flags); in peer_abort()
2900 read_tcb(ep); in peer_abort()
2906 if (ep->com.cm_id && ep->com.qp) { in peer_abort()
2908 ret = c4iw_modify_qp(ep->com.qp->rhp, in peer_abort()
2909 ep->com.qp, C4IW_QP_ATTR_NEXT_STATE, in peer_abort()
2914 peer_abort_upcall(ep); in peer_abort()
2920 mutex_unlock(&ep->com.mutex); in peer_abort()
2923 WARN_ONCE(1, "Bad endpoint state %u\n", ep->com.state); in peer_abort()
2926 dst_confirm(ep->dst); in peer_abort()
2927 if (ep->com.state != ABORTING) { in peer_abort()
2928 __state_set(&ep->com, DEAD); in peer_abort()
2930 if (!ep->retry_with_mpa_v1) in peer_abort()
2933 mutex_unlock(&ep->com.mutex); in peer_abort()
2935 rpl_skb = skb_dequeue(&ep->com.ep_skb_list); in peer_abort()
2941 cxgb_mk_abort_rpl(rpl_skb, len, ep->hwtid, ep->txq_idx); in peer_abort()
2943 c4iw_ofld_send(&ep->com.dev->rdev, rpl_skb); in peer_abort()
2946 release_ep_resources(ep); in peer_abort()
2947 else if (ep->retry_with_mpa_v1) { in peer_abort()
2948 if (ep->com.remote_addr.ss_family == AF_INET6) { in peer_abort()
2951 &ep->com.local_addr; in peer_abort()
2953 ep->com.dev->rdev.lldi.ports[0], in peer_abort()
2957 xa_erase_irq(&ep->com.dev->hwtids, ep->hwtid); in peer_abort()
2958 cxgb4_remove_tid(ep->com.dev->rdev.lldi.tids, 0, ep->hwtid, in peer_abort()
2959 ep->com.local_addr.ss_family); in peer_abort()
2960 dst_release(ep->dst); in peer_abort()
2961 cxgb4_l2t_release(ep->l2t); in peer_abort()
2962 c4iw_reconnect(ep); in peer_abort()
2966 c4iw_put_ep(&ep->com); in peer_abort()
2968 c4iw_put_ep(&ep->com); in peer_abort()
2974 struct c4iw_ep *ep; in close_con_rpl() local
2980 ep = get_ep_from_tid(dev, tid); in close_con_rpl()
2981 if (!ep) in close_con_rpl()
2984 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in close_con_rpl()
2987 mutex_lock(&ep->com.mutex); in close_con_rpl()
2988 set_bit(CLOSE_CON_RPL, &ep->com.history); in close_con_rpl()
2989 switch (ep->com.state) { in close_con_rpl()
2991 __state_set(&ep->com, MORIBUND); in close_con_rpl()
2994 (void)stop_ep_timer(ep); in close_con_rpl()
2995 if ((ep->com.cm_id) && (ep->com.qp)) { in close_con_rpl()
2997 c4iw_modify_qp(ep->com.qp->rhp, in close_con_rpl()
2998 ep->com.qp, in close_con_rpl()
3002 close_complete_upcall(ep, 0); in close_con_rpl()
3003 __state_set(&ep->com, DEAD); in close_con_rpl()
3010 WARN_ONCE(1, "Bad endpoint state %u\n", ep->com.state); in close_con_rpl()
3013 mutex_unlock(&ep->com.mutex); in close_con_rpl()
3015 release_ep_resources(ep); in close_con_rpl()
3016 c4iw_put_ep(&ep->com); in close_con_rpl()
3024 struct c4iw_ep *ep; in terminate() local
3027 ep = get_ep_from_tid(dev, tid); in terminate()
3029 if (ep) { in terminate()
3030 if (ep->com.qp) { in terminate()
3032 ep->com.qp->wq.sq.qid); in terminate()
3034 c4iw_modify_qp(ep->com.qp->rhp, ep->com.qp, in terminate()
3041 c4iw_ep_disconnect(ep, 1, GFP_KERNEL); in terminate()
3042 c4iw_put_ep(&ep->com); in terminate()
3056 struct c4iw_ep *ep; in fw4_ack() local
3062 ep = get_ep_from_tid(dev, tid); in fw4_ack()
3063 if (!ep) in fw4_ack()
3066 ep, ep->hwtid, credits); in fw4_ack()
3069 ep, ep->hwtid, state_read(&ep->com)); in fw4_ack()
3073 dst_confirm(ep->dst); in fw4_ack()
3074 if (ep->mpa_skb) { in fw4_ack()
3076 ep, ep->hwtid, state_read(&ep->com), in fw4_ack()
3077 ep->mpa_attr.initiator ? 1 : 0); in fw4_ack()
3078 mutex_lock(&ep->com.mutex); in fw4_ack()
3079 kfree_skb(ep->mpa_skb); in fw4_ack()
3080 ep->mpa_skb = NULL; in fw4_ack()
3081 if (test_bit(STOP_MPA_TIMER, &ep->com.flags)) in fw4_ack()
3082 stop_ep_timer(ep); in fw4_ack()
3083 mutex_unlock(&ep->com.mutex); in fw4_ack()
3086 c4iw_put_ep(&ep->com); in fw4_ack()
3093 struct c4iw_ep *ep = to_ep(cm_id); in c4iw_reject_cr() local
3095 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in c4iw_reject_cr()
3097 mutex_lock(&ep->com.mutex); in c4iw_reject_cr()
3098 if (ep->com.state != MPA_REQ_RCVD) { in c4iw_reject_cr()
3099 mutex_unlock(&ep->com.mutex); in c4iw_reject_cr()
3100 c4iw_put_ep(&ep->com); in c4iw_reject_cr()
3103 set_bit(ULP_REJECT, &ep->com.history); in c4iw_reject_cr()
3107 abort = send_mpa_reject(ep, pdata, pdata_len); in c4iw_reject_cr()
3108 mutex_unlock(&ep->com.mutex); in c4iw_reject_cr()
3110 stop_ep_timer(ep); in c4iw_reject_cr()
3111 c4iw_ep_disconnect(ep, abort != 0, GFP_KERNEL); in c4iw_reject_cr()
3112 c4iw_put_ep(&ep->com); in c4iw_reject_cr()
3121 struct c4iw_ep *ep = to_ep(cm_id); in c4iw_accept_cr() local
3126 pr_debug("ep %p tid %u\n", ep, ep->hwtid); in c4iw_accept_cr()
3128 mutex_lock(&ep->com.mutex); in c4iw_accept_cr()
3129 if (ep->com.state != MPA_REQ_RCVD) { in c4iw_accept_cr()
3139 set_bit(ULP_ACCEPT, &ep->com.history); in c4iw_accept_cr()
3140 if ((conn_param->ord > cur_max_read_depth(ep->com.dev)) || in c4iw_accept_cr()
3141 (conn_param->ird > cur_max_read_depth(ep->com.dev))) { in c4iw_accept_cr()
3146 if (ep->mpa_attr.version == 2 && ep->mpa_attr.enhanced_rdma_conn) { in c4iw_accept_cr()
3147 if (conn_param->ord > ep->ird) { in c4iw_accept_cr()
3149 conn_param->ord = ep->ird; in c4iw_accept_cr()
3151 ep->ird = conn_param->ird; in c4iw_accept_cr()
3152 ep->ord = conn_param->ord; in c4iw_accept_cr()
3153 send_mpa_reject(ep, conn_param->private_data, in c4iw_accept_cr()
3159 if (conn_param->ird < ep->ord) { in c4iw_accept_cr()
3161 ep->ord <= h->rdev.lldi.max_ordird_qp) { in c4iw_accept_cr()
3162 conn_param->ird = ep->ord; in c4iw_accept_cr()
3169 ep->ird = conn_param->ird; in c4iw_accept_cr()
3170 ep->ord = conn_param->ord; in c4iw_accept_cr()
3172 if (ep->mpa_attr.version == 1) { in c4iw_accept_cr()
3173 if (peer2peer && ep->ird == 0) in c4iw_accept_cr()
3174 ep->ird = 1; in c4iw_accept_cr()
3177 (ep->mpa_attr.p2p_type != FW_RI_INIT_P2PTYPE_DISABLED) && in c4iw_accept_cr()
3178 (p2p_type == FW_RI_INIT_P2PTYPE_READ_REQ) && ep->ird == 0) in c4iw_accept_cr()
3179 ep->ird = 1; in c4iw_accept_cr()
3182 pr_debug("ird %d ord %d\n", ep->ird, ep->ord); in c4iw_accept_cr()
3184 ep->com.cm_id = cm_id; in c4iw_accept_cr()
3185 ref_cm_id(&ep->com); in c4iw_accept_cr()
3186 ep->com.qp = qp; in c4iw_accept_cr()
3187 ref_qp(ep); in c4iw_accept_cr()
3190 attrs.mpa_attr = ep->mpa_attr; in c4iw_accept_cr()
3191 attrs.max_ird = ep->ird; in c4iw_accept_cr()
3192 attrs.max_ord = ep->ord; in c4iw_accept_cr()
3193 attrs.llp_stream_handle = ep; in c4iw_accept_cr()
3203 err = c4iw_modify_qp(ep->com.qp->rhp, in c4iw_accept_cr()
3204 ep->com.qp, mask, &attrs, 1); in c4iw_accept_cr()
3208 set_bit(STOP_MPA_TIMER, &ep->com.flags); in c4iw_accept_cr()
3209 err = send_mpa_reply(ep, conn_param->private_data, in c4iw_accept_cr()
3214 __state_set(&ep->com, FPDU_MODE); in c4iw_accept_cr()
3215 established_upcall(ep); in c4iw_accept_cr()
3216 mutex_unlock(&ep->com.mutex); in c4iw_accept_cr()
3217 c4iw_put_ep(&ep->com); in c4iw_accept_cr()
3220 deref_cm_id(&ep->com); in c4iw_accept_cr()
3224 mutex_unlock(&ep->com.mutex); in c4iw_accept_cr()
3226 c4iw_ep_disconnect(ep, 1, GFP_KERNEL); in c4iw_accept_cr()
3227 c4iw_put_ep(&ep->com); in c4iw_accept_cr()
3300 struct c4iw_ep *ep; in c4iw_connect() local
3314 ep = alloc_ep(sizeof(*ep), GFP_KERNEL); in c4iw_connect()
3315 if (!ep) { in c4iw_connect()
3321 skb_queue_head_init(&ep->com.ep_skb_list); in c4iw_connect()
3322 if (alloc_ep_skb_list(&ep->com.ep_skb_list, CN_MAX_CON_BUF)) { in c4iw_connect()
3327 timer_setup(&ep->timer, ep_timeout, 0); in c4iw_connect()
3328 ep->plen = conn_param->private_data_len; in c4iw_connect()
3329 if (ep->plen) in c4iw_connect()
3330 memcpy(ep->mpa_pkt + sizeof(struct mpa_message), in c4iw_connect()
3331 conn_param->private_data, ep->plen); in c4iw_connect()
3332 ep->ird = conn_param->ird; in c4iw_connect()
3333 ep->ord = conn_param->ord; in c4iw_connect()
3335 if (peer2peer && ep->ord == 0) in c4iw_connect()
3336 ep->ord = 1; in c4iw_connect()
3338 ep->com.cm_id = cm_id; in c4iw_connect()
3339 ref_cm_id(&ep->com); in c4iw_connect()
3340 cm_id->provider_data = ep; in c4iw_connect()
3341 ep->com.dev = dev; in c4iw_connect()
3342 ep->com.qp = get_qhp(dev, conn_param->qpn); in c4iw_connect()
3343 if (!ep->com.qp) { in c4iw_connect()
3348 ref_qp(ep); in c4iw_connect()
3350 ep->com.qp, cm_id); in c4iw_connect()
3355 ep->atid = cxgb4_alloc_atid(dev->rdev.lldi.tids, ep); in c4iw_connect()
3356 if (ep->atid == -1) { in c4iw_connect()
3361 err = xa_insert_irq(&dev->atids, ep->atid, ep, GFP_KERNEL); in c4iw_connect()
3365 memcpy(&ep->com.local_addr, &cm_id->m_local_addr, in c4iw_connect()
3366 sizeof(ep->com.local_addr)); in c4iw_connect()
3367 memcpy(&ep->com.remote_addr, &cm_id->m_remote_addr, in c4iw_connect()
3368 sizeof(ep->com.remote_addr)); in c4iw_connect()
3370 laddr = (struct sockaddr_in *)&ep->com.local_addr; in c4iw_connect()
3371 raddr = (struct sockaddr_in *)&ep->com.remote_addr; in c4iw_connect()
3372 laddr6 = (struct sockaddr_in6 *)&ep->com.local_addr; in c4iw_connect()
3373 raddr6 = (struct sockaddr_in6 *) &ep->com.remote_addr; in c4iw_connect()
3392 ep->dst = cxgb_find_route(&dev->rdev.lldi, get_real_dev, in c4iw_connect()
3415 ep->dst = cxgb_find_route6(&dev->rdev.lldi, get_real_dev, in c4iw_connect()
3422 if (!ep->dst) { in c4iw_connect()
3428 err = import_ep(ep, iptype, ra, ep->dst, ep->com.dev, true, in c4iw_connect()
3429 ep->com.dev->rdev.lldi.adapter_type, cm_id->tos); in c4iw_connect()
3436 ep->txq_idx, ep->tx_chan, ep->smac_idx, ep->rss_qid, in c4iw_connect()
3437 ep->l2t->idx); in c4iw_connect()
3439 state_set(&ep->com, CONNECTING); in c4iw_connect()
3440 ep->tos = cm_id->tos; in c4iw_connect()
3443 err = send_connect(ep); in c4iw_connect()
3447 cxgb4_l2t_release(ep->l2t); in c4iw_connect()
3449 dst_release(ep->dst); in c4iw_connect()
3451 xa_erase_irq(&ep->com.dev->atids, ep->atid); in c4iw_connect()
3453 cxgb4_free_atid(ep->com.dev->rdev.lldi.tids, ep->atid); in c4iw_connect()
3455 skb_queue_purge(&ep->com.ep_skb_list); in c4iw_connect()
3456 deref_cm_id(&ep->com); in c4iw_connect()
3458 c4iw_put_ep(&ep->com); in c4iw_connect()
3463 static int create_server6(struct c4iw_dev *dev, struct c4iw_listen_ep *ep) in create_server6() argument
3467 &ep->com.local_addr; in create_server6()
3470 err = cxgb4_clip_get(ep->com.dev->rdev.lldi.ports[0], in create_server6()
3475 c4iw_init_wr_wait(ep->com.wr_waitp); in create_server6()
3476 err = cxgb4_create_server6(ep->com.dev->rdev.lldi.ports[0], in create_server6()
3477 ep->stid, &sin6->sin6_addr, in create_server6()
3479 ep->com.dev->rdev.lldi.rxq_ids[0]); in create_server6()
3481 err = c4iw_wait_for_reply(&ep->com.dev->rdev, in create_server6()
3482 ep->com.wr_waitp, in create_server6()
3487 cxgb4_clip_release(ep->com.dev->rdev.lldi.ports[0], in create_server6()
3490 err, ep->stid, in create_server6()
3496 static int create_server4(struct c4iw_dev *dev, struct c4iw_listen_ep *ep) in create_server4() argument
3500 &ep->com.local_addr; in create_server4()
3505 ep->com.dev->rdev.lldi.ports[0], ep->stid, in create_server4()
3507 ep->com.dev->rdev.lldi.rxq_ids[0], 0, 0); in create_server4()
3509 if (c4iw_fatal_error(&ep->com.dev->rdev)) { in create_server4()
3518 c4iw_init_wr_wait(ep->com.wr_waitp); in create_server4()
3519 err = cxgb4_create_server(ep->com.dev->rdev.lldi.ports[0], in create_server4()
3520 ep->stid, sin->sin_addr.s_addr, sin->sin_port, in create_server4()
3521 0, ep->com.dev->rdev.lldi.rxq_ids[0]); in create_server4()
3523 err = c4iw_wait_for_reply(&ep->com.dev->rdev, in create_server4()
3524 ep->com.wr_waitp, in create_server4()
3531 , err, ep->stid, in create_server4()
3540 struct c4iw_listen_ep *ep; in c4iw_create_listen() local
3544 ep = alloc_ep(sizeof(*ep), GFP_KERNEL); in c4iw_create_listen()
3545 if (!ep) { in c4iw_create_listen()
3550 skb_queue_head_init(&ep->com.ep_skb_list); in c4iw_create_listen()
3551 pr_debug("ep %p\n", ep); in c4iw_create_listen()
3552 ep->com.cm_id = cm_id; in c4iw_create_listen()
3553 ref_cm_id(&ep->com); in c4iw_create_listen()
3554 ep->com.dev = dev; in c4iw_create_listen()
3555 ep->backlog = backlog; in c4iw_create_listen()
3556 memcpy(&ep->com.local_addr, &cm_id->m_local_addr, in c4iw_create_listen()
3557 sizeof(ep->com.local_addr)); in c4iw_create_listen()
3563 ep->com.local_addr.ss_family == AF_INET) in c4iw_create_listen()
3564 ep->stid = cxgb4_alloc_sftid(dev->rdev.lldi.tids, in c4iw_create_listen()
3565 cm_id->m_local_addr.ss_family, ep); in c4iw_create_listen()
3567 ep->stid = cxgb4_alloc_stid(dev->rdev.lldi.tids, in c4iw_create_listen()
3568 cm_id->m_local_addr.ss_family, ep); in c4iw_create_listen()
3570 if (ep->stid == -1) { in c4iw_create_listen()
3575 err = xa_insert_irq(&dev->stids, ep->stid, ep, GFP_KERNEL); in c4iw_create_listen()
3579 state_set(&ep->com, LISTEN); in c4iw_create_listen()
3580 if (ep->com.local_addr.ss_family == AF_INET) in c4iw_create_listen()
3581 err = create_server4(dev, ep); in c4iw_create_listen()
3583 err = create_server6(dev, ep); in c4iw_create_listen()
3585 cm_id->provider_data = ep; in c4iw_create_listen()
3588 xa_erase_irq(&ep->com.dev->stids, ep->stid); in c4iw_create_listen()
3590 cxgb4_free_stid(ep->com.dev->rdev.lldi.tids, ep->stid, in c4iw_create_listen()
3591 ep->com.local_addr.ss_family); in c4iw_create_listen()
3593 deref_cm_id(&ep->com); in c4iw_create_listen()
3594 c4iw_put_ep(&ep->com); in c4iw_create_listen()
3603 struct c4iw_listen_ep *ep = to_listen_ep(cm_id); in c4iw_destroy_listen() local
3605 pr_debug("ep %p\n", ep); in c4iw_destroy_listen()
3608 state_set(&ep->com, DEAD); in c4iw_destroy_listen()
3609 if (ep->com.dev->rdev.lldi.enable_fw_ofld_conn && in c4iw_destroy_listen()
3610 ep->com.local_addr.ss_family == AF_INET) { in c4iw_destroy_listen()
3612 ep->com.dev->rdev.lldi.ports[0], ep->stid, in c4iw_destroy_listen()
3613 ep->com.dev->rdev.lldi.rxq_ids[0], false); in c4iw_destroy_listen()
3616 c4iw_init_wr_wait(ep->com.wr_waitp); in c4iw_destroy_listen()
3618 ep->com.dev->rdev.lldi.ports[0], ep->stid, in c4iw_destroy_listen()
3619 ep->com.dev->rdev.lldi.rxq_ids[0], in c4iw_destroy_listen()
3620 ep->com.local_addr.ss_family == AF_INET6); in c4iw_destroy_listen()
3623 err = c4iw_wait_for_reply(&ep->com.dev->rdev, ep->com.wr_waitp, in c4iw_destroy_listen()
3625 sin6 = (struct sockaddr_in6 *)&ep->com.local_addr; in c4iw_destroy_listen()
3626 cxgb4_clip_release(ep->com.dev->rdev.lldi.ports[0], in c4iw_destroy_listen()
3629 xa_erase_irq(&ep->com.dev->stids, ep->stid); in c4iw_destroy_listen()
3630 cxgb4_free_stid(ep->com.dev->rdev.lldi.tids, ep->stid, in c4iw_destroy_listen()
3631 ep->com.local_addr.ss_family); in c4iw_destroy_listen()
3633 deref_cm_id(&ep->com); in c4iw_destroy_listen()
3634 c4iw_put_ep(&ep->com); in c4iw_destroy_listen()
3638 int c4iw_ep_disconnect(struct c4iw_ep *ep, int abrupt, gfp_t gfp) in c4iw_ep_disconnect() argument
3645 mutex_lock(&ep->com.mutex); in c4iw_ep_disconnect()
3647 pr_debug("ep %p state %s, abrupt %d\n", ep, in c4iw_ep_disconnect()
3648 states[ep->com.state], abrupt); in c4iw_ep_disconnect()
3654 c4iw_get_ep(&ep->com); in c4iw_ep_disconnect()
3656 rdev = &ep->com.dev->rdev; in c4iw_ep_disconnect()
3659 close_complete_upcall(ep, -EIO); in c4iw_ep_disconnect()
3660 ep->com.state = DEAD; in c4iw_ep_disconnect()
3662 switch (ep->com.state) { in c4iw_ep_disconnect()
3671 ep->com.state = ABORTING; in c4iw_ep_disconnect()
3673 ep->com.state = CLOSING; in c4iw_ep_disconnect()
3679 if (ep->mpa_skb && in c4iw_ep_disconnect()
3680 test_bit(STOP_MPA_TIMER, &ep->com.flags)) { in c4iw_ep_disconnect()
3681 clear_bit(STOP_MPA_TIMER, &ep->com.flags); in c4iw_ep_disconnect()
3682 stop_ep_timer(ep); in c4iw_ep_disconnect()
3684 start_ep_timer(ep); in c4iw_ep_disconnect()
3686 set_bit(CLOSE_SENT, &ep->com.flags); in c4iw_ep_disconnect()
3689 if (!test_and_set_bit(CLOSE_SENT, &ep->com.flags)) { in c4iw_ep_disconnect()
3692 (void)stop_ep_timer(ep); in c4iw_ep_disconnect()
3693 ep->com.state = ABORTING; in c4iw_ep_disconnect()
3695 ep->com.state = MORIBUND; in c4iw_ep_disconnect()
3702 ep, ep->com.state); in c4iw_ep_disconnect()
3705 WARN_ONCE(1, "Bad endpoint state %u\n", ep->com.state); in c4iw_ep_disconnect()
3711 set_bit(EP_DISC_ABORT, &ep->com.history); in c4iw_ep_disconnect()
3712 ret = send_abort(ep); in c4iw_ep_disconnect()
3714 set_bit(EP_DISC_CLOSE, &ep->com.history); in c4iw_ep_disconnect()
3715 ret = send_halfclose(ep); in c4iw_ep_disconnect()
3718 set_bit(EP_DISC_FAIL, &ep->com.history); in c4iw_ep_disconnect()
3720 stop_ep_timer(ep); in c4iw_ep_disconnect()
3721 close_complete_upcall(ep, -EIO); in c4iw_ep_disconnect()
3723 if (ep->com.qp) { in c4iw_ep_disconnect()
3727 ret = c4iw_modify_qp(ep->com.qp->rhp, in c4iw_ep_disconnect()
3728 ep->com.qp, in c4iw_ep_disconnect()
3738 mutex_unlock(&ep->com.mutex); in c4iw_ep_disconnect()
3739 c4iw_put_ep(&ep->com); in c4iw_ep_disconnect()
3741 release_ep_resources(ep); in c4iw_ep_disconnect()
3748 struct c4iw_ep *ep; in active_ofld_conn_reply() local
3751 ep = (struct c4iw_ep *)lookup_atid(dev->rdev.lldi.tids, in active_ofld_conn_reply()
3753 if (!ep) in active_ofld_conn_reply()
3758 set_bit(ACT_RETRY_NOMEM, &ep->com.history); in active_ofld_conn_reply()
3759 if (ep->retry_count++ < ACT_OPEN_RETRY_COUNT) { in active_ofld_conn_reply()
3760 send_fw_act_open_req(ep, atid); in active_ofld_conn_reply()
3765 set_bit(ACT_RETRY_INUSE, &ep->com.history); in active_ofld_conn_reply()
3766 if (ep->retry_count++ < ACT_OPEN_RETRY_COUNT) { in active_ofld_conn_reply()
3767 send_fw_act_open_req(ep, atid); in active_ofld_conn_reply()
3781 connect_reply_upcall(ep, status2errno(req->retval)); in active_ofld_conn_reply()
3782 state_set(&ep->com, DEAD); in active_ofld_conn_reply()
3783 if (ep->com.remote_addr.ss_family == AF_INET6) { in active_ofld_conn_reply()
3785 (struct sockaddr_in6 *)&ep->com.local_addr; in active_ofld_conn_reply()
3786 cxgb4_clip_release(ep->com.dev->rdev.lldi.ports[0], in active_ofld_conn_reply()
3791 dst_release(ep->dst); in active_ofld_conn_reply()
3792 cxgb4_l2t_release(ep->l2t); in active_ofld_conn_reply()
3793 c4iw_put_ep(&ep->com); in active_ofld_conn_reply()
3850 struct c4iw_ep *ep; in read_tcb_rpl() local
3854 ep = get_ep_from_tid(dev, tid); in read_tcb_rpl()
3855 if (!ep) in read_tcb_rpl()
3868 c4iw_put_ep(&ep->com); /* from get_ep_from_tid() */ in read_tcb_rpl()
3869 c4iw_put_ep(&ep->com); /* from read_tcb() */ in read_tcb_rpl()
3873 if (++ep->rx_pdu_out_cnt >= 2) { in read_tcb_rpl()
3877 read_tcb(ep); in read_tcb_rpl()
3881 ep->srqe_idx = t4_tcb_get_field32(tcb, TCB_RQ_START_W, TCB_RQ_START_M, in read_tcb_rpl()
3884 pr_debug("ep %p tid %u %016x\n", ep, ep->hwtid, ep->srqe_idx); in read_tcb_rpl()
3886 if (test_bit(PEER_ABORT_IN_PROGRESS, &ep->com.flags)) in read_tcb_rpl()
3887 finish_peer_abort(dev, ep); in read_tcb_rpl()
3888 else if (test_bit(ABORT_REQ_IN_PROGRESS, &ep->com.flags)) in read_tcb_rpl()
3889 send_abort_req(ep); in read_tcb_rpl()
4223 static void process_timeout(struct c4iw_ep *ep) in process_timeout() argument
4228 mutex_lock(&ep->com.mutex); in process_timeout()
4229 pr_debug("ep %p tid %u state %d\n", ep, ep->hwtid, ep->com.state); in process_timeout()
4230 set_bit(TIMEDOUT, &ep->com.history); in process_timeout()
4231 switch (ep->com.state) { in process_timeout()
4233 connect_reply_upcall(ep, -ETIMEDOUT); in process_timeout()
4242 if (ep->com.cm_id && ep->com.qp) { in process_timeout()
4244 c4iw_modify_qp(ep->com.qp->rhp, in process_timeout()
4245 ep->com.qp, C4IW_QP_ATTR_NEXT_STATE, in process_timeout()
4248 close_complete_upcall(ep, -ETIMEDOUT); in process_timeout()
4262 __func__, ep, ep->hwtid, ep->com.state); in process_timeout()
4265 mutex_unlock(&ep->com.mutex); in process_timeout()
4267 c4iw_ep_disconnect(ep, 1, GFP_KERNEL); in process_timeout()
4268 c4iw_put_ep(&ep->com); in process_timeout()
4273 struct c4iw_ep *ep; in process_timedout_eps() local
4284 ep = list_entry(tmp, struct c4iw_ep, entry); in process_timedout_eps()
4285 process_timeout(ep); in process_timedout_eps()
4322 struct c4iw_ep *ep = from_timer(ep, t, timer); in ep_timeout() local
4326 if (!test_and_set_bit(TIMEOUT, &ep->com.flags)) { in ep_timeout()
4330 if (!ep->entry.next) { in ep_timeout()
4331 list_add_tail(&ep->entry, &timeout_list); in ep_timeout()
4404 struct c4iw_ep *ep; in peer_abort_intr() local
4407 ep = get_ep_from_tid(dev, tid); in peer_abort_intr()
4409 if (!ep) { in peer_abort_intr()
4416 ep->hwtid, req->status, in peer_abort_intr()
4420 pr_debug("ep %p tid %u state %u\n", ep, ep->hwtid, ep->com.state); in peer_abort_intr()
4422 c4iw_wake_up_noref(ep->com.wr_waitp, -ECONNRESET); in peer_abort_intr()