• Home
  • Raw
  • Download

Lines Matching refs:resp

118 	struct rdma_ucm_event_resp resp;  member
274 uevent->resp.uid = uevent->mc->uid; in ucma_create_uevent()
275 uevent->resp.id = uevent->mc->id; in ucma_create_uevent()
278 uevent->resp.uid = ctx->uid; in ucma_create_uevent()
279 uevent->resp.id = ctx->id; in ucma_create_uevent()
282 uevent->resp.event = event->event; in ucma_create_uevent()
283 uevent->resp.status = event->status; in ucma_create_uevent()
285 ucma_copy_ud_event(ctx->cm_id->device, &uevent->resp.param.ud, in ucma_create_uevent()
288 ucma_copy_conn_event(&uevent->resp.param.conn, in ucma_create_uevent()
291 uevent->resp.ece.vendor_id = event->ece.vendor_id; in ucma_create_uevent()
292 uevent->resp.ece.attr_mod = event->ece.attr_mod; in ucma_create_uevent()
314 uevent->resp.id = ctx->id; in ucma_connect_event_handler()
379 if (out_len < sizeof(uevent->resp) - sizeof(uevent->resp.reserved) - in ucma_get_event()
380 sizeof(uevent->resp.ece)) in ucma_get_event()
403 &uevent->resp, in ucma_get_event()
404 min_t(size_t, out_len, sizeof(uevent->resp)))) { in ucma_get_event()
413 if (uevent->resp.event == RDMA_CM_EVENT_CONNECT_REQUEST) in ucma_get_event()
443 struct rdma_ucm_create_id_resp resp; in ucma_create_id() local
449 if (out_len < sizeof(resp)) in ucma_create_id()
471 resp.id = ctx->id; in ucma_create_id()
473 &resp, sizeof(resp))) { in ucma_create_id()
534 if (uevent->resp.event == RDMA_CM_EVENT_CONNECT_REQUEST && in ucma_cleanup_ctx_events()
595 struct rdma_ucm_destroy_id_resp resp; in ucma_destroy_id() local
599 if (out_len < sizeof(resp)) in ucma_destroy_id()
617 resp.events_reported = ucma_destroy_private_ctx(ctx); in ucma_destroy_id()
619 &resp, sizeof(resp))) in ucma_destroy_id()
752 static void ucma_copy_ib_route(struct rdma_ucm_query_route_resp *resp, in ucma_copy_ib_route() argument
757 resp->num_paths = route->num_pri_alt_paths; in ucma_copy_ib_route()
762 (union ib_gid *) &resp->ib_route[0].dgid); in ucma_copy_ib_route()
764 (union ib_gid *) &resp->ib_route[0].sgid); in ucma_copy_ib_route()
765 resp->ib_route[0].pkey = cpu_to_be16(ib_addr_get_pkey(dev_addr)); in ucma_copy_ib_route()
768 ib_copy_path_rec_to_user(&resp->ib_route[1], in ucma_copy_ib_route()
772 ib_copy_path_rec_to_user(&resp->ib_route[0], in ucma_copy_ib_route()
780 static void ucma_copy_iboe_route(struct rdma_ucm_query_route_resp *resp, in ucma_copy_iboe_route() argument
784 resp->num_paths = route->num_pri_alt_paths; in ucma_copy_iboe_route()
788 (union ib_gid *)&resp->ib_route[0].dgid); in ucma_copy_iboe_route()
790 (union ib_gid *)&resp->ib_route[0].sgid); in ucma_copy_iboe_route()
791 resp->ib_route[0].pkey = cpu_to_be16(0xffff); in ucma_copy_iboe_route()
794 ib_copy_path_rec_to_user(&resp->ib_route[1], in ucma_copy_iboe_route()
798 ib_copy_path_rec_to_user(&resp->ib_route[0], in ucma_copy_iboe_route()
806 static void ucma_copy_iw_route(struct rdma_ucm_query_route_resp *resp, in ucma_copy_iw_route() argument
812 rdma_addr_get_dgid(dev_addr, (union ib_gid *) &resp->ib_route[0].dgid); in ucma_copy_iw_route()
813 rdma_addr_get_sgid(dev_addr, (union ib_gid *) &resp->ib_route[0].sgid); in ucma_copy_iw_route()
821 struct rdma_ucm_query_route_resp resp; in ucma_query_route() local
837 memset(&resp, 0, sizeof resp); in ucma_query_route()
839 memcpy(&resp.src_addr, addr, addr->sa_family == AF_INET ? in ucma_query_route()
843 memcpy(&resp.dst_addr, addr, addr->sa_family == AF_INET ? in ucma_query_route()
849 resp.node_guid = (__force __u64) ctx->cm_id->device->node_guid; in ucma_query_route()
850 resp.ibdev_index = ctx->cm_id->device->index; in ucma_query_route()
851 resp.port_num = ctx->cm_id->port_num; in ucma_query_route()
854 ucma_copy_ib_route(&resp, &ctx->cm_id->route); in ucma_query_route()
856 ucma_copy_iboe_route(&resp, &ctx->cm_id->route); in ucma_query_route()
858 ucma_copy_iw_route(&resp, &ctx->cm_id->route); in ucma_query_route()
862 if (copy_to_user(u64_to_user_ptr(cmd.response), &resp, in ucma_query_route()
863 min_t(size_t, out_len, sizeof(resp)))) in ucma_query_route()
871 struct rdma_ucm_query_addr_resp *resp) in ucma_query_device_addr() argument
876 resp->node_guid = (__force __u64) cm_id->device->node_guid; in ucma_query_device_addr()
877 resp->ibdev_index = cm_id->device->index; in ucma_query_device_addr()
878 resp->port_num = cm_id->port_num; in ucma_query_device_addr()
879 resp->pkey = (__force __u16) cpu_to_be16( in ucma_query_device_addr()
886 struct rdma_ucm_query_addr_resp resp; in ucma_query_addr() local
893 memset(&resp, 0, sizeof resp); in ucma_query_addr()
896 resp.src_size = rdma_addr_size(addr); in ucma_query_addr()
897 memcpy(&resp.src_addr, addr, resp.src_size); in ucma_query_addr()
900 resp.dst_size = rdma_addr_size(addr); in ucma_query_addr()
901 memcpy(&resp.dst_addr, addr, resp.dst_size); in ucma_query_addr()
903 ucma_query_device_addr(ctx->cm_id, &resp); in ucma_query_addr()
905 if (copy_to_user(response, &resp, min_t(size_t, out_len, sizeof(resp)))) in ucma_query_addr()
914 struct rdma_ucm_query_path_resp *resp; in ucma_query_path() local
917 if (out_len < sizeof(*resp)) in ucma_query_path()
920 resp = kzalloc(out_len, GFP_KERNEL); in ucma_query_path()
921 if (!resp) in ucma_query_path()
924 resp->num_paths = ctx->cm_id->route.num_pri_alt_paths; in ucma_query_path()
925 for (i = 0, out_len -= sizeof(*resp); in ucma_query_path()
926 i < resp->num_paths && out_len > sizeof(struct ib_path_rec_data); in ucma_query_path()
930 resp->path_data[i].flags = IB_PATH_GMP | IB_PATH_PRIMARY | in ucma_query_path()
936 ib_sa_pack_path(&ib, &resp->path_data[i].path_rec); in ucma_query_path()
939 ib_sa_pack_path(rec, &resp->path_data[i].path_rec); in ucma_query_path()
943 if (copy_to_user(response, resp, struct_size(resp, path_data, i))) in ucma_query_path()
946 kfree(resp); in ucma_query_path()
953 struct rdma_ucm_query_addr_resp resp; in ucma_query_gid() local
960 memset(&resp, 0, sizeof resp); in ucma_query_gid()
962 ucma_query_device_addr(ctx->cm_id, &resp); in ucma_query_gid()
964 addr = (struct sockaddr_ib *) &resp.src_addr; in ucma_query_gid()
965 resp.src_size = sizeof(*addr); in ucma_query_gid()
967 memcpy(addr, &ctx->cm_id->route.addr.src_addr, resp.src_size); in ucma_query_gid()
970 addr->sib_pkey = (__force __be16) resp.pkey; in ucma_query_gid()
977 addr = (struct sockaddr_ib *) &resp.dst_addr; in ucma_query_gid()
978 resp.dst_size = sizeof(*addr); in ucma_query_gid()
980 memcpy(addr, &ctx->cm_id->route.addr.dst_addr, resp.dst_size); in ucma_query_gid()
983 addr->sib_pkey = (__force __be16) resp.pkey; in ucma_query_gid()
990 if (copy_to_user(response, &resp, min_t(size_t, out_len, sizeof(resp)))) in ucma_query_gid()
1217 struct ib_uverbs_qp_attr resp; in ucma_init_qp_attr() local
1222 if (out_len < sizeof(resp)) in ucma_init_qp_attr()
1235 resp.qp_attr_mask = 0; in ucma_init_qp_attr()
1239 ret = rdma_init_qp_attr(ctx->cm_id, &qp_attr, &resp.qp_attr_mask); in ucma_init_qp_attr()
1244 ib_copy_qp_attr_to_user(ctx->cm_id->device, &resp, &qp_attr); in ucma_init_qp_attr()
1246 &resp, sizeof(resp))) in ucma_init_qp_attr()
1439 struct rdma_ucm_create_id_resp resp; in ucma_process_join() local
1446 if (out_len < sizeof(resp)) in ucma_process_join()
1492 resp.id = mc->id; in ucma_process_join()
1494 &resp, sizeof(resp))) { in ucma_process_join()
1564 struct rdma_ucm_destroy_id_resp resp; in ucma_leave_multicast() local
1568 if (out_len < sizeof(resp)) in ucma_leave_multicast()
1600 resp.events_reported = mc->events_reported; in ucma_leave_multicast()
1604 &resp, sizeof(resp))) in ucma_leave_multicast()
1615 struct rdma_ucm_migrate_resp resp; in ucma_migrate_id() local
1667 resp.events_reported = ctx->events_reported; in ucma_migrate_id()
1676 &resp, sizeof(resp))) in ucma_migrate_id()