/net/sunrpc/ |
D | xprt.c | 65 static void xprt_init(struct rpc_xprt *xprt, struct net *net); 69 static void xprt_destroy(struct rpc_xprt *xprt); 88 #define RPC_MAXCWND(xprt) ((xprt)->max_reqs << RPC_CWNDSHIFT) argument 90 #define RPCXPRT_CONGESTED(xprt) ((xprt)->cong >= (xprt)->cwnd) argument 197 int xprt_reserve_xprt(struct rpc_xprt *xprt, struct rpc_task *task) in xprt_reserve_xprt() argument 202 if (test_and_set_bit(XPRT_LOCKED, &xprt->state)) { in xprt_reserve_xprt() 203 if (task == xprt->snd_task) in xprt_reserve_xprt() 207 xprt->snd_task = task; in xprt_reserve_xprt() 217 task->tk_pid, xprt); in xprt_reserve_xprt() 226 rpc_sleep_on_priority(&xprt->sending, task, NULL, priority); in xprt_reserve_xprt() [all …]
|
D | xprtsock.c | 52 static void xs_close(struct rpc_xprt *xprt); 218 struct rpc_xprt xprt; member 275 static inline struct sockaddr *xs_addr(struct rpc_xprt *xprt) in xs_addr() argument 277 return (struct sockaddr *) &xprt->addr; in xs_addr() 280 static inline struct sockaddr_un *xs_addr_un(struct rpc_xprt *xprt) in xs_addr_un() argument 282 return (struct sockaddr_un *) &xprt->addr; in xs_addr_un() 285 static inline struct sockaddr_in *xs_addr_in(struct rpc_xprt *xprt) in xs_addr_in() argument 287 return (struct sockaddr_in *) &xprt->addr; in xs_addr_in() 290 static inline struct sockaddr_in6 *xs_addr_in6(struct rpc_xprt *xprt) in xs_addr_in6() argument 292 return (struct sockaddr_in6 *) &xprt->addr; in xs_addr_in6() [all …]
|
D | svc_xprt.c | 21 static struct svc_deferred_req *svc_deferred_dequeue(struct svc_xprt *xprt); 25 static void svc_delete_xprt(struct svc_xprt *xprt); 128 struct svc_xprt *xprt = in svc_xprt_free() local 130 struct module *owner = xprt->xpt_class->xcl_owner; in svc_xprt_free() 131 if (test_bit(XPT_CACHE_AUTH, &xprt->xpt_flags)) in svc_xprt_free() 132 svcauth_unix_info_release(xprt); in svc_xprt_free() 133 put_net(xprt->xpt_net); in svc_xprt_free() 135 if (xprt->xpt_bc_xprt) in svc_xprt_free() 136 xprt_put(xprt->xpt_bc_xprt); in svc_xprt_free() 137 xprt->xpt_ops->xpo_free(xprt); in svc_xprt_free() [all …]
|
D | backchannel_rqst.c | 38 static inline int xprt_need_to_requeue(struct rpc_xprt *xprt) in xprt_need_to_requeue() argument 40 return xprt->bc_alloc_count > 0; in xprt_need_to_requeue() 43 static inline void xprt_inc_alloc_count(struct rpc_xprt *xprt, unsigned int n) in xprt_inc_alloc_count() argument 45 xprt->bc_alloc_count += n; in xprt_inc_alloc_count() 48 static inline int xprt_dec_alloc_count(struct rpc_xprt *xprt, unsigned int n) in xprt_dec_alloc_count() argument 50 return xprt->bc_alloc_count -= n; in xprt_dec_alloc_count() 89 int xprt_setup_backchannel(struct rpc_xprt *xprt, unsigned int min_reqs) in xprt_setup_backchannel() argument 120 req->rq_xprt = xprt; in xprt_setup_backchannel() 159 spin_lock_bh(&xprt->bc_pa_lock); in xprt_setup_backchannel() 160 list_splice(&tmp_list, &xprt->bc_pa_list); in xprt_setup_backchannel() [all …]
|
D | rpcb_clnt.c | 148 static void rpcb_wake_rpcbind_waiters(struct rpc_xprt *xprt, int status) in rpcb_wake_rpcbind_waiters() argument 150 xprt_clear_binding(xprt); in rpcb_wake_rpcbind_waiters() 151 rpc_wake_up_status(&xprt->binding, status); in rpcb_wake_rpcbind_waiters() 627 struct rpc_xprt *xprt = rcu_dereference(clnt->cl_xprt); in rpcb_find_transport_owner() local 630 if (rcu_dereference(parent->cl_xprt) != xprt) in rpcb_find_transport_owner() 652 struct rpc_xprt *xprt; in rpcb_getport_async() local 664 xprt = xprt_get(rcu_dereference(clnt->cl_xprt)); in rpcb_getport_async() 665 } while (xprt == NULL); in rpcb_getport_async() 670 xprt->servername, clnt->cl_prog, clnt->cl_vers, xprt->prot); in rpcb_getport_async() 674 rpc_sleep_on(&xprt->binding, task, NULL); in rpcb_getport_async() [all …]
|
D | clnt.c | 292 static struct rpc_clnt * rpc_new_client(const struct rpc_create_args *args, struct rpc_xprt *xprt) in rpc_new_client() argument 302 program->name, args->servername, xprt); in rpc_new_client() 321 rcu_assign_pointer(clnt->cl_xprt, xprt); in rpc_new_client() 336 if (!xprt_bound(xprt)) in rpc_new_client() 339 clnt->cl_timeout = xprt->timeout; in rpc_new_client() 385 xprt_put(xprt); in rpc_new_client() 401 struct rpc_xprt *xprt; in rpc_create() local 452 xprt = xprt_create_transport(&xprtargs); in rpc_create() 453 if (IS_ERR(xprt)) in rpc_create() 454 return (struct rpc_clnt *)xprt; in rpc_create() [all …]
|
D | svcauth_unix.c | 341 ip_map_cached_get(struct svc_xprt *xprt) in ip_map_cached_get() argument 346 if (test_bit(XPT_CACHE_AUTH, &xprt->xpt_flags)) { in ip_map_cached_get() 347 spin_lock(&xprt->xpt_lock); in ip_map_cached_get() 348 ipm = xprt->xpt_auth_cache; in ip_map_cached_get() 356 sn = net_generic(xprt->xpt_net, sunrpc_net_id); in ip_map_cached_get() 357 xprt->xpt_auth_cache = NULL; in ip_map_cached_get() 358 spin_unlock(&xprt->xpt_lock); in ip_map_cached_get() 364 spin_unlock(&xprt->xpt_lock); in ip_map_cached_get() 370 ip_map_cached_put(struct svc_xprt *xprt, struct ip_map *ipm) in ip_map_cached_put() argument 372 if (test_bit(XPT_CACHE_AUTH, &xprt->xpt_flags)) { in ip_map_cached_put() [all …]
|
D | svcsock.c | 77 static void svc_bc_sock_free(struct svc_xprt *xprt); 637 static int svc_udp_has_wspace(struct svc_xprt *xprt) in svc_udp_has_wspace() argument 639 struct svc_sock *svsk = container_of(xprt, struct svc_sock, sk_xprt); in svc_udp_has_wspace() 640 struct svc_serv *serv = xprt->xpt_server; in svc_udp_has_wspace() 655 static struct svc_xprt *svc_udp_accept(struct svc_xprt *xprt) in svc_udp_accept() argument 802 static struct svc_xprt *svc_tcp_accept(struct svc_xprt *xprt) in svc_tcp_accept() argument 804 struct svc_sock *svsk = container_of(xprt, struct svc_sock, sk_xprt); in svc_tcp_accept() 1182 static int svc_tcp_has_wspace(struct svc_xprt *xprt) in svc_tcp_has_wspace() argument 1184 struct svc_sock *svsk = container_of(xprt, struct svc_sock, sk_xprt); in svc_tcp_has_wspace() 1188 if (test_bit(XPT_LISTENER, &xprt->xpt_flags)) in svc_tcp_has_wspace() [all …]
|
D | stats.c | 183 struct rpc_xprt *xprt; in rpc_print_iostats() local 194 xprt = rcu_dereference(clnt->cl_xprt); in rpc_print_iostats() 195 if (xprt) in rpc_print_iostats() 196 xprt->ops->print_stats(xprt, seq); in rpc_print_iostats()
|
D | Makefile | 10 sunrpc-y := clnt.o xprt.o socklib.o xprtsock.o sched.o \
|
/net/sunrpc/xprtrdma/ |
D | transport.c | 155 xprt_rdma_format_addresses(struct rpc_xprt *xprt) in xprt_rdma_format_addresses() argument 158 &rpcx_to_rdmad(xprt).addr; in xprt_rdma_format_addresses() 163 xprt->address_strings[RPC_DISPLAY_ADDR] = kstrdup(buf, GFP_KERNEL); in xprt_rdma_format_addresses() 166 xprt->address_strings[RPC_DISPLAY_PORT] = kstrdup(buf, GFP_KERNEL); in xprt_rdma_format_addresses() 168 xprt->address_strings[RPC_DISPLAY_PROTO] = "rdma"; in xprt_rdma_format_addresses() 171 xprt->address_strings[RPC_DISPLAY_HEX_ADDR] = kstrdup(buf, GFP_KERNEL); in xprt_rdma_format_addresses() 174 xprt->address_strings[RPC_DISPLAY_HEX_PORT] = kstrdup(buf, GFP_KERNEL); in xprt_rdma_format_addresses() 177 xprt->address_strings[RPC_DISPLAY_NETID] = "rdma"; in xprt_rdma_format_addresses() 181 xprt_rdma_free_addresses(struct rpc_xprt *xprt) in xprt_rdma_free_addresses() argument 191 kfree(xprt->address_strings[i]); in xprt_rdma_free_addresses() [all …]
|
D | svc_rdma_transport.c | 62 static struct svc_xprt *svc_rdma_accept(struct svc_xprt *xprt); 65 static void svc_rdma_detach(struct svc_xprt *xprt); 66 static void svc_rdma_free(struct svc_xprt *xprt); 67 static int svc_rdma_has_wspace(struct svc_xprt *xprt); 68 static void rq_cq_reap(struct svcxprt_rdma *xprt); 69 static void sq_cq_reap(struct svcxprt_rdma *xprt); 94 struct svc_rdma_op_ctxt *svc_rdma_get_context(struct svcxprt_rdma *xprt) in svc_rdma_get_context() argument 104 ctxt->xprt = xprt; in svc_rdma_get_context() 108 atomic_inc(&xprt->sc_ctxt_used); in svc_rdma_get_context() 114 struct svcxprt_rdma *xprt = ctxt->xprt; in svc_rdma_unmap_dma() local [all …]
|
D | svc_rdma_sendto.c | 72 static int fast_reg_xdr(struct svcxprt_rdma *xprt, in fast_reg_xdr() argument 84 frmr = svc_rdma_get_frmr(xprt); in fast_reg_xdr() 107 ib_dma_map_page(xprt->sc_cm_id->device, in fast_reg_xdr() 112 if (ib_dma_mapping_error(xprt->sc_cm_id->device, in fast_reg_xdr() 115 atomic_inc(&xprt->sc_dma_used); in fast_reg_xdr() 135 ib_dma_map_page(xprt->sc_cm_id->device, in fast_reg_xdr() 138 if (ib_dma_mapping_error(xprt->sc_cm_id->device, in fast_reg_xdr() 142 atomic_inc(&xprt->sc_dma_used); in fast_reg_xdr() 173 ib_dma_map_page(xprt->sc_cm_id->device, virt_to_page(va), in fast_reg_xdr() 177 if (ib_dma_mapping_error(xprt->sc_cm_id->device, in fast_reg_xdr() [all …]
|
D | svc_rdma_recvfrom.c | 128 static int map_read_chunks(struct svcxprt_rdma *xprt, in map_read_chunks() argument 228 static int fast_reg_read_chunks(struct svcxprt_rdma *xprt, in fast_reg_read_chunks() argument 244 frmr = svc_rdma_get_frmr(xprt); in fast_reg_read_chunks() 266 ib_dma_map_page(xprt->sc_cm_id->device, in fast_reg_read_chunks() 269 if (ib_dma_mapping_error(xprt->sc_cm_id->device, in fast_reg_read_chunks() 272 atomic_inc(&xprt->sc_dma_used); in fast_reg_read_chunks() 293 ret = svc_rdma_fastreg(xprt, frmr); in fast_reg_read_chunks() 300 printk("svcrdma: error fast registering xdr for xprt %p", xprt); in fast_reg_read_chunks() 301 svc_rdma_put_frmr(xprt, frmr); in fast_reg_read_chunks() 305 static int rdma_set_ctxt_sge(struct svcxprt_rdma *xprt, in rdma_set_ctxt_sge() argument [all …]
|
D | rpc_rdma.c | 369 struct rpc_xprt *xprt = rqst->rq_xprt; in rpcrdma_marshal_req() local 370 struct rpcrdma_xprt *r_xprt = rpcx_to_rdmax(xprt); in rpcrdma_marshal_req() 694 struct rpc_xprt *xprt = ep->rep_xprt; in rpcrdma_conn_func() local 696 spin_lock_bh(&xprt->transport_lock); in rpcrdma_conn_func() 697 if (++xprt->connect_cookie == 0) /* maintain a reserved value */ in rpcrdma_conn_func() 698 ++xprt->connect_cookie; in rpcrdma_conn_func() 700 if (!xprt_test_and_set_connected(xprt)) in rpcrdma_conn_func() 701 xprt_wake_pending_tasks(xprt, 0); in rpcrdma_conn_func() 703 if (xprt_test_and_clear_connected(xprt)) in rpcrdma_conn_func() 704 xprt_wake_pending_tasks(xprt, -ENOTCONN); in rpcrdma_conn_func() [all …]
|
D | svc_rdma_marshal.c | 278 int svc_rdma_xdr_encode_error(struct svcxprt_rdma *xprt, in svc_rdma_xdr_encode_error() argument 286 *va++ = htonl(xprt->sc_max_requests); in svc_rdma_xdr_encode_error() 364 void svc_rdma_xdr_encode_reply_header(struct svcxprt_rdma *xprt, in svc_rdma_xdr_encode_reply_header() argument 371 rdma_resp->rm_credit = htonl(xprt->sc_max_requests); in svc_rdma_xdr_encode_reply_header()
|
D | verbs.c | 287 struct rpcrdma_xprt *xprt = id->context; in rpcrdma_conn_upcall() local 288 struct rpcrdma_ia *ia = &xprt->rx_ia; in rpcrdma_conn_upcall() 289 struct rpcrdma_ep *ep = &xprt->rx_ep; in rpcrdma_conn_upcall() 369 xprt->rx_buf.rb_max_requests, in rpcrdma_conn_upcall() 383 rpcrdma_create_id(struct rpcrdma_xprt *xprt, in rpcrdma_create_id() argument 391 id = rdma_create_id(rpcrdma_conn_upcall, xprt, RDMA_PS_TCP, IB_QPT_RC); in rpcrdma_create_id() 459 rpcrdma_ia_open(struct rpcrdma_xprt *xprt, struct sockaddr *addr, int memreg) in rpcrdma_ia_open() argument 463 struct rpcrdma_ia *ia = &xprt->rx_ia; in rpcrdma_ia_open() 465 ia->ri_id = rpcrdma_create_id(xprt, ia, addr); in rpcrdma_ia_open() 832 struct rpcrdma_xprt *xprt; in rpcrdma_ep_connect() local [all …]
|
D | xprt_rdma.h | 275 struct rpc_xprt xprt; member 284 #define rpcx_to_rdmax(x) container_of(x, struct rpcrdma_xprt, xprt)
|
/net/sched/ |
D | cls_rsvp.h | 143 u8 *xprt; in rsvp_classify() local 164 xprt = ((u8 *)nhptr) + sizeof(struct ipv6hdr); in rsvp_classify() 169 xprt = ((u8 *)nhptr) + (nhptr->ihl<<2); in rsvp_classify() 181 (*(u32 *)(xprt + s->dpi.offset) ^ s->dpi.key)) && in rsvp_classify() 191 !(f->spi.mask & (*(u32 *)(xprt + f->spi.offset) ^ f->spi.key)) in rsvp_classify() 207 nhptr = (void *)(xprt + f->tunnelhdr - sizeof(*nhptr)); in rsvp_classify()
|