/net/sunrpc/ |
D | xdr.c | 136 struct kvec *tail = xdr->tail; in xdr_inline_pages() local 146 tail->iov_base = buf + offset; in xdr_inline_pages() 147 tail->iov_len = buflen - offset; in xdr_inline_pages() 318 struct kvec *head, *tail; in xdr_shrink_bufhead() local 322 tail = buf->tail; in xdr_shrink_bufhead() 330 if (tail->iov_len != 0) { in xdr_shrink_bufhead() 331 if (tail->iov_len > len) { in xdr_shrink_bufhead() 332 copy = tail->iov_len - len; in xdr_shrink_bufhead() 333 memmove((char *)tail->iov_base + len, in xdr_shrink_bufhead() 334 tail->iov_base, copy); in xdr_shrink_bufhead() [all …]
|
D | socklib.c | 137 len = xdr->tail[0].iov_len; in xdr_partial_copy_from_skb() 139 copied += copy_actor(desc, (char *)xdr->tail[0].iov_base + base, len - base); in xdr_partial_copy_from_skb()
|
/net/sched/ |
D | sch_choke.c | 72 unsigned int tail; member 82 return (q->tail - q->head) & q->tab_mask; in choke_len() 102 if (q->head == q->tail) in choke_zap_head_holes() 111 q->tail = (q->tail - 1) & q->tab_mask; in choke_zap_tail_holes() 112 if (q->head == q->tail) in choke_zap_tail_holes() 114 } while (q->tab[q->tail] == NULL); in choke_zap_tail_holes() 127 if (idx == q->tail) in choke_drop_by_idx() 254 if (q->head == q->tail) in choke_match_random() 326 q->tab[q->tail] = skb; in choke_enqueue() 327 q->tail = (q->tail + 1) & q->tab_mask; in choke_enqueue() [all …]
|
D | sch_sfq.c | 133 struct sfq_slot *tail; /* current slot in round */ member 318 x = q->tail->next; in sfq_drop() 320 q->tail->next = slot->next; in sfq_drop() 447 if (q->tail == NULL) { /* It is the first flow */ in sfq_enqueue() 450 slot->next = q->tail->next; in sfq_enqueue() 451 q->tail->next = x; in sfq_enqueue() 457 q->tail = slot; in sfq_enqueue() 488 if (q->tail == NULL) in sfq_dequeue() 492 a = q->tail->next; in sfq_dequeue() 495 q->tail = slot; in sfq_dequeue() [all …]
|
D | ematch.c | 442 u8 *tail; in tcf_em_tree_dump() local 457 tail = skb_tail_pointer(skb); in tcf_em_tree_dump() 459 struct nlattr *match_start = (struct nlattr *)tail; in tcf_em_tree_dump() 479 tail = skb_tail_pointer(skb); in tcf_em_tree_dump() 480 match_start->nla_len = tail - (u8 *)match_start; in tcf_em_tree_dump()
|
D | sch_fq.c | 63 struct sk_buff *tail; /* last skb in the list */ member 355 flow->tail = skb; in flow_queue_add() 359 flow->tail->next = skb; in flow_queue_add() 360 flow->tail = skb; in flow_queue_add() 378 if (prev == flow->tail) in flow_queue_add() 379 flow->tail = skb; in flow_queue_add()
|
/net/core/ |
D | gen_stats.c | 70 d->tail = (struct nlattr *)skb_tail_pointer(skb); in gnet_stats_start_copy_compat() 75 if (d->tail) in gnet_stats_start_copy_compat() 165 if (d->tail) { in gnet_stats_copy_basic() 209 if (d->tail) { in gnet_stats_copy_rate_est() 286 if (d->tail) in gnet_stats_copy_queue() 317 if (d->tail) in gnet_stats_copy_app() 344 if (d->tail) in gnet_stats_finish_copy() 345 d->tail->nla_len = skb_tail_pointer(d->skb) - (u8 *)d->tail; in gnet_stats_finish_copy()
|
D | skbuff.c | 102 (unsigned long)skb->tail, (unsigned long)skb->end, in skb_panic() 175 memset(skb, 0, offsetof(struct sk_buff, tail)); in __alloc_skb_head() 245 memset(skb, 0, offsetof(struct sk_buff, tail)); in __alloc_skb() 253 skb->end = skb->tail + size; in __alloc_skb() 315 memset(skb, 0, offsetof(struct sk_buff, tail)); in __build_skb() 321 skb->end = skb->tail + size; in __build_skb() 846 C(tail); in __skb_clone() 1196 skb->tail += off; in pskb_expand_head() 1314 ntail = skb->data_len + pad - (skb->end - skb->tail); in skb_pad() 1350 unsigned char *pskb_put(struct sk_buff *skb, struct sk_buff *tail, int len) in pskb_put() argument [all …]
|
/net/sunrpc/xprtrdma/ |
D | rpc_rdma.c | 102 size_t tlen = buf->tail[0].iov_len; in rpcrdma_tail_pullup() 118 src = buf->tail[0].iov_base; in rpcrdma_tail_pullup() 190 if (xdrbuf->tail[0].iov_len) { in rpcrdma_convert_iovs() 193 if (xdrbuf->tail[0].iov_len < 4 && xprt_rdma_pad_optimize) in rpcrdma_convert_iovs() 199 seg[n].mr_offset = xdrbuf->tail[0].iov_base; in rpcrdma_convert_iovs() 200 seg[n].mr_len = xdrbuf->tail[0].iov_len; in rpcrdma_convert_iovs() 383 if (rqst->rq_snd_buf.tail[0].iov_len) { in rpcrdma_inline_pullup() 384 curlen = rqst->rq_snd_buf.tail[0].iov_len; in rpcrdma_inline_pullup() 385 if (destp + copy_len != rqst->rq_snd_buf.tail[0].iov_base) { in rpcrdma_inline_pullup() 387 rqst->rq_snd_buf.tail[0].iov_base, curlen); in rpcrdma_inline_pullup() [all …]
|
D | svc_rdma_sendto.c | 64 (xdr->head[0].iov_len + xdr->page_len + xdr->tail[0].iov_len)) { in map_xdr() 94 if (xdr->tail[0].iov_len) { in map_xdr() 95 vec->sge[sge_no].iov_base = xdr->tail[0].iov_base; in map_xdr() 96 vec->sge[sge_no].iov_len = xdr->tail[0].iov_len; in map_xdr() 103 xdr->head[0].iov_len, xdr->tail[0].iov_len); in map_xdr() 130 xdr->tail[0].iov_base & ~PAGE_MASK; in dma_map_xdr() 131 page = virt_to_page(xdr->tail[0].iov_base); in dma_map_xdr() 311 u32 xfer_len = rqstp->rq_res.page_len + rqstp->rq_res.tail[0].iov_len; in send_write_chunks() 367 return rqstp->rq_res.page_len + rqstp->rq_res.tail[0].iov_len; in send_write_chunks()
|
D | svc_rdma_recvfrom.c | 114 rqstp->rq_arg.tail[0].iov_base = NULL; in rdma_build_arg_xdr() 115 rqstp->rq_arg.tail[0].iov_len = 0; in rdma_build_arg_xdr() 439 head->arg.tail[0] = rqstp->rq_arg.tail[0]; in rdma_read_chunks() 546 rqstp->rq_arg.tail[0] = head->arg.tail[0]; in rdma_read_complete() 559 + rqstp->rq_arg.tail[0].iov_len; in rdma_read_complete() 646 + rqstp->rq_arg.tail[0].iov_len; in svc_rdma_recvfrom()
|
/net/ipv4/ |
D | tcp_cdg.c | 82 u8 tail; member 195 ca->gsum.min += gmin - ca->gradients[ca->tail].min; in tcp_cdg_grad() 196 ca->gsum.max += gmax - ca->gradients[ca->tail].max; in tcp_cdg_grad() 197 ca->gradients[ca->tail].min = gmin; in tcp_cdg_grad() 198 ca->gradients[ca->tail].max = gmax; in tcp_cdg_grad() 199 ca->tail = (ca->tail + 1) & (window - 1); in tcp_cdg_grad() 216 else if (ca->tail == 0) in tcp_cdg_grad() 219 grad = (grad * window) / (int)ca->tail; in tcp_cdg_grad()
|
D | tcp_probe.c | 82 unsigned long head, tail; member 88 return (tcp_probe.head - tcp_probe.tail) & (bufsize - 1); in tcp_probe_used() 179 tcp_probe.head = tcp_probe.tail = 0; in tcpprobe_open() 189 = tcp_probe.log + tcp_probe.tail; in tcpprobe_sprint() 221 if (tcp_probe.head == tcp_probe.tail) { in tcpprobe_read() 230 tcp_probe.tail = (tcp_probe.tail + 1) & (bufsize - 1); in tcpprobe_read()
|
D | esp4.c | 142 u8 *tail; in esp_output() local 198 tail = skb_tail_pointer(trailer); in esp_output() 200 memset(tail, 0, tfclen); in esp_output() 201 tail += tfclen; in esp_output() 206 tail[i] = i + 1; in esp_output() 208 tail[plen - 2] = plen - 2; in esp_output() 209 tail[plen - 1] = *skb_mac_header(skb); in esp_output()
|
/net/bluetooth/cmtp/ |
D | core.c | 214 unsigned int size, tail; in cmtp_process_transmit() local 227 tail = session->mtu - nskb->len; in cmtp_process_transmit() 228 if (tail < 5) { in cmtp_process_transmit() 231 tail = session->mtu; in cmtp_process_transmit() 234 size = min_t(uint, ((tail < 258) ? (tail - 2) : (tail - 3)), skb->len); in cmtp_process_transmit()
|
/net/mac80211/ |
D | wpa.c | 39 int tail; in ieee80211_tx_h_michael_mic_add() local 66 tail = MICHAEL_MIC_LEN; in ieee80211_tx_h_michael_mic_add() 68 tail += IEEE80211_TKIP_ICV_LEN; in ieee80211_tx_h_michael_mic_add() 70 if (WARN(skb_tailroom(skb) < tail || in ieee80211_tx_h_michael_mic_add() 74 skb_tailroom(skb), tail)) in ieee80211_tx_h_michael_mic_add() 194 int len, tail; in tkip_encrypt_skb() local 208 tail = 0; in tkip_encrypt_skb() 210 tail = IEEE80211_TKIP_ICV_LEN; in tkip_encrypt_skb() 212 if (WARN_ON(skb_tailroom(skb) < tail || in tkip_encrypt_skb() 409 int hdrlen, len, tail; in ccmp_encrypt_skb() local [all …]
|
/net/tipc/ |
D | msg.c | 124 struct sk_buff *tail = NULL; in tipc_buf_append() local 148 TIPC_SKB_CB(head)->tail = NULL; in tipc_buf_append() 158 tail = TIPC_SKB_CB(head)->tail; in tipc_buf_append() 162 tail->next = frag; in tipc_buf_append() 166 TIPC_SKB_CB(head)->tail = frag; in tipc_buf_append() 174 TIPC_SKB_CB(head)->tail = NULL; in tipc_buf_append()
|
/net/sunrpc/auth_gss/ |
D | svcauth_gss.c | 874 return buf->head[0].iov_len + buf->page_len + buf->tail[0].iov_len; in total_buf_len() 1607 if (resbuf->tail[0].iov_base == NULL) { in svcauth_gss_wrap_resp_integ() 1610 resbuf->tail[0].iov_base = resbuf->head[0].iov_base in svcauth_gss_wrap_resp_integ() 1612 resbuf->tail[0].iov_len = 0; in svcauth_gss_wrap_resp_integ() 1614 resv = &resbuf->tail[0]; in svcauth_gss_wrap_resp_integ() 1659 if (resbuf->tail[0].iov_base) { in svcauth_gss_wrap_resp_priv() 1660 BUG_ON(resbuf->tail[0].iov_base >= resbuf->head[0].iov_base in svcauth_gss_wrap_resp_priv() 1662 BUG_ON(resbuf->tail[0].iov_base < resbuf->head[0].iov_base); in svcauth_gss_wrap_resp_priv() 1663 if (resbuf->tail[0].iov_len + resbuf->head[0].iov_len in svcauth_gss_wrap_resp_priv() 1666 memmove(resbuf->tail[0].iov_base + RPC_MAX_AUTH_SIZE, in svcauth_gss_wrap_resp_priv() [all …]
|
D | gss_krb5_crypto.c | 678 if (buf->tail[0].iov_base != NULL) { in gss_krb5_aes_encrypt() 679 ecptr = buf->tail[0].iov_base + buf->tail[0].iov_len; in gss_krb5_aes_encrypt() 681 buf->tail[0].iov_base = buf->head[0].iov_base in gss_krb5_aes_encrypt() 683 buf->tail[0].iov_len = 0; in gss_krb5_aes_encrypt() 684 ecptr = buf->tail[0].iov_base; in gss_krb5_aes_encrypt() 689 buf->tail[0].iov_len += GSS_KRB5_TOK_HDR_LEN; in gss_krb5_aes_encrypt() 694 hmac.data = buf->tail[0].iov_base + buf->tail[0].iov_len; in gss_krb5_aes_encrypt() 750 buf->tail[0].iov_len += kctx->gk5e->cksumlength; in gss_krb5_aes_encrypt()
|
/net/vmw_vsock/ |
D | vmci_transport_notify.c | 193 u64 tail; in send_waiting_read() local 209 vmci_qpair_get_consume_indexes(vmci_trans(vsk)->qpair, &tail, &head); in send_waiting_read() 235 u64 tail; in send_waiting_write() local 245 vmci_qpair_get_produce_indexes(vmci_trans(vsk)->qpair, &tail, &head); in send_waiting_write() 246 room_left = vmci_trans(vsk)->produce_size - tail; in send_waiting_write() 252 waiting_info.offset = tail + room_needed + 1; in send_waiting_write()
|
/net/ipv6/ |
D | esp6.c | 168 u8 *tail; in esp6_output() local 215 tail = skb_tail_pointer(trailer); in esp6_output() 217 memset(tail, 0, tfclen); in esp6_output() 218 tail += tfclen; in esp6_output() 223 tail[i] = i + 1; in esp6_output() 225 tail[plen - 2] = plen - 2; in esp6_output() 226 tail[plen - 1] = *skb_mac_header(skb); in esp6_output()
|
/net/rxrpc/ |
D | ar-ack.c | 428 int tail = call->acks_tail, old_tail; in rxrpc_rotate_tx_window() local 429 int win = CIRC_CNT(call->acks_head, tail, call->acks_winsz); in rxrpc_rotate_tx_window() 437 _skb = call->acks_window[tail] & ~1; in rxrpc_rotate_tx_window() 439 old_tail = tail; in rxrpc_rotate_tx_window() 440 tail = (tail + 1) & (call->acks_winsz - 1); in rxrpc_rotate_tx_window() 441 call->acks_tail = tail; in rxrpc_rotate_tx_window() 443 call->acks_unacked = tail; in rxrpc_rotate_tx_window() 575 int tail; in rxrpc_zap_tx_window() local 581 tail = call->acks_tail; in rxrpc_zap_tx_window() 583 _skb = acks_window[tail] & ~1; in rxrpc_zap_tx_window()
|
/net/decnet/netfilter/ |
D | dn_rtmsg.c | 49 old_tail = skb->tail; in dnrmg_build_message() 60 nlh->nlmsg_len = skb->tail - old_tail; in dnrmg_build_message()
|
/net/dccp/ |
D | options.c | 409 const unsigned char *tail, *from; in dccp_insert_option_ackvec() local 435 tail = av->av_buf + DCCPAV_MAX_ACKVEC_LEN; in dccp_insert_option_ackvec() 453 if (from + copylen > tail) { in dccp_insert_option_ackvec() 454 const u16 tailsize = tail - from; in dccp_insert_option_ackvec()
|
/net/caif/ |
D | cfpkt_skbuff.c | 314 if (dst->tail + neededtailspace > dst->end) { in cfpkt_append() 331 dst->tail += addlen; in cfpkt_append() 368 skb2->tail += len2nd; in cfpkt_split()
|