/net/sunrpc/ |
D | xdr.c | 136 struct kvec *tail = xdr->tail; in xdr_inline_pages() local 146 tail->iov_base = buf + offset; in xdr_inline_pages() 147 tail->iov_len = buflen - offset; in xdr_inline_pages() 315 struct kvec *head, *tail; in xdr_shrink_bufhead() local 319 tail = buf->tail; in xdr_shrink_bufhead() 327 if (tail->iov_len != 0) { in xdr_shrink_bufhead() 328 if (tail->iov_len > len) { in xdr_shrink_bufhead() 329 copy = tail->iov_len - len; in xdr_shrink_bufhead() 330 memmove((char *)tail->iov_base + len, in xdr_shrink_bufhead() 331 tail->iov_base, copy); in xdr_shrink_bufhead() [all …]
|
D | backchannel_rqst.c | 133 xbufp->tail[0].iov_base = NULL; in xprt_setup_backchannel() 134 xbufp->tail[0].iov_len = 0; in xprt_setup_backchannel() 149 xbufp->tail[0].iov_base = NULL; in xprt_setup_backchannel() 150 xbufp->tail[0].iov_len = 0; in xprt_setup_backchannel()
|
D | socklib.c | 137 len = xdr->tail[0].iov_len; in xdr_partial_copy_from_skb() 139 copied += copy_actor(desc, (char *)xdr->tail[0].iov_base + base, len - base); in xdr_partial_copy_from_skb()
|
/net/core/ |
D | gen_stats.c | 67 d->tail = (struct nlattr *)skb_tail_pointer(skb); in gnet_stats_start_copy_compat() 72 if (d->tail) in gnet_stats_start_copy_compat() 119 if (d->tail) { in gnet_stats_copy_basic() 156 if (d->tail) in gnet_stats_copy_rate_est() 184 if (d->tail) in gnet_stats_copy_queue() 212 if (d->tail) in gnet_stats_copy_app() 234 if (d->tail) in gnet_stats_finish_copy() 235 d->tail->nla_len = skb_tail_pointer(d->skb) - (u8 *)d->tail; in gnet_stats_finish_copy()
|
/net/sched/ |
D | sch_choke.c | 73 unsigned int tail; member 89 return (q->tail - q->head) & q->tab_mask; in choke_len() 109 if (q->head == q->tail) in choke_zap_head_holes() 118 q->tail = (q->tail - 1) & q->tab_mask; in choke_zap_tail_holes() 119 if (q->head == q->tail) in choke_zap_tail_holes() 121 } while (q->tab[q->tail] == NULL); in choke_zap_tail_holes() 134 if (idx == q->tail) in choke_drop_by_idx() 255 if (q->head == q->tail) in choke_match_random() 327 q->tab[q->tail] = skb; in choke_enqueue() 328 q->tail = (q->tail + 1) & q->tab_mask; in choke_enqueue() [all …]
|
D | sch_sfq.c | 134 struct sfq_slot *tail; /* current slot in round */ member 342 x = q->tail->next; in sfq_drop() 344 q->tail->next = slot->next; in sfq_drop() 470 if (q->tail == NULL) { /* It is the first flow */ in sfq_enqueue() 473 slot->next = q->tail->next; in sfq_enqueue() 474 q->tail->next = x; in sfq_enqueue() 480 q->tail = slot; in sfq_enqueue() 509 if (q->tail == NULL) in sfq_dequeue() 513 a = q->tail->next; in sfq_dequeue() 516 q->tail = slot; in sfq_dequeue() [all …]
|
D | ematch.c | 436 u8 *tail; in tcf_em_tree_dump() local 451 tail = skb_tail_pointer(skb); in tcf_em_tree_dump() 453 struct nlattr *match_start = (struct nlattr *)tail; in tcf_em_tree_dump() 473 tail = skb_tail_pointer(skb); in tcf_em_tree_dump() 474 match_start->nla_len = tail - (u8 *)match_start; in tcf_em_tree_dump()
|
/net/sunrpc/xprtrdma/ |
D | svc_rdma_sendto.c | 151 if (0 == xdr->tail[0].iov_len) in fast_reg_xdr() 155 vec->sge[sge_no].iov_len = xdr->tail[0].iov_len; in fast_reg_xdr() 157 if (((unsigned long)xdr->tail[0].iov_base & PAGE_MASK) == in fast_reg_xdr() 163 vec->sge[sge_no].iov_base = xdr->tail[0].iov_base; in fast_reg_xdr() 168 page_off = (unsigned long)xdr->tail[0].iov_base & ~PAGE_MASK; in fast_reg_xdr() 169 va = (void *)((unsigned long)xdr->tail[0].iov_base & PAGE_MASK); in fast_reg_xdr() 209 (xdr->head[0].iov_len + xdr->page_len + xdr->tail[0].iov_len)); in map_xdr() 239 if (xdr->tail[0].iov_len) { in map_xdr() 240 vec->sge[sge_no].iov_base = xdr->tail[0].iov_base; in map_xdr() 241 vec->sge[sge_no].iov_len = xdr->tail[0].iov_len; in map_xdr() [all …]
|
D | rpc_rdma.c | 119 if (xdrbuf->tail[0].iov_len) { in rpcrdma_convert_iovs() 122 if (xdrbuf->tail[0].iov_len < 4 && xprt_rdma_pad_optimize) in rpcrdma_convert_iovs() 128 seg[n].mr_offset = xdrbuf->tail[0].iov_base; in rpcrdma_convert_iovs() 129 seg[n].mr_len = xdrbuf->tail[0].iov_len; in rpcrdma_convert_iovs() 318 if (rqst->rq_snd_buf.tail[0].iov_len) { in rpcrdma_inline_pullup() 319 curlen = rqst->rq_snd_buf.tail[0].iov_len; in rpcrdma_inline_pullup() 320 if (destp + copy_len != rqst->rq_snd_buf.tail[0].iov_base) { in rpcrdma_inline_pullup() 322 rqst->rq_snd_buf.tail[0].iov_base, curlen); in rpcrdma_inline_pullup() 656 if (copy_len && rqst->rq_rcv_buf.tail[0].iov_len) { in rpcrdma_inline_fixup() 658 if (curlen > rqst->rq_rcv_buf.tail[0].iov_len) in rpcrdma_inline_fixup() [all …]
|
D | svc_rdma_recvfrom.c | 111 rqstp->rq_arg.tail[0].iov_base = NULL; in rdma_build_arg_xdr() 112 rqstp->rq_arg.tail[0].iov_len = 0; in rdma_build_arg_xdr() 152 head->arg.tail[0] = rqstp->rq_arg.tail[0]; in map_read_chunks() 250 head->arg.tail[0] = rqstp->rq_arg.tail[0]; in fast_reg_read_chunks() 557 rqstp->rq_arg.tail[0] = head->arg.tail[0]; in rdma_read_complete() 570 + rqstp->rq_arg.tail[0].iov_len; in rdma_read_complete() 664 + rqstp->rq_arg.tail[0].iov_len; in svc_rdma_recvfrom()
|
/net/ipv4/ |
D | tcp_probe.c | 74 unsigned long head, tail; member 81 return (tcp_probe.head - tcp_probe.tail) & (bufsize - 1); in tcp_probe_used() 145 tcp_probe.head = tcp_probe.tail = 0; in tcpprobe_open() 155 = tcp_probe.log + tcp_probe.tail; in tcpprobe_sprint() 189 if (tcp_probe.head == tcp_probe.tail) { in tcpprobe_read() 198 tcp_probe.tail = (tcp_probe.tail + 1) & (bufsize - 1); in tcpprobe_read()
|
D | esp4.c | 128 u8 *tail; in esp_output() local 189 tail = skb_tail_pointer(trailer); in esp_output() 191 memset(tail, 0, tfclen); in esp_output() 192 tail += tfclen; in esp_output() 197 tail[i] = i + 1; in esp_output() 199 tail[plen - 2] = plen - 2; in esp_output() 200 tail[plen - 1] = *skb_mac_header(skb); in esp_output()
|
/net/mac802154/ |
D | wpan.c | 219 const u8 *tail = skb_tail_pointer(skb); in mac802154_header_parse() local 224 if (hdr + 3 > tail) in mac802154_header_parse() 241 if (hdr + 2 > tail) in mac802154_header_parse() 247 if (hdr + IEEE802154_ADDR_LEN > tail) in mac802154_header_parse() 254 if (hdr + 2 > tail) in mac802154_header_parse() 260 if (hdr + 2 > tail) in mac802154_header_parse() 275 if (hdr + 2 > tail) in mac802154_header_parse() 281 if (hdr + IEEE802154_ADDR_LEN > tail) in mac802154_header_parse() 289 if (hdr + 2 > tail) in mac802154_header_parse() 295 if (hdr + 2 > tail) in mac802154_header_parse()
|
/net/mac80211/ |
D | wpa.c | 36 int tail; in ieee80211_tx_h_michael_mic_add() local 63 tail = MICHAEL_MIC_LEN; in ieee80211_tx_h_michael_mic_add() 65 tail += TKIP_ICV_LEN; in ieee80211_tx_h_michael_mic_add() 67 if (WARN_ON(skb_tailroom(skb) < tail || in ieee80211_tx_h_michael_mic_add() 185 int len, tail; in tkip_encrypt_skb() local 199 tail = 0; in tkip_encrypt_skb() 201 tail = TKIP_ICV_LEN; in tkip_encrypt_skb() 203 if (WARN_ON(skb_tailroom(skb) < tail || in tkip_encrypt_skb() 405 int hdrlen, len, tail; in ccmp_encrypt_skb() local 425 tail = 0; in ccmp_encrypt_skb() [all …]
|
/net/bluetooth/cmtp/ |
D | core.c | 213 unsigned int size, tail; in cmtp_process_transmit() local 226 tail = session->mtu - nskb->len; in cmtp_process_transmit() 227 if (tail < 5) { in cmtp_process_transmit() 230 tail = session->mtu; in cmtp_process_transmit() 233 size = min_t(uint, ((tail < 258) ? (tail - 2) : (tail - 3)), skb->len); in cmtp_process_transmit()
|
/net/sunrpc/auth_gss/ |
D | gss_krb5_crypto.c | 678 if (buf->tail[0].iov_base != NULL) { in gss_krb5_aes_encrypt() 679 ecptr = buf->tail[0].iov_base + buf->tail[0].iov_len; in gss_krb5_aes_encrypt() 681 buf->tail[0].iov_base = buf->head[0].iov_base in gss_krb5_aes_encrypt() 683 buf->tail[0].iov_len = 0; in gss_krb5_aes_encrypt() 684 ecptr = buf->tail[0].iov_base; in gss_krb5_aes_encrypt() 688 buf->tail[0].iov_len += ec; in gss_krb5_aes_encrypt() 694 buf->tail[0].iov_len += GSS_KRB5_TOK_HDR_LEN; in gss_krb5_aes_encrypt() 699 hmac.data = buf->tail[0].iov_base + buf->tail[0].iov_len; in gss_krb5_aes_encrypt() 755 buf->tail[0].iov_len += kctx->gk5e->cksumlength; in gss_krb5_aes_encrypt()
|
D | svcauth_gss.c | 873 return buf->head[0].iov_len + buf->page_len + buf->tail[0].iov_len; in total_buf_len() 1636 if (resbuf->tail[0].iov_base == NULL) { in svcauth_gss_wrap_resp_integ() 1639 resbuf->tail[0].iov_base = resbuf->head[0].iov_base in svcauth_gss_wrap_resp_integ() 1641 resbuf->tail[0].iov_len = 0; in svcauth_gss_wrap_resp_integ() 1642 resv = &resbuf->tail[0]; in svcauth_gss_wrap_resp_integ() 1644 resv = &resbuf->tail[0]; in svcauth_gss_wrap_resp_integ() 1690 if (resbuf->tail[0].iov_base) { in svcauth_gss_wrap_resp_priv() 1691 BUG_ON(resbuf->tail[0].iov_base >= resbuf->head[0].iov_base in svcauth_gss_wrap_resp_priv() 1693 BUG_ON(resbuf->tail[0].iov_base < resbuf->head[0].iov_base); in svcauth_gss_wrap_resp_priv() 1694 if (resbuf->tail[0].iov_len + resbuf->head[0].iov_len in svcauth_gss_wrap_resp_priv() [all …]
|
D | gss_krb5_wrap.c | 55 if (buf->page_len || buf->tail[0].iov_len) in gss_krb5_add_padding() 56 iov = &buf->tail[0]; in gss_krb5_add_padding() 91 BUG_ON(len > buf->tail[0].iov_len); in gss_krb5_remove_padding() 92 pad = *(u8 *)(buf->tail[0].iov_base + len - 1); in gss_krb5_remove_padding()
|
/net/vmw_vsock/ |
D | vmci_transport_notify.c | 193 u64 tail; in send_waiting_read() local 209 vmci_qpair_get_consume_indexes(vmci_trans(vsk)->qpair, &tail, &head); in send_waiting_read() 235 u64 tail; in send_waiting_write() local 245 vmci_qpair_get_produce_indexes(vmci_trans(vsk)->qpair, &tail, &head); in send_waiting_write() 246 room_left = vmci_trans(vsk)->produce_size - tail; in send_waiting_write() 252 waiting_info.offset = tail + room_needed + 1; in send_waiting_write()
|
/net/ipv6/ |
D | esp6.c | 165 u8 *tail; in esp6_output() local 216 tail = skb_tail_pointer(trailer); in esp6_output() 218 memset(tail, 0, tfclen); in esp6_output() 219 tail += tfclen; in esp6_output() 224 tail[i] = i + 1; in esp6_output() 226 tail[plen - 2] = plen - 2; in esp6_output() 227 tail[plen - 1] = *skb_mac_header(skb); in esp6_output()
|
/net/rxrpc/ |
D | ar-ack.c | 379 int tail = call->acks_tail, old_tail; in rxrpc_rotate_tx_window() local 380 int win = CIRC_CNT(call->acks_head, tail, call->acks_winsz); in rxrpc_rotate_tx_window() 388 _skb = call->acks_window[tail] & ~1; in rxrpc_rotate_tx_window() 390 old_tail = tail; in rxrpc_rotate_tx_window() 391 tail = (tail + 1) & (call->acks_winsz - 1); in rxrpc_rotate_tx_window() 392 call->acks_tail = tail; in rxrpc_rotate_tx_window() 394 call->acks_unacked = tail; in rxrpc_rotate_tx_window() 526 int tail; in rxrpc_zap_tx_window() local 532 tail = call->acks_tail; in rxrpc_zap_tx_window() 534 _skb = acks_window[tail] & ~1; in rxrpc_zap_tx_window()
|
/net/decnet/netfilter/ |
D | dn_rtmsg.c | 49 old_tail = skb->tail; in dnrmg_build_message() 60 nlh->nlmsg_len = skb->tail - old_tail; in dnrmg_build_message()
|
/net/dccp/ |
D | options.c | 441 const unsigned char *tail, *from; in dccp_insert_option_ackvec() local 467 tail = av->av_buf + DCCPAV_MAX_ACKVEC_LEN; in dccp_insert_option_ackvec() 485 if (from + copylen > tail) { in dccp_insert_option_ackvec() 486 const u16 tailsize = tail - from; in dccp_insert_option_ackvec()
|
/net/xfrm/ |
D | xfrm_algo.c | 807 void *pskb_put(struct sk_buff *skb, struct sk_buff *tail, int len) in pskb_put() argument 809 if (tail != skb) { in pskb_put() 813 return skb_put(tail, len); in pskb_put()
|
/net/caif/ |
D | cfpkt_skbuff.c | 328 if (dst->tail + neededtailspace > dst->end) { in cfpkt_append() 345 dst->tail += addlen; in cfpkt_append() 382 skb2->tail += len2nd; in cfpkt_split()
|