Home
last modified time | relevance | path

Searched refs:tail (Results 1 – 25 of 67) sorted by relevance

123

/net/sunrpc/
Dxdr.c136 struct kvec *tail = xdr->tail; in xdr_inline_pages() local
146 tail->iov_base = buf + offset; in xdr_inline_pages()
147 tail->iov_len = buflen - offset; in xdr_inline_pages()
315 struct kvec *head, *tail; in xdr_shrink_bufhead() local
319 tail = buf->tail; in xdr_shrink_bufhead()
327 if (tail->iov_len != 0) { in xdr_shrink_bufhead()
328 if (tail->iov_len > len) { in xdr_shrink_bufhead()
329 copy = tail->iov_len - len; in xdr_shrink_bufhead()
330 memmove((char *)tail->iov_base + len, in xdr_shrink_bufhead()
331 tail->iov_base, copy); in xdr_shrink_bufhead()
[all …]
Dbackchannel_rqst.c133 xbufp->tail[0].iov_base = NULL; in xprt_setup_backchannel()
134 xbufp->tail[0].iov_len = 0; in xprt_setup_backchannel()
149 xbufp->tail[0].iov_base = NULL; in xprt_setup_backchannel()
150 xbufp->tail[0].iov_len = 0; in xprt_setup_backchannel()
Dsocklib.c137 len = xdr->tail[0].iov_len; in xdr_partial_copy_from_skb()
139 copied += copy_actor(desc, (char *)xdr->tail[0].iov_base + base, len - base); in xdr_partial_copy_from_skb()
/net/core/
Dgen_stats.c67 d->tail = (struct nlattr *)skb_tail_pointer(skb); in gnet_stats_start_copy_compat()
72 if (d->tail) in gnet_stats_start_copy_compat()
119 if (d->tail) { in gnet_stats_copy_basic()
156 if (d->tail) in gnet_stats_copy_rate_est()
184 if (d->tail) in gnet_stats_copy_queue()
212 if (d->tail) in gnet_stats_copy_app()
234 if (d->tail) in gnet_stats_finish_copy()
235 d->tail->nla_len = skb_tail_pointer(d->skb) - (u8 *)d->tail; in gnet_stats_finish_copy()
/net/sched/
Dsch_choke.c73 unsigned int tail; member
89 return (q->tail - q->head) & q->tab_mask; in choke_len()
109 if (q->head == q->tail) in choke_zap_head_holes()
118 q->tail = (q->tail - 1) & q->tab_mask; in choke_zap_tail_holes()
119 if (q->head == q->tail) in choke_zap_tail_holes()
121 } while (q->tab[q->tail] == NULL); in choke_zap_tail_holes()
134 if (idx == q->tail) in choke_drop_by_idx()
255 if (q->head == q->tail) in choke_match_random()
327 q->tab[q->tail] = skb; in choke_enqueue()
328 q->tail = (q->tail + 1) & q->tab_mask; in choke_enqueue()
[all …]
Dsch_sfq.c134 struct sfq_slot *tail; /* current slot in round */ member
342 x = q->tail->next; in sfq_drop()
344 q->tail->next = slot->next; in sfq_drop()
470 if (q->tail == NULL) { /* It is the first flow */ in sfq_enqueue()
473 slot->next = q->tail->next; in sfq_enqueue()
474 q->tail->next = x; in sfq_enqueue()
480 q->tail = slot; in sfq_enqueue()
509 if (q->tail == NULL) in sfq_dequeue()
513 a = q->tail->next; in sfq_dequeue()
516 q->tail = slot; in sfq_dequeue()
[all …]
Dematch.c436 u8 *tail; in tcf_em_tree_dump() local
451 tail = skb_tail_pointer(skb); in tcf_em_tree_dump()
453 struct nlattr *match_start = (struct nlattr *)tail; in tcf_em_tree_dump()
473 tail = skb_tail_pointer(skb); in tcf_em_tree_dump()
474 match_start->nla_len = tail - (u8 *)match_start; in tcf_em_tree_dump()
/net/sunrpc/xprtrdma/
Dsvc_rdma_sendto.c151 if (0 == xdr->tail[0].iov_len) in fast_reg_xdr()
155 vec->sge[sge_no].iov_len = xdr->tail[0].iov_len; in fast_reg_xdr()
157 if (((unsigned long)xdr->tail[0].iov_base & PAGE_MASK) == in fast_reg_xdr()
163 vec->sge[sge_no].iov_base = xdr->tail[0].iov_base; in fast_reg_xdr()
168 page_off = (unsigned long)xdr->tail[0].iov_base & ~PAGE_MASK; in fast_reg_xdr()
169 va = (void *)((unsigned long)xdr->tail[0].iov_base & PAGE_MASK); in fast_reg_xdr()
209 (xdr->head[0].iov_len + xdr->page_len + xdr->tail[0].iov_len)); in map_xdr()
239 if (xdr->tail[0].iov_len) { in map_xdr()
240 vec->sge[sge_no].iov_base = xdr->tail[0].iov_base; in map_xdr()
241 vec->sge[sge_no].iov_len = xdr->tail[0].iov_len; in map_xdr()
[all …]
Drpc_rdma.c119 if (xdrbuf->tail[0].iov_len) { in rpcrdma_convert_iovs()
122 if (xdrbuf->tail[0].iov_len < 4 && xprt_rdma_pad_optimize) in rpcrdma_convert_iovs()
128 seg[n].mr_offset = xdrbuf->tail[0].iov_base; in rpcrdma_convert_iovs()
129 seg[n].mr_len = xdrbuf->tail[0].iov_len; in rpcrdma_convert_iovs()
318 if (rqst->rq_snd_buf.tail[0].iov_len) { in rpcrdma_inline_pullup()
319 curlen = rqst->rq_snd_buf.tail[0].iov_len; in rpcrdma_inline_pullup()
320 if (destp + copy_len != rqst->rq_snd_buf.tail[0].iov_base) { in rpcrdma_inline_pullup()
322 rqst->rq_snd_buf.tail[0].iov_base, curlen); in rpcrdma_inline_pullup()
656 if (copy_len && rqst->rq_rcv_buf.tail[0].iov_len) { in rpcrdma_inline_fixup()
658 if (curlen > rqst->rq_rcv_buf.tail[0].iov_len) in rpcrdma_inline_fixup()
[all …]
Dsvc_rdma_recvfrom.c111 rqstp->rq_arg.tail[0].iov_base = NULL; in rdma_build_arg_xdr()
112 rqstp->rq_arg.tail[0].iov_len = 0; in rdma_build_arg_xdr()
152 head->arg.tail[0] = rqstp->rq_arg.tail[0]; in map_read_chunks()
250 head->arg.tail[0] = rqstp->rq_arg.tail[0]; in fast_reg_read_chunks()
557 rqstp->rq_arg.tail[0] = head->arg.tail[0]; in rdma_read_complete()
570 + rqstp->rq_arg.tail[0].iov_len; in rdma_read_complete()
664 + rqstp->rq_arg.tail[0].iov_len; in svc_rdma_recvfrom()
/net/ipv4/
Dtcp_probe.c74 unsigned long head, tail; member
81 return (tcp_probe.head - tcp_probe.tail) & (bufsize - 1); in tcp_probe_used()
145 tcp_probe.head = tcp_probe.tail = 0; in tcpprobe_open()
155 = tcp_probe.log + tcp_probe.tail; in tcpprobe_sprint()
189 if (tcp_probe.head == tcp_probe.tail) { in tcpprobe_read()
198 tcp_probe.tail = (tcp_probe.tail + 1) & (bufsize - 1); in tcpprobe_read()
Desp4.c128 u8 *tail; in esp_output() local
189 tail = skb_tail_pointer(trailer); in esp_output()
191 memset(tail, 0, tfclen); in esp_output()
192 tail += tfclen; in esp_output()
197 tail[i] = i + 1; in esp_output()
199 tail[plen - 2] = plen - 2; in esp_output()
200 tail[plen - 1] = *skb_mac_header(skb); in esp_output()
/net/mac802154/
Dwpan.c219 const u8 *tail = skb_tail_pointer(skb); in mac802154_header_parse() local
224 if (hdr + 3 > tail) in mac802154_header_parse()
241 if (hdr + 2 > tail) in mac802154_header_parse()
247 if (hdr + IEEE802154_ADDR_LEN > tail) in mac802154_header_parse()
254 if (hdr + 2 > tail) in mac802154_header_parse()
260 if (hdr + 2 > tail) in mac802154_header_parse()
275 if (hdr + 2 > tail) in mac802154_header_parse()
281 if (hdr + IEEE802154_ADDR_LEN > tail) in mac802154_header_parse()
289 if (hdr + 2 > tail) in mac802154_header_parse()
295 if (hdr + 2 > tail) in mac802154_header_parse()
/net/mac80211/
Dwpa.c36 int tail; in ieee80211_tx_h_michael_mic_add() local
63 tail = MICHAEL_MIC_LEN; in ieee80211_tx_h_michael_mic_add()
65 tail += TKIP_ICV_LEN; in ieee80211_tx_h_michael_mic_add()
67 if (WARN_ON(skb_tailroom(skb) < tail || in ieee80211_tx_h_michael_mic_add()
185 int len, tail; in tkip_encrypt_skb() local
199 tail = 0; in tkip_encrypt_skb()
201 tail = TKIP_ICV_LEN; in tkip_encrypt_skb()
203 if (WARN_ON(skb_tailroom(skb) < tail || in tkip_encrypt_skb()
405 int hdrlen, len, tail; in ccmp_encrypt_skb() local
425 tail = 0; in ccmp_encrypt_skb()
[all …]
/net/bluetooth/cmtp/
Dcore.c213 unsigned int size, tail; in cmtp_process_transmit() local
226 tail = session->mtu - nskb->len; in cmtp_process_transmit()
227 if (tail < 5) { in cmtp_process_transmit()
230 tail = session->mtu; in cmtp_process_transmit()
233 size = min_t(uint, ((tail < 258) ? (tail - 2) : (tail - 3)), skb->len); in cmtp_process_transmit()
/net/sunrpc/auth_gss/
Dgss_krb5_crypto.c678 if (buf->tail[0].iov_base != NULL) { in gss_krb5_aes_encrypt()
679 ecptr = buf->tail[0].iov_base + buf->tail[0].iov_len; in gss_krb5_aes_encrypt()
681 buf->tail[0].iov_base = buf->head[0].iov_base in gss_krb5_aes_encrypt()
683 buf->tail[0].iov_len = 0; in gss_krb5_aes_encrypt()
684 ecptr = buf->tail[0].iov_base; in gss_krb5_aes_encrypt()
688 buf->tail[0].iov_len += ec; in gss_krb5_aes_encrypt()
694 buf->tail[0].iov_len += GSS_KRB5_TOK_HDR_LEN; in gss_krb5_aes_encrypt()
699 hmac.data = buf->tail[0].iov_base + buf->tail[0].iov_len; in gss_krb5_aes_encrypt()
755 buf->tail[0].iov_len += kctx->gk5e->cksumlength; in gss_krb5_aes_encrypt()
Dsvcauth_gss.c873 return buf->head[0].iov_len + buf->page_len + buf->tail[0].iov_len; in total_buf_len()
1636 if (resbuf->tail[0].iov_base == NULL) { in svcauth_gss_wrap_resp_integ()
1639 resbuf->tail[0].iov_base = resbuf->head[0].iov_base in svcauth_gss_wrap_resp_integ()
1641 resbuf->tail[0].iov_len = 0; in svcauth_gss_wrap_resp_integ()
1642 resv = &resbuf->tail[0]; in svcauth_gss_wrap_resp_integ()
1644 resv = &resbuf->tail[0]; in svcauth_gss_wrap_resp_integ()
1690 if (resbuf->tail[0].iov_base) { in svcauth_gss_wrap_resp_priv()
1691 BUG_ON(resbuf->tail[0].iov_base >= resbuf->head[0].iov_base in svcauth_gss_wrap_resp_priv()
1693 BUG_ON(resbuf->tail[0].iov_base < resbuf->head[0].iov_base); in svcauth_gss_wrap_resp_priv()
1694 if (resbuf->tail[0].iov_len + resbuf->head[0].iov_len in svcauth_gss_wrap_resp_priv()
[all …]
Dgss_krb5_wrap.c55 if (buf->page_len || buf->tail[0].iov_len) in gss_krb5_add_padding()
56 iov = &buf->tail[0]; in gss_krb5_add_padding()
91 BUG_ON(len > buf->tail[0].iov_len); in gss_krb5_remove_padding()
92 pad = *(u8 *)(buf->tail[0].iov_base + len - 1); in gss_krb5_remove_padding()
/net/vmw_vsock/
Dvmci_transport_notify.c193 u64 tail; in send_waiting_read() local
209 vmci_qpair_get_consume_indexes(vmci_trans(vsk)->qpair, &tail, &head); in send_waiting_read()
235 u64 tail; in send_waiting_write() local
245 vmci_qpair_get_produce_indexes(vmci_trans(vsk)->qpair, &tail, &head); in send_waiting_write()
246 room_left = vmci_trans(vsk)->produce_size - tail; in send_waiting_write()
252 waiting_info.offset = tail + room_needed + 1; in send_waiting_write()
/net/ipv6/
Desp6.c165 u8 *tail; in esp6_output() local
216 tail = skb_tail_pointer(trailer); in esp6_output()
218 memset(tail, 0, tfclen); in esp6_output()
219 tail += tfclen; in esp6_output()
224 tail[i] = i + 1; in esp6_output()
226 tail[plen - 2] = plen - 2; in esp6_output()
227 tail[plen - 1] = *skb_mac_header(skb); in esp6_output()
/net/rxrpc/
Dar-ack.c379 int tail = call->acks_tail, old_tail; in rxrpc_rotate_tx_window() local
380 int win = CIRC_CNT(call->acks_head, tail, call->acks_winsz); in rxrpc_rotate_tx_window()
388 _skb = call->acks_window[tail] & ~1; in rxrpc_rotate_tx_window()
390 old_tail = tail; in rxrpc_rotate_tx_window()
391 tail = (tail + 1) & (call->acks_winsz - 1); in rxrpc_rotate_tx_window()
392 call->acks_tail = tail; in rxrpc_rotate_tx_window()
394 call->acks_unacked = tail; in rxrpc_rotate_tx_window()
526 int tail; in rxrpc_zap_tx_window() local
532 tail = call->acks_tail; in rxrpc_zap_tx_window()
534 _skb = acks_window[tail] & ~1; in rxrpc_zap_tx_window()
/net/decnet/netfilter/
Ddn_rtmsg.c49 old_tail = skb->tail; in dnrmg_build_message()
60 nlh->nlmsg_len = skb->tail - old_tail; in dnrmg_build_message()
/net/dccp/
Doptions.c441 const unsigned char *tail, *from; in dccp_insert_option_ackvec() local
467 tail = av->av_buf + DCCPAV_MAX_ACKVEC_LEN; in dccp_insert_option_ackvec()
485 if (from + copylen > tail) { in dccp_insert_option_ackvec()
486 const u16 tailsize = tail - from; in dccp_insert_option_ackvec()
/net/xfrm/
Dxfrm_algo.c807 void *pskb_put(struct sk_buff *skb, struct sk_buff *tail, int len) in pskb_put() argument
809 if (tail != skb) { in pskb_put()
813 return skb_put(tail, len); in pskb_put()
/net/caif/
Dcfpkt_skbuff.c328 if (dst->tail + neededtailspace > dst->end) { in cfpkt_append()
345 dst->tail += addlen; in cfpkt_append()
382 skb2->tail += len2nd; in cfpkt_split()

123