Home
last modified time | relevance | path

Searched refs:tail (Results 1 – 25 of 58) sorted by relevance

123

/net/sunrpc/
Dxdr.c136 struct kvec *tail = xdr->tail; in xdr_inline_pages() local
146 tail->iov_base = buf + offset; in xdr_inline_pages()
147 tail->iov_len = buflen - offset; in xdr_inline_pages()
318 struct kvec *head, *tail; in xdr_shrink_bufhead() local
322 tail = buf->tail; in xdr_shrink_bufhead()
330 if (tail->iov_len != 0) { in xdr_shrink_bufhead()
331 if (tail->iov_len > len) { in xdr_shrink_bufhead()
332 copy = tail->iov_len - len; in xdr_shrink_bufhead()
333 memmove((char *)tail->iov_base + len, in xdr_shrink_bufhead()
334 tail->iov_base, copy); in xdr_shrink_bufhead()
[all …]
Dsocklib.c137 len = xdr->tail[0].iov_len; in xdr_partial_copy_from_skb()
139 copied += copy_actor(desc, (char *)xdr->tail[0].iov_base + base, len - base); in xdr_partial_copy_from_skb()
/net/sched/
Dsch_choke.c72 unsigned int tail; member
82 return (q->tail - q->head) & q->tab_mask; in choke_len()
102 if (q->head == q->tail) in choke_zap_head_holes()
111 q->tail = (q->tail - 1) & q->tab_mask; in choke_zap_tail_holes()
112 if (q->head == q->tail) in choke_zap_tail_holes()
114 } while (q->tab[q->tail] == NULL); in choke_zap_tail_holes()
127 if (idx == q->tail) in choke_drop_by_idx()
254 if (q->head == q->tail) in choke_match_random()
326 q->tab[q->tail] = skb; in choke_enqueue()
327 q->tail = (q->tail + 1) & q->tab_mask; in choke_enqueue()
[all …]
Dsch_sfq.c133 struct sfq_slot *tail; /* current slot in round */ member
318 x = q->tail->next; in sfq_drop()
320 q->tail->next = slot->next; in sfq_drop()
447 if (q->tail == NULL) { /* It is the first flow */ in sfq_enqueue()
450 slot->next = q->tail->next; in sfq_enqueue()
451 q->tail->next = x; in sfq_enqueue()
457 q->tail = slot; in sfq_enqueue()
488 if (q->tail == NULL) in sfq_dequeue()
492 a = q->tail->next; in sfq_dequeue()
495 q->tail = slot; in sfq_dequeue()
[all …]
Dematch.c442 u8 *tail; in tcf_em_tree_dump() local
457 tail = skb_tail_pointer(skb); in tcf_em_tree_dump()
459 struct nlattr *match_start = (struct nlattr *)tail; in tcf_em_tree_dump()
479 tail = skb_tail_pointer(skb); in tcf_em_tree_dump()
480 match_start->nla_len = tail - (u8 *)match_start; in tcf_em_tree_dump()
Dsch_fq.c63 struct sk_buff *tail; /* last skb in the list */ member
355 flow->tail = skb; in flow_queue_add()
359 flow->tail->next = skb; in flow_queue_add()
360 flow->tail = skb; in flow_queue_add()
378 if (prev == flow->tail) in flow_queue_add()
379 flow->tail = skb; in flow_queue_add()
/net/core/
Dgen_stats.c70 d->tail = (struct nlattr *)skb_tail_pointer(skb); in gnet_stats_start_copy_compat()
75 if (d->tail) in gnet_stats_start_copy_compat()
165 if (d->tail) { in gnet_stats_copy_basic()
209 if (d->tail) { in gnet_stats_copy_rate_est()
286 if (d->tail) in gnet_stats_copy_queue()
317 if (d->tail) in gnet_stats_copy_app()
344 if (d->tail) in gnet_stats_finish_copy()
345 d->tail->nla_len = skb_tail_pointer(d->skb) - (u8 *)d->tail; in gnet_stats_finish_copy()
Dskbuff.c102 (unsigned long)skb->tail, (unsigned long)skb->end, in skb_panic()
175 memset(skb, 0, offsetof(struct sk_buff, tail)); in __alloc_skb_head()
245 memset(skb, 0, offsetof(struct sk_buff, tail)); in __alloc_skb()
253 skb->end = skb->tail + size; in __alloc_skb()
315 memset(skb, 0, offsetof(struct sk_buff, tail)); in __build_skb()
321 skb->end = skb->tail + size; in __build_skb()
846 C(tail); in __skb_clone()
1196 skb->tail += off; in pskb_expand_head()
1314 ntail = skb->data_len + pad - (skb->end - skb->tail); in skb_pad()
1350 unsigned char *pskb_put(struct sk_buff *skb, struct sk_buff *tail, int len) in pskb_put() argument
[all …]
/net/sunrpc/xprtrdma/
Drpc_rdma.c102 size_t tlen = buf->tail[0].iov_len; in rpcrdma_tail_pullup()
118 src = buf->tail[0].iov_base; in rpcrdma_tail_pullup()
190 if (xdrbuf->tail[0].iov_len) { in rpcrdma_convert_iovs()
193 if (xdrbuf->tail[0].iov_len < 4 && xprt_rdma_pad_optimize) in rpcrdma_convert_iovs()
199 seg[n].mr_offset = xdrbuf->tail[0].iov_base; in rpcrdma_convert_iovs()
200 seg[n].mr_len = xdrbuf->tail[0].iov_len; in rpcrdma_convert_iovs()
383 if (rqst->rq_snd_buf.tail[0].iov_len) { in rpcrdma_inline_pullup()
384 curlen = rqst->rq_snd_buf.tail[0].iov_len; in rpcrdma_inline_pullup()
385 if (destp + copy_len != rqst->rq_snd_buf.tail[0].iov_base) { in rpcrdma_inline_pullup()
387 rqst->rq_snd_buf.tail[0].iov_base, curlen); in rpcrdma_inline_pullup()
[all …]
Dsvc_rdma_sendto.c64 (xdr->head[0].iov_len + xdr->page_len + xdr->tail[0].iov_len)) { in map_xdr()
94 if (xdr->tail[0].iov_len) { in map_xdr()
95 vec->sge[sge_no].iov_base = xdr->tail[0].iov_base; in map_xdr()
96 vec->sge[sge_no].iov_len = xdr->tail[0].iov_len; in map_xdr()
103 xdr->head[0].iov_len, xdr->tail[0].iov_len); in map_xdr()
130 xdr->tail[0].iov_base & ~PAGE_MASK; in dma_map_xdr()
131 page = virt_to_page(xdr->tail[0].iov_base); in dma_map_xdr()
311 u32 xfer_len = rqstp->rq_res.page_len + rqstp->rq_res.tail[0].iov_len; in send_write_chunks()
367 return rqstp->rq_res.page_len + rqstp->rq_res.tail[0].iov_len; in send_write_chunks()
Dsvc_rdma_recvfrom.c114 rqstp->rq_arg.tail[0].iov_base = NULL; in rdma_build_arg_xdr()
115 rqstp->rq_arg.tail[0].iov_len = 0; in rdma_build_arg_xdr()
439 head->arg.tail[0] = rqstp->rq_arg.tail[0]; in rdma_read_chunks()
546 rqstp->rq_arg.tail[0] = head->arg.tail[0]; in rdma_read_complete()
559 + rqstp->rq_arg.tail[0].iov_len; in rdma_read_complete()
646 + rqstp->rq_arg.tail[0].iov_len; in svc_rdma_recvfrom()
/net/ipv4/
Dtcp_cdg.c82 u8 tail; member
195 ca->gsum.min += gmin - ca->gradients[ca->tail].min; in tcp_cdg_grad()
196 ca->gsum.max += gmax - ca->gradients[ca->tail].max; in tcp_cdg_grad()
197 ca->gradients[ca->tail].min = gmin; in tcp_cdg_grad()
198 ca->gradients[ca->tail].max = gmax; in tcp_cdg_grad()
199 ca->tail = (ca->tail + 1) & (window - 1); in tcp_cdg_grad()
216 else if (ca->tail == 0) in tcp_cdg_grad()
219 grad = (grad * window) / (int)ca->tail; in tcp_cdg_grad()
Dtcp_probe.c82 unsigned long head, tail; member
88 return (tcp_probe.head - tcp_probe.tail) & (bufsize - 1); in tcp_probe_used()
179 tcp_probe.head = tcp_probe.tail = 0; in tcpprobe_open()
189 = tcp_probe.log + tcp_probe.tail; in tcpprobe_sprint()
221 if (tcp_probe.head == tcp_probe.tail) { in tcpprobe_read()
230 tcp_probe.tail = (tcp_probe.tail + 1) & (bufsize - 1); in tcpprobe_read()
Desp4.c142 u8 *tail; in esp_output() local
198 tail = skb_tail_pointer(trailer); in esp_output()
200 memset(tail, 0, tfclen); in esp_output()
201 tail += tfclen; in esp_output()
206 tail[i] = i + 1; in esp_output()
208 tail[plen - 2] = plen - 2; in esp_output()
209 tail[plen - 1] = *skb_mac_header(skb); in esp_output()
/net/bluetooth/cmtp/
Dcore.c214 unsigned int size, tail; in cmtp_process_transmit() local
227 tail = session->mtu - nskb->len; in cmtp_process_transmit()
228 if (tail < 5) { in cmtp_process_transmit()
231 tail = session->mtu; in cmtp_process_transmit()
234 size = min_t(uint, ((tail < 258) ? (tail - 2) : (tail - 3)), skb->len); in cmtp_process_transmit()
/net/mac80211/
Dwpa.c39 int tail; in ieee80211_tx_h_michael_mic_add() local
66 tail = MICHAEL_MIC_LEN; in ieee80211_tx_h_michael_mic_add()
68 tail += IEEE80211_TKIP_ICV_LEN; in ieee80211_tx_h_michael_mic_add()
70 if (WARN(skb_tailroom(skb) < tail || in ieee80211_tx_h_michael_mic_add()
74 skb_tailroom(skb), tail)) in ieee80211_tx_h_michael_mic_add()
194 int len, tail; in tkip_encrypt_skb() local
208 tail = 0; in tkip_encrypt_skb()
210 tail = IEEE80211_TKIP_ICV_LEN; in tkip_encrypt_skb()
212 if (WARN_ON(skb_tailroom(skb) < tail || in tkip_encrypt_skb()
409 int hdrlen, len, tail; in ccmp_encrypt_skb() local
[all …]
/net/tipc/
Dmsg.c124 struct sk_buff *tail = NULL; in tipc_buf_append() local
148 TIPC_SKB_CB(head)->tail = NULL; in tipc_buf_append()
158 tail = TIPC_SKB_CB(head)->tail; in tipc_buf_append()
162 tail->next = frag; in tipc_buf_append()
166 TIPC_SKB_CB(head)->tail = frag; in tipc_buf_append()
174 TIPC_SKB_CB(head)->tail = NULL; in tipc_buf_append()
/net/sunrpc/auth_gss/
Dsvcauth_gss.c874 return buf->head[0].iov_len + buf->page_len + buf->tail[0].iov_len; in total_buf_len()
1607 if (resbuf->tail[0].iov_base == NULL) { in svcauth_gss_wrap_resp_integ()
1610 resbuf->tail[0].iov_base = resbuf->head[0].iov_base in svcauth_gss_wrap_resp_integ()
1612 resbuf->tail[0].iov_len = 0; in svcauth_gss_wrap_resp_integ()
1614 resv = &resbuf->tail[0]; in svcauth_gss_wrap_resp_integ()
1659 if (resbuf->tail[0].iov_base) { in svcauth_gss_wrap_resp_priv()
1660 BUG_ON(resbuf->tail[0].iov_base >= resbuf->head[0].iov_base in svcauth_gss_wrap_resp_priv()
1662 BUG_ON(resbuf->tail[0].iov_base < resbuf->head[0].iov_base); in svcauth_gss_wrap_resp_priv()
1663 if (resbuf->tail[0].iov_len + resbuf->head[0].iov_len in svcauth_gss_wrap_resp_priv()
1666 memmove(resbuf->tail[0].iov_base + RPC_MAX_AUTH_SIZE, in svcauth_gss_wrap_resp_priv()
[all …]
Dgss_krb5_crypto.c678 if (buf->tail[0].iov_base != NULL) { in gss_krb5_aes_encrypt()
679 ecptr = buf->tail[0].iov_base + buf->tail[0].iov_len; in gss_krb5_aes_encrypt()
681 buf->tail[0].iov_base = buf->head[0].iov_base in gss_krb5_aes_encrypt()
683 buf->tail[0].iov_len = 0; in gss_krb5_aes_encrypt()
684 ecptr = buf->tail[0].iov_base; in gss_krb5_aes_encrypt()
689 buf->tail[0].iov_len += GSS_KRB5_TOK_HDR_LEN; in gss_krb5_aes_encrypt()
694 hmac.data = buf->tail[0].iov_base + buf->tail[0].iov_len; in gss_krb5_aes_encrypt()
750 buf->tail[0].iov_len += kctx->gk5e->cksumlength; in gss_krb5_aes_encrypt()
/net/vmw_vsock/
Dvmci_transport_notify.c193 u64 tail; in send_waiting_read() local
209 vmci_qpair_get_consume_indexes(vmci_trans(vsk)->qpair, &tail, &head); in send_waiting_read()
235 u64 tail; in send_waiting_write() local
245 vmci_qpair_get_produce_indexes(vmci_trans(vsk)->qpair, &tail, &head); in send_waiting_write()
246 room_left = vmci_trans(vsk)->produce_size - tail; in send_waiting_write()
252 waiting_info.offset = tail + room_needed + 1; in send_waiting_write()
/net/ipv6/
Desp6.c168 u8 *tail; in esp6_output() local
215 tail = skb_tail_pointer(trailer); in esp6_output()
217 memset(tail, 0, tfclen); in esp6_output()
218 tail += tfclen; in esp6_output()
223 tail[i] = i + 1; in esp6_output()
225 tail[plen - 2] = plen - 2; in esp6_output()
226 tail[plen - 1] = *skb_mac_header(skb); in esp6_output()
/net/rxrpc/
Dar-ack.c428 int tail = call->acks_tail, old_tail; in rxrpc_rotate_tx_window() local
429 int win = CIRC_CNT(call->acks_head, tail, call->acks_winsz); in rxrpc_rotate_tx_window()
437 _skb = call->acks_window[tail] & ~1; in rxrpc_rotate_tx_window()
439 old_tail = tail; in rxrpc_rotate_tx_window()
440 tail = (tail + 1) & (call->acks_winsz - 1); in rxrpc_rotate_tx_window()
441 call->acks_tail = tail; in rxrpc_rotate_tx_window()
443 call->acks_unacked = tail; in rxrpc_rotate_tx_window()
575 int tail; in rxrpc_zap_tx_window() local
581 tail = call->acks_tail; in rxrpc_zap_tx_window()
583 _skb = acks_window[tail] & ~1; in rxrpc_zap_tx_window()
/net/decnet/netfilter/
Ddn_rtmsg.c49 old_tail = skb->tail; in dnrmg_build_message()
60 nlh->nlmsg_len = skb->tail - old_tail; in dnrmg_build_message()
/net/dccp/
Doptions.c409 const unsigned char *tail, *from; in dccp_insert_option_ackvec() local
435 tail = av->av_buf + DCCPAV_MAX_ACKVEC_LEN; in dccp_insert_option_ackvec()
453 if (from + copylen > tail) { in dccp_insert_option_ackvec()
454 const u16 tailsize = tail - from; in dccp_insert_option_ackvec()
/net/caif/
Dcfpkt_skbuff.c314 if (dst->tail + neededtailspace > dst->end) { in cfpkt_append()
331 dst->tail += addlen; in cfpkt_append()
368 skb2->tail += len2nd; in cfpkt_split()

123