/net/core/ |
D | datagram.c | 422 int i, copy = start - offset, start_off = offset, n; in __skb_datagram_iter() local 426 if (copy > 0) { in __skb_datagram_iter() 427 if (copy > len) in __skb_datagram_iter() 428 copy = len; in __skb_datagram_iter() 430 skb->data + offset, copy, data, to); in __skb_datagram_iter() 432 if (n != copy) in __skb_datagram_iter() 434 if ((len -= copy) == 0) in __skb_datagram_iter() 446 if ((copy = end - offset) > 0) { in __skb_datagram_iter() 450 if (copy > len) in __skb_datagram_iter() 451 copy = len; in __skb_datagram_iter() [all …]
|
D | skbuff.c | 1399 u32 copy, done = 0; in skb_copy_ubufs() local 1407 copy = min_t(u32, PAGE_SIZE - d_off, p_len - done); in skb_copy_ubufs() 1409 vaddr + p_off + done, copy); in skb_copy_ubufs() 1410 done += copy; in skb_copy_ubufs() 1411 d_off += copy; in skb_copy_ubufs() 2220 int i, copy; in skb_copy_bits() local 2226 if ((copy = start - offset) > 0) { in skb_copy_bits() 2227 if (copy > len) in skb_copy_bits() 2228 copy = len; in skb_copy_bits() 2229 skb_copy_from_linear_data_offset(skb, offset, to, copy); in skb_copy_bits() [all …]
|
D | skmsg.c | 373 u32 copy, buf_size; in sk_msg_memcopy_from_iter() local 388 copy = (buf_size > bytes) ? bytes : buf_size; in sk_msg_memcopy_from_iter() 390 msg->sg.copybreak += copy; in sk_msg_memcopy_from_iter() 392 ret = copy_from_iter_nocache(to, copy, from); in sk_msg_memcopy_from_iter() 394 ret = copy_from_iter(to, copy, from); in sk_msg_memcopy_from_iter() 395 if (ret != copy) { in sk_msg_memcopy_from_iter() 399 bytes -= copy; in sk_msg_memcopy_from_iter()
|
/net/sunrpc/ |
D | xdr.c | 224 size_t copy; in _shift_data_right_pages() local 248 copy = len; in _shift_data_right_pages() 249 if (copy > pgto_base) in _shift_data_right_pages() 250 copy = pgto_base; in _shift_data_right_pages() 251 if (copy > pgfrom_base) in _shift_data_right_pages() 252 copy = pgfrom_base; in _shift_data_right_pages() 253 pgto_base -= copy; in _shift_data_right_pages() 254 pgfrom_base -= copy; in _shift_data_right_pages() 259 memcpy(vto + pgto_base, vfrom + pgfrom_base, copy); in _shift_data_right_pages() 262 memmove(vto + pgto_base, vto + pgfrom_base, copy); in _shift_data_right_pages() [all …]
|
/net/ipv4/ |
D | tcp_bpf.c | 65 int copy; in __tcp_bpf_recvmsg() local 68 copy = sge->length; in __tcp_bpf_recvmsg() 70 if (copied + copy > len) in __tcp_bpf_recvmsg() 71 copy = len - copied; in __tcp_bpf_recvmsg() 72 copy = copy_page_to_iter(page, sge->offset, copy, iter); in __tcp_bpf_recvmsg() 73 if (!copy) in __tcp_bpf_recvmsg() 76 copied += copy; in __tcp_bpf_recvmsg() 78 sge->offset += copy; in __tcp_bpf_recvmsg() 79 sge->length -= copy; in __tcp_bpf_recvmsg() 81 sk_mem_uncharge(sk, copy); in __tcp_bpf_recvmsg() [all …]
|
D | ip_output.c | 952 csum_page(struct page *page, int offset, int copy) in csum_page() argument 957 csum = csum_partial(kaddr + offset, copy, 0); in csum_page() 980 int copy; in __ip_append_data() local 1051 copy = mtu - skb->len; in __ip_append_data() 1052 if (copy < length) in __ip_append_data() 1053 copy = maxfraglen - skb->len; in __ip_append_data() 1054 if (copy <= 0) { in __ip_append_data() 1145 copy = datalen - transhdrlen - fraggap - pagedlen; in __ip_append_data() 1146 if (copy > 0 && getfrag(from, data + transhdrlen, offset, copy, fraggap, skb) < 0) { in __ip_append_data() 1152 offset += copy; in __ip_append_data() [all …]
|
D | tcp.c | 1006 int copy, i; in do_tcp_sendpages() local 1009 if (!skb || (copy = size_goal - skb->len) <= 0 || in do_tcp_sendpages() 1024 copy = size_goal; in do_tcp_sendpages() 1027 if (copy > size) in do_tcp_sendpages() 1028 copy = size; in do_tcp_sendpages() 1036 if (!sk_wmem_schedule(sk, copy)) in do_tcp_sendpages() 1040 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in do_tcp_sendpages() 1043 skb_fill_page_desc(skb, i, page, offset, copy); in do_tcp_sendpages() 1049 skb->len += copy; in do_tcp_sendpages() 1050 skb->data_len += copy; in do_tcp_sendpages() [all …]
|
D | raw.c | 473 int copy = min(rfv->hlen - offset, len); in raw_getfrag() local 476 memcpy(to, rfv->hdr.c + offset, copy); in raw_getfrag() 481 to, copy, 0), in raw_getfrag() 485 offset += copy; in raw_getfrag() 486 to += copy; in raw_getfrag() 487 len -= copy; in raw_getfrag()
|
D | tcp_output.c | 2111 int copy, len; in tcp_mtu_probe() local 2191 copy = min_t(int, skb->len, probe_size - len); in tcp_mtu_probe() 2192 skb_copy_bits(skb, 0, skb_put(nskb, copy), copy); in tcp_mtu_probe() 2194 if (skb->len <= copy) { in tcp_mtu_probe() 2209 skb_pull(skb, copy); in tcp_mtu_probe() 2211 __pskb_trim_head(skb, copy); in tcp_mtu_probe() 2214 TCP_SKB_CB(skb)->seq += copy; in tcp_mtu_probe() 2217 len += copy; in tcp_mtu_probe()
|
/net/ipv6/ |
D | ip6_output.c | 1407 int copy; in __ip6_append_data() local 1522 copy = (cork->length <= mtu && !(cork->flags & IPCORK_ALLFRAG) ? mtu : maxfraglen) - skb->len; in __ip6_append_data() 1523 if (copy < length) in __ip6_append_data() 1524 copy = maxfraglen - skb->len; in __ip6_append_data() 1526 if (copy <= 0) { in __ip6_append_data() 1591 copy = datalen - transhdrlen - fraggap - pagedlen; in __ip6_append_data() 1592 if (copy < 0) { in __ip6_append_data() 1637 if (copy > 0 && in __ip6_append_data() 1639 copy, fraggap, skb) < 0) { in __ip6_append_data() 1645 offset += copy; in __ip6_append_data() [all …]
|
D | raw.c | 745 int copy = min(rfv->hlen - offset, len); in raw6_getfrag() local 748 memcpy(to, rfv->c + offset, copy); in raw6_getfrag() 753 to, copy, 0), in raw6_getfrag() 757 offset += copy; in raw6_getfrag() 758 to += copy; in raw6_getfrag() 759 len -= copy; in raw6_getfrag()
|
/net/tls/ |
D | tls_device.c | 417 int copy, rc = 0; in tls_push_data() local 472 copy = min_t(size_t, size, (pfrag->size - pfrag->offset)); in tls_push_data() 473 copy = min_t(size_t, copy, (max_open_record_len - record->len)); in tls_push_data() 475 if (copy) { in tls_push_data() 477 pfrag->offset, copy, msg_iter); in tls_push_data() 480 tls_append_frag(record, pfrag, copy); in tls_push_data() 483 size -= copy; in tls_push_data() 779 int err = 0, offset = rxm->offset, copy, nsg, data_len, pos; in tls_device_reencrypt() local 815 copy = min_t(int, skb_pagelen(skb) - offset, data_len); in tls_device_reencrypt() 818 err = skb_store_bits(skb, offset, buf, copy); in tls_device_reencrypt() [all …]
|
D | tls_sw.c | 1169 size_t copy, required_size; in tls_sw_do_sendpage() local 1189 copy = size; in tls_sw_do_sendpage() 1190 if (copy >= record_room) { in tls_sw_do_sendpage() 1191 copy = record_room; in tls_sw_do_sendpage() 1195 required_size = msg_pl->sg.size + copy + prot->overhead_size; in tls_sw_do_sendpage() 1209 copy -= required_size - msg_pl->sg.size; in tls_sw_do_sendpage() 1213 sk_msg_page_add(msg_pl, page, copy, offset); in tls_sw_do_sendpage() 1216 sk_mem_charge(sk, copy); in tls_sw_do_sendpage() 1218 offset += copy; in tls_sw_do_sendpage() 1219 size -= copy; in tls_sw_do_sendpage() [all …]
|
/net/wireless/ |
D | util.c | 1395 unsigned int copy; in cfg80211_get_p2p_attr() local 1421 copy = min_t(unsigned int, attr_remaining, iedatalen); in cfg80211_get_p2p_attr() 1422 if (copy && desired_attr) { in cfg80211_get_p2p_attr() 1423 desired_len += copy; in cfg80211_get_p2p_attr() 1425 memcpy(out, iedata, min(bufsize, copy)); in cfg80211_get_p2p_attr() 1426 out += min(bufsize, copy); in cfg80211_get_p2p_attr() 1427 bufsize -= min(bufsize, copy); in cfg80211_get_p2p_attr() 1431 if (copy == attr_remaining) in cfg80211_get_p2p_attr() 1435 attr_remaining -= copy; in cfg80211_get_p2p_attr() 1439 iedatalen -= copy; in cfg80211_get_p2p_attr() [all …]
|
/net/rxrpc/ |
D | sendmsg.c | 398 int copy = skb_tailroom(skb); in rxrpc_send_data() local 399 ASSERTCMP(copy, >, 0); in rxrpc_send_data() 400 if (copy > msg_data_left(msg)) in rxrpc_send_data() 401 copy = msg_data_left(msg); in rxrpc_send_data() 402 if (copy > sp->remain) in rxrpc_send_data() 403 copy = sp->remain; in rxrpc_send_data() 406 ret = skb_add_data(skb, &msg->msg_iter, copy); in rxrpc_send_data() 410 sp->remain -= copy; in rxrpc_send_data() 411 skb->mark += copy; in rxrpc_send_data() 412 copied += copy; in rxrpc_send_data() [all …]
|
D | recvmsg.c | 317 int ix, copy, ret = -EAGAIN, ret2; in rxrpc_recvmsg_data() local 379 copy = rx_pkt_len; in rxrpc_recvmsg_data() 380 if (copy > remain) in rxrpc_recvmsg_data() 381 copy = remain; in rxrpc_recvmsg_data() 382 if (copy > 0) { in rxrpc_recvmsg_data() 384 copy); in rxrpc_recvmsg_data() 391 rx_pkt_offset += copy; in rxrpc_recvmsg_data() 392 rx_pkt_len -= copy; in rxrpc_recvmsg_data() 393 *_offset += copy; in rxrpc_recvmsg_data()
|
/net/appletalk/ |
D | ddp.c | 930 int i, copy; in atalk_sum_skb() local 933 if ((copy = start - offset) > 0) { in atalk_sum_skb() 934 if (copy > len) in atalk_sum_skb() 935 copy = len; in atalk_sum_skb() 936 sum = atalk_sum_partial(skb->data + offset, copy, sum); in atalk_sum_skb() 937 if ((len -= copy) == 0) in atalk_sum_skb() 940 offset += copy; in atalk_sum_skb() 950 if ((copy = end - offset) > 0) { in atalk_sum_skb() 953 if (copy > len) in atalk_sum_skb() 954 copy = len; in atalk_sum_skb() [all …]
|
/net/mac80211/ |
D | mesh_pathtbl.c | 158 bool copy) in mesh_path_move_to_queue() argument 189 if (copy) in mesh_path_move_to_queue() 199 if (!copy) in mesh_path_move_to_queue() 677 bool copy = false; in mesh_path_send_to_gates() local 685 mesh_path_move_to_queue(gate, from_mpath, copy); in mesh_path_send_to_gates() 687 copy = true; in mesh_path_send_to_gates()
|
/net/rds/ |
D | recv.c | 456 LIST_HEAD(copy); in rds_notify_queue_get() 478 list_move(¬ifier->n_list, ©); in rds_notify_queue_get() 486 while (!list_empty(©)) { in rds_notify_queue_get() 487 notifier = list_entry(copy.next, struct rds_notifier, n_list); in rds_notify_queue_get() 506 if (!list_empty(©)) { in rds_notify_queue_get() 508 list_splice(©, &rs->rs_notify_queue); in rds_notify_queue_get()
|
D | message.c | 78 LIST_HEAD(copy); in rds_notify_msg_zcopy_purge() 82 list_splice(&q->zcookie_head, ©); in rds_notify_msg_zcopy_purge() 86 list_for_each_entry_safe(info, tmp, ©, rs_zcookie_next) { in rds_notify_msg_zcopy_purge()
|
/net/smc/ |
D | smc_rx.c | 318 goto copy; in smc_rx_recvmsg() 326 goto copy; in smc_rx_recvmsg() 364 copy: in smc_rx_recvmsg()
|
/net/kcm/ |
D | kcmsock.c | 910 size_t copy, copied = 0; in kcm_sendmsg() local 990 copy = min_t(int, msg_data_left(msg), in kcm_sendmsg() 993 if (!sk_wmem_schedule(sk, copy)) in kcm_sendmsg() 999 copy); in kcm_sendmsg() 1005 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in kcm_sendmsg() 1008 pfrag->offset, copy); in kcm_sendmsg() 1012 pfrag->offset += copy; in kcm_sendmsg() 1013 copied += copy; in kcm_sendmsg() 1015 head->len += copy; in kcm_sendmsg() 1016 head->data_len += copy; in kcm_sendmsg()
|
/net/tipc/ |
D | socket.c | 1759 int rc, err, hlen, dlen, copy; in tipc_recvmsg() local 1806 copy = min_t(int, dlen - offset, buflen); in tipc_recvmsg() 1807 rc = skb_copy_datagram_msg(skb, hlen + offset, m, copy); in tipc_recvmsg() 1810 if (unlikely(offset + copy < dlen)) { in tipc_recvmsg() 1813 skb_cb->bytes_read = offset + copy; in tipc_recvmsg() 1824 copy = 0; in tipc_recvmsg() 1837 copy = 0; in tipc_recvmsg() 1867 return rc ? rc : copy; in tipc_recvmsg() 1890 int offset, required, copy, copied = 0; in tipc_recvstream() local 1937 copy = min_t(int, dlen - offset, buflen - copied); in tipc_recvstream() [all …]
|
/net/iucv/ |
D | iucv.c | 1115 size_t copy; in iucv_message_receive_iprmdata() local 1128 copy = min_t(size_t, size, array->length); in iucv_message_receive_iprmdata() 1130 rmmsg, copy); in iucv_message_receive_iprmdata() 1131 rmmsg += copy; in iucv_message_receive_iprmdata() 1132 size -= copy; in iucv_message_receive_iprmdata()
|
/net/ax25/ |
D | af_ax25.c | 258 struct sk_buff *copy; in ax25_send_to_raw() local 267 if ((copy = skb_clone(skb, GFP_ATOMIC)) == NULL) in ax25_send_to_raw() 269 if (sock_queue_rcv_skb(s->sk, copy) != 0) in ax25_send_to_raw() 270 kfree_skb(copy); in ax25_send_to_raw()
|