/net/core/ |
D | datagram.c | 416 int i, copy = start - offset, start_off = offset, n; in __skb_datagram_iter() local 420 if (copy > 0) { in __skb_datagram_iter() 421 if (copy > len) in __skb_datagram_iter() 422 copy = len; in __skb_datagram_iter() 423 n = cb(skb->data + offset, copy, data, to); in __skb_datagram_iter() 425 if (n != copy) in __skb_datagram_iter() 427 if ((len -= copy) == 0) in __skb_datagram_iter() 439 if ((copy = end - offset) > 0) { in __skb_datagram_iter() 443 if (copy > len) in __skb_datagram_iter() 444 copy = len; in __skb_datagram_iter() [all …]
|
D | skbuff.c | 1380 u32 copy, done = 0; in skb_copy_ubufs() local 1388 copy = min_t(u32, PAGE_SIZE - d_off, p_len - done); in skb_copy_ubufs() 1390 vaddr + p_off + done, copy); in skb_copy_ubufs() 1391 done += copy; in skb_copy_ubufs() 1392 d_off += copy; in skb_copy_ubufs() 2192 int i, copy; in skb_copy_bits() local 2198 if ((copy = start - offset) > 0) { in skb_copy_bits() 2199 if (copy > len) in skb_copy_bits() 2200 copy = len; in skb_copy_bits() 2201 skb_copy_from_linear_data_offset(skb, offset, to, copy); in skb_copy_bits() [all …]
|
D | skmsg.c | 361 u32 copy, buf_size; in sk_msg_memcopy_from_iter() local 376 copy = (buf_size > bytes) ? bytes : buf_size; in sk_msg_memcopy_from_iter() 378 msg->sg.copybreak += copy; in sk_msg_memcopy_from_iter() 380 ret = copy_from_iter_nocache(to, copy, from); in sk_msg_memcopy_from_iter() 382 ret = copy_from_iter(to, copy, from); in sk_msg_memcopy_from_iter() 383 if (ret != copy) { in sk_msg_memcopy_from_iter() 387 bytes -= copy; in sk_msg_memcopy_from_iter()
|
/net/sunrpc/ |
D | xdr.c | 224 size_t copy; in _shift_data_right_pages() local 248 copy = len; in _shift_data_right_pages() 249 if (copy > pgto_base) in _shift_data_right_pages() 250 copy = pgto_base; in _shift_data_right_pages() 251 if (copy > pgfrom_base) in _shift_data_right_pages() 252 copy = pgfrom_base; in _shift_data_right_pages() 253 pgto_base -= copy; in _shift_data_right_pages() 254 pgfrom_base -= copy; in _shift_data_right_pages() 259 memcpy(vto + pgto_base, vfrom + pgfrom_base, copy); in _shift_data_right_pages() 262 memmove(vto + pgto_base, vto + pgfrom_base, copy); in _shift_data_right_pages() [all …]
|
/net/ipv4/ |
D | tcp_bpf.c | 65 int copy; in __tcp_bpf_recvmsg() local 68 copy = sge->length; in __tcp_bpf_recvmsg() 70 if (copied + copy > len) in __tcp_bpf_recvmsg() 71 copy = len - copied; in __tcp_bpf_recvmsg() 72 ret = copy_page_to_iter(page, sge->offset, copy, iter); in __tcp_bpf_recvmsg() 73 if (ret != copy) { in __tcp_bpf_recvmsg() 78 copied += copy; in __tcp_bpf_recvmsg() 80 sge->offset += copy; in __tcp_bpf_recvmsg() 81 sge->length -= copy; in __tcp_bpf_recvmsg() 82 sk_mem_uncharge(sk, copy); in __tcp_bpf_recvmsg() [all …]
|
D | ip_output.c | 943 csum_page(struct page *page, int offset, int copy) in csum_page() argument 948 csum = csum_partial(kaddr + offset, copy, 0); in csum_page() 971 int copy; in __ip_append_data() local 1042 copy = mtu - skb->len; in __ip_append_data() 1043 if (copy < length) in __ip_append_data() 1044 copy = maxfraglen - skb->len; in __ip_append_data() 1045 if (copy <= 0) { in __ip_append_data() 1132 copy = datalen - transhdrlen - fraggap - pagedlen; in __ip_append_data() 1133 if (copy > 0 && getfrag(from, data + transhdrlen, offset, copy, fraggap, skb) < 0) { in __ip_append_data() 1139 offset += copy; in __ip_append_data() [all …]
|
D | tcp.c | 991 int copy, i; in do_tcp_sendpages() local 994 if (!skb || (copy = size_goal - skb->len) <= 0 || in do_tcp_sendpages() 1009 copy = size_goal; in do_tcp_sendpages() 1012 if (copy > size) in do_tcp_sendpages() 1013 copy = size; in do_tcp_sendpages() 1021 if (!sk_wmem_schedule(sk, copy)) in do_tcp_sendpages() 1025 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in do_tcp_sendpages() 1028 skb_fill_page_desc(skb, i, page, offset, copy); in do_tcp_sendpages() 1034 skb->len += copy; in do_tcp_sendpages() 1035 skb->data_len += copy; in do_tcp_sendpages() [all …]
|
D | raw.c | 473 int copy = min(rfv->hlen - offset, len); in raw_getfrag() local 476 memcpy(to, rfv->hdr.c + offset, copy); in raw_getfrag() 481 to, copy, 0), in raw_getfrag() 485 offset += copy; in raw_getfrag() 486 to += copy; in raw_getfrag() 487 len -= copy; in raw_getfrag()
|
D | tcp_output.c | 2097 int copy, len; in tcp_mtu_probe() local 2177 copy = min_t(int, skb->len, probe_size - len); in tcp_mtu_probe() 2178 skb_copy_bits(skb, 0, skb_put(nskb, copy), copy); in tcp_mtu_probe() 2180 if (skb->len <= copy) { in tcp_mtu_probe() 2195 skb_pull(skb, copy); in tcp_mtu_probe() 2197 __pskb_trim_head(skb, copy); in tcp_mtu_probe() 2200 TCP_SKB_CB(skb)->seq += copy; in tcp_mtu_probe() 2203 len += copy; in tcp_mtu_probe()
|
/net/ipv6/ |
D | ip6_output.c | 1329 int copy; in __ip6_append_data() local 1439 copy = (cork->length <= mtu && !(cork->flags & IPCORK_ALLFRAG) ? mtu : maxfraglen) - skb->len; in __ip6_append_data() 1440 if (copy < length) in __ip6_append_data() 1441 copy = maxfraglen - skb->len; in __ip6_append_data() 1443 if (copy <= 0) { in __ip6_append_data() 1505 copy = datalen - transhdrlen - fraggap - pagedlen; in __ip6_append_data() 1506 if (copy < 0) { in __ip6_append_data() 1552 if (copy > 0 && in __ip6_append_data() 1554 copy, fraggap, skb) < 0) { in __ip6_append_data() 1560 offset += copy; in __ip6_append_data() [all …]
|
D | raw.c | 741 int copy = min(rfv->hlen - offset, len); in raw6_getfrag() local 744 memcpy(to, rfv->c + offset, copy); in raw6_getfrag() 749 to, copy, 0), in raw6_getfrag() 753 offset += copy; in raw6_getfrag() 754 to += copy; in raw6_getfrag() 755 len -= copy; in raw6_getfrag()
|
/net/tls/ |
D | tls_device.c | 414 int copy, rc = 0; in tls_push_data() local 470 copy = min_t(size_t, size, (pfrag->size - pfrag->offset)); in tls_push_data() 471 copy = min_t(size_t, copy, (max_open_record_len - record->len)); in tls_push_data() 474 pfrag->offset, copy, msg_iter); in tls_push_data() 477 tls_append_frag(record, pfrag, copy); in tls_push_data() 479 size -= copy; in tls_push_data() 755 int err = 0, offset = rxm->offset, copy, nsg, data_len, pos; in tls_device_reencrypt() local 791 copy = min_t(int, skb_pagelen(skb) - offset, data_len); in tls_device_reencrypt() 794 err = skb_store_bits(skb, offset, buf, copy); in tls_device_reencrypt() 799 offset += copy; in tls_device_reencrypt() [all …]
|
D | tls_sw.c | 1141 size_t copy, required_size; in tls_sw_do_sendpage() local 1161 copy = size; in tls_sw_do_sendpage() 1162 if (copy >= record_room) { in tls_sw_do_sendpage() 1163 copy = record_room; in tls_sw_do_sendpage() 1167 required_size = msg_pl->sg.size + copy + prot->overhead_size; in tls_sw_do_sendpage() 1181 copy -= required_size - msg_pl->sg.size; in tls_sw_do_sendpage() 1185 sk_msg_page_add(msg_pl, page, copy, offset); in tls_sw_do_sendpage() 1186 sk_mem_charge(sk, copy); in tls_sw_do_sendpage() 1188 offset += copy; in tls_sw_do_sendpage() 1189 size -= copy; in tls_sw_do_sendpage() [all …]
|
/net/wireless/ |
D | util.c | 1329 unsigned int copy; in cfg80211_get_p2p_attr() local 1355 copy = min_t(unsigned int, attr_remaining, iedatalen); in cfg80211_get_p2p_attr() 1356 if (copy && desired_attr) { in cfg80211_get_p2p_attr() 1357 desired_len += copy; in cfg80211_get_p2p_attr() 1359 memcpy(out, iedata, min(bufsize, copy)); in cfg80211_get_p2p_attr() 1360 out += min(bufsize, copy); in cfg80211_get_p2p_attr() 1361 bufsize -= min(bufsize, copy); in cfg80211_get_p2p_attr() 1365 if (copy == attr_remaining) in cfg80211_get_p2p_attr() 1369 attr_remaining -= copy; in cfg80211_get_p2p_attr() 1373 iedatalen -= copy; in cfg80211_get_p2p_attr() [all …]
|
/net/rxrpc/ |
D | sendmsg.c | 369 int copy = skb_tailroom(skb); in rxrpc_send_data() local 370 ASSERTCMP(copy, >, 0); in rxrpc_send_data() 371 if (copy > msg_data_left(msg)) in rxrpc_send_data() 372 copy = msg_data_left(msg); in rxrpc_send_data() 373 if (copy > sp->remain) in rxrpc_send_data() 374 copy = sp->remain; in rxrpc_send_data() 377 ret = skb_add_data(skb, &msg->msg_iter, copy); in rxrpc_send_data() 381 sp->remain -= copy; in rxrpc_send_data() 382 skb->mark += copy; in rxrpc_send_data() 383 copied += copy; in rxrpc_send_data() [all …]
|
D | recvmsg.c | 319 int ix, copy, ret = -EAGAIN, ret2; in rxrpc_recvmsg_data() local 381 copy = rx_pkt_len; in rxrpc_recvmsg_data() 382 if (copy > remain) in rxrpc_recvmsg_data() 383 copy = remain; in rxrpc_recvmsg_data() 384 if (copy > 0) { in rxrpc_recvmsg_data() 386 copy); in rxrpc_recvmsg_data() 393 rx_pkt_offset += copy; in rxrpc_recvmsg_data() 394 rx_pkt_len -= copy; in rxrpc_recvmsg_data() 395 *_offset += copy; in rxrpc_recvmsg_data()
|
/net/appletalk/ |
D | ddp.c | 930 int i, copy; in atalk_sum_skb() local 933 if ((copy = start - offset) > 0) { in atalk_sum_skb() 934 if (copy > len) in atalk_sum_skb() 935 copy = len; in atalk_sum_skb() 936 sum = atalk_sum_partial(skb->data + offset, copy, sum); in atalk_sum_skb() 937 if ((len -= copy) == 0) in atalk_sum_skb() 940 offset += copy; in atalk_sum_skb() 950 if ((copy = end - offset) > 0) { in atalk_sum_skb() 953 if (copy > len) in atalk_sum_skb() 954 copy = len; in atalk_sum_skb() [all …]
|
/net/mac80211/ |
D | mesh_pathtbl.c | 162 bool copy) in mesh_path_move_to_queue() argument 193 if (copy) in mesh_path_move_to_queue() 203 if (!copy) in mesh_path_move_to_queue() 680 bool copy = false; in mesh_path_send_to_gates() local 688 mesh_path_move_to_queue(gate, from_mpath, copy); in mesh_path_send_to_gates() 690 copy = true; in mesh_path_send_to_gates()
|
/net/rds/ |
D | recv.c | 456 LIST_HEAD(copy); in rds_notify_queue_get() 477 list_move(¬ifier->n_list, ©); in rds_notify_queue_get() 485 while (!list_empty(©)) { in rds_notify_queue_get() 486 notifier = list_entry(copy.next, struct rds_notifier, n_list); in rds_notify_queue_get() 505 if (!list_empty(©)) { in rds_notify_queue_get() 507 list_splice(©, &rs->rs_notify_queue); in rds_notify_queue_get()
|
D | message.c | 78 LIST_HEAD(copy); in rds_notify_msg_zcopy_purge() 82 list_splice(&q->zcookie_head, ©); in rds_notify_msg_zcopy_purge() 86 list_for_each_entry_safe(info, tmp, ©, rs_zcookie_next) { in rds_notify_msg_zcopy_purge()
|
/net/smc/ |
D | smc_rx.c | 318 goto copy; in smc_rx_recvmsg() 326 goto copy; in smc_rx_recvmsg() 364 copy: in smc_rx_recvmsg()
|
/net/kcm/ |
D | kcmsock.c | 905 size_t copy, copied = 0; in kcm_sendmsg() local 985 copy = min_t(int, msg_data_left(msg), in kcm_sendmsg() 988 if (!sk_wmem_schedule(sk, copy)) in kcm_sendmsg() 994 copy); in kcm_sendmsg() 1000 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in kcm_sendmsg() 1003 pfrag->offset, copy); in kcm_sendmsg() 1007 pfrag->offset += copy; in kcm_sendmsg() 1008 copied += copy; in kcm_sendmsg() 1010 head->len += copy; in kcm_sendmsg() 1011 head->data_len += copy; in kcm_sendmsg()
|
/net/tipc/ |
D | socket.c | 1744 int rc, err, hlen, dlen, copy; in tipc_recvmsg() local 1787 copy = min_t(int, dlen, buflen); in tipc_recvmsg() 1788 if (unlikely(copy != dlen)) in tipc_recvmsg() 1790 rc = skb_copy_datagram_msg(skb, hlen, m, copy); in tipc_recvmsg() 1792 copy = 0; in tipc_recvmsg() 1805 copy = 0; in tipc_recvmsg() 1832 return rc ? rc : copy; in tipc_recvmsg() 1855 int offset, required, copy, copied = 0; in tipc_recvstream() local 1902 copy = min_t(int, dlen - offset, buflen - copied); in tipc_recvstream() 1903 rc = skb_copy_datagram_msg(skb, hlen + offset, m, copy); in tipc_recvstream() [all …]
|
/net/iucv/ |
D | iucv.c | 1115 size_t copy; in iucv_message_receive_iprmdata() local 1128 copy = min_t(size_t, size, array->length); in iucv_message_receive_iprmdata() 1130 rmmsg, copy); in iucv_message_receive_iprmdata() 1131 rmmsg += copy; in iucv_message_receive_iprmdata() 1132 size -= copy; in iucv_message_receive_iprmdata()
|
/net/ax25/ |
D | af_ax25.c | 242 struct sk_buff *copy; in ax25_send_to_raw() local 251 if ((copy = skb_clone(skb, GFP_ATOMIC)) == NULL) in ax25_send_to_raw() 253 if (sock_queue_rcv_skb(s->sk, copy) != 0) in ax25_send_to_raw() 254 kfree_skb(copy); in ax25_send_to_raw()
|