/net/sunrpc/ |
D | socklib.c | 90 xdr_partial_copy_from_skb(struct xdr_buf *xdr, unsigned int base, struct xdr_skb_reader *desc, xdr_… in xdr_partial_copy_from_skb() argument 98 if (base < len) { in xdr_partial_copy_from_skb() 99 len -= base; in xdr_partial_copy_from_skb() 100 ret = copy_actor(desc, (char *)xdr->head[0].iov_base + base, len); in xdr_partial_copy_from_skb() 104 base = 0; in xdr_partial_copy_from_skb() 106 base -= len; in xdr_partial_copy_from_skb() 110 if (unlikely(base >= pglen)) { in xdr_partial_copy_from_skb() 111 base -= pglen; in xdr_partial_copy_from_skb() 114 if (base || xdr->page_base) { in xdr_partial_copy_from_skb() 115 pglen -= base; in xdr_partial_copy_from_skb() [all …]
|
D | xdr.c | 179 struct page **pages, unsigned int base, unsigned int len) in xdr_inline_pages() argument 189 xdr->page_base = base; in xdr_inline_pages() 425 static void xdr_buf_iov_zero(const struct kvec *iov, unsigned int base, in xdr_buf_iov_zero() argument 428 if (base >= iov->iov_len) in xdr_buf_iov_zero() 430 if (len > iov->iov_len - base) in xdr_buf_iov_zero() 431 len = iov->iov_len - base; in xdr_buf_iov_zero() 432 memset(iov->iov_base + base, 0, len); in xdr_buf_iov_zero() 536 unsigned int base, unsigned int len, in xdr_buf_tail_copy_right() argument 540 unsigned int to = base + shift; in xdr_buf_tail_copy_right() 546 memmove(tail->iov_base + to, tail->iov_base + base, len); in xdr_buf_tail_copy_right() [all …]
|
/net/sctp/ |
D | endpointola.c | 65 ep->base.type = SCTP_EP_TYPE_SOCKET; in sctp_endpoint_init() 68 refcount_set(&ep->base.refcnt, 1); in sctp_endpoint_init() 69 ep->base.dead = false; in sctp_endpoint_init() 72 sctp_inq_init(&ep->base.inqueue); in sctp_endpoint_init() 75 sctp_inq_set_th_handler(&ep->base.inqueue, sctp_endpoint_bh_rcv); in sctp_endpoint_init() 78 sctp_bind_addr_init(&ep->base.bind_addr, 0); in sctp_endpoint_init() 112 ep->base.sk = sk; in sctp_endpoint_init() 113 ep->base.net = sock_net(sk); in sctp_endpoint_init() 114 sock_hold(ep->base.sk); in sctp_endpoint_init() 154 struct sock *sk = ep->base.sk; in sctp_endpoint_add_asoc() [all …]
|
D | associola.c | 66 asoc->base.sk = (struct sock *)sk; in sctp_association_init() 67 asoc->base.net = sock_net(sk); in sctp_association_init() 70 sock_hold(asoc->base.sk); in sctp_association_init() 73 asoc->base.type = SCTP_EP_TYPE_ASSOCIATION; in sctp_association_init() 76 refcount_set(&asoc->base.refcnt, 1); in sctp_association_init() 79 sctp_bind_addr_init(&asoc->base.bind_addr, ep->base.bind_addr.port); in sctp_association_init() 178 asoc->c.my_port = ep->base.bind_addr.port; in sctp_association_init() 223 sctp_inq_init(&asoc->base.inqueue); in sctp_association_init() 224 sctp_inq_set_th_handler(&asoc->base.inqueue, sctp_assoc_bh_rcv); in sctp_association_init() 243 if (asoc->base.sk->sk_family == PF_INET6) in sctp_association_init() [all …]
|
D | input.c | 166 rcvr = asoc ? &asoc->base : &ep->base; in sctp_rcv() 186 rcvr = &ep->base; in sctp_rcv() 507 sk = asoc->base.sk; in sctp_err_lookup() 565 struct sock *sk = asoc->base.sk; in sctp_v4_err_handle() 748 struct sock *sk = ep->base.sk; in __sctp_hash_endpoint() 752 ep->hashent = sctp_ep_hashfn(net, ep->base.bind_addr.port); in __sctp_hash_endpoint() 761 list_for_each(list, &ep->base.bind_addr.address_list) in __sctp_hash_endpoint() 765 struct sock *sk2 = ep2->base.sk; in __sctp_hash_endpoint() 812 struct sock *sk = ep->base.sk; in __sctp_unhash_endpoint() 815 ep->hashent = sctp_ep_hashfn(sock_net(sk), ep->base.bind_addr.port); in __sctp_unhash_endpoint() [all …]
|
D | ulpqueue.c | 168 struct sctp_sock *sp = sctp_sk(ulpq->asoc->base.sk); in sctp_ulpq_set_pd() 179 return sctp_clear_pd(ulpq->asoc->base.sk, ulpq->asoc); in sctp_ulpq_clear_pd() 184 struct sock *sk = ulpq->asoc->base.sk; in sctp_ulpq_tail_event() 482 if (!sctp_sk(asoc->base.sk)->frag_interleave && in sctp_ulpq_retrieve_reassembled() 483 atomic_read(&sctp_sk(asoc->base.sk)->pd_mode)) in sctp_ulpq_retrieve_reassembled() 487 pd_point = sctp_sk(asoc->base.sk)->pd_point; in sctp_ulpq_retrieve_reassembled() 489 retval = sctp_make_reassembled_event(asoc->base.net, in sctp_ulpq_retrieve_reassembled() 499 retval = sctp_make_reassembled_event(ulpq->asoc->base.net, in sctp_ulpq_retrieve_reassembled() 565 retval = sctp_make_reassembled_event(ulpq->asoc->base.net, &ulpq->reasm, in sctp_ulpq_retrieve_partial() 666 retval = sctp_make_reassembled_event(ulpq->asoc->base.net, &ulpq->reasm, in sctp_ulpq_retrieve_first() [all …]
|
D | diag.c | 34 laddr = list_entry(asoc->base.bind_addr.address_list.next, in inet_diag_msg_sctpasoc_fill() 40 r->id.idiag_sport = htons(asoc->base.bind_addr.port); in inet_diag_msg_sctpasoc_fill() 197 addr_list = asoc ? &asoc->base.bind_addr.address_list in inet_sctp_diag_fill() 198 : &ep->base.bind_addr.address_list; in inet_sctp_diag_fill() 232 list_for_each_entry_rcu(laddr, &asoc->base.bind_addr.address_list, in inet_assoc_attr_size() 249 struct sock *sk = ep->base.sk; in sctp_sock_dump_one() 289 struct sock *sk = ep->base.sk; in sctp_sock_dump() 303 if (r->id.idiag_sport != htons(assoc->base.bind_addr.port) && in sctp_sock_dump() 344 struct sock *sk = ep->base.sk; in sctp_sock_filter() 360 struct sock *sk = ep->base.sk; in sctp_ep_dump()
|
D | sm_sideeffect.c | 236 struct sock *sk = asoc->base.sk; in sctp_generate_t3_rtx_event() 273 struct sock *sk = asoc->base.sk; in sctp_generate_timeout_event() 291 if (asoc->base.dead) in sctp_generate_timeout_event() 366 struct sock *sk = asoc->base.sk; in sctp_generate_heartbeat_event() 412 struct sock *sk = asoc->base.sk; in sctp_generate_proto_unreach_event() 429 if (asoc->base.dead) in sctp_generate_proto_unreach_event() 447 struct sock *sk = asoc->base.sk; in sctp_generate_reconf_event() 483 struct sock *sk = asoc->base.sk; in sctp_generate_probe_event() 581 if (asoc->base.net->sctp.pf_enable && in sctp_do_8_2_transport_strike() 836 err = sctp_do_sm(asoc->base.net, SCTP_EVENT_T_OTHER, in sctp_cmd_process_sack() [all …]
|
/net/netfilter/ |
D | nf_conntrack_h323_asn1.c | 110 static int decode_nul(struct bitstr *bs, const struct field_t *f, char *base, int level); 111 static int decode_bool(struct bitstr *bs, const struct field_t *f, char *base, int level); 112 static int decode_oid(struct bitstr *bs, const struct field_t *f, char *base, int level); 113 static int decode_int(struct bitstr *bs, const struct field_t *f, char *base, int level); 114 static int decode_enum(struct bitstr *bs, const struct field_t *f, char *base, int level); 115 static int decode_bitstr(struct bitstr *bs, const struct field_t *f, char *base, int level); 116 static int decode_numstr(struct bitstr *bs, const struct field_t *f, char *base, int level); 117 static int decode_octstr(struct bitstr *bs, const struct field_t *f, char *base, int level); 118 static int decode_bmpstr(struct bitstr *bs, const struct field_t *f, char *base, int level); 119 static int decode_seq(struct bitstr *bs, const struct field_t *f, char *base, int level); [all …]
|
D | nft_payload.c | 133 switch (priv->base) { in nft_payload_eval() 189 priv->base = ntohl(nla_get_be32(tb[NFTA_PAYLOAD_BASE])); in nft_payload_init() 203 nla_put_be32(skb, NFTA_PAYLOAD_BASE, htonl(priv->base)) || in nft_payload_dump() 225 if (priv->base != payload->base || in nft_payload_reduce() 518 switch (priv->base) { in nft_payload_offload() 678 switch (priv->base) { in nft_payload_set_eval() 706 ((priv->base != NFT_PAYLOAD_TRANSPORT_HEADER && in nft_payload_set_eval() 707 priv->base != NFT_PAYLOAD_INNER_HEADER) || in nft_payload_set_eval() 746 priv->base = ntohl(nla_get_be32(tb[NFTA_PAYLOAD_BASE])); in nft_payload_set_init() 775 if (priv->base != NFT_PAYLOAD_TRANSPORT_HEADER) in nft_payload_set_init() [all …]
|
D | nft_set_pipapo.c | 957 static bool pipapo_step_diff(u8 *base, int step, int len) in pipapo_step_diff() argument 961 return !(BIT(step % BITS_PER_BYTE) & base[step / BITS_PER_BYTE]); in pipapo_step_diff() 964 base[len - 1 - step / BITS_PER_BYTE]); in pipapo_step_diff() 979 static bool pipapo_step_after_end(const u8 *base, const u8 *end, int step, in pipapo_step_after_end() argument 985 memcpy(tmp, base, len); in pipapo_step_after_end() 1004 static void pipapo_base_sum(u8 *base, int step, int len) argument 1016 base[i]++; 1018 base[i] += 1 << (step % BITS_PER_BYTE); 1020 if (base[i]) 1043 u8 base[NFT_PIPAPO_MAX_BYTES]; local [all …]
|
/net/ipv4/ |
D | inetpeer.c | 92 struct inet_peer_base *base, in lookup() argument 102 pp = &base->rb_root.rb_node; in lookup() 121 } else if (unlikely(read_seqretry(&base->lock, seq))) { in lookup() 140 static void inet_peer_gc(struct inet_peer_base *base, in inet_peer_gc() argument 153 if (base->total >= peer_threshold) in inet_peer_gc() 157 base->total / peer_threshold * HZ; in inet_peer_gc() 172 rb_erase(&p->rb_node, &base->rb_root); in inet_peer_gc() 173 base->total--; in inet_peer_gc() 179 struct inet_peer *inet_getpeer(struct inet_peer_base *base, in inet_getpeer() argument 192 seq = read_seqbegin(&base->lock); in inet_getpeer() [all …]
|
/net/rxrpc/ |
D | peer_event.c | 523 time64_t base, in rxrpc_peer_keepalive_dispatch() argument 545 slot = keepalive_at - base; in rxrpc_peer_keepalive_dispatch() 549 if (keepalive_at <= base || in rxrpc_peer_keepalive_dispatch() 550 keepalive_at > base + RXRPC_KEEPALIVE_TIME) { in rxrpc_peer_keepalive_dispatch() 580 time64_t base, now, delay; in rxrpc_peer_keepalive_worker() local 585 base = rxnet->peer_keepalive_base; in rxrpc_peer_keepalive_worker() 587 _enter("%lld,%u", base - now, cursor); in rxrpc_peer_keepalive_worker() 603 while (base <= now && (s8)(cursor - stop) < 0) { in rxrpc_peer_keepalive_worker() 606 base++; in rxrpc_peer_keepalive_worker() 610 base = now; in rxrpc_peer_keepalive_worker() [all …]
|
/net/ethtool/ |
D | linkmodes.c | 10 struct ethnl_req_info base; member 14 struct ethnl_reply_data base; member 21 container_of(__reply_base, struct linkmodes_reply_data, base) 36 data->lsettings = &data->ksettings.base; in linkmodes_prepare_data() 65 const struct ethtool_link_settings *lsettings = &ksettings->base; in linkmodes_reply_size() 106 const struct ethtool_link_settings *lsettings = &ksettings->base; in linkmodes_fill_reply() 201 (!req_speed || info->speed == ksettings->base.speed) && in ethnl_auto_linkmodes() 203 (!req_duplex || info->duplex == ksettings->base.duplex)) in ethnl_auto_linkmodes() 252 struct ethtool_link_settings *lsettings = &ksettings->base; in ethnl_update_linkmodes()
|
D | linkinfo.c | 7 struct ethnl_req_info base; member 11 struct ethnl_reply_data base; member 17 container_of(__reply_base, struct linkinfo_reply_data, base) 32 data->lsettings = &data->ksettings.base; in linkinfo_prepare_data() 130 lsettings = &ksettings.base; in ethnl_set_linkinfo()
|
D | common.c | 474 link_ksettings->base.speed in convert_legacy_settings_to_link_ksettings() 476 link_ksettings->base.duplex in convert_legacy_settings_to_link_ksettings() 478 link_ksettings->base.port in convert_legacy_settings_to_link_ksettings() 480 link_ksettings->base.phy_address in convert_legacy_settings_to_link_ksettings() 482 link_ksettings->base.autoneg in convert_legacy_settings_to_link_ksettings() 484 link_ksettings->base.mdio_support in convert_legacy_settings_to_link_ksettings() 486 link_ksettings->base.eth_tp_mdix in convert_legacy_settings_to_link_ksettings() 488 link_ksettings->base.eth_tp_mdix_ctrl in convert_legacy_settings_to_link_ksettings() 595 link_ksettings->base.speed = link_info->speed; in ethtool_params_from_link_mode() 597 link_ksettings->base.duplex = link_info->duplex; in ethtool_params_from_link_mode()
|
D | eeprom.c | 9 struct ethnl_req_info base; member 18 struct ethnl_reply_data base; member 24 container_of(__req_base, struct eeprom_req_info, base) 27 container_of(__reply_base, struct eeprom_reply_data, base) 57 struct net_device *dev = reply->base.dev; in eeprom_fallback()
|
D | ioctl.c | 407 ethtool_cmd_speed_set(legacy_settings, link_ksettings->base.speed); in convert_link_ksettings_to_legacy_settings() 409 = link_ksettings->base.duplex; in convert_link_ksettings_to_legacy_settings() 411 = link_ksettings->base.port; in convert_link_ksettings_to_legacy_settings() 413 = link_ksettings->base.phy_address; in convert_link_ksettings_to_legacy_settings() 415 = link_ksettings->base.autoneg; in convert_link_ksettings_to_legacy_settings() 417 = link_ksettings->base.mdio_support; in convert_link_ksettings_to_legacy_settings() 419 = link_ksettings->base.eth_tp_mdix; in convert_link_ksettings_to_legacy_settings() 421 = link_ksettings->base.eth_tp_mdix_ctrl; in convert_link_ksettings_to_legacy_settings() 423 = link_ksettings->base.transceiver; in convert_link_ksettings_to_legacy_settings() 433 struct ethtool_link_settings base; member [all …]
|
D | phc_vclocks.c | 9 struct ethnl_req_info base; member 13 struct ethnl_reply_data base; member 19 container_of(__reply_base, struct phc_vclocks_reply_data, base)
|
/net/bridge/ |
D | br_device.c | 266 cmd->base.duplex = DUPLEX_UNKNOWN; in br_get_link_ksettings() 267 cmd->base.port = PORT_OTHER; in br_get_link_ksettings() 268 cmd->base.speed = SPEED_UNKNOWN; in br_get_link_ksettings() 280 if (ecmd.base.speed == (__u32)SPEED_UNKNOWN) in br_get_link_ksettings() 283 if (cmd->base.speed == (__u32)SPEED_UNKNOWN || in br_get_link_ksettings() 284 cmd->base.speed < ecmd.base.speed) in br_get_link_ksettings() 285 cmd->base.speed = ecmd.base.speed; in br_get_link_ksettings()
|
/net/sched/ |
D | act_gate.c | 37 ktime_t now, base, cycle; in gate_get_start_time() local 40 base = ns_to_ktime(param->tcfg_basetime); in gate_get_start_time() 43 if (ktime_after(base, now)) { in gate_get_start_time() 44 *start = base; in gate_get_start_time() 50 n = div64_u64(ktime_sub_ns(now, base), cycle); in gate_get_start_time() 51 *start = ktime_add_ns(base, (n + 1) * cycle); in gate_get_start_time() 98 ktime_t cycle, base; in gate_timer_func() local 102 base = ns_to_ktime(p->tcfg_basetime); in gate_timer_func() 103 n = div64_u64(ktime_sub_ns(now, base), cycle); in gate_timer_func() 104 close_time = ktime_add_ns(base, (n + 1) * cycle); in gate_timer_func()
|
D | cls_basic.c | 146 struct basic_filter *f, unsigned long base, in basic_set_parms() argument 163 tcf_bind_filter(tp, &f->res, base); in basic_set_parms() 171 struct tcf_proto *tp, unsigned long base, u32 handle, in basic_change() argument 219 err = basic_set_parms(net, tp, fnew, base, tb, tca[TCA_RATE], flags, in basic_change() 260 unsigned long base) in basic_bind_class() argument 264 tc_cls_bind_class(classid, cl, q, &f->res, base); in basic_bind_class()
|
D | cls_fw.c | 201 struct nlattr **tca, unsigned long base, u32 flags, in fw_set_parms() argument 231 tcf_bind_filter(tp, &f->res, base); in fw_set_parms() 238 struct tcf_proto *tp, unsigned long base, in fw_change() argument 278 err = fw_set_parms(net, tp, fnew, tb, tca, base, flags, extack); in fw_change() 327 err = fw_set_parms(net, tp, f, tb, tca, base, flags, extack); in fw_change() 414 unsigned long base) in fw_bind_class() argument 418 tc_cls_bind_class(classid, cl, q, &f->res, base); in fw_bind_class()
|
/net/unix/ |
D | unix_bpf.c | 102 static void unix_dgram_bpf_rebuild_protos(struct proto *prot, const struct proto *base) in unix_dgram_bpf_rebuild_protos() argument 104 *prot = *base; in unix_dgram_bpf_rebuild_protos() 111 const struct proto *base) in unix_stream_bpf_rebuild_protos() argument 113 *prot = *base; in unix_stream_bpf_rebuild_protos()
|
/net/ipv6/ |
D | ip6_output.c | 1407 cork->base.dst = &rt->dst; in ip6_setup_cork() 1454 cork->base.fragsize = mtu; in ip6_setup_cork() 1455 cork->base.gso_size = ipc6->gso_size; in ip6_setup_cork() 1456 cork->base.tx_flags = 0; in ip6_setup_cork() 1457 cork->base.mark = ipc6->sockc.mark; in ip6_setup_cork() 1458 sock_tx_timestamp(sk, ipc6->sockc.tsflags, &cork->base.tx_flags); in ip6_setup_cork() 1461 cork->base.flags |= IPCORK_ALLFRAG; in ip6_setup_cork() 1462 cork->base.length = 0; in ip6_setup_cork() 1464 cork->base.transmit_time = ipc6->sockc.transmit_time; in ip6_setup_cork() 1480 struct inet_cork *cork = &cork_full->base; in __ip6_append_data() [all …]
|