/net/dsa/ |
D | port.c | 33 static int dsa_port_notify(const struct dsa_port *dp, unsigned long e, void *v) in dsa_port_notify() argument 35 return dsa_tree_notify(dp->ds->dst, e, v); in dsa_port_notify() 38 static void dsa_port_notify_bridge_fdb_flush(const struct dsa_port *dp, u16 vid) in dsa_port_notify_bridge_fdb_flush() argument 40 struct net_device *brport_dev = dsa_port_to_bridge_port(dp); in dsa_port_notify_bridge_fdb_flush() 55 static void dsa_port_fast_age(const struct dsa_port *dp) in dsa_port_fast_age() argument 57 struct dsa_switch *ds = dp->ds; in dsa_port_fast_age() 62 ds->ops->port_fast_age(ds, dp->index); in dsa_port_fast_age() 65 dsa_port_notify_bridge_fdb_flush(dp, 0); in dsa_port_fast_age() 68 static int dsa_port_vlan_fast_age(const struct dsa_port *dp, u16 vid) in dsa_port_vlan_fast_age() argument 70 struct dsa_switch *ds = dp->ds; in dsa_port_vlan_fast_age() [all …]
|
D | port.h | 18 bool dsa_port_supports_hwtstamp(struct dsa_port *dp); 21 int dsa_port_set_state(struct dsa_port *dp, u8 state, bool do_fast_age); 22 int dsa_port_set_mst_state(struct dsa_port *dp, 25 int dsa_port_enable_rt(struct dsa_port *dp, struct phy_device *phy); 26 int dsa_port_enable(struct dsa_port *dp, struct phy_device *phy); 27 void dsa_port_disable_rt(struct dsa_port *dp); 28 void dsa_port_disable(struct dsa_port *dp); 29 int dsa_port_bridge_join(struct dsa_port *dp, struct net_device *br, 31 void dsa_port_pre_bridge_leave(struct dsa_port *dp, struct net_device *br); 32 void dsa_port_bridge_leave(struct dsa_port *dp, struct net_device *br); [all …]
|
D | switch.c | 26 struct dsa_port *dp; in dsa_switch_fastest_ageing_time() local 28 dsa_switch_for_each_port(dp, ds) in dsa_switch_fastest_ageing_time() 29 if (dp->ageing_time && dp->ageing_time < ageing_time) in dsa_switch_fastest_ageing_time() 30 ageing_time = dp->ageing_time; in dsa_switch_fastest_ageing_time() 55 static bool dsa_port_mtu_match(struct dsa_port *dp, in dsa_port_mtu_match() argument 58 return dp == info->dp || dsa_port_is_dsa(dp) || dsa_port_is_cpu(dp); in dsa_port_mtu_match() 64 struct dsa_port *dp; in dsa_switch_mtu() local 70 dsa_switch_for_each_port(dp, ds) { in dsa_switch_mtu() 71 if (dsa_port_mtu_match(dp, info)) { in dsa_switch_mtu() 72 ret = ds->ops->port_change_mtu(ds, dp->index, in dsa_switch_mtu() [all …]
|
D | dsa.c | 108 struct dsa_port *dp; in dsa_tree_lag_find() local 110 list_for_each_entry(dp, &dst->ports, list) in dsa_tree_lag_find() 111 if (dsa_port_lag_dev_get(dp) == lag_dev) in dsa_tree_lag_find() 112 return dp->lag; in dsa_tree_lag_find() 120 struct dsa_port *dp; in dsa_tree_bridge_find() local 122 list_for_each_entry(dp, &dst->ports, list) in dsa_tree_bridge_find() 123 if (dsa_port_bridge_dev_get(dp) == br) in dsa_tree_bridge_find() 124 return dp->bridge; in dsa_tree_bridge_find() 183 struct dsa_port *dp; in dsa_switch_find() local 189 list_for_each_entry(dp, &dst->ports, list) { in dsa_switch_find() [all …]
|
D | slave.c | 88 struct dsa_port *dp = dsa_slave_to_port(dev); in dsa_slave_standalone_event_work() local 90 struct dsa_switch *ds = dp->ds; in dsa_slave_standalone_event_work() 96 err = dsa_port_standalone_host_fdb_add(dp, addr, vid); in dsa_slave_standalone_event_work() 100 dp->index, addr, vid, err); in dsa_slave_standalone_event_work() 106 err = dsa_port_standalone_host_fdb_del(dp, addr, vid); in dsa_slave_standalone_event_work() 110 dp->index, addr, vid, err); in dsa_slave_standalone_event_work() 118 err = dsa_port_standalone_host_mdb_add(dp, &mdb); in dsa_slave_standalone_event_work() 122 dp->index, addr, vid, err); in dsa_slave_standalone_event_work() 130 err = dsa_port_standalone_host_mdb_del(dp, &mdb); in dsa_slave_standalone_event_work() 134 dp->index, addr, vid, err); in dsa_slave_standalone_event_work() [all …]
|
D | tag_8021q.c | 92 u16 dsa_tag_8021q_standalone_vid(const struct dsa_port *dp) in dsa_tag_8021q_standalone_vid() argument 94 return DSA_8021Q_RSV | DSA_8021Q_SWITCH_ID(dp->ds->index) | in dsa_tag_8021q_standalone_vid() 95 DSA_8021Q_PORT(dp->index); in dsa_tag_8021q_standalone_vid() 142 static int dsa_port_do_tag_8021q_vlan_add(struct dsa_port *dp, u16 vid, in dsa_port_do_tag_8021q_vlan_add() argument 145 struct dsa_8021q_context *ctx = dp->ds->tag_8021q_ctx; in dsa_port_do_tag_8021q_vlan_add() 146 struct dsa_switch *ds = dp->ds; in dsa_port_do_tag_8021q_vlan_add() 148 int port = dp->index; in dsa_port_do_tag_8021q_vlan_add() 152 if (!(dsa_port_is_cpu(dp) || dsa_port_is_dsa(dp))) in dsa_port_do_tag_8021q_vlan_add() 179 static int dsa_port_do_tag_8021q_vlan_del(struct dsa_port *dp, u16 vid) in dsa_port_do_tag_8021q_vlan_del() argument 181 struct dsa_8021q_context *ctx = dp->ds->tag_8021q_ctx; in dsa_port_do_tag_8021q_vlan_del() [all …]
|
D | trace.h | 22 const char *dsa_port_kind(const struct dsa_port *dp); 26 TP_PROTO(const struct dsa_port *dp, const unsigned char *addr, u16 vid, 29 TP_ARGS(dp, addr, vid, db, err), 32 __string(dev, dev_name(dp->ds->dev)) 33 __string(kind, dsa_port_kind(dp)) 42 __assign_str(dev, dev_name(dp->ds->dev)); 43 __assign_str(kind, dsa_port_kind(dp)); 44 __entry->port = dp->index; 61 TP_PROTO(const struct dsa_port *dp, const unsigned char *addr, 63 TP_ARGS(dp, addr, vid, db, err)); [all …]
|
D | tag.h | 37 struct dsa_port *dp; in dsa_master_find_slave() local 39 list_for_each_entry(dp, &dst->ports, list) in dsa_master_find_slave() 40 if (dp->ds->index == device && dp->index == port && in dsa_master_find_slave() 41 dp->type == DSA_PORT_TYPE_USER) in dsa_master_find_slave() 42 return dp->slave; in dsa_master_find_slave() 52 struct dsa_port *dp = dsa_slave_to_port(skb->dev); in dsa_untag_bridge_pvid() local 53 struct net_device *br = dsa_port_bridge_dev_get(dp); in dsa_untag_bridge_pvid() 116 struct dsa_port *dp; in dsa_find_designated_bridge_port_by_vid() local 119 list_for_each_entry(dp, &dst->ports, list) { in dsa_find_designated_bridge_port_by_vid() 120 if (dp->type != DSA_PORT_TYPE_USER) in dsa_find_designated_bridge_port_by_vid() [all …]
|
D | tag_ksz.c | 122 struct dsa_port *dp = dsa_slave_to_port(dev); in ksz8795_xmit() local 133 *tag = 1 << dp->index; in ksz8795_xmit() 197 static void ksz_xmit_timestamp(struct dsa_port *dp, struct sk_buff *skb) in ksz_xmit_timestamp() argument 205 priv = ksz_tagger_private(dp->ds); in ksz_xmit_timestamp() 236 static struct sk_buff *ksz_defer_xmit(struct dsa_port *dp, struct sk_buff *skb) in ksz_defer_xmit() argument 238 struct ksz_tagger_data *tagger_data = ksz_tagger_data(dp->ds); in ksz_defer_xmit() 239 struct ksz_tagger_private *priv = ksz_tagger_private(dp->ds); in ksz_defer_xmit() 262 xmit_work->dp = dp; in ksz_defer_xmit() 275 struct dsa_port *dp = dsa_slave_to_port(dev); in ksz9477_xmit() local 284 ksz_xmit_timestamp(dp, skb); in ksz9477_xmit() [all …]
|
D | tag_sja1105.c | 140 static struct sk_buff *sja1105_defer_xmit(struct dsa_port *dp, in sja1105_defer_xmit() argument 143 struct sja1105_tagger_data *tagger_data = sja1105_tagger_data(dp->ds); in sja1105_defer_xmit() 144 struct sja1105_tagger_private *priv = sja1105_tagger_private(dp->ds); in sja1105_defer_xmit() 163 xmit_work->dp = dp; in sja1105_defer_xmit() 174 static u16 sja1105_xmit_tpid(struct dsa_port *dp) in sja1105_xmit_tpid() argument 176 struct dsa_switch *ds = dp->ds; in sja1105_xmit_tpid() 183 if (!dsa_port_is_vlan_filtering(dp)) in sja1105_xmit_tpid() 213 struct dsa_port *dp = dsa_slave_to_port(netdev); in sja1105_imprecise_xmit() local 214 unsigned int bridge_num = dsa_port_bridge_num_get(dp); in sja1105_imprecise_xmit() 215 struct net_device *br = dsa_port_bridge_dev_get(dp); in sja1105_imprecise_xmit() [all …]
|
D | tag_ocelot.c | 14 struct dsa_port *dp = dsa_slave_to_port(netdev); in ocelot_xmit_common() local 15 struct dsa_switch *ds = dp->ds; in ocelot_xmit_common() 22 ocelot_xmit_get_vlan_info(skb, dsa_port_bridge_dev_get(dp), &vlan_tci, in ocelot_xmit_common() 49 struct dsa_port *dp = dsa_slave_to_port(netdev); in ocelot_xmit() local 53 ocelot_ifh_set_dest(injection, BIT_ULL(dp->index)); in ocelot_xmit() 61 struct dsa_port *dp = dsa_slave_to_port(netdev); in seville_xmit() local 65 seville_ifh_set_dest(injection, BIT_ULL(dp->index)); in seville_xmit() 76 struct dsa_port *dp; in ocelot_rcv() local 140 dp = dsa_slave_to_port(skb->dev); in ocelot_rcv() 143 if (dsa_port_is_vlan_filtering(dp) && in ocelot_rcv()
|
D | devlink.c | 277 struct dsa_port *dp = dsa_to_port(ds, port); in dsa_devlink_port_region_create() local 279 return devlink_port_region_create(&dp->devlink_port, ops, in dsa_devlink_port_region_create() 291 int dsa_port_devlink_setup(struct dsa_port *dp) in dsa_port_devlink_setup() argument 293 struct devlink_port *dlp = &dp->devlink_port; in dsa_port_devlink_setup() 294 struct dsa_switch_tree *dst = dp->ds->dst; in dsa_port_devlink_setup() 296 struct devlink *dl = dp->ds->devlink; in dsa_port_devlink_setup() 297 struct dsa_switch *ds = dp->ds; in dsa_port_devlink_setup() 306 err = ds->ops->port_setup(ds, dp->index); in dsa_port_devlink_setup() 314 attrs.phys.port_number = dp->index; in dsa_port_devlink_setup() 318 switch (dp->type) { in dsa_port_devlink_setup() [all …]
|
D | switch.h | 47 const struct dsa_port *dp; member 55 const struct dsa_port *dp; member 71 const struct dsa_port *dp; member 78 const struct dsa_port *dp; member 86 const struct dsa_port *dp; member 93 const struct dsa_port *dp; member 104 const struct dsa_port *dp; member
|
/net/openvswitch/ |
D | datapath.c | 128 static int queue_gso_packets(struct datapath *dp, struct sk_buff *, 132 static int queue_userspace_packet(struct datapath *dp, struct sk_buff *, 142 const char *ovs_dp_name(const struct datapath *dp) in ovs_dp_name() argument 144 struct vport *vport = ovs_vport_ovsl_rcu(dp, OVSP_LOCAL); in ovs_dp_name() 148 static int get_dpifindex(const struct datapath *dp) in get_dpifindex() argument 155 local = ovs_vport_rcu(dp, OVSP_LOCAL); in get_dpifindex() 168 struct datapath *dp = container_of(rcu, struct datapath, rcu); in destroy_dp_rcu() local 170 ovs_flow_tbl_destroy(&dp->table); in destroy_dp_rcu() 171 free_percpu(dp->stats_percpu); in destroy_dp_rcu() 172 kfree(dp->ports); in destroy_dp_rcu() [all …]
|
D | datapath.h | 193 static inline struct net *ovs_dp_get_net(const struct datapath *dp) in ovs_dp_get_net() argument 195 return read_pnet(&dp->net); in ovs_dp_get_net() 198 static inline void ovs_dp_set_net(struct datapath *dp, struct net *net) in ovs_dp_set_net() argument 200 write_pnet(&dp->net, net); in ovs_dp_set_net() 203 struct vport *ovs_lookup_vport(const struct datapath *dp, u16 port_no); 205 static inline struct vport *ovs_vport_rcu(const struct datapath *dp, int port_no) in ovs_vport_rcu() argument 208 return ovs_lookup_vport(dp, port_no); in ovs_vport_rcu() 211 static inline struct vport *ovs_vport_ovsl_rcu(const struct datapath *dp, int port_no) in ovs_vport_ovsl_rcu() argument 214 return ovs_lookup_vport(dp, port_no); in ovs_vport_ovsl_rcu() 217 static inline struct vport *ovs_vport_ovsl(const struct datapath *dp, int port_no) in ovs_vport_ovsl() argument [all …]
|
D | dp_notify.c | 17 struct datapath *dp; in dp_detach_port_notify() local 19 dp = vport->dp; in dp_detach_port_notify() 20 notify = ovs_vport_cmd_build_info(vport, ovs_dp_get_net(dp), in dp_detach_port_notify() 24 genl_set_err(&dp_vport_genl_family, ovs_dp_get_net(dp), 0, in dp_detach_port_notify() 30 ovs_dp_get_net(dp), notify, 0, in dp_detach_port_notify() 37 struct datapath *dp; in ovs_dp_notify_wq() local 40 list_for_each_entry(dp, &ovs_net->dps, list_node) { in ovs_dp_notify_wq() 47 hlist_for_each_entry_safe(vport, n, &dp->ports[i], dp_hash_node) { in ovs_dp_notify_wq()
|
/net/dccp/ |
D | input.c | 172 const struct dccp_sock *dp = dccp_sk(sk); in dccp_deliver_input_to_ccids() local 176 ccid_hc_rx_packet_recv(dp->dccps_hc_rx_ccid, sk, skb); in dccp_deliver_input_to_ccids() 182 ccid_hc_tx_packet_recv(dp->dccps_hc_tx_ccid, sk, skb); in dccp_deliver_input_to_ccids() 188 struct dccp_sock *dp = dccp_sk(sk); in dccp_check_seqno() local 206 if (between48(ackno, dp->dccps_awl, dp->dccps_awh) && in dccp_check_seqno() 207 dccp_delta_seqno(dp->dccps_swl, seqno) >= 0) in dccp_check_seqno() 224 lswl = dp->dccps_swl; in dccp_check_seqno() 225 lawl = dp->dccps_awl; in dccp_check_seqno() 230 lswl = ADD48(dp->dccps_gsr, 1); in dccp_check_seqno() 231 lawl = dp->dccps_gar; in dccp_check_seqno() [all …]
|
D | proto.c | 176 struct dccp_sock *dp = dccp_sk(sk); in dccp_destruct_common() local 178 ccid_hc_tx_delete(dp->dccps_hc_tx_ccid, sk); in dccp_destruct_common() 179 dp->dccps_hc_tx_ccid = NULL; in dccp_destruct_common() 191 struct dccp_sock *dp = dccp_sk(sk); in dccp_init_sock() local 203 dp->dccps_mss_cache = 536; in dccp_init_sock() 204 dp->dccps_rate_last = jiffies; in dccp_init_sock() 205 dp->dccps_role = DCCP_ROLE_UNDEFINED; in dccp_init_sock() 206 dp->dccps_service = DCCP_SERVICE_CODE_IS_ABSENT; in dccp_init_sock() 207 dp->dccps_tx_qlen = sysctl_dccp_tx_qlen; in dccp_init_sock() 211 INIT_LIST_HEAD(&dp->dccps_featneg); in dccp_init_sock() [all …]
|
D | output.c | 47 struct dccp_sock *dp = dccp_sk(sk); in dccp_transmit_skb() local 55 u64 ackno = dp->dccps_gsr; in dccp_transmit_skb() 60 dcb->dccpd_seq = ADD48(dp->dccps_gss, 1); in dccp_transmit_skb() 74 dcb->dccpd_seq = dp->dccps_iss; in dccp_transmit_skb() 106 dh->dccph_cscov = dp->dccps_pcslen; in dccp_transmit_skb() 111 dccp_hdr_set_seq(dh, dp->dccps_gss); in dccp_transmit_skb() 118 dp->dccps_service; in dccp_transmit_skb() 123 dp->dccps_awl = dp->dccps_iss; in dccp_transmit_skb() 152 static u32 dccp_determine_ccmps(const struct dccp_sock *dp) in dccp_determine_ccmps() argument 154 const struct ccid *tx_ccid = dp->dccps_hc_tx_ccid; in dccp_determine_ccmps() [all …]
|
D | dccp.h | 255 int dccp_reqsk_init(struct request_sock *rq, struct dccp_sock const *dp, 315 const struct dccp_sock *dp = dccp_sk(sk); in dccp_bad_service_code() local 317 if (dp->dccps_service == service) in dccp_bad_service_code() 319 return !dccp_list_has_service(dp->dccps_service_list, service); in dccp_bad_service_code() 404 struct dccp_sock *dp = dccp_sk(sk); in dccp_update_gsr() local 406 if (after48(seq, dp->dccps_gsr)) in dccp_update_gsr() 407 dp->dccps_gsr = seq; in dccp_update_gsr() 409 dp->dccps_swl = SUB48(ADD48(dp->dccps_gsr, 1), dp->dccps_r_seq_win / 4); in dccp_update_gsr() 425 if (before48(dp->dccps_swl, dp->dccps_isr)) in dccp_update_gsr() 426 dp->dccps_swl = dp->dccps_isr; in dccp_update_gsr() [all …]
|
D | options.c | 51 struct dccp_sock *dp = dccp_sk(sk); in dccp_parse_options() local 58 struct dccp_options_received *opt_recv = &dp->dccps_options_received; in dccp_parse_options() 149 dp->dccps_timestamp_echo = ntohl(opt_val); in dccp_parse_options() 150 dp->dccps_timestamp_time = dccp_timestamp(); in dccp_parse_options() 215 if (ccid_hc_rx_parse_options(dp->dccps_hc_rx_ccid, sk, in dccp_parse_options() 230 if (ccid_hc_tx_parse_options(dp->dccps_hc_tx_ccid, sk, in dccp_parse_options() 311 struct dccp_sock *dp = dccp_sk(sk); in dccp_insert_option_ndp() local 312 u64 ndp = dp->dccps_ndp_count; in dccp_insert_option_ndp() 315 ++dp->dccps_ndp_count; in dccp_insert_option_ndp() 317 dp->dccps_ndp_count = 0; in dccp_insert_option_ndp() [all …]
|
D | feat.c | 38 struct dccp_sock *dp = dccp_sk(sk); in dccp_hdlr_ccid() local 45 ccid_hc_rx_delete(dp->dccps_hc_rx_ccid, sk); in dccp_hdlr_ccid() 46 dp->dccps_hc_rx_ccid = new_ccid; in dccp_hdlr_ccid() 48 ccid_hc_tx_delete(dp->dccps_hc_tx_ccid, sk); in dccp_hdlr_ccid() 49 dp->dccps_hc_tx_ccid = new_ccid; in dccp_hdlr_ccid() 56 struct dccp_sock *dp = dccp_sk(sk); in dccp_hdlr_seq_win() local 59 dp->dccps_r_seq_win = seq_win; in dccp_hdlr_seq_win() 61 dccp_update_gsr(sk, dp->dccps_gsr); in dccp_hdlr_seq_win() 63 dp->dccps_l_seq_win = seq_win; in dccp_hdlr_seq_win() 65 dccp_update_gss(sk, dp->dccps_gss); in dccp_hdlr_seq_win() [all …]
|
/net/rds/ |
D | ib_cm.c | 77 const union rds_ib_conn_priv *dp = NULL; in rds_ib_cm_connect_complete() local 84 dp = event->param.conn.private_data; in rds_ib_cm_connect_complete() 88 major = dp->ricp_v6.dp_protocol_major; in rds_ib_cm_connect_complete() 89 minor = dp->ricp_v6.dp_protocol_minor; in rds_ib_cm_connect_complete() 90 credit = dp->ricp_v6.dp_credit; in rds_ib_cm_connect_complete() 96 ack_seq = get_unaligned(&dp->ricp_v6.dp_ack_seq); in rds_ib_cm_connect_complete() 100 major = dp->ricp_v4.dp_protocol_major; in rds_ib_cm_connect_complete() 101 minor = dp->ricp_v4.dp_protocol_minor; in rds_ib_cm_connect_complete() 102 credit = dp->ricp_v4.dp_credit; in rds_ib_cm_connect_complete() 103 ack_seq = get_unaligned(&dp->ricp_v4.dp_ack_seq); in rds_ib_cm_connect_complete() [all …]
|
/net/ax25/ |
D | ax25_in.c | 189 ax25_digi dp, reverse_dp; in ax25_rcv() local 206 if (ax25_addr_parse(skb->data, skb->len, &src, &dest, &dp, &type, &dama) == NULL) in ax25_rcv() 212 if (dp.lastrepeat + 1 < dp.ndigi) /* Not yet digipeated completely */ in ax25_rcv() 213 next_digi = &dp.calls[dp.lastrepeat + 1]; in ax25_rcv() 218 skb_pull(skb, ax25_addr_size(&dp)); in ax25_rcv() 221 if (ax25cmp(&dest, dev_addr) == 0 && dp.lastrepeat + 1 == dp.ndigi) in ax25_rcv() 225 if (!mine && ax25_listen_mine(&dest, dev) && dp.lastrepeat + 1 == dp.ndigi) in ax25_rcv() 229 if ((*skb->data & ~0x10) == AX25_UI && dp.lastrepeat + 1 == dp.ndigi) { in ax25_rcv() 301 ax25_digi_invert(&dp, &reverse_dp); in ax25_rcv() 328 ax25_return_dm(dev, &src, &dest, &dp); in ax25_rcv() [all …]
|
/net/dccp/ccids/ |
D | ccid2.c | 240 struct dccp_sock *dp = dccp_sk(sk); in ccid2_hc_tx_packet_sent() local 268 hc->tx_seqh->ccid2s_seq = dp->dccps_gss; in ccid2_hc_tx_packet_sent() 318 if (dp->dccps_l_ack_ratio > 1) { in ccid2_hc_tx_packet_sent() 320 int denom = dp->dccps_l_ack_ratio * dp->dccps_l_ack_ratio - in ccid2_hc_tx_packet_sent() 321 dp->dccps_l_ack_ratio; in ccid2_hc_tx_packet_sent() 326 ccid2_change_l_ack_ratio(sk, dp->dccps_l_ack_ratio - 1); in ccid2_hc_tx_packet_sent() 438 struct dccp_sock *dp = dccp_sk(sk); in ccid2_new_ack() local 439 int r_seq_used = hc->tx_cwnd / dp->dccps_l_ack_ratio; in ccid2_new_ack() 441 if (hc->tx_cwnd < dp->dccps_l_seq_win && in ccid2_new_ack() 442 r_seq_used < dp->dccps_r_seq_win) { in ccid2_new_ack() [all …]
|