Home
last modified time | relevance | path

Searched refs:dp (Results 1 – 25 of 69) sorted by relevance

123

/net/dsa/
Dport.c33 static int dsa_port_notify(const struct dsa_port *dp, unsigned long e, void *v) in dsa_port_notify() argument
35 return dsa_tree_notify(dp->ds->dst, e, v); in dsa_port_notify()
38 static void dsa_port_notify_bridge_fdb_flush(const struct dsa_port *dp, u16 vid) in dsa_port_notify_bridge_fdb_flush() argument
40 struct net_device *brport_dev = dsa_port_to_bridge_port(dp); in dsa_port_notify_bridge_fdb_flush()
55 static void dsa_port_fast_age(const struct dsa_port *dp) in dsa_port_fast_age() argument
57 struct dsa_switch *ds = dp->ds; in dsa_port_fast_age()
62 ds->ops->port_fast_age(ds, dp->index); in dsa_port_fast_age()
65 dsa_port_notify_bridge_fdb_flush(dp, 0); in dsa_port_fast_age()
68 static int dsa_port_vlan_fast_age(const struct dsa_port *dp, u16 vid) in dsa_port_vlan_fast_age() argument
70 struct dsa_switch *ds = dp->ds; in dsa_port_vlan_fast_age()
[all …]
Dport.h18 bool dsa_port_supports_hwtstamp(struct dsa_port *dp);
21 int dsa_port_set_state(struct dsa_port *dp, u8 state, bool do_fast_age);
22 int dsa_port_set_mst_state(struct dsa_port *dp,
25 int dsa_port_enable_rt(struct dsa_port *dp, struct phy_device *phy);
26 int dsa_port_enable(struct dsa_port *dp, struct phy_device *phy);
27 void dsa_port_disable_rt(struct dsa_port *dp);
28 void dsa_port_disable(struct dsa_port *dp);
29 int dsa_port_bridge_join(struct dsa_port *dp, struct net_device *br,
31 void dsa_port_pre_bridge_leave(struct dsa_port *dp, struct net_device *br);
32 void dsa_port_bridge_leave(struct dsa_port *dp, struct net_device *br);
[all …]
Dswitch.c26 struct dsa_port *dp; in dsa_switch_fastest_ageing_time() local
28 dsa_switch_for_each_port(dp, ds) in dsa_switch_fastest_ageing_time()
29 if (dp->ageing_time && dp->ageing_time < ageing_time) in dsa_switch_fastest_ageing_time()
30 ageing_time = dp->ageing_time; in dsa_switch_fastest_ageing_time()
55 static bool dsa_port_mtu_match(struct dsa_port *dp, in dsa_port_mtu_match() argument
58 return dp == info->dp || dsa_port_is_dsa(dp) || dsa_port_is_cpu(dp); in dsa_port_mtu_match()
64 struct dsa_port *dp; in dsa_switch_mtu() local
70 dsa_switch_for_each_port(dp, ds) { in dsa_switch_mtu()
71 if (dsa_port_mtu_match(dp, info)) { in dsa_switch_mtu()
72 ret = ds->ops->port_change_mtu(ds, dp->index, in dsa_switch_mtu()
[all …]
Ddsa.c108 struct dsa_port *dp; in dsa_tree_lag_find() local
110 list_for_each_entry(dp, &dst->ports, list) in dsa_tree_lag_find()
111 if (dsa_port_lag_dev_get(dp) == lag_dev) in dsa_tree_lag_find()
112 return dp->lag; in dsa_tree_lag_find()
120 struct dsa_port *dp; in dsa_tree_bridge_find() local
122 list_for_each_entry(dp, &dst->ports, list) in dsa_tree_bridge_find()
123 if (dsa_port_bridge_dev_get(dp) == br) in dsa_tree_bridge_find()
124 return dp->bridge; in dsa_tree_bridge_find()
183 struct dsa_port *dp; in dsa_switch_find() local
189 list_for_each_entry(dp, &dst->ports, list) { in dsa_switch_find()
[all …]
Dslave.c88 struct dsa_port *dp = dsa_slave_to_port(dev); in dsa_slave_standalone_event_work() local
90 struct dsa_switch *ds = dp->ds; in dsa_slave_standalone_event_work()
96 err = dsa_port_standalone_host_fdb_add(dp, addr, vid); in dsa_slave_standalone_event_work()
100 dp->index, addr, vid, err); in dsa_slave_standalone_event_work()
106 err = dsa_port_standalone_host_fdb_del(dp, addr, vid); in dsa_slave_standalone_event_work()
110 dp->index, addr, vid, err); in dsa_slave_standalone_event_work()
118 err = dsa_port_standalone_host_mdb_add(dp, &mdb); in dsa_slave_standalone_event_work()
122 dp->index, addr, vid, err); in dsa_slave_standalone_event_work()
130 err = dsa_port_standalone_host_mdb_del(dp, &mdb); in dsa_slave_standalone_event_work()
134 dp->index, addr, vid, err); in dsa_slave_standalone_event_work()
[all …]
Dtag_8021q.c92 u16 dsa_tag_8021q_standalone_vid(const struct dsa_port *dp) in dsa_tag_8021q_standalone_vid() argument
94 return DSA_8021Q_RSV | DSA_8021Q_SWITCH_ID(dp->ds->index) | in dsa_tag_8021q_standalone_vid()
95 DSA_8021Q_PORT(dp->index); in dsa_tag_8021q_standalone_vid()
142 static int dsa_port_do_tag_8021q_vlan_add(struct dsa_port *dp, u16 vid, in dsa_port_do_tag_8021q_vlan_add() argument
145 struct dsa_8021q_context *ctx = dp->ds->tag_8021q_ctx; in dsa_port_do_tag_8021q_vlan_add()
146 struct dsa_switch *ds = dp->ds; in dsa_port_do_tag_8021q_vlan_add()
148 int port = dp->index; in dsa_port_do_tag_8021q_vlan_add()
152 if (!(dsa_port_is_cpu(dp) || dsa_port_is_dsa(dp))) in dsa_port_do_tag_8021q_vlan_add()
179 static int dsa_port_do_tag_8021q_vlan_del(struct dsa_port *dp, u16 vid) in dsa_port_do_tag_8021q_vlan_del() argument
181 struct dsa_8021q_context *ctx = dp->ds->tag_8021q_ctx; in dsa_port_do_tag_8021q_vlan_del()
[all …]
Dtrace.h22 const char *dsa_port_kind(const struct dsa_port *dp);
26 TP_PROTO(const struct dsa_port *dp, const unsigned char *addr, u16 vid,
29 TP_ARGS(dp, addr, vid, db, err),
32 __string(dev, dev_name(dp->ds->dev))
33 __string(kind, dsa_port_kind(dp))
42 __assign_str(dev, dev_name(dp->ds->dev));
43 __assign_str(kind, dsa_port_kind(dp));
44 __entry->port = dp->index;
61 TP_PROTO(const struct dsa_port *dp, const unsigned char *addr,
63 TP_ARGS(dp, addr, vid, db, err));
[all …]
Dtag.h37 struct dsa_port *dp; in dsa_master_find_slave() local
39 list_for_each_entry(dp, &dst->ports, list) in dsa_master_find_slave()
40 if (dp->ds->index == device && dp->index == port && in dsa_master_find_slave()
41 dp->type == DSA_PORT_TYPE_USER) in dsa_master_find_slave()
42 return dp->slave; in dsa_master_find_slave()
52 struct dsa_port *dp = dsa_slave_to_port(skb->dev); in dsa_untag_bridge_pvid() local
53 struct net_device *br = dsa_port_bridge_dev_get(dp); in dsa_untag_bridge_pvid()
116 struct dsa_port *dp; in dsa_find_designated_bridge_port_by_vid() local
119 list_for_each_entry(dp, &dst->ports, list) { in dsa_find_designated_bridge_port_by_vid()
120 if (dp->type != DSA_PORT_TYPE_USER) in dsa_find_designated_bridge_port_by_vid()
[all …]
Dtag_ksz.c122 struct dsa_port *dp = dsa_slave_to_port(dev); in ksz8795_xmit() local
133 *tag = 1 << dp->index; in ksz8795_xmit()
197 static void ksz_xmit_timestamp(struct dsa_port *dp, struct sk_buff *skb) in ksz_xmit_timestamp() argument
205 priv = ksz_tagger_private(dp->ds); in ksz_xmit_timestamp()
236 static struct sk_buff *ksz_defer_xmit(struct dsa_port *dp, struct sk_buff *skb) in ksz_defer_xmit() argument
238 struct ksz_tagger_data *tagger_data = ksz_tagger_data(dp->ds); in ksz_defer_xmit()
239 struct ksz_tagger_private *priv = ksz_tagger_private(dp->ds); in ksz_defer_xmit()
262 xmit_work->dp = dp; in ksz_defer_xmit()
275 struct dsa_port *dp = dsa_slave_to_port(dev); in ksz9477_xmit() local
284 ksz_xmit_timestamp(dp, skb); in ksz9477_xmit()
[all …]
Dtag_sja1105.c140 static struct sk_buff *sja1105_defer_xmit(struct dsa_port *dp, in sja1105_defer_xmit() argument
143 struct sja1105_tagger_data *tagger_data = sja1105_tagger_data(dp->ds); in sja1105_defer_xmit()
144 struct sja1105_tagger_private *priv = sja1105_tagger_private(dp->ds); in sja1105_defer_xmit()
163 xmit_work->dp = dp; in sja1105_defer_xmit()
174 static u16 sja1105_xmit_tpid(struct dsa_port *dp) in sja1105_xmit_tpid() argument
176 struct dsa_switch *ds = dp->ds; in sja1105_xmit_tpid()
183 if (!dsa_port_is_vlan_filtering(dp)) in sja1105_xmit_tpid()
213 struct dsa_port *dp = dsa_slave_to_port(netdev); in sja1105_imprecise_xmit() local
214 unsigned int bridge_num = dsa_port_bridge_num_get(dp); in sja1105_imprecise_xmit()
215 struct net_device *br = dsa_port_bridge_dev_get(dp); in sja1105_imprecise_xmit()
[all …]
Dtag_ocelot.c14 struct dsa_port *dp = dsa_slave_to_port(netdev); in ocelot_xmit_common() local
15 struct dsa_switch *ds = dp->ds; in ocelot_xmit_common()
22 ocelot_xmit_get_vlan_info(skb, dsa_port_bridge_dev_get(dp), &vlan_tci, in ocelot_xmit_common()
49 struct dsa_port *dp = dsa_slave_to_port(netdev); in ocelot_xmit() local
53 ocelot_ifh_set_dest(injection, BIT_ULL(dp->index)); in ocelot_xmit()
61 struct dsa_port *dp = dsa_slave_to_port(netdev); in seville_xmit() local
65 seville_ifh_set_dest(injection, BIT_ULL(dp->index)); in seville_xmit()
76 struct dsa_port *dp; in ocelot_rcv() local
140 dp = dsa_slave_to_port(skb->dev); in ocelot_rcv()
143 if (dsa_port_is_vlan_filtering(dp) && in ocelot_rcv()
Ddevlink.c277 struct dsa_port *dp = dsa_to_port(ds, port); in dsa_devlink_port_region_create() local
279 return devlink_port_region_create(&dp->devlink_port, ops, in dsa_devlink_port_region_create()
291 int dsa_port_devlink_setup(struct dsa_port *dp) in dsa_port_devlink_setup() argument
293 struct devlink_port *dlp = &dp->devlink_port; in dsa_port_devlink_setup()
294 struct dsa_switch_tree *dst = dp->ds->dst; in dsa_port_devlink_setup()
296 struct devlink *dl = dp->ds->devlink; in dsa_port_devlink_setup()
297 struct dsa_switch *ds = dp->ds; in dsa_port_devlink_setup()
306 err = ds->ops->port_setup(ds, dp->index); in dsa_port_devlink_setup()
314 attrs.phys.port_number = dp->index; in dsa_port_devlink_setup()
318 switch (dp->type) { in dsa_port_devlink_setup()
[all …]
Dswitch.h47 const struct dsa_port *dp; member
55 const struct dsa_port *dp; member
71 const struct dsa_port *dp; member
78 const struct dsa_port *dp; member
86 const struct dsa_port *dp; member
93 const struct dsa_port *dp; member
104 const struct dsa_port *dp; member
/net/openvswitch/
Ddatapath.c128 static int queue_gso_packets(struct datapath *dp, struct sk_buff *,
132 static int queue_userspace_packet(struct datapath *dp, struct sk_buff *,
142 const char *ovs_dp_name(const struct datapath *dp) in ovs_dp_name() argument
144 struct vport *vport = ovs_vport_ovsl_rcu(dp, OVSP_LOCAL); in ovs_dp_name()
148 static int get_dpifindex(const struct datapath *dp) in get_dpifindex() argument
155 local = ovs_vport_rcu(dp, OVSP_LOCAL); in get_dpifindex()
168 struct datapath *dp = container_of(rcu, struct datapath, rcu); in destroy_dp_rcu() local
170 ovs_flow_tbl_destroy(&dp->table); in destroy_dp_rcu()
171 free_percpu(dp->stats_percpu); in destroy_dp_rcu()
172 kfree(dp->ports); in destroy_dp_rcu()
[all …]
Ddatapath.h193 static inline struct net *ovs_dp_get_net(const struct datapath *dp) in ovs_dp_get_net() argument
195 return read_pnet(&dp->net); in ovs_dp_get_net()
198 static inline void ovs_dp_set_net(struct datapath *dp, struct net *net) in ovs_dp_set_net() argument
200 write_pnet(&dp->net, net); in ovs_dp_set_net()
203 struct vport *ovs_lookup_vport(const struct datapath *dp, u16 port_no);
205 static inline struct vport *ovs_vport_rcu(const struct datapath *dp, int port_no) in ovs_vport_rcu() argument
208 return ovs_lookup_vport(dp, port_no); in ovs_vport_rcu()
211 static inline struct vport *ovs_vport_ovsl_rcu(const struct datapath *dp, int port_no) in ovs_vport_ovsl_rcu() argument
214 return ovs_lookup_vport(dp, port_no); in ovs_vport_ovsl_rcu()
217 static inline struct vport *ovs_vport_ovsl(const struct datapath *dp, int port_no) in ovs_vport_ovsl() argument
[all …]
Ddp_notify.c17 struct datapath *dp; in dp_detach_port_notify() local
19 dp = vport->dp; in dp_detach_port_notify()
20 notify = ovs_vport_cmd_build_info(vport, ovs_dp_get_net(dp), in dp_detach_port_notify()
24 genl_set_err(&dp_vport_genl_family, ovs_dp_get_net(dp), 0, in dp_detach_port_notify()
30 ovs_dp_get_net(dp), notify, 0, in dp_detach_port_notify()
37 struct datapath *dp; in ovs_dp_notify_wq() local
40 list_for_each_entry(dp, &ovs_net->dps, list_node) { in ovs_dp_notify_wq()
47 hlist_for_each_entry_safe(vport, n, &dp->ports[i], dp_hash_node) { in ovs_dp_notify_wq()
/net/dccp/
Dinput.c172 const struct dccp_sock *dp = dccp_sk(sk); in dccp_deliver_input_to_ccids() local
176 ccid_hc_rx_packet_recv(dp->dccps_hc_rx_ccid, sk, skb); in dccp_deliver_input_to_ccids()
182 ccid_hc_tx_packet_recv(dp->dccps_hc_tx_ccid, sk, skb); in dccp_deliver_input_to_ccids()
188 struct dccp_sock *dp = dccp_sk(sk); in dccp_check_seqno() local
206 if (between48(ackno, dp->dccps_awl, dp->dccps_awh) && in dccp_check_seqno()
207 dccp_delta_seqno(dp->dccps_swl, seqno) >= 0) in dccp_check_seqno()
224 lswl = dp->dccps_swl; in dccp_check_seqno()
225 lawl = dp->dccps_awl; in dccp_check_seqno()
230 lswl = ADD48(dp->dccps_gsr, 1); in dccp_check_seqno()
231 lawl = dp->dccps_gar; in dccp_check_seqno()
[all …]
Dproto.c176 struct dccp_sock *dp = dccp_sk(sk); in dccp_destruct_common() local
178 ccid_hc_tx_delete(dp->dccps_hc_tx_ccid, sk); in dccp_destruct_common()
179 dp->dccps_hc_tx_ccid = NULL; in dccp_destruct_common()
191 struct dccp_sock *dp = dccp_sk(sk); in dccp_init_sock() local
203 dp->dccps_mss_cache = 536; in dccp_init_sock()
204 dp->dccps_rate_last = jiffies; in dccp_init_sock()
205 dp->dccps_role = DCCP_ROLE_UNDEFINED; in dccp_init_sock()
206 dp->dccps_service = DCCP_SERVICE_CODE_IS_ABSENT; in dccp_init_sock()
207 dp->dccps_tx_qlen = sysctl_dccp_tx_qlen; in dccp_init_sock()
211 INIT_LIST_HEAD(&dp->dccps_featneg); in dccp_init_sock()
[all …]
Doutput.c47 struct dccp_sock *dp = dccp_sk(sk); in dccp_transmit_skb() local
55 u64 ackno = dp->dccps_gsr; in dccp_transmit_skb()
60 dcb->dccpd_seq = ADD48(dp->dccps_gss, 1); in dccp_transmit_skb()
74 dcb->dccpd_seq = dp->dccps_iss; in dccp_transmit_skb()
106 dh->dccph_cscov = dp->dccps_pcslen; in dccp_transmit_skb()
111 dccp_hdr_set_seq(dh, dp->dccps_gss); in dccp_transmit_skb()
118 dp->dccps_service; in dccp_transmit_skb()
123 dp->dccps_awl = dp->dccps_iss; in dccp_transmit_skb()
152 static u32 dccp_determine_ccmps(const struct dccp_sock *dp) in dccp_determine_ccmps() argument
154 const struct ccid *tx_ccid = dp->dccps_hc_tx_ccid; in dccp_determine_ccmps()
[all …]
Ddccp.h255 int dccp_reqsk_init(struct request_sock *rq, struct dccp_sock const *dp,
315 const struct dccp_sock *dp = dccp_sk(sk); in dccp_bad_service_code() local
317 if (dp->dccps_service == service) in dccp_bad_service_code()
319 return !dccp_list_has_service(dp->dccps_service_list, service); in dccp_bad_service_code()
404 struct dccp_sock *dp = dccp_sk(sk); in dccp_update_gsr() local
406 if (after48(seq, dp->dccps_gsr)) in dccp_update_gsr()
407 dp->dccps_gsr = seq; in dccp_update_gsr()
409 dp->dccps_swl = SUB48(ADD48(dp->dccps_gsr, 1), dp->dccps_r_seq_win / 4); in dccp_update_gsr()
425 if (before48(dp->dccps_swl, dp->dccps_isr)) in dccp_update_gsr()
426 dp->dccps_swl = dp->dccps_isr; in dccp_update_gsr()
[all …]
Doptions.c51 struct dccp_sock *dp = dccp_sk(sk); in dccp_parse_options() local
58 struct dccp_options_received *opt_recv = &dp->dccps_options_received; in dccp_parse_options()
149 dp->dccps_timestamp_echo = ntohl(opt_val); in dccp_parse_options()
150 dp->dccps_timestamp_time = dccp_timestamp(); in dccp_parse_options()
215 if (ccid_hc_rx_parse_options(dp->dccps_hc_rx_ccid, sk, in dccp_parse_options()
230 if (ccid_hc_tx_parse_options(dp->dccps_hc_tx_ccid, sk, in dccp_parse_options()
311 struct dccp_sock *dp = dccp_sk(sk); in dccp_insert_option_ndp() local
312 u64 ndp = dp->dccps_ndp_count; in dccp_insert_option_ndp()
315 ++dp->dccps_ndp_count; in dccp_insert_option_ndp()
317 dp->dccps_ndp_count = 0; in dccp_insert_option_ndp()
[all …]
Dfeat.c38 struct dccp_sock *dp = dccp_sk(sk); in dccp_hdlr_ccid() local
45 ccid_hc_rx_delete(dp->dccps_hc_rx_ccid, sk); in dccp_hdlr_ccid()
46 dp->dccps_hc_rx_ccid = new_ccid; in dccp_hdlr_ccid()
48 ccid_hc_tx_delete(dp->dccps_hc_tx_ccid, sk); in dccp_hdlr_ccid()
49 dp->dccps_hc_tx_ccid = new_ccid; in dccp_hdlr_ccid()
56 struct dccp_sock *dp = dccp_sk(sk); in dccp_hdlr_seq_win() local
59 dp->dccps_r_seq_win = seq_win; in dccp_hdlr_seq_win()
61 dccp_update_gsr(sk, dp->dccps_gsr); in dccp_hdlr_seq_win()
63 dp->dccps_l_seq_win = seq_win; in dccp_hdlr_seq_win()
65 dccp_update_gss(sk, dp->dccps_gss); in dccp_hdlr_seq_win()
[all …]
/net/rds/
Dib_cm.c77 const union rds_ib_conn_priv *dp = NULL; in rds_ib_cm_connect_complete() local
84 dp = event->param.conn.private_data; in rds_ib_cm_connect_complete()
88 major = dp->ricp_v6.dp_protocol_major; in rds_ib_cm_connect_complete()
89 minor = dp->ricp_v6.dp_protocol_minor; in rds_ib_cm_connect_complete()
90 credit = dp->ricp_v6.dp_credit; in rds_ib_cm_connect_complete()
96 ack_seq = get_unaligned(&dp->ricp_v6.dp_ack_seq); in rds_ib_cm_connect_complete()
100 major = dp->ricp_v4.dp_protocol_major; in rds_ib_cm_connect_complete()
101 minor = dp->ricp_v4.dp_protocol_minor; in rds_ib_cm_connect_complete()
102 credit = dp->ricp_v4.dp_credit; in rds_ib_cm_connect_complete()
103 ack_seq = get_unaligned(&dp->ricp_v4.dp_ack_seq); in rds_ib_cm_connect_complete()
[all …]
/net/ax25/
Dax25_in.c189 ax25_digi dp, reverse_dp; in ax25_rcv() local
206 if (ax25_addr_parse(skb->data, skb->len, &src, &dest, &dp, &type, &dama) == NULL) in ax25_rcv()
212 if (dp.lastrepeat + 1 < dp.ndigi) /* Not yet digipeated completely */ in ax25_rcv()
213 next_digi = &dp.calls[dp.lastrepeat + 1]; in ax25_rcv()
218 skb_pull(skb, ax25_addr_size(&dp)); in ax25_rcv()
221 if (ax25cmp(&dest, dev_addr) == 0 && dp.lastrepeat + 1 == dp.ndigi) in ax25_rcv()
225 if (!mine && ax25_listen_mine(&dest, dev) && dp.lastrepeat + 1 == dp.ndigi) in ax25_rcv()
229 if ((*skb->data & ~0x10) == AX25_UI && dp.lastrepeat + 1 == dp.ndigi) { in ax25_rcv()
301 ax25_digi_invert(&dp, &reverse_dp); in ax25_rcv()
328 ax25_return_dm(dev, &src, &dest, &dp); in ax25_rcv()
[all …]
/net/dccp/ccids/
Dccid2.c240 struct dccp_sock *dp = dccp_sk(sk); in ccid2_hc_tx_packet_sent() local
268 hc->tx_seqh->ccid2s_seq = dp->dccps_gss; in ccid2_hc_tx_packet_sent()
318 if (dp->dccps_l_ack_ratio > 1) { in ccid2_hc_tx_packet_sent()
320 int denom = dp->dccps_l_ack_ratio * dp->dccps_l_ack_ratio - in ccid2_hc_tx_packet_sent()
321 dp->dccps_l_ack_ratio; in ccid2_hc_tx_packet_sent()
326 ccid2_change_l_ack_ratio(sk, dp->dccps_l_ack_ratio - 1); in ccid2_hc_tx_packet_sent()
438 struct dccp_sock *dp = dccp_sk(sk); in ccid2_new_ack() local
439 int r_seq_used = hc->tx_cwnd / dp->dccps_l_ack_ratio; in ccid2_new_ack()
441 if (hc->tx_cwnd < dp->dccps_l_seq_win && in ccid2_new_ack()
442 r_seq_used < dp->dccps_r_seq_win) { in ccid2_new_ack()
[all …]

123