1 // SPDX-License-Identifier: GPL-2.0
2 #include <linux/kernel.h>
3 #include <linux/list.h>
4 #include <linux/netdevice.h>
5 #include <linux/rtnetlink.h>
6 #include <linux/skbuff.h>
7 #include <net/ip.h>
8 #include <net/switchdev.h>
9
10 #include "br_private.h"
11
12 static struct static_key_false br_switchdev_tx_fwd_offload;
13
nbp_switchdev_can_offload_tx_fwd(const struct net_bridge_port * p,const struct sk_buff * skb)14 static bool nbp_switchdev_can_offload_tx_fwd(const struct net_bridge_port *p,
15 const struct sk_buff *skb)
16 {
17 if (!static_branch_unlikely(&br_switchdev_tx_fwd_offload))
18 return false;
19
20 if (br_multicast_igmp_type(skb))
21 return false;
22
23 return (p->flags & BR_TX_FWD_OFFLOAD) &&
24 (p->hwdom != BR_INPUT_SKB_CB(skb)->src_hwdom);
25 }
26
br_switchdev_frame_uses_tx_fwd_offload(struct sk_buff * skb)27 bool br_switchdev_frame_uses_tx_fwd_offload(struct sk_buff *skb)
28 {
29 if (!static_branch_unlikely(&br_switchdev_tx_fwd_offload))
30 return false;
31
32 return BR_INPUT_SKB_CB(skb)->tx_fwd_offload;
33 }
34
br_switchdev_frame_set_offload_fwd_mark(struct sk_buff * skb)35 void br_switchdev_frame_set_offload_fwd_mark(struct sk_buff *skb)
36 {
37 skb->offload_fwd_mark = br_switchdev_frame_uses_tx_fwd_offload(skb);
38 }
39
40 /* Mark the frame for TX forwarding offload if this egress port supports it */
nbp_switchdev_frame_mark_tx_fwd_offload(const struct net_bridge_port * p,struct sk_buff * skb)41 void nbp_switchdev_frame_mark_tx_fwd_offload(const struct net_bridge_port *p,
42 struct sk_buff *skb)
43 {
44 if (nbp_switchdev_can_offload_tx_fwd(p, skb))
45 BR_INPUT_SKB_CB(skb)->tx_fwd_offload = true;
46 }
47
48 /* Lazily adds the hwdom of the egress bridge port to the bit mask of hwdoms
49 * that the skb has been already forwarded to, to avoid further cloning to
50 * other ports in the same hwdom by making nbp_switchdev_allowed_egress()
51 * return false.
52 */
nbp_switchdev_frame_mark_tx_fwd_to_hwdom(const struct net_bridge_port * p,struct sk_buff * skb)53 void nbp_switchdev_frame_mark_tx_fwd_to_hwdom(const struct net_bridge_port *p,
54 struct sk_buff *skb)
55 {
56 if (nbp_switchdev_can_offload_tx_fwd(p, skb))
57 set_bit(p->hwdom, &BR_INPUT_SKB_CB(skb)->fwd_hwdoms);
58 }
59
nbp_switchdev_frame_mark(const struct net_bridge_port * p,struct sk_buff * skb)60 void nbp_switchdev_frame_mark(const struct net_bridge_port *p,
61 struct sk_buff *skb)
62 {
63 if (p->hwdom)
64 BR_INPUT_SKB_CB(skb)->src_hwdom = p->hwdom;
65 }
66
nbp_switchdev_allowed_egress(const struct net_bridge_port * p,const struct sk_buff * skb)67 bool nbp_switchdev_allowed_egress(const struct net_bridge_port *p,
68 const struct sk_buff *skb)
69 {
70 struct br_input_skb_cb *cb = BR_INPUT_SKB_CB(skb);
71
72 return !test_bit(p->hwdom, &cb->fwd_hwdoms) &&
73 (!skb->offload_fwd_mark || cb->src_hwdom != p->hwdom);
74 }
75
76 /* Flags that can be offloaded to hardware */
77 #define BR_PORT_FLAGS_HW_OFFLOAD (BR_LEARNING | BR_FLOOD | BR_PORT_MAB | \
78 BR_MCAST_FLOOD | BR_BCAST_FLOOD | BR_PORT_LOCKED | \
79 BR_HAIRPIN_MODE | BR_ISOLATED | BR_MULTICAST_TO_UNICAST)
80
br_switchdev_set_port_flag(struct net_bridge_port * p,unsigned long flags,unsigned long mask,struct netlink_ext_ack * extack)81 int br_switchdev_set_port_flag(struct net_bridge_port *p,
82 unsigned long flags,
83 unsigned long mask,
84 struct netlink_ext_ack *extack)
85 {
86 struct switchdev_attr attr = {
87 .orig_dev = p->dev,
88 };
89 struct switchdev_notifier_port_attr_info info = {
90 .attr = &attr,
91 };
92 int err;
93
94 mask &= BR_PORT_FLAGS_HW_OFFLOAD;
95 if (!mask)
96 return 0;
97
98 attr.id = SWITCHDEV_ATTR_ID_PORT_PRE_BRIDGE_FLAGS;
99 attr.u.brport_flags.val = flags;
100 attr.u.brport_flags.mask = mask;
101
102 /* We run from atomic context here */
103 err = call_switchdev_notifiers(SWITCHDEV_PORT_ATTR_SET, p->dev,
104 &info.info, extack);
105 err = notifier_to_errno(err);
106 if (err == -EOPNOTSUPP)
107 return 0;
108
109 if (err) {
110 NL_SET_ERR_MSG_WEAK_MOD(extack,
111 "bridge flag offload is not supported");
112 return -EOPNOTSUPP;
113 }
114
115 attr.id = SWITCHDEV_ATTR_ID_PORT_BRIDGE_FLAGS;
116 attr.flags = SWITCHDEV_F_DEFER;
117
118 err = switchdev_port_attr_set(p->dev, &attr, extack);
119 if (err) {
120 NL_SET_ERR_MSG_WEAK_MOD(extack,
121 "error setting offload flag on port");
122 return err;
123 }
124
125 return 0;
126 }
127
br_switchdev_fdb_populate(struct net_bridge * br,struct switchdev_notifier_fdb_info * item,const struct net_bridge_fdb_entry * fdb,const void * ctx)128 static void br_switchdev_fdb_populate(struct net_bridge *br,
129 struct switchdev_notifier_fdb_info *item,
130 const struct net_bridge_fdb_entry *fdb,
131 const void *ctx)
132 {
133 const struct net_bridge_port *p = READ_ONCE(fdb->dst);
134
135 item->addr = fdb->key.addr.addr;
136 item->vid = fdb->key.vlan_id;
137 item->added_by_user = test_bit(BR_FDB_ADDED_BY_USER, &fdb->flags);
138 item->offloaded = test_bit(BR_FDB_OFFLOADED, &fdb->flags);
139 item->is_local = test_bit(BR_FDB_LOCAL, &fdb->flags);
140 item->locked = false;
141 item->info.dev = (!p || item->is_local) ? br->dev : p->dev;
142 item->info.ctx = ctx;
143 }
144
145 void
br_switchdev_fdb_notify(struct net_bridge * br,const struct net_bridge_fdb_entry * fdb,int type)146 br_switchdev_fdb_notify(struct net_bridge *br,
147 const struct net_bridge_fdb_entry *fdb, int type)
148 {
149 struct switchdev_notifier_fdb_info item;
150
151 if (test_bit(BR_FDB_LOCKED, &fdb->flags))
152 return;
153
154 /* Entries with these flags were created using ndm_state == NUD_REACHABLE,
155 * ndm_flags == NTF_MASTER( | NTF_STICKY), ext_flags == 0 by something
156 * equivalent to 'bridge fdb add ... master dynamic (sticky)'.
157 * Drivers don't know how to deal with these, so don't notify them to
158 * avoid confusing them.
159 */
160 if (test_bit(BR_FDB_ADDED_BY_USER, &fdb->flags) &&
161 !test_bit(BR_FDB_STATIC, &fdb->flags) &&
162 !test_bit(BR_FDB_ADDED_BY_EXT_LEARN, &fdb->flags))
163 return;
164
165 br_switchdev_fdb_populate(br, &item, fdb, NULL);
166
167 switch (type) {
168 case RTM_DELNEIGH:
169 call_switchdev_notifiers(SWITCHDEV_FDB_DEL_TO_DEVICE,
170 item.info.dev, &item.info, NULL);
171 break;
172 case RTM_NEWNEIGH:
173 call_switchdev_notifiers(SWITCHDEV_FDB_ADD_TO_DEVICE,
174 item.info.dev, &item.info, NULL);
175 break;
176 }
177 }
178
br_switchdev_port_vlan_add(struct net_device * dev,u16 vid,u16 flags,bool changed,struct netlink_ext_ack * extack)179 int br_switchdev_port_vlan_add(struct net_device *dev, u16 vid, u16 flags,
180 bool changed, struct netlink_ext_ack *extack)
181 {
182 struct switchdev_obj_port_vlan v = {
183 .obj.orig_dev = dev,
184 .obj.id = SWITCHDEV_OBJ_ID_PORT_VLAN,
185 .flags = flags,
186 .vid = vid,
187 .changed = changed,
188 };
189
190 return switchdev_port_obj_add(dev, &v.obj, extack);
191 }
192
br_switchdev_port_vlan_del(struct net_device * dev,u16 vid)193 int br_switchdev_port_vlan_del(struct net_device *dev, u16 vid)
194 {
195 struct switchdev_obj_port_vlan v = {
196 .obj.orig_dev = dev,
197 .obj.id = SWITCHDEV_OBJ_ID_PORT_VLAN,
198 .vid = vid,
199 };
200
201 return switchdev_port_obj_del(dev, &v.obj);
202 }
203
nbp_switchdev_hwdom_set(struct net_bridge_port * joining)204 static int nbp_switchdev_hwdom_set(struct net_bridge_port *joining)
205 {
206 struct net_bridge *br = joining->br;
207 struct net_bridge_port *p;
208 int hwdom;
209
210 /* joining is yet to be added to the port list. */
211 list_for_each_entry(p, &br->port_list, list) {
212 if (netdev_phys_item_id_same(&joining->ppid, &p->ppid)) {
213 joining->hwdom = p->hwdom;
214 return 0;
215 }
216 }
217
218 hwdom = find_next_zero_bit(&br->busy_hwdoms, BR_HWDOM_MAX, 1);
219 if (hwdom >= BR_HWDOM_MAX)
220 return -EBUSY;
221
222 set_bit(hwdom, &br->busy_hwdoms);
223 joining->hwdom = hwdom;
224 return 0;
225 }
226
nbp_switchdev_hwdom_put(struct net_bridge_port * leaving)227 static void nbp_switchdev_hwdom_put(struct net_bridge_port *leaving)
228 {
229 struct net_bridge *br = leaving->br;
230 struct net_bridge_port *p;
231
232 /* leaving is no longer in the port list. */
233 list_for_each_entry(p, &br->port_list, list) {
234 if (p->hwdom == leaving->hwdom)
235 return;
236 }
237
238 clear_bit(leaving->hwdom, &br->busy_hwdoms);
239 }
240
nbp_switchdev_add(struct net_bridge_port * p,struct netdev_phys_item_id ppid,bool tx_fwd_offload,struct netlink_ext_ack * extack)241 static int nbp_switchdev_add(struct net_bridge_port *p,
242 struct netdev_phys_item_id ppid,
243 bool tx_fwd_offload,
244 struct netlink_ext_ack *extack)
245 {
246 int err;
247
248 if (p->offload_count) {
249 /* Prevent unsupported configurations such as a bridge port
250 * which is a bonding interface, and the member ports are from
251 * different hardware switches.
252 */
253 if (!netdev_phys_item_id_same(&p->ppid, &ppid)) {
254 NL_SET_ERR_MSG_MOD(extack,
255 "Same bridge port cannot be offloaded by two physical switches");
256 return -EBUSY;
257 }
258
259 /* Tolerate drivers that call switchdev_bridge_port_offload()
260 * more than once for the same bridge port, such as when the
261 * bridge port is an offloaded bonding/team interface.
262 */
263 p->offload_count++;
264
265 return 0;
266 }
267
268 p->ppid = ppid;
269 p->offload_count = 1;
270
271 err = nbp_switchdev_hwdom_set(p);
272 if (err)
273 return err;
274
275 if (tx_fwd_offload) {
276 p->flags |= BR_TX_FWD_OFFLOAD;
277 static_branch_inc(&br_switchdev_tx_fwd_offload);
278 }
279
280 return 0;
281 }
282
nbp_switchdev_del(struct net_bridge_port * p)283 static void nbp_switchdev_del(struct net_bridge_port *p)
284 {
285 if (WARN_ON(!p->offload_count))
286 return;
287
288 p->offload_count--;
289
290 if (p->offload_count)
291 return;
292
293 if (p->hwdom)
294 nbp_switchdev_hwdom_put(p);
295
296 if (p->flags & BR_TX_FWD_OFFLOAD) {
297 p->flags &= ~BR_TX_FWD_OFFLOAD;
298 static_branch_dec(&br_switchdev_tx_fwd_offload);
299 }
300 }
301
302 static int
br_switchdev_fdb_replay_one(struct net_bridge * br,struct notifier_block * nb,const struct net_bridge_fdb_entry * fdb,unsigned long action,const void * ctx)303 br_switchdev_fdb_replay_one(struct net_bridge *br, struct notifier_block *nb,
304 const struct net_bridge_fdb_entry *fdb,
305 unsigned long action, const void *ctx)
306 {
307 struct switchdev_notifier_fdb_info item;
308 int err;
309
310 br_switchdev_fdb_populate(br, &item, fdb, ctx);
311
312 err = nb->notifier_call(nb, action, &item);
313 return notifier_to_errno(err);
314 }
315
316 static int
br_switchdev_fdb_replay(const struct net_device * br_dev,const void * ctx,bool adding,struct notifier_block * nb)317 br_switchdev_fdb_replay(const struct net_device *br_dev, const void *ctx,
318 bool adding, struct notifier_block *nb)
319 {
320 struct net_bridge_fdb_entry *fdb;
321 struct net_bridge *br;
322 unsigned long action;
323 int err = 0;
324
325 if (!nb)
326 return 0;
327
328 if (!netif_is_bridge_master(br_dev))
329 return -EINVAL;
330
331 br = netdev_priv(br_dev);
332
333 if (adding)
334 action = SWITCHDEV_FDB_ADD_TO_DEVICE;
335 else
336 action = SWITCHDEV_FDB_DEL_TO_DEVICE;
337
338 rcu_read_lock();
339
340 hlist_for_each_entry_rcu(fdb, &br->fdb_list, fdb_node) {
341 err = br_switchdev_fdb_replay_one(br, nb, fdb, action, ctx);
342 if (err)
343 break;
344 }
345
346 rcu_read_unlock();
347
348 return err;
349 }
350
br_switchdev_vlan_attr_replay(struct net_device * br_dev,const void * ctx,struct notifier_block * nb,struct netlink_ext_ack * extack)351 static int br_switchdev_vlan_attr_replay(struct net_device *br_dev,
352 const void *ctx,
353 struct notifier_block *nb,
354 struct netlink_ext_ack *extack)
355 {
356 struct switchdev_notifier_port_attr_info attr_info = {
357 .info = {
358 .dev = br_dev,
359 .extack = extack,
360 .ctx = ctx,
361 },
362 };
363 struct net_bridge *br = netdev_priv(br_dev);
364 struct net_bridge_vlan_group *vg;
365 struct switchdev_attr attr;
366 struct net_bridge_vlan *v;
367 int err;
368
369 attr_info.attr = &attr;
370 attr.orig_dev = br_dev;
371
372 vg = br_vlan_group(br);
373 if (!vg)
374 return 0;
375
376 list_for_each_entry(v, &vg->vlan_list, vlist) {
377 if (v->msti) {
378 attr.id = SWITCHDEV_ATTR_ID_VLAN_MSTI;
379 attr.u.vlan_msti.vid = v->vid;
380 attr.u.vlan_msti.msti = v->msti;
381
382 err = nb->notifier_call(nb, SWITCHDEV_PORT_ATTR_SET,
383 &attr_info);
384 err = notifier_to_errno(err);
385 if (err)
386 return err;
387 }
388 }
389
390 return 0;
391 }
392
393 static int
br_switchdev_vlan_replay_one(struct notifier_block * nb,struct net_device * dev,struct switchdev_obj_port_vlan * vlan,const void * ctx,unsigned long action,struct netlink_ext_ack * extack)394 br_switchdev_vlan_replay_one(struct notifier_block *nb,
395 struct net_device *dev,
396 struct switchdev_obj_port_vlan *vlan,
397 const void *ctx, unsigned long action,
398 struct netlink_ext_ack *extack)
399 {
400 struct switchdev_notifier_port_obj_info obj_info = {
401 .info = {
402 .dev = dev,
403 .extack = extack,
404 .ctx = ctx,
405 },
406 .obj = &vlan->obj,
407 };
408 int err;
409
410 err = nb->notifier_call(nb, action, &obj_info);
411 return notifier_to_errno(err);
412 }
413
br_switchdev_vlan_replay_group(struct notifier_block * nb,struct net_device * dev,struct net_bridge_vlan_group * vg,const void * ctx,unsigned long action,struct netlink_ext_ack * extack)414 static int br_switchdev_vlan_replay_group(struct notifier_block *nb,
415 struct net_device *dev,
416 struct net_bridge_vlan_group *vg,
417 const void *ctx, unsigned long action,
418 struct netlink_ext_ack *extack)
419 {
420 struct net_bridge_vlan *v;
421 int err = 0;
422 u16 pvid;
423
424 if (!vg)
425 return 0;
426
427 pvid = br_get_pvid(vg);
428
429 list_for_each_entry(v, &vg->vlan_list, vlist) {
430 struct switchdev_obj_port_vlan vlan = {
431 .obj.orig_dev = dev,
432 .obj.id = SWITCHDEV_OBJ_ID_PORT_VLAN,
433 .flags = br_vlan_flags(v, pvid),
434 .vid = v->vid,
435 };
436
437 if (!br_vlan_should_use(v))
438 continue;
439
440 err = br_switchdev_vlan_replay_one(nb, dev, &vlan, ctx,
441 action, extack);
442 if (err)
443 return err;
444 }
445
446 return 0;
447 }
448
br_switchdev_vlan_replay(struct net_device * br_dev,const void * ctx,bool adding,struct notifier_block * nb,struct netlink_ext_ack * extack)449 static int br_switchdev_vlan_replay(struct net_device *br_dev,
450 const void *ctx, bool adding,
451 struct notifier_block *nb,
452 struct netlink_ext_ack *extack)
453 {
454 struct net_bridge *br = netdev_priv(br_dev);
455 struct net_bridge_port *p;
456 unsigned long action;
457 int err;
458
459 ASSERT_RTNL();
460
461 if (!nb)
462 return 0;
463
464 if (!netif_is_bridge_master(br_dev))
465 return -EINVAL;
466
467 if (adding)
468 action = SWITCHDEV_PORT_OBJ_ADD;
469 else
470 action = SWITCHDEV_PORT_OBJ_DEL;
471
472 err = br_switchdev_vlan_replay_group(nb, br_dev, br_vlan_group(br),
473 ctx, action, extack);
474 if (err)
475 return err;
476
477 list_for_each_entry(p, &br->port_list, list) {
478 struct net_device *dev = p->dev;
479
480 err = br_switchdev_vlan_replay_group(nb, dev,
481 nbp_vlan_group(p),
482 ctx, action, extack);
483 if (err)
484 return err;
485 }
486
487 if (adding) {
488 err = br_switchdev_vlan_attr_replay(br_dev, ctx, nb, extack);
489 if (err)
490 return err;
491 }
492
493 return 0;
494 }
495
496 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
497 struct br_switchdev_mdb_complete_info {
498 struct net_bridge_port *port;
499 struct br_ip ip;
500 };
501
br_switchdev_mdb_complete(struct net_device * dev,int err,void * priv)502 static void br_switchdev_mdb_complete(struct net_device *dev, int err, void *priv)
503 {
504 struct br_switchdev_mdb_complete_info *data = priv;
505 struct net_bridge_port_group __rcu **pp;
506 struct net_bridge_port_group *p;
507 struct net_bridge_mdb_entry *mp;
508 struct net_bridge_port *port = data->port;
509 struct net_bridge *br = port->br;
510
511 if (err)
512 goto err;
513
514 spin_lock_bh(&br->multicast_lock);
515 mp = br_mdb_ip_get(br, &data->ip);
516 if (!mp)
517 goto out;
518 for (pp = &mp->ports; (p = mlock_dereference(*pp, br)) != NULL;
519 pp = &p->next) {
520 if (p->key.port != port)
521 continue;
522 p->flags |= MDB_PG_FLAGS_OFFLOAD;
523 }
524 out:
525 spin_unlock_bh(&br->multicast_lock);
526 err:
527 kfree(priv);
528 }
529
br_switchdev_mdb_populate(struct switchdev_obj_port_mdb * mdb,const struct net_bridge_mdb_entry * mp)530 static void br_switchdev_mdb_populate(struct switchdev_obj_port_mdb *mdb,
531 const struct net_bridge_mdb_entry *mp)
532 {
533 if (mp->addr.proto == htons(ETH_P_IP))
534 ip_eth_mc_map(mp->addr.dst.ip4, mdb->addr);
535 #if IS_ENABLED(CONFIG_IPV6)
536 else if (mp->addr.proto == htons(ETH_P_IPV6))
537 ipv6_eth_mc_map(&mp->addr.dst.ip6, mdb->addr);
538 #endif
539 else
540 ether_addr_copy(mdb->addr, mp->addr.dst.mac_addr);
541
542 mdb->vid = mp->addr.vid;
543 }
544
br_switchdev_host_mdb_one(struct net_device * dev,struct net_device * lower_dev,struct net_bridge_mdb_entry * mp,int type)545 static void br_switchdev_host_mdb_one(struct net_device *dev,
546 struct net_device *lower_dev,
547 struct net_bridge_mdb_entry *mp,
548 int type)
549 {
550 struct switchdev_obj_port_mdb mdb = {
551 .obj = {
552 .id = SWITCHDEV_OBJ_ID_HOST_MDB,
553 .flags = SWITCHDEV_F_DEFER,
554 .orig_dev = dev,
555 },
556 };
557
558 br_switchdev_mdb_populate(&mdb, mp);
559
560 switch (type) {
561 case RTM_NEWMDB:
562 switchdev_port_obj_add(lower_dev, &mdb.obj, NULL);
563 break;
564 case RTM_DELMDB:
565 switchdev_port_obj_del(lower_dev, &mdb.obj);
566 break;
567 }
568 }
569
br_switchdev_host_mdb(struct net_device * dev,struct net_bridge_mdb_entry * mp,int type)570 static void br_switchdev_host_mdb(struct net_device *dev,
571 struct net_bridge_mdb_entry *mp, int type)
572 {
573 struct net_device *lower_dev;
574 struct list_head *iter;
575
576 netdev_for_each_lower_dev(dev, lower_dev, iter)
577 br_switchdev_host_mdb_one(dev, lower_dev, mp, type);
578 }
579
580 static int
br_switchdev_mdb_replay_one(struct notifier_block * nb,struct net_device * dev,const struct switchdev_obj_port_mdb * mdb,unsigned long action,const void * ctx,struct netlink_ext_ack * extack)581 br_switchdev_mdb_replay_one(struct notifier_block *nb, struct net_device *dev,
582 const struct switchdev_obj_port_mdb *mdb,
583 unsigned long action, const void *ctx,
584 struct netlink_ext_ack *extack)
585 {
586 struct switchdev_notifier_port_obj_info obj_info = {
587 .info = {
588 .dev = dev,
589 .extack = extack,
590 .ctx = ctx,
591 },
592 .obj = &mdb->obj,
593 };
594 int err;
595
596 err = nb->notifier_call(nb, action, &obj_info);
597 return notifier_to_errno(err);
598 }
599
br_switchdev_mdb_queue_one(struct list_head * mdb_list,struct net_device * dev,unsigned long action,enum switchdev_obj_id id,const struct net_bridge_mdb_entry * mp,struct net_device * orig_dev)600 static int br_switchdev_mdb_queue_one(struct list_head *mdb_list,
601 struct net_device *dev,
602 unsigned long action,
603 enum switchdev_obj_id id,
604 const struct net_bridge_mdb_entry *mp,
605 struct net_device *orig_dev)
606 {
607 struct switchdev_obj_port_mdb mdb = {
608 .obj = {
609 .id = id,
610 .orig_dev = orig_dev,
611 },
612 };
613 struct switchdev_obj_port_mdb *pmdb;
614
615 br_switchdev_mdb_populate(&mdb, mp);
616
617 if (action == SWITCHDEV_PORT_OBJ_ADD &&
618 switchdev_port_obj_act_is_deferred(dev, action, &mdb.obj)) {
619 /* This event is already in the deferred queue of
620 * events, so this replay must be elided, lest the
621 * driver receives duplicate events for it. This can
622 * only happen when replaying additions, since
623 * modifications are always immediately visible in
624 * br->mdb_list, whereas actual event delivery may be
625 * delayed.
626 */
627 return 0;
628 }
629
630 pmdb = kmemdup(&mdb, sizeof(mdb), GFP_ATOMIC);
631 if (!pmdb)
632 return -ENOMEM;
633
634 list_add_tail(&pmdb->obj.list, mdb_list);
635 return 0;
636 }
637
br_switchdev_mdb_notify(struct net_device * dev,struct net_bridge_mdb_entry * mp,struct net_bridge_port_group * pg,int type)638 void br_switchdev_mdb_notify(struct net_device *dev,
639 struct net_bridge_mdb_entry *mp,
640 struct net_bridge_port_group *pg,
641 int type)
642 {
643 struct br_switchdev_mdb_complete_info *complete_info;
644 struct switchdev_obj_port_mdb mdb = {
645 .obj = {
646 .id = SWITCHDEV_OBJ_ID_PORT_MDB,
647 .flags = SWITCHDEV_F_DEFER,
648 },
649 };
650
651 if (!pg)
652 return br_switchdev_host_mdb(dev, mp, type);
653
654 br_switchdev_mdb_populate(&mdb, mp);
655
656 mdb.obj.orig_dev = pg->key.port->dev;
657 switch (type) {
658 case RTM_NEWMDB:
659 complete_info = kmalloc(sizeof(*complete_info), GFP_ATOMIC);
660 if (!complete_info)
661 break;
662 complete_info->port = pg->key.port;
663 complete_info->ip = mp->addr;
664 mdb.obj.complete_priv = complete_info;
665 mdb.obj.complete = br_switchdev_mdb_complete;
666 if (switchdev_port_obj_add(pg->key.port->dev, &mdb.obj, NULL))
667 kfree(complete_info);
668 break;
669 case RTM_DELMDB:
670 switchdev_port_obj_del(pg->key.port->dev, &mdb.obj);
671 break;
672 }
673 }
674 #endif
675
676 static int
br_switchdev_mdb_replay(struct net_device * br_dev,struct net_device * dev,const void * ctx,bool adding,struct notifier_block * nb,struct netlink_ext_ack * extack)677 br_switchdev_mdb_replay(struct net_device *br_dev, struct net_device *dev,
678 const void *ctx, bool adding, struct notifier_block *nb,
679 struct netlink_ext_ack *extack)
680 {
681 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
682 const struct net_bridge_mdb_entry *mp;
683 struct switchdev_obj *obj, *tmp;
684 struct net_bridge *br;
685 unsigned long action;
686 LIST_HEAD(mdb_list);
687 int err = 0;
688
689 ASSERT_RTNL();
690
691 if (!nb)
692 return 0;
693
694 if (!netif_is_bridge_master(br_dev) || !netif_is_bridge_port(dev))
695 return -EINVAL;
696
697 br = netdev_priv(br_dev);
698
699 if (!br_opt_get(br, BROPT_MULTICAST_ENABLED))
700 return 0;
701
702 if (adding)
703 action = SWITCHDEV_PORT_OBJ_ADD;
704 else
705 action = SWITCHDEV_PORT_OBJ_DEL;
706
707 /* br_switchdev_mdb_queue_one() will take care to not queue a
708 * replay of an event that is already pending in the switchdev
709 * deferred queue. In order to safely determine that, there
710 * must be no new deferred MDB notifications enqueued for the
711 * duration of the MDB scan. Therefore, grab the write-side
712 * lock to avoid racing with any concurrent IGMP/MLD snooping.
713 */
714 spin_lock_bh(&br->multicast_lock);
715
716 hlist_for_each_entry(mp, &br->mdb_list, mdb_node) {
717 struct net_bridge_port_group __rcu * const *pp;
718 const struct net_bridge_port_group *p;
719
720 if (mp->host_joined) {
721 err = br_switchdev_mdb_queue_one(&mdb_list, dev, action,
722 SWITCHDEV_OBJ_ID_HOST_MDB,
723 mp, br_dev);
724 if (err) {
725 spin_unlock_bh(&br->multicast_lock);
726 goto out_free_mdb;
727 }
728 }
729
730 for (pp = &mp->ports; (p = mlock_dereference(*pp, br)) != NULL;
731 pp = &p->next) {
732 if (p->key.port->dev != dev)
733 continue;
734
735 err = br_switchdev_mdb_queue_one(&mdb_list, dev, action,
736 SWITCHDEV_OBJ_ID_PORT_MDB,
737 mp, dev);
738 if (err) {
739 spin_unlock_bh(&br->multicast_lock);
740 goto out_free_mdb;
741 }
742 }
743 }
744
745 spin_unlock_bh(&br->multicast_lock);
746
747 list_for_each_entry(obj, &mdb_list, list) {
748 err = br_switchdev_mdb_replay_one(nb, dev,
749 SWITCHDEV_OBJ_PORT_MDB(obj),
750 action, ctx, extack);
751 if (err == -EOPNOTSUPP)
752 err = 0;
753 if (err)
754 goto out_free_mdb;
755 }
756
757 out_free_mdb:
758 list_for_each_entry_safe(obj, tmp, &mdb_list, list) {
759 list_del(&obj->list);
760 kfree(SWITCHDEV_OBJ_PORT_MDB(obj));
761 }
762
763 if (err)
764 return err;
765 #endif
766
767 return 0;
768 }
769
nbp_switchdev_sync_objs(struct net_bridge_port * p,const void * ctx,struct notifier_block * atomic_nb,struct notifier_block * blocking_nb,struct netlink_ext_ack * extack)770 static int nbp_switchdev_sync_objs(struct net_bridge_port *p, const void *ctx,
771 struct notifier_block *atomic_nb,
772 struct notifier_block *blocking_nb,
773 struct netlink_ext_ack *extack)
774 {
775 struct net_device *br_dev = p->br->dev;
776 struct net_device *dev = p->dev;
777 int err;
778
779 err = br_switchdev_vlan_replay(br_dev, ctx, true, blocking_nb, extack);
780 if (err && err != -EOPNOTSUPP)
781 return err;
782
783 err = br_switchdev_mdb_replay(br_dev, dev, ctx, true, blocking_nb,
784 extack);
785 if (err) {
786 /* -EOPNOTSUPP not propagated from MDB replay. */
787 return err;
788 }
789
790 err = br_switchdev_fdb_replay(br_dev, ctx, true, atomic_nb);
791 if (err && err != -EOPNOTSUPP)
792 return err;
793
794 return 0;
795 }
796
nbp_switchdev_unsync_objs(struct net_bridge_port * p,const void * ctx,struct notifier_block * atomic_nb,struct notifier_block * blocking_nb)797 static void nbp_switchdev_unsync_objs(struct net_bridge_port *p,
798 const void *ctx,
799 struct notifier_block *atomic_nb,
800 struct notifier_block *blocking_nb)
801 {
802 struct net_device *br_dev = p->br->dev;
803 struct net_device *dev = p->dev;
804
805 br_switchdev_fdb_replay(br_dev, ctx, false, atomic_nb);
806
807 br_switchdev_mdb_replay(br_dev, dev, ctx, false, blocking_nb, NULL);
808
809 br_switchdev_vlan_replay(br_dev, ctx, false, blocking_nb, NULL);
810
811 /* Make sure that the device leaving this bridge has seen all
812 * relevant events before it is disassociated. In the normal
813 * case, when the device is directly attached to the bridge,
814 * this is covered by del_nbp(). If the association was indirect
815 * however, e.g. via a team or bond, and the device is leaving
816 * that intermediate device, then the bridge port remains in
817 * place.
818 */
819 switchdev_deferred_process();
820 }
821
822 /* Let the bridge know that this port is offloaded, so that it can assign a
823 * switchdev hardware domain to it.
824 */
br_switchdev_port_offload(struct net_bridge_port * p,struct net_device * dev,const void * ctx,struct notifier_block * atomic_nb,struct notifier_block * blocking_nb,bool tx_fwd_offload,struct netlink_ext_ack * extack)825 int br_switchdev_port_offload(struct net_bridge_port *p,
826 struct net_device *dev, const void *ctx,
827 struct notifier_block *atomic_nb,
828 struct notifier_block *blocking_nb,
829 bool tx_fwd_offload,
830 struct netlink_ext_ack *extack)
831 {
832 struct netdev_phys_item_id ppid;
833 int err;
834
835 err = dev_get_port_parent_id(dev, &ppid, false);
836 if (err)
837 return err;
838
839 err = nbp_switchdev_add(p, ppid, tx_fwd_offload, extack);
840 if (err)
841 return err;
842
843 err = nbp_switchdev_sync_objs(p, ctx, atomic_nb, blocking_nb, extack);
844 if (err)
845 goto out_switchdev_del;
846
847 return 0;
848
849 out_switchdev_del:
850 nbp_switchdev_del(p);
851
852 return err;
853 }
854
br_switchdev_port_unoffload(struct net_bridge_port * p,const void * ctx,struct notifier_block * atomic_nb,struct notifier_block * blocking_nb)855 void br_switchdev_port_unoffload(struct net_bridge_port *p, const void *ctx,
856 struct notifier_block *atomic_nb,
857 struct notifier_block *blocking_nb)
858 {
859 nbp_switchdev_unsync_objs(p, ctx, atomic_nb, blocking_nb);
860
861 nbp_switchdev_del(p);
862 }
863
br_switchdev_port_replay(struct net_bridge_port * p,struct net_device * dev,const void * ctx,struct notifier_block * atomic_nb,struct notifier_block * blocking_nb,struct netlink_ext_ack * extack)864 int br_switchdev_port_replay(struct net_bridge_port *p,
865 struct net_device *dev, const void *ctx,
866 struct notifier_block *atomic_nb,
867 struct notifier_block *blocking_nb,
868 struct netlink_ext_ack *extack)
869 {
870 return nbp_switchdev_sync_objs(p, ctx, atomic_nb, blocking_nb, extack);
871 }
872