1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3 * Device handling code
4 * Linux ethernet bridge
5 *
6 * Authors:
7 * Lennert Buytenhek <buytenh@gnu.org>
8 */
9
10 #include <linux/kernel.h>
11 #include <linux/netdevice.h>
12 #include <linux/netpoll.h>
13 #include <linux/etherdevice.h>
14 #include <linux/ethtool.h>
15 #include <linux/list.h>
16 #include <linux/netfilter_bridge.h>
17
18 #include <linux/uaccess.h>
19 #include "br_private.h"
20
21 #define COMMON_FEATURES (NETIF_F_SG | NETIF_F_FRAGLIST | NETIF_F_HIGHDMA | \
22 NETIF_F_GSO_MASK | NETIF_F_HW_CSUM)
23
24 const struct nf_br_ops __rcu *nf_br_ops __read_mostly;
25 EXPORT_SYMBOL_GPL(nf_br_ops);
26
27 /* net device transmit always called with BH disabled */
br_dev_xmit(struct sk_buff * skb,struct net_device * dev)28 netdev_tx_t br_dev_xmit(struct sk_buff *skb, struct net_device *dev)
29 {
30 struct net_bridge *br = netdev_priv(dev);
31 struct net_bridge_fdb_entry *dst;
32 struct net_bridge_mdb_entry *mdst;
33 struct pcpu_sw_netstats *brstats = this_cpu_ptr(br->stats);
34 const struct nf_br_ops *nf_ops;
35 const unsigned char *dest;
36 u16 vid = 0;
37
38 memset(skb->cb, 0, sizeof(struct br_input_skb_cb));
39
40 rcu_read_lock();
41 nf_ops = rcu_dereference(nf_br_ops);
42 if (nf_ops && nf_ops->br_dev_xmit_hook(skb)) {
43 rcu_read_unlock();
44 return NETDEV_TX_OK;
45 }
46
47 u64_stats_update_begin(&brstats->syncp);
48 brstats->tx_packets++;
49 brstats->tx_bytes += skb->len;
50 u64_stats_update_end(&brstats->syncp);
51
52 br_switchdev_frame_unmark(skb);
53 BR_INPUT_SKB_CB(skb)->brdev = dev;
54 BR_INPUT_SKB_CB(skb)->frag_max_size = 0;
55
56 skb_reset_mac_header(skb);
57 skb_pull(skb, ETH_HLEN);
58
59 if (!br_allowed_ingress(br, br_vlan_group_rcu(br), skb, &vid))
60 goto out;
61
62 if (IS_ENABLED(CONFIG_INET) &&
63 (eth_hdr(skb)->h_proto == htons(ETH_P_ARP) ||
64 eth_hdr(skb)->h_proto == htons(ETH_P_RARP)) &&
65 br_opt_get(br, BROPT_NEIGH_SUPPRESS_ENABLED)) {
66 br_do_proxy_suppress_arp(skb, br, vid, NULL);
67 } else if (IS_ENABLED(CONFIG_IPV6) &&
68 skb->protocol == htons(ETH_P_IPV6) &&
69 br_opt_get(br, BROPT_NEIGH_SUPPRESS_ENABLED) &&
70 pskb_may_pull(skb, sizeof(struct ipv6hdr) +
71 sizeof(struct nd_msg)) &&
72 ipv6_hdr(skb)->nexthdr == IPPROTO_ICMPV6) {
73 struct nd_msg *msg, _msg;
74
75 msg = br_is_nd_neigh_msg(skb, &_msg);
76 if (msg)
77 br_do_suppress_nd(skb, br, vid, NULL, msg);
78 }
79
80 dest = eth_hdr(skb)->h_dest;
81 if (is_broadcast_ether_addr(dest)) {
82 br_flood(br, skb, BR_PKT_BROADCAST, false, true);
83 } else if (is_multicast_ether_addr(dest)) {
84 if (unlikely(netpoll_tx_running(dev))) {
85 br_flood(br, skb, BR_PKT_MULTICAST, false, true);
86 goto out;
87 }
88 if (br_multicast_rcv(br, NULL, skb, vid)) {
89 kfree_skb(skb);
90 goto out;
91 }
92
93 mdst = br_mdb_get(br, skb, vid);
94 if ((mdst || BR_INPUT_SKB_CB_MROUTERS_ONLY(skb)) &&
95 br_multicast_querier_exists(br, eth_hdr(skb)))
96 br_multicast_flood(mdst, skb, false, true);
97 else
98 br_flood(br, skb, BR_PKT_MULTICAST, false, true);
99 } else if ((dst = br_fdb_find_rcu(br, dest, vid)) != NULL) {
100 br_forward(dst->dst, skb, false, true);
101 } else {
102 br_flood(br, skb, BR_PKT_UNICAST, false, true);
103 }
104 out:
105 rcu_read_unlock();
106 return NETDEV_TX_OK;
107 }
108
br_dev_init(struct net_device * dev)109 static int br_dev_init(struct net_device *dev)
110 {
111 struct net_bridge *br = netdev_priv(dev);
112 int err;
113
114 br->stats = netdev_alloc_pcpu_stats(struct pcpu_sw_netstats);
115 if (!br->stats)
116 return -ENOMEM;
117
118 err = br_fdb_hash_init(br);
119 if (err) {
120 free_percpu(br->stats);
121 return err;
122 }
123
124 err = br_mdb_hash_init(br);
125 if (err) {
126 free_percpu(br->stats);
127 br_fdb_hash_fini(br);
128 return err;
129 }
130
131 err = br_vlan_init(br);
132 if (err) {
133 free_percpu(br->stats);
134 br_mdb_hash_fini(br);
135 br_fdb_hash_fini(br);
136 return err;
137 }
138
139 err = br_multicast_init_stats(br);
140 if (err) {
141 free_percpu(br->stats);
142 br_vlan_flush(br);
143 br_mdb_hash_fini(br);
144 br_fdb_hash_fini(br);
145 }
146
147 return err;
148 }
149
br_dev_uninit(struct net_device * dev)150 static void br_dev_uninit(struct net_device *dev)
151 {
152 struct net_bridge *br = netdev_priv(dev);
153
154 br_multicast_dev_del(br);
155 br_multicast_uninit_stats(br);
156 br_vlan_flush(br);
157 br_mdb_hash_fini(br);
158 br_fdb_hash_fini(br);
159 free_percpu(br->stats);
160 }
161
br_dev_open(struct net_device * dev)162 static int br_dev_open(struct net_device *dev)
163 {
164 struct net_bridge *br = netdev_priv(dev);
165
166 netdev_update_features(dev);
167 netif_start_queue(dev);
168 br_stp_enable_bridge(br);
169 br_multicast_open(br);
170
171 if (br_opt_get(br, BROPT_MULTICAST_ENABLED))
172 br_multicast_join_snoopers(br);
173
174 return 0;
175 }
176
br_dev_set_multicast_list(struct net_device * dev)177 static void br_dev_set_multicast_list(struct net_device *dev)
178 {
179 }
180
br_dev_change_rx_flags(struct net_device * dev,int change)181 static void br_dev_change_rx_flags(struct net_device *dev, int change)
182 {
183 if (change & IFF_PROMISC)
184 br_manage_promisc(netdev_priv(dev));
185 }
186
br_dev_stop(struct net_device * dev)187 static int br_dev_stop(struct net_device *dev)
188 {
189 struct net_bridge *br = netdev_priv(dev);
190
191 br_stp_disable_bridge(br);
192 br_multicast_stop(br);
193
194 if (br_opt_get(br, BROPT_MULTICAST_ENABLED))
195 br_multicast_leave_snoopers(br);
196
197 netif_stop_queue(dev);
198
199 return 0;
200 }
201
br_get_stats64(struct net_device * dev,struct rtnl_link_stats64 * stats)202 static void br_get_stats64(struct net_device *dev,
203 struct rtnl_link_stats64 *stats)
204 {
205 struct net_bridge *br = netdev_priv(dev);
206 struct pcpu_sw_netstats tmp, sum = { 0 };
207 unsigned int cpu;
208
209 for_each_possible_cpu(cpu) {
210 unsigned int start;
211 const struct pcpu_sw_netstats *bstats
212 = per_cpu_ptr(br->stats, cpu);
213 do {
214 start = u64_stats_fetch_begin_irq(&bstats->syncp);
215 memcpy(&tmp, bstats, sizeof(tmp));
216 } while (u64_stats_fetch_retry_irq(&bstats->syncp, start));
217 sum.tx_bytes += tmp.tx_bytes;
218 sum.tx_packets += tmp.tx_packets;
219 sum.rx_bytes += tmp.rx_bytes;
220 sum.rx_packets += tmp.rx_packets;
221 }
222
223 netdev_stats_to_stats64(stats, &dev->stats);
224 stats->tx_bytes = sum.tx_bytes;
225 stats->tx_packets = sum.tx_packets;
226 stats->rx_bytes = sum.rx_bytes;
227 stats->rx_packets = sum.rx_packets;
228 }
229
br_change_mtu(struct net_device * dev,int new_mtu)230 static int br_change_mtu(struct net_device *dev, int new_mtu)
231 {
232 struct net_bridge *br = netdev_priv(dev);
233
234 dev->mtu = new_mtu;
235
236 /* this flag will be cleared if the MTU was automatically adjusted */
237 br_opt_toggle(br, BROPT_MTU_SET_BY_USER, true);
238 #if IS_ENABLED(CONFIG_BRIDGE_NETFILTER)
239 /* remember the MTU in the rtable for PMTU */
240 dst_metric_set(&br->fake_rtable.dst, RTAX_MTU, new_mtu);
241 #endif
242
243 return 0;
244 }
245
246 /* Allow setting mac address to any valid ethernet address. */
br_set_mac_address(struct net_device * dev,void * p)247 static int br_set_mac_address(struct net_device *dev, void *p)
248 {
249 struct net_bridge *br = netdev_priv(dev);
250 struct sockaddr *addr = p;
251
252 if (!is_valid_ether_addr(addr->sa_data))
253 return -EADDRNOTAVAIL;
254
255 /* dev_set_mac_addr() can be called by a master device on bridge's
256 * NETDEV_UNREGISTER, but since it's being destroyed do nothing
257 */
258 if (dev->reg_state != NETREG_REGISTERED)
259 return -EBUSY;
260
261 spin_lock_bh(&br->lock);
262 if (!ether_addr_equal(dev->dev_addr, addr->sa_data)) {
263 /* Mac address will be changed in br_stp_change_bridge_id(). */
264 br_stp_change_bridge_id(br, addr->sa_data);
265 }
266 spin_unlock_bh(&br->lock);
267
268 return 0;
269 }
270
br_getinfo(struct net_device * dev,struct ethtool_drvinfo * info)271 static void br_getinfo(struct net_device *dev, struct ethtool_drvinfo *info)
272 {
273 strlcpy(info->driver, "bridge", sizeof(info->driver));
274 strlcpy(info->version, BR_VERSION, sizeof(info->version));
275 strlcpy(info->fw_version, "N/A", sizeof(info->fw_version));
276 strlcpy(info->bus_info, "N/A", sizeof(info->bus_info));
277 }
278
br_fix_features(struct net_device * dev,netdev_features_t features)279 static netdev_features_t br_fix_features(struct net_device *dev,
280 netdev_features_t features)
281 {
282 struct net_bridge *br = netdev_priv(dev);
283
284 return br_features_recompute(br, features);
285 }
286
287 #ifdef CONFIG_NET_POLL_CONTROLLER
br_poll_controller(struct net_device * br_dev)288 static void br_poll_controller(struct net_device *br_dev)
289 {
290 }
291
br_netpoll_cleanup(struct net_device * dev)292 static void br_netpoll_cleanup(struct net_device *dev)
293 {
294 struct net_bridge *br = netdev_priv(dev);
295 struct net_bridge_port *p;
296
297 list_for_each_entry(p, &br->port_list, list)
298 br_netpoll_disable(p);
299 }
300
__br_netpoll_enable(struct net_bridge_port * p)301 static int __br_netpoll_enable(struct net_bridge_port *p)
302 {
303 struct netpoll *np;
304 int err;
305
306 np = kzalloc(sizeof(*p->np), GFP_KERNEL);
307 if (!np)
308 return -ENOMEM;
309
310 err = __netpoll_setup(np, p->dev);
311 if (err) {
312 kfree(np);
313 return err;
314 }
315
316 p->np = np;
317 return err;
318 }
319
br_netpoll_enable(struct net_bridge_port * p)320 int br_netpoll_enable(struct net_bridge_port *p)
321 {
322 if (!p->br->dev->npinfo)
323 return 0;
324
325 return __br_netpoll_enable(p);
326 }
327
br_netpoll_setup(struct net_device * dev,struct netpoll_info * ni)328 static int br_netpoll_setup(struct net_device *dev, struct netpoll_info *ni)
329 {
330 struct net_bridge *br = netdev_priv(dev);
331 struct net_bridge_port *p;
332 int err = 0;
333
334 list_for_each_entry(p, &br->port_list, list) {
335 if (!p->dev)
336 continue;
337 err = __br_netpoll_enable(p);
338 if (err)
339 goto fail;
340 }
341
342 out:
343 return err;
344
345 fail:
346 br_netpoll_cleanup(dev);
347 goto out;
348 }
349
br_netpoll_disable(struct net_bridge_port * p)350 void br_netpoll_disable(struct net_bridge_port *p)
351 {
352 struct netpoll *np = p->np;
353
354 if (!np)
355 return;
356
357 p->np = NULL;
358
359 __netpoll_free(np);
360 }
361
362 #endif
363
br_add_slave(struct net_device * dev,struct net_device * slave_dev,struct netlink_ext_ack * extack)364 static int br_add_slave(struct net_device *dev, struct net_device *slave_dev,
365 struct netlink_ext_ack *extack)
366
367 {
368 struct net_bridge *br = netdev_priv(dev);
369
370 return br_add_if(br, slave_dev, extack);
371 }
372
br_del_slave(struct net_device * dev,struct net_device * slave_dev)373 static int br_del_slave(struct net_device *dev, struct net_device *slave_dev)
374 {
375 struct net_bridge *br = netdev_priv(dev);
376
377 return br_del_if(br, slave_dev);
378 }
379
380 static const struct ethtool_ops br_ethtool_ops = {
381 .get_drvinfo = br_getinfo,
382 .get_link = ethtool_op_get_link,
383 };
384
385 static const struct net_device_ops br_netdev_ops = {
386 .ndo_open = br_dev_open,
387 .ndo_stop = br_dev_stop,
388 .ndo_init = br_dev_init,
389 .ndo_uninit = br_dev_uninit,
390 .ndo_start_xmit = br_dev_xmit,
391 .ndo_get_stats64 = br_get_stats64,
392 .ndo_set_mac_address = br_set_mac_address,
393 .ndo_set_rx_mode = br_dev_set_multicast_list,
394 .ndo_change_rx_flags = br_dev_change_rx_flags,
395 .ndo_change_mtu = br_change_mtu,
396 .ndo_do_ioctl = br_dev_ioctl,
397 #ifdef CONFIG_NET_POLL_CONTROLLER
398 .ndo_netpoll_setup = br_netpoll_setup,
399 .ndo_netpoll_cleanup = br_netpoll_cleanup,
400 .ndo_poll_controller = br_poll_controller,
401 #endif
402 .ndo_add_slave = br_add_slave,
403 .ndo_del_slave = br_del_slave,
404 .ndo_fix_features = br_fix_features,
405 .ndo_fdb_add = br_fdb_add,
406 .ndo_fdb_del = br_fdb_delete,
407 .ndo_fdb_dump = br_fdb_dump,
408 .ndo_fdb_get = br_fdb_get,
409 .ndo_bridge_getlink = br_getlink,
410 .ndo_bridge_setlink = br_setlink,
411 .ndo_bridge_dellink = br_dellink,
412 .ndo_features_check = passthru_features_check,
413 };
414
415 static struct device_type br_type = {
416 .name = "bridge",
417 };
418
br_dev_setup(struct net_device * dev)419 void br_dev_setup(struct net_device *dev)
420 {
421 struct net_bridge *br = netdev_priv(dev);
422
423 eth_hw_addr_random(dev);
424 ether_setup(dev);
425
426 dev->netdev_ops = &br_netdev_ops;
427 dev->needs_free_netdev = true;
428 dev->ethtool_ops = &br_ethtool_ops;
429 SET_NETDEV_DEVTYPE(dev, &br_type);
430 dev->priv_flags = IFF_EBRIDGE | IFF_NO_QUEUE;
431
432 dev->features = COMMON_FEATURES | NETIF_F_LLTX | NETIF_F_NETNS_LOCAL |
433 NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_STAG_TX;
434 dev->hw_features = COMMON_FEATURES | NETIF_F_HW_VLAN_CTAG_TX |
435 NETIF_F_HW_VLAN_STAG_TX;
436 dev->vlan_features = COMMON_FEATURES;
437
438 br->dev = dev;
439 spin_lock_init(&br->lock);
440 INIT_LIST_HEAD(&br->port_list);
441 INIT_HLIST_HEAD(&br->fdb_list);
442 spin_lock_init(&br->hash_lock);
443
444 br->bridge_id.prio[0] = 0x80;
445 br->bridge_id.prio[1] = 0x00;
446
447 ether_addr_copy(br->group_addr, eth_stp_addr);
448
449 br->stp_enabled = BR_NO_STP;
450 br->group_fwd_mask = BR_GROUPFWD_DEFAULT;
451 br->group_fwd_mask_required = BR_GROUPFWD_DEFAULT;
452
453 br->designated_root = br->bridge_id;
454 br->bridge_max_age = br->max_age = 20 * HZ;
455 br->bridge_hello_time = br->hello_time = 2 * HZ;
456 br->bridge_forward_delay = br->forward_delay = 15 * HZ;
457 br->bridge_ageing_time = br->ageing_time = BR_DEFAULT_AGEING_TIME;
458 dev->max_mtu = ETH_MAX_MTU;
459
460 br_netfilter_rtable_init(br);
461 br_stp_timer_init(br);
462 br_multicast_init(br);
463 INIT_DELAYED_WORK(&br->gc_work, br_fdb_cleanup);
464 }
465