1 /* SPDX-License-Identifier: GPL-2.0-or-later */
3 * net/dsa/dsa_priv.h - Hardware switch handling
4 * Copyright (c) 2008-2009 Marvell Semiconductor
10 #include <linux/if_bridge.h>
11 #include <linux/phy.h>
12 #include <linux/netdevice.h>
13 #include <linux/netpoll.h>
15 #include <net/gro_cells.h>
17 #define DSA_MAX_NUM_OFFLOADING_BRIDGES BITS_PER_LONG
20 DSA_NOTIFIER_AGEING_TIME,
21 DSA_NOTIFIER_BRIDGE_JOIN,
22 DSA_NOTIFIER_BRIDGE_LEAVE,
25 DSA_NOTIFIER_HOST_FDB_ADD,
26 DSA_NOTIFIER_HOST_FDB_DEL,
27 DSA_NOTIFIER_HSR_JOIN,
28 DSA_NOTIFIER_HSR_LEAVE,
29 DSA_NOTIFIER_LAG_CHANGE,
30 DSA_NOTIFIER_LAG_JOIN,
31 DSA_NOTIFIER_LAG_LEAVE,
34 DSA_NOTIFIER_HOST_MDB_ADD,
35 DSA_NOTIFIER_HOST_MDB_DEL,
36 DSA_NOTIFIER_VLAN_ADD,
37 DSA_NOTIFIER_VLAN_DEL,
39 DSA_NOTIFIER_TAG_PROTO,
42 DSA_NOTIFIER_MRP_ADD_RING_ROLE,
43 DSA_NOTIFIER_MRP_DEL_RING_ROLE,
44 DSA_NOTIFIER_TAG_8021Q_VLAN_ADD,
45 DSA_NOTIFIER_TAG_8021Q_VLAN_DEL,
48 /* DSA_NOTIFIER_AGEING_TIME */
49 struct dsa_notifier_ageing_time_info {
50 unsigned int ageing_time;
53 /* DSA_NOTIFIER_BRIDGE_* */
54 struct dsa_notifier_bridge_info {
55 struct net_device *br;
61 /* DSA_NOTIFIER_FDB_* */
62 struct dsa_notifier_fdb_info {
65 const unsigned char *addr;
69 /* DSA_NOTIFIER_MDB_* */
70 struct dsa_notifier_mdb_info {
71 const struct switchdev_obj_port_mdb *mdb;
76 /* DSA_NOTIFIER_LAG_* */
77 struct dsa_notifier_lag_info {
78 struct net_device *lag;
82 struct netdev_lag_upper_info *info;
85 /* DSA_NOTIFIER_VLAN_* */
86 struct dsa_notifier_vlan_info {
87 const struct switchdev_obj_port_vlan *vlan;
90 struct netlink_ext_ack *extack;
93 /* DSA_NOTIFIER_MTU */
94 struct dsa_notifier_mtu_info {
101 /* DSA_NOTIFIER_TAG_PROTO_* */
102 struct dsa_notifier_tag_proto_info {
103 const struct dsa_device_ops *tag_ops;
106 /* DSA_NOTIFIER_MRP_* */
107 struct dsa_notifier_mrp_info {
108 const struct switchdev_obj_mrp *mrp;
113 /* DSA_NOTIFIER_MRP_* */
114 struct dsa_notifier_mrp_ring_role_info {
115 const struct switchdev_obj_ring_role_mrp *mrp;
120 /* DSA_NOTIFIER_TAG_8021Q_VLAN_* */
121 struct dsa_notifier_tag_8021q_vlan_info {
128 struct dsa_switchdev_event_work {
129 struct dsa_switch *ds;
131 struct net_device *dev;
132 struct work_struct work;
134 /* Specific for SWITCHDEV_FDB_ADD_TO_DEVICE and
135 * SWITCHDEV_FDB_DEL_TO_DEVICE
137 unsigned char addr[ETH_ALEN];
142 /* DSA_NOTIFIER_HSR_* */
143 struct dsa_notifier_hsr_info {
144 struct net_device *hsr;
149 struct dsa_slave_priv {
150 /* Copy of CPU port xmit for faster access in slave transmit hot path */
151 struct sk_buff * (*xmit)(struct sk_buff *skb,
152 struct net_device *dev);
154 struct gro_cells gcells;
156 /* DSA port data, such as switch, port index, etc. */
159 #ifdef CONFIG_NET_POLL_CONTROLLER
160 struct netpoll *netpoll;
164 struct list_head mall_tc_list;
168 const struct dsa_device_ops *dsa_tag_driver_get(int tag_protocol);
169 void dsa_tag_driver_put(const struct dsa_device_ops *ops);
170 const struct dsa_device_ops *dsa_find_tagger_by_name(const char *buf);
172 bool dsa_schedule_work(struct work_struct *work);
173 void dsa_flush_workqueue(void);
174 const char *dsa_tag_protocol_to_str(const struct dsa_device_ops *ops);
176 static inline int dsa_tag_protocol_overhead(const struct dsa_device_ops *ops)
178 return ops->needed_headroom + ops->needed_tailroom;
182 int dsa_master_setup(struct net_device *dev, struct dsa_port *cpu_dp);
183 void dsa_master_teardown(struct net_device *dev);
185 static inline struct net_device *dsa_master_find_slave(struct net_device *dev,
186 int device, int port)
188 struct dsa_port *cpu_dp = dev->dsa_ptr;
189 struct dsa_switch_tree *dst = cpu_dp->dst;
192 list_for_each_entry(dp, &dst->ports, list)
193 if (dp->ds->index == device && dp->index == port &&
194 dp->type == DSA_PORT_TYPE_USER)
201 void dsa_port_set_tag_protocol(struct dsa_port *cpu_dp,
202 const struct dsa_device_ops *tag_ops);
203 int dsa_port_set_state(struct dsa_port *dp, u8 state, bool do_fast_age);
204 int dsa_port_enable_rt(struct dsa_port *dp, struct phy_device *phy);
205 int dsa_port_enable(struct dsa_port *dp, struct phy_device *phy);
206 void dsa_port_disable_rt(struct dsa_port *dp);
207 void dsa_port_disable(struct dsa_port *dp);
208 int dsa_port_bridge_join(struct dsa_port *dp, struct net_device *br,
209 struct netlink_ext_ack *extack);
210 void dsa_port_pre_bridge_leave(struct dsa_port *dp, struct net_device *br);
211 void dsa_port_bridge_leave(struct dsa_port *dp, struct net_device *br);
212 int dsa_port_lag_change(struct dsa_port *dp,
213 struct netdev_lag_lower_state_info *linfo);
214 int dsa_port_lag_join(struct dsa_port *dp, struct net_device *lag_dev,
215 struct netdev_lag_upper_info *uinfo,
216 struct netlink_ext_ack *extack);
217 void dsa_port_pre_lag_leave(struct dsa_port *dp, struct net_device *lag_dev);
218 void dsa_port_lag_leave(struct dsa_port *dp, struct net_device *lag_dev);
219 int dsa_port_vlan_filtering(struct dsa_port *dp, bool vlan_filtering,
220 struct netlink_ext_ack *extack);
221 bool dsa_port_skip_vlan_configuration(struct dsa_port *dp);
222 int dsa_port_ageing_time(struct dsa_port *dp, clock_t ageing_clock);
223 int dsa_port_mtu_change(struct dsa_port *dp, int new_mtu,
224 bool targeted_match);
225 int dsa_port_fdb_add(struct dsa_port *dp, const unsigned char *addr,
227 int dsa_port_fdb_del(struct dsa_port *dp, const unsigned char *addr,
229 int dsa_port_host_fdb_add(struct dsa_port *dp, const unsigned char *addr,
231 int dsa_port_host_fdb_del(struct dsa_port *dp, const unsigned char *addr,
233 int dsa_port_fdb_dump(struct dsa_port *dp, dsa_fdb_dump_cb_t *cb, void *data);
234 int dsa_port_mdb_add(const struct dsa_port *dp,
235 const struct switchdev_obj_port_mdb *mdb);
236 int dsa_port_mdb_del(const struct dsa_port *dp,
237 const struct switchdev_obj_port_mdb *mdb);
238 int dsa_port_host_mdb_add(const struct dsa_port *dp,
239 const struct switchdev_obj_port_mdb *mdb);
240 int dsa_port_host_mdb_del(const struct dsa_port *dp,
241 const struct switchdev_obj_port_mdb *mdb);
242 int dsa_port_pre_bridge_flags(const struct dsa_port *dp,
243 struct switchdev_brport_flags flags,
244 struct netlink_ext_ack *extack);
245 int dsa_port_bridge_flags(struct dsa_port *dp,
246 struct switchdev_brport_flags flags,
247 struct netlink_ext_ack *extack);
248 int dsa_port_vlan_add(struct dsa_port *dp,
249 const struct switchdev_obj_port_vlan *vlan,
250 struct netlink_ext_ack *extack);
251 int dsa_port_vlan_del(struct dsa_port *dp,
252 const struct switchdev_obj_port_vlan *vlan);
253 int dsa_port_mrp_add(const struct dsa_port *dp,
254 const struct switchdev_obj_mrp *mrp);
255 int dsa_port_mrp_del(const struct dsa_port *dp,
256 const struct switchdev_obj_mrp *mrp);
257 int dsa_port_mrp_add_ring_role(const struct dsa_port *dp,
258 const struct switchdev_obj_ring_role_mrp *mrp);
259 int dsa_port_mrp_del_ring_role(const struct dsa_port *dp,
260 const struct switchdev_obj_ring_role_mrp *mrp);
261 int dsa_port_link_register_of(struct dsa_port *dp);
262 void dsa_port_link_unregister_of(struct dsa_port *dp);
263 int dsa_port_hsr_join(struct dsa_port *dp, struct net_device *hsr);
264 void dsa_port_hsr_leave(struct dsa_port *dp, struct net_device *hsr);
265 int dsa_port_tag_8021q_vlan_add(struct dsa_port *dp, u16 vid, bool broadcast);
266 void dsa_port_tag_8021q_vlan_del(struct dsa_port *dp, u16 vid, bool broadcast);
267 extern const struct phylink_mac_ops dsa_port_phylink_mac_ops;
269 static inline bool dsa_port_offloads_bridge_port(struct dsa_port *dp,
270 const struct net_device *dev)
272 return dsa_port_to_bridge_port(dp) == dev;
275 static inline bool dsa_port_offloads_bridge(struct dsa_port *dp,
276 const struct net_device *bridge_dev)
278 /* DSA ports connected to a bridge, and event was emitted
281 return dp->bridge_dev == bridge_dev;
284 /* Returns true if any port of this tree offloads the given net_device */
285 static inline bool dsa_tree_offloads_bridge_port(struct dsa_switch_tree *dst,
286 const struct net_device *dev)
290 list_for_each_entry(dp, &dst->ports, list)
291 if (dsa_port_offloads_bridge_port(dp, dev))
297 /* Returns true if any port of this tree offloads the given bridge */
298 static inline bool dsa_tree_offloads_bridge(struct dsa_switch_tree *dst,
299 const struct net_device *bridge_dev)
303 list_for_each_entry(dp, &dst->ports, list)
304 if (dsa_port_offloads_bridge(dp, bridge_dev))
311 extern const struct dsa_device_ops notag_netdev_ops;
312 extern struct notifier_block dsa_slave_switchdev_notifier;
313 extern struct notifier_block dsa_slave_switchdev_blocking_notifier;
315 void dsa_slave_mii_bus_init(struct dsa_switch *ds);
316 int dsa_slave_create(struct dsa_port *dp);
317 void dsa_slave_destroy(struct net_device *slave_dev);
318 int dsa_slave_suspend(struct net_device *slave_dev);
319 int dsa_slave_resume(struct net_device *slave_dev);
320 int dsa_slave_register_notifier(void);
321 void dsa_slave_unregister_notifier(void);
322 void dsa_slave_setup_tagger(struct net_device *slave);
323 int dsa_slave_change_mtu(struct net_device *dev, int new_mtu);
324 int dsa_slave_manage_vlan_filtering(struct net_device *dev,
325 bool vlan_filtering);
327 static inline struct dsa_port *dsa_slave_to_port(const struct net_device *dev)
329 struct dsa_slave_priv *p = netdev_priv(dev);
334 static inline struct net_device *
335 dsa_slave_to_master(const struct net_device *dev)
337 struct dsa_port *dp = dsa_slave_to_port(dev);
339 return dp->cpu_dp->master;
342 /* If under a bridge with vlan_filtering=0, make sure to send pvid-tagged
343 * frames as untagged, since the bridge will not untag them.
345 static inline struct sk_buff *dsa_untag_bridge_pvid(struct sk_buff *skb)
347 struct dsa_port *dp = dsa_slave_to_port(skb->dev);
348 struct net_device *br = dp->bridge_dev;
349 struct net_device *dev = skb->dev;
350 struct net_device *upper_dev;
351 u16 vid, pvid, proto;
354 if (!br || br_vlan_enabled(br))
357 err = br_vlan_get_proto(br, &proto);
361 /* Move VLAN tag from data to hwaccel */
362 if (!skb_vlan_tag_present(skb) && skb->protocol == htons(proto)) {
363 skb = skb_vlan_untag(skb);
368 if (!skb_vlan_tag_present(skb))
371 vid = skb_vlan_tag_get_id(skb);
373 /* We already run under an RCU read-side critical section since
374 * we are called from netif_receive_skb_list_internal().
376 err = br_vlan_get_pvid_rcu(dev, &pvid);
383 /* The sad part about attempting to untag from DSA is that we
384 * don't know, unless we check, if the skb will end up in
385 * the bridge's data path - br_allowed_ingress() - or not.
386 * For example, there might be an 8021q upper for the
387 * default_pvid of the bridge, which will steal VLAN-tagged traffic
388 * from the bridge's data path. This is a configuration that DSA
389 * supports because vlan_filtering is 0. In that case, we should
390 * definitely keep the tag, to make sure it keeps working.
392 upper_dev = __vlan_find_dev_deep_rcu(br, htons(proto), vid);
396 __vlan_hwaccel_clear_tag(skb);
401 /* For switches without hardware support for DSA tagging to be able
402 * to support termination through the bridge.
404 static inline struct net_device *
405 dsa_find_designated_bridge_port_by_vid(struct net_device *master, u16 vid)
407 struct dsa_port *cpu_dp = master->dsa_ptr;
408 struct dsa_switch_tree *dst = cpu_dp->dst;
409 struct bridge_vlan_info vinfo;
410 struct net_device *slave;
414 list_for_each_entry(dp, &dst->ports, list) {
415 if (dp->type != DSA_PORT_TYPE_USER)
421 if (dp->stp_state != BR_STATE_LEARNING &&
422 dp->stp_state != BR_STATE_FORWARDING)
425 /* Since the bridge might learn this packet, keep the CPU port
426 * affinity with the port that will be used for the reply on
429 if (dp->cpu_dp != cpu_dp)
434 err = br_vlan_get_info_rcu(slave, vid, &vinfo);
444 /* If the ingress port offloads the bridge, we mark the frame as autonomously
445 * forwarded by hardware, so the software bridge doesn't forward in twice, back
446 * to us, because we already did. However, if we're in fallback mode and we do
447 * software bridging, we are not offloading it, therefore the dp->bridge_dev
448 * pointer is not populated, and flooding needs to be done by software (we are
449 * effectively operating in standalone ports mode).
451 static inline void dsa_default_offload_fwd_mark(struct sk_buff *skb)
453 struct dsa_port *dp = dsa_slave_to_port(skb->dev);
455 skb->offload_fwd_mark = !!(dp->bridge_dev);
458 /* Helper for removing DSA header tags from packets in the RX path.
459 * Must not be called before skb_pull(len).
463 * | | | | | | | | | | | | | | | | | | |
464 * +-----------------------+-----------------------+---------------+-------+
465 * | Destination MAC | Source MAC | DSA header | EType |
466 * +-----------------------+-----------------------+---------------+-------+
468 * <----- len -----> <----- len ----->
471 * >>>>>>> | | | | | | | | | | | | | | |
472 * >>>>>>> +-----------------------+-----------------------+-------+
473 * >>>>>>> | Destination MAC | Source MAC | EType |
474 * +-----------------------+-----------------------+-------+
479 static inline void dsa_strip_etype_header(struct sk_buff *skb, int len)
481 memmove(skb->data - ETH_HLEN, skb->data - ETH_HLEN - len, 2 * ETH_ALEN);
484 /* Helper for creating space for DSA header tags in TX path packets.
485 * Must not be called before skb_push(len).
489 * <<<<<<< | | | | | | | | | | | | | | |
490 * ^ <<<<<<< +-----------------------+-----------------------+-------+
491 * | <<<<<<< | Destination MAC | Source MAC | EType |
492 * | +-----------------------+-----------------------+-------+
500 * | | | | | | | | | | | | | | | | | | |
501 * +-----------------------+-----------------------+---------------+-------+
502 * | Destination MAC | Source MAC | DSA header | EType |
503 * +-----------------------+-----------------------+---------------+-------+
505 * | <----- len ----->
508 static inline void dsa_alloc_etype_header(struct sk_buff *skb, int len)
510 memmove(skb->data, skb->data + len, 2 * ETH_ALEN);
513 /* On RX, eth_type_trans() on the DSA master pulls ETH_HLEN bytes starting from
514 * skb_mac_header(skb), which leaves skb->data pointing at the first byte after
515 * what the DSA master perceives as the EtherType (the beginning of the L3
516 * protocol). Since DSA EtherType header taggers treat the EtherType as part of
517 * the DSA tag itself, and the EtherType is 2 bytes in length, the DSA header
518 * is located 2 bytes behind skb->data. Note that EtherType in this context
519 * means the first 2 bytes of the DSA header, not the encapsulated EtherType
520 * that will become visible after the DSA header is stripped.
522 static inline void *dsa_etype_header_pos_rx(struct sk_buff *skb)
524 return skb->data - 2;
527 /* On TX, skb->data points to skb_mac_header(skb), which means that EtherType
528 * header taggers start exactly where the EtherType is (the EtherType is
529 * treated as part of the DSA header).
531 static inline void *dsa_etype_header_pos_tx(struct sk_buff *skb)
533 return skb->data + 2 * ETH_ALEN;
537 int dsa_switch_register_notifier(struct dsa_switch *ds);
538 void dsa_switch_unregister_notifier(struct dsa_switch *ds);
541 void dsa_lag_map(struct dsa_switch_tree *dst, struct net_device *lag);
542 void dsa_lag_unmap(struct dsa_switch_tree *dst, struct net_device *lag);
543 int dsa_tree_notify(struct dsa_switch_tree *dst, unsigned long e, void *v);
544 int dsa_broadcast(unsigned long e, void *v);
545 int dsa_tree_change_tag_proto(struct dsa_switch_tree *dst,
546 struct net_device *master,
547 const struct dsa_device_ops *tag_ops,
548 const struct dsa_device_ops *old_tag_ops);
549 int dsa_bridge_num_get(const struct net_device *bridge_dev, int max);
550 void dsa_bridge_num_put(const struct net_device *bridge_dev, int bridge_num);
553 int dsa_tag_8021q_bridge_join(struct dsa_switch *ds,
554 struct dsa_notifier_bridge_info *info);
555 int dsa_tag_8021q_bridge_leave(struct dsa_switch *ds,
556 struct dsa_notifier_bridge_info *info);
557 int dsa_switch_tag_8021q_vlan_add(struct dsa_switch *ds,
558 struct dsa_notifier_tag_8021q_vlan_info *info);
559 int dsa_switch_tag_8021q_vlan_del(struct dsa_switch *ds,
560 struct dsa_notifier_tag_8021q_vlan_info *info);
562 extern struct list_head dsa_tree_list;