1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * Handling of a single switch chip, part of a switch fabric
5 * Copyright (c) 2017 Savoir-faire Linux Inc.
9 #include <linux/if_bridge.h>
10 #include <linux/netdevice.h>
11 #include <linux/notifier.h>
12 #include <linux/if_vlan.h>
13 #include <net/switchdev.h>
17 static unsigned int dsa_switch_fastest_ageing_time(struct dsa_switch *ds,
18 unsigned int ageing_time)
22 dsa_switch_for_each_port(dp, ds)
23 if (dp->ageing_time && dp->ageing_time < ageing_time)
24 ageing_time = dp->ageing_time;
29 static int dsa_switch_ageing_time(struct dsa_switch *ds,
30 struct dsa_notifier_ageing_time_info *info)
32 unsigned int ageing_time = info->ageing_time;
34 if (ds->ageing_time_min && ageing_time < ds->ageing_time_min)
37 if (ds->ageing_time_max && ageing_time > ds->ageing_time_max)
40 /* Program the fastest ageing time in case of multiple bridges */
41 ageing_time = dsa_switch_fastest_ageing_time(ds, ageing_time);
43 if (ds->ops->set_ageing_time)
44 return ds->ops->set_ageing_time(ds, ageing_time);
49 static bool dsa_port_mtu_match(struct dsa_port *dp,
50 struct dsa_notifier_mtu_info *info)
52 return dp == info->dp || dsa_port_is_dsa(dp) || dsa_port_is_cpu(dp);
55 static int dsa_switch_mtu(struct dsa_switch *ds,
56 struct dsa_notifier_mtu_info *info)
61 if (!ds->ops->port_change_mtu)
64 dsa_switch_for_each_port(dp, ds) {
65 if (dsa_port_mtu_match(dp, info)) {
66 ret = ds->ops->port_change_mtu(ds, dp->index,
76 static int dsa_switch_bridge_join(struct dsa_switch *ds,
77 struct dsa_notifier_bridge_info *info)
81 if (info->dp->ds == ds) {
82 if (!ds->ops->port_bridge_join)
85 err = ds->ops->port_bridge_join(ds, info->dp->index,
87 &info->tx_fwd_offload,
93 if (info->dp->ds != ds && ds->ops->crosschip_bridge_join) {
94 err = ds->ops->crosschip_bridge_join(ds,
95 info->dp->ds->dst->index,
107 static int dsa_switch_bridge_leave(struct dsa_switch *ds,
108 struct dsa_notifier_bridge_info *info)
110 if (info->dp->ds == ds && ds->ops->port_bridge_leave)
111 ds->ops->port_bridge_leave(ds, info->dp->index, info->bridge);
113 if (info->dp->ds != ds && ds->ops->crosschip_bridge_leave)
114 ds->ops->crosschip_bridge_leave(ds, info->dp->ds->dst->index,
122 /* Matches for all upstream-facing ports (the CPU port and all upstream-facing
123 * DSA links) that sit between the targeted port on which the notifier was
124 * emitted and its dedicated CPU port.
126 static bool dsa_port_host_address_match(struct dsa_port *dp,
127 const struct dsa_port *targeted_dp)
129 struct dsa_port *cpu_dp = targeted_dp->cpu_dp;
131 if (dsa_switch_is_upstream_of(dp->ds, targeted_dp->ds))
132 return dp->index == dsa_towards_port(dp->ds, cpu_dp->ds->index,
138 static struct dsa_mac_addr *dsa_mac_addr_find(struct list_head *addr_list,
139 const unsigned char *addr, u16 vid,
142 struct dsa_mac_addr *a;
144 list_for_each_entry(a, addr_list, list)
145 if (ether_addr_equal(a->addr, addr) && a->vid == vid &&
146 dsa_db_equal(&a->db, &db))
152 static int dsa_port_do_mdb_add(struct dsa_port *dp,
153 const struct switchdev_obj_port_mdb *mdb,
156 struct dsa_switch *ds = dp->ds;
157 struct dsa_mac_addr *a;
158 int port = dp->index;
161 /* No need to bother with refcounting for user ports */
162 if (!(dsa_port_is_cpu(dp) || dsa_port_is_dsa(dp)))
163 return ds->ops->port_mdb_add(ds, port, mdb, db);
165 mutex_lock(&dp->addr_lists_lock);
167 a = dsa_mac_addr_find(&dp->mdbs, mdb->addr, mdb->vid, db);
169 refcount_inc(&a->refcount);
173 a = kzalloc(sizeof(*a), GFP_KERNEL);
179 err = ds->ops->port_mdb_add(ds, port, mdb, db);
185 ether_addr_copy(a->addr, mdb->addr);
188 refcount_set(&a->refcount, 1);
189 list_add_tail(&a->list, &dp->mdbs);
192 mutex_unlock(&dp->addr_lists_lock);
197 static int dsa_port_do_mdb_del(struct dsa_port *dp,
198 const struct switchdev_obj_port_mdb *mdb,
201 struct dsa_switch *ds = dp->ds;
202 struct dsa_mac_addr *a;
203 int port = dp->index;
206 /* No need to bother with refcounting for user ports */
207 if (!(dsa_port_is_cpu(dp) || dsa_port_is_dsa(dp)))
208 return ds->ops->port_mdb_del(ds, port, mdb, db);
210 mutex_lock(&dp->addr_lists_lock);
212 a = dsa_mac_addr_find(&dp->mdbs, mdb->addr, mdb->vid, db);
218 if (!refcount_dec_and_test(&a->refcount))
221 err = ds->ops->port_mdb_del(ds, port, mdb, db);
223 refcount_set(&a->refcount, 1);
231 mutex_unlock(&dp->addr_lists_lock);
236 static int dsa_port_do_fdb_add(struct dsa_port *dp, const unsigned char *addr,
237 u16 vid, struct dsa_db db)
239 struct dsa_switch *ds = dp->ds;
240 struct dsa_mac_addr *a;
241 int port = dp->index;
244 /* No need to bother with refcounting for user ports */
245 if (!(dsa_port_is_cpu(dp) || dsa_port_is_dsa(dp)))
246 return ds->ops->port_fdb_add(ds, port, addr, vid, db);
248 mutex_lock(&dp->addr_lists_lock);
250 a = dsa_mac_addr_find(&dp->fdbs, addr, vid, db);
252 refcount_inc(&a->refcount);
256 a = kzalloc(sizeof(*a), GFP_KERNEL);
262 err = ds->ops->port_fdb_add(ds, port, addr, vid, db);
268 ether_addr_copy(a->addr, addr);
271 refcount_set(&a->refcount, 1);
272 list_add_tail(&a->list, &dp->fdbs);
275 mutex_unlock(&dp->addr_lists_lock);
280 static int dsa_port_do_fdb_del(struct dsa_port *dp, const unsigned char *addr,
281 u16 vid, struct dsa_db db)
283 struct dsa_switch *ds = dp->ds;
284 struct dsa_mac_addr *a;
285 int port = dp->index;
288 /* No need to bother with refcounting for user ports */
289 if (!(dsa_port_is_cpu(dp) || dsa_port_is_dsa(dp)))
290 return ds->ops->port_fdb_del(ds, port, addr, vid, db);
292 mutex_lock(&dp->addr_lists_lock);
294 a = dsa_mac_addr_find(&dp->fdbs, addr, vid, db);
300 if (!refcount_dec_and_test(&a->refcount))
303 err = ds->ops->port_fdb_del(ds, port, addr, vid, db);
305 refcount_set(&a->refcount, 1);
313 mutex_unlock(&dp->addr_lists_lock);
318 static int dsa_switch_do_lag_fdb_add(struct dsa_switch *ds, struct dsa_lag *lag,
319 const unsigned char *addr, u16 vid,
322 struct dsa_mac_addr *a;
325 mutex_lock(&lag->fdb_lock);
327 a = dsa_mac_addr_find(&lag->fdbs, addr, vid, db);
329 refcount_inc(&a->refcount);
333 a = kzalloc(sizeof(*a), GFP_KERNEL);
339 err = ds->ops->lag_fdb_add(ds, *lag, addr, vid, db);
345 ether_addr_copy(a->addr, addr);
348 refcount_set(&a->refcount, 1);
349 list_add_tail(&a->list, &lag->fdbs);
352 mutex_unlock(&lag->fdb_lock);
357 static int dsa_switch_do_lag_fdb_del(struct dsa_switch *ds, struct dsa_lag *lag,
358 const unsigned char *addr, u16 vid,
361 struct dsa_mac_addr *a;
364 mutex_lock(&lag->fdb_lock);
366 a = dsa_mac_addr_find(&lag->fdbs, addr, vid, db);
372 if (!refcount_dec_and_test(&a->refcount))
375 err = ds->ops->lag_fdb_del(ds, *lag, addr, vid, db);
377 refcount_set(&a->refcount, 1);
385 mutex_unlock(&lag->fdb_lock);
390 static int dsa_switch_host_fdb_add(struct dsa_switch *ds,
391 struct dsa_notifier_fdb_info *info)
396 if (!ds->ops->port_fdb_add)
399 dsa_switch_for_each_port(dp, ds) {
400 if (dsa_port_host_address_match(dp, info->dp)) {
401 if (dsa_port_is_cpu(dp) && info->dp->cpu_port_in_lag) {
402 err = dsa_switch_do_lag_fdb_add(ds, dp->lag,
407 err = dsa_port_do_fdb_add(dp, info->addr,
408 info->vid, info->db);
418 static int dsa_switch_host_fdb_del(struct dsa_switch *ds,
419 struct dsa_notifier_fdb_info *info)
424 if (!ds->ops->port_fdb_del)
427 dsa_switch_for_each_port(dp, ds) {
428 if (dsa_port_host_address_match(dp, info->dp)) {
429 if (dsa_port_is_cpu(dp) && info->dp->cpu_port_in_lag) {
430 err = dsa_switch_do_lag_fdb_del(ds, dp->lag,
435 err = dsa_port_do_fdb_del(dp, info->addr,
436 info->vid, info->db);
446 static int dsa_switch_fdb_add(struct dsa_switch *ds,
447 struct dsa_notifier_fdb_info *info)
449 int port = dsa_towards_port(ds, info->dp->ds->index, info->dp->index);
450 struct dsa_port *dp = dsa_to_port(ds, port);
452 if (!ds->ops->port_fdb_add)
455 return dsa_port_do_fdb_add(dp, info->addr, info->vid, info->db);
458 static int dsa_switch_fdb_del(struct dsa_switch *ds,
459 struct dsa_notifier_fdb_info *info)
461 int port = dsa_towards_port(ds, info->dp->ds->index, info->dp->index);
462 struct dsa_port *dp = dsa_to_port(ds, port);
464 if (!ds->ops->port_fdb_del)
467 return dsa_port_do_fdb_del(dp, info->addr, info->vid, info->db);
470 static int dsa_switch_lag_fdb_add(struct dsa_switch *ds,
471 struct dsa_notifier_lag_fdb_info *info)
475 if (!ds->ops->lag_fdb_add)
478 /* Notify switch only if it has a port in this LAG */
479 dsa_switch_for_each_port(dp, ds)
480 if (dsa_port_offloads_lag(dp, info->lag))
481 return dsa_switch_do_lag_fdb_add(ds, info->lag,
482 info->addr, info->vid,
488 static int dsa_switch_lag_fdb_del(struct dsa_switch *ds,
489 struct dsa_notifier_lag_fdb_info *info)
493 if (!ds->ops->lag_fdb_del)
496 /* Notify switch only if it has a port in this LAG */
497 dsa_switch_for_each_port(dp, ds)
498 if (dsa_port_offloads_lag(dp, info->lag))
499 return dsa_switch_do_lag_fdb_del(ds, info->lag,
500 info->addr, info->vid,
506 static int dsa_switch_lag_change(struct dsa_switch *ds,
507 struct dsa_notifier_lag_info *info)
509 if (info->dp->ds == ds && ds->ops->port_lag_change)
510 return ds->ops->port_lag_change(ds, info->dp->index);
512 if (info->dp->ds != ds && ds->ops->crosschip_lag_change)
513 return ds->ops->crosschip_lag_change(ds, info->dp->ds->index,
519 static int dsa_switch_lag_join(struct dsa_switch *ds,
520 struct dsa_notifier_lag_info *info)
522 if (info->dp->ds == ds && ds->ops->port_lag_join)
523 return ds->ops->port_lag_join(ds, info->dp->index, info->lag,
524 info->info, info->extack);
526 if (info->dp->ds != ds && ds->ops->crosschip_lag_join)
527 return ds->ops->crosschip_lag_join(ds, info->dp->ds->index,
528 info->dp->index, info->lag,
529 info->info, info->extack);
534 static int dsa_switch_lag_leave(struct dsa_switch *ds,
535 struct dsa_notifier_lag_info *info)
537 if (info->dp->ds == ds && ds->ops->port_lag_leave)
538 return ds->ops->port_lag_leave(ds, info->dp->index, info->lag);
540 if (info->dp->ds != ds && ds->ops->crosschip_lag_leave)
541 return ds->ops->crosschip_lag_leave(ds, info->dp->ds->index,
542 info->dp->index, info->lag);
547 static int dsa_switch_mdb_add(struct dsa_switch *ds,
548 struct dsa_notifier_mdb_info *info)
550 int port = dsa_towards_port(ds, info->dp->ds->index, info->dp->index);
551 struct dsa_port *dp = dsa_to_port(ds, port);
553 if (!ds->ops->port_mdb_add)
556 return dsa_port_do_mdb_add(dp, info->mdb, info->db);
559 static int dsa_switch_mdb_del(struct dsa_switch *ds,
560 struct dsa_notifier_mdb_info *info)
562 int port = dsa_towards_port(ds, info->dp->ds->index, info->dp->index);
563 struct dsa_port *dp = dsa_to_port(ds, port);
565 if (!ds->ops->port_mdb_del)
568 return dsa_port_do_mdb_del(dp, info->mdb, info->db);
571 static int dsa_switch_host_mdb_add(struct dsa_switch *ds,
572 struct dsa_notifier_mdb_info *info)
577 if (!ds->ops->port_mdb_add)
580 dsa_switch_for_each_port(dp, ds) {
581 if (dsa_port_host_address_match(dp, info->dp)) {
582 err = dsa_port_do_mdb_add(dp, info->mdb, info->db);
591 static int dsa_switch_host_mdb_del(struct dsa_switch *ds,
592 struct dsa_notifier_mdb_info *info)
597 if (!ds->ops->port_mdb_del)
600 dsa_switch_for_each_port(dp, ds) {
601 if (dsa_port_host_address_match(dp, info->dp)) {
602 err = dsa_port_do_mdb_del(dp, info->mdb, info->db);
611 /* Port VLANs match on the targeted port and on all DSA ports */
612 static bool dsa_port_vlan_match(struct dsa_port *dp,
613 struct dsa_notifier_vlan_info *info)
615 return dsa_port_is_dsa(dp) || dp == info->dp;
618 /* Host VLANs match on the targeted port's CPU port, and on all DSA ports
619 * (upstream and downstream) of that switch and its upstream switches.
621 static bool dsa_port_host_vlan_match(struct dsa_port *dp,
622 const struct dsa_port *targeted_dp)
624 struct dsa_port *cpu_dp = targeted_dp->cpu_dp;
626 if (dsa_switch_is_upstream_of(dp->ds, targeted_dp->ds))
627 return dsa_port_is_dsa(dp) || dp == cpu_dp;
632 static struct dsa_vlan *dsa_vlan_find(struct list_head *vlan_list,
633 const struct switchdev_obj_port_vlan *vlan)
637 list_for_each_entry(v, vlan_list, list)
638 if (v->vid == vlan->vid)
644 static int dsa_port_do_vlan_add(struct dsa_port *dp,
645 const struct switchdev_obj_port_vlan *vlan,
646 struct netlink_ext_ack *extack)
648 struct dsa_switch *ds = dp->ds;
649 int port = dp->index;
653 /* No need to bother with refcounting for user ports. */
654 if (!(dsa_port_is_cpu(dp) || dsa_port_is_dsa(dp)))
655 return ds->ops->port_vlan_add(ds, port, vlan, extack);
657 /* No need to propagate on shared ports the existing VLANs that were
658 * re-notified after just the flags have changed. This would cause a
659 * refcount bump which we need to avoid, since it unbalances the
660 * additions with the deletions.
665 mutex_lock(&dp->vlans_lock);
667 v = dsa_vlan_find(&dp->vlans, vlan);
669 refcount_inc(&v->refcount);
673 v = kzalloc(sizeof(*v), GFP_KERNEL);
679 err = ds->ops->port_vlan_add(ds, port, vlan, extack);
686 refcount_set(&v->refcount, 1);
687 list_add_tail(&v->list, &dp->vlans);
690 mutex_unlock(&dp->vlans_lock);
695 static int dsa_port_do_vlan_del(struct dsa_port *dp,
696 const struct switchdev_obj_port_vlan *vlan)
698 struct dsa_switch *ds = dp->ds;
699 int port = dp->index;
703 /* No need to bother with refcounting for user ports */
704 if (!(dsa_port_is_cpu(dp) || dsa_port_is_dsa(dp)))
705 return ds->ops->port_vlan_del(ds, port, vlan);
707 mutex_lock(&dp->vlans_lock);
709 v = dsa_vlan_find(&dp->vlans, vlan);
715 if (!refcount_dec_and_test(&v->refcount))
718 err = ds->ops->port_vlan_del(ds, port, vlan);
720 refcount_set(&v->refcount, 1);
728 mutex_unlock(&dp->vlans_lock);
733 static int dsa_switch_vlan_add(struct dsa_switch *ds,
734 struct dsa_notifier_vlan_info *info)
739 if (!ds->ops->port_vlan_add)
742 dsa_switch_for_each_port(dp, ds) {
743 if (dsa_port_vlan_match(dp, info)) {
744 err = dsa_port_do_vlan_add(dp, info->vlan,
754 static int dsa_switch_vlan_del(struct dsa_switch *ds,
755 struct dsa_notifier_vlan_info *info)
760 if (!ds->ops->port_vlan_del)
763 dsa_switch_for_each_port(dp, ds) {
764 if (dsa_port_vlan_match(dp, info)) {
765 err = dsa_port_do_vlan_del(dp, info->vlan);
774 static int dsa_switch_host_vlan_add(struct dsa_switch *ds,
775 struct dsa_notifier_vlan_info *info)
780 if (!ds->ops->port_vlan_add)
783 dsa_switch_for_each_port(dp, ds) {
784 if (dsa_port_host_vlan_match(dp, info->dp)) {
785 err = dsa_port_do_vlan_add(dp, info->vlan,
795 static int dsa_switch_host_vlan_del(struct dsa_switch *ds,
796 struct dsa_notifier_vlan_info *info)
801 if (!ds->ops->port_vlan_del)
804 dsa_switch_for_each_port(dp, ds) {
805 if (dsa_port_host_vlan_match(dp, info->dp)) {
806 err = dsa_port_do_vlan_del(dp, info->vlan);
815 static int dsa_switch_change_tag_proto(struct dsa_switch *ds,
816 struct dsa_notifier_tag_proto_info *info)
818 const struct dsa_device_ops *tag_ops = info->tag_ops;
819 struct dsa_port *dp, *cpu_dp;
822 if (!ds->ops->change_tag_protocol)
827 err = ds->ops->change_tag_protocol(ds, tag_ops->proto);
831 dsa_switch_for_each_cpu_port(cpu_dp, ds)
832 dsa_port_set_tag_protocol(cpu_dp, tag_ops);
834 /* Now that changing the tag protocol can no longer fail, let's update
835 * the remaining bits which are "duplicated for faster access", and the
836 * bits that depend on the tagger, such as the MTU.
838 dsa_switch_for_each_user_port(dp, ds) {
839 struct net_device *slave = dp->slave;
841 dsa_slave_setup_tagger(slave);
843 /* rtnl_mutex is held in dsa_tree_change_tag_proto */
844 dsa_slave_change_mtu(slave, slave->mtu);
850 /* We use the same cross-chip notifiers to inform both the tagger side, as well
851 * as the switch side, of connection and disconnection events.
852 * Since ds->tagger_data is owned by the tagger, it isn't a hard error if the
853 * switch side doesn't support connecting to this tagger, and therefore, the
854 * fact that we don't disconnect the tagger side doesn't constitute a memory
855 * leak: the tagger will still operate with persistent per-switch memory, just
856 * with the switch side unconnected to it. What does constitute a hard error is
857 * when the switch side supports connecting but fails.
860 dsa_switch_connect_tag_proto(struct dsa_switch *ds,
861 struct dsa_notifier_tag_proto_info *info)
863 const struct dsa_device_ops *tag_ops = info->tag_ops;
866 /* Notify the new tagger about the connection to this switch */
867 if (tag_ops->connect) {
868 err = tag_ops->connect(ds);
873 if (!ds->ops->connect_tag_protocol)
876 /* Notify the switch about the connection to the new tagger */
877 err = ds->ops->connect_tag_protocol(ds, tag_ops->proto);
879 /* Revert the new tagger's connection to this tree */
880 if (tag_ops->disconnect)
881 tag_ops->disconnect(ds);
889 dsa_switch_disconnect_tag_proto(struct dsa_switch *ds,
890 struct dsa_notifier_tag_proto_info *info)
892 const struct dsa_device_ops *tag_ops = info->tag_ops;
894 /* Notify the tagger about the disconnection from this switch */
895 if (tag_ops->disconnect && ds->tagger_data)
896 tag_ops->disconnect(ds);
898 /* No need to notify the switch, since it shouldn't have any
899 * resources to tear down
905 dsa_switch_master_state_change(struct dsa_switch *ds,
906 struct dsa_notifier_master_state_info *info)
908 if (!ds->ops->master_state_change)
911 ds->ops->master_state_change(ds, info->master, info->operational);
916 static int dsa_switch_event(struct notifier_block *nb,
917 unsigned long event, void *info)
919 struct dsa_switch *ds = container_of(nb, struct dsa_switch, nb);
923 case DSA_NOTIFIER_AGEING_TIME:
924 err = dsa_switch_ageing_time(ds, info);
926 case DSA_NOTIFIER_BRIDGE_JOIN:
927 err = dsa_switch_bridge_join(ds, info);
929 case DSA_NOTIFIER_BRIDGE_LEAVE:
930 err = dsa_switch_bridge_leave(ds, info);
932 case DSA_NOTIFIER_FDB_ADD:
933 err = dsa_switch_fdb_add(ds, info);
935 case DSA_NOTIFIER_FDB_DEL:
936 err = dsa_switch_fdb_del(ds, info);
938 case DSA_NOTIFIER_HOST_FDB_ADD:
939 err = dsa_switch_host_fdb_add(ds, info);
941 case DSA_NOTIFIER_HOST_FDB_DEL:
942 err = dsa_switch_host_fdb_del(ds, info);
944 case DSA_NOTIFIER_LAG_FDB_ADD:
945 err = dsa_switch_lag_fdb_add(ds, info);
947 case DSA_NOTIFIER_LAG_FDB_DEL:
948 err = dsa_switch_lag_fdb_del(ds, info);
950 case DSA_NOTIFIER_LAG_CHANGE:
951 err = dsa_switch_lag_change(ds, info);
953 case DSA_NOTIFIER_LAG_JOIN:
954 err = dsa_switch_lag_join(ds, info);
956 case DSA_NOTIFIER_LAG_LEAVE:
957 err = dsa_switch_lag_leave(ds, info);
959 case DSA_NOTIFIER_MDB_ADD:
960 err = dsa_switch_mdb_add(ds, info);
962 case DSA_NOTIFIER_MDB_DEL:
963 err = dsa_switch_mdb_del(ds, info);
965 case DSA_NOTIFIER_HOST_MDB_ADD:
966 err = dsa_switch_host_mdb_add(ds, info);
968 case DSA_NOTIFIER_HOST_MDB_DEL:
969 err = dsa_switch_host_mdb_del(ds, info);
971 case DSA_NOTIFIER_VLAN_ADD:
972 err = dsa_switch_vlan_add(ds, info);
974 case DSA_NOTIFIER_VLAN_DEL:
975 err = dsa_switch_vlan_del(ds, info);
977 case DSA_NOTIFIER_HOST_VLAN_ADD:
978 err = dsa_switch_host_vlan_add(ds, info);
980 case DSA_NOTIFIER_HOST_VLAN_DEL:
981 err = dsa_switch_host_vlan_del(ds, info);
983 case DSA_NOTIFIER_MTU:
984 err = dsa_switch_mtu(ds, info);
986 case DSA_NOTIFIER_TAG_PROTO:
987 err = dsa_switch_change_tag_proto(ds, info);
989 case DSA_NOTIFIER_TAG_PROTO_CONNECT:
990 err = dsa_switch_connect_tag_proto(ds, info);
992 case DSA_NOTIFIER_TAG_PROTO_DISCONNECT:
993 err = dsa_switch_disconnect_tag_proto(ds, info);
995 case DSA_NOTIFIER_TAG_8021Q_VLAN_ADD:
996 err = dsa_switch_tag_8021q_vlan_add(ds, info);
998 case DSA_NOTIFIER_TAG_8021Q_VLAN_DEL:
999 err = dsa_switch_tag_8021q_vlan_del(ds, info);
1001 case DSA_NOTIFIER_MASTER_STATE_CHANGE:
1002 err = dsa_switch_master_state_change(ds, info);
1010 dev_dbg(ds->dev, "breaking chain for DSA event %lu (%d)\n",
1013 return notifier_from_errno(err);
1016 int dsa_switch_register_notifier(struct dsa_switch *ds)
1018 ds->nb.notifier_call = dsa_switch_event;
1020 return raw_notifier_chain_register(&ds->dst->nh, &ds->nb);
1023 void dsa_switch_unregister_notifier(struct dsa_switch *ds)
1027 err = raw_notifier_chain_unregister(&ds->dst->nh, &ds->nb);
1029 dev_err(ds->dev, "failed to unregister notifier (%d)\n", err);