debd2c466f11cbd60e355dee0190f6e4053aa585
[platform/kernel/linux-starfive.git] / drivers / net / ethernet / mellanox / mlxsw / spectrum_router.c
1 // SPDX-License-Identifier: BSD-3-Clause OR GPL-2.0
2 /* Copyright (c) 2016-2018 Mellanox Technologies. All rights reserved */
3
4 #include <linux/kernel.h>
5 #include <linux/types.h>
6 #include <linux/rhashtable.h>
7 #include <linux/bitops.h>
8 #include <linux/in6.h>
9 #include <linux/notifier.h>
10 #include <linux/inetdevice.h>
11 #include <linux/netdevice.h>
12 #include <linux/if_bridge.h>
13 #include <linux/socket.h>
14 #include <linux/route.h>
15 #include <linux/gcd.h>
16 #include <linux/if_macvlan.h>
17 #include <linux/refcount.h>
18 #include <linux/jhash.h>
19 #include <linux/net_namespace.h>
20 #include <linux/mutex.h>
21 #include <linux/genalloc.h>
22 #include <net/netevent.h>
23 #include <net/neighbour.h>
24 #include <net/arp.h>
25 #include <net/inet_dscp.h>
26 #include <net/ip_fib.h>
27 #include <net/ip6_fib.h>
28 #include <net/nexthop.h>
29 #include <net/fib_rules.h>
30 #include <net/ip_tunnels.h>
31 #include <net/l3mdev.h>
32 #include <net/addrconf.h>
33 #include <net/ndisc.h>
34 #include <net/ipv6.h>
35 #include <net/fib_notifier.h>
36 #include <net/switchdev.h>
37
38 #include "spectrum.h"
39 #include "core.h"
40 #include "reg.h"
41 #include "spectrum_cnt.h"
42 #include "spectrum_dpipe.h"
43 #include "spectrum_ipip.h"
44 #include "spectrum_mr.h"
45 #include "spectrum_mr_tcam.h"
46 #include "spectrum_router.h"
47 #include "spectrum_span.h"
48
49 struct mlxsw_sp_fib;
50 struct mlxsw_sp_vr;
51 struct mlxsw_sp_lpm_tree;
52 struct mlxsw_sp_rif_ops;
53
54 struct mlxsw_sp_crif_key {
55         struct net_device *dev;
56 };
57
58 struct mlxsw_sp_crif {
59         struct mlxsw_sp_crif_key key;
60         struct rhash_head ht_node;
61         bool can_destroy;
62         struct list_head nexthop_list;
63         struct mlxsw_sp_rif *rif;
64 };
65
66 static const struct rhashtable_params mlxsw_sp_crif_ht_params = {
67         .key_offset = offsetof(struct mlxsw_sp_crif, key),
68         .key_len = sizeof_field(struct mlxsw_sp_crif, key),
69         .head_offset = offsetof(struct mlxsw_sp_crif, ht_node),
70 };
71
72 struct mlxsw_sp_rif {
73         struct mlxsw_sp_crif *crif; /* NULL for underlay RIF */
74         netdevice_tracker dev_tracker;
75         struct list_head neigh_list;
76         struct mlxsw_sp_fid *fid;
77         unsigned char addr[ETH_ALEN];
78         int mtu;
79         u16 rif_index;
80         u8 mac_profile_id;
81         u8 rif_entries;
82         u16 vr_id;
83         const struct mlxsw_sp_rif_ops *ops;
84         struct mlxsw_sp *mlxsw_sp;
85
86         unsigned int counter_ingress;
87         bool counter_ingress_valid;
88         unsigned int counter_egress;
89         bool counter_egress_valid;
90 };
91
92 static struct net_device *mlxsw_sp_rif_dev(const struct mlxsw_sp_rif *rif)
93 {
94         if (!rif->crif)
95                 return NULL;
96         return rif->crif->key.dev;
97 }
98
99 struct mlxsw_sp_rif_params {
100         struct net_device *dev;
101         union {
102                 u16 system_port;
103                 u16 lag_id;
104         };
105         u16 vid;
106         bool lag;
107         bool double_entry;
108 };
109
110 struct mlxsw_sp_rif_subport {
111         struct mlxsw_sp_rif common;
112         refcount_t ref_count;
113         union {
114                 u16 system_port;
115                 u16 lag_id;
116         };
117         u16 vid;
118         bool lag;
119 };
120
121 struct mlxsw_sp_rif_ipip_lb {
122         struct mlxsw_sp_rif common;
123         struct mlxsw_sp_rif_ipip_lb_config lb_config;
124         u16 ul_vr_id;   /* Spectrum-1. */
125         u16 ul_rif_id;  /* Spectrum-2+. */
126 };
127
128 struct mlxsw_sp_rif_params_ipip_lb {
129         struct mlxsw_sp_rif_params common;
130         struct mlxsw_sp_rif_ipip_lb_config lb_config;
131 };
132
133 struct mlxsw_sp_rif_ops {
134         enum mlxsw_sp_rif_type type;
135         size_t rif_size;
136
137         void (*setup)(struct mlxsw_sp_rif *rif,
138                       const struct mlxsw_sp_rif_params *params);
139         int (*configure)(struct mlxsw_sp_rif *rif,
140                          struct netlink_ext_ack *extack);
141         void (*deconfigure)(struct mlxsw_sp_rif *rif);
142         struct mlxsw_sp_fid * (*fid_get)(struct mlxsw_sp_rif *rif,
143                                          const struct mlxsw_sp_rif_params *params,
144                                          struct netlink_ext_ack *extack);
145         void (*fdb_del)(struct mlxsw_sp_rif *rif, const char *mac);
146 };
147
148 struct mlxsw_sp_rif_mac_profile {
149         unsigned char mac_prefix[ETH_ALEN];
150         refcount_t ref_count;
151         u8 id;
152 };
153
154 struct mlxsw_sp_router_ops {
155         int (*init)(struct mlxsw_sp *mlxsw_sp);
156         int (*ipips_init)(struct mlxsw_sp *mlxsw_sp);
157 };
158
159 static struct mlxsw_sp_rif *
160 mlxsw_sp_rif_find_by_dev(const struct mlxsw_sp *mlxsw_sp,
161                          const struct net_device *dev);
162 static void mlxsw_sp_rif_destroy(struct mlxsw_sp_rif *rif);
163 static void mlxsw_sp_lpm_tree_hold(struct mlxsw_sp_lpm_tree *lpm_tree);
164 static void mlxsw_sp_lpm_tree_put(struct mlxsw_sp *mlxsw_sp,
165                                   struct mlxsw_sp_lpm_tree *lpm_tree);
166 static int mlxsw_sp_vr_lpm_tree_bind(struct mlxsw_sp *mlxsw_sp,
167                                      const struct mlxsw_sp_fib *fib,
168                                      u8 tree_id);
169 static int mlxsw_sp_vr_lpm_tree_unbind(struct mlxsw_sp *mlxsw_sp,
170                                        const struct mlxsw_sp_fib *fib);
171
172 static unsigned int *
173 mlxsw_sp_rif_p_counter_get(struct mlxsw_sp_rif *rif,
174                            enum mlxsw_sp_rif_counter_dir dir)
175 {
176         switch (dir) {
177         case MLXSW_SP_RIF_COUNTER_EGRESS:
178                 return &rif->counter_egress;
179         case MLXSW_SP_RIF_COUNTER_INGRESS:
180                 return &rif->counter_ingress;
181         }
182         return NULL;
183 }
184
185 static bool
186 mlxsw_sp_rif_counter_valid_get(struct mlxsw_sp_rif *rif,
187                                enum mlxsw_sp_rif_counter_dir dir)
188 {
189         switch (dir) {
190         case MLXSW_SP_RIF_COUNTER_EGRESS:
191                 return rif->counter_egress_valid;
192         case MLXSW_SP_RIF_COUNTER_INGRESS:
193                 return rif->counter_ingress_valid;
194         }
195         return false;
196 }
197
198 static void
199 mlxsw_sp_rif_counter_valid_set(struct mlxsw_sp_rif *rif,
200                                enum mlxsw_sp_rif_counter_dir dir,
201                                bool valid)
202 {
203         switch (dir) {
204         case MLXSW_SP_RIF_COUNTER_EGRESS:
205                 rif->counter_egress_valid = valid;
206                 break;
207         case MLXSW_SP_RIF_COUNTER_INGRESS:
208                 rif->counter_ingress_valid = valid;
209                 break;
210         }
211 }
212
213 static int mlxsw_sp_rif_counter_edit(struct mlxsw_sp *mlxsw_sp, u16 rif_index,
214                                      unsigned int counter_index, bool enable,
215                                      enum mlxsw_sp_rif_counter_dir dir)
216 {
217         char ritr_pl[MLXSW_REG_RITR_LEN];
218         bool is_egress = false;
219         int err;
220
221         if (dir == MLXSW_SP_RIF_COUNTER_EGRESS)
222                 is_egress = true;
223         mlxsw_reg_ritr_rif_pack(ritr_pl, rif_index);
224         err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
225         if (err)
226                 return err;
227
228         mlxsw_reg_ritr_counter_pack(ritr_pl, counter_index, enable,
229                                     is_egress);
230         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
231 }
232
233 int mlxsw_sp_rif_counter_value_get(struct mlxsw_sp *mlxsw_sp,
234                                    struct mlxsw_sp_rif *rif,
235                                    enum mlxsw_sp_rif_counter_dir dir, u64 *cnt)
236 {
237         char ricnt_pl[MLXSW_REG_RICNT_LEN];
238         unsigned int *p_counter_index;
239         bool valid;
240         int err;
241
242         valid = mlxsw_sp_rif_counter_valid_get(rif, dir);
243         if (!valid)
244                 return -EINVAL;
245
246         p_counter_index = mlxsw_sp_rif_p_counter_get(rif, dir);
247         if (!p_counter_index)
248                 return -EINVAL;
249         mlxsw_reg_ricnt_pack(ricnt_pl, *p_counter_index,
250                              MLXSW_REG_RICNT_OPCODE_NOP);
251         err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(ricnt), ricnt_pl);
252         if (err)
253                 return err;
254         *cnt = mlxsw_reg_ricnt_good_unicast_packets_get(ricnt_pl);
255         return 0;
256 }
257
258 struct mlxsw_sp_rif_counter_set_basic {
259         u64 good_unicast_packets;
260         u64 good_multicast_packets;
261         u64 good_broadcast_packets;
262         u64 good_unicast_bytes;
263         u64 good_multicast_bytes;
264         u64 good_broadcast_bytes;
265         u64 error_packets;
266         u64 discard_packets;
267         u64 error_bytes;
268         u64 discard_bytes;
269 };
270
271 static int
272 mlxsw_sp_rif_counter_fetch_clear(struct mlxsw_sp_rif *rif,
273                                  enum mlxsw_sp_rif_counter_dir dir,
274                                  struct mlxsw_sp_rif_counter_set_basic *set)
275 {
276         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
277         char ricnt_pl[MLXSW_REG_RICNT_LEN];
278         unsigned int *p_counter_index;
279         int err;
280
281         if (!mlxsw_sp_rif_counter_valid_get(rif, dir))
282                 return -EINVAL;
283
284         p_counter_index = mlxsw_sp_rif_p_counter_get(rif, dir);
285         if (!p_counter_index)
286                 return -EINVAL;
287
288         mlxsw_reg_ricnt_pack(ricnt_pl, *p_counter_index,
289                              MLXSW_REG_RICNT_OPCODE_CLEAR);
290         err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(ricnt), ricnt_pl);
291         if (err)
292                 return err;
293
294         if (!set)
295                 return 0;
296
297 #define MLXSW_SP_RIF_COUNTER_EXTRACT(NAME)                              \
298                 (set->NAME = mlxsw_reg_ricnt_ ## NAME ## _get(ricnt_pl))
299
300         MLXSW_SP_RIF_COUNTER_EXTRACT(good_unicast_packets);
301         MLXSW_SP_RIF_COUNTER_EXTRACT(good_multicast_packets);
302         MLXSW_SP_RIF_COUNTER_EXTRACT(good_broadcast_packets);
303         MLXSW_SP_RIF_COUNTER_EXTRACT(good_unicast_bytes);
304         MLXSW_SP_RIF_COUNTER_EXTRACT(good_multicast_bytes);
305         MLXSW_SP_RIF_COUNTER_EXTRACT(good_broadcast_bytes);
306         MLXSW_SP_RIF_COUNTER_EXTRACT(error_packets);
307         MLXSW_SP_RIF_COUNTER_EXTRACT(discard_packets);
308         MLXSW_SP_RIF_COUNTER_EXTRACT(error_bytes);
309         MLXSW_SP_RIF_COUNTER_EXTRACT(discard_bytes);
310
311 #undef MLXSW_SP_RIF_COUNTER_EXTRACT
312
313         return 0;
314 }
315
316 static int mlxsw_sp_rif_counter_clear(struct mlxsw_sp *mlxsw_sp,
317                                       unsigned int counter_index)
318 {
319         char ricnt_pl[MLXSW_REG_RICNT_LEN];
320
321         mlxsw_reg_ricnt_pack(ricnt_pl, counter_index,
322                              MLXSW_REG_RICNT_OPCODE_CLEAR);
323         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ricnt), ricnt_pl);
324 }
325
326 int mlxsw_sp_rif_counter_alloc(struct mlxsw_sp_rif *rif,
327                                enum mlxsw_sp_rif_counter_dir dir)
328 {
329         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
330         unsigned int *p_counter_index;
331         int err;
332
333         if (mlxsw_sp_rif_counter_valid_get(rif, dir))
334                 return 0;
335
336         p_counter_index = mlxsw_sp_rif_p_counter_get(rif, dir);
337         if (!p_counter_index)
338                 return -EINVAL;
339
340         err = mlxsw_sp_counter_alloc(mlxsw_sp, MLXSW_SP_COUNTER_SUB_POOL_RIF,
341                                      p_counter_index);
342         if (err)
343                 return err;
344
345         err = mlxsw_sp_rif_counter_clear(mlxsw_sp, *p_counter_index);
346         if (err)
347                 goto err_counter_clear;
348
349         err = mlxsw_sp_rif_counter_edit(mlxsw_sp, rif->rif_index,
350                                         *p_counter_index, true, dir);
351         if (err)
352                 goto err_counter_edit;
353         mlxsw_sp_rif_counter_valid_set(rif, dir, true);
354         return 0;
355
356 err_counter_edit:
357 err_counter_clear:
358         mlxsw_sp_counter_free(mlxsw_sp, MLXSW_SP_COUNTER_SUB_POOL_RIF,
359                               *p_counter_index);
360         return err;
361 }
362
363 void mlxsw_sp_rif_counter_free(struct mlxsw_sp_rif *rif,
364                                enum mlxsw_sp_rif_counter_dir dir)
365 {
366         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
367         unsigned int *p_counter_index;
368
369         if (!mlxsw_sp_rif_counter_valid_get(rif, dir))
370                 return;
371
372         p_counter_index = mlxsw_sp_rif_p_counter_get(rif, dir);
373         if (WARN_ON(!p_counter_index))
374                 return;
375         mlxsw_sp_rif_counter_edit(mlxsw_sp, rif->rif_index,
376                                   *p_counter_index, false, dir);
377         mlxsw_sp_counter_free(mlxsw_sp, MLXSW_SP_COUNTER_SUB_POOL_RIF,
378                               *p_counter_index);
379         mlxsw_sp_rif_counter_valid_set(rif, dir, false);
380 }
381
382 static void mlxsw_sp_rif_counters_alloc(struct mlxsw_sp_rif *rif)
383 {
384         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
385         struct devlink *devlink;
386
387         devlink = priv_to_devlink(mlxsw_sp->core);
388         if (!devlink_dpipe_table_counter_enabled(devlink,
389                                                  MLXSW_SP_DPIPE_TABLE_NAME_ERIF))
390                 return;
391         mlxsw_sp_rif_counter_alloc(rif, MLXSW_SP_RIF_COUNTER_EGRESS);
392 }
393
394 static void mlxsw_sp_rif_counters_free(struct mlxsw_sp_rif *rif)
395 {
396         mlxsw_sp_rif_counter_free(rif, MLXSW_SP_RIF_COUNTER_EGRESS);
397 }
398
399 #define MLXSW_SP_PREFIX_COUNT (sizeof(struct in6_addr) * BITS_PER_BYTE + 1)
400
401 struct mlxsw_sp_prefix_usage {
402         DECLARE_BITMAP(b, MLXSW_SP_PREFIX_COUNT);
403 };
404
405 #define mlxsw_sp_prefix_usage_for_each(prefix, prefix_usage) \
406         for_each_set_bit(prefix, (prefix_usage)->b, MLXSW_SP_PREFIX_COUNT)
407
408 static bool
409 mlxsw_sp_prefix_usage_eq(struct mlxsw_sp_prefix_usage *prefix_usage1,
410                          struct mlxsw_sp_prefix_usage *prefix_usage2)
411 {
412         return !memcmp(prefix_usage1, prefix_usage2, sizeof(*prefix_usage1));
413 }
414
415 static void
416 mlxsw_sp_prefix_usage_cpy(struct mlxsw_sp_prefix_usage *prefix_usage1,
417                           struct mlxsw_sp_prefix_usage *prefix_usage2)
418 {
419         memcpy(prefix_usage1, prefix_usage2, sizeof(*prefix_usage1));
420 }
421
422 static void
423 mlxsw_sp_prefix_usage_set(struct mlxsw_sp_prefix_usage *prefix_usage,
424                           unsigned char prefix_len)
425 {
426         set_bit(prefix_len, prefix_usage->b);
427 }
428
429 static void
430 mlxsw_sp_prefix_usage_clear(struct mlxsw_sp_prefix_usage *prefix_usage,
431                             unsigned char prefix_len)
432 {
433         clear_bit(prefix_len, prefix_usage->b);
434 }
435
436 struct mlxsw_sp_fib_key {
437         unsigned char addr[sizeof(struct in6_addr)];
438         unsigned char prefix_len;
439 };
440
441 enum mlxsw_sp_fib_entry_type {
442         MLXSW_SP_FIB_ENTRY_TYPE_REMOTE,
443         MLXSW_SP_FIB_ENTRY_TYPE_LOCAL,
444         MLXSW_SP_FIB_ENTRY_TYPE_TRAP,
445         MLXSW_SP_FIB_ENTRY_TYPE_BLACKHOLE,
446         MLXSW_SP_FIB_ENTRY_TYPE_UNREACHABLE,
447
448         /* This is a special case of local delivery, where a packet should be
449          * decapsulated on reception. Note that there is no corresponding ENCAP,
450          * because that's a type of next hop, not of FIB entry. (There can be
451          * several next hops in a REMOTE entry, and some of them may be
452          * encapsulating entries.)
453          */
454         MLXSW_SP_FIB_ENTRY_TYPE_IPIP_DECAP,
455         MLXSW_SP_FIB_ENTRY_TYPE_NVE_DECAP,
456 };
457
458 struct mlxsw_sp_nexthop_group_info;
459 struct mlxsw_sp_nexthop_group;
460 struct mlxsw_sp_fib_entry;
461
462 struct mlxsw_sp_fib_node {
463         struct mlxsw_sp_fib_entry *fib_entry;
464         struct list_head list;
465         struct rhash_head ht_node;
466         struct mlxsw_sp_fib *fib;
467         struct mlxsw_sp_fib_key key;
468 };
469
470 struct mlxsw_sp_fib_entry_decap {
471         struct mlxsw_sp_ipip_entry *ipip_entry;
472         u32 tunnel_index;
473 };
474
475 struct mlxsw_sp_fib_entry {
476         struct mlxsw_sp_fib_node *fib_node;
477         enum mlxsw_sp_fib_entry_type type;
478         struct list_head nexthop_group_node;
479         struct mlxsw_sp_nexthop_group *nh_group;
480         struct mlxsw_sp_fib_entry_decap decap; /* Valid for decap entries. */
481 };
482
483 struct mlxsw_sp_fib4_entry {
484         struct mlxsw_sp_fib_entry common;
485         struct fib_info *fi;
486         u32 tb_id;
487         dscp_t dscp;
488         u8 type;
489 };
490
491 struct mlxsw_sp_fib6_entry {
492         struct mlxsw_sp_fib_entry common;
493         struct list_head rt6_list;
494         unsigned int nrt6;
495 };
496
497 struct mlxsw_sp_rt6 {
498         struct list_head list;
499         struct fib6_info *rt;
500 };
501
502 struct mlxsw_sp_lpm_tree {
503         u8 id; /* tree ID */
504         unsigned int ref_count;
505         enum mlxsw_sp_l3proto proto;
506         unsigned long prefix_ref_count[MLXSW_SP_PREFIX_COUNT];
507         struct mlxsw_sp_prefix_usage prefix_usage;
508 };
509
510 struct mlxsw_sp_fib {
511         struct rhashtable ht;
512         struct list_head node_list;
513         struct mlxsw_sp_vr *vr;
514         struct mlxsw_sp_lpm_tree *lpm_tree;
515         enum mlxsw_sp_l3proto proto;
516 };
517
518 struct mlxsw_sp_vr {
519         u16 id; /* virtual router ID */
520         u32 tb_id; /* kernel fib table id */
521         unsigned int rif_count;
522         struct mlxsw_sp_fib *fib4;
523         struct mlxsw_sp_fib *fib6;
524         struct mlxsw_sp_mr_table *mr_table[MLXSW_SP_L3_PROTO_MAX];
525         struct mlxsw_sp_rif *ul_rif;
526         refcount_t ul_rif_refcnt;
527 };
528
529 static const struct rhashtable_params mlxsw_sp_fib_ht_params;
530
531 static struct mlxsw_sp_fib *mlxsw_sp_fib_create(struct mlxsw_sp *mlxsw_sp,
532                                                 struct mlxsw_sp_vr *vr,
533                                                 enum mlxsw_sp_l3proto proto)
534 {
535         struct mlxsw_sp_lpm_tree *lpm_tree;
536         struct mlxsw_sp_fib *fib;
537         int err;
538
539         lpm_tree = mlxsw_sp->router->lpm.proto_trees[proto];
540         fib = kzalloc(sizeof(*fib), GFP_KERNEL);
541         if (!fib)
542                 return ERR_PTR(-ENOMEM);
543         err = rhashtable_init(&fib->ht, &mlxsw_sp_fib_ht_params);
544         if (err)
545                 goto err_rhashtable_init;
546         INIT_LIST_HEAD(&fib->node_list);
547         fib->proto = proto;
548         fib->vr = vr;
549         fib->lpm_tree = lpm_tree;
550         mlxsw_sp_lpm_tree_hold(lpm_tree);
551         err = mlxsw_sp_vr_lpm_tree_bind(mlxsw_sp, fib, lpm_tree->id);
552         if (err)
553                 goto err_lpm_tree_bind;
554         return fib;
555
556 err_lpm_tree_bind:
557         mlxsw_sp_lpm_tree_put(mlxsw_sp, lpm_tree);
558 err_rhashtable_init:
559         kfree(fib);
560         return ERR_PTR(err);
561 }
562
563 static void mlxsw_sp_fib_destroy(struct mlxsw_sp *mlxsw_sp,
564                                  struct mlxsw_sp_fib *fib)
565 {
566         mlxsw_sp_vr_lpm_tree_unbind(mlxsw_sp, fib);
567         mlxsw_sp_lpm_tree_put(mlxsw_sp, fib->lpm_tree);
568         WARN_ON(!list_empty(&fib->node_list));
569         rhashtable_destroy(&fib->ht);
570         kfree(fib);
571 }
572
573 static struct mlxsw_sp_lpm_tree *
574 mlxsw_sp_lpm_tree_find_unused(struct mlxsw_sp *mlxsw_sp)
575 {
576         static struct mlxsw_sp_lpm_tree *lpm_tree;
577         int i;
578
579         for (i = 0; i < mlxsw_sp->router->lpm.tree_count; i++) {
580                 lpm_tree = &mlxsw_sp->router->lpm.trees[i];
581                 if (lpm_tree->ref_count == 0)
582                         return lpm_tree;
583         }
584         return NULL;
585 }
586
587 static int mlxsw_sp_lpm_tree_alloc(struct mlxsw_sp *mlxsw_sp,
588                                    struct mlxsw_sp_lpm_tree *lpm_tree)
589 {
590         char ralta_pl[MLXSW_REG_RALTA_LEN];
591
592         mlxsw_reg_ralta_pack(ralta_pl, true,
593                              (enum mlxsw_reg_ralxx_protocol) lpm_tree->proto,
594                              lpm_tree->id);
595         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralta), ralta_pl);
596 }
597
598 static void mlxsw_sp_lpm_tree_free(struct mlxsw_sp *mlxsw_sp,
599                                    struct mlxsw_sp_lpm_tree *lpm_tree)
600 {
601         char ralta_pl[MLXSW_REG_RALTA_LEN];
602
603         mlxsw_reg_ralta_pack(ralta_pl, false,
604                              (enum mlxsw_reg_ralxx_protocol) lpm_tree->proto,
605                              lpm_tree->id);
606         mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralta), ralta_pl);
607 }
608
609 static int
610 mlxsw_sp_lpm_tree_left_struct_set(struct mlxsw_sp *mlxsw_sp,
611                                   struct mlxsw_sp_prefix_usage *prefix_usage,
612                                   struct mlxsw_sp_lpm_tree *lpm_tree)
613 {
614         char ralst_pl[MLXSW_REG_RALST_LEN];
615         u8 root_bin = 0;
616         u8 prefix;
617         u8 last_prefix = MLXSW_REG_RALST_BIN_NO_CHILD;
618
619         mlxsw_sp_prefix_usage_for_each(prefix, prefix_usage)
620                 root_bin = prefix;
621
622         mlxsw_reg_ralst_pack(ralst_pl, root_bin, lpm_tree->id);
623         mlxsw_sp_prefix_usage_for_each(prefix, prefix_usage) {
624                 if (prefix == 0)
625                         continue;
626                 mlxsw_reg_ralst_bin_pack(ralst_pl, prefix, last_prefix,
627                                          MLXSW_REG_RALST_BIN_NO_CHILD);
628                 last_prefix = prefix;
629         }
630         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralst), ralst_pl);
631 }
632
633 static struct mlxsw_sp_lpm_tree *
634 mlxsw_sp_lpm_tree_create(struct mlxsw_sp *mlxsw_sp,
635                          struct mlxsw_sp_prefix_usage *prefix_usage,
636                          enum mlxsw_sp_l3proto proto)
637 {
638         struct mlxsw_sp_lpm_tree *lpm_tree;
639         int err;
640
641         lpm_tree = mlxsw_sp_lpm_tree_find_unused(mlxsw_sp);
642         if (!lpm_tree)
643                 return ERR_PTR(-EBUSY);
644         lpm_tree->proto = proto;
645         err = mlxsw_sp_lpm_tree_alloc(mlxsw_sp, lpm_tree);
646         if (err)
647                 return ERR_PTR(err);
648
649         err = mlxsw_sp_lpm_tree_left_struct_set(mlxsw_sp, prefix_usage,
650                                                 lpm_tree);
651         if (err)
652                 goto err_left_struct_set;
653         memcpy(&lpm_tree->prefix_usage, prefix_usage,
654                sizeof(lpm_tree->prefix_usage));
655         memset(&lpm_tree->prefix_ref_count, 0,
656                sizeof(lpm_tree->prefix_ref_count));
657         lpm_tree->ref_count = 1;
658         return lpm_tree;
659
660 err_left_struct_set:
661         mlxsw_sp_lpm_tree_free(mlxsw_sp, lpm_tree);
662         return ERR_PTR(err);
663 }
664
665 static void mlxsw_sp_lpm_tree_destroy(struct mlxsw_sp *mlxsw_sp,
666                                       struct mlxsw_sp_lpm_tree *lpm_tree)
667 {
668         mlxsw_sp_lpm_tree_free(mlxsw_sp, lpm_tree);
669 }
670
671 static struct mlxsw_sp_lpm_tree *
672 mlxsw_sp_lpm_tree_get(struct mlxsw_sp *mlxsw_sp,
673                       struct mlxsw_sp_prefix_usage *prefix_usage,
674                       enum mlxsw_sp_l3proto proto)
675 {
676         struct mlxsw_sp_lpm_tree *lpm_tree;
677         int i;
678
679         for (i = 0; i < mlxsw_sp->router->lpm.tree_count; i++) {
680                 lpm_tree = &mlxsw_sp->router->lpm.trees[i];
681                 if (lpm_tree->ref_count != 0 &&
682                     lpm_tree->proto == proto &&
683                     mlxsw_sp_prefix_usage_eq(&lpm_tree->prefix_usage,
684                                              prefix_usage)) {
685                         mlxsw_sp_lpm_tree_hold(lpm_tree);
686                         return lpm_tree;
687                 }
688         }
689         return mlxsw_sp_lpm_tree_create(mlxsw_sp, prefix_usage, proto);
690 }
691
692 static void mlxsw_sp_lpm_tree_hold(struct mlxsw_sp_lpm_tree *lpm_tree)
693 {
694         lpm_tree->ref_count++;
695 }
696
697 static void mlxsw_sp_lpm_tree_put(struct mlxsw_sp *mlxsw_sp,
698                                   struct mlxsw_sp_lpm_tree *lpm_tree)
699 {
700         if (--lpm_tree->ref_count == 0)
701                 mlxsw_sp_lpm_tree_destroy(mlxsw_sp, lpm_tree);
702 }
703
704 #define MLXSW_SP_LPM_TREE_MIN 1 /* tree 0 is reserved */
705
706 static int mlxsw_sp_lpm_init(struct mlxsw_sp *mlxsw_sp)
707 {
708         struct mlxsw_sp_prefix_usage req_prefix_usage = {{ 0 } };
709         struct mlxsw_sp_lpm_tree *lpm_tree;
710         u64 max_trees;
711         int err, i;
712
713         if (!MLXSW_CORE_RES_VALID(mlxsw_sp->core, MAX_LPM_TREES))
714                 return -EIO;
715
716         max_trees = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_LPM_TREES);
717         mlxsw_sp->router->lpm.tree_count = max_trees - MLXSW_SP_LPM_TREE_MIN;
718         mlxsw_sp->router->lpm.trees = kcalloc(mlxsw_sp->router->lpm.tree_count,
719                                              sizeof(struct mlxsw_sp_lpm_tree),
720                                              GFP_KERNEL);
721         if (!mlxsw_sp->router->lpm.trees)
722                 return -ENOMEM;
723
724         for (i = 0; i < mlxsw_sp->router->lpm.tree_count; i++) {
725                 lpm_tree = &mlxsw_sp->router->lpm.trees[i];
726                 lpm_tree->id = i + MLXSW_SP_LPM_TREE_MIN;
727         }
728
729         lpm_tree = mlxsw_sp_lpm_tree_get(mlxsw_sp, &req_prefix_usage,
730                                          MLXSW_SP_L3_PROTO_IPV4);
731         if (IS_ERR(lpm_tree)) {
732                 err = PTR_ERR(lpm_tree);
733                 goto err_ipv4_tree_get;
734         }
735         mlxsw_sp->router->lpm.proto_trees[MLXSW_SP_L3_PROTO_IPV4] = lpm_tree;
736
737         lpm_tree = mlxsw_sp_lpm_tree_get(mlxsw_sp, &req_prefix_usage,
738                                          MLXSW_SP_L3_PROTO_IPV6);
739         if (IS_ERR(lpm_tree)) {
740                 err = PTR_ERR(lpm_tree);
741                 goto err_ipv6_tree_get;
742         }
743         mlxsw_sp->router->lpm.proto_trees[MLXSW_SP_L3_PROTO_IPV6] = lpm_tree;
744
745         return 0;
746
747 err_ipv6_tree_get:
748         lpm_tree = mlxsw_sp->router->lpm.proto_trees[MLXSW_SP_L3_PROTO_IPV4];
749         mlxsw_sp_lpm_tree_put(mlxsw_sp, lpm_tree);
750 err_ipv4_tree_get:
751         kfree(mlxsw_sp->router->lpm.trees);
752         return err;
753 }
754
755 static void mlxsw_sp_lpm_fini(struct mlxsw_sp *mlxsw_sp)
756 {
757         struct mlxsw_sp_lpm_tree *lpm_tree;
758
759         lpm_tree = mlxsw_sp->router->lpm.proto_trees[MLXSW_SP_L3_PROTO_IPV6];
760         mlxsw_sp_lpm_tree_put(mlxsw_sp, lpm_tree);
761
762         lpm_tree = mlxsw_sp->router->lpm.proto_trees[MLXSW_SP_L3_PROTO_IPV4];
763         mlxsw_sp_lpm_tree_put(mlxsw_sp, lpm_tree);
764
765         kfree(mlxsw_sp->router->lpm.trees);
766 }
767
768 static bool mlxsw_sp_vr_is_used(const struct mlxsw_sp_vr *vr)
769 {
770         return !!vr->fib4 || !!vr->fib6 ||
771                !!vr->mr_table[MLXSW_SP_L3_PROTO_IPV4] ||
772                !!vr->mr_table[MLXSW_SP_L3_PROTO_IPV6];
773 }
774
775 static struct mlxsw_sp_vr *mlxsw_sp_vr_find_unused(struct mlxsw_sp *mlxsw_sp)
776 {
777         int max_vrs = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_VRS);
778         struct mlxsw_sp_vr *vr;
779         int i;
780
781         for (i = 0; i < max_vrs; i++) {
782                 vr = &mlxsw_sp->router->vrs[i];
783                 if (!mlxsw_sp_vr_is_used(vr))
784                         return vr;
785         }
786         return NULL;
787 }
788
789 static int mlxsw_sp_vr_lpm_tree_bind(struct mlxsw_sp *mlxsw_sp,
790                                      const struct mlxsw_sp_fib *fib, u8 tree_id)
791 {
792         char raltb_pl[MLXSW_REG_RALTB_LEN];
793
794         mlxsw_reg_raltb_pack(raltb_pl, fib->vr->id,
795                              (enum mlxsw_reg_ralxx_protocol) fib->proto,
796                              tree_id);
797         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(raltb), raltb_pl);
798 }
799
800 static int mlxsw_sp_vr_lpm_tree_unbind(struct mlxsw_sp *mlxsw_sp,
801                                        const struct mlxsw_sp_fib *fib)
802 {
803         char raltb_pl[MLXSW_REG_RALTB_LEN];
804
805         /* Bind to tree 0 which is default */
806         mlxsw_reg_raltb_pack(raltb_pl, fib->vr->id,
807                              (enum mlxsw_reg_ralxx_protocol) fib->proto, 0);
808         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(raltb), raltb_pl);
809 }
810
811 static u32 mlxsw_sp_fix_tb_id(u32 tb_id)
812 {
813         /* For our purpose, squash main, default and local tables into one */
814         if (tb_id == RT_TABLE_LOCAL || tb_id == RT_TABLE_DEFAULT)
815                 tb_id = RT_TABLE_MAIN;
816         return tb_id;
817 }
818
819 static struct mlxsw_sp_vr *mlxsw_sp_vr_find(struct mlxsw_sp *mlxsw_sp,
820                                             u32 tb_id)
821 {
822         int max_vrs = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_VRS);
823         struct mlxsw_sp_vr *vr;
824         int i;
825
826         tb_id = mlxsw_sp_fix_tb_id(tb_id);
827
828         for (i = 0; i < max_vrs; i++) {
829                 vr = &mlxsw_sp->router->vrs[i];
830                 if (mlxsw_sp_vr_is_used(vr) && vr->tb_id == tb_id)
831                         return vr;
832         }
833         return NULL;
834 }
835
836 int mlxsw_sp_router_tb_id_vr_id(struct mlxsw_sp *mlxsw_sp, u32 tb_id,
837                                 u16 *vr_id)
838 {
839         struct mlxsw_sp_vr *vr;
840         int err = 0;
841
842         mutex_lock(&mlxsw_sp->router->lock);
843         vr = mlxsw_sp_vr_find(mlxsw_sp, tb_id);
844         if (!vr) {
845                 err = -ESRCH;
846                 goto out;
847         }
848         *vr_id = vr->id;
849 out:
850         mutex_unlock(&mlxsw_sp->router->lock);
851         return err;
852 }
853
854 static struct mlxsw_sp_fib *mlxsw_sp_vr_fib(const struct mlxsw_sp_vr *vr,
855                                             enum mlxsw_sp_l3proto proto)
856 {
857         switch (proto) {
858         case MLXSW_SP_L3_PROTO_IPV4:
859                 return vr->fib4;
860         case MLXSW_SP_L3_PROTO_IPV6:
861                 return vr->fib6;
862         }
863         return NULL;
864 }
865
866 static struct mlxsw_sp_vr *mlxsw_sp_vr_create(struct mlxsw_sp *mlxsw_sp,
867                                               u32 tb_id,
868                                               struct netlink_ext_ack *extack)
869 {
870         struct mlxsw_sp_mr_table *mr4_table, *mr6_table;
871         struct mlxsw_sp_fib *fib4;
872         struct mlxsw_sp_fib *fib6;
873         struct mlxsw_sp_vr *vr;
874         int err;
875
876         vr = mlxsw_sp_vr_find_unused(mlxsw_sp);
877         if (!vr) {
878                 NL_SET_ERR_MSG_MOD(extack, "Exceeded number of supported virtual routers");
879                 return ERR_PTR(-EBUSY);
880         }
881         fib4 = mlxsw_sp_fib_create(mlxsw_sp, vr, MLXSW_SP_L3_PROTO_IPV4);
882         if (IS_ERR(fib4))
883                 return ERR_CAST(fib4);
884         fib6 = mlxsw_sp_fib_create(mlxsw_sp, vr, MLXSW_SP_L3_PROTO_IPV6);
885         if (IS_ERR(fib6)) {
886                 err = PTR_ERR(fib6);
887                 goto err_fib6_create;
888         }
889         mr4_table = mlxsw_sp_mr_table_create(mlxsw_sp, vr->id,
890                                              MLXSW_SP_L3_PROTO_IPV4);
891         if (IS_ERR(mr4_table)) {
892                 err = PTR_ERR(mr4_table);
893                 goto err_mr4_table_create;
894         }
895         mr6_table = mlxsw_sp_mr_table_create(mlxsw_sp, vr->id,
896                                              MLXSW_SP_L3_PROTO_IPV6);
897         if (IS_ERR(mr6_table)) {
898                 err = PTR_ERR(mr6_table);
899                 goto err_mr6_table_create;
900         }
901
902         vr->fib4 = fib4;
903         vr->fib6 = fib6;
904         vr->mr_table[MLXSW_SP_L3_PROTO_IPV4] = mr4_table;
905         vr->mr_table[MLXSW_SP_L3_PROTO_IPV6] = mr6_table;
906         vr->tb_id = tb_id;
907         return vr;
908
909 err_mr6_table_create:
910         mlxsw_sp_mr_table_destroy(mr4_table);
911 err_mr4_table_create:
912         mlxsw_sp_fib_destroy(mlxsw_sp, fib6);
913 err_fib6_create:
914         mlxsw_sp_fib_destroy(mlxsw_sp, fib4);
915         return ERR_PTR(err);
916 }
917
918 static void mlxsw_sp_vr_destroy(struct mlxsw_sp *mlxsw_sp,
919                                 struct mlxsw_sp_vr *vr)
920 {
921         mlxsw_sp_mr_table_destroy(vr->mr_table[MLXSW_SP_L3_PROTO_IPV6]);
922         vr->mr_table[MLXSW_SP_L3_PROTO_IPV6] = NULL;
923         mlxsw_sp_mr_table_destroy(vr->mr_table[MLXSW_SP_L3_PROTO_IPV4]);
924         vr->mr_table[MLXSW_SP_L3_PROTO_IPV4] = NULL;
925         mlxsw_sp_fib_destroy(mlxsw_sp, vr->fib6);
926         vr->fib6 = NULL;
927         mlxsw_sp_fib_destroy(mlxsw_sp, vr->fib4);
928         vr->fib4 = NULL;
929 }
930
931 static struct mlxsw_sp_vr *mlxsw_sp_vr_get(struct mlxsw_sp *mlxsw_sp, u32 tb_id,
932                                            struct netlink_ext_ack *extack)
933 {
934         struct mlxsw_sp_vr *vr;
935
936         tb_id = mlxsw_sp_fix_tb_id(tb_id);
937         vr = mlxsw_sp_vr_find(mlxsw_sp, tb_id);
938         if (!vr)
939                 vr = mlxsw_sp_vr_create(mlxsw_sp, tb_id, extack);
940         return vr;
941 }
942
943 static void mlxsw_sp_vr_put(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_vr *vr)
944 {
945         if (!vr->rif_count && list_empty(&vr->fib4->node_list) &&
946             list_empty(&vr->fib6->node_list) &&
947             mlxsw_sp_mr_table_empty(vr->mr_table[MLXSW_SP_L3_PROTO_IPV4]) &&
948             mlxsw_sp_mr_table_empty(vr->mr_table[MLXSW_SP_L3_PROTO_IPV6]))
949                 mlxsw_sp_vr_destroy(mlxsw_sp, vr);
950 }
951
952 static bool
953 mlxsw_sp_vr_lpm_tree_should_replace(struct mlxsw_sp_vr *vr,
954                                     enum mlxsw_sp_l3proto proto, u8 tree_id)
955 {
956         struct mlxsw_sp_fib *fib = mlxsw_sp_vr_fib(vr, proto);
957
958         if (!mlxsw_sp_vr_is_used(vr))
959                 return false;
960         if (fib->lpm_tree->id == tree_id)
961                 return true;
962         return false;
963 }
964
965 static int mlxsw_sp_vr_lpm_tree_replace(struct mlxsw_sp *mlxsw_sp,
966                                         struct mlxsw_sp_fib *fib,
967                                         struct mlxsw_sp_lpm_tree *new_tree)
968 {
969         struct mlxsw_sp_lpm_tree *old_tree = fib->lpm_tree;
970         int err;
971
972         fib->lpm_tree = new_tree;
973         mlxsw_sp_lpm_tree_hold(new_tree);
974         err = mlxsw_sp_vr_lpm_tree_bind(mlxsw_sp, fib, new_tree->id);
975         if (err)
976                 goto err_tree_bind;
977         mlxsw_sp_lpm_tree_put(mlxsw_sp, old_tree);
978         return 0;
979
980 err_tree_bind:
981         mlxsw_sp_lpm_tree_put(mlxsw_sp, new_tree);
982         fib->lpm_tree = old_tree;
983         return err;
984 }
985
986 static int mlxsw_sp_vrs_lpm_tree_replace(struct mlxsw_sp *mlxsw_sp,
987                                          struct mlxsw_sp_fib *fib,
988                                          struct mlxsw_sp_lpm_tree *new_tree)
989 {
990         int max_vrs = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_VRS);
991         enum mlxsw_sp_l3proto proto = fib->proto;
992         struct mlxsw_sp_lpm_tree *old_tree;
993         u8 old_id, new_id = new_tree->id;
994         struct mlxsw_sp_vr *vr;
995         int i, err;
996
997         old_tree = mlxsw_sp->router->lpm.proto_trees[proto];
998         old_id = old_tree->id;
999
1000         for (i = 0; i < max_vrs; i++) {
1001                 vr = &mlxsw_sp->router->vrs[i];
1002                 if (!mlxsw_sp_vr_lpm_tree_should_replace(vr, proto, old_id))
1003                         continue;
1004                 err = mlxsw_sp_vr_lpm_tree_replace(mlxsw_sp,
1005                                                    mlxsw_sp_vr_fib(vr, proto),
1006                                                    new_tree);
1007                 if (err)
1008                         goto err_tree_replace;
1009         }
1010
1011         memcpy(new_tree->prefix_ref_count, old_tree->prefix_ref_count,
1012                sizeof(new_tree->prefix_ref_count));
1013         mlxsw_sp->router->lpm.proto_trees[proto] = new_tree;
1014         mlxsw_sp_lpm_tree_put(mlxsw_sp, old_tree);
1015
1016         return 0;
1017
1018 err_tree_replace:
1019         for (i--; i >= 0; i--) {
1020                 if (!mlxsw_sp_vr_lpm_tree_should_replace(vr, proto, new_id))
1021                         continue;
1022                 mlxsw_sp_vr_lpm_tree_replace(mlxsw_sp,
1023                                              mlxsw_sp_vr_fib(vr, proto),
1024                                              old_tree);
1025         }
1026         return err;
1027 }
1028
1029 static int mlxsw_sp_vrs_init(struct mlxsw_sp *mlxsw_sp)
1030 {
1031         struct mlxsw_sp_vr *vr;
1032         u64 max_vrs;
1033         int i;
1034
1035         if (!MLXSW_CORE_RES_VALID(mlxsw_sp->core, MAX_VRS))
1036                 return -EIO;
1037
1038         max_vrs = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_VRS);
1039         mlxsw_sp->router->vrs = kcalloc(max_vrs, sizeof(struct mlxsw_sp_vr),
1040                                         GFP_KERNEL);
1041         if (!mlxsw_sp->router->vrs)
1042                 return -ENOMEM;
1043
1044         for (i = 0; i < max_vrs; i++) {
1045                 vr = &mlxsw_sp->router->vrs[i];
1046                 vr->id = i;
1047         }
1048
1049         return 0;
1050 }
1051
1052 static void mlxsw_sp_router_fib_flush(struct mlxsw_sp *mlxsw_sp);
1053
1054 static void mlxsw_sp_vrs_fini(struct mlxsw_sp *mlxsw_sp)
1055 {
1056         /* At this stage we're guaranteed not to have new incoming
1057          * FIB notifications and the work queue is free from FIBs
1058          * sitting on top of mlxsw netdevs. However, we can still
1059          * have other FIBs queued. Flush the queue before flushing
1060          * the device's tables. No need for locks, as we're the only
1061          * writer.
1062          */
1063         mlxsw_core_flush_owq();
1064         mlxsw_sp_router_fib_flush(mlxsw_sp);
1065         kfree(mlxsw_sp->router->vrs);
1066 }
1067
1068 u32 mlxsw_sp_ipip_dev_ul_tb_id(const struct net_device *ol_dev)
1069 {
1070         struct net_device *d;
1071         u32 tb_id;
1072
1073         rcu_read_lock();
1074         d = mlxsw_sp_ipip_netdev_ul_dev_get(ol_dev);
1075         if (d)
1076                 tb_id = l3mdev_fib_table(d) ? : RT_TABLE_MAIN;
1077         else
1078                 tb_id = RT_TABLE_MAIN;
1079         rcu_read_unlock();
1080
1081         return tb_id;
1082 }
1083
1084 static void
1085 mlxsw_sp_crif_init(struct mlxsw_sp_crif *crif, struct net_device *dev)
1086 {
1087         crif->key.dev = dev;
1088         INIT_LIST_HEAD(&crif->nexthop_list);
1089 }
1090
1091 static struct mlxsw_sp_crif *
1092 mlxsw_sp_crif_alloc(struct net_device *dev)
1093 {
1094         struct mlxsw_sp_crif *crif;
1095
1096         crif = kzalloc(sizeof(*crif), GFP_KERNEL);
1097         if (!crif)
1098                 return NULL;
1099
1100         mlxsw_sp_crif_init(crif, dev);
1101         return crif;
1102 }
1103
1104 static void mlxsw_sp_crif_free(struct mlxsw_sp_crif *crif)
1105 {
1106         if (WARN_ON(crif->rif))
1107                 return;
1108
1109         WARN_ON(!list_empty(&crif->nexthop_list));
1110         kfree(crif);
1111 }
1112
1113 static int mlxsw_sp_crif_insert(struct mlxsw_sp_router *router,
1114                                 struct mlxsw_sp_crif *crif)
1115 {
1116         return rhashtable_insert_fast(&router->crif_ht, &crif->ht_node,
1117                                       mlxsw_sp_crif_ht_params);
1118 }
1119
1120 static void mlxsw_sp_crif_remove(struct mlxsw_sp_router *router,
1121                                  struct mlxsw_sp_crif *crif)
1122 {
1123         rhashtable_remove_fast(&router->crif_ht, &crif->ht_node,
1124                                mlxsw_sp_crif_ht_params);
1125 }
1126
1127 static struct mlxsw_sp_crif *
1128 mlxsw_sp_crif_lookup(struct mlxsw_sp_router *router,
1129                      const struct net_device *dev)
1130 {
1131         struct mlxsw_sp_crif_key key = {
1132                 .dev = (struct net_device *)dev,
1133         };
1134
1135         return rhashtable_lookup_fast(&router->crif_ht, &key,
1136                                       mlxsw_sp_crif_ht_params);
1137 }
1138
1139 static struct mlxsw_sp_rif *
1140 mlxsw_sp_rif_create(struct mlxsw_sp *mlxsw_sp,
1141                     const struct mlxsw_sp_rif_params *params,
1142                     struct netlink_ext_ack *extack);
1143
1144 static struct mlxsw_sp_rif_ipip_lb *
1145 mlxsw_sp_ipip_ol_ipip_lb_create(struct mlxsw_sp *mlxsw_sp,
1146                                 enum mlxsw_sp_ipip_type ipipt,
1147                                 struct net_device *ol_dev,
1148                                 struct netlink_ext_ack *extack)
1149 {
1150         struct mlxsw_sp_rif_params_ipip_lb lb_params;
1151         const struct mlxsw_sp_ipip_ops *ipip_ops;
1152         struct mlxsw_sp_rif *rif;
1153
1154         ipip_ops = mlxsw_sp->router->ipip_ops_arr[ipipt];
1155         lb_params = (struct mlxsw_sp_rif_params_ipip_lb) {
1156                 .common.dev = ol_dev,
1157                 .common.lag = false,
1158                 .common.double_entry = ipip_ops->double_rif_entry,
1159                 .lb_config = ipip_ops->ol_loopback_config(mlxsw_sp, ol_dev),
1160         };
1161
1162         rif = mlxsw_sp_rif_create(mlxsw_sp, &lb_params.common, extack);
1163         if (IS_ERR(rif))
1164                 return ERR_CAST(rif);
1165         return container_of(rif, struct mlxsw_sp_rif_ipip_lb, common);
1166 }
1167
1168 static struct mlxsw_sp_ipip_entry *
1169 mlxsw_sp_ipip_entry_alloc(struct mlxsw_sp *mlxsw_sp,
1170                           enum mlxsw_sp_ipip_type ipipt,
1171                           struct net_device *ol_dev)
1172 {
1173         const struct mlxsw_sp_ipip_ops *ipip_ops;
1174         struct mlxsw_sp_ipip_entry *ipip_entry;
1175         struct mlxsw_sp_ipip_entry *ret = NULL;
1176         int err;
1177
1178         ipip_ops = mlxsw_sp->router->ipip_ops_arr[ipipt];
1179         ipip_entry = kzalloc(sizeof(*ipip_entry), GFP_KERNEL);
1180         if (!ipip_entry)
1181                 return ERR_PTR(-ENOMEM);
1182
1183         ipip_entry->ol_lb = mlxsw_sp_ipip_ol_ipip_lb_create(mlxsw_sp, ipipt,
1184                                                             ol_dev, NULL);
1185         if (IS_ERR(ipip_entry->ol_lb)) {
1186                 ret = ERR_CAST(ipip_entry->ol_lb);
1187                 goto err_ol_ipip_lb_create;
1188         }
1189
1190         ipip_entry->ipipt = ipipt;
1191         ipip_entry->ol_dev = ol_dev;
1192         ipip_entry->parms = ipip_ops->parms_init(ol_dev);
1193
1194         err = ipip_ops->rem_ip_addr_set(mlxsw_sp, ipip_entry);
1195         if (err) {
1196                 ret = ERR_PTR(err);
1197                 goto err_rem_ip_addr_set;
1198         }
1199
1200         return ipip_entry;
1201
1202 err_rem_ip_addr_set:
1203         mlxsw_sp_rif_destroy(&ipip_entry->ol_lb->common);
1204 err_ol_ipip_lb_create:
1205         kfree(ipip_entry);
1206         return ret;
1207 }
1208
1209 static void mlxsw_sp_ipip_entry_dealloc(struct mlxsw_sp *mlxsw_sp,
1210                                         struct mlxsw_sp_ipip_entry *ipip_entry)
1211 {
1212         const struct mlxsw_sp_ipip_ops *ipip_ops =
1213                 mlxsw_sp->router->ipip_ops_arr[ipip_entry->ipipt];
1214
1215         ipip_ops->rem_ip_addr_unset(mlxsw_sp, ipip_entry);
1216         mlxsw_sp_rif_destroy(&ipip_entry->ol_lb->common);
1217         kfree(ipip_entry);
1218 }
1219
1220 static bool
1221 mlxsw_sp_ipip_entry_saddr_matches(struct mlxsw_sp *mlxsw_sp,
1222                                   const enum mlxsw_sp_l3proto ul_proto,
1223                                   union mlxsw_sp_l3addr saddr,
1224                                   u32 ul_tb_id,
1225                                   struct mlxsw_sp_ipip_entry *ipip_entry)
1226 {
1227         u32 tun_ul_tb_id = mlxsw_sp_ipip_dev_ul_tb_id(ipip_entry->ol_dev);
1228         enum mlxsw_sp_ipip_type ipipt = ipip_entry->ipipt;
1229         union mlxsw_sp_l3addr tun_saddr;
1230
1231         if (mlxsw_sp->router->ipip_ops_arr[ipipt]->ul_proto != ul_proto)
1232                 return false;
1233
1234         tun_saddr = mlxsw_sp_ipip_netdev_saddr(ul_proto, ipip_entry->ol_dev);
1235         return tun_ul_tb_id == ul_tb_id &&
1236                mlxsw_sp_l3addr_eq(&tun_saddr, &saddr);
1237 }
1238
1239 static int mlxsw_sp_ipip_decap_parsing_depth_inc(struct mlxsw_sp *mlxsw_sp,
1240                                                  enum mlxsw_sp_ipip_type ipipt)
1241 {
1242         const struct mlxsw_sp_ipip_ops *ipip_ops;
1243
1244         ipip_ops = mlxsw_sp->router->ipip_ops_arr[ipipt];
1245
1246         /* Not all tunnels require to increase the default pasing depth
1247          * (96 bytes).
1248          */
1249         if (ipip_ops->inc_parsing_depth)
1250                 return mlxsw_sp_parsing_depth_inc(mlxsw_sp);
1251
1252         return 0;
1253 }
1254
1255 static void mlxsw_sp_ipip_decap_parsing_depth_dec(struct mlxsw_sp *mlxsw_sp,
1256                                                   enum mlxsw_sp_ipip_type ipipt)
1257 {
1258         const struct mlxsw_sp_ipip_ops *ipip_ops =
1259                 mlxsw_sp->router->ipip_ops_arr[ipipt];
1260
1261         if (ipip_ops->inc_parsing_depth)
1262                 mlxsw_sp_parsing_depth_dec(mlxsw_sp);
1263 }
1264
1265 static int
1266 mlxsw_sp_fib_entry_decap_init(struct mlxsw_sp *mlxsw_sp,
1267                               struct mlxsw_sp_fib_entry *fib_entry,
1268                               struct mlxsw_sp_ipip_entry *ipip_entry)
1269 {
1270         u32 tunnel_index;
1271         int err;
1272
1273         err = mlxsw_sp_kvdl_alloc(mlxsw_sp, MLXSW_SP_KVDL_ENTRY_TYPE_ADJ,
1274                                   1, &tunnel_index);
1275         if (err)
1276                 return err;
1277
1278         err = mlxsw_sp_ipip_decap_parsing_depth_inc(mlxsw_sp,
1279                                                     ipip_entry->ipipt);
1280         if (err)
1281                 goto err_parsing_depth_inc;
1282
1283         ipip_entry->decap_fib_entry = fib_entry;
1284         fib_entry->decap.ipip_entry = ipip_entry;
1285         fib_entry->decap.tunnel_index = tunnel_index;
1286
1287         return 0;
1288
1289 err_parsing_depth_inc:
1290         mlxsw_sp_kvdl_free(mlxsw_sp, MLXSW_SP_KVDL_ENTRY_TYPE_ADJ, 1,
1291                            fib_entry->decap.tunnel_index);
1292         return err;
1293 }
1294
1295 static void mlxsw_sp_fib_entry_decap_fini(struct mlxsw_sp *mlxsw_sp,
1296                                           struct mlxsw_sp_fib_entry *fib_entry)
1297 {
1298         enum mlxsw_sp_ipip_type ipipt = fib_entry->decap.ipip_entry->ipipt;
1299
1300         /* Unlink this node from the IPIP entry that it's the decap entry of. */
1301         fib_entry->decap.ipip_entry->decap_fib_entry = NULL;
1302         fib_entry->decap.ipip_entry = NULL;
1303         mlxsw_sp_ipip_decap_parsing_depth_dec(mlxsw_sp, ipipt);
1304         mlxsw_sp_kvdl_free(mlxsw_sp, MLXSW_SP_KVDL_ENTRY_TYPE_ADJ,
1305                            1, fib_entry->decap.tunnel_index);
1306 }
1307
1308 static struct mlxsw_sp_fib_node *
1309 mlxsw_sp_fib_node_lookup(struct mlxsw_sp_fib *fib, const void *addr,
1310                          size_t addr_len, unsigned char prefix_len);
1311 static int mlxsw_sp_fib_entry_update(struct mlxsw_sp *mlxsw_sp,
1312                                      struct mlxsw_sp_fib_entry *fib_entry);
1313
1314 static void
1315 mlxsw_sp_ipip_entry_demote_decap(struct mlxsw_sp *mlxsw_sp,
1316                                  struct mlxsw_sp_ipip_entry *ipip_entry)
1317 {
1318         struct mlxsw_sp_fib_entry *fib_entry = ipip_entry->decap_fib_entry;
1319
1320         mlxsw_sp_fib_entry_decap_fini(mlxsw_sp, fib_entry);
1321         fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_TRAP;
1322
1323         mlxsw_sp_fib_entry_update(mlxsw_sp, fib_entry);
1324 }
1325
1326 static void
1327 mlxsw_sp_ipip_entry_promote_decap(struct mlxsw_sp *mlxsw_sp,
1328                                   struct mlxsw_sp_ipip_entry *ipip_entry,
1329                                   struct mlxsw_sp_fib_entry *decap_fib_entry)
1330 {
1331         if (mlxsw_sp_fib_entry_decap_init(mlxsw_sp, decap_fib_entry,
1332                                           ipip_entry))
1333                 return;
1334         decap_fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_IPIP_DECAP;
1335
1336         if (mlxsw_sp_fib_entry_update(mlxsw_sp, decap_fib_entry))
1337                 mlxsw_sp_ipip_entry_demote_decap(mlxsw_sp, ipip_entry);
1338 }
1339
1340 static struct mlxsw_sp_fib_entry *
1341 mlxsw_sp_router_ip2me_fib_entry_find(struct mlxsw_sp *mlxsw_sp, u32 tb_id,
1342                                      enum mlxsw_sp_l3proto proto,
1343                                      const union mlxsw_sp_l3addr *addr,
1344                                      enum mlxsw_sp_fib_entry_type type)
1345 {
1346         struct mlxsw_sp_fib_node *fib_node;
1347         unsigned char addr_prefix_len;
1348         struct mlxsw_sp_fib *fib;
1349         struct mlxsw_sp_vr *vr;
1350         const void *addrp;
1351         size_t addr_len;
1352         u32 addr4;
1353
1354         vr = mlxsw_sp_vr_find(mlxsw_sp, tb_id);
1355         if (!vr)
1356                 return NULL;
1357         fib = mlxsw_sp_vr_fib(vr, proto);
1358
1359         switch (proto) {
1360         case MLXSW_SP_L3_PROTO_IPV4:
1361                 addr4 = be32_to_cpu(addr->addr4);
1362                 addrp = &addr4;
1363                 addr_len = 4;
1364                 addr_prefix_len = 32;
1365                 break;
1366         case MLXSW_SP_L3_PROTO_IPV6:
1367                 addrp = &addr->addr6;
1368                 addr_len = 16;
1369                 addr_prefix_len = 128;
1370                 break;
1371         default:
1372                 WARN_ON(1);
1373                 return NULL;
1374         }
1375
1376         fib_node = mlxsw_sp_fib_node_lookup(fib, addrp, addr_len,
1377                                             addr_prefix_len);
1378         if (!fib_node || fib_node->fib_entry->type != type)
1379                 return NULL;
1380
1381         return fib_node->fib_entry;
1382 }
1383
1384 /* Given an IPIP entry, find the corresponding decap route. */
1385 static struct mlxsw_sp_fib_entry *
1386 mlxsw_sp_ipip_entry_find_decap(struct mlxsw_sp *mlxsw_sp,
1387                                struct mlxsw_sp_ipip_entry *ipip_entry)
1388 {
1389         static struct mlxsw_sp_fib_node *fib_node;
1390         const struct mlxsw_sp_ipip_ops *ipip_ops;
1391         unsigned char saddr_prefix_len;
1392         union mlxsw_sp_l3addr saddr;
1393         struct mlxsw_sp_fib *ul_fib;
1394         struct mlxsw_sp_vr *ul_vr;
1395         const void *saddrp;
1396         size_t saddr_len;
1397         u32 ul_tb_id;
1398         u32 saddr4;
1399
1400         ipip_ops = mlxsw_sp->router->ipip_ops_arr[ipip_entry->ipipt];
1401
1402         ul_tb_id = mlxsw_sp_ipip_dev_ul_tb_id(ipip_entry->ol_dev);
1403         ul_vr = mlxsw_sp_vr_find(mlxsw_sp, ul_tb_id);
1404         if (!ul_vr)
1405                 return NULL;
1406
1407         ul_fib = mlxsw_sp_vr_fib(ul_vr, ipip_ops->ul_proto);
1408         saddr = mlxsw_sp_ipip_netdev_saddr(ipip_ops->ul_proto,
1409                                            ipip_entry->ol_dev);
1410
1411         switch (ipip_ops->ul_proto) {
1412         case MLXSW_SP_L3_PROTO_IPV4:
1413                 saddr4 = be32_to_cpu(saddr.addr4);
1414                 saddrp = &saddr4;
1415                 saddr_len = 4;
1416                 saddr_prefix_len = 32;
1417                 break;
1418         case MLXSW_SP_L3_PROTO_IPV6:
1419                 saddrp = &saddr.addr6;
1420                 saddr_len = 16;
1421                 saddr_prefix_len = 128;
1422                 break;
1423         default:
1424                 WARN_ON(1);
1425                 return NULL;
1426         }
1427
1428         fib_node = mlxsw_sp_fib_node_lookup(ul_fib, saddrp, saddr_len,
1429                                             saddr_prefix_len);
1430         if (!fib_node ||
1431             fib_node->fib_entry->type != MLXSW_SP_FIB_ENTRY_TYPE_TRAP)
1432                 return NULL;
1433
1434         return fib_node->fib_entry;
1435 }
1436
1437 static struct mlxsw_sp_ipip_entry *
1438 mlxsw_sp_ipip_entry_create(struct mlxsw_sp *mlxsw_sp,
1439                            enum mlxsw_sp_ipip_type ipipt,
1440                            struct net_device *ol_dev)
1441 {
1442         struct mlxsw_sp_ipip_entry *ipip_entry;
1443
1444         ipip_entry = mlxsw_sp_ipip_entry_alloc(mlxsw_sp, ipipt, ol_dev);
1445         if (IS_ERR(ipip_entry))
1446                 return ipip_entry;
1447
1448         list_add_tail(&ipip_entry->ipip_list_node,
1449                       &mlxsw_sp->router->ipip_list);
1450
1451         return ipip_entry;
1452 }
1453
1454 static void
1455 mlxsw_sp_ipip_entry_destroy(struct mlxsw_sp *mlxsw_sp,
1456                             struct mlxsw_sp_ipip_entry *ipip_entry)
1457 {
1458         list_del(&ipip_entry->ipip_list_node);
1459         mlxsw_sp_ipip_entry_dealloc(mlxsw_sp, ipip_entry);
1460 }
1461
1462 static bool
1463 mlxsw_sp_ipip_entry_matches_decap(struct mlxsw_sp *mlxsw_sp,
1464                                   const struct net_device *ul_dev,
1465                                   enum mlxsw_sp_l3proto ul_proto,
1466                                   union mlxsw_sp_l3addr ul_dip,
1467                                   struct mlxsw_sp_ipip_entry *ipip_entry)
1468 {
1469         u32 ul_tb_id = l3mdev_fib_table(ul_dev) ? : RT_TABLE_MAIN;
1470         enum mlxsw_sp_ipip_type ipipt = ipip_entry->ipipt;
1471
1472         if (mlxsw_sp->router->ipip_ops_arr[ipipt]->ul_proto != ul_proto)
1473                 return false;
1474
1475         return mlxsw_sp_ipip_entry_saddr_matches(mlxsw_sp, ul_proto, ul_dip,
1476                                                  ul_tb_id, ipip_entry);
1477 }
1478
1479 /* Given decap parameters, find the corresponding IPIP entry. */
1480 static struct mlxsw_sp_ipip_entry *
1481 mlxsw_sp_ipip_entry_find_by_decap(struct mlxsw_sp *mlxsw_sp, int ul_dev_ifindex,
1482                                   enum mlxsw_sp_l3proto ul_proto,
1483                                   union mlxsw_sp_l3addr ul_dip)
1484 {
1485         struct mlxsw_sp_ipip_entry *ipip_entry = NULL;
1486         struct net_device *ul_dev;
1487
1488         rcu_read_lock();
1489
1490         ul_dev = dev_get_by_index_rcu(mlxsw_sp_net(mlxsw_sp), ul_dev_ifindex);
1491         if (!ul_dev)
1492                 goto out_unlock;
1493
1494         list_for_each_entry(ipip_entry, &mlxsw_sp->router->ipip_list,
1495                             ipip_list_node)
1496                 if (mlxsw_sp_ipip_entry_matches_decap(mlxsw_sp, ul_dev,
1497                                                       ul_proto, ul_dip,
1498                                                       ipip_entry))
1499                         goto out_unlock;
1500
1501         rcu_read_unlock();
1502
1503         return NULL;
1504
1505 out_unlock:
1506         rcu_read_unlock();
1507         return ipip_entry;
1508 }
1509
1510 static bool mlxsw_sp_netdev_ipip_type(const struct mlxsw_sp *mlxsw_sp,
1511                                       const struct net_device *dev,
1512                                       enum mlxsw_sp_ipip_type *p_type)
1513 {
1514         struct mlxsw_sp_router *router = mlxsw_sp->router;
1515         const struct mlxsw_sp_ipip_ops *ipip_ops;
1516         enum mlxsw_sp_ipip_type ipipt;
1517
1518         for (ipipt = 0; ipipt < MLXSW_SP_IPIP_TYPE_MAX; ++ipipt) {
1519                 ipip_ops = router->ipip_ops_arr[ipipt];
1520                 if (dev->type == ipip_ops->dev_type) {
1521                         if (p_type)
1522                                 *p_type = ipipt;
1523                         return true;
1524                 }
1525         }
1526         return false;
1527 }
1528
1529 static bool mlxsw_sp_netdev_is_ipip_ol(const struct mlxsw_sp *mlxsw_sp,
1530                                        const struct net_device *dev)
1531 {
1532         return mlxsw_sp_netdev_ipip_type(mlxsw_sp, dev, NULL);
1533 }
1534
1535 static struct mlxsw_sp_ipip_entry *
1536 mlxsw_sp_ipip_entry_find_by_ol_dev(struct mlxsw_sp *mlxsw_sp,
1537                                    const struct net_device *ol_dev)
1538 {
1539         struct mlxsw_sp_ipip_entry *ipip_entry;
1540
1541         list_for_each_entry(ipip_entry, &mlxsw_sp->router->ipip_list,
1542                             ipip_list_node)
1543                 if (ipip_entry->ol_dev == ol_dev)
1544                         return ipip_entry;
1545
1546         return NULL;
1547 }
1548
1549 static struct mlxsw_sp_ipip_entry *
1550 mlxsw_sp_ipip_entry_find_by_ul_dev(const struct mlxsw_sp *mlxsw_sp,
1551                                    const struct net_device *ul_dev,
1552                                    struct mlxsw_sp_ipip_entry *start)
1553 {
1554         struct mlxsw_sp_ipip_entry *ipip_entry;
1555
1556         ipip_entry = list_prepare_entry(start, &mlxsw_sp->router->ipip_list,
1557                                         ipip_list_node);
1558         list_for_each_entry_continue(ipip_entry, &mlxsw_sp->router->ipip_list,
1559                                      ipip_list_node) {
1560                 struct net_device *ol_dev = ipip_entry->ol_dev;
1561                 struct net_device *ipip_ul_dev;
1562
1563                 rcu_read_lock();
1564                 ipip_ul_dev = mlxsw_sp_ipip_netdev_ul_dev_get(ol_dev);
1565                 rcu_read_unlock();
1566
1567                 if (ipip_ul_dev == ul_dev)
1568                         return ipip_entry;
1569         }
1570
1571         return NULL;
1572 }
1573
1574 static bool mlxsw_sp_netdev_is_ipip_ul(struct mlxsw_sp *mlxsw_sp,
1575                                        const struct net_device *dev)
1576 {
1577         return mlxsw_sp_ipip_entry_find_by_ul_dev(mlxsw_sp, dev, NULL);
1578 }
1579
1580 static bool mlxsw_sp_netdevice_ipip_can_offload(struct mlxsw_sp *mlxsw_sp,
1581                                                 const struct net_device *ol_dev,
1582                                                 enum mlxsw_sp_ipip_type ipipt)
1583 {
1584         const struct mlxsw_sp_ipip_ops *ops
1585                 = mlxsw_sp->router->ipip_ops_arr[ipipt];
1586
1587         return ops->can_offload(mlxsw_sp, ol_dev);
1588 }
1589
1590 static int mlxsw_sp_netdevice_ipip_ol_reg_event(struct mlxsw_sp *mlxsw_sp,
1591                                                 struct net_device *ol_dev)
1592 {
1593         enum mlxsw_sp_ipip_type ipipt = MLXSW_SP_IPIP_TYPE_MAX;
1594         struct mlxsw_sp_ipip_entry *ipip_entry;
1595         enum mlxsw_sp_l3proto ul_proto;
1596         union mlxsw_sp_l3addr saddr;
1597         u32 ul_tb_id;
1598
1599         mlxsw_sp_netdev_ipip_type(mlxsw_sp, ol_dev, &ipipt);
1600         if (mlxsw_sp_netdevice_ipip_can_offload(mlxsw_sp, ol_dev, ipipt)) {
1601                 ul_tb_id = mlxsw_sp_ipip_dev_ul_tb_id(ol_dev);
1602                 ul_proto = mlxsw_sp->router->ipip_ops_arr[ipipt]->ul_proto;
1603                 saddr = mlxsw_sp_ipip_netdev_saddr(ul_proto, ol_dev);
1604                 if (!mlxsw_sp_ipip_demote_tunnel_by_saddr(mlxsw_sp, ul_proto,
1605                                                           saddr, ul_tb_id,
1606                                                           NULL)) {
1607                         ipip_entry = mlxsw_sp_ipip_entry_create(mlxsw_sp, ipipt,
1608                                                                 ol_dev);
1609                         if (IS_ERR(ipip_entry))
1610                                 return PTR_ERR(ipip_entry);
1611                 }
1612         }
1613
1614         return 0;
1615 }
1616
1617 static void mlxsw_sp_netdevice_ipip_ol_unreg_event(struct mlxsw_sp *mlxsw_sp,
1618                                                    struct net_device *ol_dev)
1619 {
1620         struct mlxsw_sp_ipip_entry *ipip_entry;
1621
1622         ipip_entry = mlxsw_sp_ipip_entry_find_by_ol_dev(mlxsw_sp, ol_dev);
1623         if (ipip_entry)
1624                 mlxsw_sp_ipip_entry_destroy(mlxsw_sp, ipip_entry);
1625 }
1626
1627 static void
1628 mlxsw_sp_ipip_entry_ol_up_event(struct mlxsw_sp *mlxsw_sp,
1629                                 struct mlxsw_sp_ipip_entry *ipip_entry)
1630 {
1631         struct mlxsw_sp_fib_entry *decap_fib_entry;
1632
1633         decap_fib_entry = mlxsw_sp_ipip_entry_find_decap(mlxsw_sp, ipip_entry);
1634         if (decap_fib_entry)
1635                 mlxsw_sp_ipip_entry_promote_decap(mlxsw_sp, ipip_entry,
1636                                                   decap_fib_entry);
1637 }
1638
1639 static int
1640 mlxsw_sp_rif_ipip_lb_op(struct mlxsw_sp_rif_ipip_lb *lb_rif, u16 ul_vr_id,
1641                         u16 ul_rif_id, bool enable)
1642 {
1643         struct mlxsw_sp_rif_ipip_lb_config lb_cf = lb_rif->lb_config;
1644         struct net_device *dev = mlxsw_sp_rif_dev(&lb_rif->common);
1645         enum mlxsw_reg_ritr_loopback_ipip_options ipip_options;
1646         struct mlxsw_sp_rif *rif = &lb_rif->common;
1647         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
1648         char ritr_pl[MLXSW_REG_RITR_LEN];
1649         struct in6_addr *saddr6;
1650         u32 saddr4;
1651
1652         ipip_options = MLXSW_REG_RITR_LOOPBACK_IPIP_OPTIONS_GRE_KEY_PRESET;
1653         switch (lb_cf.ul_protocol) {
1654         case MLXSW_SP_L3_PROTO_IPV4:
1655                 saddr4 = be32_to_cpu(lb_cf.saddr.addr4);
1656                 mlxsw_reg_ritr_pack(ritr_pl, enable, MLXSW_REG_RITR_LOOPBACK_IF,
1657                                     rif->rif_index, rif->vr_id, dev->mtu);
1658                 mlxsw_reg_ritr_loopback_ipip4_pack(ritr_pl, lb_cf.lb_ipipt,
1659                                                    ipip_options, ul_vr_id,
1660                                                    ul_rif_id, saddr4,
1661                                                    lb_cf.okey);
1662                 break;
1663
1664         case MLXSW_SP_L3_PROTO_IPV6:
1665                 saddr6 = &lb_cf.saddr.addr6;
1666                 mlxsw_reg_ritr_pack(ritr_pl, enable, MLXSW_REG_RITR_LOOPBACK_IF,
1667                                     rif->rif_index, rif->vr_id, dev->mtu);
1668                 mlxsw_reg_ritr_loopback_ipip6_pack(ritr_pl, lb_cf.lb_ipipt,
1669                                                    ipip_options, ul_vr_id,
1670                                                    ul_rif_id, saddr6,
1671                                                    lb_cf.okey);
1672                 break;
1673         }
1674
1675         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
1676 }
1677
1678 static int mlxsw_sp_netdevice_ipip_ol_update_mtu(struct mlxsw_sp *mlxsw_sp,
1679                                                  struct net_device *ol_dev)
1680 {
1681         struct mlxsw_sp_ipip_entry *ipip_entry;
1682         struct mlxsw_sp_rif_ipip_lb *lb_rif;
1683         int err = 0;
1684
1685         ipip_entry = mlxsw_sp_ipip_entry_find_by_ol_dev(mlxsw_sp, ol_dev);
1686         if (ipip_entry) {
1687                 lb_rif = ipip_entry->ol_lb;
1688                 err = mlxsw_sp_rif_ipip_lb_op(lb_rif, lb_rif->ul_vr_id,
1689                                               lb_rif->ul_rif_id, true);
1690                 if (err)
1691                         goto out;
1692                 lb_rif->common.mtu = ol_dev->mtu;
1693         }
1694
1695 out:
1696         return err;
1697 }
1698
1699 static void mlxsw_sp_netdevice_ipip_ol_up_event(struct mlxsw_sp *mlxsw_sp,
1700                                                 struct net_device *ol_dev)
1701 {
1702         struct mlxsw_sp_ipip_entry *ipip_entry;
1703
1704         ipip_entry = mlxsw_sp_ipip_entry_find_by_ol_dev(mlxsw_sp, ol_dev);
1705         if (ipip_entry)
1706                 mlxsw_sp_ipip_entry_ol_up_event(mlxsw_sp, ipip_entry);
1707 }
1708
1709 static void
1710 mlxsw_sp_ipip_entry_ol_down_event(struct mlxsw_sp *mlxsw_sp,
1711                                   struct mlxsw_sp_ipip_entry *ipip_entry)
1712 {
1713         if (ipip_entry->decap_fib_entry)
1714                 mlxsw_sp_ipip_entry_demote_decap(mlxsw_sp, ipip_entry);
1715 }
1716
1717 static void mlxsw_sp_netdevice_ipip_ol_down_event(struct mlxsw_sp *mlxsw_sp,
1718                                                   struct net_device *ol_dev)
1719 {
1720         struct mlxsw_sp_ipip_entry *ipip_entry;
1721
1722         ipip_entry = mlxsw_sp_ipip_entry_find_by_ol_dev(mlxsw_sp, ol_dev);
1723         if (ipip_entry)
1724                 mlxsw_sp_ipip_entry_ol_down_event(mlxsw_sp, ipip_entry);
1725 }
1726
1727 static void mlxsw_sp_nexthop_rif_update(struct mlxsw_sp *mlxsw_sp,
1728                                         struct mlxsw_sp_rif *rif);
1729
1730 static void mlxsw_sp_rif_migrate_destroy(struct mlxsw_sp *mlxsw_sp,
1731                                          struct mlxsw_sp_rif *old_rif,
1732                                          struct mlxsw_sp_rif *new_rif,
1733                                          bool migrate_nhs)
1734 {
1735         struct mlxsw_sp_crif *crif = old_rif->crif;
1736         struct mlxsw_sp_crif mock_crif = {};
1737
1738         if (migrate_nhs)
1739                 mlxsw_sp_nexthop_rif_update(mlxsw_sp, new_rif);
1740
1741         /* Plant a mock CRIF so that destroying the old RIF doesn't unoffload
1742          * our nexthops and IPIP tunnels, and doesn't sever the crif->rif link.
1743          */
1744         mlxsw_sp_crif_init(&mock_crif, crif->key.dev);
1745         old_rif->crif = &mock_crif;
1746         mock_crif.rif = old_rif;
1747         mlxsw_sp_rif_destroy(old_rif);
1748 }
1749
1750 static int
1751 mlxsw_sp_ipip_entry_ol_lb_update(struct mlxsw_sp *mlxsw_sp,
1752                                  struct mlxsw_sp_ipip_entry *ipip_entry,
1753                                  bool keep_encap,
1754                                  struct netlink_ext_ack *extack)
1755 {
1756         struct mlxsw_sp_rif_ipip_lb *old_lb_rif = ipip_entry->ol_lb;
1757         struct mlxsw_sp_rif_ipip_lb *new_lb_rif;
1758
1759         new_lb_rif = mlxsw_sp_ipip_ol_ipip_lb_create(mlxsw_sp,
1760                                                      ipip_entry->ipipt,
1761                                                      ipip_entry->ol_dev,
1762                                                      extack);
1763         if (IS_ERR(new_lb_rif))
1764                 return PTR_ERR(new_lb_rif);
1765         ipip_entry->ol_lb = new_lb_rif;
1766
1767         mlxsw_sp_rif_migrate_destroy(mlxsw_sp, &old_lb_rif->common,
1768                                      &new_lb_rif->common, keep_encap);
1769         return 0;
1770 }
1771
1772 /**
1773  * __mlxsw_sp_ipip_entry_update_tunnel - Update offload related to IPIP entry.
1774  * @mlxsw_sp: mlxsw_sp.
1775  * @ipip_entry: IPIP entry.
1776  * @recreate_loopback: Recreates the associated loopback RIF.
1777  * @keep_encap: Updates next hops that use the tunnel netdevice. This is only
1778  *              relevant when recreate_loopback is true.
1779  * @update_nexthops: Updates next hops, keeping the current loopback RIF. This
1780  *                   is only relevant when recreate_loopback is false.
1781  * @extack: extack.
1782  *
1783  * Return: Non-zero value on failure.
1784  */
1785 int __mlxsw_sp_ipip_entry_update_tunnel(struct mlxsw_sp *mlxsw_sp,
1786                                         struct mlxsw_sp_ipip_entry *ipip_entry,
1787                                         bool recreate_loopback,
1788                                         bool keep_encap,
1789                                         bool update_nexthops,
1790                                         struct netlink_ext_ack *extack)
1791 {
1792         int err;
1793
1794         /* RIFs can't be edited, so to update loopback, we need to destroy and
1795          * recreate it. That creates a window of opportunity where RALUE and
1796          * RATR registers end up referencing a RIF that's already gone. RATRs
1797          * are handled in mlxsw_sp_ipip_entry_ol_lb_update(), and to take care
1798          * of RALUE, demote the decap route back.
1799          */
1800         if (ipip_entry->decap_fib_entry)
1801                 mlxsw_sp_ipip_entry_demote_decap(mlxsw_sp, ipip_entry);
1802
1803         if (recreate_loopback) {
1804                 err = mlxsw_sp_ipip_entry_ol_lb_update(mlxsw_sp, ipip_entry,
1805                                                        keep_encap, extack);
1806                 if (err)
1807                         return err;
1808         } else if (update_nexthops) {
1809                 mlxsw_sp_nexthop_rif_update(mlxsw_sp,
1810                                             &ipip_entry->ol_lb->common);
1811         }
1812
1813         if (ipip_entry->ol_dev->flags & IFF_UP)
1814                 mlxsw_sp_ipip_entry_ol_up_event(mlxsw_sp, ipip_entry);
1815
1816         return 0;
1817 }
1818
1819 static int mlxsw_sp_netdevice_ipip_ol_vrf_event(struct mlxsw_sp *mlxsw_sp,
1820                                                 struct net_device *ol_dev,
1821                                                 struct netlink_ext_ack *extack)
1822 {
1823         struct mlxsw_sp_ipip_entry *ipip_entry =
1824                 mlxsw_sp_ipip_entry_find_by_ol_dev(mlxsw_sp, ol_dev);
1825
1826         if (!ipip_entry)
1827                 return 0;
1828
1829         return __mlxsw_sp_ipip_entry_update_tunnel(mlxsw_sp, ipip_entry,
1830                                                    true, false, false, extack);
1831 }
1832
1833 static int
1834 mlxsw_sp_netdevice_ipip_ul_vrf_event(struct mlxsw_sp *mlxsw_sp,
1835                                      struct mlxsw_sp_ipip_entry *ipip_entry,
1836                                      struct net_device *ul_dev,
1837                                      bool *demote_this,
1838                                      struct netlink_ext_ack *extack)
1839 {
1840         u32 ul_tb_id = l3mdev_fib_table(ul_dev) ? : RT_TABLE_MAIN;
1841         enum mlxsw_sp_l3proto ul_proto;
1842         union mlxsw_sp_l3addr saddr;
1843
1844         /* Moving underlay to a different VRF might cause local address
1845          * conflict, and the conflicting tunnels need to be demoted.
1846          */
1847         ul_proto = mlxsw_sp->router->ipip_ops_arr[ipip_entry->ipipt]->ul_proto;
1848         saddr = mlxsw_sp_ipip_netdev_saddr(ul_proto, ipip_entry->ol_dev);
1849         if (mlxsw_sp_ipip_demote_tunnel_by_saddr(mlxsw_sp, ul_proto,
1850                                                  saddr, ul_tb_id,
1851                                                  ipip_entry)) {
1852                 *demote_this = true;
1853                 return 0;
1854         }
1855
1856         return __mlxsw_sp_ipip_entry_update_tunnel(mlxsw_sp, ipip_entry,
1857                                                    true, true, false, extack);
1858 }
1859
1860 static int
1861 mlxsw_sp_netdevice_ipip_ul_up_event(struct mlxsw_sp *mlxsw_sp,
1862                                     struct mlxsw_sp_ipip_entry *ipip_entry,
1863                                     struct net_device *ul_dev)
1864 {
1865         return __mlxsw_sp_ipip_entry_update_tunnel(mlxsw_sp, ipip_entry,
1866                                                    false, false, true, NULL);
1867 }
1868
1869 static int
1870 mlxsw_sp_netdevice_ipip_ul_down_event(struct mlxsw_sp *mlxsw_sp,
1871                                       struct mlxsw_sp_ipip_entry *ipip_entry,
1872                                       struct net_device *ul_dev)
1873 {
1874         /* A down underlay device causes encapsulated packets to not be
1875          * forwarded, but decap still works. So refresh next hops without
1876          * touching anything else.
1877          */
1878         return __mlxsw_sp_ipip_entry_update_tunnel(mlxsw_sp, ipip_entry,
1879                                                    false, false, true, NULL);
1880 }
1881
1882 static int
1883 mlxsw_sp_netdevice_ipip_ol_change_event(struct mlxsw_sp *mlxsw_sp,
1884                                         struct net_device *ol_dev,
1885                                         struct netlink_ext_ack *extack)
1886 {
1887         const struct mlxsw_sp_ipip_ops *ipip_ops;
1888         struct mlxsw_sp_ipip_entry *ipip_entry;
1889         int err;
1890
1891         ipip_entry = mlxsw_sp_ipip_entry_find_by_ol_dev(mlxsw_sp, ol_dev);
1892         if (!ipip_entry)
1893                 /* A change might make a tunnel eligible for offloading, but
1894                  * that is currently not implemented. What falls to slow path
1895                  * stays there.
1896                  */
1897                 return 0;
1898
1899         /* A change might make a tunnel not eligible for offloading. */
1900         if (!mlxsw_sp_netdevice_ipip_can_offload(mlxsw_sp, ol_dev,
1901                                                  ipip_entry->ipipt)) {
1902                 mlxsw_sp_ipip_entry_demote_tunnel(mlxsw_sp, ipip_entry);
1903                 return 0;
1904         }
1905
1906         ipip_ops = mlxsw_sp->router->ipip_ops_arr[ipip_entry->ipipt];
1907         err = ipip_ops->ol_netdev_change(mlxsw_sp, ipip_entry, extack);
1908         return err;
1909 }
1910
1911 void mlxsw_sp_ipip_entry_demote_tunnel(struct mlxsw_sp *mlxsw_sp,
1912                                        struct mlxsw_sp_ipip_entry *ipip_entry)
1913 {
1914         struct net_device *ol_dev = ipip_entry->ol_dev;
1915
1916         if (ol_dev->flags & IFF_UP)
1917                 mlxsw_sp_ipip_entry_ol_down_event(mlxsw_sp, ipip_entry);
1918         mlxsw_sp_ipip_entry_destroy(mlxsw_sp, ipip_entry);
1919 }
1920
1921 /* The configuration where several tunnels have the same local address in the
1922  * same underlay table needs special treatment in the HW. That is currently not
1923  * implemented in the driver. This function finds and demotes the first tunnel
1924  * with a given source address, except the one passed in the argument
1925  * `except'.
1926  */
1927 bool
1928 mlxsw_sp_ipip_demote_tunnel_by_saddr(struct mlxsw_sp *mlxsw_sp,
1929                                      enum mlxsw_sp_l3proto ul_proto,
1930                                      union mlxsw_sp_l3addr saddr,
1931                                      u32 ul_tb_id,
1932                                      const struct mlxsw_sp_ipip_entry *except)
1933 {
1934         struct mlxsw_sp_ipip_entry *ipip_entry, *tmp;
1935
1936         list_for_each_entry_safe(ipip_entry, tmp, &mlxsw_sp->router->ipip_list,
1937                                  ipip_list_node) {
1938                 if (ipip_entry != except &&
1939                     mlxsw_sp_ipip_entry_saddr_matches(mlxsw_sp, ul_proto, saddr,
1940                                                       ul_tb_id, ipip_entry)) {
1941                         mlxsw_sp_ipip_entry_demote_tunnel(mlxsw_sp, ipip_entry);
1942                         return true;
1943                 }
1944         }
1945
1946         return false;
1947 }
1948
1949 static void mlxsw_sp_ipip_demote_tunnel_by_ul_netdev(struct mlxsw_sp *mlxsw_sp,
1950                                                      struct net_device *ul_dev)
1951 {
1952         struct mlxsw_sp_ipip_entry *ipip_entry, *tmp;
1953
1954         list_for_each_entry_safe(ipip_entry, tmp, &mlxsw_sp->router->ipip_list,
1955                                  ipip_list_node) {
1956                 struct net_device *ol_dev = ipip_entry->ol_dev;
1957                 struct net_device *ipip_ul_dev;
1958
1959                 rcu_read_lock();
1960                 ipip_ul_dev = mlxsw_sp_ipip_netdev_ul_dev_get(ol_dev);
1961                 rcu_read_unlock();
1962                 if (ipip_ul_dev == ul_dev)
1963                         mlxsw_sp_ipip_entry_demote_tunnel(mlxsw_sp, ipip_entry);
1964         }
1965 }
1966
1967 static int mlxsw_sp_netdevice_ipip_ol_event(struct mlxsw_sp *mlxsw_sp,
1968                                             struct net_device *ol_dev,
1969                                             unsigned long event,
1970                                             struct netdev_notifier_info *info)
1971 {
1972         struct netdev_notifier_changeupper_info *chup;
1973         struct netlink_ext_ack *extack;
1974         int err = 0;
1975
1976         switch (event) {
1977         case NETDEV_REGISTER:
1978                 err = mlxsw_sp_netdevice_ipip_ol_reg_event(mlxsw_sp, ol_dev);
1979                 break;
1980         case NETDEV_UNREGISTER:
1981                 mlxsw_sp_netdevice_ipip_ol_unreg_event(mlxsw_sp, ol_dev);
1982                 break;
1983         case NETDEV_UP:
1984                 mlxsw_sp_netdevice_ipip_ol_up_event(mlxsw_sp, ol_dev);
1985                 break;
1986         case NETDEV_DOWN:
1987                 mlxsw_sp_netdevice_ipip_ol_down_event(mlxsw_sp, ol_dev);
1988                 break;
1989         case NETDEV_CHANGEUPPER:
1990                 chup = container_of(info, typeof(*chup), info);
1991                 extack = info->extack;
1992                 if (netif_is_l3_master(chup->upper_dev))
1993                         err = mlxsw_sp_netdevice_ipip_ol_vrf_event(mlxsw_sp,
1994                                                                    ol_dev,
1995                                                                    extack);
1996                 break;
1997         case NETDEV_CHANGE:
1998                 extack = info->extack;
1999                 err = mlxsw_sp_netdevice_ipip_ol_change_event(mlxsw_sp,
2000                                                               ol_dev, extack);
2001                 break;
2002         case NETDEV_CHANGEMTU:
2003                 err = mlxsw_sp_netdevice_ipip_ol_update_mtu(mlxsw_sp, ol_dev);
2004                 break;
2005         }
2006         return err;
2007 }
2008
2009 static int
2010 __mlxsw_sp_netdevice_ipip_ul_event(struct mlxsw_sp *mlxsw_sp,
2011                                    struct mlxsw_sp_ipip_entry *ipip_entry,
2012                                    struct net_device *ul_dev,
2013                                    bool *demote_this,
2014                                    unsigned long event,
2015                                    struct netdev_notifier_info *info)
2016 {
2017         struct netdev_notifier_changeupper_info *chup;
2018         struct netlink_ext_ack *extack;
2019
2020         switch (event) {
2021         case NETDEV_CHANGEUPPER:
2022                 chup = container_of(info, typeof(*chup), info);
2023                 extack = info->extack;
2024                 if (netif_is_l3_master(chup->upper_dev))
2025                         return mlxsw_sp_netdevice_ipip_ul_vrf_event(mlxsw_sp,
2026                                                                     ipip_entry,
2027                                                                     ul_dev,
2028                                                                     demote_this,
2029                                                                     extack);
2030                 break;
2031
2032         case NETDEV_UP:
2033                 return mlxsw_sp_netdevice_ipip_ul_up_event(mlxsw_sp, ipip_entry,
2034                                                            ul_dev);
2035         case NETDEV_DOWN:
2036                 return mlxsw_sp_netdevice_ipip_ul_down_event(mlxsw_sp,
2037                                                              ipip_entry,
2038                                                              ul_dev);
2039         }
2040         return 0;
2041 }
2042
2043 static int
2044 mlxsw_sp_netdevice_ipip_ul_event(struct mlxsw_sp *mlxsw_sp,
2045                                  struct net_device *ul_dev,
2046                                  unsigned long event,
2047                                  struct netdev_notifier_info *info)
2048 {
2049         struct mlxsw_sp_ipip_entry *ipip_entry = NULL;
2050         int err;
2051
2052         while ((ipip_entry = mlxsw_sp_ipip_entry_find_by_ul_dev(mlxsw_sp,
2053                                                                 ul_dev,
2054                                                                 ipip_entry))) {
2055                 struct mlxsw_sp_ipip_entry *prev;
2056                 bool demote_this = false;
2057
2058                 err = __mlxsw_sp_netdevice_ipip_ul_event(mlxsw_sp, ipip_entry,
2059                                                          ul_dev, &demote_this,
2060                                                          event, info);
2061                 if (err) {
2062                         mlxsw_sp_ipip_demote_tunnel_by_ul_netdev(mlxsw_sp,
2063                                                                  ul_dev);
2064                         return err;
2065                 }
2066
2067                 if (demote_this) {
2068                         if (list_is_first(&ipip_entry->ipip_list_node,
2069                                           &mlxsw_sp->router->ipip_list))
2070                                 prev = NULL;
2071                         else
2072                                 /* This can't be cached from previous iteration,
2073                                  * because that entry could be gone now.
2074                                  */
2075                                 prev = list_prev_entry(ipip_entry,
2076                                                        ipip_list_node);
2077                         mlxsw_sp_ipip_entry_demote_tunnel(mlxsw_sp, ipip_entry);
2078                         ipip_entry = prev;
2079                 }
2080         }
2081
2082         return 0;
2083 }
2084
2085 int mlxsw_sp_router_nve_promote_decap(struct mlxsw_sp *mlxsw_sp, u32 ul_tb_id,
2086                                       enum mlxsw_sp_l3proto ul_proto,
2087                                       const union mlxsw_sp_l3addr *ul_sip,
2088                                       u32 tunnel_index)
2089 {
2090         enum mlxsw_sp_fib_entry_type type = MLXSW_SP_FIB_ENTRY_TYPE_TRAP;
2091         struct mlxsw_sp_router *router = mlxsw_sp->router;
2092         struct mlxsw_sp_fib_entry *fib_entry;
2093         int err = 0;
2094
2095         mutex_lock(&mlxsw_sp->router->lock);
2096
2097         if (WARN_ON_ONCE(router->nve_decap_config.valid)) {
2098                 err = -EINVAL;
2099                 goto out;
2100         }
2101
2102         router->nve_decap_config.ul_tb_id = ul_tb_id;
2103         router->nve_decap_config.tunnel_index = tunnel_index;
2104         router->nve_decap_config.ul_proto = ul_proto;
2105         router->nve_decap_config.ul_sip = *ul_sip;
2106         router->nve_decap_config.valid = true;
2107
2108         /* It is valid to create a tunnel with a local IP and only later
2109          * assign this IP address to a local interface
2110          */
2111         fib_entry = mlxsw_sp_router_ip2me_fib_entry_find(mlxsw_sp, ul_tb_id,
2112                                                          ul_proto, ul_sip,
2113                                                          type);
2114         if (!fib_entry)
2115                 goto out;
2116
2117         fib_entry->decap.tunnel_index = tunnel_index;
2118         fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_NVE_DECAP;
2119
2120         err = mlxsw_sp_fib_entry_update(mlxsw_sp, fib_entry);
2121         if (err)
2122                 goto err_fib_entry_update;
2123
2124         goto out;
2125
2126 err_fib_entry_update:
2127         fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_TRAP;
2128         mlxsw_sp_fib_entry_update(mlxsw_sp, fib_entry);
2129 out:
2130         mutex_unlock(&mlxsw_sp->router->lock);
2131         return err;
2132 }
2133
2134 void mlxsw_sp_router_nve_demote_decap(struct mlxsw_sp *mlxsw_sp, u32 ul_tb_id,
2135                                       enum mlxsw_sp_l3proto ul_proto,
2136                                       const union mlxsw_sp_l3addr *ul_sip)
2137 {
2138         enum mlxsw_sp_fib_entry_type type = MLXSW_SP_FIB_ENTRY_TYPE_NVE_DECAP;
2139         struct mlxsw_sp_router *router = mlxsw_sp->router;
2140         struct mlxsw_sp_fib_entry *fib_entry;
2141
2142         mutex_lock(&mlxsw_sp->router->lock);
2143
2144         if (WARN_ON_ONCE(!router->nve_decap_config.valid))
2145                 goto out;
2146
2147         router->nve_decap_config.valid = false;
2148
2149         fib_entry = mlxsw_sp_router_ip2me_fib_entry_find(mlxsw_sp, ul_tb_id,
2150                                                          ul_proto, ul_sip,
2151                                                          type);
2152         if (!fib_entry)
2153                 goto out;
2154
2155         fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_TRAP;
2156         mlxsw_sp_fib_entry_update(mlxsw_sp, fib_entry);
2157 out:
2158         mutex_unlock(&mlxsw_sp->router->lock);
2159 }
2160
2161 static bool mlxsw_sp_router_nve_is_decap(struct mlxsw_sp *mlxsw_sp,
2162                                          u32 ul_tb_id,
2163                                          enum mlxsw_sp_l3proto ul_proto,
2164                                          const union mlxsw_sp_l3addr *ul_sip)
2165 {
2166         struct mlxsw_sp_router *router = mlxsw_sp->router;
2167
2168         return router->nve_decap_config.valid &&
2169                router->nve_decap_config.ul_tb_id == ul_tb_id &&
2170                router->nve_decap_config.ul_proto == ul_proto &&
2171                !memcmp(&router->nve_decap_config.ul_sip, ul_sip,
2172                        sizeof(*ul_sip));
2173 }
2174
2175 struct mlxsw_sp_neigh_key {
2176         struct neighbour *n;
2177 };
2178
2179 struct mlxsw_sp_neigh_entry {
2180         struct list_head rif_list_node;
2181         struct rhash_head ht_node;
2182         struct mlxsw_sp_neigh_key key;
2183         u16 rif;
2184         bool connected;
2185         unsigned char ha[ETH_ALEN];
2186         struct list_head nexthop_list; /* list of nexthops using
2187                                         * this neigh entry
2188                                         */
2189         struct list_head nexthop_neighs_list_node;
2190         unsigned int counter_index;
2191         bool counter_valid;
2192 };
2193
2194 static const struct rhashtable_params mlxsw_sp_neigh_ht_params = {
2195         .key_offset = offsetof(struct mlxsw_sp_neigh_entry, key),
2196         .head_offset = offsetof(struct mlxsw_sp_neigh_entry, ht_node),
2197         .key_len = sizeof(struct mlxsw_sp_neigh_key),
2198 };
2199
2200 struct mlxsw_sp_neigh_entry *
2201 mlxsw_sp_rif_neigh_next(struct mlxsw_sp_rif *rif,
2202                         struct mlxsw_sp_neigh_entry *neigh_entry)
2203 {
2204         if (!neigh_entry) {
2205                 if (list_empty(&rif->neigh_list))
2206                         return NULL;
2207                 else
2208                         return list_first_entry(&rif->neigh_list,
2209                                                 typeof(*neigh_entry),
2210                                                 rif_list_node);
2211         }
2212         if (list_is_last(&neigh_entry->rif_list_node, &rif->neigh_list))
2213                 return NULL;
2214         return list_next_entry(neigh_entry, rif_list_node);
2215 }
2216
2217 int mlxsw_sp_neigh_entry_type(struct mlxsw_sp_neigh_entry *neigh_entry)
2218 {
2219         return neigh_entry->key.n->tbl->family;
2220 }
2221
2222 unsigned char *
2223 mlxsw_sp_neigh_entry_ha(struct mlxsw_sp_neigh_entry *neigh_entry)
2224 {
2225         return neigh_entry->ha;
2226 }
2227
2228 u32 mlxsw_sp_neigh4_entry_dip(struct mlxsw_sp_neigh_entry *neigh_entry)
2229 {
2230         struct neighbour *n;
2231
2232         n = neigh_entry->key.n;
2233         return ntohl(*((__be32 *) n->primary_key));
2234 }
2235
2236 struct in6_addr *
2237 mlxsw_sp_neigh6_entry_dip(struct mlxsw_sp_neigh_entry *neigh_entry)
2238 {
2239         struct neighbour *n;
2240
2241         n = neigh_entry->key.n;
2242         return (struct in6_addr *) &n->primary_key;
2243 }
2244
2245 int mlxsw_sp_neigh_counter_get(struct mlxsw_sp *mlxsw_sp,
2246                                struct mlxsw_sp_neigh_entry *neigh_entry,
2247                                u64 *p_counter)
2248 {
2249         if (!neigh_entry->counter_valid)
2250                 return -EINVAL;
2251
2252         return mlxsw_sp_flow_counter_get(mlxsw_sp, neigh_entry->counter_index,
2253                                          p_counter, NULL);
2254 }
2255
2256 static struct mlxsw_sp_neigh_entry *
2257 mlxsw_sp_neigh_entry_alloc(struct mlxsw_sp *mlxsw_sp, struct neighbour *n,
2258                            u16 rif)
2259 {
2260         struct mlxsw_sp_neigh_entry *neigh_entry;
2261
2262         neigh_entry = kzalloc(sizeof(*neigh_entry), GFP_KERNEL);
2263         if (!neigh_entry)
2264                 return NULL;
2265
2266         neigh_entry->key.n = n;
2267         neigh_entry->rif = rif;
2268         INIT_LIST_HEAD(&neigh_entry->nexthop_list);
2269
2270         return neigh_entry;
2271 }
2272
2273 static void mlxsw_sp_neigh_entry_free(struct mlxsw_sp_neigh_entry *neigh_entry)
2274 {
2275         kfree(neigh_entry);
2276 }
2277
2278 static int
2279 mlxsw_sp_neigh_entry_insert(struct mlxsw_sp *mlxsw_sp,
2280                             struct mlxsw_sp_neigh_entry *neigh_entry)
2281 {
2282         return rhashtable_insert_fast(&mlxsw_sp->router->neigh_ht,
2283                                       &neigh_entry->ht_node,
2284                                       mlxsw_sp_neigh_ht_params);
2285 }
2286
2287 static void
2288 mlxsw_sp_neigh_entry_remove(struct mlxsw_sp *mlxsw_sp,
2289                             struct mlxsw_sp_neigh_entry *neigh_entry)
2290 {
2291         rhashtable_remove_fast(&mlxsw_sp->router->neigh_ht,
2292                                &neigh_entry->ht_node,
2293                                mlxsw_sp_neigh_ht_params);
2294 }
2295
2296 static bool
2297 mlxsw_sp_neigh_counter_should_alloc(struct mlxsw_sp *mlxsw_sp,
2298                                     struct mlxsw_sp_neigh_entry *neigh_entry)
2299 {
2300         struct devlink *devlink;
2301         const char *table_name;
2302
2303         switch (mlxsw_sp_neigh_entry_type(neigh_entry)) {
2304         case AF_INET:
2305                 table_name = MLXSW_SP_DPIPE_TABLE_NAME_HOST4;
2306                 break;
2307         case AF_INET6:
2308                 table_name = MLXSW_SP_DPIPE_TABLE_NAME_HOST6;
2309                 break;
2310         default:
2311                 WARN_ON(1);
2312                 return false;
2313         }
2314
2315         devlink = priv_to_devlink(mlxsw_sp->core);
2316         return devlink_dpipe_table_counter_enabled(devlink, table_name);
2317 }
2318
2319 static void
2320 mlxsw_sp_neigh_counter_alloc(struct mlxsw_sp *mlxsw_sp,
2321                              struct mlxsw_sp_neigh_entry *neigh_entry)
2322 {
2323         if (!mlxsw_sp_neigh_counter_should_alloc(mlxsw_sp, neigh_entry))
2324                 return;
2325
2326         if (mlxsw_sp_flow_counter_alloc(mlxsw_sp, &neigh_entry->counter_index))
2327                 return;
2328
2329         neigh_entry->counter_valid = true;
2330 }
2331
2332 static void
2333 mlxsw_sp_neigh_counter_free(struct mlxsw_sp *mlxsw_sp,
2334                             struct mlxsw_sp_neigh_entry *neigh_entry)
2335 {
2336         if (!neigh_entry->counter_valid)
2337                 return;
2338         mlxsw_sp_flow_counter_free(mlxsw_sp,
2339                                    neigh_entry->counter_index);
2340         neigh_entry->counter_valid = false;
2341 }
2342
2343 static struct mlxsw_sp_neigh_entry *
2344 mlxsw_sp_neigh_entry_create(struct mlxsw_sp *mlxsw_sp, struct neighbour *n)
2345 {
2346         struct mlxsw_sp_neigh_entry *neigh_entry;
2347         struct mlxsw_sp_rif *rif;
2348         int err;
2349
2350         rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, n->dev);
2351         if (!rif)
2352                 return ERR_PTR(-EINVAL);
2353
2354         neigh_entry = mlxsw_sp_neigh_entry_alloc(mlxsw_sp, n, rif->rif_index);
2355         if (!neigh_entry)
2356                 return ERR_PTR(-ENOMEM);
2357
2358         err = mlxsw_sp_neigh_entry_insert(mlxsw_sp, neigh_entry);
2359         if (err)
2360                 goto err_neigh_entry_insert;
2361
2362         mlxsw_sp_neigh_counter_alloc(mlxsw_sp, neigh_entry);
2363         atomic_inc(&mlxsw_sp->router->neighs_update.neigh_count);
2364         list_add(&neigh_entry->rif_list_node, &rif->neigh_list);
2365
2366         return neigh_entry;
2367
2368 err_neigh_entry_insert:
2369         mlxsw_sp_neigh_entry_free(neigh_entry);
2370         return ERR_PTR(err);
2371 }
2372
2373 static void
2374 mlxsw_sp_neigh_entry_destroy(struct mlxsw_sp *mlxsw_sp,
2375                              struct mlxsw_sp_neigh_entry *neigh_entry)
2376 {
2377         list_del(&neigh_entry->rif_list_node);
2378         atomic_dec(&mlxsw_sp->router->neighs_update.neigh_count);
2379         mlxsw_sp_neigh_counter_free(mlxsw_sp, neigh_entry);
2380         mlxsw_sp_neigh_entry_remove(mlxsw_sp, neigh_entry);
2381         mlxsw_sp_neigh_entry_free(neigh_entry);
2382 }
2383
2384 static struct mlxsw_sp_neigh_entry *
2385 mlxsw_sp_neigh_entry_lookup(struct mlxsw_sp *mlxsw_sp, struct neighbour *n)
2386 {
2387         struct mlxsw_sp_neigh_key key;
2388
2389         key.n = n;
2390         return rhashtable_lookup_fast(&mlxsw_sp->router->neigh_ht,
2391                                       &key, mlxsw_sp_neigh_ht_params);
2392 }
2393
2394 static void
2395 mlxsw_sp_router_neighs_update_interval_init(struct mlxsw_sp *mlxsw_sp)
2396 {
2397         unsigned long interval;
2398
2399 #if IS_ENABLED(CONFIG_IPV6)
2400         interval = min_t(unsigned long,
2401                          NEIGH_VAR(&arp_tbl.parms, DELAY_PROBE_TIME),
2402                          NEIGH_VAR(&nd_tbl.parms, DELAY_PROBE_TIME));
2403 #else
2404         interval = NEIGH_VAR(&arp_tbl.parms, DELAY_PROBE_TIME);
2405 #endif
2406         mlxsw_sp->router->neighs_update.interval = jiffies_to_msecs(interval);
2407 }
2408
2409 static void mlxsw_sp_router_neigh_ent_ipv4_process(struct mlxsw_sp *mlxsw_sp,
2410                                                    char *rauhtd_pl,
2411                                                    int ent_index)
2412 {
2413         u64 max_rifs = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_RIFS);
2414         struct net_device *dev;
2415         struct neighbour *n;
2416         __be32 dipn;
2417         u32 dip;
2418         u16 rif;
2419
2420         mlxsw_reg_rauhtd_ent_ipv4_unpack(rauhtd_pl, ent_index, &rif, &dip);
2421
2422         if (WARN_ON_ONCE(rif >= max_rifs))
2423                 return;
2424         if (!mlxsw_sp->router->rifs[rif]) {
2425                 dev_err_ratelimited(mlxsw_sp->bus_info->dev, "Incorrect RIF in neighbour entry\n");
2426                 return;
2427         }
2428
2429         dipn = htonl(dip);
2430         dev = mlxsw_sp_rif_dev(mlxsw_sp->router->rifs[rif]);
2431         n = neigh_lookup(&arp_tbl, &dipn, dev);
2432         if (!n)
2433                 return;
2434
2435         netdev_dbg(dev, "Updating neighbour with IP=%pI4h\n", &dip);
2436         neigh_event_send(n, NULL);
2437         neigh_release(n);
2438 }
2439
2440 #if IS_ENABLED(CONFIG_IPV6)
2441 static void mlxsw_sp_router_neigh_ent_ipv6_process(struct mlxsw_sp *mlxsw_sp,
2442                                                    char *rauhtd_pl,
2443                                                    int rec_index)
2444 {
2445         struct net_device *dev;
2446         struct neighbour *n;
2447         struct in6_addr dip;
2448         u16 rif;
2449
2450         mlxsw_reg_rauhtd_ent_ipv6_unpack(rauhtd_pl, rec_index, &rif,
2451                                          (char *) &dip);
2452
2453         if (!mlxsw_sp->router->rifs[rif]) {
2454                 dev_err_ratelimited(mlxsw_sp->bus_info->dev, "Incorrect RIF in neighbour entry\n");
2455                 return;
2456         }
2457
2458         dev = mlxsw_sp_rif_dev(mlxsw_sp->router->rifs[rif]);
2459         n = neigh_lookup(&nd_tbl, &dip, dev);
2460         if (!n)
2461                 return;
2462
2463         netdev_dbg(dev, "Updating neighbour with IP=%pI6c\n", &dip);
2464         neigh_event_send(n, NULL);
2465         neigh_release(n);
2466 }
2467 #else
2468 static void mlxsw_sp_router_neigh_ent_ipv6_process(struct mlxsw_sp *mlxsw_sp,
2469                                                    char *rauhtd_pl,
2470                                                    int rec_index)
2471 {
2472 }
2473 #endif
2474
2475 static void mlxsw_sp_router_neigh_rec_ipv4_process(struct mlxsw_sp *mlxsw_sp,
2476                                                    char *rauhtd_pl,
2477                                                    int rec_index)
2478 {
2479         u8 num_entries;
2480         int i;
2481
2482         num_entries = mlxsw_reg_rauhtd_ipv4_rec_num_entries_get(rauhtd_pl,
2483                                                                 rec_index);
2484         /* Hardware starts counting at 0, so add 1. */
2485         num_entries++;
2486
2487         /* Each record consists of several neighbour entries. */
2488         for (i = 0; i < num_entries; i++) {
2489                 int ent_index;
2490
2491                 ent_index = rec_index * MLXSW_REG_RAUHTD_IPV4_ENT_PER_REC + i;
2492                 mlxsw_sp_router_neigh_ent_ipv4_process(mlxsw_sp, rauhtd_pl,
2493                                                        ent_index);
2494         }
2495
2496 }
2497
2498 static void mlxsw_sp_router_neigh_rec_ipv6_process(struct mlxsw_sp *mlxsw_sp,
2499                                                    char *rauhtd_pl,
2500                                                    int rec_index)
2501 {
2502         /* One record contains one entry. */
2503         mlxsw_sp_router_neigh_ent_ipv6_process(mlxsw_sp, rauhtd_pl,
2504                                                rec_index);
2505 }
2506
2507 static void mlxsw_sp_router_neigh_rec_process(struct mlxsw_sp *mlxsw_sp,
2508                                               char *rauhtd_pl, int rec_index)
2509 {
2510         switch (mlxsw_reg_rauhtd_rec_type_get(rauhtd_pl, rec_index)) {
2511         case MLXSW_REG_RAUHTD_TYPE_IPV4:
2512                 mlxsw_sp_router_neigh_rec_ipv4_process(mlxsw_sp, rauhtd_pl,
2513                                                        rec_index);
2514                 break;
2515         case MLXSW_REG_RAUHTD_TYPE_IPV6:
2516                 mlxsw_sp_router_neigh_rec_ipv6_process(mlxsw_sp, rauhtd_pl,
2517                                                        rec_index);
2518                 break;
2519         }
2520 }
2521
2522 static bool mlxsw_sp_router_rauhtd_is_full(char *rauhtd_pl)
2523 {
2524         u8 num_rec, last_rec_index, num_entries;
2525
2526         num_rec = mlxsw_reg_rauhtd_num_rec_get(rauhtd_pl);
2527         last_rec_index = num_rec - 1;
2528
2529         if (num_rec < MLXSW_REG_RAUHTD_REC_MAX_NUM)
2530                 return false;
2531         if (mlxsw_reg_rauhtd_rec_type_get(rauhtd_pl, last_rec_index) ==
2532             MLXSW_REG_RAUHTD_TYPE_IPV6)
2533                 return true;
2534
2535         num_entries = mlxsw_reg_rauhtd_ipv4_rec_num_entries_get(rauhtd_pl,
2536                                                                 last_rec_index);
2537         if (++num_entries == MLXSW_REG_RAUHTD_IPV4_ENT_PER_REC)
2538                 return true;
2539         return false;
2540 }
2541
2542 static int
2543 __mlxsw_sp_router_neighs_update_rauhtd(struct mlxsw_sp *mlxsw_sp,
2544                                        char *rauhtd_pl,
2545                                        enum mlxsw_reg_rauhtd_type type)
2546 {
2547         int i, num_rec;
2548         int err;
2549
2550         /* Ensure the RIF we read from the device does not change mid-dump. */
2551         mutex_lock(&mlxsw_sp->router->lock);
2552         do {
2553                 mlxsw_reg_rauhtd_pack(rauhtd_pl, type);
2554                 err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(rauhtd),
2555                                       rauhtd_pl);
2556                 if (err) {
2557                         dev_err_ratelimited(mlxsw_sp->bus_info->dev, "Failed to dump neighbour table\n");
2558                         break;
2559                 }
2560                 num_rec = mlxsw_reg_rauhtd_num_rec_get(rauhtd_pl);
2561                 for (i = 0; i < num_rec; i++)
2562                         mlxsw_sp_router_neigh_rec_process(mlxsw_sp, rauhtd_pl,
2563                                                           i);
2564         } while (mlxsw_sp_router_rauhtd_is_full(rauhtd_pl));
2565         mutex_unlock(&mlxsw_sp->router->lock);
2566
2567         return err;
2568 }
2569
2570 static int mlxsw_sp_router_neighs_update_rauhtd(struct mlxsw_sp *mlxsw_sp)
2571 {
2572         enum mlxsw_reg_rauhtd_type type;
2573         char *rauhtd_pl;
2574         int err;
2575
2576         if (!atomic_read(&mlxsw_sp->router->neighs_update.neigh_count))
2577                 return 0;
2578
2579         rauhtd_pl = kmalloc(MLXSW_REG_RAUHTD_LEN, GFP_KERNEL);
2580         if (!rauhtd_pl)
2581                 return -ENOMEM;
2582
2583         type = MLXSW_REG_RAUHTD_TYPE_IPV4;
2584         err = __mlxsw_sp_router_neighs_update_rauhtd(mlxsw_sp, rauhtd_pl, type);
2585         if (err)
2586                 goto out;
2587
2588         type = MLXSW_REG_RAUHTD_TYPE_IPV6;
2589         err = __mlxsw_sp_router_neighs_update_rauhtd(mlxsw_sp, rauhtd_pl, type);
2590 out:
2591         kfree(rauhtd_pl);
2592         return err;
2593 }
2594
2595 static void mlxsw_sp_router_neighs_update_nh(struct mlxsw_sp *mlxsw_sp)
2596 {
2597         struct mlxsw_sp_neigh_entry *neigh_entry;
2598
2599         mutex_lock(&mlxsw_sp->router->lock);
2600         list_for_each_entry(neigh_entry, &mlxsw_sp->router->nexthop_neighs_list,
2601                             nexthop_neighs_list_node)
2602                 /* If this neigh have nexthops, make the kernel think this neigh
2603                  * is active regardless of the traffic.
2604                  */
2605                 neigh_event_send(neigh_entry->key.n, NULL);
2606         mutex_unlock(&mlxsw_sp->router->lock);
2607 }
2608
2609 static void
2610 mlxsw_sp_router_neighs_update_work_schedule(struct mlxsw_sp *mlxsw_sp)
2611 {
2612         unsigned long interval = mlxsw_sp->router->neighs_update.interval;
2613
2614         mlxsw_core_schedule_dw(&mlxsw_sp->router->neighs_update.dw,
2615                                msecs_to_jiffies(interval));
2616 }
2617
2618 static void mlxsw_sp_router_neighs_update_work(struct work_struct *work)
2619 {
2620         struct mlxsw_sp_router *router;
2621         int err;
2622
2623         router = container_of(work, struct mlxsw_sp_router,
2624                               neighs_update.dw.work);
2625         err = mlxsw_sp_router_neighs_update_rauhtd(router->mlxsw_sp);
2626         if (err)
2627                 dev_err(router->mlxsw_sp->bus_info->dev, "Could not update kernel for neigh activity");
2628
2629         mlxsw_sp_router_neighs_update_nh(router->mlxsw_sp);
2630
2631         mlxsw_sp_router_neighs_update_work_schedule(router->mlxsw_sp);
2632 }
2633
2634 static void mlxsw_sp_router_probe_unresolved_nexthops(struct work_struct *work)
2635 {
2636         struct mlxsw_sp_neigh_entry *neigh_entry;
2637         struct mlxsw_sp_router *router;
2638
2639         router = container_of(work, struct mlxsw_sp_router,
2640                               nexthop_probe_dw.work);
2641         /* Iterate over nexthop neighbours, find those who are unresolved and
2642          * send arp on them. This solves the chicken-egg problem when
2643          * the nexthop wouldn't get offloaded until the neighbor is resolved
2644          * but it wouldn't get resolved ever in case traffic is flowing in HW
2645          * using different nexthop.
2646          */
2647         mutex_lock(&router->lock);
2648         list_for_each_entry(neigh_entry, &router->nexthop_neighs_list,
2649                             nexthop_neighs_list_node)
2650                 if (!neigh_entry->connected)
2651                         neigh_event_send(neigh_entry->key.n, NULL);
2652         mutex_unlock(&router->lock);
2653
2654         mlxsw_core_schedule_dw(&router->nexthop_probe_dw,
2655                                MLXSW_SP_UNRESOLVED_NH_PROBE_INTERVAL);
2656 }
2657
2658 static void
2659 mlxsw_sp_nexthop_neigh_update(struct mlxsw_sp *mlxsw_sp,
2660                               struct mlxsw_sp_neigh_entry *neigh_entry,
2661                               bool removing, bool dead);
2662
2663 static enum mlxsw_reg_rauht_op mlxsw_sp_rauht_op(bool adding)
2664 {
2665         return adding ? MLXSW_REG_RAUHT_OP_WRITE_ADD :
2666                         MLXSW_REG_RAUHT_OP_WRITE_DELETE;
2667 }
2668
2669 static int
2670 mlxsw_sp_router_neigh_entry_op4(struct mlxsw_sp *mlxsw_sp,
2671                                 struct mlxsw_sp_neigh_entry *neigh_entry,
2672                                 enum mlxsw_reg_rauht_op op)
2673 {
2674         struct neighbour *n = neigh_entry->key.n;
2675         u32 dip = ntohl(*((__be32 *) n->primary_key));
2676         char rauht_pl[MLXSW_REG_RAUHT_LEN];
2677
2678         mlxsw_reg_rauht_pack4(rauht_pl, op, neigh_entry->rif, neigh_entry->ha,
2679                               dip);
2680         if (neigh_entry->counter_valid)
2681                 mlxsw_reg_rauht_pack_counter(rauht_pl,
2682                                              neigh_entry->counter_index);
2683         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rauht), rauht_pl);
2684 }
2685
2686 static int
2687 mlxsw_sp_router_neigh_entry_op6(struct mlxsw_sp *mlxsw_sp,
2688                                 struct mlxsw_sp_neigh_entry *neigh_entry,
2689                                 enum mlxsw_reg_rauht_op op)
2690 {
2691         struct neighbour *n = neigh_entry->key.n;
2692         char rauht_pl[MLXSW_REG_RAUHT_LEN];
2693         const char *dip = n->primary_key;
2694
2695         mlxsw_reg_rauht_pack6(rauht_pl, op, neigh_entry->rif, neigh_entry->ha,
2696                               dip);
2697         if (neigh_entry->counter_valid)
2698                 mlxsw_reg_rauht_pack_counter(rauht_pl,
2699                                              neigh_entry->counter_index);
2700         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rauht), rauht_pl);
2701 }
2702
2703 bool mlxsw_sp_neigh_ipv6_ignore(struct mlxsw_sp_neigh_entry *neigh_entry)
2704 {
2705         struct neighbour *n = neigh_entry->key.n;
2706
2707         /* Packets with a link-local destination address are trapped
2708          * after LPM lookup and never reach the neighbour table, so
2709          * there is no need to program such neighbours to the device.
2710          */
2711         if (ipv6_addr_type((struct in6_addr *) &n->primary_key) &
2712             IPV6_ADDR_LINKLOCAL)
2713                 return true;
2714         return false;
2715 }
2716
2717 static void
2718 mlxsw_sp_neigh_entry_update(struct mlxsw_sp *mlxsw_sp,
2719                             struct mlxsw_sp_neigh_entry *neigh_entry,
2720                             bool adding)
2721 {
2722         enum mlxsw_reg_rauht_op op = mlxsw_sp_rauht_op(adding);
2723         int err;
2724
2725         if (!adding && !neigh_entry->connected)
2726                 return;
2727         neigh_entry->connected = adding;
2728         if (neigh_entry->key.n->tbl->family == AF_INET) {
2729                 err = mlxsw_sp_router_neigh_entry_op4(mlxsw_sp, neigh_entry,
2730                                                       op);
2731                 if (err)
2732                         return;
2733         } else if (neigh_entry->key.n->tbl->family == AF_INET6) {
2734                 if (mlxsw_sp_neigh_ipv6_ignore(neigh_entry))
2735                         return;
2736                 err = mlxsw_sp_router_neigh_entry_op6(mlxsw_sp, neigh_entry,
2737                                                       op);
2738                 if (err)
2739                         return;
2740         } else {
2741                 WARN_ON_ONCE(1);
2742                 return;
2743         }
2744
2745         if (adding)
2746                 neigh_entry->key.n->flags |= NTF_OFFLOADED;
2747         else
2748                 neigh_entry->key.n->flags &= ~NTF_OFFLOADED;
2749 }
2750
2751 void
2752 mlxsw_sp_neigh_entry_counter_update(struct mlxsw_sp *mlxsw_sp,
2753                                     struct mlxsw_sp_neigh_entry *neigh_entry,
2754                                     bool adding)
2755 {
2756         if (adding)
2757                 mlxsw_sp_neigh_counter_alloc(mlxsw_sp, neigh_entry);
2758         else
2759                 mlxsw_sp_neigh_counter_free(mlxsw_sp, neigh_entry);
2760         mlxsw_sp_neigh_entry_update(mlxsw_sp, neigh_entry, true);
2761 }
2762
2763 struct mlxsw_sp_netevent_work {
2764         struct work_struct work;
2765         struct mlxsw_sp *mlxsw_sp;
2766         struct neighbour *n;
2767 };
2768
2769 static void mlxsw_sp_router_neigh_event_work(struct work_struct *work)
2770 {
2771         struct mlxsw_sp_netevent_work *net_work =
2772                 container_of(work, struct mlxsw_sp_netevent_work, work);
2773         struct mlxsw_sp *mlxsw_sp = net_work->mlxsw_sp;
2774         struct mlxsw_sp_neigh_entry *neigh_entry;
2775         struct neighbour *n = net_work->n;
2776         unsigned char ha[ETH_ALEN];
2777         bool entry_connected;
2778         u8 nud_state, dead;
2779
2780         /* If these parameters are changed after we release the lock,
2781          * then we are guaranteed to receive another event letting us
2782          * know about it.
2783          */
2784         read_lock_bh(&n->lock);
2785         memcpy(ha, n->ha, ETH_ALEN);
2786         nud_state = n->nud_state;
2787         dead = n->dead;
2788         read_unlock_bh(&n->lock);
2789
2790         mutex_lock(&mlxsw_sp->router->lock);
2791         mlxsw_sp_span_respin(mlxsw_sp);
2792
2793         entry_connected = nud_state & NUD_VALID && !dead;
2794         neigh_entry = mlxsw_sp_neigh_entry_lookup(mlxsw_sp, n);
2795         if (!entry_connected && !neigh_entry)
2796                 goto out;
2797         if (!neigh_entry) {
2798                 neigh_entry = mlxsw_sp_neigh_entry_create(mlxsw_sp, n);
2799                 if (IS_ERR(neigh_entry))
2800                         goto out;
2801         }
2802
2803         if (neigh_entry->connected && entry_connected &&
2804             !memcmp(neigh_entry->ha, ha, ETH_ALEN))
2805                 goto out;
2806
2807         memcpy(neigh_entry->ha, ha, ETH_ALEN);
2808         mlxsw_sp_neigh_entry_update(mlxsw_sp, neigh_entry, entry_connected);
2809         mlxsw_sp_nexthop_neigh_update(mlxsw_sp, neigh_entry, !entry_connected,
2810                                       dead);
2811
2812         if (!neigh_entry->connected && list_empty(&neigh_entry->nexthop_list))
2813                 mlxsw_sp_neigh_entry_destroy(mlxsw_sp, neigh_entry);
2814
2815 out:
2816         mutex_unlock(&mlxsw_sp->router->lock);
2817         neigh_release(n);
2818         kfree(net_work);
2819 }
2820
2821 static int mlxsw_sp_mp_hash_init(struct mlxsw_sp *mlxsw_sp);
2822
2823 static void mlxsw_sp_router_mp_hash_event_work(struct work_struct *work)
2824 {
2825         struct mlxsw_sp_netevent_work *net_work =
2826                 container_of(work, struct mlxsw_sp_netevent_work, work);
2827         struct mlxsw_sp *mlxsw_sp = net_work->mlxsw_sp;
2828
2829         mlxsw_sp_mp_hash_init(mlxsw_sp);
2830         kfree(net_work);
2831 }
2832
2833 static int __mlxsw_sp_router_init(struct mlxsw_sp *mlxsw_sp);
2834
2835 static void mlxsw_sp_router_update_priority_work(struct work_struct *work)
2836 {
2837         struct mlxsw_sp_netevent_work *net_work =
2838                 container_of(work, struct mlxsw_sp_netevent_work, work);
2839         struct mlxsw_sp *mlxsw_sp = net_work->mlxsw_sp;
2840
2841         __mlxsw_sp_router_init(mlxsw_sp);
2842         kfree(net_work);
2843 }
2844
2845 static int mlxsw_sp_router_schedule_work(struct net *net,
2846                                          struct mlxsw_sp_router *router,
2847                                          struct neighbour *n,
2848                                          void (*cb)(struct work_struct *))
2849 {
2850         struct mlxsw_sp_netevent_work *net_work;
2851
2852         if (!net_eq(net, mlxsw_sp_net(router->mlxsw_sp)))
2853                 return NOTIFY_DONE;
2854
2855         net_work = kzalloc(sizeof(*net_work), GFP_ATOMIC);
2856         if (!net_work)
2857                 return NOTIFY_BAD;
2858
2859         INIT_WORK(&net_work->work, cb);
2860         net_work->mlxsw_sp = router->mlxsw_sp;
2861         net_work->n = n;
2862         mlxsw_core_schedule_work(&net_work->work);
2863         return NOTIFY_DONE;
2864 }
2865
2866 static bool mlxsw_sp_dev_lower_is_port(struct net_device *dev)
2867 {
2868         struct mlxsw_sp_port *mlxsw_sp_port;
2869
2870         rcu_read_lock();
2871         mlxsw_sp_port = mlxsw_sp_port_dev_lower_find_rcu(dev);
2872         rcu_read_unlock();
2873         return !!mlxsw_sp_port;
2874 }
2875
2876 static int mlxsw_sp_router_schedule_neigh_work(struct mlxsw_sp_router *router,
2877                                                struct neighbour *n)
2878 {
2879         struct net *net;
2880
2881         net = neigh_parms_net(n->parms);
2882
2883         /* Take a reference to ensure the neighbour won't be destructed until we
2884          * drop the reference in delayed work.
2885          */
2886         neigh_clone(n);
2887         return mlxsw_sp_router_schedule_work(net, router, n,
2888                                              mlxsw_sp_router_neigh_event_work);
2889 }
2890
2891 static int mlxsw_sp_router_netevent_event(struct notifier_block *nb,
2892                                           unsigned long event, void *ptr)
2893 {
2894         struct mlxsw_sp_router *router;
2895         unsigned long interval;
2896         struct neigh_parms *p;
2897         struct neighbour *n;
2898
2899         router = container_of(nb, struct mlxsw_sp_router, netevent_nb);
2900
2901         switch (event) {
2902         case NETEVENT_DELAY_PROBE_TIME_UPDATE:
2903                 p = ptr;
2904
2905                 /* We don't care about changes in the default table. */
2906                 if (!p->dev || (p->tbl->family != AF_INET &&
2907                                 p->tbl->family != AF_INET6))
2908                         return NOTIFY_DONE;
2909
2910                 /* We are in atomic context and can't take RTNL mutex,
2911                  * so use RCU variant to walk the device chain.
2912                  */
2913                 if (!mlxsw_sp_dev_lower_is_port(p->dev))
2914                         return NOTIFY_DONE;
2915
2916                 interval = jiffies_to_msecs(NEIGH_VAR(p, DELAY_PROBE_TIME));
2917                 router->neighs_update.interval = interval;
2918                 break;
2919         case NETEVENT_NEIGH_UPDATE:
2920                 n = ptr;
2921
2922                 if (n->tbl->family != AF_INET && n->tbl->family != AF_INET6)
2923                         return NOTIFY_DONE;
2924
2925                 if (!mlxsw_sp_dev_lower_is_port(n->dev))
2926                         return NOTIFY_DONE;
2927
2928                 return mlxsw_sp_router_schedule_neigh_work(router, n);
2929
2930         case NETEVENT_IPV4_MPATH_HASH_UPDATE:
2931         case NETEVENT_IPV6_MPATH_HASH_UPDATE:
2932                 return mlxsw_sp_router_schedule_work(ptr, router, NULL,
2933                                 mlxsw_sp_router_mp_hash_event_work);
2934
2935         case NETEVENT_IPV4_FWD_UPDATE_PRIORITY_UPDATE:
2936                 return mlxsw_sp_router_schedule_work(ptr, router, NULL,
2937                                 mlxsw_sp_router_update_priority_work);
2938         }
2939
2940         return NOTIFY_DONE;
2941 }
2942
2943 static int mlxsw_sp_neigh_init(struct mlxsw_sp *mlxsw_sp)
2944 {
2945         int err;
2946
2947         err = rhashtable_init(&mlxsw_sp->router->neigh_ht,
2948                               &mlxsw_sp_neigh_ht_params);
2949         if (err)
2950                 return err;
2951
2952         /* Initialize the polling interval according to the default
2953          * table.
2954          */
2955         mlxsw_sp_router_neighs_update_interval_init(mlxsw_sp);
2956
2957         /* Create the delayed works for the activity_update */
2958         INIT_DELAYED_WORK(&mlxsw_sp->router->neighs_update.dw,
2959                           mlxsw_sp_router_neighs_update_work);
2960         INIT_DELAYED_WORK(&mlxsw_sp->router->nexthop_probe_dw,
2961                           mlxsw_sp_router_probe_unresolved_nexthops);
2962         atomic_set(&mlxsw_sp->router->neighs_update.neigh_count, 0);
2963         mlxsw_core_schedule_dw(&mlxsw_sp->router->neighs_update.dw, 0);
2964         mlxsw_core_schedule_dw(&mlxsw_sp->router->nexthop_probe_dw, 0);
2965         return 0;
2966 }
2967
2968 static void mlxsw_sp_neigh_fini(struct mlxsw_sp *mlxsw_sp)
2969 {
2970         cancel_delayed_work_sync(&mlxsw_sp->router->neighs_update.dw);
2971         cancel_delayed_work_sync(&mlxsw_sp->router->nexthop_probe_dw);
2972         rhashtable_destroy(&mlxsw_sp->router->neigh_ht);
2973 }
2974
2975 static void mlxsw_sp_neigh_rif_gone_sync(struct mlxsw_sp *mlxsw_sp,
2976                                          struct mlxsw_sp_rif *rif)
2977 {
2978         struct mlxsw_sp_neigh_entry *neigh_entry, *tmp;
2979
2980         list_for_each_entry_safe(neigh_entry, tmp, &rif->neigh_list,
2981                                  rif_list_node) {
2982                 mlxsw_sp_neigh_entry_update(mlxsw_sp, neigh_entry, false);
2983                 mlxsw_sp_neigh_entry_destroy(mlxsw_sp, neigh_entry);
2984         }
2985 }
2986
2987 struct mlxsw_sp_neigh_rif_made_sync {
2988         struct mlxsw_sp *mlxsw_sp;
2989         struct mlxsw_sp_rif *rif;
2990         int err;
2991 };
2992
2993 static void mlxsw_sp_neigh_rif_made_sync_each(struct neighbour *n, void *data)
2994 {
2995         struct mlxsw_sp_neigh_rif_made_sync *rms = data;
2996         int rc;
2997
2998         if (rms->err)
2999                 return;
3000         if (n->dev != mlxsw_sp_rif_dev(rms->rif))
3001                 return;
3002         rc = mlxsw_sp_router_schedule_neigh_work(rms->mlxsw_sp->router, n);
3003         if (rc != NOTIFY_DONE)
3004                 rms->err = -ENOMEM;
3005 }
3006
3007 static int mlxsw_sp_neigh_rif_made_sync(struct mlxsw_sp *mlxsw_sp,
3008                                         struct mlxsw_sp_rif *rif)
3009 {
3010         struct mlxsw_sp_neigh_rif_made_sync rms = {
3011                 .mlxsw_sp = mlxsw_sp,
3012                 .rif = rif,
3013         };
3014
3015         neigh_for_each(&arp_tbl, mlxsw_sp_neigh_rif_made_sync_each, &rms);
3016         if (rms.err)
3017                 goto err_arp;
3018
3019 #if IS_ENABLED(CONFIG_IPV6)
3020         neigh_for_each(&nd_tbl, mlxsw_sp_neigh_rif_made_sync_each, &rms);
3021 #endif
3022         if (rms.err)
3023                 goto err_nd;
3024
3025         return 0;
3026
3027 err_nd:
3028 err_arp:
3029         mlxsw_sp_neigh_rif_gone_sync(mlxsw_sp, rif);
3030         return rms.err;
3031 }
3032
3033 enum mlxsw_sp_nexthop_type {
3034         MLXSW_SP_NEXTHOP_TYPE_ETH,
3035         MLXSW_SP_NEXTHOP_TYPE_IPIP,
3036 };
3037
3038 enum mlxsw_sp_nexthop_action {
3039         /* Nexthop forwards packets to an egress RIF */
3040         MLXSW_SP_NEXTHOP_ACTION_FORWARD,
3041         /* Nexthop discards packets */
3042         MLXSW_SP_NEXTHOP_ACTION_DISCARD,
3043         /* Nexthop traps packets */
3044         MLXSW_SP_NEXTHOP_ACTION_TRAP,
3045 };
3046
3047 struct mlxsw_sp_nexthop_key {
3048         struct fib_nh *fib_nh;
3049 };
3050
3051 struct mlxsw_sp_nexthop {
3052         struct list_head neigh_list_node; /* member of neigh entry list */
3053         struct list_head crif_list_node;
3054         struct list_head router_list_node;
3055         struct mlxsw_sp_nexthop_group_info *nhgi; /* pointer back to the group
3056                                                    * this nexthop belongs to
3057                                                    */
3058         struct rhash_head ht_node;
3059         struct neigh_table *neigh_tbl;
3060         struct mlxsw_sp_nexthop_key key;
3061         unsigned char gw_addr[sizeof(struct in6_addr)];
3062         int ifindex;
3063         int nh_weight;
3064         int norm_nh_weight;
3065         int num_adj_entries;
3066         struct mlxsw_sp_crif *crif;
3067         u8 should_offload:1, /* set indicates this nexthop should be written
3068                               * to the adjacency table.
3069                               */
3070            offloaded:1, /* set indicates this nexthop was written to the
3071                          * adjacency table.
3072                          */
3073            update:1; /* set indicates this nexthop should be updated in the
3074                       * adjacency table (f.e., its MAC changed).
3075                       */
3076         enum mlxsw_sp_nexthop_action action;
3077         enum mlxsw_sp_nexthop_type type;
3078         union {
3079                 struct mlxsw_sp_neigh_entry *neigh_entry;
3080                 struct mlxsw_sp_ipip_entry *ipip_entry;
3081         };
3082         unsigned int counter_index;
3083         bool counter_valid;
3084 };
3085
3086 static struct net_device *
3087 mlxsw_sp_nexthop_dev(const struct mlxsw_sp_nexthop *nh)
3088 {
3089         if (!nh->crif)
3090                 return NULL;
3091         return nh->crif->key.dev;
3092 }
3093
3094 enum mlxsw_sp_nexthop_group_type {
3095         MLXSW_SP_NEXTHOP_GROUP_TYPE_IPV4,
3096         MLXSW_SP_NEXTHOP_GROUP_TYPE_IPV6,
3097         MLXSW_SP_NEXTHOP_GROUP_TYPE_OBJ,
3098 };
3099
3100 struct mlxsw_sp_nexthop_group_info {
3101         struct mlxsw_sp_nexthop_group *nh_grp;
3102         u32 adj_index;
3103         u16 ecmp_size;
3104         u16 count;
3105         int sum_norm_weight;
3106         u8 adj_index_valid:1,
3107            gateway:1, /* routes using the group use a gateway */
3108            is_resilient:1;
3109         struct list_head list; /* member in nh_res_grp_list */
3110         struct mlxsw_sp_nexthop nexthops[];
3111 };
3112
3113 static struct mlxsw_sp_rif *
3114 mlxsw_sp_nhgi_rif(const struct mlxsw_sp_nexthop_group_info *nhgi)
3115 {
3116         struct mlxsw_sp_crif *crif = nhgi->nexthops[0].crif;
3117
3118         if (!crif)
3119                 return NULL;
3120         return crif->rif;
3121 }
3122
3123 struct mlxsw_sp_nexthop_group_vr_key {
3124         u16 vr_id;
3125         enum mlxsw_sp_l3proto proto;
3126 };
3127
3128 struct mlxsw_sp_nexthop_group_vr_entry {
3129         struct list_head list; /* member in vr_list */
3130         struct rhash_head ht_node; /* member in vr_ht */
3131         refcount_t ref_count;
3132         struct mlxsw_sp_nexthop_group_vr_key key;
3133 };
3134
3135 struct mlxsw_sp_nexthop_group {
3136         struct rhash_head ht_node;
3137         struct list_head fib_list; /* list of fib entries that use this group */
3138         union {
3139                 struct {
3140                         struct fib_info *fi;
3141                 } ipv4;
3142                 struct {
3143                         u32 id;
3144                 } obj;
3145         };
3146         struct mlxsw_sp_nexthop_group_info *nhgi;
3147         struct list_head vr_list;
3148         struct rhashtable vr_ht;
3149         enum mlxsw_sp_nexthop_group_type type;
3150         bool can_destroy;
3151 };
3152
3153 void mlxsw_sp_nexthop_counter_alloc(struct mlxsw_sp *mlxsw_sp,
3154                                     struct mlxsw_sp_nexthop *nh)
3155 {
3156         struct devlink *devlink;
3157
3158         devlink = priv_to_devlink(mlxsw_sp->core);
3159         if (!devlink_dpipe_table_counter_enabled(devlink,
3160                                                  MLXSW_SP_DPIPE_TABLE_NAME_ADJ))
3161                 return;
3162
3163         if (mlxsw_sp_flow_counter_alloc(mlxsw_sp, &nh->counter_index))
3164                 return;
3165
3166         nh->counter_valid = true;
3167 }
3168
3169 void mlxsw_sp_nexthop_counter_free(struct mlxsw_sp *mlxsw_sp,
3170                                    struct mlxsw_sp_nexthop *nh)
3171 {
3172         if (!nh->counter_valid)
3173                 return;
3174         mlxsw_sp_flow_counter_free(mlxsw_sp, nh->counter_index);
3175         nh->counter_valid = false;
3176 }
3177
3178 int mlxsw_sp_nexthop_counter_get(struct mlxsw_sp *mlxsw_sp,
3179                                  struct mlxsw_sp_nexthop *nh, u64 *p_counter)
3180 {
3181         if (!nh->counter_valid)
3182                 return -EINVAL;
3183
3184         return mlxsw_sp_flow_counter_get(mlxsw_sp, nh->counter_index,
3185                                          p_counter, NULL);
3186 }
3187
3188 struct mlxsw_sp_nexthop *mlxsw_sp_nexthop_next(struct mlxsw_sp_router *router,
3189                                                struct mlxsw_sp_nexthop *nh)
3190 {
3191         if (!nh) {
3192                 if (list_empty(&router->nexthop_list))
3193                         return NULL;
3194                 else
3195                         return list_first_entry(&router->nexthop_list,
3196                                                 typeof(*nh), router_list_node);
3197         }
3198         if (list_is_last(&nh->router_list_node, &router->nexthop_list))
3199                 return NULL;
3200         return list_next_entry(nh, router_list_node);
3201 }
3202
3203 bool mlxsw_sp_nexthop_is_forward(const struct mlxsw_sp_nexthop *nh)
3204 {
3205         return nh->offloaded && nh->action == MLXSW_SP_NEXTHOP_ACTION_FORWARD;
3206 }
3207
3208 unsigned char *mlxsw_sp_nexthop_ha(struct mlxsw_sp_nexthop *nh)
3209 {
3210         if (nh->type != MLXSW_SP_NEXTHOP_TYPE_ETH ||
3211             !mlxsw_sp_nexthop_is_forward(nh))
3212                 return NULL;
3213         return nh->neigh_entry->ha;
3214 }
3215
3216 int mlxsw_sp_nexthop_indexes(struct mlxsw_sp_nexthop *nh, u32 *p_adj_index,
3217                              u32 *p_adj_size, u32 *p_adj_hash_index)
3218 {
3219         struct mlxsw_sp_nexthop_group_info *nhgi = nh->nhgi;
3220         u32 adj_hash_index = 0;
3221         int i;
3222
3223         if (!nh->offloaded || !nhgi->adj_index_valid)
3224                 return -EINVAL;
3225
3226         *p_adj_index = nhgi->adj_index;
3227         *p_adj_size = nhgi->ecmp_size;
3228
3229         for (i = 0; i < nhgi->count; i++) {
3230                 struct mlxsw_sp_nexthop *nh_iter = &nhgi->nexthops[i];
3231
3232                 if (nh_iter == nh)
3233                         break;
3234                 if (nh_iter->offloaded)
3235                         adj_hash_index += nh_iter->num_adj_entries;
3236         }
3237
3238         *p_adj_hash_index = adj_hash_index;
3239         return 0;
3240 }
3241
3242 struct mlxsw_sp_rif *mlxsw_sp_nexthop_rif(struct mlxsw_sp_nexthop *nh)
3243 {
3244         if (WARN_ON(!nh->crif))
3245                 return NULL;
3246         return nh->crif->rif;
3247 }
3248
3249 bool mlxsw_sp_nexthop_group_has_ipip(struct mlxsw_sp_nexthop *nh)
3250 {
3251         struct mlxsw_sp_nexthop_group_info *nhgi = nh->nhgi;
3252         int i;
3253
3254         for (i = 0; i < nhgi->count; i++) {
3255                 struct mlxsw_sp_nexthop *nh_iter = &nhgi->nexthops[i];
3256
3257                 if (nh_iter->type == MLXSW_SP_NEXTHOP_TYPE_IPIP)
3258                         return true;
3259         }
3260         return false;
3261 }
3262
3263 static const struct rhashtable_params mlxsw_sp_nexthop_group_vr_ht_params = {
3264         .key_offset = offsetof(struct mlxsw_sp_nexthop_group_vr_entry, key),
3265         .head_offset = offsetof(struct mlxsw_sp_nexthop_group_vr_entry, ht_node),
3266         .key_len = sizeof(struct mlxsw_sp_nexthop_group_vr_key),
3267         .automatic_shrinking = true,
3268 };
3269
3270 static struct mlxsw_sp_nexthop_group_vr_entry *
3271 mlxsw_sp_nexthop_group_vr_entry_lookup(struct mlxsw_sp_nexthop_group *nh_grp,
3272                                        const struct mlxsw_sp_fib *fib)
3273 {
3274         struct mlxsw_sp_nexthop_group_vr_key key;
3275
3276         memset(&key, 0, sizeof(key));
3277         key.vr_id = fib->vr->id;
3278         key.proto = fib->proto;
3279         return rhashtable_lookup_fast(&nh_grp->vr_ht, &key,
3280                                       mlxsw_sp_nexthop_group_vr_ht_params);
3281 }
3282
3283 static int
3284 mlxsw_sp_nexthop_group_vr_entry_create(struct mlxsw_sp_nexthop_group *nh_grp,
3285                                        const struct mlxsw_sp_fib *fib)
3286 {
3287         struct mlxsw_sp_nexthop_group_vr_entry *vr_entry;
3288         int err;
3289
3290         vr_entry = kzalloc(sizeof(*vr_entry), GFP_KERNEL);
3291         if (!vr_entry)
3292                 return -ENOMEM;
3293
3294         vr_entry->key.vr_id = fib->vr->id;
3295         vr_entry->key.proto = fib->proto;
3296         refcount_set(&vr_entry->ref_count, 1);
3297
3298         err = rhashtable_insert_fast(&nh_grp->vr_ht, &vr_entry->ht_node,
3299                                      mlxsw_sp_nexthop_group_vr_ht_params);
3300         if (err)
3301                 goto err_hashtable_insert;
3302
3303         list_add(&vr_entry->list, &nh_grp->vr_list);
3304
3305         return 0;
3306
3307 err_hashtable_insert:
3308         kfree(vr_entry);
3309         return err;
3310 }
3311
3312 static void
3313 mlxsw_sp_nexthop_group_vr_entry_destroy(struct mlxsw_sp_nexthop_group *nh_grp,
3314                                         struct mlxsw_sp_nexthop_group_vr_entry *vr_entry)
3315 {
3316         list_del(&vr_entry->list);
3317         rhashtable_remove_fast(&nh_grp->vr_ht, &vr_entry->ht_node,
3318                                mlxsw_sp_nexthop_group_vr_ht_params);
3319         kfree(vr_entry);
3320 }
3321
3322 static int
3323 mlxsw_sp_nexthop_group_vr_link(struct mlxsw_sp_nexthop_group *nh_grp,
3324                                const struct mlxsw_sp_fib *fib)
3325 {
3326         struct mlxsw_sp_nexthop_group_vr_entry *vr_entry;
3327
3328         vr_entry = mlxsw_sp_nexthop_group_vr_entry_lookup(nh_grp, fib);
3329         if (vr_entry) {
3330                 refcount_inc(&vr_entry->ref_count);
3331                 return 0;
3332         }
3333
3334         return mlxsw_sp_nexthop_group_vr_entry_create(nh_grp, fib);
3335 }
3336
3337 static void
3338 mlxsw_sp_nexthop_group_vr_unlink(struct mlxsw_sp_nexthop_group *nh_grp,
3339                                  const struct mlxsw_sp_fib *fib)
3340 {
3341         struct mlxsw_sp_nexthop_group_vr_entry *vr_entry;
3342
3343         vr_entry = mlxsw_sp_nexthop_group_vr_entry_lookup(nh_grp, fib);
3344         if (WARN_ON_ONCE(!vr_entry))
3345                 return;
3346
3347         if (!refcount_dec_and_test(&vr_entry->ref_count))
3348                 return;
3349
3350         mlxsw_sp_nexthop_group_vr_entry_destroy(nh_grp, vr_entry);
3351 }
3352
3353 struct mlxsw_sp_nexthop_group_cmp_arg {
3354         enum mlxsw_sp_nexthop_group_type type;
3355         union {
3356                 struct fib_info *fi;
3357                 struct mlxsw_sp_fib6_entry *fib6_entry;
3358                 u32 id;
3359         };
3360 };
3361
3362 static bool
3363 mlxsw_sp_nexthop6_group_has_nexthop(const struct mlxsw_sp_nexthop_group *nh_grp,
3364                                     const struct in6_addr *gw, int ifindex,
3365                                     int weight)
3366 {
3367         int i;
3368
3369         for (i = 0; i < nh_grp->nhgi->count; i++) {
3370                 const struct mlxsw_sp_nexthop *nh;
3371
3372                 nh = &nh_grp->nhgi->nexthops[i];
3373                 if (nh->ifindex == ifindex && nh->nh_weight == weight &&
3374                     ipv6_addr_equal(gw, (struct in6_addr *) nh->gw_addr))
3375                         return true;
3376         }
3377
3378         return false;
3379 }
3380
3381 static bool
3382 mlxsw_sp_nexthop6_group_cmp(const struct mlxsw_sp_nexthop_group *nh_grp,
3383                             const struct mlxsw_sp_fib6_entry *fib6_entry)
3384 {
3385         struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
3386
3387         if (nh_grp->nhgi->count != fib6_entry->nrt6)
3388                 return false;
3389
3390         list_for_each_entry(mlxsw_sp_rt6, &fib6_entry->rt6_list, list) {
3391                 struct fib6_nh *fib6_nh = mlxsw_sp_rt6->rt->fib6_nh;
3392                 struct in6_addr *gw;
3393                 int ifindex, weight;
3394
3395                 ifindex = fib6_nh->fib_nh_dev->ifindex;
3396                 weight = fib6_nh->fib_nh_weight;
3397                 gw = &fib6_nh->fib_nh_gw6;
3398                 if (!mlxsw_sp_nexthop6_group_has_nexthop(nh_grp, gw, ifindex,
3399                                                          weight))
3400                         return false;
3401         }
3402
3403         return true;
3404 }
3405
3406 static int
3407 mlxsw_sp_nexthop_group_cmp(struct rhashtable_compare_arg *arg, const void *ptr)
3408 {
3409         const struct mlxsw_sp_nexthop_group_cmp_arg *cmp_arg = arg->key;
3410         const struct mlxsw_sp_nexthop_group *nh_grp = ptr;
3411
3412         if (nh_grp->type != cmp_arg->type)
3413                 return 1;
3414
3415         switch (cmp_arg->type) {
3416         case MLXSW_SP_NEXTHOP_GROUP_TYPE_IPV4:
3417                 return cmp_arg->fi != nh_grp->ipv4.fi;
3418         case MLXSW_SP_NEXTHOP_GROUP_TYPE_IPV6:
3419                 return !mlxsw_sp_nexthop6_group_cmp(nh_grp,
3420                                                     cmp_arg->fib6_entry);
3421         case MLXSW_SP_NEXTHOP_GROUP_TYPE_OBJ:
3422                 return cmp_arg->id != nh_grp->obj.id;
3423         default:
3424                 WARN_ON(1);
3425                 return 1;
3426         }
3427 }
3428
3429 static u32 mlxsw_sp_nexthop_group_hash_obj(const void *data, u32 len, u32 seed)
3430 {
3431         const struct mlxsw_sp_nexthop_group *nh_grp = data;
3432         const struct mlxsw_sp_nexthop *nh;
3433         struct fib_info *fi;
3434         unsigned int val;
3435         int i;
3436
3437         switch (nh_grp->type) {
3438         case MLXSW_SP_NEXTHOP_GROUP_TYPE_IPV4:
3439                 fi = nh_grp->ipv4.fi;
3440                 return jhash(&fi, sizeof(fi), seed);
3441         case MLXSW_SP_NEXTHOP_GROUP_TYPE_IPV6:
3442                 val = nh_grp->nhgi->count;
3443                 for (i = 0; i < nh_grp->nhgi->count; i++) {
3444                         nh = &nh_grp->nhgi->nexthops[i];
3445                         val ^= jhash(&nh->ifindex, sizeof(nh->ifindex), seed);
3446                         val ^= jhash(&nh->gw_addr, sizeof(nh->gw_addr), seed);
3447                 }
3448                 return jhash(&val, sizeof(val), seed);
3449         case MLXSW_SP_NEXTHOP_GROUP_TYPE_OBJ:
3450                 return jhash(&nh_grp->obj.id, sizeof(nh_grp->obj.id), seed);
3451         default:
3452                 WARN_ON(1);
3453                 return 0;
3454         }
3455 }
3456
3457 static u32
3458 mlxsw_sp_nexthop6_group_hash(struct mlxsw_sp_fib6_entry *fib6_entry, u32 seed)
3459 {
3460         unsigned int val = fib6_entry->nrt6;
3461         struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
3462
3463         list_for_each_entry(mlxsw_sp_rt6, &fib6_entry->rt6_list, list) {
3464                 struct fib6_nh *fib6_nh = mlxsw_sp_rt6->rt->fib6_nh;
3465                 struct net_device *dev = fib6_nh->fib_nh_dev;
3466                 struct in6_addr *gw = &fib6_nh->fib_nh_gw6;
3467
3468                 val ^= jhash(&dev->ifindex, sizeof(dev->ifindex), seed);
3469                 val ^= jhash(gw, sizeof(*gw), seed);
3470         }
3471
3472         return jhash(&val, sizeof(val), seed);
3473 }
3474
3475 static u32
3476 mlxsw_sp_nexthop_group_hash(const void *data, u32 len, u32 seed)
3477 {
3478         const struct mlxsw_sp_nexthop_group_cmp_arg *cmp_arg = data;
3479
3480         switch (cmp_arg->type) {
3481         case MLXSW_SP_NEXTHOP_GROUP_TYPE_IPV4:
3482                 return jhash(&cmp_arg->fi, sizeof(cmp_arg->fi), seed);
3483         case MLXSW_SP_NEXTHOP_GROUP_TYPE_IPV6:
3484                 return mlxsw_sp_nexthop6_group_hash(cmp_arg->fib6_entry, seed);
3485         case MLXSW_SP_NEXTHOP_GROUP_TYPE_OBJ:
3486                 return jhash(&cmp_arg->id, sizeof(cmp_arg->id), seed);
3487         default:
3488                 WARN_ON(1);
3489                 return 0;
3490         }
3491 }
3492
3493 static const struct rhashtable_params mlxsw_sp_nexthop_group_ht_params = {
3494         .head_offset = offsetof(struct mlxsw_sp_nexthop_group, ht_node),
3495         .hashfn      = mlxsw_sp_nexthop_group_hash,
3496         .obj_hashfn  = mlxsw_sp_nexthop_group_hash_obj,
3497         .obj_cmpfn   = mlxsw_sp_nexthop_group_cmp,
3498 };
3499
3500 static int mlxsw_sp_nexthop_group_insert(struct mlxsw_sp *mlxsw_sp,
3501                                          struct mlxsw_sp_nexthop_group *nh_grp)
3502 {
3503         if (nh_grp->type == MLXSW_SP_NEXTHOP_GROUP_TYPE_IPV6 &&
3504             !nh_grp->nhgi->gateway)
3505                 return 0;
3506
3507         return rhashtable_insert_fast(&mlxsw_sp->router->nexthop_group_ht,
3508                                       &nh_grp->ht_node,
3509                                       mlxsw_sp_nexthop_group_ht_params);
3510 }
3511
3512 static void mlxsw_sp_nexthop_group_remove(struct mlxsw_sp *mlxsw_sp,
3513                                           struct mlxsw_sp_nexthop_group *nh_grp)
3514 {
3515         if (nh_grp->type == MLXSW_SP_NEXTHOP_GROUP_TYPE_IPV6 &&
3516             !nh_grp->nhgi->gateway)
3517                 return;
3518
3519         rhashtable_remove_fast(&mlxsw_sp->router->nexthop_group_ht,
3520                                &nh_grp->ht_node,
3521                                mlxsw_sp_nexthop_group_ht_params);
3522 }
3523
3524 static struct mlxsw_sp_nexthop_group *
3525 mlxsw_sp_nexthop4_group_lookup(struct mlxsw_sp *mlxsw_sp,
3526                                struct fib_info *fi)
3527 {
3528         struct mlxsw_sp_nexthop_group_cmp_arg cmp_arg;
3529
3530         cmp_arg.type = MLXSW_SP_NEXTHOP_GROUP_TYPE_IPV4;
3531         cmp_arg.fi = fi;
3532         return rhashtable_lookup_fast(&mlxsw_sp->router->nexthop_group_ht,
3533                                       &cmp_arg,
3534                                       mlxsw_sp_nexthop_group_ht_params);
3535 }
3536
3537 static struct mlxsw_sp_nexthop_group *
3538 mlxsw_sp_nexthop6_group_lookup(struct mlxsw_sp *mlxsw_sp,
3539                                struct mlxsw_sp_fib6_entry *fib6_entry)
3540 {
3541         struct mlxsw_sp_nexthop_group_cmp_arg cmp_arg;
3542
3543         cmp_arg.type = MLXSW_SP_NEXTHOP_GROUP_TYPE_IPV6;
3544         cmp_arg.fib6_entry = fib6_entry;
3545         return rhashtable_lookup_fast(&mlxsw_sp->router->nexthop_group_ht,
3546                                       &cmp_arg,
3547                                       mlxsw_sp_nexthop_group_ht_params);
3548 }
3549
3550 static const struct rhashtable_params mlxsw_sp_nexthop_ht_params = {
3551         .key_offset = offsetof(struct mlxsw_sp_nexthop, key),
3552         .head_offset = offsetof(struct mlxsw_sp_nexthop, ht_node),
3553         .key_len = sizeof(struct mlxsw_sp_nexthop_key),
3554 };
3555
3556 static int mlxsw_sp_nexthop_insert(struct mlxsw_sp *mlxsw_sp,
3557                                    struct mlxsw_sp_nexthop *nh)
3558 {
3559         return rhashtable_insert_fast(&mlxsw_sp->router->nexthop_ht,
3560                                       &nh->ht_node, mlxsw_sp_nexthop_ht_params);
3561 }
3562
3563 static void mlxsw_sp_nexthop_remove(struct mlxsw_sp *mlxsw_sp,
3564                                     struct mlxsw_sp_nexthop *nh)
3565 {
3566         rhashtable_remove_fast(&mlxsw_sp->router->nexthop_ht, &nh->ht_node,
3567                                mlxsw_sp_nexthop_ht_params);
3568 }
3569
3570 static struct mlxsw_sp_nexthop *
3571 mlxsw_sp_nexthop_lookup(struct mlxsw_sp *mlxsw_sp,
3572                         struct mlxsw_sp_nexthop_key key)
3573 {
3574         return rhashtable_lookup_fast(&mlxsw_sp->router->nexthop_ht, &key,
3575                                       mlxsw_sp_nexthop_ht_params);
3576 }
3577
3578 static int mlxsw_sp_adj_index_mass_update_vr(struct mlxsw_sp *mlxsw_sp,
3579                                              enum mlxsw_sp_l3proto proto,
3580                                              u16 vr_id,
3581                                              u32 adj_index, u16 ecmp_size,
3582                                              u32 new_adj_index,
3583                                              u16 new_ecmp_size)
3584 {
3585         char raleu_pl[MLXSW_REG_RALEU_LEN];
3586
3587         mlxsw_reg_raleu_pack(raleu_pl,
3588                              (enum mlxsw_reg_ralxx_protocol) proto, vr_id,
3589                              adj_index, ecmp_size, new_adj_index,
3590                              new_ecmp_size);
3591         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(raleu), raleu_pl);
3592 }
3593
3594 static int mlxsw_sp_adj_index_mass_update(struct mlxsw_sp *mlxsw_sp,
3595                                           struct mlxsw_sp_nexthop_group *nh_grp,
3596                                           u32 old_adj_index, u16 old_ecmp_size)
3597 {
3598         struct mlxsw_sp_nexthop_group_info *nhgi = nh_grp->nhgi;
3599         struct mlxsw_sp_nexthop_group_vr_entry *vr_entry;
3600         int err;
3601
3602         list_for_each_entry(vr_entry, &nh_grp->vr_list, list) {
3603                 err = mlxsw_sp_adj_index_mass_update_vr(mlxsw_sp,
3604                                                         vr_entry->key.proto,
3605                                                         vr_entry->key.vr_id,
3606                                                         old_adj_index,
3607                                                         old_ecmp_size,
3608                                                         nhgi->adj_index,
3609                                                         nhgi->ecmp_size);
3610                 if (err)
3611                         goto err_mass_update_vr;
3612         }
3613         return 0;
3614
3615 err_mass_update_vr:
3616         list_for_each_entry_continue_reverse(vr_entry, &nh_grp->vr_list, list)
3617                 mlxsw_sp_adj_index_mass_update_vr(mlxsw_sp, vr_entry->key.proto,
3618                                                   vr_entry->key.vr_id,
3619                                                   nhgi->adj_index,
3620                                                   nhgi->ecmp_size,
3621                                                   old_adj_index, old_ecmp_size);
3622         return err;
3623 }
3624
3625 static int __mlxsw_sp_nexthop_eth_update(struct mlxsw_sp *mlxsw_sp,
3626                                          u32 adj_index,
3627                                          struct mlxsw_sp_nexthop *nh,
3628                                          bool force, char *ratr_pl)
3629 {
3630         struct mlxsw_sp_neigh_entry *neigh_entry = nh->neigh_entry;
3631         struct mlxsw_sp_rif *rif = mlxsw_sp_nexthop_rif(nh);
3632         enum mlxsw_reg_ratr_op op;
3633         u16 rif_index;
3634
3635         rif_index = rif ? rif->rif_index :
3636                           mlxsw_sp->router->lb_crif->rif->rif_index;
3637         op = force ? MLXSW_REG_RATR_OP_WRITE_WRITE_ENTRY :
3638                      MLXSW_REG_RATR_OP_WRITE_WRITE_ENTRY_ON_ACTIVITY;
3639         mlxsw_reg_ratr_pack(ratr_pl, op, true, MLXSW_REG_RATR_TYPE_ETHERNET,
3640                             adj_index, rif_index);
3641         switch (nh->action) {
3642         case MLXSW_SP_NEXTHOP_ACTION_FORWARD:
3643                 mlxsw_reg_ratr_eth_entry_pack(ratr_pl, neigh_entry->ha);
3644                 break;
3645         case MLXSW_SP_NEXTHOP_ACTION_DISCARD:
3646                 mlxsw_reg_ratr_trap_action_set(ratr_pl,
3647                                                MLXSW_REG_RATR_TRAP_ACTION_DISCARD_ERRORS);
3648                 break;
3649         case MLXSW_SP_NEXTHOP_ACTION_TRAP:
3650                 mlxsw_reg_ratr_trap_action_set(ratr_pl,
3651                                                MLXSW_REG_RATR_TRAP_ACTION_TRAP);
3652                 mlxsw_reg_ratr_trap_id_set(ratr_pl, MLXSW_TRAP_ID_RTR_EGRESS0);
3653                 break;
3654         default:
3655                 WARN_ON_ONCE(1);
3656                 return -EINVAL;
3657         }
3658         if (nh->counter_valid)
3659                 mlxsw_reg_ratr_counter_pack(ratr_pl, nh->counter_index, true);
3660         else
3661                 mlxsw_reg_ratr_counter_pack(ratr_pl, 0, false);
3662
3663         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ratr), ratr_pl);
3664 }
3665
3666 int mlxsw_sp_nexthop_eth_update(struct mlxsw_sp *mlxsw_sp, u32 adj_index,
3667                                 struct mlxsw_sp_nexthop *nh, bool force,
3668                                 char *ratr_pl)
3669 {
3670         int i;
3671
3672         for (i = 0; i < nh->num_adj_entries; i++) {
3673                 int err;
3674
3675                 err = __mlxsw_sp_nexthop_eth_update(mlxsw_sp, adj_index + i,
3676                                                     nh, force, ratr_pl);
3677                 if (err)
3678                         return err;
3679         }
3680
3681         return 0;
3682 }
3683
3684 static int __mlxsw_sp_nexthop_ipip_update(struct mlxsw_sp *mlxsw_sp,
3685                                           u32 adj_index,
3686                                           struct mlxsw_sp_nexthop *nh,
3687                                           bool force, char *ratr_pl)
3688 {
3689         const struct mlxsw_sp_ipip_ops *ipip_ops;
3690
3691         ipip_ops = mlxsw_sp->router->ipip_ops_arr[nh->ipip_entry->ipipt];
3692         return ipip_ops->nexthop_update(mlxsw_sp, adj_index, nh->ipip_entry,
3693                                         force, ratr_pl);
3694 }
3695
3696 static int mlxsw_sp_nexthop_ipip_update(struct mlxsw_sp *mlxsw_sp,
3697                                         u32 adj_index,
3698                                         struct mlxsw_sp_nexthop *nh, bool force,
3699                                         char *ratr_pl)
3700 {
3701         int i;
3702
3703         for (i = 0; i < nh->num_adj_entries; i++) {
3704                 int err;
3705
3706                 err = __mlxsw_sp_nexthop_ipip_update(mlxsw_sp, adj_index + i,
3707                                                      nh, force, ratr_pl);
3708                 if (err)
3709                         return err;
3710         }
3711
3712         return 0;
3713 }
3714
3715 static int mlxsw_sp_nexthop_update(struct mlxsw_sp *mlxsw_sp, u32 adj_index,
3716                                    struct mlxsw_sp_nexthop *nh, bool force,
3717                                    char *ratr_pl)
3718 {
3719         /* When action is discard or trap, the nexthop must be
3720          * programmed as an Ethernet nexthop.
3721          */
3722         if (nh->type == MLXSW_SP_NEXTHOP_TYPE_ETH ||
3723             nh->action == MLXSW_SP_NEXTHOP_ACTION_DISCARD ||
3724             nh->action == MLXSW_SP_NEXTHOP_ACTION_TRAP)
3725                 return mlxsw_sp_nexthop_eth_update(mlxsw_sp, adj_index, nh,
3726                                                    force, ratr_pl);
3727         else
3728                 return mlxsw_sp_nexthop_ipip_update(mlxsw_sp, adj_index, nh,
3729                                                     force, ratr_pl);
3730 }
3731
3732 static int
3733 mlxsw_sp_nexthop_group_update(struct mlxsw_sp *mlxsw_sp,
3734                               struct mlxsw_sp_nexthop_group_info *nhgi,
3735                               bool reallocate)
3736 {
3737         char ratr_pl[MLXSW_REG_RATR_LEN];
3738         u32 adj_index = nhgi->adj_index; /* base */
3739         struct mlxsw_sp_nexthop *nh;
3740         int i;
3741
3742         for (i = 0; i < nhgi->count; i++) {
3743                 nh = &nhgi->nexthops[i];
3744
3745                 if (!nh->should_offload) {
3746                         nh->offloaded = 0;
3747                         continue;
3748                 }
3749
3750                 if (nh->update || reallocate) {
3751                         int err = 0;
3752
3753                         err = mlxsw_sp_nexthop_update(mlxsw_sp, adj_index, nh,
3754                                                       true, ratr_pl);
3755                         if (err)
3756                                 return err;
3757                         nh->update = 0;
3758                         nh->offloaded = 1;
3759                 }
3760                 adj_index += nh->num_adj_entries;
3761         }
3762         return 0;
3763 }
3764
3765 static int
3766 mlxsw_sp_nexthop_fib_entries_update(struct mlxsw_sp *mlxsw_sp,
3767                                     struct mlxsw_sp_nexthop_group *nh_grp)
3768 {
3769         struct mlxsw_sp_fib_entry *fib_entry;
3770         int err;
3771
3772         list_for_each_entry(fib_entry, &nh_grp->fib_list, nexthop_group_node) {
3773                 err = mlxsw_sp_fib_entry_update(mlxsw_sp, fib_entry);
3774                 if (err)
3775                         return err;
3776         }
3777         return 0;
3778 }
3779
3780 struct mlxsw_sp_adj_grp_size_range {
3781         u16 start; /* Inclusive */
3782         u16 end; /* Inclusive */
3783 };
3784
3785 /* Ordered by range start value */
3786 static const struct mlxsw_sp_adj_grp_size_range
3787 mlxsw_sp1_adj_grp_size_ranges[] = {
3788         { .start = 1, .end = 64 },
3789         { .start = 512, .end = 512 },
3790         { .start = 1024, .end = 1024 },
3791         { .start = 2048, .end = 2048 },
3792         { .start = 4096, .end = 4096 },
3793 };
3794
3795 /* Ordered by range start value */
3796 static const struct mlxsw_sp_adj_grp_size_range
3797 mlxsw_sp2_adj_grp_size_ranges[] = {
3798         { .start = 1, .end = 128 },
3799         { .start = 256, .end = 256 },
3800         { .start = 512, .end = 512 },
3801         { .start = 1024, .end = 1024 },
3802         { .start = 2048, .end = 2048 },
3803         { .start = 4096, .end = 4096 },
3804 };
3805
3806 static void mlxsw_sp_adj_grp_size_round_up(const struct mlxsw_sp *mlxsw_sp,
3807                                            u16 *p_adj_grp_size)
3808 {
3809         int i;
3810
3811         for (i = 0; i < mlxsw_sp->router->adj_grp_size_ranges_count; i++) {
3812                 const struct mlxsw_sp_adj_grp_size_range *size_range;
3813
3814                 size_range = &mlxsw_sp->router->adj_grp_size_ranges[i];
3815
3816                 if (*p_adj_grp_size >= size_range->start &&
3817                     *p_adj_grp_size <= size_range->end)
3818                         return;
3819
3820                 if (*p_adj_grp_size <= size_range->end) {
3821                         *p_adj_grp_size = size_range->end;
3822                         return;
3823                 }
3824         }
3825 }
3826
3827 static void mlxsw_sp_adj_grp_size_round_down(const struct mlxsw_sp *mlxsw_sp,
3828                                              u16 *p_adj_grp_size,
3829                                              unsigned int alloc_size)
3830 {
3831         int i;
3832
3833         for (i = mlxsw_sp->router->adj_grp_size_ranges_count - 1; i >= 0; i--) {
3834                 const struct mlxsw_sp_adj_grp_size_range *size_range;
3835
3836                 size_range = &mlxsw_sp->router->adj_grp_size_ranges[i];
3837
3838                 if (alloc_size >= size_range->end) {
3839                         *p_adj_grp_size = size_range->end;
3840                         return;
3841                 }
3842         }
3843 }
3844
3845 static int mlxsw_sp_fix_adj_grp_size(struct mlxsw_sp *mlxsw_sp,
3846                                      u16 *p_adj_grp_size)
3847 {
3848         unsigned int alloc_size;
3849         int err;
3850
3851         /* Round up the requested group size to the next size supported
3852          * by the device and make sure the request can be satisfied.
3853          */
3854         mlxsw_sp_adj_grp_size_round_up(mlxsw_sp, p_adj_grp_size);
3855         err = mlxsw_sp_kvdl_alloc_count_query(mlxsw_sp,
3856                                               MLXSW_SP_KVDL_ENTRY_TYPE_ADJ,
3857                                               *p_adj_grp_size, &alloc_size);
3858         if (err)
3859                 return err;
3860         /* It is possible the allocation results in more allocated
3861          * entries than requested. Try to use as much of them as
3862          * possible.
3863          */
3864         mlxsw_sp_adj_grp_size_round_down(mlxsw_sp, p_adj_grp_size, alloc_size);
3865
3866         return 0;
3867 }
3868
3869 static void
3870 mlxsw_sp_nexthop_group_normalize(struct mlxsw_sp_nexthop_group_info *nhgi)
3871 {
3872         int i, g = 0, sum_norm_weight = 0;
3873         struct mlxsw_sp_nexthop *nh;
3874
3875         for (i = 0; i < nhgi->count; i++) {
3876                 nh = &nhgi->nexthops[i];
3877
3878                 if (!nh->should_offload)
3879                         continue;
3880                 if (g > 0)
3881                         g = gcd(nh->nh_weight, g);
3882                 else
3883                         g = nh->nh_weight;
3884         }
3885
3886         for (i = 0; i < nhgi->count; i++) {
3887                 nh = &nhgi->nexthops[i];
3888
3889                 if (!nh->should_offload)
3890                         continue;
3891                 nh->norm_nh_weight = nh->nh_weight / g;
3892                 sum_norm_weight += nh->norm_nh_weight;
3893         }
3894
3895         nhgi->sum_norm_weight = sum_norm_weight;
3896 }
3897
3898 static void
3899 mlxsw_sp_nexthop_group_rebalance(struct mlxsw_sp_nexthop_group_info *nhgi)
3900 {
3901         int i, weight = 0, lower_bound = 0;
3902         int total = nhgi->sum_norm_weight;
3903         u16 ecmp_size = nhgi->ecmp_size;
3904
3905         for (i = 0; i < nhgi->count; i++) {
3906                 struct mlxsw_sp_nexthop *nh = &nhgi->nexthops[i];
3907                 int upper_bound;
3908
3909                 if (!nh->should_offload)
3910                         continue;
3911                 weight += nh->norm_nh_weight;
3912                 upper_bound = DIV_ROUND_CLOSEST(ecmp_size * weight, total);
3913                 nh->num_adj_entries = upper_bound - lower_bound;
3914                 lower_bound = upper_bound;
3915         }
3916 }
3917
3918 static struct mlxsw_sp_nexthop *
3919 mlxsw_sp_rt6_nexthop(struct mlxsw_sp_nexthop_group *nh_grp,
3920                      const struct mlxsw_sp_rt6 *mlxsw_sp_rt6);
3921
3922 static void
3923 mlxsw_sp_nexthop4_group_offload_refresh(struct mlxsw_sp *mlxsw_sp,
3924                                         struct mlxsw_sp_nexthop_group *nh_grp)
3925 {
3926         int i;
3927
3928         for (i = 0; i < nh_grp->nhgi->count; i++) {
3929                 struct mlxsw_sp_nexthop *nh = &nh_grp->nhgi->nexthops[i];
3930
3931                 if (nh->offloaded)
3932                         nh->key.fib_nh->fib_nh_flags |= RTNH_F_OFFLOAD;
3933                 else
3934                         nh->key.fib_nh->fib_nh_flags &= ~RTNH_F_OFFLOAD;
3935         }
3936 }
3937
3938 static void
3939 __mlxsw_sp_nexthop6_group_offload_refresh(struct mlxsw_sp_nexthop_group *nh_grp,
3940                                           struct mlxsw_sp_fib6_entry *fib6_entry)
3941 {
3942         struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
3943
3944         list_for_each_entry(mlxsw_sp_rt6, &fib6_entry->rt6_list, list) {
3945                 struct fib6_nh *fib6_nh = mlxsw_sp_rt6->rt->fib6_nh;
3946                 struct mlxsw_sp_nexthop *nh;
3947
3948                 nh = mlxsw_sp_rt6_nexthop(nh_grp, mlxsw_sp_rt6);
3949                 if (nh && nh->offloaded)
3950                         fib6_nh->fib_nh_flags |= RTNH_F_OFFLOAD;
3951                 else
3952                         fib6_nh->fib_nh_flags &= ~RTNH_F_OFFLOAD;
3953         }
3954 }
3955
3956 static void
3957 mlxsw_sp_nexthop6_group_offload_refresh(struct mlxsw_sp *mlxsw_sp,
3958                                         struct mlxsw_sp_nexthop_group *nh_grp)
3959 {
3960         struct mlxsw_sp_fib6_entry *fib6_entry;
3961
3962         /* Unfortunately, in IPv6 the route and the nexthop are described by
3963          * the same struct, so we need to iterate over all the routes using the
3964          * nexthop group and set / clear the offload indication for them.
3965          */
3966         list_for_each_entry(fib6_entry, &nh_grp->fib_list,
3967                             common.nexthop_group_node)
3968                 __mlxsw_sp_nexthop6_group_offload_refresh(nh_grp, fib6_entry);
3969 }
3970
3971 static void
3972 mlxsw_sp_nexthop_bucket_offload_refresh(struct mlxsw_sp *mlxsw_sp,
3973                                         const struct mlxsw_sp_nexthop *nh,
3974                                         u16 bucket_index)
3975 {
3976         struct mlxsw_sp_nexthop_group *nh_grp = nh->nhgi->nh_grp;
3977         bool offload = false, trap = false;
3978
3979         if (nh->offloaded) {
3980                 if (nh->action == MLXSW_SP_NEXTHOP_ACTION_TRAP)
3981                         trap = true;
3982                 else
3983                         offload = true;
3984         }
3985         nexthop_bucket_set_hw_flags(mlxsw_sp_net(mlxsw_sp), nh_grp->obj.id,
3986                                     bucket_index, offload, trap);
3987 }
3988
3989 static void
3990 mlxsw_sp_nexthop_obj_group_offload_refresh(struct mlxsw_sp *mlxsw_sp,
3991                                            struct mlxsw_sp_nexthop_group *nh_grp)
3992 {
3993         int i;
3994
3995         /* Do not update the flags if the nexthop group is being destroyed
3996          * since:
3997          * 1. The nexthop objects is being deleted, in which case the flags are
3998          * irrelevant.
3999          * 2. The nexthop group was replaced by a newer group, in which case
4000          * the flags of the nexthop object were already updated based on the
4001          * new group.
4002          */
4003         if (nh_grp->can_destroy)
4004                 return;
4005
4006         nexthop_set_hw_flags(mlxsw_sp_net(mlxsw_sp), nh_grp->obj.id,
4007                              nh_grp->nhgi->adj_index_valid, false);
4008
4009         /* Update flags of individual nexthop buckets in case of a resilient
4010          * nexthop group.
4011          */
4012         if (!nh_grp->nhgi->is_resilient)
4013                 return;
4014
4015         for (i = 0; i < nh_grp->nhgi->count; i++) {
4016                 struct mlxsw_sp_nexthop *nh = &nh_grp->nhgi->nexthops[i];
4017
4018                 mlxsw_sp_nexthop_bucket_offload_refresh(mlxsw_sp, nh, i);
4019         }
4020 }
4021
4022 static void
4023 mlxsw_sp_nexthop_group_offload_refresh(struct mlxsw_sp *mlxsw_sp,
4024                                        struct mlxsw_sp_nexthop_group *nh_grp)
4025 {
4026         switch (nh_grp->type) {
4027         case MLXSW_SP_NEXTHOP_GROUP_TYPE_IPV4:
4028                 mlxsw_sp_nexthop4_group_offload_refresh(mlxsw_sp, nh_grp);
4029                 break;
4030         case MLXSW_SP_NEXTHOP_GROUP_TYPE_IPV6:
4031                 mlxsw_sp_nexthop6_group_offload_refresh(mlxsw_sp, nh_grp);
4032                 break;
4033         case MLXSW_SP_NEXTHOP_GROUP_TYPE_OBJ:
4034                 mlxsw_sp_nexthop_obj_group_offload_refresh(mlxsw_sp, nh_grp);
4035                 break;
4036         }
4037 }
4038
4039 static int
4040 mlxsw_sp_nexthop_group_refresh(struct mlxsw_sp *mlxsw_sp,
4041                                struct mlxsw_sp_nexthop_group *nh_grp)
4042 {
4043         struct mlxsw_sp_nexthop_group_info *nhgi = nh_grp->nhgi;
4044         u16 ecmp_size, old_ecmp_size;
4045         struct mlxsw_sp_nexthop *nh;
4046         bool offload_change = false;
4047         u32 adj_index;
4048         bool old_adj_index_valid;
4049         u32 old_adj_index;
4050         int i, err2, err;
4051
4052         if (!nhgi->gateway)
4053                 return mlxsw_sp_nexthop_fib_entries_update(mlxsw_sp, nh_grp);
4054
4055         for (i = 0; i < nhgi->count; i++) {
4056                 nh = &nhgi->nexthops[i];
4057
4058                 if (nh->should_offload != nh->offloaded) {
4059                         offload_change = true;
4060                         if (nh->should_offload)
4061                                 nh->update = 1;
4062                 }
4063         }
4064         if (!offload_change) {
4065                 /* Nothing was added or removed, so no need to reallocate. Just
4066                  * update MAC on existing adjacency indexes.
4067                  */
4068                 err = mlxsw_sp_nexthop_group_update(mlxsw_sp, nhgi, false);
4069                 if (err) {
4070                         dev_warn(mlxsw_sp->bus_info->dev, "Failed to update neigh MAC in adjacency table.\n");
4071                         goto set_trap;
4072                 }
4073                 /* Flags of individual nexthop buckets might need to be
4074                  * updated.
4075                  */
4076                 mlxsw_sp_nexthop_group_offload_refresh(mlxsw_sp, nh_grp);
4077                 return 0;
4078         }
4079         mlxsw_sp_nexthop_group_normalize(nhgi);
4080         if (!nhgi->sum_norm_weight) {
4081                 /* No neigh of this group is connected so we just set
4082                  * the trap and let everthing flow through kernel.
4083                  */
4084                 err = 0;
4085                 goto set_trap;
4086         }
4087
4088         ecmp_size = nhgi->sum_norm_weight;
4089         err = mlxsw_sp_fix_adj_grp_size(mlxsw_sp, &ecmp_size);
4090         if (err)
4091                 /* No valid allocation size available. */
4092                 goto set_trap;
4093
4094         err = mlxsw_sp_kvdl_alloc(mlxsw_sp, MLXSW_SP_KVDL_ENTRY_TYPE_ADJ,
4095                                   ecmp_size, &adj_index);
4096         if (err) {
4097                 /* We ran out of KVD linear space, just set the
4098                  * trap and let everything flow through kernel.
4099                  */
4100                 dev_warn(mlxsw_sp->bus_info->dev, "Failed to allocate KVD linear area for nexthop group.\n");
4101                 goto set_trap;
4102         }
4103         old_adj_index_valid = nhgi->adj_index_valid;
4104         old_adj_index = nhgi->adj_index;
4105         old_ecmp_size = nhgi->ecmp_size;
4106         nhgi->adj_index_valid = 1;
4107         nhgi->adj_index = adj_index;
4108         nhgi->ecmp_size = ecmp_size;
4109         mlxsw_sp_nexthop_group_rebalance(nhgi);
4110         err = mlxsw_sp_nexthop_group_update(mlxsw_sp, nhgi, true);
4111         if (err) {
4112                 dev_warn(mlxsw_sp->bus_info->dev, "Failed to update neigh MAC in adjacency table.\n");
4113                 goto set_trap;
4114         }
4115
4116         mlxsw_sp_nexthop_group_offload_refresh(mlxsw_sp, nh_grp);
4117
4118         if (!old_adj_index_valid) {
4119                 /* The trap was set for fib entries, so we have to call
4120                  * fib entry update to unset it and use adjacency index.
4121                  */
4122                 err = mlxsw_sp_nexthop_fib_entries_update(mlxsw_sp, nh_grp);
4123                 if (err) {
4124                         dev_warn(mlxsw_sp->bus_info->dev, "Failed to add adjacency index to fib entries.\n");
4125                         goto set_trap;
4126                 }
4127                 return 0;
4128         }
4129
4130         err = mlxsw_sp_adj_index_mass_update(mlxsw_sp, nh_grp,
4131                                              old_adj_index, old_ecmp_size);
4132         mlxsw_sp_kvdl_free(mlxsw_sp, MLXSW_SP_KVDL_ENTRY_TYPE_ADJ,
4133                            old_ecmp_size, old_adj_index);
4134         if (err) {
4135                 dev_warn(mlxsw_sp->bus_info->dev, "Failed to mass-update adjacency index for nexthop group.\n");
4136                 goto set_trap;
4137         }
4138
4139         return 0;
4140
4141 set_trap:
4142         old_adj_index_valid = nhgi->adj_index_valid;
4143         nhgi->adj_index_valid = 0;
4144         for (i = 0; i < nhgi->count; i++) {
4145                 nh = &nhgi->nexthops[i];
4146                 nh->offloaded = 0;
4147         }
4148         err2 = mlxsw_sp_nexthop_fib_entries_update(mlxsw_sp, nh_grp);
4149         if (err2)
4150                 dev_warn(mlxsw_sp->bus_info->dev, "Failed to set traps for fib entries.\n");
4151         mlxsw_sp_nexthop_group_offload_refresh(mlxsw_sp, nh_grp);
4152         if (old_adj_index_valid)
4153                 mlxsw_sp_kvdl_free(mlxsw_sp, MLXSW_SP_KVDL_ENTRY_TYPE_ADJ,
4154                                    nhgi->ecmp_size, nhgi->adj_index);
4155         return err;
4156 }
4157
4158 static void __mlxsw_sp_nexthop_neigh_update(struct mlxsw_sp_nexthop *nh,
4159                                             bool removing)
4160 {
4161         if (!removing) {
4162                 nh->action = MLXSW_SP_NEXTHOP_ACTION_FORWARD;
4163                 nh->should_offload = 1;
4164         } else if (nh->nhgi->is_resilient) {
4165                 nh->action = MLXSW_SP_NEXTHOP_ACTION_TRAP;
4166                 nh->should_offload = 1;
4167         } else {
4168                 nh->should_offload = 0;
4169         }
4170         nh->update = 1;
4171 }
4172
4173 static int
4174 mlxsw_sp_nexthop_dead_neigh_replace(struct mlxsw_sp *mlxsw_sp,
4175                                     struct mlxsw_sp_neigh_entry *neigh_entry)
4176 {
4177         struct neighbour *n, *old_n = neigh_entry->key.n;
4178         struct mlxsw_sp_nexthop *nh;
4179         struct net_device *dev;
4180         bool entry_connected;
4181         u8 nud_state, dead;
4182         int err;
4183
4184         nh = list_first_entry(&neigh_entry->nexthop_list,
4185                               struct mlxsw_sp_nexthop, neigh_list_node);
4186         dev = mlxsw_sp_nexthop_dev(nh);
4187
4188         n = neigh_lookup(nh->neigh_tbl, &nh->gw_addr, dev);
4189         if (!n) {
4190                 n = neigh_create(nh->neigh_tbl, &nh->gw_addr, dev);
4191                 if (IS_ERR(n))
4192                         return PTR_ERR(n);
4193                 neigh_event_send(n, NULL);
4194         }
4195
4196         mlxsw_sp_neigh_entry_remove(mlxsw_sp, neigh_entry);
4197         neigh_entry->key.n = n;
4198         err = mlxsw_sp_neigh_entry_insert(mlxsw_sp, neigh_entry);
4199         if (err)
4200                 goto err_neigh_entry_insert;
4201
4202         read_lock_bh(&n->lock);
4203         nud_state = n->nud_state;
4204         dead = n->dead;
4205         read_unlock_bh(&n->lock);
4206         entry_connected = nud_state & NUD_VALID && !dead;
4207
4208         list_for_each_entry(nh, &neigh_entry->nexthop_list,
4209                             neigh_list_node) {
4210                 neigh_release(old_n);
4211                 neigh_clone(n);
4212                 __mlxsw_sp_nexthop_neigh_update(nh, !entry_connected);
4213                 mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh->nhgi->nh_grp);
4214         }
4215
4216         neigh_release(n);
4217
4218         return 0;
4219
4220 err_neigh_entry_insert:
4221         neigh_entry->key.n = old_n;
4222         mlxsw_sp_neigh_entry_insert(mlxsw_sp, neigh_entry);
4223         neigh_release(n);
4224         return err;
4225 }
4226
4227 static void
4228 mlxsw_sp_nexthop_neigh_update(struct mlxsw_sp *mlxsw_sp,
4229                               struct mlxsw_sp_neigh_entry *neigh_entry,
4230                               bool removing, bool dead)
4231 {
4232         struct mlxsw_sp_nexthop *nh;
4233
4234         if (list_empty(&neigh_entry->nexthop_list))
4235                 return;
4236
4237         if (dead) {
4238                 int err;
4239
4240                 err = mlxsw_sp_nexthop_dead_neigh_replace(mlxsw_sp,
4241                                                           neigh_entry);
4242                 if (err)
4243                         dev_err(mlxsw_sp->bus_info->dev, "Failed to replace dead neigh\n");
4244                 return;
4245         }
4246
4247         list_for_each_entry(nh, &neigh_entry->nexthop_list,
4248                             neigh_list_node) {
4249                 __mlxsw_sp_nexthop_neigh_update(nh, removing);
4250                 mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh->nhgi->nh_grp);
4251         }
4252 }
4253
4254 static void mlxsw_sp_nexthop_crif_init(struct mlxsw_sp_nexthop *nh,
4255                                        struct mlxsw_sp_crif *crif)
4256 {
4257         if (nh->crif)
4258                 return;
4259
4260         nh->crif = crif;
4261         list_add(&nh->crif_list_node, &crif->nexthop_list);
4262 }
4263
4264 static void mlxsw_sp_nexthop_crif_fini(struct mlxsw_sp_nexthop *nh)
4265 {
4266         if (!nh->crif)
4267                 return;
4268
4269         list_del(&nh->crif_list_node);
4270         nh->crif = NULL;
4271 }
4272
4273 static int mlxsw_sp_nexthop_neigh_init(struct mlxsw_sp *mlxsw_sp,
4274                                        struct mlxsw_sp_nexthop *nh)
4275 {
4276         struct mlxsw_sp_neigh_entry *neigh_entry;
4277         struct net_device *dev;
4278         struct neighbour *n;
4279         u8 nud_state, dead;
4280         int err;
4281
4282         if (WARN_ON(!nh->crif->rif))
4283                 return 0;
4284
4285         if (!nh->nhgi->gateway || nh->neigh_entry)
4286                 return 0;
4287         dev = mlxsw_sp_nexthop_dev(nh);
4288
4289         /* Take a reference of neigh here ensuring that neigh would
4290          * not be destructed before the nexthop entry is finished.
4291          * The reference is taken either in neigh_lookup() or
4292          * in neigh_create() in case n is not found.
4293          */
4294         n = neigh_lookup(nh->neigh_tbl, &nh->gw_addr, dev);
4295         if (!n) {
4296                 n = neigh_create(nh->neigh_tbl, &nh->gw_addr, dev);
4297                 if (IS_ERR(n))
4298                         return PTR_ERR(n);
4299                 neigh_event_send(n, NULL);
4300         }
4301         neigh_entry = mlxsw_sp_neigh_entry_lookup(mlxsw_sp, n);
4302         if (!neigh_entry) {
4303                 neigh_entry = mlxsw_sp_neigh_entry_create(mlxsw_sp, n);
4304                 if (IS_ERR(neigh_entry)) {
4305                         err = -EINVAL;
4306                         goto err_neigh_entry_create;
4307                 }
4308         }
4309
4310         /* If that is the first nexthop connected to that neigh, add to
4311          * nexthop_neighs_list
4312          */
4313         if (list_empty(&neigh_entry->nexthop_list))
4314                 list_add_tail(&neigh_entry->nexthop_neighs_list_node,
4315                               &mlxsw_sp->router->nexthop_neighs_list);
4316
4317         nh->neigh_entry = neigh_entry;
4318         list_add_tail(&nh->neigh_list_node, &neigh_entry->nexthop_list);
4319         read_lock_bh(&n->lock);
4320         nud_state = n->nud_state;
4321         dead = n->dead;
4322         read_unlock_bh(&n->lock);
4323         __mlxsw_sp_nexthop_neigh_update(nh, !(nud_state & NUD_VALID && !dead));
4324
4325         return 0;
4326
4327 err_neigh_entry_create:
4328         neigh_release(n);
4329         return err;
4330 }
4331
4332 static void mlxsw_sp_nexthop_neigh_fini(struct mlxsw_sp *mlxsw_sp,
4333                                         struct mlxsw_sp_nexthop *nh)
4334 {
4335         struct mlxsw_sp_neigh_entry *neigh_entry = nh->neigh_entry;
4336         struct neighbour *n;
4337
4338         if (!neigh_entry)
4339                 return;
4340         n = neigh_entry->key.n;
4341
4342         __mlxsw_sp_nexthop_neigh_update(nh, true);
4343         list_del(&nh->neigh_list_node);
4344         nh->neigh_entry = NULL;
4345
4346         /* If that is the last nexthop connected to that neigh, remove from
4347          * nexthop_neighs_list
4348          */
4349         if (list_empty(&neigh_entry->nexthop_list))
4350                 list_del(&neigh_entry->nexthop_neighs_list_node);
4351
4352         if (!neigh_entry->connected && list_empty(&neigh_entry->nexthop_list))
4353                 mlxsw_sp_neigh_entry_destroy(mlxsw_sp, neigh_entry);
4354
4355         neigh_release(n);
4356 }
4357
4358 static bool mlxsw_sp_ipip_netdev_ul_up(struct net_device *ol_dev)
4359 {
4360         struct net_device *ul_dev;
4361         bool is_up;
4362
4363         rcu_read_lock();
4364         ul_dev = mlxsw_sp_ipip_netdev_ul_dev_get(ol_dev);
4365         is_up = ul_dev ? (ul_dev->flags & IFF_UP) : true;
4366         rcu_read_unlock();
4367
4368         return is_up;
4369 }
4370
4371 static void mlxsw_sp_nexthop_ipip_init(struct mlxsw_sp *mlxsw_sp,
4372                                        struct mlxsw_sp_nexthop *nh,
4373                                        struct mlxsw_sp_ipip_entry *ipip_entry)
4374 {
4375         struct mlxsw_sp_crif *crif;
4376         bool removing;
4377
4378         if (!nh->nhgi->gateway || nh->ipip_entry)
4379                 return;
4380
4381         crif = mlxsw_sp_crif_lookup(mlxsw_sp->router, ipip_entry->ol_dev);
4382         if (WARN_ON(!crif))
4383                 return;
4384
4385         nh->ipip_entry = ipip_entry;
4386         removing = !mlxsw_sp_ipip_netdev_ul_up(ipip_entry->ol_dev);
4387         __mlxsw_sp_nexthop_neigh_update(nh, removing);
4388         mlxsw_sp_nexthop_crif_init(nh, crif);
4389 }
4390
4391 static void mlxsw_sp_nexthop_ipip_fini(struct mlxsw_sp *mlxsw_sp,
4392                                        struct mlxsw_sp_nexthop *nh)
4393 {
4394         struct mlxsw_sp_ipip_entry *ipip_entry = nh->ipip_entry;
4395
4396         if (!ipip_entry)
4397                 return;
4398
4399         __mlxsw_sp_nexthop_neigh_update(nh, true);
4400         nh->ipip_entry = NULL;
4401 }
4402
4403 static bool mlxsw_sp_nexthop4_ipip_type(const struct mlxsw_sp *mlxsw_sp,
4404                                         const struct fib_nh *fib_nh,
4405                                         enum mlxsw_sp_ipip_type *p_ipipt)
4406 {
4407         struct net_device *dev = fib_nh->fib_nh_dev;
4408
4409         return dev &&
4410                fib_nh->nh_parent->fib_type == RTN_UNICAST &&
4411                mlxsw_sp_netdev_ipip_type(mlxsw_sp, dev, p_ipipt);
4412 }
4413
4414 static int mlxsw_sp_nexthop_type_init(struct mlxsw_sp *mlxsw_sp,
4415                                       struct mlxsw_sp_nexthop *nh,
4416                                       const struct net_device *dev)
4417 {
4418         const struct mlxsw_sp_ipip_ops *ipip_ops;
4419         struct mlxsw_sp_ipip_entry *ipip_entry;
4420         struct mlxsw_sp_crif *crif;
4421         int err;
4422
4423         ipip_entry = mlxsw_sp_ipip_entry_find_by_ol_dev(mlxsw_sp, dev);
4424         if (ipip_entry) {
4425                 ipip_ops = mlxsw_sp->router->ipip_ops_arr[ipip_entry->ipipt];
4426                 if (ipip_ops->can_offload(mlxsw_sp, dev)) {
4427                         nh->type = MLXSW_SP_NEXTHOP_TYPE_IPIP;
4428                         mlxsw_sp_nexthop_ipip_init(mlxsw_sp, nh, ipip_entry);
4429                         return 0;
4430                 }
4431         }
4432
4433         nh->type = MLXSW_SP_NEXTHOP_TYPE_ETH;
4434         crif = mlxsw_sp_crif_lookup(mlxsw_sp->router, dev);
4435         if (!crif)
4436                 return 0;
4437
4438         mlxsw_sp_nexthop_crif_init(nh, crif);
4439
4440         if (!crif->rif)
4441                 return 0;
4442
4443         err = mlxsw_sp_nexthop_neigh_init(mlxsw_sp, nh);
4444         if (err)
4445                 goto err_neigh_init;
4446
4447         return 0;
4448
4449 err_neigh_init:
4450         mlxsw_sp_nexthop_crif_fini(nh);
4451         return err;
4452 }
4453
4454 static int mlxsw_sp_nexthop_type_rif_made(struct mlxsw_sp *mlxsw_sp,
4455                                           struct mlxsw_sp_nexthop *nh)
4456 {
4457         switch (nh->type) {
4458         case MLXSW_SP_NEXTHOP_TYPE_ETH:
4459                 return mlxsw_sp_nexthop_neigh_init(mlxsw_sp, nh);
4460         case MLXSW_SP_NEXTHOP_TYPE_IPIP:
4461                 break;
4462         }
4463
4464         return 0;
4465 }
4466
4467 static void mlxsw_sp_nexthop_type_rif_gone(struct mlxsw_sp *mlxsw_sp,
4468                                            struct mlxsw_sp_nexthop *nh)
4469 {
4470         switch (nh->type) {
4471         case MLXSW_SP_NEXTHOP_TYPE_ETH:
4472                 mlxsw_sp_nexthop_neigh_fini(mlxsw_sp, nh);
4473                 break;
4474         case MLXSW_SP_NEXTHOP_TYPE_IPIP:
4475                 mlxsw_sp_nexthop_ipip_fini(mlxsw_sp, nh);
4476                 break;
4477         }
4478 }
4479
4480 static void mlxsw_sp_nexthop_type_fini(struct mlxsw_sp *mlxsw_sp,
4481                                        struct mlxsw_sp_nexthop *nh)
4482 {
4483         mlxsw_sp_nexthop_type_rif_gone(mlxsw_sp, nh);
4484         mlxsw_sp_nexthop_crif_fini(nh);
4485 }
4486
4487 static int mlxsw_sp_nexthop4_init(struct mlxsw_sp *mlxsw_sp,
4488                                   struct mlxsw_sp_nexthop_group *nh_grp,
4489                                   struct mlxsw_sp_nexthop *nh,
4490                                   struct fib_nh *fib_nh)
4491 {
4492         struct net_device *dev = fib_nh->fib_nh_dev;
4493         struct in_device *in_dev;
4494         int err;
4495
4496         nh->nhgi = nh_grp->nhgi;
4497         nh->key.fib_nh = fib_nh;
4498 #ifdef CONFIG_IP_ROUTE_MULTIPATH
4499         nh->nh_weight = fib_nh->fib_nh_weight;
4500 #else
4501         nh->nh_weight = 1;
4502 #endif
4503         memcpy(&nh->gw_addr, &fib_nh->fib_nh_gw4, sizeof(fib_nh->fib_nh_gw4));
4504         nh->neigh_tbl = &arp_tbl;
4505         err = mlxsw_sp_nexthop_insert(mlxsw_sp, nh);
4506         if (err)
4507                 return err;
4508
4509         mlxsw_sp_nexthop_counter_alloc(mlxsw_sp, nh);
4510         list_add_tail(&nh->router_list_node, &mlxsw_sp->router->nexthop_list);
4511
4512         if (!dev)
4513                 return 0;
4514         nh->ifindex = dev->ifindex;
4515
4516         rcu_read_lock();
4517         in_dev = __in_dev_get_rcu(dev);
4518         if (in_dev && IN_DEV_IGNORE_ROUTES_WITH_LINKDOWN(in_dev) &&
4519             fib_nh->fib_nh_flags & RTNH_F_LINKDOWN) {
4520                 rcu_read_unlock();
4521                 return 0;
4522         }
4523         rcu_read_unlock();
4524
4525         err = mlxsw_sp_nexthop_type_init(mlxsw_sp, nh, dev);
4526         if (err)
4527                 goto err_nexthop_neigh_init;
4528
4529         return 0;
4530
4531 err_nexthop_neigh_init:
4532         list_del(&nh->router_list_node);
4533         mlxsw_sp_nexthop_counter_free(mlxsw_sp, nh);
4534         mlxsw_sp_nexthop_remove(mlxsw_sp, nh);
4535         return err;
4536 }
4537
4538 static void mlxsw_sp_nexthop4_fini(struct mlxsw_sp *mlxsw_sp,
4539                                    struct mlxsw_sp_nexthop *nh)
4540 {
4541         mlxsw_sp_nexthop_type_fini(mlxsw_sp, nh);
4542         list_del(&nh->router_list_node);
4543         mlxsw_sp_nexthop_counter_free(mlxsw_sp, nh);
4544         mlxsw_sp_nexthop_remove(mlxsw_sp, nh);
4545 }
4546
4547 static void mlxsw_sp_nexthop4_event(struct mlxsw_sp *mlxsw_sp,
4548                                     unsigned long event, struct fib_nh *fib_nh)
4549 {
4550         struct mlxsw_sp_nexthop_key key;
4551         struct mlxsw_sp_nexthop *nh;
4552
4553         key.fib_nh = fib_nh;
4554         nh = mlxsw_sp_nexthop_lookup(mlxsw_sp, key);
4555         if (!nh)
4556                 return;
4557
4558         switch (event) {
4559         case FIB_EVENT_NH_ADD:
4560                 mlxsw_sp_nexthop_type_init(mlxsw_sp, nh, fib_nh->fib_nh_dev);
4561                 break;
4562         case FIB_EVENT_NH_DEL:
4563                 mlxsw_sp_nexthop_type_fini(mlxsw_sp, nh);
4564                 break;
4565         }
4566
4567         mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh->nhgi->nh_grp);
4568 }
4569
4570 static void mlxsw_sp_nexthop_rif_update(struct mlxsw_sp *mlxsw_sp,
4571                                         struct mlxsw_sp_rif *rif)
4572 {
4573         struct net_device *dev = mlxsw_sp_rif_dev(rif);
4574         struct mlxsw_sp_nexthop *nh;
4575         bool removing;
4576
4577         list_for_each_entry(nh, &rif->crif->nexthop_list, crif_list_node) {
4578                 switch (nh->type) {
4579                 case MLXSW_SP_NEXTHOP_TYPE_ETH:
4580                         removing = false;
4581                         break;
4582                 case MLXSW_SP_NEXTHOP_TYPE_IPIP:
4583                         removing = !mlxsw_sp_ipip_netdev_ul_up(dev);
4584                         break;
4585                 default:
4586                         WARN_ON(1);
4587                         continue;
4588                 }
4589
4590                 __mlxsw_sp_nexthop_neigh_update(nh, removing);
4591                 mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh->nhgi->nh_grp);
4592         }
4593 }
4594
4595 static int mlxsw_sp_nexthop_rif_made_sync(struct mlxsw_sp *mlxsw_sp,
4596                                           struct mlxsw_sp_rif *rif)
4597 {
4598         struct mlxsw_sp_nexthop *nh, *tmp;
4599         unsigned int n = 0;
4600         int err;
4601
4602         list_for_each_entry_safe(nh, tmp, &rif->crif->nexthop_list,
4603                                  crif_list_node) {
4604                 err = mlxsw_sp_nexthop_type_rif_made(mlxsw_sp, nh);
4605                 if (err)
4606                         goto err_nexthop_type_rif;
4607                 mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh->nhgi->nh_grp);
4608                 n++;
4609         }
4610
4611         return 0;
4612
4613 err_nexthop_type_rif:
4614         list_for_each_entry_safe(nh, tmp, &rif->crif->nexthop_list,
4615                                  crif_list_node) {
4616                 if (!n--)
4617                         break;
4618                 mlxsw_sp_nexthop_type_rif_gone(mlxsw_sp, nh);
4619                 mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh->nhgi->nh_grp);
4620         }
4621         return err;
4622 }
4623
4624 static void mlxsw_sp_nexthop_rif_gone_sync(struct mlxsw_sp *mlxsw_sp,
4625                                            struct mlxsw_sp_rif *rif)
4626 {
4627         struct mlxsw_sp_nexthop *nh, *tmp;
4628
4629         list_for_each_entry_safe(nh, tmp, &rif->crif->nexthop_list,
4630                                  crif_list_node) {
4631                 mlxsw_sp_nexthop_type_rif_gone(mlxsw_sp, nh);
4632                 mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh->nhgi->nh_grp);
4633         }
4634 }
4635
4636 static int mlxsw_sp_adj_trap_entry_init(struct mlxsw_sp *mlxsw_sp)
4637 {
4638         enum mlxsw_reg_ratr_trap_action trap_action;
4639         char ratr_pl[MLXSW_REG_RATR_LEN];
4640         int err;
4641
4642         err = mlxsw_sp_kvdl_alloc(mlxsw_sp, MLXSW_SP_KVDL_ENTRY_TYPE_ADJ, 1,
4643                                   &mlxsw_sp->router->adj_trap_index);
4644         if (err)
4645                 return err;
4646
4647         trap_action = MLXSW_REG_RATR_TRAP_ACTION_TRAP;
4648         mlxsw_reg_ratr_pack(ratr_pl, MLXSW_REG_RATR_OP_WRITE_WRITE_ENTRY, true,
4649                             MLXSW_REG_RATR_TYPE_ETHERNET,
4650                             mlxsw_sp->router->adj_trap_index,
4651                             mlxsw_sp->router->lb_crif->rif->rif_index);
4652         mlxsw_reg_ratr_trap_action_set(ratr_pl, trap_action);
4653         mlxsw_reg_ratr_trap_id_set(ratr_pl, MLXSW_TRAP_ID_RTR_EGRESS0);
4654         err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ratr), ratr_pl);
4655         if (err)
4656                 goto err_ratr_write;
4657
4658         return 0;
4659
4660 err_ratr_write:
4661         mlxsw_sp_kvdl_free(mlxsw_sp, MLXSW_SP_KVDL_ENTRY_TYPE_ADJ, 1,
4662                            mlxsw_sp->router->adj_trap_index);
4663         return err;
4664 }
4665
4666 static void mlxsw_sp_adj_trap_entry_fini(struct mlxsw_sp *mlxsw_sp)
4667 {
4668         mlxsw_sp_kvdl_free(mlxsw_sp, MLXSW_SP_KVDL_ENTRY_TYPE_ADJ, 1,
4669                            mlxsw_sp->router->adj_trap_index);
4670 }
4671
4672 static int mlxsw_sp_nexthop_group_inc(struct mlxsw_sp *mlxsw_sp)
4673 {
4674         int err;
4675
4676         if (refcount_inc_not_zero(&mlxsw_sp->router->num_groups))
4677                 return 0;
4678
4679         err = mlxsw_sp_adj_trap_entry_init(mlxsw_sp);
4680         if (err)
4681                 return err;
4682
4683         refcount_set(&mlxsw_sp->router->num_groups, 1);
4684
4685         return 0;
4686 }
4687
4688 static void mlxsw_sp_nexthop_group_dec(struct mlxsw_sp *mlxsw_sp)
4689 {
4690         if (!refcount_dec_and_test(&mlxsw_sp->router->num_groups))
4691                 return;
4692
4693         mlxsw_sp_adj_trap_entry_fini(mlxsw_sp);
4694 }
4695
4696 static void
4697 mlxsw_sp_nh_grp_activity_get(struct mlxsw_sp *mlxsw_sp,
4698                              const struct mlxsw_sp_nexthop_group *nh_grp,
4699                              unsigned long *activity)
4700 {
4701         char *ratrad_pl;
4702         int i, err;
4703
4704         ratrad_pl = kmalloc(MLXSW_REG_RATRAD_LEN, GFP_KERNEL);
4705         if (!ratrad_pl)
4706                 return;
4707
4708         mlxsw_reg_ratrad_pack(ratrad_pl, nh_grp->nhgi->adj_index,
4709                               nh_grp->nhgi->count);
4710         err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(ratrad), ratrad_pl);
4711         if (err)
4712                 goto out;
4713
4714         for (i = 0; i < nh_grp->nhgi->count; i++) {
4715                 if (!mlxsw_reg_ratrad_activity_vector_get(ratrad_pl, i))
4716                         continue;
4717                 bitmap_set(activity, i, 1);
4718         }
4719
4720 out:
4721         kfree(ratrad_pl);
4722 }
4723
4724 #define MLXSW_SP_NH_GRP_ACTIVITY_UPDATE_INTERVAL 1000 /* ms */
4725
4726 static void
4727 mlxsw_sp_nh_grp_activity_update(struct mlxsw_sp *mlxsw_sp,
4728                                 const struct mlxsw_sp_nexthop_group *nh_grp)
4729 {
4730         unsigned long *activity;
4731
4732         activity = bitmap_zalloc(nh_grp->nhgi->count, GFP_KERNEL);
4733         if (!activity)
4734                 return;
4735
4736         mlxsw_sp_nh_grp_activity_get(mlxsw_sp, nh_grp, activity);
4737         nexthop_res_grp_activity_update(mlxsw_sp_net(mlxsw_sp), nh_grp->obj.id,
4738                                         nh_grp->nhgi->count, activity);
4739
4740         bitmap_free(activity);
4741 }
4742
4743 static void
4744 mlxsw_sp_nh_grp_activity_work_schedule(struct mlxsw_sp *mlxsw_sp)
4745 {
4746         unsigned int interval = MLXSW_SP_NH_GRP_ACTIVITY_UPDATE_INTERVAL;
4747
4748         mlxsw_core_schedule_dw(&mlxsw_sp->router->nh_grp_activity_dw,
4749                                msecs_to_jiffies(interval));
4750 }
4751
4752 static void mlxsw_sp_nh_grp_activity_work(struct work_struct *work)
4753 {
4754         struct mlxsw_sp_nexthop_group_info *nhgi;
4755         struct mlxsw_sp_router *router;
4756         bool reschedule = false;
4757
4758         router = container_of(work, struct mlxsw_sp_router,
4759                               nh_grp_activity_dw.work);
4760
4761         mutex_lock(&router->lock);
4762
4763         list_for_each_entry(nhgi, &router->nh_res_grp_list, list) {
4764                 mlxsw_sp_nh_grp_activity_update(router->mlxsw_sp, nhgi->nh_grp);
4765                 reschedule = true;
4766         }
4767
4768         mutex_unlock(&router->lock);
4769
4770         if (!reschedule)
4771                 return;
4772         mlxsw_sp_nh_grp_activity_work_schedule(router->mlxsw_sp);
4773 }
4774
4775 static int
4776 mlxsw_sp_nexthop_obj_single_validate(struct mlxsw_sp *mlxsw_sp,
4777                                      const struct nh_notifier_single_info *nh,
4778                                      struct netlink_ext_ack *extack)
4779 {
4780         int err = -EINVAL;
4781
4782         if (nh->is_fdb)
4783                 NL_SET_ERR_MSG_MOD(extack, "FDB nexthops are not supported");
4784         else if (nh->has_encap)
4785                 NL_SET_ERR_MSG_MOD(extack, "Encapsulating nexthops are not supported");
4786         else
4787                 err = 0;
4788
4789         return err;
4790 }
4791
4792 static int
4793 mlxsw_sp_nexthop_obj_group_entry_validate(struct mlxsw_sp *mlxsw_sp,
4794                                           const struct nh_notifier_single_info *nh,
4795                                           struct netlink_ext_ack *extack)
4796 {
4797         int err;
4798
4799         err = mlxsw_sp_nexthop_obj_single_validate(mlxsw_sp, nh, extack);
4800         if (err)
4801                 return err;
4802
4803         /* Device only nexthops with an IPIP device are programmed as
4804          * encapsulating adjacency entries.
4805          */
4806         if (!nh->gw_family && !nh->is_reject &&
4807             !mlxsw_sp_netdev_ipip_type(mlxsw_sp, nh->dev, NULL)) {
4808                 NL_SET_ERR_MSG_MOD(extack, "Nexthop group entry does not have a gateway");
4809                 return -EINVAL;
4810         }
4811
4812         return 0;
4813 }
4814
4815 static int
4816 mlxsw_sp_nexthop_obj_group_validate(struct mlxsw_sp *mlxsw_sp,
4817                                     const struct nh_notifier_grp_info *nh_grp,
4818                                     struct netlink_ext_ack *extack)
4819 {
4820         int i;
4821
4822         if (nh_grp->is_fdb) {
4823                 NL_SET_ERR_MSG_MOD(extack, "FDB nexthop groups are not supported");
4824                 return -EINVAL;
4825         }
4826
4827         for (i = 0; i < nh_grp->num_nh; i++) {
4828                 const struct nh_notifier_single_info *nh;
4829                 int err;
4830
4831                 nh = &nh_grp->nh_entries[i].nh;
4832                 err = mlxsw_sp_nexthop_obj_group_entry_validate(mlxsw_sp, nh,
4833                                                                 extack);
4834                 if (err)
4835                         return err;
4836         }
4837
4838         return 0;
4839 }
4840
4841 static int
4842 mlxsw_sp_nexthop_obj_res_group_size_validate(struct mlxsw_sp *mlxsw_sp,
4843                                              const struct nh_notifier_res_table_info *nh_res_table,
4844                                              struct netlink_ext_ack *extack)
4845 {
4846         unsigned int alloc_size;
4847         bool valid_size = false;
4848         int err, i;
4849
4850         if (nh_res_table->num_nh_buckets < 32) {
4851                 NL_SET_ERR_MSG_MOD(extack, "Minimum number of buckets is 32");
4852                 return -EINVAL;
4853         }
4854
4855         for (i = 0; i < mlxsw_sp->router->adj_grp_size_ranges_count; i++) {
4856                 const struct mlxsw_sp_adj_grp_size_range *size_range;
4857
4858                 size_range = &mlxsw_sp->router->adj_grp_size_ranges[i];
4859
4860                 if (nh_res_table->num_nh_buckets >= size_range->start &&
4861                     nh_res_table->num_nh_buckets <= size_range->end) {
4862                         valid_size = true;
4863                         break;
4864                 }
4865         }
4866
4867         if (!valid_size) {
4868                 NL_SET_ERR_MSG_MOD(extack, "Invalid number of buckets");
4869                 return -EINVAL;
4870         }
4871
4872         err = mlxsw_sp_kvdl_alloc_count_query(mlxsw_sp,
4873                                               MLXSW_SP_KVDL_ENTRY_TYPE_ADJ,
4874                                               nh_res_table->num_nh_buckets,
4875                                               &alloc_size);
4876         if (err || nh_res_table->num_nh_buckets != alloc_size) {
4877                 NL_SET_ERR_MSG_MOD(extack, "Number of buckets does not fit allocation size of any KVDL partition");
4878                 return -EINVAL;
4879         }
4880
4881         return 0;
4882 }
4883
4884 static int
4885 mlxsw_sp_nexthop_obj_res_group_validate(struct mlxsw_sp *mlxsw_sp,
4886                                         const struct nh_notifier_res_table_info *nh_res_table,
4887                                         struct netlink_ext_ack *extack)
4888 {
4889         int err;
4890         u16 i;
4891
4892         err = mlxsw_sp_nexthop_obj_res_group_size_validate(mlxsw_sp,
4893                                                            nh_res_table,
4894                                                            extack);
4895         if (err)
4896                 return err;
4897
4898         for (i = 0; i < nh_res_table->num_nh_buckets; i++) {
4899                 const struct nh_notifier_single_info *nh;
4900                 int err;
4901
4902                 nh = &nh_res_table->nhs[i];
4903                 err = mlxsw_sp_nexthop_obj_group_entry_validate(mlxsw_sp, nh,
4904                                                                 extack);
4905                 if (err)
4906                         return err;
4907         }
4908
4909         return 0;
4910 }
4911
4912 static int mlxsw_sp_nexthop_obj_validate(struct mlxsw_sp *mlxsw_sp,
4913                                          unsigned long event,
4914                                          struct nh_notifier_info *info)
4915 {
4916         struct nh_notifier_single_info *nh;
4917
4918         if (event != NEXTHOP_EVENT_REPLACE &&
4919             event != NEXTHOP_EVENT_RES_TABLE_PRE_REPLACE &&
4920             event != NEXTHOP_EVENT_BUCKET_REPLACE)
4921                 return 0;
4922
4923         switch (info->type) {
4924         case NH_NOTIFIER_INFO_TYPE_SINGLE:
4925                 return mlxsw_sp_nexthop_obj_single_validate(mlxsw_sp, info->nh,
4926                                                             info->extack);
4927         case NH_NOTIFIER_INFO_TYPE_GRP:
4928                 return mlxsw_sp_nexthop_obj_group_validate(mlxsw_sp,
4929                                                            info->nh_grp,
4930                                                            info->extack);
4931         case NH_NOTIFIER_INFO_TYPE_RES_TABLE:
4932                 return mlxsw_sp_nexthop_obj_res_group_validate(mlxsw_sp,
4933                                                                info->nh_res_table,
4934                                                                info->extack);
4935         case NH_NOTIFIER_INFO_TYPE_RES_BUCKET:
4936                 nh = &info->nh_res_bucket->new_nh;
4937                 return mlxsw_sp_nexthop_obj_group_entry_validate(mlxsw_sp, nh,
4938                                                                  info->extack);
4939         default:
4940                 NL_SET_ERR_MSG_MOD(info->extack, "Unsupported nexthop type");
4941                 return -EOPNOTSUPP;
4942         }
4943 }
4944
4945 static bool mlxsw_sp_nexthop_obj_is_gateway(struct mlxsw_sp *mlxsw_sp,
4946                                             const struct nh_notifier_info *info)
4947 {
4948         const struct net_device *dev;
4949
4950         switch (info->type) {
4951         case NH_NOTIFIER_INFO_TYPE_SINGLE:
4952                 dev = info->nh->dev;
4953                 return info->nh->gw_family || info->nh->is_reject ||
4954                        mlxsw_sp_netdev_ipip_type(mlxsw_sp, dev, NULL);
4955         case NH_NOTIFIER_INFO_TYPE_GRP:
4956         case NH_NOTIFIER_INFO_TYPE_RES_TABLE:
4957                 /* Already validated earlier. */
4958                 return true;
4959         default:
4960                 return false;
4961         }
4962 }
4963
4964 static void mlxsw_sp_nexthop_obj_blackhole_init(struct mlxsw_sp *mlxsw_sp,
4965                                                 struct mlxsw_sp_nexthop *nh)
4966 {
4967         nh->action = MLXSW_SP_NEXTHOP_ACTION_DISCARD;
4968         nh->should_offload = 1;
4969         /* While nexthops that discard packets do not forward packets
4970          * via an egress RIF, they still need to be programmed using a
4971          * valid RIF, so use the loopback RIF created during init.
4972          */
4973         nh->crif = mlxsw_sp->router->lb_crif;
4974 }
4975
4976 static void mlxsw_sp_nexthop_obj_blackhole_fini(struct mlxsw_sp *mlxsw_sp,
4977                                                 struct mlxsw_sp_nexthop *nh)
4978 {
4979         nh->crif = NULL;
4980         nh->should_offload = 0;
4981 }
4982
4983 static int
4984 mlxsw_sp_nexthop_obj_init(struct mlxsw_sp *mlxsw_sp,
4985                           struct mlxsw_sp_nexthop_group *nh_grp,
4986                           struct mlxsw_sp_nexthop *nh,
4987                           struct nh_notifier_single_info *nh_obj, int weight)
4988 {
4989         struct net_device *dev = nh_obj->dev;
4990         int err;
4991
4992         nh->nhgi = nh_grp->nhgi;
4993         nh->nh_weight = weight;
4994
4995         switch (nh_obj->gw_family) {
4996         case AF_INET:
4997                 memcpy(&nh->gw_addr, &nh_obj->ipv4, sizeof(nh_obj->ipv4));
4998                 nh->neigh_tbl = &arp_tbl;
4999                 break;
5000         case AF_INET6:
5001                 memcpy(&nh->gw_addr, &nh_obj->ipv6, sizeof(nh_obj->ipv6));
5002 #if IS_ENABLED(CONFIG_IPV6)
5003                 nh->neigh_tbl = &nd_tbl;
5004 #endif
5005                 break;
5006         }
5007
5008         mlxsw_sp_nexthop_counter_alloc(mlxsw_sp, nh);
5009         list_add_tail(&nh->router_list_node, &mlxsw_sp->router->nexthop_list);
5010         nh->ifindex = dev->ifindex;
5011
5012         err = mlxsw_sp_nexthop_type_init(mlxsw_sp, nh, dev);
5013         if (err)
5014                 goto err_type_init;
5015
5016         if (nh_obj->is_reject)
5017                 mlxsw_sp_nexthop_obj_blackhole_init(mlxsw_sp, nh);
5018
5019         /* In a resilient nexthop group, all the nexthops must be written to
5020          * the adjacency table. Even if they do not have a valid neighbour or
5021          * RIF.
5022          */
5023         if (nh_grp->nhgi->is_resilient && !nh->should_offload) {
5024                 nh->action = MLXSW_SP_NEXTHOP_ACTION_TRAP;
5025                 nh->should_offload = 1;
5026         }
5027
5028         return 0;
5029
5030 err_type_init:
5031         list_del(&nh->router_list_node);
5032         mlxsw_sp_nexthop_counter_free(mlxsw_sp, nh);
5033         return err;
5034 }
5035
5036 static void mlxsw_sp_nexthop_obj_fini(struct mlxsw_sp *mlxsw_sp,
5037                                       struct mlxsw_sp_nexthop *nh)
5038 {
5039         if (nh->action == MLXSW_SP_NEXTHOP_ACTION_DISCARD)
5040                 mlxsw_sp_nexthop_obj_blackhole_fini(mlxsw_sp, nh);
5041         mlxsw_sp_nexthop_type_fini(mlxsw_sp, nh);
5042         list_del(&nh->router_list_node);
5043         mlxsw_sp_nexthop_counter_free(mlxsw_sp, nh);
5044         nh->should_offload = 0;
5045 }
5046
5047 static int
5048 mlxsw_sp_nexthop_obj_group_info_init(struct mlxsw_sp *mlxsw_sp,
5049                                      struct mlxsw_sp_nexthop_group *nh_grp,
5050                                      struct nh_notifier_info *info)
5051 {
5052         struct mlxsw_sp_nexthop_group_info *nhgi;
5053         struct mlxsw_sp_nexthop *nh;
5054         bool is_resilient = false;
5055         unsigned int nhs;
5056         int err, i;
5057
5058         switch (info->type) {
5059         case NH_NOTIFIER_INFO_TYPE_SINGLE:
5060                 nhs = 1;
5061                 break;
5062         case NH_NOTIFIER_INFO_TYPE_GRP:
5063                 nhs = info->nh_grp->num_nh;
5064                 break;
5065         case NH_NOTIFIER_INFO_TYPE_RES_TABLE:
5066                 nhs = info->nh_res_table->num_nh_buckets;
5067                 is_resilient = true;
5068                 break;
5069         default:
5070                 return -EINVAL;
5071         }
5072
5073         nhgi = kzalloc(struct_size(nhgi, nexthops, nhs), GFP_KERNEL);
5074         if (!nhgi)
5075                 return -ENOMEM;
5076         nh_grp->nhgi = nhgi;
5077         nhgi->nh_grp = nh_grp;
5078         nhgi->gateway = mlxsw_sp_nexthop_obj_is_gateway(mlxsw_sp, info);
5079         nhgi->is_resilient = is_resilient;
5080         nhgi->count = nhs;
5081         for (i = 0; i < nhgi->count; i++) {
5082                 struct nh_notifier_single_info *nh_obj;
5083                 int weight;
5084
5085                 nh = &nhgi->nexthops[i];
5086                 switch (info->type) {
5087                 case NH_NOTIFIER_INFO_TYPE_SINGLE:
5088                         nh_obj = info->nh;
5089                         weight = 1;
5090                         break;
5091                 case NH_NOTIFIER_INFO_TYPE_GRP:
5092                         nh_obj = &info->nh_grp->nh_entries[i].nh;
5093                         weight = info->nh_grp->nh_entries[i].weight;
5094                         break;
5095                 case NH_NOTIFIER_INFO_TYPE_RES_TABLE:
5096                         nh_obj = &info->nh_res_table->nhs[i];
5097                         weight = 1;
5098                         break;
5099                 default:
5100                         err = -EINVAL;
5101                         goto err_nexthop_obj_init;
5102                 }
5103                 err = mlxsw_sp_nexthop_obj_init(mlxsw_sp, nh_grp, nh, nh_obj,
5104                                                 weight);
5105                 if (err)
5106                         goto err_nexthop_obj_init;
5107         }
5108         err = mlxsw_sp_nexthop_group_inc(mlxsw_sp);
5109         if (err)
5110                 goto err_group_inc;
5111         err = mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh_grp);
5112         if (err) {
5113                 NL_SET_ERR_MSG_MOD(info->extack, "Failed to write adjacency entries to the device");
5114                 goto err_group_refresh;
5115         }
5116
5117         /* Add resilient nexthop groups to a list so that the activity of their
5118          * nexthop buckets will be periodically queried and cleared.
5119          */
5120         if (nhgi->is_resilient) {
5121                 if (list_empty(&mlxsw_sp->router->nh_res_grp_list))
5122                         mlxsw_sp_nh_grp_activity_work_schedule(mlxsw_sp);
5123                 list_add(&nhgi->list, &mlxsw_sp->router->nh_res_grp_list);
5124         }
5125
5126         return 0;
5127
5128 err_group_refresh:
5129         mlxsw_sp_nexthop_group_dec(mlxsw_sp);
5130 err_group_inc:
5131         i = nhgi->count;
5132 err_nexthop_obj_init:
5133         for (i--; i >= 0; i--) {
5134                 nh = &nhgi->nexthops[i];
5135                 mlxsw_sp_nexthop_obj_fini(mlxsw_sp, nh);
5136         }
5137         kfree(nhgi);
5138         return err;
5139 }
5140
5141 static void
5142 mlxsw_sp_nexthop_obj_group_info_fini(struct mlxsw_sp *mlxsw_sp,
5143                                      struct mlxsw_sp_nexthop_group *nh_grp)
5144 {
5145         struct mlxsw_sp_nexthop_group_info *nhgi = nh_grp->nhgi;
5146         struct mlxsw_sp_router *router = mlxsw_sp->router;
5147         int i;
5148
5149         if (nhgi->is_resilient) {
5150                 list_del(&nhgi->list);
5151                 if (list_empty(&mlxsw_sp->router->nh_res_grp_list))
5152                         cancel_delayed_work(&router->nh_grp_activity_dw);
5153         }
5154
5155         mlxsw_sp_nexthop_group_dec(mlxsw_sp);
5156         for (i = nhgi->count - 1; i >= 0; i--) {
5157                 struct mlxsw_sp_nexthop *nh = &nhgi->nexthops[i];
5158
5159                 mlxsw_sp_nexthop_obj_fini(mlxsw_sp, nh);
5160         }
5161         mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh_grp);
5162         WARN_ON_ONCE(nhgi->adj_index_valid);
5163         kfree(nhgi);
5164 }
5165
5166 static struct mlxsw_sp_nexthop_group *
5167 mlxsw_sp_nexthop_obj_group_create(struct mlxsw_sp *mlxsw_sp,
5168                                   struct nh_notifier_info *info)
5169 {
5170         struct mlxsw_sp_nexthop_group *nh_grp;
5171         int err;
5172
5173         nh_grp = kzalloc(sizeof(*nh_grp), GFP_KERNEL);
5174         if (!nh_grp)
5175                 return ERR_PTR(-ENOMEM);
5176         INIT_LIST_HEAD(&nh_grp->vr_list);
5177         err = rhashtable_init(&nh_grp->vr_ht,
5178                               &mlxsw_sp_nexthop_group_vr_ht_params);
5179         if (err)
5180                 goto err_nexthop_group_vr_ht_init;
5181         INIT_LIST_HEAD(&nh_grp->fib_list);
5182         nh_grp->type = MLXSW_SP_NEXTHOP_GROUP_TYPE_OBJ;
5183         nh_grp->obj.id = info->id;
5184
5185         err = mlxsw_sp_nexthop_obj_group_info_init(mlxsw_sp, nh_grp, info);
5186         if (err)
5187                 goto err_nexthop_group_info_init;
5188
5189         nh_grp->can_destroy = false;
5190
5191         return nh_grp;
5192
5193 err_nexthop_group_info_init:
5194         rhashtable_destroy(&nh_grp->vr_ht);
5195 err_nexthop_group_vr_ht_init:
5196         kfree(nh_grp);
5197         return ERR_PTR(err);
5198 }
5199
5200 static void
5201 mlxsw_sp_nexthop_obj_group_destroy(struct mlxsw_sp *mlxsw_sp,
5202                                    struct mlxsw_sp_nexthop_group *nh_grp)
5203 {
5204         if (!nh_grp->can_destroy)
5205                 return;
5206         mlxsw_sp_nexthop_obj_group_info_fini(mlxsw_sp, nh_grp);
5207         WARN_ON_ONCE(!list_empty(&nh_grp->fib_list));
5208         WARN_ON_ONCE(!list_empty(&nh_grp->vr_list));
5209         rhashtable_destroy(&nh_grp->vr_ht);
5210         kfree(nh_grp);
5211 }
5212
5213 static struct mlxsw_sp_nexthop_group *
5214 mlxsw_sp_nexthop_obj_group_lookup(struct mlxsw_sp *mlxsw_sp, u32 id)
5215 {
5216         struct mlxsw_sp_nexthop_group_cmp_arg cmp_arg;
5217
5218         cmp_arg.type = MLXSW_SP_NEXTHOP_GROUP_TYPE_OBJ;
5219         cmp_arg.id = id;
5220         return rhashtable_lookup_fast(&mlxsw_sp->router->nexthop_group_ht,
5221                                       &cmp_arg,
5222                                       mlxsw_sp_nexthop_group_ht_params);
5223 }
5224
5225 static int mlxsw_sp_nexthop_obj_group_add(struct mlxsw_sp *mlxsw_sp,
5226                                           struct mlxsw_sp_nexthop_group *nh_grp)
5227 {
5228         return mlxsw_sp_nexthop_group_insert(mlxsw_sp, nh_grp);
5229 }
5230
5231 static int
5232 mlxsw_sp_nexthop_obj_group_replace(struct mlxsw_sp *mlxsw_sp,
5233                                    struct mlxsw_sp_nexthop_group *nh_grp,
5234                                    struct mlxsw_sp_nexthop_group *old_nh_grp,
5235                                    struct netlink_ext_ack *extack)
5236 {
5237         struct mlxsw_sp_nexthop_group_info *old_nhgi = old_nh_grp->nhgi;
5238         struct mlxsw_sp_nexthop_group_info *new_nhgi = nh_grp->nhgi;
5239         int err;
5240
5241         old_nh_grp->nhgi = new_nhgi;
5242         new_nhgi->nh_grp = old_nh_grp;
5243         nh_grp->nhgi = old_nhgi;
5244         old_nhgi->nh_grp = nh_grp;
5245
5246         if (old_nhgi->adj_index_valid && new_nhgi->adj_index_valid) {
5247                 /* Both the old adjacency index and the new one are valid.
5248                  * Routes are currently using the old one. Tell the device to
5249                  * replace the old adjacency index with the new one.
5250                  */
5251                 err = mlxsw_sp_adj_index_mass_update(mlxsw_sp, old_nh_grp,
5252                                                      old_nhgi->adj_index,
5253                                                      old_nhgi->ecmp_size);
5254                 if (err) {
5255                         NL_SET_ERR_MSG_MOD(extack, "Failed to replace old adjacency index with new one");
5256                         goto err_out;
5257                 }
5258         } else if (old_nhgi->adj_index_valid && !new_nhgi->adj_index_valid) {
5259                 /* The old adjacency index is valid, while the new one is not.
5260                  * Iterate over all the routes using the group and change them
5261                  * to trap packets to the CPU.
5262                  */
5263                 err = mlxsw_sp_nexthop_fib_entries_update(mlxsw_sp, old_nh_grp);
5264                 if (err) {
5265                         NL_SET_ERR_MSG_MOD(extack, "Failed to update routes to trap packets");
5266                         goto err_out;
5267                 }
5268         } else if (!old_nhgi->adj_index_valid && new_nhgi->adj_index_valid) {
5269                 /* The old adjacency index is invalid, while the new one is.
5270                  * Iterate over all the routes using the group and change them
5271                  * to forward packets using the new valid index.
5272                  */
5273                 err = mlxsw_sp_nexthop_fib_entries_update(mlxsw_sp, old_nh_grp);
5274                 if (err) {
5275                         NL_SET_ERR_MSG_MOD(extack, "Failed to update routes to forward packets");
5276                         goto err_out;
5277                 }
5278         }
5279
5280         /* Make sure the flags are set / cleared based on the new nexthop group
5281          * information.
5282          */
5283         mlxsw_sp_nexthop_obj_group_offload_refresh(mlxsw_sp, old_nh_grp);
5284
5285         /* At this point 'nh_grp' is just a shell that is not used by anyone
5286          * and its nexthop group info is the old info that was just replaced
5287          * with the new one. Remove it.
5288          */
5289         nh_grp->can_destroy = true;
5290         mlxsw_sp_nexthop_obj_group_destroy(mlxsw_sp, nh_grp);
5291
5292         return 0;
5293
5294 err_out:
5295         old_nhgi->nh_grp = old_nh_grp;
5296         nh_grp->nhgi = new_nhgi;
5297         new_nhgi->nh_grp = nh_grp;
5298         old_nh_grp->nhgi = old_nhgi;
5299         return err;
5300 }
5301
5302 static int mlxsw_sp_nexthop_obj_new(struct mlxsw_sp *mlxsw_sp,
5303                                     struct nh_notifier_info *info)
5304 {
5305         struct mlxsw_sp_nexthop_group *nh_grp, *old_nh_grp;
5306         struct netlink_ext_ack *extack = info->extack;
5307         int err;
5308
5309         nh_grp = mlxsw_sp_nexthop_obj_group_create(mlxsw_sp, info);
5310         if (IS_ERR(nh_grp))
5311                 return PTR_ERR(nh_grp);
5312
5313         old_nh_grp = mlxsw_sp_nexthop_obj_group_lookup(mlxsw_sp, info->id);
5314         if (!old_nh_grp)
5315                 err = mlxsw_sp_nexthop_obj_group_add(mlxsw_sp, nh_grp);
5316         else
5317                 err = mlxsw_sp_nexthop_obj_group_replace(mlxsw_sp, nh_grp,
5318                                                          old_nh_grp, extack);
5319
5320         if (err) {
5321                 nh_grp->can_destroy = true;
5322                 mlxsw_sp_nexthop_obj_group_destroy(mlxsw_sp, nh_grp);
5323         }
5324
5325         return err;
5326 }
5327
5328 static void mlxsw_sp_nexthop_obj_del(struct mlxsw_sp *mlxsw_sp,
5329                                      struct nh_notifier_info *info)
5330 {
5331         struct mlxsw_sp_nexthop_group *nh_grp;
5332
5333         nh_grp = mlxsw_sp_nexthop_obj_group_lookup(mlxsw_sp, info->id);
5334         if (!nh_grp)
5335                 return;
5336
5337         nh_grp->can_destroy = true;
5338         mlxsw_sp_nexthop_group_remove(mlxsw_sp, nh_grp);
5339
5340         /* If the group still has routes using it, then defer the delete
5341          * operation until the last route using it is deleted.
5342          */
5343         if (!list_empty(&nh_grp->fib_list))
5344                 return;
5345         mlxsw_sp_nexthop_obj_group_destroy(mlxsw_sp, nh_grp);
5346 }
5347
5348 static int mlxsw_sp_nexthop_obj_bucket_query(struct mlxsw_sp *mlxsw_sp,
5349                                              u32 adj_index, char *ratr_pl)
5350 {
5351         MLXSW_REG_ZERO(ratr, ratr_pl);
5352         mlxsw_reg_ratr_op_set(ratr_pl, MLXSW_REG_RATR_OP_QUERY_READ);
5353         mlxsw_reg_ratr_adjacency_index_low_set(ratr_pl, adj_index);
5354         mlxsw_reg_ratr_adjacency_index_high_set(ratr_pl, adj_index >> 16);
5355
5356         return mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(ratr), ratr_pl);
5357 }
5358
5359 static int mlxsw_sp_nexthop_obj_bucket_compare(char *ratr_pl, char *ratr_pl_new)
5360 {
5361         /* Clear the opcode and activity on both the old and new payload as
5362          * they are irrelevant for the comparison.
5363          */
5364         mlxsw_reg_ratr_op_set(ratr_pl, MLXSW_REG_RATR_OP_QUERY_READ);
5365         mlxsw_reg_ratr_a_set(ratr_pl, 0);
5366         mlxsw_reg_ratr_op_set(ratr_pl_new, MLXSW_REG_RATR_OP_QUERY_READ);
5367         mlxsw_reg_ratr_a_set(ratr_pl_new, 0);
5368
5369         /* If the contents of the adjacency entry are consistent with the
5370          * replacement request, then replacement was successful.
5371          */
5372         if (!memcmp(ratr_pl, ratr_pl_new, MLXSW_REG_RATR_LEN))
5373                 return 0;
5374
5375         return -EINVAL;
5376 }
5377
5378 static int
5379 mlxsw_sp_nexthop_obj_bucket_adj_update(struct mlxsw_sp *mlxsw_sp,
5380                                        struct mlxsw_sp_nexthop *nh,
5381                                        struct nh_notifier_info *info)
5382 {
5383         u16 bucket_index = info->nh_res_bucket->bucket_index;
5384         struct netlink_ext_ack *extack = info->extack;
5385         bool force = info->nh_res_bucket->force;
5386         char ratr_pl_new[MLXSW_REG_RATR_LEN];
5387         char ratr_pl[MLXSW_REG_RATR_LEN];
5388         u32 adj_index;
5389         int err;
5390
5391         /* No point in trying an atomic replacement if the idle timer interval
5392          * is smaller than the interval in which we query and clear activity.
5393          */
5394         if (!force && info->nh_res_bucket->idle_timer_ms <
5395             MLXSW_SP_NH_GRP_ACTIVITY_UPDATE_INTERVAL)
5396                 force = true;
5397
5398         adj_index = nh->nhgi->adj_index + bucket_index;
5399         err = mlxsw_sp_nexthop_update(mlxsw_sp, adj_index, nh, force, ratr_pl);
5400         if (err) {
5401                 NL_SET_ERR_MSG_MOD(extack, "Failed to overwrite nexthop bucket");
5402                 return err;
5403         }
5404
5405         if (!force) {
5406                 err = mlxsw_sp_nexthop_obj_bucket_query(mlxsw_sp, adj_index,
5407                                                         ratr_pl_new);
5408                 if (err) {
5409                         NL_SET_ERR_MSG_MOD(extack, "Failed to query nexthop bucket state after replacement. State might be inconsistent");
5410                         return err;
5411                 }
5412
5413                 err = mlxsw_sp_nexthop_obj_bucket_compare(ratr_pl, ratr_pl_new);
5414                 if (err) {
5415                         NL_SET_ERR_MSG_MOD(extack, "Nexthop bucket was not replaced because it was active during replacement");
5416                         return err;
5417                 }
5418         }
5419
5420         nh->update = 0;
5421         nh->offloaded = 1;
5422         mlxsw_sp_nexthop_bucket_offload_refresh(mlxsw_sp, nh, bucket_index);
5423
5424         return 0;
5425 }
5426
5427 static int mlxsw_sp_nexthop_obj_bucket_replace(struct mlxsw_sp *mlxsw_sp,
5428                                                struct nh_notifier_info *info)
5429 {
5430         u16 bucket_index = info->nh_res_bucket->bucket_index;
5431         struct netlink_ext_ack *extack = info->extack;
5432         struct mlxsw_sp_nexthop_group_info *nhgi;
5433         struct nh_notifier_single_info *nh_obj;
5434         struct mlxsw_sp_nexthop_group *nh_grp;
5435         struct mlxsw_sp_nexthop *nh;
5436         int err;
5437
5438         nh_grp = mlxsw_sp_nexthop_obj_group_lookup(mlxsw_sp, info->id);
5439         if (!nh_grp) {
5440                 NL_SET_ERR_MSG_MOD(extack, "Nexthop group was not found");
5441                 return -EINVAL;
5442         }
5443
5444         nhgi = nh_grp->nhgi;
5445
5446         if (bucket_index >= nhgi->count) {
5447                 NL_SET_ERR_MSG_MOD(extack, "Nexthop bucket index out of range");
5448                 return -EINVAL;
5449         }
5450
5451         nh = &nhgi->nexthops[bucket_index];
5452         mlxsw_sp_nexthop_obj_fini(mlxsw_sp, nh);
5453
5454         nh_obj = &info->nh_res_bucket->new_nh;
5455         err = mlxsw_sp_nexthop_obj_init(mlxsw_sp, nh_grp, nh, nh_obj, 1);
5456         if (err) {
5457                 NL_SET_ERR_MSG_MOD(extack, "Failed to initialize nexthop object for nexthop bucket replacement");
5458                 goto err_nexthop_obj_init;
5459         }
5460
5461         err = mlxsw_sp_nexthop_obj_bucket_adj_update(mlxsw_sp, nh, info);
5462         if (err)
5463                 goto err_nexthop_obj_bucket_adj_update;
5464
5465         return 0;
5466
5467 err_nexthop_obj_bucket_adj_update:
5468         mlxsw_sp_nexthop_obj_fini(mlxsw_sp, nh);
5469 err_nexthop_obj_init:
5470         nh_obj = &info->nh_res_bucket->old_nh;
5471         mlxsw_sp_nexthop_obj_init(mlxsw_sp, nh_grp, nh, nh_obj, 1);
5472         /* The old adjacency entry was not overwritten */
5473         nh->update = 0;
5474         nh->offloaded = 1;
5475         return err;
5476 }
5477
5478 static int mlxsw_sp_nexthop_obj_event(struct notifier_block *nb,
5479                                       unsigned long event, void *ptr)
5480 {
5481         struct nh_notifier_info *info = ptr;
5482         struct mlxsw_sp_router *router;
5483         int err = 0;
5484
5485         router = container_of(nb, struct mlxsw_sp_router, nexthop_nb);
5486         err = mlxsw_sp_nexthop_obj_validate(router->mlxsw_sp, event, info);
5487         if (err)
5488                 goto out;
5489
5490         mutex_lock(&router->lock);
5491
5492         switch (event) {
5493         case NEXTHOP_EVENT_REPLACE:
5494                 err = mlxsw_sp_nexthop_obj_new(router->mlxsw_sp, info);
5495                 break;
5496         case NEXTHOP_EVENT_DEL:
5497                 mlxsw_sp_nexthop_obj_del(router->mlxsw_sp, info);
5498                 break;
5499         case NEXTHOP_EVENT_BUCKET_REPLACE:
5500                 err = mlxsw_sp_nexthop_obj_bucket_replace(router->mlxsw_sp,
5501                                                           info);
5502                 break;
5503         default:
5504                 break;
5505         }
5506
5507         mutex_unlock(&router->lock);
5508
5509 out:
5510         return notifier_from_errno(err);
5511 }
5512
5513 static bool mlxsw_sp_fi_is_gateway(const struct mlxsw_sp *mlxsw_sp,
5514                                    struct fib_info *fi)
5515 {
5516         const struct fib_nh *nh = fib_info_nh(fi, 0);
5517
5518         return nh->fib_nh_gw_family ||
5519                mlxsw_sp_nexthop4_ipip_type(mlxsw_sp, nh, NULL);
5520 }
5521
5522 static int
5523 mlxsw_sp_nexthop4_group_info_init(struct mlxsw_sp *mlxsw_sp,
5524                                   struct mlxsw_sp_nexthop_group *nh_grp)
5525 {
5526         unsigned int nhs = fib_info_num_path(nh_grp->ipv4.fi);
5527         struct mlxsw_sp_nexthop_group_info *nhgi;
5528         struct mlxsw_sp_nexthop *nh;
5529         int err, i;
5530
5531         nhgi = kzalloc(struct_size(nhgi, nexthops, nhs), GFP_KERNEL);
5532         if (!nhgi)
5533                 return -ENOMEM;
5534         nh_grp->nhgi = nhgi;
5535         nhgi->nh_grp = nh_grp;
5536         nhgi->gateway = mlxsw_sp_fi_is_gateway(mlxsw_sp, nh_grp->ipv4.fi);
5537         nhgi->count = nhs;
5538         for (i = 0; i < nhgi->count; i++) {
5539                 struct fib_nh *fib_nh;
5540
5541                 nh = &nhgi->nexthops[i];
5542                 fib_nh = fib_info_nh(nh_grp->ipv4.fi, i);
5543                 err = mlxsw_sp_nexthop4_init(mlxsw_sp, nh_grp, nh, fib_nh);
5544                 if (err)
5545                         goto err_nexthop4_init;
5546         }
5547         err = mlxsw_sp_nexthop_group_inc(mlxsw_sp);
5548         if (err)
5549                 goto err_group_inc;
5550         err = mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh_grp);
5551         if (err)
5552                 goto err_group_refresh;
5553
5554         return 0;
5555
5556 err_group_refresh:
5557         mlxsw_sp_nexthop_group_dec(mlxsw_sp);
5558 err_group_inc:
5559         i = nhgi->count;
5560 err_nexthop4_init:
5561         for (i--; i >= 0; i--) {
5562                 nh = &nhgi->nexthops[i];
5563                 mlxsw_sp_nexthop4_fini(mlxsw_sp, nh);
5564         }
5565         kfree(nhgi);
5566         return err;
5567 }
5568
5569 static void
5570 mlxsw_sp_nexthop4_group_info_fini(struct mlxsw_sp *mlxsw_sp,
5571                                   struct mlxsw_sp_nexthop_group *nh_grp)
5572 {
5573         struct mlxsw_sp_nexthop_group_info *nhgi = nh_grp->nhgi;
5574         int i;
5575
5576         mlxsw_sp_nexthop_group_dec(mlxsw_sp);
5577         for (i = nhgi->count - 1; i >= 0; i--) {
5578                 struct mlxsw_sp_nexthop *nh = &nhgi->nexthops[i];
5579
5580                 mlxsw_sp_nexthop4_fini(mlxsw_sp, nh);
5581         }
5582         mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh_grp);
5583         WARN_ON_ONCE(nhgi->adj_index_valid);
5584         kfree(nhgi);
5585 }
5586
5587 static struct mlxsw_sp_nexthop_group *
5588 mlxsw_sp_nexthop4_group_create(struct mlxsw_sp *mlxsw_sp, struct fib_info *fi)
5589 {
5590         struct mlxsw_sp_nexthop_group *nh_grp;
5591         int err;
5592
5593         nh_grp = kzalloc(sizeof(*nh_grp), GFP_KERNEL);
5594         if (!nh_grp)
5595                 return ERR_PTR(-ENOMEM);
5596         INIT_LIST_HEAD(&nh_grp->vr_list);
5597         err = rhashtable_init(&nh_grp->vr_ht,
5598                               &mlxsw_sp_nexthop_group_vr_ht_params);
5599         if (err)
5600                 goto err_nexthop_group_vr_ht_init;
5601         INIT_LIST_HEAD(&nh_grp->fib_list);
5602         nh_grp->type = MLXSW_SP_NEXTHOP_GROUP_TYPE_IPV4;
5603         nh_grp->ipv4.fi = fi;
5604         fib_info_hold(fi);
5605
5606         err = mlxsw_sp_nexthop4_group_info_init(mlxsw_sp, nh_grp);
5607         if (err)
5608                 goto err_nexthop_group_info_init;
5609
5610         err = mlxsw_sp_nexthop_group_insert(mlxsw_sp, nh_grp);
5611         if (err)
5612                 goto err_nexthop_group_insert;
5613
5614         nh_grp->can_destroy = true;
5615
5616         return nh_grp;
5617
5618 err_nexthop_group_insert:
5619         mlxsw_sp_nexthop4_group_info_fini(mlxsw_sp, nh_grp);
5620 err_nexthop_group_info_init:
5621         fib_info_put(fi);
5622         rhashtable_destroy(&nh_grp->vr_ht);
5623 err_nexthop_group_vr_ht_init:
5624         kfree(nh_grp);
5625         return ERR_PTR(err);
5626 }
5627
5628 static void
5629 mlxsw_sp_nexthop4_group_destroy(struct mlxsw_sp *mlxsw_sp,
5630                                 struct mlxsw_sp_nexthop_group *nh_grp)
5631 {
5632         if (!nh_grp->can_destroy)
5633                 return;
5634         mlxsw_sp_nexthop_group_remove(mlxsw_sp, nh_grp);
5635         mlxsw_sp_nexthop4_group_info_fini(mlxsw_sp, nh_grp);
5636         fib_info_put(nh_grp->ipv4.fi);
5637         WARN_ON_ONCE(!list_empty(&nh_grp->vr_list));
5638         rhashtable_destroy(&nh_grp->vr_ht);
5639         kfree(nh_grp);
5640 }
5641
5642 static int mlxsw_sp_nexthop4_group_get(struct mlxsw_sp *mlxsw_sp,
5643                                        struct mlxsw_sp_fib_entry *fib_entry,
5644                                        struct fib_info *fi)
5645 {
5646         struct mlxsw_sp_nexthop_group *nh_grp;
5647
5648         if (fi->nh) {
5649                 nh_grp = mlxsw_sp_nexthop_obj_group_lookup(mlxsw_sp,
5650                                                            fi->nh->id);
5651                 if (WARN_ON_ONCE(!nh_grp))
5652                         return -EINVAL;
5653                 goto out;
5654         }
5655
5656         nh_grp = mlxsw_sp_nexthop4_group_lookup(mlxsw_sp, fi);
5657         if (!nh_grp) {
5658                 nh_grp = mlxsw_sp_nexthop4_group_create(mlxsw_sp, fi);
5659                 if (IS_ERR(nh_grp))
5660                         return PTR_ERR(nh_grp);
5661         }
5662 out:
5663         list_add_tail(&fib_entry->nexthop_group_node, &nh_grp->fib_list);
5664         fib_entry->nh_group = nh_grp;
5665         return 0;
5666 }
5667
5668 static void mlxsw_sp_nexthop4_group_put(struct mlxsw_sp *mlxsw_sp,
5669                                         struct mlxsw_sp_fib_entry *fib_entry)
5670 {
5671         struct mlxsw_sp_nexthop_group *nh_grp = fib_entry->nh_group;
5672
5673         list_del(&fib_entry->nexthop_group_node);
5674         if (!list_empty(&nh_grp->fib_list))
5675                 return;
5676
5677         if (nh_grp->type == MLXSW_SP_NEXTHOP_GROUP_TYPE_OBJ) {
5678                 mlxsw_sp_nexthop_obj_group_destroy(mlxsw_sp, nh_grp);
5679                 return;
5680         }
5681
5682         mlxsw_sp_nexthop4_group_destroy(mlxsw_sp, nh_grp);
5683 }
5684
5685 static bool
5686 mlxsw_sp_fib4_entry_should_offload(const struct mlxsw_sp_fib_entry *fib_entry)
5687 {
5688         struct mlxsw_sp_fib4_entry *fib4_entry;
5689
5690         fib4_entry = container_of(fib_entry, struct mlxsw_sp_fib4_entry,
5691                                   common);
5692         return !fib4_entry->dscp;
5693 }
5694
5695 static bool
5696 mlxsw_sp_fib_entry_should_offload(const struct mlxsw_sp_fib_entry *fib_entry)
5697 {
5698         struct mlxsw_sp_nexthop_group *nh_group = fib_entry->nh_group;
5699
5700         switch (fib_entry->fib_node->fib->proto) {
5701         case MLXSW_SP_L3_PROTO_IPV4:
5702                 if (!mlxsw_sp_fib4_entry_should_offload(fib_entry))
5703                         return false;
5704                 break;
5705         case MLXSW_SP_L3_PROTO_IPV6:
5706                 break;
5707         }
5708
5709         switch (fib_entry->type) {
5710         case MLXSW_SP_FIB_ENTRY_TYPE_REMOTE:
5711                 return !!nh_group->nhgi->adj_index_valid;
5712         case MLXSW_SP_FIB_ENTRY_TYPE_LOCAL:
5713                 return !!mlxsw_sp_nhgi_rif(nh_group->nhgi);
5714         case MLXSW_SP_FIB_ENTRY_TYPE_BLACKHOLE:
5715         case MLXSW_SP_FIB_ENTRY_TYPE_IPIP_DECAP:
5716         case MLXSW_SP_FIB_ENTRY_TYPE_NVE_DECAP:
5717                 return true;
5718         default:
5719                 return false;
5720         }
5721 }
5722
5723 static struct mlxsw_sp_nexthop *
5724 mlxsw_sp_rt6_nexthop(struct mlxsw_sp_nexthop_group *nh_grp,
5725                      const struct mlxsw_sp_rt6 *mlxsw_sp_rt6)
5726 {
5727         int i;
5728
5729         for (i = 0; i < nh_grp->nhgi->count; i++) {
5730                 struct mlxsw_sp_nexthop *nh = &nh_grp->nhgi->nexthops[i];
5731                 struct net_device *dev = mlxsw_sp_nexthop_dev(nh);
5732                 struct fib6_info *rt = mlxsw_sp_rt6->rt;
5733
5734                 if (dev && dev == rt->fib6_nh->fib_nh_dev &&
5735                     ipv6_addr_equal((const struct in6_addr *) &nh->gw_addr,
5736                                     &rt->fib6_nh->fib_nh_gw6))
5737                         return nh;
5738         }
5739
5740         return NULL;
5741 }
5742
5743 static void
5744 mlxsw_sp_fib4_offload_failed_flag_set(struct mlxsw_sp *mlxsw_sp,
5745                                       struct fib_entry_notifier_info *fen_info)
5746 {
5747         u32 *p_dst = (u32 *) &fen_info->dst;
5748         struct fib_rt_info fri;
5749
5750         fri.fi = fen_info->fi;
5751         fri.tb_id = fen_info->tb_id;
5752         fri.dst = cpu_to_be32(*p_dst);
5753         fri.dst_len = fen_info->dst_len;
5754         fri.dscp = fen_info->dscp;
5755         fri.type = fen_info->type;
5756         fri.offload = false;
5757         fri.trap = false;
5758         fri.offload_failed = true;
5759         fib_alias_hw_flags_set(mlxsw_sp_net(mlxsw_sp), &fri);
5760 }
5761
5762 static void
5763 mlxsw_sp_fib4_entry_hw_flags_set(struct mlxsw_sp *mlxsw_sp,
5764                                  struct mlxsw_sp_fib_entry *fib_entry)
5765 {
5766         u32 *p_dst = (u32 *) fib_entry->fib_node->key.addr;
5767         int dst_len = fib_entry->fib_node->key.prefix_len;
5768         struct mlxsw_sp_fib4_entry *fib4_entry;
5769         struct fib_rt_info fri;
5770         bool should_offload;
5771
5772         should_offload = mlxsw_sp_fib_entry_should_offload(fib_entry);
5773         fib4_entry = container_of(fib_entry, struct mlxsw_sp_fib4_entry,
5774                                   common);
5775         fri.fi = fib4_entry->fi;
5776         fri.tb_id = fib4_entry->tb_id;
5777         fri.dst = cpu_to_be32(*p_dst);
5778         fri.dst_len = dst_len;
5779         fri.dscp = fib4_entry->dscp;
5780         fri.type = fib4_entry->type;
5781         fri.offload = should_offload;
5782         fri.trap = !should_offload;
5783         fri.offload_failed = false;
5784         fib_alias_hw_flags_set(mlxsw_sp_net(mlxsw_sp), &fri);
5785 }
5786
5787 static void
5788 mlxsw_sp_fib4_entry_hw_flags_clear(struct mlxsw_sp *mlxsw_sp,
5789                                    struct mlxsw_sp_fib_entry *fib_entry)
5790 {
5791         u32 *p_dst = (u32 *) fib_entry->fib_node->key.addr;
5792         int dst_len = fib_entry->fib_node->key.prefix_len;
5793         struct mlxsw_sp_fib4_entry *fib4_entry;
5794         struct fib_rt_info fri;
5795
5796         fib4_entry = container_of(fib_entry, struct mlxsw_sp_fib4_entry,
5797                                   common);
5798         fri.fi = fib4_entry->fi;
5799         fri.tb_id = fib4_entry->tb_id;
5800         fri.dst = cpu_to_be32(*p_dst);
5801         fri.dst_len = dst_len;
5802         fri.dscp = fib4_entry->dscp;
5803         fri.type = fib4_entry->type;
5804         fri.offload = false;
5805         fri.trap = false;
5806         fri.offload_failed = false;
5807         fib_alias_hw_flags_set(mlxsw_sp_net(mlxsw_sp), &fri);
5808 }
5809
5810 #if IS_ENABLED(CONFIG_IPV6)
5811 static void
5812 mlxsw_sp_fib6_offload_failed_flag_set(struct mlxsw_sp *mlxsw_sp,
5813                                       struct fib6_info **rt_arr,
5814                                       unsigned int nrt6)
5815 {
5816         int i;
5817
5818         /* In IPv6 a multipath route is represented using multiple routes, so
5819          * we need to set the flags on all of them.
5820          */
5821         for (i = 0; i < nrt6; i++)
5822                 fib6_info_hw_flags_set(mlxsw_sp_net(mlxsw_sp), rt_arr[i],
5823                                        false, false, true);
5824 }
5825 #else
5826 static void
5827 mlxsw_sp_fib6_offload_failed_flag_set(struct mlxsw_sp *mlxsw_sp,
5828                                       struct fib6_info **rt_arr,
5829                                       unsigned int nrt6)
5830 {
5831 }
5832 #endif
5833
5834 #if IS_ENABLED(CONFIG_IPV6)
5835 static void
5836 mlxsw_sp_fib6_entry_hw_flags_set(struct mlxsw_sp *mlxsw_sp,
5837                                  struct mlxsw_sp_fib_entry *fib_entry)
5838 {
5839         struct mlxsw_sp_fib6_entry *fib6_entry;
5840         struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
5841         bool should_offload;
5842
5843         should_offload = mlxsw_sp_fib_entry_should_offload(fib_entry);
5844
5845         /* In IPv6 a multipath route is represented using multiple routes, so
5846          * we need to set the flags on all of them.
5847          */
5848         fib6_entry = container_of(fib_entry, struct mlxsw_sp_fib6_entry,
5849                                   common);
5850         list_for_each_entry(mlxsw_sp_rt6, &fib6_entry->rt6_list, list)
5851                 fib6_info_hw_flags_set(mlxsw_sp_net(mlxsw_sp), mlxsw_sp_rt6->rt,
5852                                        should_offload, !should_offload, false);
5853 }
5854 #else
5855 static void
5856 mlxsw_sp_fib6_entry_hw_flags_set(struct mlxsw_sp *mlxsw_sp,
5857                                  struct mlxsw_sp_fib_entry *fib_entry)
5858 {
5859 }
5860 #endif
5861
5862 #if IS_ENABLED(CONFIG_IPV6)
5863 static void
5864 mlxsw_sp_fib6_entry_hw_flags_clear(struct mlxsw_sp *mlxsw_sp,
5865                                    struct mlxsw_sp_fib_entry *fib_entry)
5866 {
5867         struct mlxsw_sp_fib6_entry *fib6_entry;
5868         struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
5869
5870         fib6_entry = container_of(fib_entry, struct mlxsw_sp_fib6_entry,
5871                                   common);
5872         list_for_each_entry(mlxsw_sp_rt6, &fib6_entry->rt6_list, list)
5873                 fib6_info_hw_flags_set(mlxsw_sp_net(mlxsw_sp), mlxsw_sp_rt6->rt,
5874                                        false, false, false);
5875 }
5876 #else
5877 static void
5878 mlxsw_sp_fib6_entry_hw_flags_clear(struct mlxsw_sp *mlxsw_sp,
5879                                    struct mlxsw_sp_fib_entry *fib_entry)
5880 {
5881 }
5882 #endif
5883
5884 static void
5885 mlxsw_sp_fib_entry_hw_flags_set(struct mlxsw_sp *mlxsw_sp,
5886                                 struct mlxsw_sp_fib_entry *fib_entry)
5887 {
5888         switch (fib_entry->fib_node->fib->proto) {
5889         case MLXSW_SP_L3_PROTO_IPV4:
5890                 mlxsw_sp_fib4_entry_hw_flags_set(mlxsw_sp, fib_entry);
5891                 break;
5892         case MLXSW_SP_L3_PROTO_IPV6:
5893                 mlxsw_sp_fib6_entry_hw_flags_set(mlxsw_sp, fib_entry);
5894                 break;
5895         }
5896 }
5897
5898 static void
5899 mlxsw_sp_fib_entry_hw_flags_clear(struct mlxsw_sp *mlxsw_sp,
5900                                   struct mlxsw_sp_fib_entry *fib_entry)
5901 {
5902         switch (fib_entry->fib_node->fib->proto) {
5903         case MLXSW_SP_L3_PROTO_IPV4:
5904                 mlxsw_sp_fib4_entry_hw_flags_clear(mlxsw_sp, fib_entry);
5905                 break;
5906         case MLXSW_SP_L3_PROTO_IPV6:
5907                 mlxsw_sp_fib6_entry_hw_flags_clear(mlxsw_sp, fib_entry);
5908                 break;
5909         }
5910 }
5911
5912 static void
5913 mlxsw_sp_fib_entry_hw_flags_refresh(struct mlxsw_sp *mlxsw_sp,
5914                                     struct mlxsw_sp_fib_entry *fib_entry,
5915                                     enum mlxsw_reg_ralue_op op)
5916 {
5917         switch (op) {
5918         case MLXSW_REG_RALUE_OP_WRITE_WRITE:
5919                 mlxsw_sp_fib_entry_hw_flags_set(mlxsw_sp, fib_entry);
5920                 break;
5921         case MLXSW_REG_RALUE_OP_WRITE_DELETE:
5922                 mlxsw_sp_fib_entry_hw_flags_clear(mlxsw_sp, fib_entry);
5923                 break;
5924         default:
5925                 break;
5926         }
5927 }
5928
5929 static void
5930 mlxsw_sp_fib_entry_ralue_pack(char *ralue_pl,
5931                               const struct mlxsw_sp_fib_entry *fib_entry,
5932                               enum mlxsw_reg_ralue_op op)
5933 {
5934         struct mlxsw_sp_fib *fib = fib_entry->fib_node->fib;
5935         enum mlxsw_reg_ralxx_protocol proto;
5936         u32 *p_dip;
5937
5938         proto = (enum mlxsw_reg_ralxx_protocol) fib->proto;
5939
5940         switch (fib->proto) {
5941         case MLXSW_SP_L3_PROTO_IPV4:
5942                 p_dip = (u32 *) fib_entry->fib_node->key.addr;
5943                 mlxsw_reg_ralue_pack4(ralue_pl, proto, op, fib->vr->id,
5944                                       fib_entry->fib_node->key.prefix_len,
5945                                       *p_dip);
5946                 break;
5947         case MLXSW_SP_L3_PROTO_IPV6:
5948                 mlxsw_reg_ralue_pack6(ralue_pl, proto, op, fib->vr->id,
5949                                       fib_entry->fib_node->key.prefix_len,
5950                                       fib_entry->fib_node->key.addr);
5951                 break;
5952         }
5953 }
5954
5955 static int mlxsw_sp_fib_entry_op_remote(struct mlxsw_sp *mlxsw_sp,
5956                                         struct mlxsw_sp_fib_entry *fib_entry,
5957                                         enum mlxsw_reg_ralue_op op)
5958 {
5959         struct mlxsw_sp_nexthop_group *nh_group = fib_entry->nh_group;
5960         struct mlxsw_sp_nexthop_group_info *nhgi = nh_group->nhgi;
5961         char ralue_pl[MLXSW_REG_RALUE_LEN];
5962         enum mlxsw_reg_ralue_trap_action trap_action;
5963         u16 trap_id = 0;
5964         u32 adjacency_index = 0;
5965         u16 ecmp_size = 0;
5966
5967         /* In case the nexthop group adjacency index is valid, use it
5968          * with provided ECMP size. Otherwise, setup trap and pass
5969          * traffic to kernel.
5970          */
5971         if (mlxsw_sp_fib_entry_should_offload(fib_entry)) {
5972                 trap_action = MLXSW_REG_RALUE_TRAP_ACTION_NOP;
5973                 adjacency_index = nhgi->adj_index;
5974                 ecmp_size = nhgi->ecmp_size;
5975         } else if (!nhgi->adj_index_valid && nhgi->count &&
5976                    mlxsw_sp_nhgi_rif(nhgi)) {
5977                 trap_action = MLXSW_REG_RALUE_TRAP_ACTION_NOP;
5978                 adjacency_index = mlxsw_sp->router->adj_trap_index;
5979                 ecmp_size = 1;
5980         } else {
5981                 trap_action = MLXSW_REG_RALUE_TRAP_ACTION_TRAP;
5982                 trap_id = MLXSW_TRAP_ID_RTR_INGRESS0;
5983         }
5984
5985         mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
5986         mlxsw_reg_ralue_act_remote_pack(ralue_pl, trap_action, trap_id,
5987                                         adjacency_index, ecmp_size);
5988         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralue), ralue_pl);
5989 }
5990
5991 static int mlxsw_sp_fib_entry_op_local(struct mlxsw_sp *mlxsw_sp,
5992                                        struct mlxsw_sp_fib_entry *fib_entry,
5993                                        enum mlxsw_reg_ralue_op op)
5994 {
5995         struct mlxsw_sp_rif *rif = mlxsw_sp_nhgi_rif(fib_entry->nh_group->nhgi);
5996         enum mlxsw_reg_ralue_trap_action trap_action;
5997         char ralue_pl[MLXSW_REG_RALUE_LEN];
5998         u16 trap_id = 0;
5999         u16 rif_index = 0;
6000
6001         if (mlxsw_sp_fib_entry_should_offload(fib_entry)) {
6002                 trap_action = MLXSW_REG_RALUE_TRAP_ACTION_NOP;
6003                 rif_index = rif->rif_index;
6004         } else {
6005                 trap_action = MLXSW_REG_RALUE_TRAP_ACTION_TRAP;
6006                 trap_id = MLXSW_TRAP_ID_RTR_INGRESS0;
6007         }
6008
6009         mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
6010         mlxsw_reg_ralue_act_local_pack(ralue_pl, trap_action, trap_id,
6011                                        rif_index);
6012         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralue), ralue_pl);
6013 }
6014
6015 static int mlxsw_sp_fib_entry_op_trap(struct mlxsw_sp *mlxsw_sp,
6016                                       struct mlxsw_sp_fib_entry *fib_entry,
6017                                       enum mlxsw_reg_ralue_op op)
6018 {
6019         char ralue_pl[MLXSW_REG_RALUE_LEN];
6020
6021         mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
6022         mlxsw_reg_ralue_act_ip2me_pack(ralue_pl);
6023         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralue), ralue_pl);
6024 }
6025
6026 static int mlxsw_sp_fib_entry_op_blackhole(struct mlxsw_sp *mlxsw_sp,
6027                                            struct mlxsw_sp_fib_entry *fib_entry,
6028                                            enum mlxsw_reg_ralue_op op)
6029 {
6030         enum mlxsw_reg_ralue_trap_action trap_action;
6031         char ralue_pl[MLXSW_REG_RALUE_LEN];
6032
6033         trap_action = MLXSW_REG_RALUE_TRAP_ACTION_DISCARD_ERROR;
6034         mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
6035         mlxsw_reg_ralue_act_local_pack(ralue_pl, trap_action, 0, 0);
6036         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralue), ralue_pl);
6037 }
6038
6039 static int
6040 mlxsw_sp_fib_entry_op_unreachable(struct mlxsw_sp *mlxsw_sp,
6041                                   struct mlxsw_sp_fib_entry *fib_entry,
6042                                   enum mlxsw_reg_ralue_op op)
6043 {
6044         enum mlxsw_reg_ralue_trap_action trap_action;
6045         char ralue_pl[MLXSW_REG_RALUE_LEN];
6046         u16 trap_id;
6047
6048         trap_action = MLXSW_REG_RALUE_TRAP_ACTION_TRAP;
6049         trap_id = MLXSW_TRAP_ID_RTR_INGRESS1;
6050
6051         mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
6052         mlxsw_reg_ralue_act_local_pack(ralue_pl, trap_action, trap_id, 0);
6053         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralue), ralue_pl);
6054 }
6055
6056 static int
6057 mlxsw_sp_fib_entry_op_ipip_decap(struct mlxsw_sp *mlxsw_sp,
6058                                  struct mlxsw_sp_fib_entry *fib_entry,
6059                                  enum mlxsw_reg_ralue_op op)
6060 {
6061         struct mlxsw_sp_ipip_entry *ipip_entry = fib_entry->decap.ipip_entry;
6062         const struct mlxsw_sp_ipip_ops *ipip_ops;
6063         char ralue_pl[MLXSW_REG_RALUE_LEN];
6064         int err;
6065
6066         if (WARN_ON(!ipip_entry))
6067                 return -EINVAL;
6068
6069         ipip_ops = mlxsw_sp->router->ipip_ops_arr[ipip_entry->ipipt];
6070         err = ipip_ops->decap_config(mlxsw_sp, ipip_entry,
6071                                      fib_entry->decap.tunnel_index);
6072         if (err)
6073                 return err;
6074
6075         mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
6076         mlxsw_reg_ralue_act_ip2me_tun_pack(ralue_pl,
6077                                            fib_entry->decap.tunnel_index);
6078         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralue), ralue_pl);
6079 }
6080
6081 static int mlxsw_sp_fib_entry_op_nve_decap(struct mlxsw_sp *mlxsw_sp,
6082                                            struct mlxsw_sp_fib_entry *fib_entry,
6083                                            enum mlxsw_reg_ralue_op op)
6084 {
6085         char ralue_pl[MLXSW_REG_RALUE_LEN];
6086
6087         mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
6088         mlxsw_reg_ralue_act_ip2me_tun_pack(ralue_pl,
6089                                            fib_entry->decap.tunnel_index);
6090         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralue), ralue_pl);
6091 }
6092
6093 static int __mlxsw_sp_fib_entry_op(struct mlxsw_sp *mlxsw_sp,
6094                                    struct mlxsw_sp_fib_entry *fib_entry,
6095                                    enum mlxsw_reg_ralue_op op)
6096 {
6097         switch (fib_entry->type) {
6098         case MLXSW_SP_FIB_ENTRY_TYPE_REMOTE:
6099                 return mlxsw_sp_fib_entry_op_remote(mlxsw_sp, fib_entry, op);
6100         case MLXSW_SP_FIB_ENTRY_TYPE_LOCAL:
6101                 return mlxsw_sp_fib_entry_op_local(mlxsw_sp, fib_entry, op);
6102         case MLXSW_SP_FIB_ENTRY_TYPE_TRAP:
6103                 return mlxsw_sp_fib_entry_op_trap(mlxsw_sp, fib_entry, op);
6104         case MLXSW_SP_FIB_ENTRY_TYPE_BLACKHOLE:
6105                 return mlxsw_sp_fib_entry_op_blackhole(mlxsw_sp, fib_entry, op);
6106         case MLXSW_SP_FIB_ENTRY_TYPE_UNREACHABLE:
6107                 return mlxsw_sp_fib_entry_op_unreachable(mlxsw_sp, fib_entry,
6108                                                          op);
6109         case MLXSW_SP_FIB_ENTRY_TYPE_IPIP_DECAP:
6110                 return mlxsw_sp_fib_entry_op_ipip_decap(mlxsw_sp,
6111                                                         fib_entry, op);
6112         case MLXSW_SP_FIB_ENTRY_TYPE_NVE_DECAP:
6113                 return mlxsw_sp_fib_entry_op_nve_decap(mlxsw_sp, fib_entry, op);
6114         }
6115         return -EINVAL;
6116 }
6117
6118 static int mlxsw_sp_fib_entry_op(struct mlxsw_sp *mlxsw_sp,
6119                                  struct mlxsw_sp_fib_entry *fib_entry,
6120                                  enum mlxsw_reg_ralue_op op)
6121 {
6122         int err = __mlxsw_sp_fib_entry_op(mlxsw_sp, fib_entry, op);
6123
6124         if (err)
6125                 return err;
6126
6127         mlxsw_sp_fib_entry_hw_flags_refresh(mlxsw_sp, fib_entry, op);
6128
6129         return err;
6130 }
6131
6132 static int mlxsw_sp_fib_entry_update(struct mlxsw_sp *mlxsw_sp,
6133                                      struct mlxsw_sp_fib_entry *fib_entry)
6134 {
6135         return mlxsw_sp_fib_entry_op(mlxsw_sp, fib_entry,
6136                                      MLXSW_REG_RALUE_OP_WRITE_WRITE);
6137 }
6138
6139 static int mlxsw_sp_fib_entry_del(struct mlxsw_sp *mlxsw_sp,
6140                                   struct mlxsw_sp_fib_entry *fib_entry)
6141 {
6142         return mlxsw_sp_fib_entry_op(mlxsw_sp, fib_entry,
6143                                      MLXSW_REG_RALUE_OP_WRITE_DELETE);
6144 }
6145
6146 static int
6147 mlxsw_sp_fib4_entry_type_set(struct mlxsw_sp *mlxsw_sp,
6148                              const struct fib_entry_notifier_info *fen_info,
6149                              struct mlxsw_sp_fib_entry *fib_entry)
6150 {
6151         struct mlxsw_sp_nexthop_group_info *nhgi = fib_entry->nh_group->nhgi;
6152         union mlxsw_sp_l3addr dip = { .addr4 = htonl(fen_info->dst) };
6153         struct mlxsw_sp_router *router = mlxsw_sp->router;
6154         u32 tb_id = mlxsw_sp_fix_tb_id(fen_info->tb_id);
6155         int ifindex = nhgi->nexthops[0].ifindex;
6156         struct mlxsw_sp_ipip_entry *ipip_entry;
6157
6158         switch (fen_info->type) {
6159         case RTN_LOCAL:
6160                 ipip_entry = mlxsw_sp_ipip_entry_find_by_decap(mlxsw_sp, ifindex,
6161                                                                MLXSW_SP_L3_PROTO_IPV4, dip);
6162                 if (ipip_entry && ipip_entry->ol_dev->flags & IFF_UP) {
6163                         fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_IPIP_DECAP;
6164                         return mlxsw_sp_fib_entry_decap_init(mlxsw_sp,
6165                                                              fib_entry,
6166                                                              ipip_entry);
6167                 }
6168                 if (mlxsw_sp_router_nve_is_decap(mlxsw_sp, tb_id,
6169                                                  MLXSW_SP_L3_PROTO_IPV4,
6170                                                  &dip)) {
6171                         u32 tunnel_index;
6172
6173                         tunnel_index = router->nve_decap_config.tunnel_index;
6174                         fib_entry->decap.tunnel_index = tunnel_index;
6175                         fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_NVE_DECAP;
6176                         return 0;
6177                 }
6178                 fallthrough;
6179         case RTN_BROADCAST:
6180                 fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_TRAP;
6181                 return 0;
6182         case RTN_BLACKHOLE:
6183                 fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_BLACKHOLE;
6184                 return 0;
6185         case RTN_UNREACHABLE:
6186         case RTN_PROHIBIT:
6187                 /* Packets hitting these routes need to be trapped, but
6188                  * can do so with a lower priority than packets directed
6189                  * at the host, so use action type local instead of trap.
6190                  */
6191                 fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_UNREACHABLE;
6192                 return 0;
6193         case RTN_UNICAST:
6194                 if (nhgi->gateway)
6195                         fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_REMOTE;
6196                 else
6197                         fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_LOCAL;
6198                 return 0;
6199         default:
6200                 return -EINVAL;
6201         }
6202 }
6203
6204 static void
6205 mlxsw_sp_fib_entry_type_unset(struct mlxsw_sp *mlxsw_sp,
6206                               struct mlxsw_sp_fib_entry *fib_entry)
6207 {
6208         switch (fib_entry->type) {
6209         case MLXSW_SP_FIB_ENTRY_TYPE_IPIP_DECAP:
6210                 mlxsw_sp_fib_entry_decap_fini(mlxsw_sp, fib_entry);
6211                 break;
6212         default:
6213                 break;
6214         }
6215 }
6216
6217 static void
6218 mlxsw_sp_fib4_entry_type_unset(struct mlxsw_sp *mlxsw_sp,
6219                                struct mlxsw_sp_fib4_entry *fib4_entry)
6220 {
6221         mlxsw_sp_fib_entry_type_unset(mlxsw_sp, &fib4_entry->common);
6222 }
6223
6224 static struct mlxsw_sp_fib4_entry *
6225 mlxsw_sp_fib4_entry_create(struct mlxsw_sp *mlxsw_sp,
6226                            struct mlxsw_sp_fib_node *fib_node,
6227                            const struct fib_entry_notifier_info *fen_info)
6228 {
6229         struct mlxsw_sp_fib4_entry *fib4_entry;
6230         struct mlxsw_sp_fib_entry *fib_entry;
6231         int err;
6232
6233         fib4_entry = kzalloc(sizeof(*fib4_entry), GFP_KERNEL);
6234         if (!fib4_entry)
6235                 return ERR_PTR(-ENOMEM);
6236         fib_entry = &fib4_entry->common;
6237
6238         err = mlxsw_sp_nexthop4_group_get(mlxsw_sp, fib_entry, fen_info->fi);
6239         if (err)
6240                 goto err_nexthop4_group_get;
6241
6242         err = mlxsw_sp_nexthop_group_vr_link(fib_entry->nh_group,
6243                                              fib_node->fib);
6244         if (err)
6245                 goto err_nexthop_group_vr_link;
6246
6247         err = mlxsw_sp_fib4_entry_type_set(mlxsw_sp, fen_info, fib_entry);
6248         if (err)
6249                 goto err_fib4_entry_type_set;
6250
6251         fib4_entry->fi = fen_info->fi;
6252         fib_info_hold(fib4_entry->fi);
6253         fib4_entry->tb_id = fen_info->tb_id;
6254         fib4_entry->type = fen_info->type;
6255         fib4_entry->dscp = fen_info->dscp;
6256
6257         fib_entry->fib_node = fib_node;
6258
6259         return fib4_entry;
6260
6261 err_fib4_entry_type_set:
6262         mlxsw_sp_nexthop_group_vr_unlink(fib_entry->nh_group, fib_node->fib);
6263 err_nexthop_group_vr_link:
6264         mlxsw_sp_nexthop4_group_put(mlxsw_sp, &fib4_entry->common);
6265 err_nexthop4_group_get:
6266         kfree(fib4_entry);
6267         return ERR_PTR(err);
6268 }
6269
6270 static void mlxsw_sp_fib4_entry_destroy(struct mlxsw_sp *mlxsw_sp,
6271                                         struct mlxsw_sp_fib4_entry *fib4_entry)
6272 {
6273         struct mlxsw_sp_fib_node *fib_node = fib4_entry->common.fib_node;
6274
6275         fib_info_put(fib4_entry->fi);
6276         mlxsw_sp_fib4_entry_type_unset(mlxsw_sp, fib4_entry);
6277         mlxsw_sp_nexthop_group_vr_unlink(fib4_entry->common.nh_group,
6278                                          fib_node->fib);
6279         mlxsw_sp_nexthop4_group_put(mlxsw_sp, &fib4_entry->common);
6280         kfree(fib4_entry);
6281 }
6282
6283 static struct mlxsw_sp_fib4_entry *
6284 mlxsw_sp_fib4_entry_lookup(struct mlxsw_sp *mlxsw_sp,
6285                            const struct fib_entry_notifier_info *fen_info)
6286 {
6287         struct mlxsw_sp_fib4_entry *fib4_entry;
6288         struct mlxsw_sp_fib_node *fib_node;
6289         struct mlxsw_sp_fib *fib;
6290         struct mlxsw_sp_vr *vr;
6291
6292         vr = mlxsw_sp_vr_find(mlxsw_sp, fen_info->tb_id);
6293         if (!vr)
6294                 return NULL;
6295         fib = mlxsw_sp_vr_fib(vr, MLXSW_SP_L3_PROTO_IPV4);
6296
6297         fib_node = mlxsw_sp_fib_node_lookup(fib, &fen_info->dst,
6298                                             sizeof(fen_info->dst),
6299                                             fen_info->dst_len);
6300         if (!fib_node)
6301                 return NULL;
6302
6303         fib4_entry = container_of(fib_node->fib_entry,
6304                                   struct mlxsw_sp_fib4_entry, common);
6305         if (fib4_entry->tb_id == fen_info->tb_id &&
6306             fib4_entry->dscp == fen_info->dscp &&
6307             fib4_entry->type == fen_info->type &&
6308             fib4_entry->fi == fen_info->fi)
6309                 return fib4_entry;
6310
6311         return NULL;
6312 }
6313
6314 static const struct rhashtable_params mlxsw_sp_fib_ht_params = {
6315         .key_offset = offsetof(struct mlxsw_sp_fib_node, key),
6316         .head_offset = offsetof(struct mlxsw_sp_fib_node, ht_node),
6317         .key_len = sizeof(struct mlxsw_sp_fib_key),
6318         .automatic_shrinking = true,
6319 };
6320
6321 static int mlxsw_sp_fib_node_insert(struct mlxsw_sp_fib *fib,
6322                                     struct mlxsw_sp_fib_node *fib_node)
6323 {
6324         return rhashtable_insert_fast(&fib->ht, &fib_node->ht_node,
6325                                       mlxsw_sp_fib_ht_params);
6326 }
6327
6328 static void mlxsw_sp_fib_node_remove(struct mlxsw_sp_fib *fib,
6329                                      struct mlxsw_sp_fib_node *fib_node)
6330 {
6331         rhashtable_remove_fast(&fib->ht, &fib_node->ht_node,
6332                                mlxsw_sp_fib_ht_params);
6333 }
6334
6335 static struct mlxsw_sp_fib_node *
6336 mlxsw_sp_fib_node_lookup(struct mlxsw_sp_fib *fib, const void *addr,
6337                          size_t addr_len, unsigned char prefix_len)
6338 {
6339         struct mlxsw_sp_fib_key key;
6340
6341         memset(&key, 0, sizeof(key));
6342         memcpy(key.addr, addr, addr_len);
6343         key.prefix_len = prefix_len;
6344         return rhashtable_lookup_fast(&fib->ht, &key, mlxsw_sp_fib_ht_params);
6345 }
6346
6347 static struct mlxsw_sp_fib_node *
6348 mlxsw_sp_fib_node_create(struct mlxsw_sp_fib *fib, const void *addr,
6349                          size_t addr_len, unsigned char prefix_len)
6350 {
6351         struct mlxsw_sp_fib_node *fib_node;
6352
6353         fib_node = kzalloc(sizeof(*fib_node), GFP_KERNEL);
6354         if (!fib_node)
6355                 return NULL;
6356
6357         list_add(&fib_node->list, &fib->node_list);
6358         memcpy(fib_node->key.addr, addr, addr_len);
6359         fib_node->key.prefix_len = prefix_len;
6360
6361         return fib_node;
6362 }
6363
6364 static void mlxsw_sp_fib_node_destroy(struct mlxsw_sp_fib_node *fib_node)
6365 {
6366         list_del(&fib_node->list);
6367         kfree(fib_node);
6368 }
6369
6370 static int mlxsw_sp_fib_lpm_tree_link(struct mlxsw_sp *mlxsw_sp,
6371                                       struct mlxsw_sp_fib_node *fib_node)
6372 {
6373         struct mlxsw_sp_prefix_usage req_prefix_usage;
6374         struct mlxsw_sp_fib *fib = fib_node->fib;
6375         struct mlxsw_sp_lpm_tree *lpm_tree;
6376         int err;
6377
6378         lpm_tree = mlxsw_sp->router->lpm.proto_trees[fib->proto];
6379         if (lpm_tree->prefix_ref_count[fib_node->key.prefix_len] != 0)
6380                 goto out;
6381
6382         mlxsw_sp_prefix_usage_cpy(&req_prefix_usage, &lpm_tree->prefix_usage);
6383         mlxsw_sp_prefix_usage_set(&req_prefix_usage, fib_node->key.prefix_len);
6384         lpm_tree = mlxsw_sp_lpm_tree_get(mlxsw_sp, &req_prefix_usage,
6385                                          fib->proto);
6386         if (IS_ERR(lpm_tree))
6387                 return PTR_ERR(lpm_tree);
6388
6389         err = mlxsw_sp_vrs_lpm_tree_replace(mlxsw_sp, fib, lpm_tree);
6390         if (err)
6391                 goto err_lpm_tree_replace;
6392
6393 out:
6394         lpm_tree->prefix_ref_count[fib_node->key.prefix_len]++;
6395         return 0;
6396
6397 err_lpm_tree_replace:
6398         mlxsw_sp_lpm_tree_put(mlxsw_sp, lpm_tree);
6399         return err;
6400 }
6401
6402 static void mlxsw_sp_fib_lpm_tree_unlink(struct mlxsw_sp *mlxsw_sp,
6403                                          struct mlxsw_sp_fib_node *fib_node)
6404 {
6405         struct mlxsw_sp_lpm_tree *lpm_tree = fib_node->fib->lpm_tree;
6406         struct mlxsw_sp_prefix_usage req_prefix_usage;
6407         struct mlxsw_sp_fib *fib = fib_node->fib;
6408         int err;
6409
6410         if (--lpm_tree->prefix_ref_count[fib_node->key.prefix_len] != 0)
6411                 return;
6412         /* Try to construct a new LPM tree from the current prefix usage
6413          * minus the unused one. If we fail, continue using the old one.
6414          */
6415         mlxsw_sp_prefix_usage_cpy(&req_prefix_usage, &lpm_tree->prefix_usage);
6416         mlxsw_sp_prefix_usage_clear(&req_prefix_usage,
6417                                     fib_node->key.prefix_len);
6418         lpm_tree = mlxsw_sp_lpm_tree_get(mlxsw_sp, &req_prefix_usage,
6419                                          fib->proto);
6420         if (IS_ERR(lpm_tree))
6421                 return;
6422
6423         err = mlxsw_sp_vrs_lpm_tree_replace(mlxsw_sp, fib, lpm_tree);
6424         if (err)
6425                 goto err_lpm_tree_replace;
6426
6427         return;
6428
6429 err_lpm_tree_replace:
6430         mlxsw_sp_lpm_tree_put(mlxsw_sp, lpm_tree);
6431 }
6432
6433 static int mlxsw_sp_fib_node_init(struct mlxsw_sp *mlxsw_sp,
6434                                   struct mlxsw_sp_fib_node *fib_node,
6435                                   struct mlxsw_sp_fib *fib)
6436 {
6437         int err;
6438
6439         err = mlxsw_sp_fib_node_insert(fib, fib_node);
6440         if (err)
6441                 return err;
6442         fib_node->fib = fib;
6443
6444         err = mlxsw_sp_fib_lpm_tree_link(mlxsw_sp, fib_node);
6445         if (err)
6446                 goto err_fib_lpm_tree_link;
6447
6448         return 0;
6449
6450 err_fib_lpm_tree_link:
6451         fib_node->fib = NULL;
6452         mlxsw_sp_fib_node_remove(fib, fib_node);
6453         return err;
6454 }
6455
6456 static void mlxsw_sp_fib_node_fini(struct mlxsw_sp *mlxsw_sp,
6457                                    struct mlxsw_sp_fib_node *fib_node)
6458 {
6459         struct mlxsw_sp_fib *fib = fib_node->fib;
6460
6461         mlxsw_sp_fib_lpm_tree_unlink(mlxsw_sp, fib_node);
6462         fib_node->fib = NULL;
6463         mlxsw_sp_fib_node_remove(fib, fib_node);
6464 }
6465
6466 static struct mlxsw_sp_fib_node *
6467 mlxsw_sp_fib_node_get(struct mlxsw_sp *mlxsw_sp, u32 tb_id, const void *addr,
6468                       size_t addr_len, unsigned char prefix_len,
6469                       enum mlxsw_sp_l3proto proto)
6470 {
6471         struct mlxsw_sp_fib_node *fib_node;
6472         struct mlxsw_sp_fib *fib;
6473         struct mlxsw_sp_vr *vr;
6474         int err;
6475
6476         vr = mlxsw_sp_vr_get(mlxsw_sp, tb_id, NULL);
6477         if (IS_ERR(vr))
6478                 return ERR_CAST(vr);
6479         fib = mlxsw_sp_vr_fib(vr, proto);
6480
6481         fib_node = mlxsw_sp_fib_node_lookup(fib, addr, addr_len, prefix_len);
6482         if (fib_node)
6483                 return fib_node;
6484
6485         fib_node = mlxsw_sp_fib_node_create(fib, addr, addr_len, prefix_len);
6486         if (!fib_node) {
6487                 err = -ENOMEM;
6488                 goto err_fib_node_create;
6489         }
6490
6491         err = mlxsw_sp_fib_node_init(mlxsw_sp, fib_node, fib);
6492         if (err)
6493                 goto err_fib_node_init;
6494
6495         return fib_node;
6496
6497 err_fib_node_init:
6498         mlxsw_sp_fib_node_destroy(fib_node);
6499 err_fib_node_create:
6500         mlxsw_sp_vr_put(mlxsw_sp, vr);
6501         return ERR_PTR(err);
6502 }
6503
6504 static void mlxsw_sp_fib_node_put(struct mlxsw_sp *mlxsw_sp,
6505                                   struct mlxsw_sp_fib_node *fib_node)
6506 {
6507         struct mlxsw_sp_vr *vr = fib_node->fib->vr;
6508
6509         if (fib_node->fib_entry)
6510                 return;
6511         mlxsw_sp_fib_node_fini(mlxsw_sp, fib_node);
6512         mlxsw_sp_fib_node_destroy(fib_node);
6513         mlxsw_sp_vr_put(mlxsw_sp, vr);
6514 }
6515
6516 static int mlxsw_sp_fib_node_entry_link(struct mlxsw_sp *mlxsw_sp,
6517                                         struct mlxsw_sp_fib_entry *fib_entry)
6518 {
6519         struct mlxsw_sp_fib_node *fib_node = fib_entry->fib_node;
6520         int err;
6521
6522         fib_node->fib_entry = fib_entry;
6523
6524         err = mlxsw_sp_fib_entry_update(mlxsw_sp, fib_entry);
6525         if (err)
6526                 goto err_fib_entry_update;
6527
6528         return 0;
6529
6530 err_fib_entry_update:
6531         fib_node->fib_entry = NULL;
6532         return err;
6533 }
6534
6535 static void
6536 mlxsw_sp_fib_node_entry_unlink(struct mlxsw_sp *mlxsw_sp,
6537                                struct mlxsw_sp_fib_entry *fib_entry)
6538 {
6539         struct mlxsw_sp_fib_node *fib_node = fib_entry->fib_node;
6540
6541         mlxsw_sp_fib_entry_del(mlxsw_sp, fib_entry);
6542         fib_node->fib_entry = NULL;
6543 }
6544
6545 static bool mlxsw_sp_fib4_allow_replace(struct mlxsw_sp_fib4_entry *fib4_entry)
6546 {
6547         struct mlxsw_sp_fib_node *fib_node = fib4_entry->common.fib_node;
6548         struct mlxsw_sp_fib4_entry *fib4_replaced;
6549
6550         if (!fib_node->fib_entry)
6551                 return true;
6552
6553         fib4_replaced = container_of(fib_node->fib_entry,
6554                                      struct mlxsw_sp_fib4_entry, common);
6555         if (fib4_entry->tb_id == RT_TABLE_MAIN &&
6556             fib4_replaced->tb_id == RT_TABLE_LOCAL)
6557                 return false;
6558
6559         return true;
6560 }
6561
6562 static int
6563 mlxsw_sp_router_fib4_replace(struct mlxsw_sp *mlxsw_sp,
6564                              const struct fib_entry_notifier_info *fen_info)
6565 {
6566         struct mlxsw_sp_fib4_entry *fib4_entry, *fib4_replaced;
6567         struct mlxsw_sp_fib_entry *replaced;
6568         struct mlxsw_sp_fib_node *fib_node;
6569         int err;
6570
6571         if (fen_info->fi->nh &&
6572             !mlxsw_sp_nexthop_obj_group_lookup(mlxsw_sp, fen_info->fi->nh->id))
6573                 return 0;
6574
6575         fib_node = mlxsw_sp_fib_node_get(mlxsw_sp, fen_info->tb_id,
6576                                          &fen_info->dst, sizeof(fen_info->dst),
6577                                          fen_info->dst_len,
6578                                          MLXSW_SP_L3_PROTO_IPV4);
6579         if (IS_ERR(fib_node)) {
6580                 dev_warn(mlxsw_sp->bus_info->dev, "Failed to get FIB node\n");
6581                 return PTR_ERR(fib_node);
6582         }
6583
6584         fib4_entry = mlxsw_sp_fib4_entry_create(mlxsw_sp, fib_node, fen_info);
6585         if (IS_ERR(fib4_entry)) {
6586                 dev_warn(mlxsw_sp->bus_info->dev, "Failed to create FIB entry\n");
6587                 err = PTR_ERR(fib4_entry);
6588                 goto err_fib4_entry_create;
6589         }
6590
6591         if (!mlxsw_sp_fib4_allow_replace(fib4_entry)) {
6592                 mlxsw_sp_fib4_entry_destroy(mlxsw_sp, fib4_entry);
6593                 mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
6594                 return 0;
6595         }
6596
6597         replaced = fib_node->fib_entry;
6598         err = mlxsw_sp_fib_node_entry_link(mlxsw_sp, &fib4_entry->common);
6599         if (err) {
6600                 dev_warn(mlxsw_sp->bus_info->dev, "Failed to link FIB entry to node\n");
6601                 goto err_fib_node_entry_link;
6602         }
6603
6604         /* Nothing to replace */
6605         if (!replaced)
6606                 return 0;
6607
6608         mlxsw_sp_fib_entry_hw_flags_clear(mlxsw_sp, replaced);
6609         fib4_replaced = container_of(replaced, struct mlxsw_sp_fib4_entry,
6610                                      common);
6611         mlxsw_sp_fib4_entry_destroy(mlxsw_sp, fib4_replaced);
6612
6613         return 0;
6614
6615 err_fib_node_entry_link:
6616         fib_node->fib_entry = replaced;
6617         mlxsw_sp_fib4_entry_destroy(mlxsw_sp, fib4_entry);
6618 err_fib4_entry_create:
6619         mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
6620         return err;
6621 }
6622
6623 static void mlxsw_sp_router_fib4_del(struct mlxsw_sp *mlxsw_sp,
6624                                      struct fib_entry_notifier_info *fen_info)
6625 {
6626         struct mlxsw_sp_fib4_entry *fib4_entry;
6627         struct mlxsw_sp_fib_node *fib_node;
6628
6629         fib4_entry = mlxsw_sp_fib4_entry_lookup(mlxsw_sp, fen_info);
6630         if (!fib4_entry)
6631                 return;
6632         fib_node = fib4_entry->common.fib_node;
6633
6634         mlxsw_sp_fib_node_entry_unlink(mlxsw_sp, &fib4_entry->common);
6635         mlxsw_sp_fib4_entry_destroy(mlxsw_sp, fib4_entry);
6636         mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
6637 }
6638
6639 static bool mlxsw_sp_fib6_rt_should_ignore(const struct fib6_info *rt)
6640 {
6641         /* Multicast routes aren't supported, so ignore them. Neighbour
6642          * Discovery packets are specifically trapped.
6643          */
6644         if (ipv6_addr_type(&rt->fib6_dst.addr) & IPV6_ADDR_MULTICAST)
6645                 return true;
6646
6647         /* Cloned routes are irrelevant in the forwarding path. */
6648         if (rt->fib6_flags & RTF_CACHE)
6649                 return true;
6650
6651         return false;
6652 }
6653
6654 static struct mlxsw_sp_rt6 *mlxsw_sp_rt6_create(struct fib6_info *rt)
6655 {
6656         struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
6657
6658         mlxsw_sp_rt6 = kzalloc(sizeof(*mlxsw_sp_rt6), GFP_KERNEL);
6659         if (!mlxsw_sp_rt6)
6660                 return ERR_PTR(-ENOMEM);
6661
6662         /* In case of route replace, replaced route is deleted with
6663          * no notification. Take reference to prevent accessing freed
6664          * memory.
6665          */
6666         mlxsw_sp_rt6->rt = rt;
6667         fib6_info_hold(rt);
6668
6669         return mlxsw_sp_rt6;
6670 }
6671
6672 #if IS_ENABLED(CONFIG_IPV6)
6673 static void mlxsw_sp_rt6_release(struct fib6_info *rt)
6674 {
6675         fib6_info_release(rt);
6676 }
6677 #else
6678 static void mlxsw_sp_rt6_release(struct fib6_info *rt)
6679 {
6680 }
6681 #endif
6682
6683 static void mlxsw_sp_rt6_destroy(struct mlxsw_sp_rt6 *mlxsw_sp_rt6)
6684 {
6685         struct fib6_nh *fib6_nh = mlxsw_sp_rt6->rt->fib6_nh;
6686
6687         if (!mlxsw_sp_rt6->rt->nh)
6688                 fib6_nh->fib_nh_flags &= ~RTNH_F_OFFLOAD;
6689         mlxsw_sp_rt6_release(mlxsw_sp_rt6->rt);
6690         kfree(mlxsw_sp_rt6);
6691 }
6692
6693 static struct fib6_info *
6694 mlxsw_sp_fib6_entry_rt(const struct mlxsw_sp_fib6_entry *fib6_entry)
6695 {
6696         return list_first_entry(&fib6_entry->rt6_list, struct mlxsw_sp_rt6,
6697                                 list)->rt;
6698 }
6699
6700 static struct mlxsw_sp_rt6 *
6701 mlxsw_sp_fib6_entry_rt_find(const struct mlxsw_sp_fib6_entry *fib6_entry,
6702                             const struct fib6_info *rt)
6703 {
6704         struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
6705
6706         list_for_each_entry(mlxsw_sp_rt6, &fib6_entry->rt6_list, list) {
6707                 if (mlxsw_sp_rt6->rt == rt)
6708                         return mlxsw_sp_rt6;
6709         }
6710
6711         return NULL;
6712 }
6713
6714 static bool mlxsw_sp_nexthop6_ipip_type(const struct mlxsw_sp *mlxsw_sp,
6715                                         const struct fib6_info *rt,
6716                                         enum mlxsw_sp_ipip_type *ret)
6717 {
6718         return rt->fib6_nh->fib_nh_dev &&
6719                mlxsw_sp_netdev_ipip_type(mlxsw_sp, rt->fib6_nh->fib_nh_dev, ret);
6720 }
6721
6722 static int mlxsw_sp_nexthop6_init(struct mlxsw_sp *mlxsw_sp,
6723                                   struct mlxsw_sp_nexthop_group *nh_grp,
6724                                   struct mlxsw_sp_nexthop *nh,
6725                                   const struct fib6_info *rt)
6726 {
6727         struct net_device *dev = rt->fib6_nh->fib_nh_dev;
6728         int err;
6729
6730         nh->nhgi = nh_grp->nhgi;
6731         nh->nh_weight = rt->fib6_nh->fib_nh_weight;
6732         memcpy(&nh->gw_addr, &rt->fib6_nh->fib_nh_gw6, sizeof(nh->gw_addr));
6733 #if IS_ENABLED(CONFIG_IPV6)
6734         nh->neigh_tbl = &nd_tbl;
6735 #endif
6736         mlxsw_sp_nexthop_counter_alloc(mlxsw_sp, nh);
6737
6738         list_add_tail(&nh->router_list_node, &mlxsw_sp->router->nexthop_list);
6739
6740         if (!dev)
6741                 return 0;
6742         nh->ifindex = dev->ifindex;
6743
6744         err = mlxsw_sp_nexthop_type_init(mlxsw_sp, nh, dev);
6745         if (err)
6746                 goto err_nexthop_type_init;
6747
6748         return 0;
6749
6750 err_nexthop_type_init:
6751         list_del(&nh->router_list_node);
6752         mlxsw_sp_nexthop_counter_free(mlxsw_sp, nh);
6753         return err;
6754 }
6755
6756 static void mlxsw_sp_nexthop6_fini(struct mlxsw_sp *mlxsw_sp,
6757                                    struct mlxsw_sp_nexthop *nh)
6758 {
6759         mlxsw_sp_nexthop_type_fini(mlxsw_sp, nh);
6760         list_del(&nh->router_list_node);
6761         mlxsw_sp_nexthop_counter_free(mlxsw_sp, nh);
6762 }
6763
6764 static bool mlxsw_sp_rt6_is_gateway(const struct mlxsw_sp *mlxsw_sp,
6765                                     const struct fib6_info *rt)
6766 {
6767         return rt->fib6_nh->fib_nh_gw_family ||
6768                mlxsw_sp_nexthop6_ipip_type(mlxsw_sp, rt, NULL);
6769 }
6770
6771 static int
6772 mlxsw_sp_nexthop6_group_info_init(struct mlxsw_sp *mlxsw_sp,
6773                                   struct mlxsw_sp_nexthop_group *nh_grp,
6774                                   struct mlxsw_sp_fib6_entry *fib6_entry)
6775 {
6776         struct mlxsw_sp_nexthop_group_info *nhgi;
6777         struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
6778         struct mlxsw_sp_nexthop *nh;
6779         int err, i;
6780
6781         nhgi = kzalloc(struct_size(nhgi, nexthops, fib6_entry->nrt6),
6782                        GFP_KERNEL);
6783         if (!nhgi)
6784                 return -ENOMEM;
6785         nh_grp->nhgi = nhgi;
6786         nhgi->nh_grp = nh_grp;
6787         mlxsw_sp_rt6 = list_first_entry(&fib6_entry->rt6_list,
6788                                         struct mlxsw_sp_rt6, list);
6789         nhgi->gateway = mlxsw_sp_rt6_is_gateway(mlxsw_sp, mlxsw_sp_rt6->rt);
6790         nhgi->count = fib6_entry->nrt6;
6791         for (i = 0; i < nhgi->count; i++) {
6792                 struct fib6_info *rt = mlxsw_sp_rt6->rt;
6793
6794                 nh = &nhgi->nexthops[i];
6795                 err = mlxsw_sp_nexthop6_init(mlxsw_sp, nh_grp, nh, rt);
6796                 if (err)
6797                         goto err_nexthop6_init;
6798                 mlxsw_sp_rt6 = list_next_entry(mlxsw_sp_rt6, list);
6799         }
6800         nh_grp->nhgi = nhgi;
6801         err = mlxsw_sp_nexthop_group_inc(mlxsw_sp);
6802         if (err)
6803                 goto err_group_inc;
6804         err = mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh_grp);
6805         if (err)
6806                 goto err_group_refresh;
6807
6808         return 0;
6809
6810 err_group_refresh:
6811         mlxsw_sp_nexthop_group_dec(mlxsw_sp);
6812 err_group_inc:
6813         i = nhgi->count;
6814 err_nexthop6_init:
6815         for (i--; i >= 0; i--) {
6816                 nh = &nhgi->nexthops[i];
6817                 mlxsw_sp_nexthop6_fini(mlxsw_sp, nh);
6818         }
6819         kfree(nhgi);
6820         return err;
6821 }
6822
6823 static void
6824 mlxsw_sp_nexthop6_group_info_fini(struct mlxsw_sp *mlxsw_sp,
6825                                   struct mlxsw_sp_nexthop_group *nh_grp)
6826 {
6827         struct mlxsw_sp_nexthop_group_info *nhgi = nh_grp->nhgi;
6828         int i;
6829
6830         mlxsw_sp_nexthop_group_dec(mlxsw_sp);
6831         for (i = nhgi->count - 1; i >= 0; i--) {
6832                 struct mlxsw_sp_nexthop *nh = &nhgi->nexthops[i];
6833
6834                 mlxsw_sp_nexthop6_fini(mlxsw_sp, nh);
6835         }
6836         mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh_grp);
6837         WARN_ON_ONCE(nhgi->adj_index_valid);
6838         kfree(nhgi);
6839 }
6840
6841 static struct mlxsw_sp_nexthop_group *
6842 mlxsw_sp_nexthop6_group_create(struct mlxsw_sp *mlxsw_sp,
6843                                struct mlxsw_sp_fib6_entry *fib6_entry)
6844 {
6845         struct mlxsw_sp_nexthop_group *nh_grp;
6846         int err;
6847
6848         nh_grp = kzalloc(sizeof(*nh_grp), GFP_KERNEL);
6849         if (!nh_grp)
6850                 return ERR_PTR(-ENOMEM);
6851         INIT_LIST_HEAD(&nh_grp->vr_list);
6852         err = rhashtable_init(&nh_grp->vr_ht,
6853                               &mlxsw_sp_nexthop_group_vr_ht_params);
6854         if (err)
6855                 goto err_nexthop_group_vr_ht_init;
6856         INIT_LIST_HEAD(&nh_grp->fib_list);
6857         nh_grp->type = MLXSW_SP_NEXTHOP_GROUP_TYPE_IPV6;
6858
6859         err = mlxsw_sp_nexthop6_group_info_init(mlxsw_sp, nh_grp, fib6_entry);
6860         if (err)
6861                 goto err_nexthop_group_info_init;
6862
6863         err = mlxsw_sp_nexthop_group_insert(mlxsw_sp, nh_grp);
6864         if (err)
6865                 goto err_nexthop_group_insert;
6866
6867         nh_grp->can_destroy = true;
6868
6869         return nh_grp;
6870
6871 err_nexthop_group_insert:
6872         mlxsw_sp_nexthop6_group_info_fini(mlxsw_sp, nh_grp);
6873 err_nexthop_group_info_init:
6874         rhashtable_destroy(&nh_grp->vr_ht);
6875 err_nexthop_group_vr_ht_init:
6876         kfree(nh_grp);
6877         return ERR_PTR(err);
6878 }
6879
6880 static void
6881 mlxsw_sp_nexthop6_group_destroy(struct mlxsw_sp *mlxsw_sp,
6882                                 struct mlxsw_sp_nexthop_group *nh_grp)
6883 {
6884         if (!nh_grp->can_destroy)
6885                 return;
6886         mlxsw_sp_nexthop_group_remove(mlxsw_sp, nh_grp);
6887         mlxsw_sp_nexthop6_group_info_fini(mlxsw_sp, nh_grp);
6888         WARN_ON_ONCE(!list_empty(&nh_grp->vr_list));
6889         rhashtable_destroy(&nh_grp->vr_ht);
6890         kfree(nh_grp);
6891 }
6892
6893 static int mlxsw_sp_nexthop6_group_get(struct mlxsw_sp *mlxsw_sp,
6894                                        struct mlxsw_sp_fib6_entry *fib6_entry)
6895 {
6896         struct fib6_info *rt = mlxsw_sp_fib6_entry_rt(fib6_entry);
6897         struct mlxsw_sp_nexthop_group *nh_grp;
6898
6899         if (rt->nh) {
6900                 nh_grp = mlxsw_sp_nexthop_obj_group_lookup(mlxsw_sp,
6901                                                            rt->nh->id);
6902                 if (WARN_ON_ONCE(!nh_grp))
6903                         return -EINVAL;
6904                 goto out;
6905         }
6906
6907         nh_grp = mlxsw_sp_nexthop6_group_lookup(mlxsw_sp, fib6_entry);
6908         if (!nh_grp) {
6909                 nh_grp = mlxsw_sp_nexthop6_group_create(mlxsw_sp, fib6_entry);
6910                 if (IS_ERR(nh_grp))
6911                         return PTR_ERR(nh_grp);
6912         }
6913
6914         /* The route and the nexthop are described by the same struct, so we
6915          * need to the update the nexthop offload indication for the new route.
6916          */
6917         __mlxsw_sp_nexthop6_group_offload_refresh(nh_grp, fib6_entry);
6918
6919 out:
6920         list_add_tail(&fib6_entry->common.nexthop_group_node,
6921                       &nh_grp->fib_list);
6922         fib6_entry->common.nh_group = nh_grp;
6923
6924         return 0;
6925 }
6926
6927 static void mlxsw_sp_nexthop6_group_put(struct mlxsw_sp *mlxsw_sp,
6928                                         struct mlxsw_sp_fib_entry *fib_entry)
6929 {
6930         struct mlxsw_sp_nexthop_group *nh_grp = fib_entry->nh_group;
6931
6932         list_del(&fib_entry->nexthop_group_node);
6933         if (!list_empty(&nh_grp->fib_list))
6934                 return;
6935
6936         if (nh_grp->type == MLXSW_SP_NEXTHOP_GROUP_TYPE_OBJ) {
6937                 mlxsw_sp_nexthop_obj_group_destroy(mlxsw_sp, nh_grp);
6938                 return;
6939         }
6940
6941         mlxsw_sp_nexthop6_group_destroy(mlxsw_sp, nh_grp);
6942 }
6943
6944 static int
6945 mlxsw_sp_nexthop6_group_update(struct mlxsw_sp *mlxsw_sp,
6946                                struct mlxsw_sp_fib6_entry *fib6_entry)
6947 {
6948         struct mlxsw_sp_nexthop_group *old_nh_grp = fib6_entry->common.nh_group;
6949         struct mlxsw_sp_fib_node *fib_node = fib6_entry->common.fib_node;
6950         int err;
6951
6952         mlxsw_sp_nexthop_group_vr_unlink(old_nh_grp, fib_node->fib);
6953         fib6_entry->common.nh_group = NULL;
6954         list_del(&fib6_entry->common.nexthop_group_node);
6955
6956         err = mlxsw_sp_nexthop6_group_get(mlxsw_sp, fib6_entry);
6957         if (err)
6958                 goto err_nexthop6_group_get;
6959
6960         err = mlxsw_sp_nexthop_group_vr_link(fib6_entry->common.nh_group,
6961                                              fib_node->fib);
6962         if (err)
6963                 goto err_nexthop_group_vr_link;
6964
6965         /* In case this entry is offloaded, then the adjacency index
6966          * currently associated with it in the device's table is that
6967          * of the old group. Start using the new one instead.
6968          */
6969         err = mlxsw_sp_fib_entry_update(mlxsw_sp, &fib6_entry->common);
6970         if (err)
6971                 goto err_fib_entry_update;
6972
6973         if (list_empty(&old_nh_grp->fib_list))
6974                 mlxsw_sp_nexthop6_group_destroy(mlxsw_sp, old_nh_grp);
6975
6976         return 0;
6977
6978 err_fib_entry_update:
6979         mlxsw_sp_nexthop_group_vr_unlink(fib6_entry->common.nh_group,
6980                                          fib_node->fib);
6981 err_nexthop_group_vr_link:
6982         mlxsw_sp_nexthop6_group_put(mlxsw_sp, &fib6_entry->common);
6983 err_nexthop6_group_get:
6984         list_add_tail(&fib6_entry->common.nexthop_group_node,
6985                       &old_nh_grp->fib_list);
6986         fib6_entry->common.nh_group = old_nh_grp;
6987         mlxsw_sp_nexthop_group_vr_link(old_nh_grp, fib_node->fib);
6988         return err;
6989 }
6990
6991 static int
6992 mlxsw_sp_fib6_entry_nexthop_add(struct mlxsw_sp *mlxsw_sp,
6993                                 struct mlxsw_sp_fib6_entry *fib6_entry,
6994                                 struct fib6_info **rt_arr, unsigned int nrt6)
6995 {
6996         struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
6997         int err, i;
6998
6999         for (i = 0; i < nrt6; i++) {
7000                 mlxsw_sp_rt6 = mlxsw_sp_rt6_create(rt_arr[i]);
7001                 if (IS_ERR(mlxsw_sp_rt6)) {
7002                         err = PTR_ERR(mlxsw_sp_rt6);
7003                         goto err_rt6_unwind;
7004                 }
7005
7006                 list_add_tail(&mlxsw_sp_rt6->list, &fib6_entry->rt6_list);
7007                 fib6_entry->nrt6++;
7008         }
7009
7010         err = mlxsw_sp_nexthop6_group_update(mlxsw_sp, fib6_entry);
7011         if (err)
7012                 goto err_rt6_unwind;
7013
7014         return 0;
7015
7016 err_rt6_unwind:
7017         for (; i > 0; i--) {
7018                 fib6_entry->nrt6--;
7019                 mlxsw_sp_rt6 = list_last_entry(&fib6_entry->rt6_list,
7020                                                struct mlxsw_sp_rt6, list);
7021                 list_del(&mlxsw_sp_rt6->list);
7022                 mlxsw_sp_rt6_destroy(mlxsw_sp_rt6);
7023         }
7024         return err;
7025 }
7026
7027 static void
7028 mlxsw_sp_fib6_entry_nexthop_del(struct mlxsw_sp *mlxsw_sp,
7029                                 struct mlxsw_sp_fib6_entry *fib6_entry,
7030                                 struct fib6_info **rt_arr, unsigned int nrt6)
7031 {
7032         struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
7033         int i;
7034
7035         for (i = 0; i < nrt6; i++) {
7036                 mlxsw_sp_rt6 = mlxsw_sp_fib6_entry_rt_find(fib6_entry,
7037                                                            rt_arr[i]);
7038                 if (WARN_ON_ONCE(!mlxsw_sp_rt6))
7039                         continue;
7040
7041                 fib6_entry->nrt6--;
7042                 list_del(&mlxsw_sp_rt6->list);
7043                 mlxsw_sp_rt6_destroy(mlxsw_sp_rt6);
7044         }
7045
7046         mlxsw_sp_nexthop6_group_update(mlxsw_sp, fib6_entry);
7047 }
7048
7049 static int
7050 mlxsw_sp_fib6_entry_type_set_local(struct mlxsw_sp *mlxsw_sp,
7051                                    struct mlxsw_sp_fib_entry *fib_entry,
7052                                    const struct fib6_info *rt)
7053 {
7054         struct mlxsw_sp_nexthop_group_info *nhgi = fib_entry->nh_group->nhgi;
7055         union mlxsw_sp_l3addr dip = { .addr6 = rt->fib6_dst.addr };
7056         u32 tb_id = mlxsw_sp_fix_tb_id(rt->fib6_table->tb6_id);
7057         struct mlxsw_sp_router *router = mlxsw_sp->router;
7058         int ifindex = nhgi->nexthops[0].ifindex;
7059         struct mlxsw_sp_ipip_entry *ipip_entry;
7060
7061         fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_TRAP;
7062         ipip_entry = mlxsw_sp_ipip_entry_find_by_decap(mlxsw_sp, ifindex,
7063                                                        MLXSW_SP_L3_PROTO_IPV6,
7064                                                        dip);
7065
7066         if (ipip_entry && ipip_entry->ol_dev->flags & IFF_UP) {
7067                 fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_IPIP_DECAP;
7068                 return mlxsw_sp_fib_entry_decap_init(mlxsw_sp, fib_entry,
7069                                                      ipip_entry);
7070         }
7071         if (mlxsw_sp_router_nve_is_decap(mlxsw_sp, tb_id,
7072                                          MLXSW_SP_L3_PROTO_IPV6, &dip)) {
7073                 u32 tunnel_index;
7074
7075                 tunnel_index = router->nve_decap_config.tunnel_index;
7076                 fib_entry->decap.tunnel_index = tunnel_index;
7077                 fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_NVE_DECAP;
7078         }
7079
7080         return 0;
7081 }
7082
7083 static int mlxsw_sp_fib6_entry_type_set(struct mlxsw_sp *mlxsw_sp,
7084                                         struct mlxsw_sp_fib_entry *fib_entry,
7085                                         const struct fib6_info *rt)
7086 {
7087         if (rt->fib6_flags & RTF_LOCAL)
7088                 return mlxsw_sp_fib6_entry_type_set_local(mlxsw_sp, fib_entry,
7089                                                           rt);
7090         if (rt->fib6_flags & RTF_ANYCAST)
7091                 fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_TRAP;
7092         else if (rt->fib6_type == RTN_BLACKHOLE)
7093                 fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_BLACKHOLE;
7094         else if (rt->fib6_flags & RTF_REJECT)
7095                 fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_UNREACHABLE;
7096         else if (fib_entry->nh_group->nhgi->gateway)
7097                 fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_REMOTE;
7098         else
7099                 fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_LOCAL;
7100
7101         return 0;
7102 }
7103
7104 static void
7105 mlxsw_sp_fib6_entry_rt_destroy_all(struct mlxsw_sp_fib6_entry *fib6_entry)
7106 {
7107         struct mlxsw_sp_rt6 *mlxsw_sp_rt6, *tmp;
7108
7109         list_for_each_entry_safe(mlxsw_sp_rt6, tmp, &fib6_entry->rt6_list,
7110                                  list) {
7111                 fib6_entry->nrt6--;
7112                 list_del(&mlxsw_sp_rt6->list);
7113                 mlxsw_sp_rt6_destroy(mlxsw_sp_rt6);
7114         }
7115 }
7116
7117 static struct mlxsw_sp_fib6_entry *
7118 mlxsw_sp_fib6_entry_create(struct mlxsw_sp *mlxsw_sp,
7119                            struct mlxsw_sp_fib_node *fib_node,
7120                            struct fib6_info **rt_arr, unsigned int nrt6)
7121 {
7122         struct mlxsw_sp_fib6_entry *fib6_entry;
7123         struct mlxsw_sp_fib_entry *fib_entry;
7124         struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
7125         int err, i;
7126
7127         fib6_entry = kzalloc(sizeof(*fib6_entry), GFP_KERNEL);
7128         if (!fib6_entry)
7129                 return ERR_PTR(-ENOMEM);
7130         fib_entry = &fib6_entry->common;
7131
7132         INIT_LIST_HEAD(&fib6_entry->rt6_list);
7133
7134         for (i = 0; i < nrt6; i++) {
7135                 mlxsw_sp_rt6 = mlxsw_sp_rt6_create(rt_arr[i]);
7136                 if (IS_ERR(mlxsw_sp_rt6)) {
7137                         err = PTR_ERR(mlxsw_sp_rt6);
7138                         goto err_rt6_unwind;
7139                 }
7140                 list_add_tail(&mlxsw_sp_rt6->list, &fib6_entry->rt6_list);
7141                 fib6_entry->nrt6++;
7142         }
7143
7144         err = mlxsw_sp_nexthop6_group_get(mlxsw_sp, fib6_entry);
7145         if (err)
7146                 goto err_rt6_unwind;
7147
7148         err = mlxsw_sp_nexthop_group_vr_link(fib_entry->nh_group,
7149                                              fib_node->fib);
7150         if (err)
7151                 goto err_nexthop_group_vr_link;
7152
7153         err = mlxsw_sp_fib6_entry_type_set(mlxsw_sp, fib_entry, rt_arr[0]);
7154         if (err)
7155                 goto err_fib6_entry_type_set;
7156
7157         fib_entry->fib_node = fib_node;
7158
7159         return fib6_entry;
7160
7161 err_fib6_entry_type_set:
7162         mlxsw_sp_nexthop_group_vr_unlink(fib_entry->nh_group, fib_node->fib);
7163 err_nexthop_group_vr_link:
7164         mlxsw_sp_nexthop6_group_put(mlxsw_sp, fib_entry);
7165 err_rt6_unwind:
7166         for (; i > 0; i--) {
7167                 fib6_entry->nrt6--;
7168                 mlxsw_sp_rt6 = list_last_entry(&fib6_entry->rt6_list,
7169                                                struct mlxsw_sp_rt6, list);
7170                 list_del(&mlxsw_sp_rt6->list);
7171                 mlxsw_sp_rt6_destroy(mlxsw_sp_rt6);
7172         }
7173         kfree(fib6_entry);
7174         return ERR_PTR(err);
7175 }
7176
7177 static void
7178 mlxsw_sp_fib6_entry_type_unset(struct mlxsw_sp *mlxsw_sp,
7179                                struct mlxsw_sp_fib6_entry *fib6_entry)
7180 {
7181         mlxsw_sp_fib_entry_type_unset(mlxsw_sp, &fib6_entry->common);
7182 }
7183
7184 static void mlxsw_sp_fib6_entry_destroy(struct mlxsw_sp *mlxsw_sp,
7185                                         struct mlxsw_sp_fib6_entry *fib6_entry)
7186 {
7187         struct mlxsw_sp_fib_node *fib_node = fib6_entry->common.fib_node;
7188
7189         mlxsw_sp_fib6_entry_type_unset(mlxsw_sp, fib6_entry);
7190         mlxsw_sp_nexthop_group_vr_unlink(fib6_entry->common.nh_group,
7191                                          fib_node->fib);
7192         mlxsw_sp_nexthop6_group_put(mlxsw_sp, &fib6_entry->common);
7193         mlxsw_sp_fib6_entry_rt_destroy_all(fib6_entry);
7194         WARN_ON(fib6_entry->nrt6);
7195         kfree(fib6_entry);
7196 }
7197
7198 static struct mlxsw_sp_fib6_entry *
7199 mlxsw_sp_fib6_entry_lookup(struct mlxsw_sp *mlxsw_sp,
7200                            const struct fib6_info *rt)
7201 {
7202         struct mlxsw_sp_fib6_entry *fib6_entry;
7203         struct mlxsw_sp_fib_node *fib_node;
7204         struct mlxsw_sp_fib *fib;
7205         struct fib6_info *cmp_rt;
7206         struct mlxsw_sp_vr *vr;
7207
7208         vr = mlxsw_sp_vr_find(mlxsw_sp, rt->fib6_table->tb6_id);
7209         if (!vr)
7210                 return NULL;
7211         fib = mlxsw_sp_vr_fib(vr, MLXSW_SP_L3_PROTO_IPV6);
7212
7213         fib_node = mlxsw_sp_fib_node_lookup(fib, &rt->fib6_dst.addr,
7214                                             sizeof(rt->fib6_dst.addr),
7215                                             rt->fib6_dst.plen);
7216         if (!fib_node)
7217                 return NULL;
7218
7219         fib6_entry = container_of(fib_node->fib_entry,
7220                                   struct mlxsw_sp_fib6_entry, common);
7221         cmp_rt = mlxsw_sp_fib6_entry_rt(fib6_entry);
7222         if (rt->fib6_table->tb6_id == cmp_rt->fib6_table->tb6_id &&
7223             rt->fib6_metric == cmp_rt->fib6_metric &&
7224             mlxsw_sp_fib6_entry_rt_find(fib6_entry, rt))
7225                 return fib6_entry;
7226
7227         return NULL;
7228 }
7229
7230 static bool mlxsw_sp_fib6_allow_replace(struct mlxsw_sp_fib6_entry *fib6_entry)
7231 {
7232         struct mlxsw_sp_fib_node *fib_node = fib6_entry->common.fib_node;
7233         struct mlxsw_sp_fib6_entry *fib6_replaced;
7234         struct fib6_info *rt, *rt_replaced;
7235
7236         if (!fib_node->fib_entry)
7237                 return true;
7238
7239         fib6_replaced = container_of(fib_node->fib_entry,
7240                                      struct mlxsw_sp_fib6_entry,
7241                                      common);
7242         rt = mlxsw_sp_fib6_entry_rt(fib6_entry);
7243         rt_replaced = mlxsw_sp_fib6_entry_rt(fib6_replaced);
7244         if (rt->fib6_table->tb6_id == RT_TABLE_MAIN &&
7245             rt_replaced->fib6_table->tb6_id == RT_TABLE_LOCAL)
7246                 return false;
7247
7248         return true;
7249 }
7250
7251 static int mlxsw_sp_router_fib6_replace(struct mlxsw_sp *mlxsw_sp,
7252                                         struct fib6_info **rt_arr,
7253                                         unsigned int nrt6)
7254 {
7255         struct mlxsw_sp_fib6_entry *fib6_entry, *fib6_replaced;
7256         struct mlxsw_sp_fib_entry *replaced;
7257         struct mlxsw_sp_fib_node *fib_node;
7258         struct fib6_info *rt = rt_arr[0];
7259         int err;
7260
7261         if (rt->fib6_src.plen)
7262                 return -EINVAL;
7263
7264         if (mlxsw_sp_fib6_rt_should_ignore(rt))
7265                 return 0;
7266
7267         if (rt->nh && !mlxsw_sp_nexthop_obj_group_lookup(mlxsw_sp, rt->nh->id))
7268                 return 0;
7269
7270         fib_node = mlxsw_sp_fib_node_get(mlxsw_sp, rt->fib6_table->tb6_id,
7271                                          &rt->fib6_dst.addr,
7272                                          sizeof(rt->fib6_dst.addr),
7273                                          rt->fib6_dst.plen,
7274                                          MLXSW_SP_L3_PROTO_IPV6);
7275         if (IS_ERR(fib_node))
7276                 return PTR_ERR(fib_node);
7277
7278         fib6_entry = mlxsw_sp_fib6_entry_create(mlxsw_sp, fib_node, rt_arr,
7279                                                 nrt6);
7280         if (IS_ERR(fib6_entry)) {
7281                 err = PTR_ERR(fib6_entry);
7282                 goto err_fib6_entry_create;
7283         }
7284
7285         if (!mlxsw_sp_fib6_allow_replace(fib6_entry)) {
7286                 mlxsw_sp_fib6_entry_destroy(mlxsw_sp, fib6_entry);
7287                 mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
7288                 return 0;
7289         }
7290
7291         replaced = fib_node->fib_entry;
7292         err = mlxsw_sp_fib_node_entry_link(mlxsw_sp, &fib6_entry->common);
7293         if (err)
7294                 goto err_fib_node_entry_link;
7295
7296         /* Nothing to replace */
7297         if (!replaced)
7298                 return 0;
7299
7300         mlxsw_sp_fib_entry_hw_flags_clear(mlxsw_sp, replaced);
7301         fib6_replaced = container_of(replaced, struct mlxsw_sp_fib6_entry,
7302                                      common);
7303         mlxsw_sp_fib6_entry_destroy(mlxsw_sp, fib6_replaced);
7304
7305         return 0;
7306
7307 err_fib_node_entry_link:
7308         fib_node->fib_entry = replaced;
7309         mlxsw_sp_fib6_entry_destroy(mlxsw_sp, fib6_entry);
7310 err_fib6_entry_create:
7311         mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
7312         return err;
7313 }
7314
7315 static int mlxsw_sp_router_fib6_append(struct mlxsw_sp *mlxsw_sp,
7316                                        struct fib6_info **rt_arr,
7317                                        unsigned int nrt6)
7318 {
7319         struct mlxsw_sp_fib6_entry *fib6_entry;
7320         struct mlxsw_sp_fib_node *fib_node;
7321         struct fib6_info *rt = rt_arr[0];
7322         int err;
7323
7324         if (rt->fib6_src.plen)
7325                 return -EINVAL;
7326
7327         if (mlxsw_sp_fib6_rt_should_ignore(rt))
7328                 return 0;
7329
7330         fib_node = mlxsw_sp_fib_node_get(mlxsw_sp, rt->fib6_table->tb6_id,
7331                                          &rt->fib6_dst.addr,
7332                                          sizeof(rt->fib6_dst.addr),
7333                                          rt->fib6_dst.plen,
7334                                          MLXSW_SP_L3_PROTO_IPV6);
7335         if (IS_ERR(fib_node))
7336                 return PTR_ERR(fib_node);
7337
7338         if (WARN_ON_ONCE(!fib_node->fib_entry)) {
7339                 mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
7340                 return -EINVAL;
7341         }
7342
7343         fib6_entry = container_of(fib_node->fib_entry,
7344                                   struct mlxsw_sp_fib6_entry, common);
7345         err = mlxsw_sp_fib6_entry_nexthop_add(mlxsw_sp, fib6_entry, rt_arr,
7346                                               nrt6);
7347         if (err)
7348                 goto err_fib6_entry_nexthop_add;
7349
7350         return 0;
7351
7352 err_fib6_entry_nexthop_add:
7353         mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
7354         return err;
7355 }
7356
7357 static void mlxsw_sp_router_fib6_del(struct mlxsw_sp *mlxsw_sp,
7358                                      struct fib6_info **rt_arr,
7359                                      unsigned int nrt6)
7360 {
7361         struct mlxsw_sp_fib6_entry *fib6_entry;
7362         struct mlxsw_sp_fib_node *fib_node;
7363         struct fib6_info *rt = rt_arr[0];
7364
7365         if (mlxsw_sp_fib6_rt_should_ignore(rt))
7366                 return;
7367
7368         /* Multipath routes are first added to the FIB trie and only then
7369          * notified. If we vetoed the addition, we will get a delete
7370          * notification for a route we do not have. Therefore, do not warn if
7371          * route was not found.
7372          */
7373         fib6_entry = mlxsw_sp_fib6_entry_lookup(mlxsw_sp, rt);
7374         if (!fib6_entry)
7375                 return;
7376
7377         /* If not all the nexthops are deleted, then only reduce the nexthop
7378          * group.
7379          */
7380         if (nrt6 != fib6_entry->nrt6) {
7381                 mlxsw_sp_fib6_entry_nexthop_del(mlxsw_sp, fib6_entry, rt_arr,
7382                                                 nrt6);
7383                 return;
7384         }
7385
7386         fib_node = fib6_entry->common.fib_node;
7387
7388         mlxsw_sp_fib_node_entry_unlink(mlxsw_sp, &fib6_entry->common);
7389         mlxsw_sp_fib6_entry_destroy(mlxsw_sp, fib6_entry);
7390         mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
7391 }
7392
7393 static struct mlxsw_sp_mr_table *
7394 mlxsw_sp_router_fibmr_family_to_table(struct mlxsw_sp_vr *vr, int family)
7395 {
7396         if (family == RTNL_FAMILY_IPMR)
7397                 return vr->mr_table[MLXSW_SP_L3_PROTO_IPV4];
7398         else
7399                 return vr->mr_table[MLXSW_SP_L3_PROTO_IPV6];
7400 }
7401
7402 static int mlxsw_sp_router_fibmr_add(struct mlxsw_sp *mlxsw_sp,
7403                                      struct mfc_entry_notifier_info *men_info,
7404                                      bool replace)
7405 {
7406         struct mlxsw_sp_mr_table *mrt;
7407         struct mlxsw_sp_vr *vr;
7408
7409         vr = mlxsw_sp_vr_get(mlxsw_sp, men_info->tb_id, NULL);
7410         if (IS_ERR(vr))
7411                 return PTR_ERR(vr);
7412
7413         mrt = mlxsw_sp_router_fibmr_family_to_table(vr, men_info->info.family);
7414         return mlxsw_sp_mr_route_add(mrt, men_info->mfc, replace);
7415 }
7416
7417 static void mlxsw_sp_router_fibmr_del(struct mlxsw_sp *mlxsw_sp,
7418                                       struct mfc_entry_notifier_info *men_info)
7419 {
7420         struct mlxsw_sp_mr_table *mrt;
7421         struct mlxsw_sp_vr *vr;
7422
7423         vr = mlxsw_sp_vr_find(mlxsw_sp, men_info->tb_id);
7424         if (WARN_ON(!vr))
7425                 return;
7426
7427         mrt = mlxsw_sp_router_fibmr_family_to_table(vr, men_info->info.family);
7428         mlxsw_sp_mr_route_del(mrt, men_info->mfc);
7429         mlxsw_sp_vr_put(mlxsw_sp, vr);
7430 }
7431
7432 static int
7433 mlxsw_sp_router_fibmr_vif_add(struct mlxsw_sp *mlxsw_sp,
7434                               struct vif_entry_notifier_info *ven_info)
7435 {
7436         struct mlxsw_sp_mr_table *mrt;
7437         struct mlxsw_sp_rif *rif;
7438         struct mlxsw_sp_vr *vr;
7439
7440         vr = mlxsw_sp_vr_get(mlxsw_sp, ven_info->tb_id, NULL);
7441         if (IS_ERR(vr))
7442                 return PTR_ERR(vr);
7443
7444         mrt = mlxsw_sp_router_fibmr_family_to_table(vr, ven_info->info.family);
7445         rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, ven_info->dev);
7446         return mlxsw_sp_mr_vif_add(mrt, ven_info->dev,
7447                                    ven_info->vif_index,
7448                                    ven_info->vif_flags, rif);
7449 }
7450
7451 static void
7452 mlxsw_sp_router_fibmr_vif_del(struct mlxsw_sp *mlxsw_sp,
7453                               struct vif_entry_notifier_info *ven_info)
7454 {
7455         struct mlxsw_sp_mr_table *mrt;
7456         struct mlxsw_sp_vr *vr;
7457
7458         vr = mlxsw_sp_vr_find(mlxsw_sp, ven_info->tb_id);
7459         if (WARN_ON(!vr))
7460                 return;
7461
7462         mrt = mlxsw_sp_router_fibmr_family_to_table(vr, ven_info->info.family);
7463         mlxsw_sp_mr_vif_del(mrt, ven_info->vif_index);
7464         mlxsw_sp_vr_put(mlxsw_sp, vr);
7465 }
7466
7467 static void mlxsw_sp_fib4_node_flush(struct mlxsw_sp *mlxsw_sp,
7468                                      struct mlxsw_sp_fib_node *fib_node)
7469 {
7470         struct mlxsw_sp_fib4_entry *fib4_entry;
7471
7472         fib4_entry = container_of(fib_node->fib_entry,
7473                                   struct mlxsw_sp_fib4_entry, common);
7474         mlxsw_sp_fib_node_entry_unlink(mlxsw_sp, fib_node->fib_entry);
7475         mlxsw_sp_fib4_entry_destroy(mlxsw_sp, fib4_entry);
7476         mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
7477 }
7478
7479 static void mlxsw_sp_fib6_node_flush(struct mlxsw_sp *mlxsw_sp,
7480                                      struct mlxsw_sp_fib_node *fib_node)
7481 {
7482         struct mlxsw_sp_fib6_entry *fib6_entry;
7483
7484         fib6_entry = container_of(fib_node->fib_entry,
7485                                   struct mlxsw_sp_fib6_entry, common);
7486         mlxsw_sp_fib_node_entry_unlink(mlxsw_sp, fib_node->fib_entry);
7487         mlxsw_sp_fib6_entry_destroy(mlxsw_sp, fib6_entry);
7488         mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
7489 }
7490
7491 static void mlxsw_sp_fib_node_flush(struct mlxsw_sp *mlxsw_sp,
7492                                     struct mlxsw_sp_fib_node *fib_node)
7493 {
7494         switch (fib_node->fib->proto) {
7495         case MLXSW_SP_L3_PROTO_IPV4:
7496                 mlxsw_sp_fib4_node_flush(mlxsw_sp, fib_node);
7497                 break;
7498         case MLXSW_SP_L3_PROTO_IPV6:
7499                 mlxsw_sp_fib6_node_flush(mlxsw_sp, fib_node);
7500                 break;
7501         }
7502 }
7503
7504 static void mlxsw_sp_vr_fib_flush(struct mlxsw_sp *mlxsw_sp,
7505                                   struct mlxsw_sp_vr *vr,
7506                                   enum mlxsw_sp_l3proto proto)
7507 {
7508         struct mlxsw_sp_fib *fib = mlxsw_sp_vr_fib(vr, proto);
7509         struct mlxsw_sp_fib_node *fib_node, *tmp;
7510
7511         list_for_each_entry_safe(fib_node, tmp, &fib->node_list, list) {
7512                 bool do_break = &tmp->list == &fib->node_list;
7513
7514                 mlxsw_sp_fib_node_flush(mlxsw_sp, fib_node);
7515                 if (do_break)
7516                         break;
7517         }
7518 }
7519
7520 static void mlxsw_sp_router_fib_flush(struct mlxsw_sp *mlxsw_sp)
7521 {
7522         int max_vrs = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_VRS);
7523         int i, j;
7524
7525         for (i = 0; i < max_vrs; i++) {
7526                 struct mlxsw_sp_vr *vr = &mlxsw_sp->router->vrs[i];
7527
7528                 if (!mlxsw_sp_vr_is_used(vr))
7529                         continue;
7530
7531                 for (j = 0; j < MLXSW_SP_L3_PROTO_MAX; j++)
7532                         mlxsw_sp_mr_table_flush(vr->mr_table[j]);
7533                 mlxsw_sp_vr_fib_flush(mlxsw_sp, vr, MLXSW_SP_L3_PROTO_IPV4);
7534
7535                 /* If virtual router was only used for IPv4, then it's no
7536                  * longer used.
7537                  */
7538                 if (!mlxsw_sp_vr_is_used(vr))
7539                         continue;
7540                 mlxsw_sp_vr_fib_flush(mlxsw_sp, vr, MLXSW_SP_L3_PROTO_IPV6);
7541         }
7542 }
7543
7544 struct mlxsw_sp_fib6_event_work {
7545         struct fib6_info **rt_arr;
7546         unsigned int nrt6;
7547 };
7548
7549 struct mlxsw_sp_fib_event_work {
7550         struct work_struct work;
7551         netdevice_tracker dev_tracker;
7552         union {
7553                 struct mlxsw_sp_fib6_event_work fib6_work;
7554                 struct fib_entry_notifier_info fen_info;
7555                 struct fib_rule_notifier_info fr_info;
7556                 struct fib_nh_notifier_info fnh_info;
7557                 struct mfc_entry_notifier_info men_info;
7558                 struct vif_entry_notifier_info ven_info;
7559         };
7560         struct mlxsw_sp *mlxsw_sp;
7561         unsigned long event;
7562 };
7563
7564 static int
7565 mlxsw_sp_router_fib6_work_init(struct mlxsw_sp_fib6_event_work *fib6_work,
7566                                struct fib6_entry_notifier_info *fen6_info)
7567 {
7568         struct fib6_info *rt = fen6_info->rt;
7569         struct fib6_info **rt_arr;
7570         struct fib6_info *iter;
7571         unsigned int nrt6;
7572         int i = 0;
7573
7574         nrt6 = fen6_info->nsiblings + 1;
7575
7576         rt_arr = kcalloc(nrt6, sizeof(struct fib6_info *), GFP_ATOMIC);
7577         if (!rt_arr)
7578                 return -ENOMEM;
7579
7580         fib6_work->rt_arr = rt_arr;
7581         fib6_work->nrt6 = nrt6;
7582
7583         rt_arr[0] = rt;
7584         fib6_info_hold(rt);
7585
7586         if (!fen6_info->nsiblings)
7587                 return 0;
7588
7589         list_for_each_entry(iter, &rt->fib6_siblings, fib6_siblings) {
7590                 if (i == fen6_info->nsiblings)
7591                         break;
7592
7593                 rt_arr[i + 1] = iter;
7594                 fib6_info_hold(iter);
7595                 i++;
7596         }
7597         WARN_ON_ONCE(i != fen6_info->nsiblings);
7598
7599         return 0;
7600 }
7601
7602 static void
7603 mlxsw_sp_router_fib6_work_fini(struct mlxsw_sp_fib6_event_work *fib6_work)
7604 {
7605         int i;
7606
7607         for (i = 0; i < fib6_work->nrt6; i++)
7608                 mlxsw_sp_rt6_release(fib6_work->rt_arr[i]);
7609         kfree(fib6_work->rt_arr);
7610 }
7611
7612 static void mlxsw_sp_router_fib4_event_work(struct work_struct *work)
7613 {
7614         struct mlxsw_sp_fib_event_work *fib_work =
7615                 container_of(work, struct mlxsw_sp_fib_event_work, work);
7616         struct mlxsw_sp *mlxsw_sp = fib_work->mlxsw_sp;
7617         int err;
7618
7619         mutex_lock(&mlxsw_sp->router->lock);
7620         mlxsw_sp_span_respin(mlxsw_sp);
7621
7622         switch (fib_work->event) {
7623         case FIB_EVENT_ENTRY_REPLACE:
7624                 err = mlxsw_sp_router_fib4_replace(mlxsw_sp,
7625                                                    &fib_work->fen_info);
7626                 if (err) {
7627                         dev_warn(mlxsw_sp->bus_info->dev, "FIB replace failed.\n");
7628                         mlxsw_sp_fib4_offload_failed_flag_set(mlxsw_sp,
7629                                                               &fib_work->fen_info);
7630                 }
7631                 fib_info_put(fib_work->fen_info.fi);
7632                 break;
7633         case FIB_EVENT_ENTRY_DEL:
7634                 mlxsw_sp_router_fib4_del(mlxsw_sp, &fib_work->fen_info);
7635                 fib_info_put(fib_work->fen_info.fi);
7636                 break;
7637         case FIB_EVENT_NH_ADD:
7638         case FIB_EVENT_NH_DEL:
7639                 mlxsw_sp_nexthop4_event(mlxsw_sp, fib_work->event,
7640                                         fib_work->fnh_info.fib_nh);
7641                 fib_info_put(fib_work->fnh_info.fib_nh->nh_parent);
7642                 break;
7643         }
7644         mutex_unlock(&mlxsw_sp->router->lock);
7645         kfree(fib_work);
7646 }
7647
7648 static void mlxsw_sp_router_fib6_event_work(struct work_struct *work)
7649 {
7650         struct mlxsw_sp_fib_event_work *fib_work =
7651                     container_of(work, struct mlxsw_sp_fib_event_work, work);
7652         struct mlxsw_sp_fib6_event_work *fib6_work = &fib_work->fib6_work;
7653         struct mlxsw_sp *mlxsw_sp = fib_work->mlxsw_sp;
7654         int err;
7655
7656         mutex_lock(&mlxsw_sp->router->lock);
7657         mlxsw_sp_span_respin(mlxsw_sp);
7658
7659         switch (fib_work->event) {
7660         case FIB_EVENT_ENTRY_REPLACE:
7661                 err = mlxsw_sp_router_fib6_replace(mlxsw_sp,
7662                                                    fib6_work->rt_arr,
7663                                                    fib6_work->nrt6);
7664                 if (err) {
7665                         dev_warn(mlxsw_sp->bus_info->dev, "FIB replace failed.\n");
7666                         mlxsw_sp_fib6_offload_failed_flag_set(mlxsw_sp,
7667                                                               fib6_work->rt_arr,
7668                                                               fib6_work->nrt6);
7669                 }
7670                 mlxsw_sp_router_fib6_work_fini(fib6_work);
7671                 break;
7672         case FIB_EVENT_ENTRY_APPEND:
7673                 err = mlxsw_sp_router_fib6_append(mlxsw_sp,
7674                                                   fib6_work->rt_arr,
7675                                                   fib6_work->nrt6);
7676                 if (err) {
7677                         dev_warn(mlxsw_sp->bus_info->dev, "FIB append failed.\n");
7678                         mlxsw_sp_fib6_offload_failed_flag_set(mlxsw_sp,
7679                                                               fib6_work->rt_arr,
7680                                                               fib6_work->nrt6);
7681                 }
7682                 mlxsw_sp_router_fib6_work_fini(fib6_work);
7683                 break;
7684         case FIB_EVENT_ENTRY_DEL:
7685                 mlxsw_sp_router_fib6_del(mlxsw_sp,
7686                                          fib6_work->rt_arr,
7687                                          fib6_work->nrt6);
7688                 mlxsw_sp_router_fib6_work_fini(fib6_work);
7689                 break;
7690         }
7691         mutex_unlock(&mlxsw_sp->router->lock);
7692         kfree(fib_work);
7693 }
7694
7695 static void mlxsw_sp_router_fibmr_event_work(struct work_struct *work)
7696 {
7697         struct mlxsw_sp_fib_event_work *fib_work =
7698                 container_of(work, struct mlxsw_sp_fib_event_work, work);
7699         struct mlxsw_sp *mlxsw_sp = fib_work->mlxsw_sp;
7700         bool replace;
7701         int err;
7702
7703         rtnl_lock();
7704         mutex_lock(&mlxsw_sp->router->lock);
7705         switch (fib_work->event) {
7706         case FIB_EVENT_ENTRY_REPLACE:
7707         case FIB_EVENT_ENTRY_ADD:
7708                 replace = fib_work->event == FIB_EVENT_ENTRY_REPLACE;
7709
7710                 err = mlxsw_sp_router_fibmr_add(mlxsw_sp, &fib_work->men_info,
7711                                                 replace);
7712                 if (err)
7713                         dev_warn(mlxsw_sp->bus_info->dev, "MR entry add failed.\n");
7714                 mr_cache_put(fib_work->men_info.mfc);
7715                 break;
7716         case FIB_EVENT_ENTRY_DEL:
7717                 mlxsw_sp_router_fibmr_del(mlxsw_sp, &fib_work->men_info);
7718                 mr_cache_put(fib_work->men_info.mfc);
7719                 break;
7720         case FIB_EVENT_VIF_ADD:
7721                 err = mlxsw_sp_router_fibmr_vif_add(mlxsw_sp,
7722                                                     &fib_work->ven_info);
7723                 if (err)
7724                         dev_warn(mlxsw_sp->bus_info->dev, "MR VIF add failed.\n");
7725                 netdev_put(fib_work->ven_info.dev, &fib_work->dev_tracker);
7726                 break;
7727         case FIB_EVENT_VIF_DEL:
7728                 mlxsw_sp_router_fibmr_vif_del(mlxsw_sp,
7729                                               &fib_work->ven_info);
7730                 netdev_put(fib_work->ven_info.dev, &fib_work->dev_tracker);
7731                 break;
7732         }
7733         mutex_unlock(&mlxsw_sp->router->lock);
7734         rtnl_unlock();
7735         kfree(fib_work);
7736 }
7737
7738 static void mlxsw_sp_router_fib4_event(struct mlxsw_sp_fib_event_work *fib_work,
7739                                        struct fib_notifier_info *info)
7740 {
7741         struct fib_entry_notifier_info *fen_info;
7742         struct fib_nh_notifier_info *fnh_info;
7743
7744         switch (fib_work->event) {
7745         case FIB_EVENT_ENTRY_REPLACE:
7746         case FIB_EVENT_ENTRY_DEL:
7747                 fen_info = container_of(info, struct fib_entry_notifier_info,
7748                                         info);
7749                 fib_work->fen_info = *fen_info;
7750                 /* Take reference on fib_info to prevent it from being
7751                  * freed while work is queued. Release it afterwards.
7752                  */
7753                 fib_info_hold(fib_work->fen_info.fi);
7754                 break;
7755         case FIB_EVENT_NH_ADD:
7756         case FIB_EVENT_NH_DEL:
7757                 fnh_info = container_of(info, struct fib_nh_notifier_info,
7758                                         info);
7759                 fib_work->fnh_info = *fnh_info;
7760                 fib_info_hold(fib_work->fnh_info.fib_nh->nh_parent);
7761                 break;
7762         }
7763 }
7764
7765 static int mlxsw_sp_router_fib6_event(struct mlxsw_sp_fib_event_work *fib_work,
7766                                       struct fib_notifier_info *info)
7767 {
7768         struct fib6_entry_notifier_info *fen6_info;
7769         int err;
7770
7771         switch (fib_work->event) {
7772         case FIB_EVENT_ENTRY_REPLACE:
7773         case FIB_EVENT_ENTRY_APPEND:
7774         case FIB_EVENT_ENTRY_DEL:
7775                 fen6_info = container_of(info, struct fib6_entry_notifier_info,
7776                                          info);
7777                 err = mlxsw_sp_router_fib6_work_init(&fib_work->fib6_work,
7778                                                      fen6_info);
7779                 if (err)
7780                         return err;
7781                 break;
7782         }
7783
7784         return 0;
7785 }
7786
7787 static void
7788 mlxsw_sp_router_fibmr_event(struct mlxsw_sp_fib_event_work *fib_work,
7789                             struct fib_notifier_info *info)
7790 {
7791         switch (fib_work->event) {
7792         case FIB_EVENT_ENTRY_REPLACE:
7793         case FIB_EVENT_ENTRY_ADD:
7794         case FIB_EVENT_ENTRY_DEL:
7795                 memcpy(&fib_work->men_info, info, sizeof(fib_work->men_info));
7796                 mr_cache_hold(fib_work->men_info.mfc);
7797                 break;
7798         case FIB_EVENT_VIF_ADD:
7799         case FIB_EVENT_VIF_DEL:
7800                 memcpy(&fib_work->ven_info, info, sizeof(fib_work->ven_info));
7801                 netdev_hold(fib_work->ven_info.dev, &fib_work->dev_tracker,
7802                             GFP_ATOMIC);
7803                 break;
7804         }
7805 }
7806
7807 static int mlxsw_sp_router_fib_rule_event(unsigned long event,
7808                                           struct fib_notifier_info *info,
7809                                           struct mlxsw_sp *mlxsw_sp)
7810 {
7811         struct netlink_ext_ack *extack = info->extack;
7812         struct fib_rule_notifier_info *fr_info;
7813         struct fib_rule *rule;
7814         int err = 0;
7815
7816         /* nothing to do at the moment */
7817         if (event == FIB_EVENT_RULE_DEL)
7818                 return 0;
7819
7820         fr_info = container_of(info, struct fib_rule_notifier_info, info);
7821         rule = fr_info->rule;
7822
7823         /* Rule only affects locally generated traffic */
7824         if (rule->iifindex == mlxsw_sp_net(mlxsw_sp)->loopback_dev->ifindex)
7825                 return 0;
7826
7827         switch (info->family) {
7828         case AF_INET:
7829                 if (!fib4_rule_default(rule) && !rule->l3mdev)
7830                         err = -EOPNOTSUPP;
7831                 break;
7832         case AF_INET6:
7833                 if (!fib6_rule_default(rule) && !rule->l3mdev)
7834                         err = -EOPNOTSUPP;
7835                 break;
7836         case RTNL_FAMILY_IPMR:
7837                 if (!ipmr_rule_default(rule) && !rule->l3mdev)
7838                         err = -EOPNOTSUPP;
7839                 break;
7840         case RTNL_FAMILY_IP6MR:
7841                 if (!ip6mr_rule_default(rule) && !rule->l3mdev)
7842                         err = -EOPNOTSUPP;
7843                 break;
7844         }
7845
7846         if (err < 0)
7847                 NL_SET_ERR_MSG_MOD(extack, "FIB rules not supported");
7848
7849         return err;
7850 }
7851
7852 /* Called with rcu_read_lock() */
7853 static int mlxsw_sp_router_fib_event(struct notifier_block *nb,
7854                                      unsigned long event, void *ptr)
7855 {
7856         struct mlxsw_sp_fib_event_work *fib_work;
7857         struct fib_notifier_info *info = ptr;
7858         struct mlxsw_sp_router *router;
7859         int err;
7860
7861         if ((info->family != AF_INET && info->family != AF_INET6 &&
7862              info->family != RTNL_FAMILY_IPMR &&
7863              info->family != RTNL_FAMILY_IP6MR))
7864                 return NOTIFY_DONE;
7865
7866         router = container_of(nb, struct mlxsw_sp_router, fib_nb);
7867
7868         switch (event) {
7869         case FIB_EVENT_RULE_ADD:
7870         case FIB_EVENT_RULE_DEL:
7871                 err = mlxsw_sp_router_fib_rule_event(event, info,
7872                                                      router->mlxsw_sp);
7873                 return notifier_from_errno(err);
7874         case FIB_EVENT_ENTRY_ADD:
7875         case FIB_EVENT_ENTRY_REPLACE:
7876         case FIB_EVENT_ENTRY_APPEND:
7877                 if (info->family == AF_INET) {
7878                         struct fib_entry_notifier_info *fen_info = ptr;
7879
7880                         if (fen_info->fi->fib_nh_is_v6) {
7881                                 NL_SET_ERR_MSG_MOD(info->extack, "IPv6 gateway with IPv4 route is not supported");
7882                                 return notifier_from_errno(-EINVAL);
7883                         }
7884                 }
7885                 break;
7886         }
7887
7888         fib_work = kzalloc(sizeof(*fib_work), GFP_ATOMIC);
7889         if (!fib_work)
7890                 return NOTIFY_BAD;
7891
7892         fib_work->mlxsw_sp = router->mlxsw_sp;
7893         fib_work->event = event;
7894
7895         switch (info->family) {
7896         case AF_INET:
7897                 INIT_WORK(&fib_work->work, mlxsw_sp_router_fib4_event_work);
7898                 mlxsw_sp_router_fib4_event(fib_work, info);
7899                 break;
7900         case AF_INET6:
7901                 INIT_WORK(&fib_work->work, mlxsw_sp_router_fib6_event_work);
7902                 err = mlxsw_sp_router_fib6_event(fib_work, info);
7903                 if (err)
7904                         goto err_fib_event;
7905                 break;
7906         case RTNL_FAMILY_IP6MR:
7907         case RTNL_FAMILY_IPMR:
7908                 INIT_WORK(&fib_work->work, mlxsw_sp_router_fibmr_event_work);
7909                 mlxsw_sp_router_fibmr_event(fib_work, info);
7910                 break;
7911         }
7912
7913         mlxsw_core_schedule_work(&fib_work->work);
7914
7915         return NOTIFY_DONE;
7916
7917 err_fib_event:
7918         kfree(fib_work);
7919         return NOTIFY_BAD;
7920 }
7921
7922 static struct mlxsw_sp_rif *
7923 mlxsw_sp_rif_find_by_dev(const struct mlxsw_sp *mlxsw_sp,
7924                          const struct net_device *dev)
7925 {
7926         int max_rifs = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_RIFS);
7927         int i;
7928
7929         for (i = 0; i < max_rifs; i++)
7930                 if (mlxsw_sp->router->rifs[i] &&
7931                     mlxsw_sp_rif_dev_is(mlxsw_sp->router->rifs[i], dev))
7932                         return mlxsw_sp->router->rifs[i];
7933
7934         return NULL;
7935 }
7936
7937 bool mlxsw_sp_rif_exists(struct mlxsw_sp *mlxsw_sp,
7938                          const struct net_device *dev)
7939 {
7940         struct mlxsw_sp_rif *rif;
7941
7942         mutex_lock(&mlxsw_sp->router->lock);
7943         rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
7944         mutex_unlock(&mlxsw_sp->router->lock);
7945
7946         return rif;
7947 }
7948
7949 u16 mlxsw_sp_rif_vid(struct mlxsw_sp *mlxsw_sp, const struct net_device *dev)
7950 {
7951         struct mlxsw_sp_rif *rif;
7952         u16 vid = 0;
7953
7954         mutex_lock(&mlxsw_sp->router->lock);
7955         rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
7956         if (!rif)
7957                 goto out;
7958
7959         /* We only return the VID for VLAN RIFs. Otherwise we return an
7960          * invalid value (0).
7961          */
7962         if (rif->ops->type != MLXSW_SP_RIF_TYPE_VLAN)
7963                 goto out;
7964
7965         vid = mlxsw_sp_fid_8021q_vid(rif->fid);
7966
7967 out:
7968         mutex_unlock(&mlxsw_sp->router->lock);
7969         return vid;
7970 }
7971
7972 static int mlxsw_sp_router_rif_disable(struct mlxsw_sp *mlxsw_sp, u16 rif)
7973 {
7974         char ritr_pl[MLXSW_REG_RITR_LEN];
7975         int err;
7976
7977         mlxsw_reg_ritr_rif_pack(ritr_pl, rif);
7978         err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
7979         if (err)
7980                 return err;
7981
7982         mlxsw_reg_ritr_enable_set(ritr_pl, false);
7983         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
7984 }
7985
7986 static int mlxsw_sp_router_rif_made_sync(struct mlxsw_sp *mlxsw_sp,
7987                                          struct mlxsw_sp_rif *rif)
7988 {
7989         int err;
7990
7991         err = mlxsw_sp_neigh_rif_made_sync(mlxsw_sp, rif);
7992         if (err)
7993                 return err;
7994
7995         err = mlxsw_sp_nexthop_rif_made_sync(mlxsw_sp, rif);
7996         if (err)
7997                 goto err_nexthop;
7998
7999         return 0;
8000
8001 err_nexthop:
8002         mlxsw_sp_neigh_rif_gone_sync(mlxsw_sp, rif);
8003         return err;
8004 }
8005
8006 static void mlxsw_sp_router_rif_gone_sync(struct mlxsw_sp *mlxsw_sp,
8007                                           struct mlxsw_sp_rif *rif)
8008 {
8009         /* Signal to nexthop cleanup that the RIF is going away. */
8010         rif->crif->rif = NULL;
8011
8012         mlxsw_sp_router_rif_disable(mlxsw_sp, rif->rif_index);
8013         mlxsw_sp_nexthop_rif_gone_sync(mlxsw_sp, rif);
8014         mlxsw_sp_neigh_rif_gone_sync(mlxsw_sp, rif);
8015 }
8016
8017 static bool __mlxsw_sp_dev_addr_list_empty(const struct net_device *dev)
8018 {
8019         struct inet6_dev *inet6_dev;
8020         struct in_device *idev;
8021
8022         idev = __in_dev_get_rcu(dev);
8023         if (idev && idev->ifa_list)
8024                 return false;
8025
8026         inet6_dev = __in6_dev_get(dev);
8027         if (inet6_dev && !list_empty(&inet6_dev->addr_list))
8028                 return false;
8029
8030         return true;
8031 }
8032
8033 static bool mlxsw_sp_dev_addr_list_empty(const struct net_device *dev)
8034 {
8035         bool addr_list_empty;
8036
8037         rcu_read_lock();
8038         addr_list_empty = __mlxsw_sp_dev_addr_list_empty(dev);
8039         rcu_read_unlock();
8040
8041         return addr_list_empty;
8042 }
8043
8044 static bool
8045 mlxsw_sp_rif_should_config(struct mlxsw_sp_rif *rif, struct net_device *dev,
8046                            unsigned long event)
8047 {
8048         bool addr_list_empty;
8049
8050         switch (event) {
8051         case NETDEV_UP:
8052                 return rif == NULL;
8053         case NETDEV_DOWN:
8054                 addr_list_empty = mlxsw_sp_dev_addr_list_empty(dev);
8055
8056                 /* macvlans do not have a RIF, but rather piggy back on the
8057                  * RIF of their lower device.
8058                  */
8059                 if (netif_is_macvlan(dev) && addr_list_empty)
8060                         return true;
8061
8062                 if (rif && addr_list_empty &&
8063                     !netif_is_l3_slave(mlxsw_sp_rif_dev(rif)))
8064                         return true;
8065                 /* It is possible we already removed the RIF ourselves
8066                  * if it was assigned to a netdev that is now a bridge
8067                  * or LAG slave.
8068                  */
8069                 return false;
8070         }
8071
8072         return false;
8073 }
8074
8075 static enum mlxsw_sp_rif_type
8076 mlxsw_sp_dev_rif_type(const struct mlxsw_sp *mlxsw_sp,
8077                       const struct net_device *dev)
8078 {
8079         enum mlxsw_sp_fid_type type;
8080
8081         if (mlxsw_sp_netdev_ipip_type(mlxsw_sp, dev, NULL))
8082                 return MLXSW_SP_RIF_TYPE_IPIP_LB;
8083
8084         /* Otherwise RIF type is derived from the type of the underlying FID. */
8085         if (is_vlan_dev(dev) && netif_is_bridge_master(vlan_dev_real_dev(dev)))
8086                 type = MLXSW_SP_FID_TYPE_8021Q;
8087         else if (netif_is_bridge_master(dev) && br_vlan_enabled(dev))
8088                 type = MLXSW_SP_FID_TYPE_8021Q;
8089         else if (netif_is_bridge_master(dev))
8090                 type = MLXSW_SP_FID_TYPE_8021D;
8091         else
8092                 type = MLXSW_SP_FID_TYPE_RFID;
8093
8094         return mlxsw_sp_fid_type_rif_type(mlxsw_sp, type);
8095 }
8096
8097 static int mlxsw_sp_rif_index_alloc(struct mlxsw_sp *mlxsw_sp, u16 *p_rif_index,
8098                                     u8 rif_entries)
8099 {
8100         *p_rif_index = gen_pool_alloc(mlxsw_sp->router->rifs_table,
8101                                       rif_entries);
8102         if (*p_rif_index == 0)
8103                 return -ENOBUFS;
8104         *p_rif_index -= MLXSW_SP_ROUTER_GENALLOC_OFFSET;
8105
8106         /* RIF indexes must be aligned to the allocation size. */
8107         WARN_ON_ONCE(*p_rif_index % rif_entries);
8108
8109         return 0;
8110 }
8111
8112 static void mlxsw_sp_rif_index_free(struct mlxsw_sp *mlxsw_sp, u16 rif_index,
8113                                     u8 rif_entries)
8114 {
8115         gen_pool_free(mlxsw_sp->router->rifs_table,
8116                       MLXSW_SP_ROUTER_GENALLOC_OFFSET + rif_index, rif_entries);
8117 }
8118
8119 static struct mlxsw_sp_rif *mlxsw_sp_rif_alloc(size_t rif_size, u16 rif_index,
8120                                                u16 vr_id,
8121                                                struct mlxsw_sp_crif *crif)
8122 {
8123         struct net_device *l3_dev = crif ? crif->key.dev : NULL;
8124         struct mlxsw_sp_rif *rif;
8125
8126         rif = kzalloc(rif_size, GFP_KERNEL);
8127         if (!rif)
8128                 return NULL;
8129
8130         INIT_LIST_HEAD(&rif->neigh_list);
8131         if (l3_dev) {
8132                 ether_addr_copy(rif->addr, l3_dev->dev_addr);
8133                 rif->mtu = l3_dev->mtu;
8134         }
8135         rif->vr_id = vr_id;
8136         rif->rif_index = rif_index;
8137         if (crif) {
8138                 rif->crif = crif;
8139                 crif->rif = rif;
8140         }
8141
8142         return rif;
8143 }
8144
8145 static void mlxsw_sp_rif_free(struct mlxsw_sp_rif *rif)
8146 {
8147         WARN_ON(!list_empty(&rif->neigh_list));
8148
8149         if (rif->crif)
8150                 rif->crif->rif = NULL;
8151         kfree(rif);
8152 }
8153
8154 struct mlxsw_sp_rif *mlxsw_sp_rif_by_index(const struct mlxsw_sp *mlxsw_sp,
8155                                            u16 rif_index)
8156 {
8157         return mlxsw_sp->router->rifs[rif_index];
8158 }
8159
8160 u16 mlxsw_sp_rif_index(const struct mlxsw_sp_rif *rif)
8161 {
8162         return rif->rif_index;
8163 }
8164
8165 u16 mlxsw_sp_ipip_lb_rif_index(const struct mlxsw_sp_rif_ipip_lb *lb_rif)
8166 {
8167         return lb_rif->common.rif_index;
8168 }
8169
8170 u16 mlxsw_sp_ipip_lb_ul_vr_id(const struct mlxsw_sp_rif_ipip_lb *lb_rif)
8171 {
8172         struct net_device *dev = mlxsw_sp_rif_dev(&lb_rif->common);
8173         u32 ul_tb_id = mlxsw_sp_ipip_dev_ul_tb_id(dev);
8174         struct mlxsw_sp_vr *ul_vr;
8175
8176         ul_vr = mlxsw_sp_vr_get(lb_rif->common.mlxsw_sp, ul_tb_id, NULL);
8177         if (WARN_ON(IS_ERR(ul_vr)))
8178                 return 0;
8179
8180         return ul_vr->id;
8181 }
8182
8183 u16 mlxsw_sp_ipip_lb_ul_rif_id(const struct mlxsw_sp_rif_ipip_lb *lb_rif)
8184 {
8185         return lb_rif->ul_rif_id;
8186 }
8187
8188 static bool
8189 mlxsw_sp_router_port_l3_stats_enabled(struct mlxsw_sp_rif *rif)
8190 {
8191         return mlxsw_sp_rif_counter_valid_get(rif,
8192                                               MLXSW_SP_RIF_COUNTER_EGRESS) &&
8193                mlxsw_sp_rif_counter_valid_get(rif,
8194                                               MLXSW_SP_RIF_COUNTER_INGRESS);
8195 }
8196
8197 static int
8198 mlxsw_sp_router_port_l3_stats_enable(struct mlxsw_sp_rif *rif)
8199 {
8200         int err;
8201
8202         err = mlxsw_sp_rif_counter_alloc(rif, MLXSW_SP_RIF_COUNTER_INGRESS);
8203         if (err)
8204                 return err;
8205
8206         /* Clear stale data. */
8207         err = mlxsw_sp_rif_counter_fetch_clear(rif,
8208                                                MLXSW_SP_RIF_COUNTER_INGRESS,
8209                                                NULL);
8210         if (err)
8211                 goto err_clear_ingress;
8212
8213         err = mlxsw_sp_rif_counter_alloc(rif, MLXSW_SP_RIF_COUNTER_EGRESS);
8214         if (err)
8215                 goto err_alloc_egress;
8216
8217         /* Clear stale data. */
8218         err = mlxsw_sp_rif_counter_fetch_clear(rif,
8219                                                MLXSW_SP_RIF_COUNTER_EGRESS,
8220                                                NULL);
8221         if (err)
8222                 goto err_clear_egress;
8223
8224         return 0;
8225
8226 err_clear_egress:
8227         mlxsw_sp_rif_counter_free(rif, MLXSW_SP_RIF_COUNTER_EGRESS);
8228 err_alloc_egress:
8229 err_clear_ingress:
8230         mlxsw_sp_rif_counter_free(rif, MLXSW_SP_RIF_COUNTER_INGRESS);
8231         return err;
8232 }
8233
8234 static void
8235 mlxsw_sp_router_port_l3_stats_disable(struct mlxsw_sp_rif *rif)
8236 {
8237         mlxsw_sp_rif_counter_free(rif, MLXSW_SP_RIF_COUNTER_EGRESS);
8238         mlxsw_sp_rif_counter_free(rif, MLXSW_SP_RIF_COUNTER_INGRESS);
8239 }
8240
8241 static void
8242 mlxsw_sp_router_port_l3_stats_report_used(struct mlxsw_sp_rif *rif,
8243                                           struct netdev_notifier_offload_xstats_info *info)
8244 {
8245         if (!mlxsw_sp_router_port_l3_stats_enabled(rif))
8246                 return;
8247         netdev_offload_xstats_report_used(info->report_used);
8248 }
8249
8250 static int
8251 mlxsw_sp_router_port_l3_stats_fetch(struct mlxsw_sp_rif *rif,
8252                                     struct rtnl_hw_stats64 *p_stats)
8253 {
8254         struct mlxsw_sp_rif_counter_set_basic ingress;
8255         struct mlxsw_sp_rif_counter_set_basic egress;
8256         int err;
8257
8258         err = mlxsw_sp_rif_counter_fetch_clear(rif,
8259                                                MLXSW_SP_RIF_COUNTER_INGRESS,
8260                                                &ingress);
8261         if (err)
8262                 return err;
8263
8264         err = mlxsw_sp_rif_counter_fetch_clear(rif,
8265                                                MLXSW_SP_RIF_COUNTER_EGRESS,
8266                                                &egress);
8267         if (err)
8268                 return err;
8269
8270 #define MLXSW_SP_ROUTER_ALL_GOOD(SET, SFX)              \
8271                 ((SET.good_unicast_ ## SFX) +           \
8272                  (SET.good_multicast_ ## SFX) +         \
8273                  (SET.good_broadcast_ ## SFX))
8274
8275         p_stats->rx_packets = MLXSW_SP_ROUTER_ALL_GOOD(ingress, packets);
8276         p_stats->tx_packets = MLXSW_SP_ROUTER_ALL_GOOD(egress, packets);
8277         p_stats->rx_bytes = MLXSW_SP_ROUTER_ALL_GOOD(ingress, bytes);
8278         p_stats->tx_bytes = MLXSW_SP_ROUTER_ALL_GOOD(egress, bytes);
8279         p_stats->rx_errors = ingress.error_packets;
8280         p_stats->tx_errors = egress.error_packets;
8281         p_stats->rx_dropped = ingress.discard_packets;
8282         p_stats->tx_dropped = egress.discard_packets;
8283         p_stats->multicast = ingress.good_multicast_packets +
8284                              ingress.good_broadcast_packets;
8285
8286 #undef MLXSW_SP_ROUTER_ALL_GOOD
8287
8288         return 0;
8289 }
8290
8291 static int
8292 mlxsw_sp_router_port_l3_stats_report_delta(struct mlxsw_sp_rif *rif,
8293                                            struct netdev_notifier_offload_xstats_info *info)
8294 {
8295         struct rtnl_hw_stats64 stats = {};
8296         int err;
8297
8298         if (!mlxsw_sp_router_port_l3_stats_enabled(rif))
8299                 return 0;
8300
8301         err = mlxsw_sp_router_port_l3_stats_fetch(rif, &stats);
8302         if (err)
8303                 return err;
8304
8305         netdev_offload_xstats_report_delta(info->report_delta, &stats);
8306         return 0;
8307 }
8308
8309 struct mlxsw_sp_router_hwstats_notify_work {
8310         struct work_struct work;
8311         struct net_device *dev;
8312         netdevice_tracker dev_tracker;
8313 };
8314
8315 static void mlxsw_sp_router_hwstats_notify_work(struct work_struct *work)
8316 {
8317         struct mlxsw_sp_router_hwstats_notify_work *hws_work =
8318                 container_of(work, struct mlxsw_sp_router_hwstats_notify_work,
8319                              work);
8320
8321         rtnl_lock();
8322         rtnl_offload_xstats_notify(hws_work->dev);
8323         rtnl_unlock();
8324         netdev_put(hws_work->dev, &hws_work->dev_tracker);
8325         kfree(hws_work);
8326 }
8327
8328 static void
8329 mlxsw_sp_router_hwstats_notify_schedule(struct net_device *dev)
8330 {
8331         struct mlxsw_sp_router_hwstats_notify_work *hws_work;
8332
8333         /* To collect notification payload, the core ends up sending another
8334          * notifier block message, which would deadlock on the attempt to
8335          * acquire the router lock again. Just postpone the notification until
8336          * later.
8337          */
8338
8339         hws_work = kzalloc(sizeof(*hws_work), GFP_KERNEL);
8340         if (!hws_work)
8341                 return;
8342
8343         INIT_WORK(&hws_work->work, mlxsw_sp_router_hwstats_notify_work);
8344         netdev_hold(dev, &hws_work->dev_tracker, GFP_KERNEL);
8345         hws_work->dev = dev;
8346         mlxsw_core_schedule_work(&hws_work->work);
8347 }
8348
8349 int mlxsw_sp_rif_dev_ifindex(const struct mlxsw_sp_rif *rif)
8350 {
8351         return mlxsw_sp_rif_dev(rif)->ifindex;
8352 }
8353
8354 bool mlxsw_sp_rif_has_dev(const struct mlxsw_sp_rif *rif)
8355 {
8356         return !!mlxsw_sp_rif_dev(rif);
8357 }
8358
8359 bool mlxsw_sp_rif_dev_is(const struct mlxsw_sp_rif *rif,
8360                          const struct net_device *dev)
8361 {
8362         return mlxsw_sp_rif_dev(rif) == dev;
8363 }
8364
8365 static void mlxsw_sp_rif_push_l3_stats(struct mlxsw_sp_rif *rif)
8366 {
8367         struct rtnl_hw_stats64 stats = {};
8368
8369         if (!mlxsw_sp_router_port_l3_stats_fetch(rif, &stats))
8370                 netdev_offload_xstats_push_delta(mlxsw_sp_rif_dev(rif),
8371                                                  NETDEV_OFFLOAD_XSTATS_TYPE_L3,
8372                                                  &stats);
8373 }
8374
8375 static struct mlxsw_sp_rif *
8376 mlxsw_sp_rif_create(struct mlxsw_sp *mlxsw_sp,
8377                     const struct mlxsw_sp_rif_params *params,
8378                     struct netlink_ext_ack *extack)
8379 {
8380         u8 rif_entries = params->double_entry ? 2 : 1;
8381         u32 tb_id = l3mdev_fib_table(params->dev);
8382         const struct mlxsw_sp_rif_ops *ops;
8383         struct mlxsw_sp_fid *fid = NULL;
8384         enum mlxsw_sp_rif_type type;
8385         struct mlxsw_sp_crif *crif;
8386         struct mlxsw_sp_rif *rif;
8387         struct mlxsw_sp_vr *vr;
8388         u16 rif_index;
8389         int i, err;
8390
8391         type = mlxsw_sp_dev_rif_type(mlxsw_sp, params->dev);
8392         ops = mlxsw_sp->router->rif_ops_arr[type];
8393
8394         vr = mlxsw_sp_vr_get(mlxsw_sp, tb_id ? : RT_TABLE_MAIN, extack);
8395         if (IS_ERR(vr))
8396                 return ERR_CAST(vr);
8397         vr->rif_count++;
8398
8399         err = mlxsw_sp_rif_index_alloc(mlxsw_sp, &rif_index, rif_entries);
8400         if (err) {
8401                 NL_SET_ERR_MSG_MOD(extack, "Exceeded number of supported router interfaces");
8402                 goto err_rif_index_alloc;
8403         }
8404
8405         crif = mlxsw_sp_crif_lookup(mlxsw_sp->router, params->dev);
8406         if (WARN_ON(!crif)) {
8407                 err = -ENOENT;
8408                 goto err_crif_lookup;
8409         }
8410
8411         rif = mlxsw_sp_rif_alloc(ops->rif_size, rif_index, vr->id, crif);
8412         if (!rif) {
8413                 err = -ENOMEM;
8414                 goto err_rif_alloc;
8415         }
8416         netdev_hold(params->dev, &rif->dev_tracker, GFP_KERNEL);
8417         mlxsw_sp->router->rifs[rif_index] = rif;
8418         rif->mlxsw_sp = mlxsw_sp;
8419         rif->ops = ops;
8420         rif->rif_entries = rif_entries;
8421
8422         if (ops->fid_get) {
8423                 fid = ops->fid_get(rif, params, extack);
8424                 if (IS_ERR(fid)) {
8425                         err = PTR_ERR(fid);
8426                         goto err_fid_get;
8427                 }
8428                 rif->fid = fid;
8429         }
8430
8431         if (ops->setup)
8432                 ops->setup(rif, params);
8433
8434         err = ops->configure(rif, extack);
8435         if (err)
8436                 goto err_configure;
8437
8438         for (i = 0; i < MLXSW_SP_L3_PROTO_MAX; i++) {
8439                 err = mlxsw_sp_mr_rif_add(vr->mr_table[i], rif);
8440                 if (err)
8441                         goto err_mr_rif_add;
8442         }
8443
8444         err = mlxsw_sp_router_rif_made_sync(mlxsw_sp, rif);
8445         if (err)
8446                 goto err_rif_made_sync;
8447
8448         if (netdev_offload_xstats_enabled(params->dev,
8449                                           NETDEV_OFFLOAD_XSTATS_TYPE_L3)) {
8450                 err = mlxsw_sp_router_port_l3_stats_enable(rif);
8451                 if (err)
8452                         goto err_stats_enable;
8453                 mlxsw_sp_router_hwstats_notify_schedule(params->dev);
8454         } else {
8455                 mlxsw_sp_rif_counters_alloc(rif);
8456         }
8457
8458         atomic_add(rif_entries, &mlxsw_sp->router->rifs_count);
8459         return rif;
8460
8461 err_stats_enable:
8462         mlxsw_sp_router_rif_gone_sync(mlxsw_sp, rif);
8463 err_rif_made_sync:
8464 err_mr_rif_add:
8465         for (i--; i >= 0; i--)
8466                 mlxsw_sp_mr_rif_del(vr->mr_table[i], rif);
8467         ops->deconfigure(rif);
8468 err_configure:
8469         if (fid)
8470                 mlxsw_sp_fid_put(fid);
8471 err_fid_get:
8472         mlxsw_sp->router->rifs[rif_index] = NULL;
8473         netdev_put(params->dev, &rif->dev_tracker);
8474         mlxsw_sp_rif_free(rif);
8475 err_rif_alloc:
8476 err_crif_lookup:
8477         mlxsw_sp_rif_index_free(mlxsw_sp, rif_index, rif_entries);
8478 err_rif_index_alloc:
8479         vr->rif_count--;
8480         mlxsw_sp_vr_put(mlxsw_sp, vr);
8481         return ERR_PTR(err);
8482 }
8483
8484 static void mlxsw_sp_rif_destroy(struct mlxsw_sp_rif *rif)
8485 {
8486         struct net_device *dev = mlxsw_sp_rif_dev(rif);
8487         const struct mlxsw_sp_rif_ops *ops = rif->ops;
8488         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
8489         struct mlxsw_sp_crif *crif = rif->crif;
8490         struct mlxsw_sp_fid *fid = rif->fid;
8491         u8 rif_entries = rif->rif_entries;
8492         u16 rif_index = rif->rif_index;
8493         struct mlxsw_sp_vr *vr;
8494         int i;
8495
8496         atomic_sub(rif_entries, &mlxsw_sp->router->rifs_count);
8497         mlxsw_sp_router_rif_gone_sync(mlxsw_sp, rif);
8498         vr = &mlxsw_sp->router->vrs[rif->vr_id];
8499
8500         if (netdev_offload_xstats_enabled(dev, NETDEV_OFFLOAD_XSTATS_TYPE_L3)) {
8501                 mlxsw_sp_rif_push_l3_stats(rif);
8502                 mlxsw_sp_router_port_l3_stats_disable(rif);
8503                 mlxsw_sp_router_hwstats_notify_schedule(dev);
8504         } else {
8505                 mlxsw_sp_rif_counters_free(rif);
8506         }
8507
8508         for (i = 0; i < MLXSW_SP_L3_PROTO_MAX; i++)
8509                 mlxsw_sp_mr_rif_del(vr->mr_table[i], rif);
8510         ops->deconfigure(rif);
8511         if (fid)
8512                 /* Loopback RIFs are not associated with a FID. */
8513                 mlxsw_sp_fid_put(fid);
8514         mlxsw_sp->router->rifs[rif->rif_index] = NULL;
8515         netdev_put(dev, &rif->dev_tracker);
8516         mlxsw_sp_rif_free(rif);
8517         mlxsw_sp_rif_index_free(mlxsw_sp, rif_index, rif_entries);
8518         vr->rif_count--;
8519         mlxsw_sp_vr_put(mlxsw_sp, vr);
8520
8521         if (crif->can_destroy)
8522                 mlxsw_sp_crif_free(crif);
8523 }
8524
8525 void mlxsw_sp_rif_destroy_by_dev(struct mlxsw_sp *mlxsw_sp,
8526                                  struct net_device *dev)
8527 {
8528         struct mlxsw_sp_rif *rif;
8529
8530         mutex_lock(&mlxsw_sp->router->lock);
8531         rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
8532         if (!rif)
8533                 goto out;
8534         mlxsw_sp_rif_destroy(rif);
8535 out:
8536         mutex_unlock(&mlxsw_sp->router->lock);
8537 }
8538
8539 static void mlxsw_sp_rif_destroy_vlan_upper(struct mlxsw_sp *mlxsw_sp,
8540                                             struct net_device *br_dev,
8541                                             u16 vid)
8542 {
8543         struct net_device *upper_dev;
8544         struct mlxsw_sp_crif *crif;
8545
8546         rcu_read_lock();
8547         upper_dev = __vlan_find_dev_deep_rcu(br_dev, htons(ETH_P_8021Q), vid);
8548         rcu_read_unlock();
8549
8550         if (!upper_dev)
8551                 return;
8552
8553         crif = mlxsw_sp_crif_lookup(mlxsw_sp->router, upper_dev);
8554         if (!crif || !crif->rif)
8555                 return;
8556
8557         mlxsw_sp_rif_destroy(crif->rif);
8558 }
8559
8560 static int mlxsw_sp_inetaddr_bridge_event(struct mlxsw_sp *mlxsw_sp,
8561                                           struct net_device *l3_dev,
8562                                           int lower_pvid,
8563                                           unsigned long event,
8564                                           struct netlink_ext_ack *extack);
8565
8566 int mlxsw_sp_router_bridge_vlan_add(struct mlxsw_sp *mlxsw_sp,
8567                                     struct net_device *br_dev,
8568                                     u16 new_vid, bool is_pvid,
8569                                     struct netlink_ext_ack *extack)
8570 {
8571         struct mlxsw_sp_rif *old_rif;
8572         struct mlxsw_sp_rif *new_rif;
8573         struct net_device *upper_dev;
8574         u16 old_pvid = 0;
8575         u16 new_pvid;
8576         int err = 0;
8577
8578         mutex_lock(&mlxsw_sp->router->lock);
8579         old_rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, br_dev);
8580         if (old_rif) {
8581                 /* If the RIF on the bridge is not a VLAN RIF, we shouldn't have
8582                  * gotten a PVID notification.
8583                  */
8584                 if (WARN_ON(old_rif->ops->type != MLXSW_SP_RIF_TYPE_VLAN))
8585                         old_rif = NULL;
8586                 else
8587                         old_pvid = mlxsw_sp_fid_8021q_vid(old_rif->fid);
8588         }
8589
8590         if (is_pvid)
8591                 new_pvid = new_vid;
8592         else if (old_pvid == new_vid)
8593                 new_pvid = 0;
8594         else
8595                 goto out;
8596
8597         if (old_pvid == new_pvid)
8598                 goto out;
8599
8600         if (new_pvid) {
8601                 struct mlxsw_sp_rif_params params = {
8602                         .dev = br_dev,
8603                         .vid = new_pvid,
8604                 };
8605
8606                 /* If there is a VLAN upper with the same VID as the new PVID,
8607                  * kill its RIF, if there is one.
8608                  */
8609                 mlxsw_sp_rif_destroy_vlan_upper(mlxsw_sp, br_dev, new_pvid);
8610
8611                 if (mlxsw_sp_dev_addr_list_empty(br_dev))
8612                         goto out;
8613                 new_rif = mlxsw_sp_rif_create(mlxsw_sp, &params, extack);
8614                 if (IS_ERR(new_rif)) {
8615                         err = PTR_ERR(new_rif);
8616                         goto out;
8617                 }
8618
8619                 if (old_pvid)
8620                         mlxsw_sp_rif_migrate_destroy(mlxsw_sp, old_rif, new_rif,
8621                                                      true);
8622         } else {
8623                 mlxsw_sp_rif_destroy(old_rif);
8624         }
8625
8626         if (old_pvid) {
8627                 rcu_read_lock();
8628                 upper_dev = __vlan_find_dev_deep_rcu(br_dev, htons(ETH_P_8021Q),
8629                                                      old_pvid);
8630                 rcu_read_unlock();
8631                 if (upper_dev)
8632                         err = mlxsw_sp_inetaddr_bridge_event(mlxsw_sp,
8633                                                              upper_dev,
8634                                                              new_pvid,
8635                                                              NETDEV_UP, extack);
8636         }
8637
8638 out:
8639         mutex_unlock(&mlxsw_sp->router->lock);
8640         return err;
8641 }
8642
8643 static void
8644 mlxsw_sp_rif_subport_params_init(struct mlxsw_sp_rif_params *params,
8645                                  struct mlxsw_sp_port_vlan *mlxsw_sp_port_vlan)
8646 {
8647         struct mlxsw_sp_port *mlxsw_sp_port = mlxsw_sp_port_vlan->mlxsw_sp_port;
8648
8649         params->vid = mlxsw_sp_port_vlan->vid;
8650         params->lag = mlxsw_sp_port->lagged;
8651         if (params->lag)
8652                 params->lag_id = mlxsw_sp_port->lag_id;
8653         else
8654                 params->system_port = mlxsw_sp_port->local_port;
8655 }
8656
8657 static struct mlxsw_sp_rif_subport *
8658 mlxsw_sp_rif_subport_rif(const struct mlxsw_sp_rif *rif)
8659 {
8660         return container_of(rif, struct mlxsw_sp_rif_subport, common);
8661 }
8662
8663 static struct mlxsw_sp_rif *
8664 mlxsw_sp_rif_subport_get(struct mlxsw_sp *mlxsw_sp,
8665                          const struct mlxsw_sp_rif_params *params,
8666                          struct netlink_ext_ack *extack)
8667 {
8668         struct mlxsw_sp_rif_subport *rif_subport;
8669         struct mlxsw_sp_rif *rif;
8670
8671         rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, params->dev);
8672         if (!rif)
8673                 return mlxsw_sp_rif_create(mlxsw_sp, params, extack);
8674
8675         rif_subport = mlxsw_sp_rif_subport_rif(rif);
8676         refcount_inc(&rif_subport->ref_count);
8677         return rif;
8678 }
8679
8680 static void mlxsw_sp_rif_subport_put(struct mlxsw_sp_rif *rif)
8681 {
8682         struct mlxsw_sp_rif_subport *rif_subport;
8683
8684         rif_subport = mlxsw_sp_rif_subport_rif(rif);
8685         if (!refcount_dec_and_test(&rif_subport->ref_count))
8686                 return;
8687
8688         mlxsw_sp_rif_destroy(rif);
8689 }
8690
8691 static int mlxsw_sp_rif_mac_profile_index_alloc(struct mlxsw_sp *mlxsw_sp,
8692                                                 struct mlxsw_sp_rif_mac_profile *profile,
8693                                                 struct netlink_ext_ack *extack)
8694 {
8695         u8 max_rif_mac_profiles = mlxsw_sp->router->max_rif_mac_profile;
8696         struct mlxsw_sp_router *router = mlxsw_sp->router;
8697         int id;
8698
8699         id = idr_alloc(&router->rif_mac_profiles_idr, profile, 0,
8700                        max_rif_mac_profiles, GFP_KERNEL);
8701
8702         if (id >= 0) {
8703                 profile->id = id;
8704                 return 0;
8705         }
8706
8707         if (id == -ENOSPC)
8708                 NL_SET_ERR_MSG_MOD(extack,
8709                                    "Exceeded number of supported router interface MAC profiles");
8710
8711         return id;
8712 }
8713
8714 static struct mlxsw_sp_rif_mac_profile *
8715 mlxsw_sp_rif_mac_profile_index_free(struct mlxsw_sp *mlxsw_sp, u8 mac_profile)
8716 {
8717         struct mlxsw_sp_rif_mac_profile *profile;
8718
8719         profile = idr_remove(&mlxsw_sp->router->rif_mac_profiles_idr,
8720                              mac_profile);
8721         WARN_ON(!profile);
8722         return profile;
8723 }
8724
8725 static struct mlxsw_sp_rif_mac_profile *
8726 mlxsw_sp_rif_mac_profile_alloc(const char *mac)
8727 {
8728         struct mlxsw_sp_rif_mac_profile *profile;
8729
8730         profile = kzalloc(sizeof(*profile), GFP_KERNEL);
8731         if (!profile)
8732                 return NULL;
8733
8734         ether_addr_copy(profile->mac_prefix, mac);
8735         refcount_set(&profile->ref_count, 1);
8736         return profile;
8737 }
8738
8739 static struct mlxsw_sp_rif_mac_profile *
8740 mlxsw_sp_rif_mac_profile_find(const struct mlxsw_sp *mlxsw_sp, const char *mac)
8741 {
8742         struct mlxsw_sp_router *router = mlxsw_sp->router;
8743         struct mlxsw_sp_rif_mac_profile *profile;
8744         int id;
8745
8746         idr_for_each_entry(&router->rif_mac_profiles_idr, profile, id) {
8747                 if (ether_addr_equal_masked(profile->mac_prefix, mac,
8748                                             mlxsw_sp->mac_mask))
8749                         return profile;
8750         }
8751
8752         return NULL;
8753 }
8754
8755 static u64 mlxsw_sp_rif_mac_profiles_occ_get(void *priv)
8756 {
8757         const struct mlxsw_sp *mlxsw_sp = priv;
8758
8759         return atomic_read(&mlxsw_sp->router->rif_mac_profiles_count);
8760 }
8761
8762 static u64 mlxsw_sp_rifs_occ_get(void *priv)
8763 {
8764         const struct mlxsw_sp *mlxsw_sp = priv;
8765
8766         return atomic_read(&mlxsw_sp->router->rifs_count);
8767 }
8768
8769 static struct mlxsw_sp_rif_mac_profile *
8770 mlxsw_sp_rif_mac_profile_create(struct mlxsw_sp *mlxsw_sp, const char *mac,
8771                                 struct netlink_ext_ack *extack)
8772 {
8773         struct mlxsw_sp_rif_mac_profile *profile;
8774         int err;
8775
8776         profile = mlxsw_sp_rif_mac_profile_alloc(mac);
8777         if (!profile)
8778                 return ERR_PTR(-ENOMEM);
8779
8780         err = mlxsw_sp_rif_mac_profile_index_alloc(mlxsw_sp, profile, extack);
8781         if (err)
8782                 goto profile_index_alloc_err;
8783
8784         atomic_inc(&mlxsw_sp->router->rif_mac_profiles_count);
8785         return profile;
8786
8787 profile_index_alloc_err:
8788         kfree(profile);
8789         return ERR_PTR(err);
8790 }
8791
8792 static void mlxsw_sp_rif_mac_profile_destroy(struct mlxsw_sp *mlxsw_sp,
8793                                              u8 mac_profile)
8794 {
8795         struct mlxsw_sp_rif_mac_profile *profile;
8796
8797         atomic_dec(&mlxsw_sp->router->rif_mac_profiles_count);
8798         profile = mlxsw_sp_rif_mac_profile_index_free(mlxsw_sp, mac_profile);
8799         kfree(profile);
8800 }
8801
8802 static int mlxsw_sp_rif_mac_profile_get(struct mlxsw_sp *mlxsw_sp,
8803                                         const char *mac, u8 *p_mac_profile,
8804                                         struct netlink_ext_ack *extack)
8805 {
8806         struct mlxsw_sp_rif_mac_profile *profile;
8807
8808         profile = mlxsw_sp_rif_mac_profile_find(mlxsw_sp, mac);
8809         if (profile) {
8810                 refcount_inc(&profile->ref_count);
8811                 goto out;
8812         }
8813
8814         profile = mlxsw_sp_rif_mac_profile_create(mlxsw_sp, mac, extack);
8815         if (IS_ERR(profile))
8816                 return PTR_ERR(profile);
8817
8818 out:
8819         *p_mac_profile = profile->id;
8820         return 0;
8821 }
8822
8823 static void mlxsw_sp_rif_mac_profile_put(struct mlxsw_sp *mlxsw_sp,
8824                                          u8 mac_profile)
8825 {
8826         struct mlxsw_sp_rif_mac_profile *profile;
8827
8828         profile = idr_find(&mlxsw_sp->router->rif_mac_profiles_idr,
8829                            mac_profile);
8830         if (WARN_ON(!profile))
8831                 return;
8832
8833         if (!refcount_dec_and_test(&profile->ref_count))
8834                 return;
8835
8836         mlxsw_sp_rif_mac_profile_destroy(mlxsw_sp, mac_profile);
8837 }
8838
8839 static bool mlxsw_sp_rif_mac_profile_is_shared(const struct mlxsw_sp_rif *rif)
8840 {
8841         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
8842         struct mlxsw_sp_rif_mac_profile *profile;
8843
8844         profile = idr_find(&mlxsw_sp->router->rif_mac_profiles_idr,
8845                            rif->mac_profile_id);
8846         if (WARN_ON(!profile))
8847                 return false;
8848
8849         return refcount_read(&profile->ref_count) > 1;
8850 }
8851
8852 static int mlxsw_sp_rif_mac_profile_edit(struct mlxsw_sp_rif *rif,
8853                                          const char *new_mac)
8854 {
8855         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
8856         struct mlxsw_sp_rif_mac_profile *profile;
8857
8858         profile = idr_find(&mlxsw_sp->router->rif_mac_profiles_idr,
8859                            rif->mac_profile_id);
8860         if (WARN_ON(!profile))
8861                 return -EINVAL;
8862
8863         ether_addr_copy(profile->mac_prefix, new_mac);
8864         return 0;
8865 }
8866
8867 static int
8868 mlxsw_sp_rif_mac_profile_replace(struct mlxsw_sp *mlxsw_sp,
8869                                  struct mlxsw_sp_rif *rif,
8870                                  const char *new_mac,
8871                                  struct netlink_ext_ack *extack)
8872 {
8873         u8 mac_profile;
8874         int err;
8875
8876         if (!mlxsw_sp_rif_mac_profile_is_shared(rif) &&
8877             !mlxsw_sp_rif_mac_profile_find(mlxsw_sp, new_mac))
8878                 return mlxsw_sp_rif_mac_profile_edit(rif, new_mac);
8879
8880         err = mlxsw_sp_rif_mac_profile_get(mlxsw_sp, new_mac,
8881                                            &mac_profile, extack);
8882         if (err)
8883                 return err;
8884
8885         mlxsw_sp_rif_mac_profile_put(mlxsw_sp, rif->mac_profile_id);
8886         rif->mac_profile_id = mac_profile;
8887         return 0;
8888 }
8889
8890 static int
8891 __mlxsw_sp_port_vlan_router_join(struct mlxsw_sp_port_vlan *mlxsw_sp_port_vlan,
8892                                  struct net_device *l3_dev,
8893                                  struct netlink_ext_ack *extack)
8894 {
8895         struct mlxsw_sp_port *mlxsw_sp_port = mlxsw_sp_port_vlan->mlxsw_sp_port;
8896         struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp;
8897         struct mlxsw_sp_rif_params params;
8898         u16 vid = mlxsw_sp_port_vlan->vid;
8899         struct mlxsw_sp_rif *rif;
8900         struct mlxsw_sp_fid *fid;
8901         int err;
8902
8903         params = (struct mlxsw_sp_rif_params) {
8904                 .dev = l3_dev,
8905                 .vid = vid,
8906         };
8907
8908         mlxsw_sp_rif_subport_params_init(&params, mlxsw_sp_port_vlan);
8909         rif = mlxsw_sp_rif_subport_get(mlxsw_sp, &params, extack);
8910         if (IS_ERR(rif))
8911                 return PTR_ERR(rif);
8912
8913         /* FID was already created, just take a reference */
8914         fid = rif->ops->fid_get(rif, &params, extack);
8915         err = mlxsw_sp_fid_port_vid_map(fid, mlxsw_sp_port, vid);
8916         if (err)
8917                 goto err_fid_port_vid_map;
8918
8919         err = mlxsw_sp_port_vid_learning_set(mlxsw_sp_port, vid, false);
8920         if (err)
8921                 goto err_port_vid_learning_set;
8922
8923         err = mlxsw_sp_port_vid_stp_set(mlxsw_sp_port, vid,
8924                                         BR_STATE_FORWARDING);
8925         if (err)
8926                 goto err_port_vid_stp_set;
8927
8928         mlxsw_sp_port_vlan->fid = fid;
8929
8930         return 0;
8931
8932 err_port_vid_stp_set:
8933         mlxsw_sp_port_vid_learning_set(mlxsw_sp_port, vid, true);
8934 err_port_vid_learning_set:
8935         mlxsw_sp_fid_port_vid_unmap(fid, mlxsw_sp_port, vid);
8936 err_fid_port_vid_map:
8937         mlxsw_sp_fid_put(fid);
8938         mlxsw_sp_rif_subport_put(rif);
8939         return err;
8940 }
8941
8942 static void
8943 __mlxsw_sp_port_vlan_router_leave(struct mlxsw_sp_port_vlan *mlxsw_sp_port_vlan)
8944 {
8945         struct mlxsw_sp_port *mlxsw_sp_port = mlxsw_sp_port_vlan->mlxsw_sp_port;
8946         struct mlxsw_sp_fid *fid = mlxsw_sp_port_vlan->fid;
8947         struct mlxsw_sp_rif *rif = mlxsw_sp_fid_rif(fid);
8948         u16 vid = mlxsw_sp_port_vlan->vid;
8949
8950         if (WARN_ON(mlxsw_sp_fid_type(fid) != MLXSW_SP_FID_TYPE_RFID))
8951                 return;
8952
8953         mlxsw_sp_port_vlan->fid = NULL;
8954         mlxsw_sp_port_vid_stp_set(mlxsw_sp_port, vid, BR_STATE_BLOCKING);
8955         mlxsw_sp_port_vid_learning_set(mlxsw_sp_port, vid, true);
8956         mlxsw_sp_fid_port_vid_unmap(fid, mlxsw_sp_port, vid);
8957         mlxsw_sp_fid_put(fid);
8958         mlxsw_sp_rif_subport_put(rif);
8959 }
8960
8961 static int
8962 mlxsw_sp_port_vlan_router_join_existing(struct mlxsw_sp_port_vlan *mlxsw_sp_port_vlan,
8963                                         struct net_device *l3_dev,
8964                                         struct netlink_ext_ack *extack)
8965 {
8966         struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port_vlan->mlxsw_sp_port->mlxsw_sp;
8967
8968         lockdep_assert_held(&mlxsw_sp->router->lock);
8969
8970         if (!mlxsw_sp_rif_find_by_dev(mlxsw_sp, l3_dev))
8971                 return 0;
8972
8973         return __mlxsw_sp_port_vlan_router_join(mlxsw_sp_port_vlan, l3_dev,
8974                                                 extack);
8975 }
8976
8977 void
8978 mlxsw_sp_port_vlan_router_leave(struct mlxsw_sp_port_vlan *mlxsw_sp_port_vlan)
8979 {
8980         struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port_vlan->mlxsw_sp_port->mlxsw_sp;
8981
8982         mutex_lock(&mlxsw_sp->router->lock);
8983         __mlxsw_sp_port_vlan_router_leave(mlxsw_sp_port_vlan);
8984         mutex_unlock(&mlxsw_sp->router->lock);
8985 }
8986
8987 static int mlxsw_sp_inetaddr_port_vlan_event(struct net_device *l3_dev,
8988                                              struct net_device *port_dev,
8989                                              unsigned long event, u16 vid,
8990                                              struct netlink_ext_ack *extack)
8991 {
8992         struct mlxsw_sp_port *mlxsw_sp_port = netdev_priv(port_dev);
8993         struct mlxsw_sp_port_vlan *mlxsw_sp_port_vlan;
8994
8995         mlxsw_sp_port_vlan = mlxsw_sp_port_vlan_find_by_vid(mlxsw_sp_port, vid);
8996         if (WARN_ON(!mlxsw_sp_port_vlan))
8997                 return -EINVAL;
8998
8999         switch (event) {
9000         case NETDEV_UP:
9001                 return __mlxsw_sp_port_vlan_router_join(mlxsw_sp_port_vlan,
9002                                                         l3_dev, extack);
9003         case NETDEV_DOWN:
9004                 __mlxsw_sp_port_vlan_router_leave(mlxsw_sp_port_vlan);
9005                 break;
9006         }
9007
9008         return 0;
9009 }
9010
9011 static int mlxsw_sp_inetaddr_port_event(struct net_device *port_dev,
9012                                         unsigned long event, bool nomaster,
9013                                         struct netlink_ext_ack *extack)
9014 {
9015         if (!nomaster && (netif_is_any_bridge_port(port_dev) ||
9016                           netif_is_lag_port(port_dev)))
9017                 return 0;
9018
9019         return mlxsw_sp_inetaddr_port_vlan_event(port_dev, port_dev, event,
9020                                                  MLXSW_SP_DEFAULT_VID, extack);
9021 }
9022
9023 static int __mlxsw_sp_inetaddr_lag_event(struct net_device *l3_dev,
9024                                          struct net_device *lag_dev,
9025                                          unsigned long event, u16 vid,
9026                                          struct netlink_ext_ack *extack)
9027 {
9028         struct net_device *port_dev;
9029         struct list_head *iter;
9030         int err;
9031
9032         netdev_for_each_lower_dev(lag_dev, port_dev, iter) {
9033                 if (mlxsw_sp_port_dev_check(port_dev)) {
9034                         err = mlxsw_sp_inetaddr_port_vlan_event(l3_dev,
9035                                                                 port_dev,
9036                                                                 event, vid,
9037                                                                 extack);
9038                         if (err)
9039                                 return err;
9040                 }
9041         }
9042
9043         return 0;
9044 }
9045
9046 static int mlxsw_sp_inetaddr_lag_event(struct net_device *lag_dev,
9047                                        unsigned long event, bool nomaster,
9048                                        struct netlink_ext_ack *extack)
9049 {
9050         if (!nomaster && netif_is_bridge_port(lag_dev))
9051                 return 0;
9052
9053         return __mlxsw_sp_inetaddr_lag_event(lag_dev, lag_dev, event,
9054                                              MLXSW_SP_DEFAULT_VID, extack);
9055 }
9056
9057 static int mlxsw_sp_inetaddr_bridge_event(struct mlxsw_sp *mlxsw_sp,
9058                                           struct net_device *l3_dev,
9059                                           int lower_pvid,
9060                                           unsigned long event,
9061                                           struct netlink_ext_ack *extack)
9062 {
9063         struct mlxsw_sp_rif_params params = {
9064                 .dev = l3_dev,
9065         };
9066         struct mlxsw_sp_rif *rif;
9067         int err;
9068
9069         switch (event) {
9070         case NETDEV_UP:
9071                 if (netif_is_bridge_master(l3_dev) && br_vlan_enabled(l3_dev)) {
9072                         u16 proto;
9073
9074                         br_vlan_get_proto(l3_dev, &proto);
9075                         if (proto == ETH_P_8021AD) {
9076                                 NL_SET_ERR_MSG_MOD(extack, "Adding an IP address to 802.1ad bridge is not supported");
9077                                 return -EOPNOTSUPP;
9078                         }
9079                         err = br_vlan_get_pvid(l3_dev, &params.vid);
9080                         if (err)
9081                                 return err;
9082                         if (!params.vid)
9083                                 return 0;
9084                 } else if (is_vlan_dev(l3_dev)) {
9085                         params.vid = vlan_dev_vlan_id(l3_dev);
9086
9087                         /* If the VID matches PVID of the bridge below, the
9088                          * bridge owns the RIF for this VLAN. Don't do anything.
9089                          */
9090                         if ((int)params.vid == lower_pvid)
9091                                 return 0;
9092                 }
9093
9094                 rif = mlxsw_sp_rif_create(mlxsw_sp, &params, extack);
9095                 if (IS_ERR(rif))
9096                         return PTR_ERR(rif);
9097                 break;
9098         case NETDEV_DOWN:
9099                 rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, l3_dev);
9100                 mlxsw_sp_rif_destroy(rif);
9101                 break;
9102         }
9103
9104         return 0;
9105 }
9106
9107 static int mlxsw_sp_inetaddr_vlan_event(struct mlxsw_sp *mlxsw_sp,
9108                                         struct net_device *vlan_dev,
9109                                         unsigned long event, bool nomaster,
9110                                         struct netlink_ext_ack *extack)
9111 {
9112         struct net_device *real_dev = vlan_dev_real_dev(vlan_dev);
9113         u16 vid = vlan_dev_vlan_id(vlan_dev);
9114         u16 lower_pvid;
9115         int err;
9116
9117         if (!nomaster && netif_is_bridge_port(vlan_dev))
9118                 return 0;
9119
9120         if (mlxsw_sp_port_dev_check(real_dev)) {
9121                 return mlxsw_sp_inetaddr_port_vlan_event(vlan_dev, real_dev,
9122                                                          event, vid, extack);
9123         } else if (netif_is_lag_master(real_dev)) {
9124                 return __mlxsw_sp_inetaddr_lag_event(vlan_dev, real_dev, event,
9125                                                      vid, extack);
9126         } else if (netif_is_bridge_master(real_dev) &&
9127                    br_vlan_enabled(real_dev)) {
9128                 err = br_vlan_get_pvid(real_dev, &lower_pvid);
9129                 if (err)
9130                         return err;
9131                 return mlxsw_sp_inetaddr_bridge_event(mlxsw_sp, vlan_dev,
9132                                                       lower_pvid, event,
9133                                                       extack);
9134         }
9135
9136         return 0;
9137 }
9138
9139 static bool mlxsw_sp_rif_macvlan_is_vrrp4(const u8 *mac)
9140 {
9141         u8 vrrp4[ETH_ALEN] = { 0x00, 0x00, 0x5e, 0x00, 0x01, 0x00 };
9142         u8 mask[ETH_ALEN] = { 0xff, 0xff, 0xff, 0xff, 0xff, 0x00 };
9143
9144         return ether_addr_equal_masked(mac, vrrp4, mask);
9145 }
9146
9147 static bool mlxsw_sp_rif_macvlan_is_vrrp6(const u8 *mac)
9148 {
9149         u8 vrrp6[ETH_ALEN] = { 0x00, 0x00, 0x5e, 0x00, 0x02, 0x00 };
9150         u8 mask[ETH_ALEN] = { 0xff, 0xff, 0xff, 0xff, 0xff, 0x00 };
9151
9152         return ether_addr_equal_masked(mac, vrrp6, mask);
9153 }
9154
9155 static int mlxsw_sp_rif_vrrp_op(struct mlxsw_sp *mlxsw_sp, u16 rif_index,
9156                                 const u8 *mac, bool adding)
9157 {
9158         char ritr_pl[MLXSW_REG_RITR_LEN];
9159         u8 vrrp_id = adding ? mac[5] : 0;
9160         int err;
9161
9162         if (!mlxsw_sp_rif_macvlan_is_vrrp4(mac) &&
9163             !mlxsw_sp_rif_macvlan_is_vrrp6(mac))
9164                 return 0;
9165
9166         mlxsw_reg_ritr_rif_pack(ritr_pl, rif_index);
9167         err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
9168         if (err)
9169                 return err;
9170
9171         if (mlxsw_sp_rif_macvlan_is_vrrp4(mac))
9172                 mlxsw_reg_ritr_if_vrrp_id_ipv4_set(ritr_pl, vrrp_id);
9173         else
9174                 mlxsw_reg_ritr_if_vrrp_id_ipv6_set(ritr_pl, vrrp_id);
9175
9176         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
9177 }
9178
9179 static int mlxsw_sp_rif_macvlan_add(struct mlxsw_sp *mlxsw_sp,
9180                                     const struct net_device *macvlan_dev,
9181                                     struct netlink_ext_ack *extack)
9182 {
9183         struct macvlan_dev *vlan = netdev_priv(macvlan_dev);
9184         struct mlxsw_sp_rif *rif;
9185         int err;
9186
9187         rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, vlan->lowerdev);
9188         if (!rif)
9189                 return 0;
9190
9191         err = mlxsw_sp_rif_fdb_op(mlxsw_sp, macvlan_dev->dev_addr,
9192                                   mlxsw_sp_fid_index(rif->fid), true);
9193         if (err)
9194                 return err;
9195
9196         err = mlxsw_sp_rif_vrrp_op(mlxsw_sp, rif->rif_index,
9197                                    macvlan_dev->dev_addr, true);
9198         if (err)
9199                 goto err_rif_vrrp_add;
9200
9201         /* Make sure the bridge driver does not have this MAC pointing at
9202          * some other port.
9203          */
9204         if (rif->ops->fdb_del)
9205                 rif->ops->fdb_del(rif, macvlan_dev->dev_addr);
9206
9207         return 0;
9208
9209 err_rif_vrrp_add:
9210         mlxsw_sp_rif_fdb_op(mlxsw_sp, macvlan_dev->dev_addr,
9211                             mlxsw_sp_fid_index(rif->fid), false);
9212         return err;
9213 }
9214
9215 static void __mlxsw_sp_rif_macvlan_del(struct mlxsw_sp *mlxsw_sp,
9216                                        const struct net_device *macvlan_dev)
9217 {
9218         struct macvlan_dev *vlan = netdev_priv(macvlan_dev);
9219         struct mlxsw_sp_rif *rif;
9220
9221         rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, vlan->lowerdev);
9222         /* If we do not have a RIF, then we already took care of
9223          * removing the macvlan's MAC during RIF deletion.
9224          */
9225         if (!rif)
9226                 return;
9227         mlxsw_sp_rif_vrrp_op(mlxsw_sp, rif->rif_index, macvlan_dev->dev_addr,
9228                              false);
9229         mlxsw_sp_rif_fdb_op(mlxsw_sp, macvlan_dev->dev_addr,
9230                             mlxsw_sp_fid_index(rif->fid), false);
9231 }
9232
9233 void mlxsw_sp_rif_macvlan_del(struct mlxsw_sp *mlxsw_sp,
9234                               const struct net_device *macvlan_dev)
9235 {
9236         mutex_lock(&mlxsw_sp->router->lock);
9237         __mlxsw_sp_rif_macvlan_del(mlxsw_sp, macvlan_dev);
9238         mutex_unlock(&mlxsw_sp->router->lock);
9239 }
9240
9241 static int mlxsw_sp_inetaddr_macvlan_event(struct mlxsw_sp *mlxsw_sp,
9242                                            struct net_device *macvlan_dev,
9243                                            unsigned long event,
9244                                            struct netlink_ext_ack *extack)
9245 {
9246         switch (event) {
9247         case NETDEV_UP:
9248                 return mlxsw_sp_rif_macvlan_add(mlxsw_sp, macvlan_dev, extack);
9249         case NETDEV_DOWN:
9250                 __mlxsw_sp_rif_macvlan_del(mlxsw_sp, macvlan_dev);
9251                 break;
9252         }
9253
9254         return 0;
9255 }
9256
9257 static int __mlxsw_sp_inetaddr_event(struct mlxsw_sp *mlxsw_sp,
9258                                      struct net_device *dev,
9259                                      unsigned long event, bool nomaster,
9260                                      struct netlink_ext_ack *extack)
9261 {
9262         if (mlxsw_sp_port_dev_check(dev))
9263                 return mlxsw_sp_inetaddr_port_event(dev, event, nomaster,
9264                                                     extack);
9265         else if (netif_is_lag_master(dev))
9266                 return mlxsw_sp_inetaddr_lag_event(dev, event, nomaster,
9267                                                    extack);
9268         else if (netif_is_bridge_master(dev))
9269                 return mlxsw_sp_inetaddr_bridge_event(mlxsw_sp, dev, -1, event,
9270                                                       extack);
9271         else if (is_vlan_dev(dev))
9272                 return mlxsw_sp_inetaddr_vlan_event(mlxsw_sp, dev, event,
9273                                                     nomaster, extack);
9274         else if (netif_is_macvlan(dev))
9275                 return mlxsw_sp_inetaddr_macvlan_event(mlxsw_sp, dev, event,
9276                                                        extack);
9277         else
9278                 return 0;
9279 }
9280
9281 static int mlxsw_sp_inetaddr_event(struct notifier_block *nb,
9282                                    unsigned long event, void *ptr)
9283 {
9284         struct in_ifaddr *ifa = (struct in_ifaddr *) ptr;
9285         struct net_device *dev = ifa->ifa_dev->dev;
9286         struct mlxsw_sp_router *router;
9287         struct mlxsw_sp_rif *rif;
9288         int err = 0;
9289
9290         /* NETDEV_UP event is handled by mlxsw_sp_inetaddr_valid_event */
9291         if (event == NETDEV_UP)
9292                 return NOTIFY_DONE;
9293
9294         router = container_of(nb, struct mlxsw_sp_router, inetaddr_nb);
9295         mutex_lock(&router->lock);
9296         rif = mlxsw_sp_rif_find_by_dev(router->mlxsw_sp, dev);
9297         if (!mlxsw_sp_rif_should_config(rif, dev, event))
9298                 goto out;
9299
9300         err = __mlxsw_sp_inetaddr_event(router->mlxsw_sp, dev, event, false,
9301                                         NULL);
9302 out:
9303         mutex_unlock(&router->lock);
9304         return notifier_from_errno(err);
9305 }
9306
9307 static int mlxsw_sp_inetaddr_valid_event(struct notifier_block *unused,
9308                                          unsigned long event, void *ptr)
9309 {
9310         struct in_validator_info *ivi = (struct in_validator_info *) ptr;
9311         struct net_device *dev = ivi->ivi_dev->dev;
9312         struct mlxsw_sp *mlxsw_sp;
9313         struct mlxsw_sp_rif *rif;
9314         int err = 0;
9315
9316         mlxsw_sp = mlxsw_sp_lower_get(dev);
9317         if (!mlxsw_sp)
9318                 return NOTIFY_DONE;
9319
9320         mutex_lock(&mlxsw_sp->router->lock);
9321         rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
9322         if (!mlxsw_sp_rif_should_config(rif, dev, event))
9323                 goto out;
9324
9325         err = __mlxsw_sp_inetaddr_event(mlxsw_sp, dev, event, false,
9326                                         ivi->extack);
9327 out:
9328         mutex_unlock(&mlxsw_sp->router->lock);
9329         return notifier_from_errno(err);
9330 }
9331
9332 struct mlxsw_sp_inet6addr_event_work {
9333         struct work_struct work;
9334         struct mlxsw_sp *mlxsw_sp;
9335         struct net_device *dev;
9336         netdevice_tracker dev_tracker;
9337         unsigned long event;
9338 };
9339
9340 static void mlxsw_sp_inet6addr_event_work(struct work_struct *work)
9341 {
9342         struct mlxsw_sp_inet6addr_event_work *inet6addr_work =
9343                 container_of(work, struct mlxsw_sp_inet6addr_event_work, work);
9344         struct mlxsw_sp *mlxsw_sp = inet6addr_work->mlxsw_sp;
9345         struct net_device *dev = inet6addr_work->dev;
9346         unsigned long event = inet6addr_work->event;
9347         struct mlxsw_sp_rif *rif;
9348
9349         rtnl_lock();
9350         mutex_lock(&mlxsw_sp->router->lock);
9351
9352         rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
9353         if (!mlxsw_sp_rif_should_config(rif, dev, event))
9354                 goto out;
9355
9356         __mlxsw_sp_inetaddr_event(mlxsw_sp, dev, event, false, NULL);
9357 out:
9358         mutex_unlock(&mlxsw_sp->router->lock);
9359         rtnl_unlock();
9360         netdev_put(dev, &inet6addr_work->dev_tracker);
9361         kfree(inet6addr_work);
9362 }
9363
9364 /* Called with rcu_read_lock() */
9365 static int mlxsw_sp_inet6addr_event(struct notifier_block *nb,
9366                                     unsigned long event, void *ptr)
9367 {
9368         struct inet6_ifaddr *if6 = (struct inet6_ifaddr *) ptr;
9369         struct mlxsw_sp_inet6addr_event_work *inet6addr_work;
9370         struct net_device *dev = if6->idev->dev;
9371         struct mlxsw_sp_router *router;
9372
9373         /* NETDEV_UP event is handled by mlxsw_sp_inet6addr_valid_event */
9374         if (event == NETDEV_UP)
9375                 return NOTIFY_DONE;
9376
9377         inet6addr_work = kzalloc(sizeof(*inet6addr_work), GFP_ATOMIC);
9378         if (!inet6addr_work)
9379                 return NOTIFY_BAD;
9380
9381         router = container_of(nb, struct mlxsw_sp_router, inet6addr_nb);
9382         INIT_WORK(&inet6addr_work->work, mlxsw_sp_inet6addr_event_work);
9383         inet6addr_work->mlxsw_sp = router->mlxsw_sp;
9384         inet6addr_work->dev = dev;
9385         inet6addr_work->event = event;
9386         netdev_hold(dev, &inet6addr_work->dev_tracker, GFP_ATOMIC);
9387         mlxsw_core_schedule_work(&inet6addr_work->work);
9388
9389         return NOTIFY_DONE;
9390 }
9391
9392 static int mlxsw_sp_inet6addr_valid_event(struct notifier_block *unused,
9393                                           unsigned long event, void *ptr)
9394 {
9395         struct in6_validator_info *i6vi = (struct in6_validator_info *) ptr;
9396         struct net_device *dev = i6vi->i6vi_dev->dev;
9397         struct mlxsw_sp *mlxsw_sp;
9398         struct mlxsw_sp_rif *rif;
9399         int err = 0;
9400
9401         mlxsw_sp = mlxsw_sp_lower_get(dev);
9402         if (!mlxsw_sp)
9403                 return NOTIFY_DONE;
9404
9405         mutex_lock(&mlxsw_sp->router->lock);
9406         rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
9407         if (!mlxsw_sp_rif_should_config(rif, dev, event))
9408                 goto out;
9409
9410         err = __mlxsw_sp_inetaddr_event(mlxsw_sp, dev, event, false,
9411                                         i6vi->extack);
9412 out:
9413         mutex_unlock(&mlxsw_sp->router->lock);
9414         return notifier_from_errno(err);
9415 }
9416
9417 static int mlxsw_sp_rif_edit(struct mlxsw_sp *mlxsw_sp, u16 rif_index,
9418                              const char *mac, int mtu, u8 mac_profile)
9419 {
9420         char ritr_pl[MLXSW_REG_RITR_LEN];
9421         int err;
9422
9423         mlxsw_reg_ritr_rif_pack(ritr_pl, rif_index);
9424         err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
9425         if (err)
9426                 return err;
9427
9428         mlxsw_reg_ritr_mtu_set(ritr_pl, mtu);
9429         mlxsw_reg_ritr_if_mac_memcpy_to(ritr_pl, mac);
9430         mlxsw_reg_ritr_if_mac_profile_id_set(ritr_pl, mac_profile);
9431         mlxsw_reg_ritr_op_set(ritr_pl, MLXSW_REG_RITR_RIF_CREATE);
9432         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
9433 }
9434
9435 static int
9436 mlxsw_sp_router_port_change_event(struct mlxsw_sp *mlxsw_sp,
9437                                   struct mlxsw_sp_rif *rif,
9438                                   struct netlink_ext_ack *extack)
9439 {
9440         struct net_device *dev = mlxsw_sp_rif_dev(rif);
9441         u8 old_mac_profile;
9442         u16 fid_index;
9443         int err;
9444
9445         fid_index = mlxsw_sp_fid_index(rif->fid);
9446
9447         err = mlxsw_sp_rif_fdb_op(mlxsw_sp, rif->addr, fid_index, false);
9448         if (err)
9449                 return err;
9450
9451         old_mac_profile = rif->mac_profile_id;
9452         err = mlxsw_sp_rif_mac_profile_replace(mlxsw_sp, rif, dev->dev_addr,
9453                                                extack);
9454         if (err)
9455                 goto err_rif_mac_profile_replace;
9456
9457         err = mlxsw_sp_rif_edit(mlxsw_sp, rif->rif_index, dev->dev_addr,
9458                                 dev->mtu, rif->mac_profile_id);
9459         if (err)
9460                 goto err_rif_edit;
9461
9462         err = mlxsw_sp_rif_fdb_op(mlxsw_sp, dev->dev_addr, fid_index, true);
9463         if (err)
9464                 goto err_rif_fdb_op;
9465
9466         if (rif->mtu != dev->mtu) {
9467                 struct mlxsw_sp_vr *vr;
9468                 int i;
9469
9470                 /* The RIF is relevant only to its mr_table instance, as unlike
9471                  * unicast routing, in multicast routing a RIF cannot be shared
9472                  * between several multicast routing tables.
9473                  */
9474                 vr = &mlxsw_sp->router->vrs[rif->vr_id];
9475                 for (i = 0; i < MLXSW_SP_L3_PROTO_MAX; i++)
9476                         mlxsw_sp_mr_rif_mtu_update(vr->mr_table[i],
9477                                                    rif, dev->mtu);
9478         }
9479
9480         ether_addr_copy(rif->addr, dev->dev_addr);
9481         rif->mtu = dev->mtu;
9482
9483         netdev_dbg(dev, "Updated RIF=%d\n", rif->rif_index);
9484
9485         return 0;
9486
9487 err_rif_fdb_op:
9488         mlxsw_sp_rif_edit(mlxsw_sp, rif->rif_index, rif->addr, rif->mtu,
9489                           old_mac_profile);
9490 err_rif_edit:
9491         mlxsw_sp_rif_mac_profile_replace(mlxsw_sp, rif, rif->addr, extack);
9492 err_rif_mac_profile_replace:
9493         mlxsw_sp_rif_fdb_op(mlxsw_sp, rif->addr, fid_index, true);
9494         return err;
9495 }
9496
9497 static int mlxsw_sp_router_port_pre_changeaddr_event(struct mlxsw_sp_rif *rif,
9498                             struct netdev_notifier_pre_changeaddr_info *info)
9499 {
9500         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
9501         struct mlxsw_sp_rif_mac_profile *profile;
9502         struct netlink_ext_ack *extack;
9503         u8 max_rif_mac_profiles;
9504         u64 occ;
9505
9506         extack = netdev_notifier_info_to_extack(&info->info);
9507
9508         profile = mlxsw_sp_rif_mac_profile_find(mlxsw_sp, info->dev_addr);
9509         if (profile)
9510                 return 0;
9511
9512         max_rif_mac_profiles = mlxsw_sp->router->max_rif_mac_profile;
9513         occ = mlxsw_sp_rif_mac_profiles_occ_get(mlxsw_sp);
9514         if (occ < max_rif_mac_profiles)
9515                 return 0;
9516
9517         if (!mlxsw_sp_rif_mac_profile_is_shared(rif))
9518                 return 0;
9519
9520         NL_SET_ERR_MSG_MOD(extack, "Exceeded number of supported router interface MAC profiles");
9521         return -ENOBUFS;
9522 }
9523
9524 static bool mlxsw_sp_router_netdevice_interesting(struct mlxsw_sp *mlxsw_sp,
9525                                                   struct net_device *dev)
9526 {
9527         struct vlan_dev_priv *vlan;
9528
9529         if (netif_is_lag_master(dev) ||
9530             netif_is_bridge_master(dev) ||
9531             mlxsw_sp_port_dev_check(dev) ||
9532             mlxsw_sp_netdev_is_ipip_ol(mlxsw_sp, dev) ||
9533             netif_is_l3_master(dev))
9534                 return true;
9535
9536         if (!is_vlan_dev(dev))
9537                 return false;
9538
9539         vlan = vlan_dev_priv(dev);
9540         return netif_is_lag_master(vlan->real_dev) ||
9541                netif_is_bridge_master(vlan->real_dev) ||
9542                mlxsw_sp_port_dev_check(vlan->real_dev);
9543 }
9544
9545 static struct mlxsw_sp_crif *
9546 mlxsw_sp_crif_register(struct mlxsw_sp_router *router, struct net_device *dev)
9547 {
9548         struct mlxsw_sp_crif *crif;
9549         int err;
9550
9551         if (WARN_ON(mlxsw_sp_crif_lookup(router, dev)))
9552                 return NULL;
9553
9554         crif = mlxsw_sp_crif_alloc(dev);
9555         if (!crif)
9556                 return ERR_PTR(-ENOMEM);
9557
9558         err = mlxsw_sp_crif_insert(router, crif);
9559         if (err)
9560                 goto err_netdev_insert;
9561
9562         return crif;
9563
9564 err_netdev_insert:
9565         mlxsw_sp_crif_free(crif);
9566         return ERR_PTR(err);
9567 }
9568
9569 static void mlxsw_sp_crif_unregister(struct mlxsw_sp_router *router,
9570                                      struct mlxsw_sp_crif *crif)
9571 {
9572         struct mlxsw_sp_nexthop *nh, *tmp;
9573
9574         mlxsw_sp_crif_remove(router, crif);
9575
9576         list_for_each_entry_safe(nh, tmp, &crif->nexthop_list, crif_list_node)
9577                 mlxsw_sp_nexthop_type_fini(router->mlxsw_sp, nh);
9578
9579         if (crif->rif)
9580                 crif->can_destroy = true;
9581         else
9582                 mlxsw_sp_crif_free(crif);
9583 }
9584
9585 static int mlxsw_sp_netdevice_register(struct mlxsw_sp_router *router,
9586                                        struct net_device *dev)
9587 {
9588         struct mlxsw_sp_crif *crif;
9589
9590         if (!mlxsw_sp_router_netdevice_interesting(router->mlxsw_sp, dev))
9591                 return 0;
9592
9593         crif = mlxsw_sp_crif_register(router, dev);
9594         return PTR_ERR_OR_ZERO(crif);
9595 }
9596
9597 static void mlxsw_sp_netdevice_unregister(struct mlxsw_sp_router *router,
9598                                           struct net_device *dev)
9599 {
9600         struct mlxsw_sp_crif *crif;
9601
9602         if (!mlxsw_sp_router_netdevice_interesting(router->mlxsw_sp, dev))
9603                 return;
9604
9605         /* netdev_run_todo(), by way of netdev_wait_allrefs_any(), rebroadcasts
9606          * the NETDEV_UNREGISTER message, so we can get here twice. If that's
9607          * what happened, the netdevice state is NETREG_UNREGISTERED. In that
9608          * case, we expect to have collected the CRIF already, and warn if it
9609          * still exists. Otherwise we expect the CRIF to exist.
9610          */
9611         crif = mlxsw_sp_crif_lookup(router, dev);
9612         if (dev->reg_state == NETREG_UNREGISTERED) {
9613                 if (!WARN_ON(crif))
9614                         return;
9615         }
9616         if (WARN_ON(!crif))
9617                 return;
9618
9619         mlxsw_sp_crif_unregister(router, crif);
9620 }
9621
9622 static bool mlxsw_sp_is_offload_xstats_event(unsigned long event)
9623 {
9624         switch (event) {
9625         case NETDEV_OFFLOAD_XSTATS_ENABLE:
9626         case NETDEV_OFFLOAD_XSTATS_DISABLE:
9627         case NETDEV_OFFLOAD_XSTATS_REPORT_USED:
9628         case NETDEV_OFFLOAD_XSTATS_REPORT_DELTA:
9629                 return true;
9630         }
9631
9632         return false;
9633 }
9634
9635 static int
9636 mlxsw_sp_router_port_offload_xstats_cmd(struct mlxsw_sp_rif *rif,
9637                                         unsigned long event,
9638                                         struct netdev_notifier_offload_xstats_info *info)
9639 {
9640         switch (info->type) {
9641         case NETDEV_OFFLOAD_XSTATS_TYPE_L3:
9642                 break;
9643         default:
9644                 return 0;
9645         }
9646
9647         switch (event) {
9648         case NETDEV_OFFLOAD_XSTATS_ENABLE:
9649                 return mlxsw_sp_router_port_l3_stats_enable(rif);
9650         case NETDEV_OFFLOAD_XSTATS_DISABLE:
9651                 mlxsw_sp_router_port_l3_stats_disable(rif);
9652                 return 0;
9653         case NETDEV_OFFLOAD_XSTATS_REPORT_USED:
9654                 mlxsw_sp_router_port_l3_stats_report_used(rif, info);
9655                 return 0;
9656         case NETDEV_OFFLOAD_XSTATS_REPORT_DELTA:
9657                 return mlxsw_sp_router_port_l3_stats_report_delta(rif, info);
9658         }
9659
9660         WARN_ON_ONCE(1);
9661         return 0;
9662 }
9663
9664 static int
9665 mlxsw_sp_netdevice_offload_xstats_cmd(struct mlxsw_sp *mlxsw_sp,
9666                                       struct net_device *dev,
9667                                       unsigned long event,
9668                                       struct netdev_notifier_offload_xstats_info *info)
9669 {
9670         struct mlxsw_sp_rif *rif;
9671
9672         rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
9673         if (!rif)
9674                 return 0;
9675
9676         return mlxsw_sp_router_port_offload_xstats_cmd(rif, event, info);
9677 }
9678
9679 static bool mlxsw_sp_is_router_event(unsigned long event)
9680 {
9681         switch (event) {
9682         case NETDEV_PRE_CHANGEADDR:
9683         case NETDEV_CHANGEADDR:
9684         case NETDEV_CHANGEMTU:
9685                 return true;
9686         default:
9687                 return false;
9688         }
9689 }
9690
9691 static int mlxsw_sp_netdevice_router_port_event(struct net_device *dev,
9692                                                 unsigned long event, void *ptr)
9693 {
9694         struct netlink_ext_ack *extack = netdev_notifier_info_to_extack(ptr);
9695         struct mlxsw_sp *mlxsw_sp;
9696         struct mlxsw_sp_rif *rif;
9697
9698         mlxsw_sp = mlxsw_sp_lower_get(dev);
9699         if (!mlxsw_sp)
9700                 return 0;
9701
9702         rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
9703         if (!rif)
9704                 return 0;
9705
9706         switch (event) {
9707         case NETDEV_CHANGEMTU:
9708         case NETDEV_CHANGEADDR:
9709                 return mlxsw_sp_router_port_change_event(mlxsw_sp, rif, extack);
9710         case NETDEV_PRE_CHANGEADDR:
9711                 return mlxsw_sp_router_port_pre_changeaddr_event(rif, ptr);
9712         default:
9713                 WARN_ON_ONCE(1);
9714                 break;
9715         }
9716
9717         return 0;
9718 }
9719
9720 static int mlxsw_sp_port_vrf_join(struct mlxsw_sp *mlxsw_sp,
9721                                   struct net_device *l3_dev,
9722                                   struct netlink_ext_ack *extack)
9723 {
9724         struct mlxsw_sp_rif *rif;
9725
9726         /* If netdev is already associated with a RIF, then we need to
9727          * destroy it and create a new one with the new virtual router ID.
9728          */
9729         rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, l3_dev);
9730         if (rif)
9731                 __mlxsw_sp_inetaddr_event(mlxsw_sp, l3_dev, NETDEV_DOWN, false,
9732                                           extack);
9733
9734         return __mlxsw_sp_inetaddr_event(mlxsw_sp, l3_dev, NETDEV_UP, false,
9735                                          extack);
9736 }
9737
9738 static void mlxsw_sp_port_vrf_leave(struct mlxsw_sp *mlxsw_sp,
9739                                     struct net_device *l3_dev)
9740 {
9741         struct mlxsw_sp_rif *rif;
9742
9743         rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, l3_dev);
9744         if (!rif)
9745                 return;
9746         __mlxsw_sp_inetaddr_event(mlxsw_sp, l3_dev, NETDEV_DOWN, false, NULL);
9747 }
9748
9749 static bool mlxsw_sp_is_vrf_event(unsigned long event, void *ptr)
9750 {
9751         struct netdev_notifier_changeupper_info *info = ptr;
9752
9753         if (event != NETDEV_PRECHANGEUPPER && event != NETDEV_CHANGEUPPER)
9754                 return false;
9755         return netif_is_l3_master(info->upper_dev);
9756 }
9757
9758 static int
9759 mlxsw_sp_netdevice_vrf_event(struct net_device *l3_dev, unsigned long event,
9760                              struct netdev_notifier_changeupper_info *info)
9761 {
9762         struct mlxsw_sp *mlxsw_sp = mlxsw_sp_lower_get(l3_dev);
9763         int err = 0;
9764
9765         /* We do not create a RIF for a macvlan, but only use it to
9766          * direct more MAC addresses to the router.
9767          */
9768         if (!mlxsw_sp || netif_is_macvlan(l3_dev))
9769                 return 0;
9770
9771         switch (event) {
9772         case NETDEV_PRECHANGEUPPER:
9773                 break;
9774         case NETDEV_CHANGEUPPER:
9775                 if (info->linking) {
9776                         struct netlink_ext_ack *extack;
9777
9778                         extack = netdev_notifier_info_to_extack(&info->info);
9779                         err = mlxsw_sp_port_vrf_join(mlxsw_sp, l3_dev, extack);
9780                 } else {
9781                         mlxsw_sp_port_vrf_leave(mlxsw_sp, l3_dev);
9782                 }
9783                 break;
9784         }
9785
9786         return err;
9787 }
9788
9789 struct mlxsw_sp_router_replay_inetaddr_up {
9790         struct mlxsw_sp *mlxsw_sp;
9791         struct netlink_ext_ack *extack;
9792         unsigned int done;
9793         bool deslavement;
9794 };
9795
9796 static int mlxsw_sp_router_replay_inetaddr_up(struct net_device *dev,
9797                                               struct netdev_nested_priv *priv)
9798 {
9799         struct mlxsw_sp_router_replay_inetaddr_up *ctx = priv->data;
9800         bool nomaster = ctx->deslavement;
9801         struct mlxsw_sp_crif *crif;
9802         int err;
9803
9804         if (mlxsw_sp_dev_addr_list_empty(dev))
9805                 return 0;
9806
9807         crif = mlxsw_sp_crif_lookup(ctx->mlxsw_sp->router, dev);
9808         if (!crif || crif->rif)
9809                 return 0;
9810
9811         if (!mlxsw_sp_rif_should_config(crif->rif, dev, NETDEV_UP))
9812                 return 0;
9813
9814         err = __mlxsw_sp_inetaddr_event(ctx->mlxsw_sp, dev, NETDEV_UP,
9815                                         nomaster, ctx->extack);
9816         if (err)
9817                 return err;
9818
9819         ctx->done++;
9820         return 0;
9821 }
9822
9823 static int mlxsw_sp_router_unreplay_inetaddr_up(struct net_device *dev,
9824                                                 struct netdev_nested_priv *priv)
9825 {
9826         struct mlxsw_sp_router_replay_inetaddr_up *ctx = priv->data;
9827         bool nomaster = ctx->deslavement;
9828         struct mlxsw_sp_crif *crif;
9829
9830         if (!ctx->done)
9831                 return 0;
9832
9833         if (mlxsw_sp_dev_addr_list_empty(dev))
9834                 return 0;
9835
9836         crif = mlxsw_sp_crif_lookup(ctx->mlxsw_sp->router, dev);
9837         if (!crif || !crif->rif)
9838                 return 0;
9839
9840         /* We are rolling back NETDEV_UP, so ask for that. */
9841         if (!mlxsw_sp_rif_should_config(crif->rif, dev, NETDEV_UP))
9842                 return 0;
9843
9844         __mlxsw_sp_inetaddr_event(ctx->mlxsw_sp, dev, NETDEV_DOWN, nomaster,
9845                                   NULL);
9846
9847         ctx->done--;
9848         return 0;
9849 }
9850
9851 int mlxsw_sp_netdevice_enslavement_replay(struct mlxsw_sp *mlxsw_sp,
9852                                           struct net_device *upper_dev,
9853                                           struct netlink_ext_ack *extack)
9854 {
9855         struct mlxsw_sp_router_replay_inetaddr_up ctx = {
9856                 .mlxsw_sp = mlxsw_sp,
9857                 .extack = extack,
9858                 .deslavement = false,
9859         };
9860         struct netdev_nested_priv priv = {
9861                 .data = &ctx,
9862         };
9863         int err;
9864
9865         err = mlxsw_sp_router_replay_inetaddr_up(upper_dev, &priv);
9866         if (err)
9867                 return err;
9868
9869         err = netdev_walk_all_upper_dev_rcu(upper_dev,
9870                                             mlxsw_sp_router_replay_inetaddr_up,
9871                                             &priv);
9872         if (err)
9873                 goto err_replay_up;
9874
9875         return 0;
9876
9877 err_replay_up:
9878         netdev_walk_all_upper_dev_rcu(upper_dev,
9879                                       mlxsw_sp_router_unreplay_inetaddr_up,
9880                                       &priv);
9881         mlxsw_sp_router_unreplay_inetaddr_up(upper_dev, &priv);
9882         return err;
9883 }
9884
9885 void mlxsw_sp_netdevice_deslavement_replay(struct mlxsw_sp *mlxsw_sp,
9886                                            struct net_device *dev)
9887 {
9888         struct mlxsw_sp_router_replay_inetaddr_up ctx = {
9889                 .mlxsw_sp = mlxsw_sp,
9890                 .deslavement = true,
9891         };
9892         struct netdev_nested_priv priv = {
9893                 .data = &ctx,
9894         };
9895
9896         mlxsw_sp_router_replay_inetaddr_up(dev, &priv);
9897 }
9898
9899 static int
9900 mlxsw_sp_port_vid_router_join_existing(struct mlxsw_sp_port *mlxsw_sp_port,
9901                                        u16 vid, struct net_device *dev,
9902                                        struct netlink_ext_ack *extack)
9903 {
9904         struct mlxsw_sp_port_vlan *mlxsw_sp_port_vlan;
9905
9906         mlxsw_sp_port_vlan = mlxsw_sp_port_vlan_find_by_vid(mlxsw_sp_port,
9907                                                             vid);
9908         if (WARN_ON(!mlxsw_sp_port_vlan))
9909                 return -EINVAL;
9910
9911         return mlxsw_sp_port_vlan_router_join_existing(mlxsw_sp_port_vlan,
9912                                                        dev, extack);
9913 }
9914
9915 static void
9916 mlxsw_sp_port_vid_router_leave(struct mlxsw_sp_port *mlxsw_sp_port, u16 vid,
9917                                struct net_device *dev)
9918 {
9919         struct mlxsw_sp_port_vlan *mlxsw_sp_port_vlan;
9920
9921         mlxsw_sp_port_vlan = mlxsw_sp_port_vlan_find_by_vid(mlxsw_sp_port,
9922                                                             vid);
9923         if (WARN_ON(!mlxsw_sp_port_vlan))
9924                 return;
9925
9926         __mlxsw_sp_port_vlan_router_leave(mlxsw_sp_port_vlan);
9927 }
9928
9929 static int __mlxsw_sp_router_port_join_lag(struct mlxsw_sp_port *mlxsw_sp_port,
9930                                            struct net_device *lag_dev,
9931                                            struct netlink_ext_ack *extack)
9932 {
9933         u16 default_vid = MLXSW_SP_DEFAULT_VID;
9934         struct net_device *upper_dev;
9935         struct list_head *iter;
9936         int done = 0;
9937         u16 vid;
9938         int err;
9939
9940         err = mlxsw_sp_port_vid_router_join_existing(mlxsw_sp_port, default_vid,
9941                                                      lag_dev, extack);
9942         if (err)
9943                 return err;
9944
9945         netdev_for_each_upper_dev_rcu(lag_dev, upper_dev, iter) {
9946                 if (!is_vlan_dev(upper_dev))
9947                         continue;
9948
9949                 vid = vlan_dev_vlan_id(upper_dev);
9950                 err = mlxsw_sp_port_vid_router_join_existing(mlxsw_sp_port, vid,
9951                                                              upper_dev, extack);
9952                 if (err)
9953                         goto err_router_join_dev;
9954
9955                 ++done;
9956         }
9957
9958         return 0;
9959
9960 err_router_join_dev:
9961         netdev_for_each_upper_dev_rcu(lag_dev, upper_dev, iter) {
9962                 if (!is_vlan_dev(upper_dev))
9963                         continue;
9964                 if (!done--)
9965                         break;
9966
9967                 vid = vlan_dev_vlan_id(upper_dev);
9968                 mlxsw_sp_port_vid_router_leave(mlxsw_sp_port, vid, upper_dev);
9969         }
9970
9971         mlxsw_sp_port_vid_router_leave(mlxsw_sp_port, default_vid, lag_dev);
9972         return err;
9973 }
9974
9975 static void
9976 __mlxsw_sp_router_port_leave_lag(struct mlxsw_sp_port *mlxsw_sp_port,
9977                                  struct net_device *lag_dev)
9978 {
9979         u16 default_vid = MLXSW_SP_DEFAULT_VID;
9980         struct net_device *upper_dev;
9981         struct list_head *iter;
9982         u16 vid;
9983
9984         netdev_for_each_upper_dev_rcu(lag_dev, upper_dev, iter) {
9985                 if (!is_vlan_dev(upper_dev))
9986                         continue;
9987
9988                 vid = vlan_dev_vlan_id(upper_dev);
9989                 mlxsw_sp_port_vid_router_leave(mlxsw_sp_port, vid, upper_dev);
9990         }
9991
9992         mlxsw_sp_port_vid_router_leave(mlxsw_sp_port, default_vid, lag_dev);
9993 }
9994
9995 int mlxsw_sp_router_port_join_lag(struct mlxsw_sp_port *mlxsw_sp_port,
9996                                   struct net_device *lag_dev,
9997                                   struct netlink_ext_ack *extack)
9998 {
9999         int err;
10000
10001         mutex_lock(&mlxsw_sp_port->mlxsw_sp->router->lock);
10002         err = __mlxsw_sp_router_port_join_lag(mlxsw_sp_port, lag_dev, extack);
10003         mutex_unlock(&mlxsw_sp_port->mlxsw_sp->router->lock);
10004
10005         return err;
10006 }
10007
10008 void mlxsw_sp_router_port_leave_lag(struct mlxsw_sp_port *mlxsw_sp_port,
10009                                     struct net_device *lag_dev)
10010 {
10011         mutex_lock(&mlxsw_sp_port->mlxsw_sp->router->lock);
10012         __mlxsw_sp_router_port_leave_lag(mlxsw_sp_port, lag_dev);
10013         mutex_unlock(&mlxsw_sp_port->mlxsw_sp->router->lock);
10014 }
10015
10016 static int mlxsw_sp_router_netdevice_event(struct notifier_block *nb,
10017                                            unsigned long event, void *ptr)
10018 {
10019         struct net_device *dev = netdev_notifier_info_to_dev(ptr);
10020         struct mlxsw_sp_router *router;
10021         struct mlxsw_sp *mlxsw_sp;
10022         int err = 0;
10023
10024         router = container_of(nb, struct mlxsw_sp_router, netdevice_nb);
10025         mlxsw_sp = router->mlxsw_sp;
10026
10027         mutex_lock(&mlxsw_sp->router->lock);
10028
10029         if (event == NETDEV_REGISTER) {
10030                 err = mlxsw_sp_netdevice_register(router, dev);
10031                 if (err)
10032                         /* No need to roll this back, UNREGISTER will collect it
10033                          * anyhow.
10034                          */
10035                         goto out;
10036         }
10037
10038         if (mlxsw_sp_is_offload_xstats_event(event))
10039                 err = mlxsw_sp_netdevice_offload_xstats_cmd(mlxsw_sp, dev,
10040                                                             event, ptr);
10041         else if (mlxsw_sp_netdev_is_ipip_ol(mlxsw_sp, dev))
10042                 err = mlxsw_sp_netdevice_ipip_ol_event(mlxsw_sp, dev,
10043                                                        event, ptr);
10044         else if (mlxsw_sp_netdev_is_ipip_ul(mlxsw_sp, dev))
10045                 err = mlxsw_sp_netdevice_ipip_ul_event(mlxsw_sp, dev,
10046                                                        event, ptr);
10047         else if (mlxsw_sp_is_router_event(event))
10048                 err = mlxsw_sp_netdevice_router_port_event(dev, event, ptr);
10049         else if (mlxsw_sp_is_vrf_event(event, ptr))
10050                 err = mlxsw_sp_netdevice_vrf_event(dev, event, ptr);
10051
10052         if (event == NETDEV_UNREGISTER)
10053                 mlxsw_sp_netdevice_unregister(router, dev);
10054
10055 out:
10056         mutex_unlock(&mlxsw_sp->router->lock);
10057
10058         return notifier_from_errno(err);
10059 }
10060
10061 struct mlxsw_sp_macvlan_replay {
10062         struct mlxsw_sp *mlxsw_sp;
10063         struct netlink_ext_ack *extack;
10064 };
10065
10066 static int mlxsw_sp_macvlan_replay_upper(struct net_device *dev,
10067                                          struct netdev_nested_priv *priv)
10068 {
10069         const struct mlxsw_sp_macvlan_replay *rms = priv->data;
10070         struct netlink_ext_ack *extack = rms->extack;
10071         struct mlxsw_sp *mlxsw_sp = rms->mlxsw_sp;
10072
10073         if (!netif_is_macvlan(dev))
10074                 return 0;
10075
10076         return mlxsw_sp_rif_macvlan_add(mlxsw_sp, dev, extack);
10077 }
10078
10079 static int mlxsw_sp_macvlan_replay(struct mlxsw_sp_rif *rif,
10080                                    struct netlink_ext_ack *extack)
10081 {
10082         struct mlxsw_sp_macvlan_replay rms = {
10083                 .mlxsw_sp = rif->mlxsw_sp,
10084                 .extack = extack,
10085         };
10086         struct netdev_nested_priv priv = {
10087                 .data = &rms,
10088         };
10089
10090         return netdev_walk_all_upper_dev_rcu(mlxsw_sp_rif_dev(rif),
10091                                              mlxsw_sp_macvlan_replay_upper,
10092                                              &priv);
10093 }
10094
10095 static int __mlxsw_sp_rif_macvlan_flush(struct net_device *dev,
10096                                         struct netdev_nested_priv *priv)
10097 {
10098         struct mlxsw_sp_rif *rif = (struct mlxsw_sp_rif *)priv->data;
10099
10100         if (!netif_is_macvlan(dev))
10101                 return 0;
10102
10103         return mlxsw_sp_rif_fdb_op(rif->mlxsw_sp, dev->dev_addr,
10104                                    mlxsw_sp_fid_index(rif->fid), false);
10105 }
10106
10107 static int mlxsw_sp_rif_macvlan_flush(struct mlxsw_sp_rif *rif)
10108 {
10109         struct net_device *dev = mlxsw_sp_rif_dev(rif);
10110         struct netdev_nested_priv priv = {
10111                 .data = (void *)rif,
10112         };
10113
10114         if (!netif_is_macvlan_port(dev))
10115                 return 0;
10116
10117         return netdev_walk_all_upper_dev_rcu(dev,
10118                                              __mlxsw_sp_rif_macvlan_flush, &priv);
10119 }
10120
10121 static void mlxsw_sp_rif_subport_setup(struct mlxsw_sp_rif *rif,
10122                                        const struct mlxsw_sp_rif_params *params)
10123 {
10124         struct mlxsw_sp_rif_subport *rif_subport;
10125
10126         rif_subport = mlxsw_sp_rif_subport_rif(rif);
10127         refcount_set(&rif_subport->ref_count, 1);
10128         rif_subport->vid = params->vid;
10129         rif_subport->lag = params->lag;
10130         if (params->lag)
10131                 rif_subport->lag_id = params->lag_id;
10132         else
10133                 rif_subport->system_port = params->system_port;
10134 }
10135
10136 static int mlxsw_sp_rif_subport_op(struct mlxsw_sp_rif *rif, bool enable)
10137 {
10138         struct net_device *dev = mlxsw_sp_rif_dev(rif);
10139         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
10140         struct mlxsw_sp_rif_subport *rif_subport;
10141         char ritr_pl[MLXSW_REG_RITR_LEN];
10142         u16 efid;
10143
10144         rif_subport = mlxsw_sp_rif_subport_rif(rif);
10145         mlxsw_reg_ritr_pack(ritr_pl, enable, MLXSW_REG_RITR_SP_IF,
10146                             rif->rif_index, rif->vr_id, dev->mtu);
10147         mlxsw_reg_ritr_mac_pack(ritr_pl, dev->dev_addr);
10148         mlxsw_reg_ritr_if_mac_profile_id_set(ritr_pl, rif->mac_profile_id);
10149         efid = mlxsw_sp_fid_index(rif->fid);
10150         mlxsw_reg_ritr_sp_if_pack(ritr_pl, rif_subport->lag,
10151                                   rif_subport->lag ? rif_subport->lag_id :
10152                                                      rif_subport->system_port,
10153                                   efid, 0);
10154         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
10155 }
10156
10157 static int mlxsw_sp_rif_subport_configure(struct mlxsw_sp_rif *rif,
10158                                           struct netlink_ext_ack *extack)
10159 {
10160         struct net_device *dev = mlxsw_sp_rif_dev(rif);
10161         u8 mac_profile;
10162         int err;
10163
10164         err = mlxsw_sp_rif_mac_profile_get(rif->mlxsw_sp, rif->addr,
10165                                            &mac_profile, extack);
10166         if (err)
10167                 return err;
10168         rif->mac_profile_id = mac_profile;
10169
10170         err = mlxsw_sp_rif_subport_op(rif, true);
10171         if (err)
10172                 goto err_rif_subport_op;
10173
10174         err = mlxsw_sp_macvlan_replay(rif, extack);
10175         if (err)
10176                 goto err_macvlan_replay;
10177
10178         err = mlxsw_sp_rif_fdb_op(rif->mlxsw_sp, dev->dev_addr,
10179                                   mlxsw_sp_fid_index(rif->fid), true);
10180         if (err)
10181                 goto err_rif_fdb_op;
10182
10183         err = mlxsw_sp_fid_rif_set(rif->fid, rif);
10184         if (err)
10185                 goto err_fid_rif_set;
10186
10187         return 0;
10188
10189 err_fid_rif_set:
10190         mlxsw_sp_rif_fdb_op(rif->mlxsw_sp, dev->dev_addr,
10191                             mlxsw_sp_fid_index(rif->fid), false);
10192 err_rif_fdb_op:
10193         mlxsw_sp_rif_macvlan_flush(rif);
10194 err_macvlan_replay:
10195         mlxsw_sp_rif_subport_op(rif, false);
10196 err_rif_subport_op:
10197         mlxsw_sp_rif_mac_profile_put(rif->mlxsw_sp, mac_profile);
10198         return err;
10199 }
10200
10201 static void mlxsw_sp_rif_subport_deconfigure(struct mlxsw_sp_rif *rif)
10202 {
10203         struct net_device *dev = mlxsw_sp_rif_dev(rif);
10204         struct mlxsw_sp_fid *fid = rif->fid;
10205
10206         mlxsw_sp_fid_rif_unset(fid);
10207         mlxsw_sp_rif_fdb_op(rif->mlxsw_sp, dev->dev_addr,
10208                             mlxsw_sp_fid_index(fid), false);
10209         mlxsw_sp_rif_macvlan_flush(rif);
10210         mlxsw_sp_rif_subport_op(rif, false);
10211         mlxsw_sp_rif_mac_profile_put(rif->mlxsw_sp, rif->mac_profile_id);
10212 }
10213
10214 static struct mlxsw_sp_fid *
10215 mlxsw_sp_rif_subport_fid_get(struct mlxsw_sp_rif *rif,
10216                              const struct mlxsw_sp_rif_params *params,
10217                              struct netlink_ext_ack *extack)
10218 {
10219         return mlxsw_sp_fid_rfid_get(rif->mlxsw_sp, rif->rif_index);
10220 }
10221
10222 static const struct mlxsw_sp_rif_ops mlxsw_sp_rif_subport_ops = {
10223         .type                   = MLXSW_SP_RIF_TYPE_SUBPORT,
10224         .rif_size               = sizeof(struct mlxsw_sp_rif_subport),
10225         .setup                  = mlxsw_sp_rif_subport_setup,
10226         .configure              = mlxsw_sp_rif_subport_configure,
10227         .deconfigure            = mlxsw_sp_rif_subport_deconfigure,
10228         .fid_get                = mlxsw_sp_rif_subport_fid_get,
10229 };
10230
10231 static int mlxsw_sp_rif_fid_op(struct mlxsw_sp_rif *rif, u16 fid, bool enable)
10232 {
10233         enum mlxsw_reg_ritr_if_type type = MLXSW_REG_RITR_FID_IF;
10234         struct net_device *dev = mlxsw_sp_rif_dev(rif);
10235         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
10236         char ritr_pl[MLXSW_REG_RITR_LEN];
10237
10238         mlxsw_reg_ritr_pack(ritr_pl, enable, type, rif->rif_index, rif->vr_id,
10239                             dev->mtu);
10240         mlxsw_reg_ritr_mac_pack(ritr_pl, dev->dev_addr);
10241         mlxsw_reg_ritr_if_mac_profile_id_set(ritr_pl, rif->mac_profile_id);
10242         mlxsw_reg_ritr_fid_if_fid_set(ritr_pl, fid);
10243
10244         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
10245 }
10246
10247 u16 mlxsw_sp_router_port(const struct mlxsw_sp *mlxsw_sp)
10248 {
10249         return mlxsw_core_max_ports(mlxsw_sp->core) + 1;
10250 }
10251
10252 static int mlxsw_sp_rif_fid_configure(struct mlxsw_sp_rif *rif,
10253                                       struct netlink_ext_ack *extack)
10254 {
10255         struct net_device *dev = mlxsw_sp_rif_dev(rif);
10256         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
10257         u16 fid_index = mlxsw_sp_fid_index(rif->fid);
10258         u8 mac_profile;
10259         int err;
10260
10261         err = mlxsw_sp_rif_mac_profile_get(mlxsw_sp, rif->addr,
10262                                            &mac_profile, extack);
10263         if (err)
10264                 return err;
10265         rif->mac_profile_id = mac_profile;
10266
10267         err = mlxsw_sp_rif_fid_op(rif, fid_index, true);
10268         if (err)
10269                 goto err_rif_fid_op;
10270
10271         err = mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_MC,
10272                                      mlxsw_sp_router_port(mlxsw_sp), true);
10273         if (err)
10274                 goto err_fid_mc_flood_set;
10275
10276         err = mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_BC,
10277                                      mlxsw_sp_router_port(mlxsw_sp), true);
10278         if (err)
10279                 goto err_fid_bc_flood_set;
10280
10281         err = mlxsw_sp_macvlan_replay(rif, extack);
10282         if (err)
10283                 goto err_macvlan_replay;
10284
10285         err = mlxsw_sp_rif_fdb_op(rif->mlxsw_sp, dev->dev_addr,
10286                                   mlxsw_sp_fid_index(rif->fid), true);
10287         if (err)
10288                 goto err_rif_fdb_op;
10289
10290         err = mlxsw_sp_fid_rif_set(rif->fid, rif);
10291         if (err)
10292                 goto err_fid_rif_set;
10293
10294         return 0;
10295
10296 err_fid_rif_set:
10297         mlxsw_sp_rif_fdb_op(rif->mlxsw_sp, dev->dev_addr,
10298                             mlxsw_sp_fid_index(rif->fid), false);
10299 err_rif_fdb_op:
10300         mlxsw_sp_rif_macvlan_flush(rif);
10301 err_macvlan_replay:
10302         mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_BC,
10303                                mlxsw_sp_router_port(mlxsw_sp), false);
10304 err_fid_bc_flood_set:
10305         mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_MC,
10306                                mlxsw_sp_router_port(mlxsw_sp), false);
10307 err_fid_mc_flood_set:
10308         mlxsw_sp_rif_fid_op(rif, fid_index, false);
10309 err_rif_fid_op:
10310         mlxsw_sp_rif_mac_profile_put(mlxsw_sp, mac_profile);
10311         return err;
10312 }
10313
10314 static void mlxsw_sp_rif_fid_deconfigure(struct mlxsw_sp_rif *rif)
10315 {
10316         struct net_device *dev = mlxsw_sp_rif_dev(rif);
10317         u16 fid_index = mlxsw_sp_fid_index(rif->fid);
10318         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
10319         struct mlxsw_sp_fid *fid = rif->fid;
10320
10321         mlxsw_sp_fid_rif_unset(fid);
10322         mlxsw_sp_rif_fdb_op(rif->mlxsw_sp, dev->dev_addr,
10323                             mlxsw_sp_fid_index(fid), false);
10324         mlxsw_sp_rif_macvlan_flush(rif);
10325         mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_BC,
10326                                mlxsw_sp_router_port(mlxsw_sp), false);
10327         mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_MC,
10328                                mlxsw_sp_router_port(mlxsw_sp), false);
10329         mlxsw_sp_rif_fid_op(rif, fid_index, false);
10330         mlxsw_sp_rif_mac_profile_put(rif->mlxsw_sp, rif->mac_profile_id);
10331 }
10332
10333 static struct mlxsw_sp_fid *
10334 mlxsw_sp_rif_fid_fid_get(struct mlxsw_sp_rif *rif,
10335                          const struct mlxsw_sp_rif_params *params,
10336                          struct netlink_ext_ack *extack)
10337 {
10338         int rif_ifindex = mlxsw_sp_rif_dev_ifindex(rif);
10339
10340         return mlxsw_sp_fid_8021d_get(rif->mlxsw_sp, rif_ifindex);
10341 }
10342
10343 static void mlxsw_sp_rif_fid_fdb_del(struct mlxsw_sp_rif *rif, const char *mac)
10344 {
10345         struct switchdev_notifier_fdb_info info = {};
10346         struct net_device *dev;
10347
10348         dev = br_fdb_find_port(mlxsw_sp_rif_dev(rif), mac, 0);
10349         if (!dev)
10350                 return;
10351
10352         info.addr = mac;
10353         info.vid = 0;
10354         call_switchdev_notifiers(SWITCHDEV_FDB_DEL_TO_BRIDGE, dev, &info.info,
10355                                  NULL);
10356 }
10357
10358 static const struct mlxsw_sp_rif_ops mlxsw_sp_rif_fid_ops = {
10359         .type                   = MLXSW_SP_RIF_TYPE_FID,
10360         .rif_size               = sizeof(struct mlxsw_sp_rif),
10361         .configure              = mlxsw_sp_rif_fid_configure,
10362         .deconfigure            = mlxsw_sp_rif_fid_deconfigure,
10363         .fid_get                = mlxsw_sp_rif_fid_fid_get,
10364         .fdb_del                = mlxsw_sp_rif_fid_fdb_del,
10365 };
10366
10367 static struct mlxsw_sp_fid *
10368 mlxsw_sp_rif_vlan_fid_get(struct mlxsw_sp_rif *rif,
10369                           const struct mlxsw_sp_rif_params *params,
10370                           struct netlink_ext_ack *extack)
10371 {
10372         struct net_device *dev = mlxsw_sp_rif_dev(rif);
10373         struct net_device *br_dev;
10374
10375         if (WARN_ON(!params->vid))
10376                 return ERR_PTR(-EINVAL);
10377
10378         if (is_vlan_dev(dev)) {
10379                 br_dev = vlan_dev_real_dev(dev);
10380                 if (WARN_ON(!netif_is_bridge_master(br_dev)))
10381                         return ERR_PTR(-EINVAL);
10382         }
10383
10384         return mlxsw_sp_fid_8021q_get(rif->mlxsw_sp, params->vid);
10385 }
10386
10387 static void mlxsw_sp_rif_vlan_fdb_del(struct mlxsw_sp_rif *rif, const char *mac)
10388 {
10389         struct net_device *rif_dev = mlxsw_sp_rif_dev(rif);
10390         struct switchdev_notifier_fdb_info info = {};
10391         u16 vid = mlxsw_sp_fid_8021q_vid(rif->fid);
10392         struct net_device *br_dev;
10393         struct net_device *dev;
10394
10395         br_dev = is_vlan_dev(rif_dev) ? vlan_dev_real_dev(rif_dev) : rif_dev;
10396         dev = br_fdb_find_port(br_dev, mac, vid);
10397         if (!dev)
10398                 return;
10399
10400         info.addr = mac;
10401         info.vid = vid;
10402         call_switchdev_notifiers(SWITCHDEV_FDB_DEL_TO_BRIDGE, dev, &info.info,
10403                                  NULL);
10404 }
10405
10406 static int mlxsw_sp_rif_vlan_op(struct mlxsw_sp_rif *rif, u16 vid, u16 efid,
10407                                 bool enable)
10408 {
10409         struct net_device *dev = mlxsw_sp_rif_dev(rif);
10410         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
10411         char ritr_pl[MLXSW_REG_RITR_LEN];
10412
10413         mlxsw_reg_ritr_vlan_if_pack(ritr_pl, enable, rif->rif_index, rif->vr_id,
10414                                     dev->mtu, dev->dev_addr,
10415                                     rif->mac_profile_id, vid, efid);
10416
10417         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
10418 }
10419
10420 static int mlxsw_sp_rif_vlan_configure(struct mlxsw_sp_rif *rif, u16 efid,
10421                                        struct netlink_ext_ack *extack)
10422 {
10423         struct net_device *dev = mlxsw_sp_rif_dev(rif);
10424         u16 vid = mlxsw_sp_fid_8021q_vid(rif->fid);
10425         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
10426         u8 mac_profile;
10427         int err;
10428
10429         err = mlxsw_sp_rif_mac_profile_get(mlxsw_sp, rif->addr,
10430                                            &mac_profile, extack);
10431         if (err)
10432                 return err;
10433         rif->mac_profile_id = mac_profile;
10434
10435         err = mlxsw_sp_rif_vlan_op(rif, vid, efid, true);
10436         if (err)
10437                 goto err_rif_vlan_fid_op;
10438
10439         err = mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_MC,
10440                                      mlxsw_sp_router_port(mlxsw_sp), true);
10441         if (err)
10442                 goto err_fid_mc_flood_set;
10443
10444         err = mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_BC,
10445                                      mlxsw_sp_router_port(mlxsw_sp), true);
10446         if (err)
10447                 goto err_fid_bc_flood_set;
10448
10449         err = mlxsw_sp_macvlan_replay(rif, extack);
10450         if (err)
10451                 goto err_macvlan_replay;
10452
10453         err = mlxsw_sp_rif_fdb_op(rif->mlxsw_sp, dev->dev_addr,
10454                                   mlxsw_sp_fid_index(rif->fid), true);
10455         if (err)
10456                 goto err_rif_fdb_op;
10457
10458         err = mlxsw_sp_fid_rif_set(rif->fid, rif);
10459         if (err)
10460                 goto err_fid_rif_set;
10461
10462         return 0;
10463
10464 err_fid_rif_set:
10465         mlxsw_sp_rif_fdb_op(rif->mlxsw_sp, dev->dev_addr,
10466                             mlxsw_sp_fid_index(rif->fid), false);
10467 err_rif_fdb_op:
10468         mlxsw_sp_rif_macvlan_flush(rif);
10469 err_macvlan_replay:
10470         mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_BC,
10471                                mlxsw_sp_router_port(mlxsw_sp), false);
10472 err_fid_bc_flood_set:
10473         mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_MC,
10474                                mlxsw_sp_router_port(mlxsw_sp), false);
10475 err_fid_mc_flood_set:
10476         mlxsw_sp_rif_vlan_op(rif, vid, 0, false);
10477 err_rif_vlan_fid_op:
10478         mlxsw_sp_rif_mac_profile_put(mlxsw_sp, mac_profile);
10479         return err;
10480 }
10481
10482 static void mlxsw_sp_rif_vlan_deconfigure(struct mlxsw_sp_rif *rif)
10483 {
10484         struct net_device *dev = mlxsw_sp_rif_dev(rif);
10485         u16 vid = mlxsw_sp_fid_8021q_vid(rif->fid);
10486         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
10487
10488         mlxsw_sp_fid_rif_unset(rif->fid);
10489         mlxsw_sp_rif_fdb_op(rif->mlxsw_sp, dev->dev_addr,
10490                             mlxsw_sp_fid_index(rif->fid), false);
10491         mlxsw_sp_rif_macvlan_flush(rif);
10492         mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_BC,
10493                                mlxsw_sp_router_port(mlxsw_sp), false);
10494         mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_MC,
10495                                mlxsw_sp_router_port(mlxsw_sp), false);
10496         mlxsw_sp_rif_vlan_op(rif, vid, 0, false);
10497         mlxsw_sp_rif_mac_profile_put(rif->mlxsw_sp, rif->mac_profile_id);
10498 }
10499
10500 static int mlxsw_sp1_rif_vlan_configure(struct mlxsw_sp_rif *rif,
10501                                         struct netlink_ext_ack *extack)
10502 {
10503         return mlxsw_sp_rif_vlan_configure(rif, 0, extack);
10504 }
10505
10506 static const struct mlxsw_sp_rif_ops mlxsw_sp1_rif_vlan_ops = {
10507         .type                   = MLXSW_SP_RIF_TYPE_VLAN,
10508         .rif_size               = sizeof(struct mlxsw_sp_rif),
10509         .configure              = mlxsw_sp1_rif_vlan_configure,
10510         .deconfigure            = mlxsw_sp_rif_vlan_deconfigure,
10511         .fid_get                = mlxsw_sp_rif_vlan_fid_get,
10512         .fdb_del                = mlxsw_sp_rif_vlan_fdb_del,
10513 };
10514
10515 static int mlxsw_sp2_rif_vlan_configure(struct mlxsw_sp_rif *rif,
10516                                         struct netlink_ext_ack *extack)
10517 {
10518         u16 efid = mlxsw_sp_fid_index(rif->fid);
10519
10520         return mlxsw_sp_rif_vlan_configure(rif, efid, extack);
10521 }
10522
10523 static const struct mlxsw_sp_rif_ops mlxsw_sp2_rif_vlan_ops = {
10524         .type                   = MLXSW_SP_RIF_TYPE_VLAN,
10525         .rif_size               = sizeof(struct mlxsw_sp_rif),
10526         .configure              = mlxsw_sp2_rif_vlan_configure,
10527         .deconfigure            = mlxsw_sp_rif_vlan_deconfigure,
10528         .fid_get                = mlxsw_sp_rif_vlan_fid_get,
10529         .fdb_del                = mlxsw_sp_rif_vlan_fdb_del,
10530 };
10531
10532 static struct mlxsw_sp_rif_ipip_lb *
10533 mlxsw_sp_rif_ipip_lb_rif(struct mlxsw_sp_rif *rif)
10534 {
10535         return container_of(rif, struct mlxsw_sp_rif_ipip_lb, common);
10536 }
10537
10538 static void
10539 mlxsw_sp_rif_ipip_lb_setup(struct mlxsw_sp_rif *rif,
10540                            const struct mlxsw_sp_rif_params *params)
10541 {
10542         struct mlxsw_sp_rif_params_ipip_lb *params_lb;
10543         struct mlxsw_sp_rif_ipip_lb *rif_lb;
10544
10545         params_lb = container_of(params, struct mlxsw_sp_rif_params_ipip_lb,
10546                                  common);
10547         rif_lb = mlxsw_sp_rif_ipip_lb_rif(rif);
10548         rif_lb->lb_config = params_lb->lb_config;
10549 }
10550
10551 static int
10552 mlxsw_sp1_rif_ipip_lb_configure(struct mlxsw_sp_rif *rif,
10553                                 struct netlink_ext_ack *extack)
10554 {
10555         struct mlxsw_sp_rif_ipip_lb *lb_rif = mlxsw_sp_rif_ipip_lb_rif(rif);
10556         struct net_device *dev = mlxsw_sp_rif_dev(rif);
10557         u32 ul_tb_id = mlxsw_sp_ipip_dev_ul_tb_id(dev);
10558         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
10559         struct mlxsw_sp_vr *ul_vr;
10560         int err;
10561
10562         ul_vr = mlxsw_sp_vr_get(mlxsw_sp, ul_tb_id, extack);
10563         if (IS_ERR(ul_vr))
10564                 return PTR_ERR(ul_vr);
10565
10566         err = mlxsw_sp_rif_ipip_lb_op(lb_rif, ul_vr->id, 0, true);
10567         if (err)
10568                 goto err_loopback_op;
10569
10570         lb_rif->ul_vr_id = ul_vr->id;
10571         lb_rif->ul_rif_id = 0;
10572         ++ul_vr->rif_count;
10573         return 0;
10574
10575 err_loopback_op:
10576         mlxsw_sp_vr_put(mlxsw_sp, ul_vr);
10577         return err;
10578 }
10579
10580 static void mlxsw_sp1_rif_ipip_lb_deconfigure(struct mlxsw_sp_rif *rif)
10581 {
10582         struct mlxsw_sp_rif_ipip_lb *lb_rif = mlxsw_sp_rif_ipip_lb_rif(rif);
10583         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
10584         struct mlxsw_sp_vr *ul_vr;
10585
10586         ul_vr = &mlxsw_sp->router->vrs[lb_rif->ul_vr_id];
10587         mlxsw_sp_rif_ipip_lb_op(lb_rif, ul_vr->id, 0, false);
10588
10589         --ul_vr->rif_count;
10590         mlxsw_sp_vr_put(mlxsw_sp, ul_vr);
10591 }
10592
10593 static const struct mlxsw_sp_rif_ops mlxsw_sp1_rif_ipip_lb_ops = {
10594         .type                   = MLXSW_SP_RIF_TYPE_IPIP_LB,
10595         .rif_size               = sizeof(struct mlxsw_sp_rif_ipip_lb),
10596         .setup                  = mlxsw_sp_rif_ipip_lb_setup,
10597         .configure              = mlxsw_sp1_rif_ipip_lb_configure,
10598         .deconfigure            = mlxsw_sp1_rif_ipip_lb_deconfigure,
10599 };
10600
10601 static const struct mlxsw_sp_rif_ops *mlxsw_sp1_rif_ops_arr[] = {
10602         [MLXSW_SP_RIF_TYPE_SUBPORT]     = &mlxsw_sp_rif_subport_ops,
10603         [MLXSW_SP_RIF_TYPE_VLAN]        = &mlxsw_sp1_rif_vlan_ops,
10604         [MLXSW_SP_RIF_TYPE_FID]         = &mlxsw_sp_rif_fid_ops,
10605         [MLXSW_SP_RIF_TYPE_IPIP_LB]     = &mlxsw_sp1_rif_ipip_lb_ops,
10606 };
10607
10608 static int
10609 mlxsw_sp_rif_ipip_lb_ul_rif_op(struct mlxsw_sp_rif *ul_rif, bool enable)
10610 {
10611         struct mlxsw_sp *mlxsw_sp = ul_rif->mlxsw_sp;
10612         char ritr_pl[MLXSW_REG_RITR_LEN];
10613
10614         mlxsw_reg_ritr_pack(ritr_pl, enable, MLXSW_REG_RITR_LOOPBACK_IF,
10615                             ul_rif->rif_index, ul_rif->vr_id, IP_MAX_MTU);
10616         mlxsw_reg_ritr_loopback_protocol_set(ritr_pl,
10617                                              MLXSW_REG_RITR_LOOPBACK_GENERIC);
10618
10619         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
10620 }
10621
10622 static struct mlxsw_sp_rif *
10623 mlxsw_sp_ul_rif_create(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_vr *vr,
10624                        struct mlxsw_sp_crif *ul_crif,
10625                        struct netlink_ext_ack *extack)
10626 {
10627         struct mlxsw_sp_rif *ul_rif;
10628         u8 rif_entries = 1;
10629         u16 rif_index;
10630         int err;
10631
10632         err = mlxsw_sp_rif_index_alloc(mlxsw_sp, &rif_index, rif_entries);
10633         if (err) {
10634                 NL_SET_ERR_MSG_MOD(extack, "Exceeded number of supported router interfaces");
10635                 return ERR_PTR(err);
10636         }
10637
10638         ul_rif = mlxsw_sp_rif_alloc(sizeof(*ul_rif), rif_index, vr->id,
10639                                     ul_crif);
10640         if (!ul_rif) {
10641                 err = -ENOMEM;
10642                 goto err_rif_alloc;
10643         }
10644
10645         mlxsw_sp->router->rifs[rif_index] = ul_rif;
10646         ul_rif->mlxsw_sp = mlxsw_sp;
10647         ul_rif->rif_entries = rif_entries;
10648         err = mlxsw_sp_rif_ipip_lb_ul_rif_op(ul_rif, true);
10649         if (err)
10650                 goto ul_rif_op_err;
10651
10652         atomic_add(rif_entries, &mlxsw_sp->router->rifs_count);
10653         return ul_rif;
10654
10655 ul_rif_op_err:
10656         mlxsw_sp->router->rifs[rif_index] = NULL;
10657         mlxsw_sp_rif_free(ul_rif);
10658 err_rif_alloc:
10659         mlxsw_sp_rif_index_free(mlxsw_sp, rif_index, rif_entries);
10660         return ERR_PTR(err);
10661 }
10662
10663 static void mlxsw_sp_ul_rif_destroy(struct mlxsw_sp_rif *ul_rif)
10664 {
10665         struct mlxsw_sp *mlxsw_sp = ul_rif->mlxsw_sp;
10666         u8 rif_entries = ul_rif->rif_entries;
10667         u16 rif_index = ul_rif->rif_index;
10668
10669         atomic_sub(rif_entries, &mlxsw_sp->router->rifs_count);
10670         mlxsw_sp_rif_ipip_lb_ul_rif_op(ul_rif, false);
10671         mlxsw_sp->router->rifs[ul_rif->rif_index] = NULL;
10672         mlxsw_sp_rif_free(ul_rif);
10673         mlxsw_sp_rif_index_free(mlxsw_sp, rif_index, rif_entries);
10674 }
10675
10676 static struct mlxsw_sp_rif *
10677 mlxsw_sp_ul_rif_get(struct mlxsw_sp *mlxsw_sp, u32 tb_id,
10678                     struct mlxsw_sp_crif *ul_crif,
10679                     struct netlink_ext_ack *extack)
10680 {
10681         struct mlxsw_sp_vr *vr;
10682         int err;
10683
10684         vr = mlxsw_sp_vr_get(mlxsw_sp, tb_id, extack);
10685         if (IS_ERR(vr))
10686                 return ERR_CAST(vr);
10687
10688         if (refcount_inc_not_zero(&vr->ul_rif_refcnt))
10689                 return vr->ul_rif;
10690
10691         vr->ul_rif = mlxsw_sp_ul_rif_create(mlxsw_sp, vr, ul_crif, extack);
10692         if (IS_ERR(vr->ul_rif)) {
10693                 err = PTR_ERR(vr->ul_rif);
10694                 goto err_ul_rif_create;
10695         }
10696
10697         vr->rif_count++;
10698         refcount_set(&vr->ul_rif_refcnt, 1);
10699
10700         return vr->ul_rif;
10701
10702 err_ul_rif_create:
10703         mlxsw_sp_vr_put(mlxsw_sp, vr);
10704         return ERR_PTR(err);
10705 }
10706
10707 static void mlxsw_sp_ul_rif_put(struct mlxsw_sp_rif *ul_rif)
10708 {
10709         struct mlxsw_sp *mlxsw_sp = ul_rif->mlxsw_sp;
10710         struct mlxsw_sp_vr *vr;
10711
10712         vr = &mlxsw_sp->router->vrs[ul_rif->vr_id];
10713
10714         if (!refcount_dec_and_test(&vr->ul_rif_refcnt))
10715                 return;
10716
10717         vr->rif_count--;
10718         mlxsw_sp_ul_rif_destroy(ul_rif);
10719         mlxsw_sp_vr_put(mlxsw_sp, vr);
10720 }
10721
10722 int mlxsw_sp_router_ul_rif_get(struct mlxsw_sp *mlxsw_sp, u32 ul_tb_id,
10723                                u16 *ul_rif_index)
10724 {
10725         struct mlxsw_sp_rif *ul_rif;
10726         int err = 0;
10727
10728         mutex_lock(&mlxsw_sp->router->lock);
10729         ul_rif = mlxsw_sp_ul_rif_get(mlxsw_sp, ul_tb_id, NULL, NULL);
10730         if (IS_ERR(ul_rif)) {
10731                 err = PTR_ERR(ul_rif);
10732                 goto out;
10733         }
10734         *ul_rif_index = ul_rif->rif_index;
10735 out:
10736         mutex_unlock(&mlxsw_sp->router->lock);
10737         return err;
10738 }
10739
10740 void mlxsw_sp_router_ul_rif_put(struct mlxsw_sp *mlxsw_sp, u16 ul_rif_index)
10741 {
10742         struct mlxsw_sp_rif *ul_rif;
10743
10744         mutex_lock(&mlxsw_sp->router->lock);
10745         ul_rif = mlxsw_sp->router->rifs[ul_rif_index];
10746         if (WARN_ON(!ul_rif))
10747                 goto out;
10748
10749         mlxsw_sp_ul_rif_put(ul_rif);
10750 out:
10751         mutex_unlock(&mlxsw_sp->router->lock);
10752 }
10753
10754 static int
10755 mlxsw_sp2_rif_ipip_lb_configure(struct mlxsw_sp_rif *rif,
10756                                 struct netlink_ext_ack *extack)
10757 {
10758         struct mlxsw_sp_rif_ipip_lb *lb_rif = mlxsw_sp_rif_ipip_lb_rif(rif);
10759         struct net_device *dev = mlxsw_sp_rif_dev(rif);
10760         u32 ul_tb_id = mlxsw_sp_ipip_dev_ul_tb_id(dev);
10761         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
10762         struct mlxsw_sp_rif *ul_rif;
10763         int err;
10764
10765         ul_rif = mlxsw_sp_ul_rif_get(mlxsw_sp, ul_tb_id, NULL, extack);
10766         if (IS_ERR(ul_rif))
10767                 return PTR_ERR(ul_rif);
10768
10769         err = mlxsw_sp_rif_ipip_lb_op(lb_rif, 0, ul_rif->rif_index, true);
10770         if (err)
10771                 goto err_loopback_op;
10772
10773         lb_rif->ul_vr_id = 0;
10774         lb_rif->ul_rif_id = ul_rif->rif_index;
10775
10776         return 0;
10777
10778 err_loopback_op:
10779         mlxsw_sp_ul_rif_put(ul_rif);
10780         return err;
10781 }
10782
10783 static void mlxsw_sp2_rif_ipip_lb_deconfigure(struct mlxsw_sp_rif *rif)
10784 {
10785         struct mlxsw_sp_rif_ipip_lb *lb_rif = mlxsw_sp_rif_ipip_lb_rif(rif);
10786         struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
10787         struct mlxsw_sp_rif *ul_rif;
10788
10789         ul_rif = mlxsw_sp_rif_by_index(mlxsw_sp, lb_rif->ul_rif_id);
10790         mlxsw_sp_rif_ipip_lb_op(lb_rif, 0, lb_rif->ul_rif_id, false);
10791         mlxsw_sp_ul_rif_put(ul_rif);
10792 }
10793
10794 static const struct mlxsw_sp_rif_ops mlxsw_sp2_rif_ipip_lb_ops = {
10795         .type                   = MLXSW_SP_RIF_TYPE_IPIP_LB,
10796         .rif_size               = sizeof(struct mlxsw_sp_rif_ipip_lb),
10797         .setup                  = mlxsw_sp_rif_ipip_lb_setup,
10798         .configure              = mlxsw_sp2_rif_ipip_lb_configure,
10799         .deconfigure            = mlxsw_sp2_rif_ipip_lb_deconfigure,
10800 };
10801
10802 static const struct mlxsw_sp_rif_ops *mlxsw_sp2_rif_ops_arr[] = {
10803         [MLXSW_SP_RIF_TYPE_SUBPORT]     = &mlxsw_sp_rif_subport_ops,
10804         [MLXSW_SP_RIF_TYPE_VLAN]        = &mlxsw_sp2_rif_vlan_ops,
10805         [MLXSW_SP_RIF_TYPE_FID]         = &mlxsw_sp_rif_fid_ops,
10806         [MLXSW_SP_RIF_TYPE_IPIP_LB]     = &mlxsw_sp2_rif_ipip_lb_ops,
10807 };
10808
10809 static int mlxsw_sp_rifs_table_init(struct mlxsw_sp *mlxsw_sp)
10810 {
10811         struct gen_pool *rifs_table;
10812         int err;
10813
10814         rifs_table = gen_pool_create(0, -1);
10815         if (!rifs_table)
10816                 return -ENOMEM;
10817
10818         gen_pool_set_algo(rifs_table, gen_pool_first_fit_order_align,
10819                           NULL);
10820
10821         err = gen_pool_add(rifs_table, MLXSW_SP_ROUTER_GENALLOC_OFFSET,
10822                            MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_RIFS), -1);
10823         if (err)
10824                 goto err_gen_pool_add;
10825
10826         mlxsw_sp->router->rifs_table = rifs_table;
10827
10828         return 0;
10829
10830 err_gen_pool_add:
10831         gen_pool_destroy(rifs_table);
10832         return err;
10833 }
10834
10835 static void mlxsw_sp_rifs_table_fini(struct mlxsw_sp *mlxsw_sp)
10836 {
10837         gen_pool_destroy(mlxsw_sp->router->rifs_table);
10838 }
10839
10840 static int mlxsw_sp_rifs_init(struct mlxsw_sp *mlxsw_sp)
10841 {
10842         u64 max_rifs = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_RIFS);
10843         struct devlink *devlink = priv_to_devlink(mlxsw_sp->core);
10844         struct mlxsw_core *core = mlxsw_sp->core;
10845         int err;
10846
10847         if (!MLXSW_CORE_RES_VALID(core, MAX_RIF_MAC_PROFILES))
10848                 return -EIO;
10849         mlxsw_sp->router->max_rif_mac_profile =
10850                 MLXSW_CORE_RES_GET(core, MAX_RIF_MAC_PROFILES);
10851
10852         mlxsw_sp->router->rifs = kcalloc(max_rifs,
10853                                          sizeof(struct mlxsw_sp_rif *),
10854                                          GFP_KERNEL);
10855         if (!mlxsw_sp->router->rifs)
10856                 return -ENOMEM;
10857
10858         err = mlxsw_sp_rifs_table_init(mlxsw_sp);
10859         if (err)
10860                 goto err_rifs_table_init;
10861
10862         idr_init(&mlxsw_sp->router->rif_mac_profiles_idr);
10863         atomic_set(&mlxsw_sp->router->rif_mac_profiles_count, 0);
10864         atomic_set(&mlxsw_sp->router->rifs_count, 0);
10865         devl_resource_occ_get_register(devlink,
10866                                        MLXSW_SP_RESOURCE_RIF_MAC_PROFILES,
10867                                        mlxsw_sp_rif_mac_profiles_occ_get,
10868                                        mlxsw_sp);
10869         devl_resource_occ_get_register(devlink,
10870                                        MLXSW_SP_RESOURCE_RIFS,
10871                                        mlxsw_sp_rifs_occ_get,
10872                                        mlxsw_sp);
10873
10874         return 0;
10875
10876 err_rifs_table_init:
10877         kfree(mlxsw_sp->router->rifs);
10878         return err;
10879 }
10880
10881 static void mlxsw_sp_rifs_fini(struct mlxsw_sp *mlxsw_sp)
10882 {
10883         int max_rifs = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_RIFS);
10884         struct devlink *devlink = priv_to_devlink(mlxsw_sp->core);
10885         int i;
10886
10887         WARN_ON_ONCE(atomic_read(&mlxsw_sp->router->rifs_count));
10888         for (i = 0; i < max_rifs; i++)
10889                 WARN_ON_ONCE(mlxsw_sp->router->rifs[i]);
10890
10891         devl_resource_occ_get_unregister(devlink, MLXSW_SP_RESOURCE_RIFS);
10892         devl_resource_occ_get_unregister(devlink,
10893                                          MLXSW_SP_RESOURCE_RIF_MAC_PROFILES);
10894         WARN_ON(!idr_is_empty(&mlxsw_sp->router->rif_mac_profiles_idr));
10895         idr_destroy(&mlxsw_sp->router->rif_mac_profiles_idr);
10896         mlxsw_sp_rifs_table_fini(mlxsw_sp);
10897         kfree(mlxsw_sp->router->rifs);
10898 }
10899
10900 static int
10901 mlxsw_sp_ipip_config_tigcr(struct mlxsw_sp *mlxsw_sp)
10902 {
10903         char tigcr_pl[MLXSW_REG_TIGCR_LEN];
10904
10905         mlxsw_reg_tigcr_pack(tigcr_pl, true, 0);
10906         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(tigcr), tigcr_pl);
10907 }
10908
10909 static int mlxsw_sp_ipips_init(struct mlxsw_sp *mlxsw_sp)
10910 {
10911         int err;
10912
10913         INIT_LIST_HEAD(&mlxsw_sp->router->ipip_list);
10914
10915         err = mlxsw_sp_ipip_ecn_encap_init(mlxsw_sp);
10916         if (err)
10917                 return err;
10918         err = mlxsw_sp_ipip_ecn_decap_init(mlxsw_sp);
10919         if (err)
10920                 return err;
10921
10922         return mlxsw_sp_ipip_config_tigcr(mlxsw_sp);
10923 }
10924
10925 static int mlxsw_sp1_ipips_init(struct mlxsw_sp *mlxsw_sp)
10926 {
10927         mlxsw_sp->router->ipip_ops_arr = mlxsw_sp1_ipip_ops_arr;
10928         return mlxsw_sp_ipips_init(mlxsw_sp);
10929 }
10930
10931 static int mlxsw_sp2_ipips_init(struct mlxsw_sp *mlxsw_sp)
10932 {
10933         mlxsw_sp->router->ipip_ops_arr = mlxsw_sp2_ipip_ops_arr;
10934         return mlxsw_sp_ipips_init(mlxsw_sp);
10935 }
10936
10937 static void mlxsw_sp_ipips_fini(struct mlxsw_sp *mlxsw_sp)
10938 {
10939         WARN_ON(!list_empty(&mlxsw_sp->router->ipip_list));
10940 }
10941
10942 static void mlxsw_sp_router_fib_dump_flush(struct notifier_block *nb)
10943 {
10944         struct mlxsw_sp_router *router;
10945
10946         /* Flush pending FIB notifications and then flush the device's
10947          * table before requesting another dump. The FIB notification
10948          * block is unregistered, so no need to take RTNL.
10949          */
10950         mlxsw_core_flush_owq();
10951         router = container_of(nb, struct mlxsw_sp_router, fib_nb);
10952         mlxsw_sp_router_fib_flush(router->mlxsw_sp);
10953 }
10954
10955 #ifdef CONFIG_IP_ROUTE_MULTIPATH
10956 struct mlxsw_sp_mp_hash_config {
10957         DECLARE_BITMAP(headers, __MLXSW_REG_RECR2_HEADER_CNT);
10958         DECLARE_BITMAP(fields, __MLXSW_REG_RECR2_FIELD_CNT);
10959         DECLARE_BITMAP(inner_headers, __MLXSW_REG_RECR2_HEADER_CNT);
10960         DECLARE_BITMAP(inner_fields, __MLXSW_REG_RECR2_INNER_FIELD_CNT);
10961         bool inc_parsing_depth;
10962 };
10963
10964 #define MLXSW_SP_MP_HASH_HEADER_SET(_headers, _header) \
10965         bitmap_set(_headers, MLXSW_REG_RECR2_##_header, 1)
10966
10967 #define MLXSW_SP_MP_HASH_FIELD_SET(_fields, _field) \
10968         bitmap_set(_fields, MLXSW_REG_RECR2_##_field, 1)
10969
10970 #define MLXSW_SP_MP_HASH_FIELD_RANGE_SET(_fields, _field, _nr) \
10971         bitmap_set(_fields, MLXSW_REG_RECR2_##_field, _nr)
10972
10973 static void mlxsw_sp_mp_hash_inner_l3(struct mlxsw_sp_mp_hash_config *config)
10974 {
10975         unsigned long *inner_headers = config->inner_headers;
10976         unsigned long *inner_fields = config->inner_fields;
10977
10978         /* IPv4 inner */
10979         MLXSW_SP_MP_HASH_HEADER_SET(inner_headers, IPV4_EN_NOT_TCP_NOT_UDP);
10980         MLXSW_SP_MP_HASH_HEADER_SET(inner_headers, IPV4_EN_TCP_UDP);
10981         MLXSW_SP_MP_HASH_FIELD_RANGE_SET(inner_fields, INNER_IPV4_SIP0, 4);
10982         MLXSW_SP_MP_HASH_FIELD_RANGE_SET(inner_fields, INNER_IPV4_DIP0, 4);
10983         /* IPv6 inner */
10984         MLXSW_SP_MP_HASH_HEADER_SET(inner_headers, IPV6_EN_NOT_TCP_NOT_UDP);
10985         MLXSW_SP_MP_HASH_HEADER_SET(inner_headers, IPV6_EN_TCP_UDP);
10986         MLXSW_SP_MP_HASH_FIELD_SET(inner_fields, INNER_IPV6_SIP0_7);
10987         MLXSW_SP_MP_HASH_FIELD_RANGE_SET(inner_fields, INNER_IPV6_SIP8, 8);
10988         MLXSW_SP_MP_HASH_FIELD_SET(inner_fields, INNER_IPV6_DIP0_7);
10989         MLXSW_SP_MP_HASH_FIELD_RANGE_SET(inner_fields, INNER_IPV6_DIP8, 8);
10990         MLXSW_SP_MP_HASH_FIELD_SET(inner_fields, INNER_IPV6_NEXT_HEADER);
10991         MLXSW_SP_MP_HASH_FIELD_SET(inner_fields, INNER_IPV6_FLOW_LABEL);
10992 }
10993
10994 static void mlxsw_sp_mp4_hash_outer_addr(struct mlxsw_sp_mp_hash_config *config)
10995 {
10996         unsigned long *headers = config->headers;
10997         unsigned long *fields = config->fields;
10998
10999         MLXSW_SP_MP_HASH_HEADER_SET(headers, IPV4_EN_NOT_TCP_NOT_UDP);
11000         MLXSW_SP_MP_HASH_HEADER_SET(headers, IPV4_EN_TCP_UDP);
11001         MLXSW_SP_MP_HASH_FIELD_RANGE_SET(fields, IPV4_SIP0, 4);
11002         MLXSW_SP_MP_HASH_FIELD_RANGE_SET(fields, IPV4_DIP0, 4);
11003 }
11004
11005 static void
11006 mlxsw_sp_mp_hash_inner_custom(struct mlxsw_sp_mp_hash_config *config,
11007                               u32 hash_fields)
11008 {
11009         unsigned long *inner_headers = config->inner_headers;
11010         unsigned long *inner_fields = config->inner_fields;
11011
11012         /* IPv4 Inner */
11013         MLXSW_SP_MP_HASH_HEADER_SET(inner_headers, IPV4_EN_NOT_TCP_NOT_UDP);
11014         MLXSW_SP_MP_HASH_HEADER_SET(inner_headers, IPV4_EN_TCP_UDP);
11015         if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_SRC_IP)
11016                 MLXSW_SP_MP_HASH_FIELD_RANGE_SET(inner_fields, INNER_IPV4_SIP0, 4);
11017         if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_DST_IP)
11018                 MLXSW_SP_MP_HASH_FIELD_RANGE_SET(inner_fields, INNER_IPV4_DIP0, 4);
11019         if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_IP_PROTO)
11020                 MLXSW_SP_MP_HASH_FIELD_SET(inner_fields, INNER_IPV4_PROTOCOL);
11021         /* IPv6 inner */
11022         MLXSW_SP_MP_HASH_HEADER_SET(inner_headers, IPV6_EN_NOT_TCP_NOT_UDP);
11023         MLXSW_SP_MP_HASH_HEADER_SET(inner_headers, IPV6_EN_TCP_UDP);
11024         if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_SRC_IP) {
11025                 MLXSW_SP_MP_HASH_FIELD_SET(inner_fields, INNER_IPV6_SIP0_7);
11026                 MLXSW_SP_MP_HASH_FIELD_RANGE_SET(inner_fields, INNER_IPV6_SIP8, 8);
11027         }
11028         if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_DST_IP) {
11029                 MLXSW_SP_MP_HASH_FIELD_SET(inner_fields, INNER_IPV6_DIP0_7);
11030                 MLXSW_SP_MP_HASH_FIELD_RANGE_SET(inner_fields, INNER_IPV6_DIP8, 8);
11031         }
11032         if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_IP_PROTO)
11033                 MLXSW_SP_MP_HASH_FIELD_SET(inner_fields, INNER_IPV6_NEXT_HEADER);
11034         if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_FLOWLABEL)
11035                 MLXSW_SP_MP_HASH_FIELD_SET(inner_fields, INNER_IPV6_FLOW_LABEL);
11036         /* L4 inner */
11037         MLXSW_SP_MP_HASH_HEADER_SET(inner_headers, TCP_UDP_EN_IPV4);
11038         MLXSW_SP_MP_HASH_HEADER_SET(inner_headers, TCP_UDP_EN_IPV6);
11039         if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_SRC_PORT)
11040                 MLXSW_SP_MP_HASH_FIELD_SET(inner_fields, INNER_TCP_UDP_SPORT);
11041         if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_DST_PORT)
11042                 MLXSW_SP_MP_HASH_FIELD_SET(inner_fields, INNER_TCP_UDP_DPORT);
11043 }
11044
11045 static void mlxsw_sp_mp4_hash_init(struct mlxsw_sp *mlxsw_sp,
11046                                    struct mlxsw_sp_mp_hash_config *config)
11047 {
11048         struct net *net = mlxsw_sp_net(mlxsw_sp);
11049         unsigned long *headers = config->headers;
11050         unsigned long *fields = config->fields;
11051         u32 hash_fields;
11052
11053         switch (READ_ONCE(net->ipv4.sysctl_fib_multipath_hash_policy)) {
11054         case 0:
11055                 mlxsw_sp_mp4_hash_outer_addr(config);
11056                 break;
11057         case 1:
11058                 mlxsw_sp_mp4_hash_outer_addr(config);
11059                 MLXSW_SP_MP_HASH_HEADER_SET(headers, TCP_UDP_EN_IPV4);
11060                 MLXSW_SP_MP_HASH_FIELD_SET(fields, IPV4_PROTOCOL);
11061                 MLXSW_SP_MP_HASH_FIELD_SET(fields, TCP_UDP_SPORT);
11062                 MLXSW_SP_MP_HASH_FIELD_SET(fields, TCP_UDP_DPORT);
11063                 break;
11064         case 2:
11065                 /* Outer */
11066                 mlxsw_sp_mp4_hash_outer_addr(config);
11067                 /* Inner */
11068                 mlxsw_sp_mp_hash_inner_l3(config);
11069                 break;
11070         case 3:
11071                 hash_fields = READ_ONCE(net->ipv4.sysctl_fib_multipath_hash_fields);
11072                 /* Outer */
11073                 MLXSW_SP_MP_HASH_HEADER_SET(headers, IPV4_EN_NOT_TCP_NOT_UDP);
11074                 MLXSW_SP_MP_HASH_HEADER_SET(headers, IPV4_EN_TCP_UDP);
11075                 MLXSW_SP_MP_HASH_HEADER_SET(headers, TCP_UDP_EN_IPV4);
11076                 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_SRC_IP)
11077                         MLXSW_SP_MP_HASH_FIELD_RANGE_SET(fields, IPV4_SIP0, 4);
11078                 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_DST_IP)
11079                         MLXSW_SP_MP_HASH_FIELD_RANGE_SET(fields, IPV4_DIP0, 4);
11080                 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_IP_PROTO)
11081                         MLXSW_SP_MP_HASH_FIELD_SET(fields, IPV4_PROTOCOL);
11082                 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_SRC_PORT)
11083                         MLXSW_SP_MP_HASH_FIELD_SET(fields, TCP_UDP_SPORT);
11084                 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_DST_PORT)
11085                         MLXSW_SP_MP_HASH_FIELD_SET(fields, TCP_UDP_DPORT);
11086                 /* Inner */
11087                 mlxsw_sp_mp_hash_inner_custom(config, hash_fields);
11088                 break;
11089         }
11090 }
11091
11092 static void mlxsw_sp_mp6_hash_outer_addr(struct mlxsw_sp_mp_hash_config *config)
11093 {
11094         unsigned long *headers = config->headers;
11095         unsigned long *fields = config->fields;
11096
11097         MLXSW_SP_MP_HASH_HEADER_SET(headers, IPV6_EN_NOT_TCP_NOT_UDP);
11098         MLXSW_SP_MP_HASH_HEADER_SET(headers, IPV6_EN_TCP_UDP);
11099         MLXSW_SP_MP_HASH_FIELD_SET(fields, IPV6_SIP0_7);
11100         MLXSW_SP_MP_HASH_FIELD_RANGE_SET(fields, IPV6_SIP8, 8);
11101         MLXSW_SP_MP_HASH_FIELD_SET(fields, IPV6_DIP0_7);
11102         MLXSW_SP_MP_HASH_FIELD_RANGE_SET(fields, IPV6_DIP8, 8);
11103 }
11104
11105 static void mlxsw_sp_mp6_hash_init(struct mlxsw_sp *mlxsw_sp,
11106                                    struct mlxsw_sp_mp_hash_config *config)
11107 {
11108         u32 hash_fields = ip6_multipath_hash_fields(mlxsw_sp_net(mlxsw_sp));
11109         unsigned long *headers = config->headers;
11110         unsigned long *fields = config->fields;
11111
11112         switch (ip6_multipath_hash_policy(mlxsw_sp_net(mlxsw_sp))) {
11113         case 0:
11114                 mlxsw_sp_mp6_hash_outer_addr(config);
11115                 MLXSW_SP_MP_HASH_FIELD_SET(fields, IPV6_NEXT_HEADER);
11116                 MLXSW_SP_MP_HASH_FIELD_SET(fields, IPV6_FLOW_LABEL);
11117                 break;
11118         case 1:
11119                 mlxsw_sp_mp6_hash_outer_addr(config);
11120                 MLXSW_SP_MP_HASH_HEADER_SET(headers, TCP_UDP_EN_IPV6);
11121                 MLXSW_SP_MP_HASH_FIELD_SET(fields, IPV6_NEXT_HEADER);
11122                 MLXSW_SP_MP_HASH_FIELD_SET(fields, TCP_UDP_SPORT);
11123                 MLXSW_SP_MP_HASH_FIELD_SET(fields, TCP_UDP_DPORT);
11124                 break;
11125         case 2:
11126                 /* Outer */
11127                 mlxsw_sp_mp6_hash_outer_addr(config);
11128                 MLXSW_SP_MP_HASH_FIELD_SET(fields, IPV6_NEXT_HEADER);
11129                 MLXSW_SP_MP_HASH_FIELD_SET(fields, IPV6_FLOW_LABEL);
11130                 /* Inner */
11131                 mlxsw_sp_mp_hash_inner_l3(config);
11132                 config->inc_parsing_depth = true;
11133                 break;
11134         case 3:
11135                 /* Outer */
11136                 MLXSW_SP_MP_HASH_HEADER_SET(headers, IPV6_EN_NOT_TCP_NOT_UDP);
11137                 MLXSW_SP_MP_HASH_HEADER_SET(headers, IPV6_EN_TCP_UDP);
11138                 MLXSW_SP_MP_HASH_HEADER_SET(headers, TCP_UDP_EN_IPV6);
11139                 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_SRC_IP) {
11140                         MLXSW_SP_MP_HASH_FIELD_SET(fields, IPV6_SIP0_7);
11141                         MLXSW_SP_MP_HASH_FIELD_RANGE_SET(fields, IPV6_SIP8, 8);
11142                 }
11143                 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_DST_IP) {
11144                         MLXSW_SP_MP_HASH_FIELD_SET(fields, IPV6_DIP0_7);
11145                         MLXSW_SP_MP_HASH_FIELD_RANGE_SET(fields, IPV6_DIP8, 8);
11146                 }
11147                 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_IP_PROTO)
11148                         MLXSW_SP_MP_HASH_FIELD_SET(fields, IPV6_NEXT_HEADER);
11149                 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_FLOWLABEL)
11150                         MLXSW_SP_MP_HASH_FIELD_SET(fields, IPV6_FLOW_LABEL);
11151                 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_SRC_PORT)
11152                         MLXSW_SP_MP_HASH_FIELD_SET(fields, TCP_UDP_SPORT);
11153                 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_DST_PORT)
11154                         MLXSW_SP_MP_HASH_FIELD_SET(fields, TCP_UDP_DPORT);
11155                 /* Inner */
11156                 mlxsw_sp_mp_hash_inner_custom(config, hash_fields);
11157                 if (hash_fields & FIB_MULTIPATH_HASH_FIELD_INNER_MASK)
11158                         config->inc_parsing_depth = true;
11159                 break;
11160         }
11161 }
11162
11163 static int mlxsw_sp_mp_hash_parsing_depth_adjust(struct mlxsw_sp *mlxsw_sp,
11164                                                  bool old_inc_parsing_depth,
11165                                                  bool new_inc_parsing_depth)
11166 {
11167         int err;
11168
11169         if (!old_inc_parsing_depth && new_inc_parsing_depth) {
11170                 err = mlxsw_sp_parsing_depth_inc(mlxsw_sp);
11171                 if (err)
11172                         return err;
11173                 mlxsw_sp->router->inc_parsing_depth = true;
11174         } else if (old_inc_parsing_depth && !new_inc_parsing_depth) {
11175                 mlxsw_sp_parsing_depth_dec(mlxsw_sp);
11176                 mlxsw_sp->router->inc_parsing_depth = false;
11177         }
11178
11179         return 0;
11180 }
11181
11182 static int mlxsw_sp_mp_hash_init(struct mlxsw_sp *mlxsw_sp)
11183 {
11184         bool old_inc_parsing_depth, new_inc_parsing_depth;
11185         struct mlxsw_sp_mp_hash_config config = {};
11186         char recr2_pl[MLXSW_REG_RECR2_LEN];
11187         unsigned long bit;
11188         u32 seed;
11189         int err;
11190
11191         seed = jhash(mlxsw_sp->base_mac, sizeof(mlxsw_sp->base_mac), 0);
11192         mlxsw_reg_recr2_pack(recr2_pl, seed);
11193         mlxsw_sp_mp4_hash_init(mlxsw_sp, &config);
11194         mlxsw_sp_mp6_hash_init(mlxsw_sp, &config);
11195
11196         old_inc_parsing_depth = mlxsw_sp->router->inc_parsing_depth;
11197         new_inc_parsing_depth = config.inc_parsing_depth;
11198         err = mlxsw_sp_mp_hash_parsing_depth_adjust(mlxsw_sp,
11199                                                     old_inc_parsing_depth,
11200                                                     new_inc_parsing_depth);
11201         if (err)
11202                 return err;
11203
11204         for_each_set_bit(bit, config.headers, __MLXSW_REG_RECR2_HEADER_CNT)
11205                 mlxsw_reg_recr2_outer_header_enables_set(recr2_pl, bit, 1);
11206         for_each_set_bit(bit, config.fields, __MLXSW_REG_RECR2_FIELD_CNT)
11207                 mlxsw_reg_recr2_outer_header_fields_enable_set(recr2_pl, bit, 1);
11208         for_each_set_bit(bit, config.inner_headers, __MLXSW_REG_RECR2_HEADER_CNT)
11209                 mlxsw_reg_recr2_inner_header_enables_set(recr2_pl, bit, 1);
11210         for_each_set_bit(bit, config.inner_fields, __MLXSW_REG_RECR2_INNER_FIELD_CNT)
11211                 mlxsw_reg_recr2_inner_header_fields_enable_set(recr2_pl, bit, 1);
11212
11213         err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(recr2), recr2_pl);
11214         if (err)
11215                 goto err_reg_write;
11216
11217         return 0;
11218
11219 err_reg_write:
11220         mlxsw_sp_mp_hash_parsing_depth_adjust(mlxsw_sp, new_inc_parsing_depth,
11221                                               old_inc_parsing_depth);
11222         return err;
11223 }
11224
11225 static void mlxsw_sp_mp_hash_fini(struct mlxsw_sp *mlxsw_sp)
11226 {
11227         bool old_inc_parsing_depth = mlxsw_sp->router->inc_parsing_depth;
11228
11229         mlxsw_sp_mp_hash_parsing_depth_adjust(mlxsw_sp, old_inc_parsing_depth,
11230                                               false);
11231 }
11232 #else
11233 static int mlxsw_sp_mp_hash_init(struct mlxsw_sp *mlxsw_sp)
11234 {
11235         return 0;
11236 }
11237
11238 static void mlxsw_sp_mp_hash_fini(struct mlxsw_sp *mlxsw_sp)
11239 {
11240 }
11241 #endif
11242
11243 static int mlxsw_sp_dscp_init(struct mlxsw_sp *mlxsw_sp)
11244 {
11245         char rdpm_pl[MLXSW_REG_RDPM_LEN];
11246         unsigned int i;
11247
11248         MLXSW_REG_ZERO(rdpm, rdpm_pl);
11249
11250         /* HW is determining switch priority based on DSCP-bits, but the
11251          * kernel is still doing that based on the ToS. Since there's a
11252          * mismatch in bits we need to make sure to translate the right
11253          * value ToS would observe, skipping the 2 least-significant ECN bits.
11254          */
11255         for (i = 0; i < MLXSW_REG_RDPM_DSCP_ENTRY_REC_MAX_COUNT; i++)
11256                 mlxsw_reg_rdpm_pack(rdpm_pl, i, rt_tos2priority(i << 2));
11257
11258         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rdpm), rdpm_pl);
11259 }
11260
11261 static int __mlxsw_sp_router_init(struct mlxsw_sp *mlxsw_sp)
11262 {
11263         struct net *net = mlxsw_sp_net(mlxsw_sp);
11264         char rgcr_pl[MLXSW_REG_RGCR_LEN];
11265         u64 max_rifs;
11266         bool usp;
11267
11268         if (!MLXSW_CORE_RES_VALID(mlxsw_sp->core, MAX_RIFS))
11269                 return -EIO;
11270         max_rifs = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_RIFS);
11271         usp = READ_ONCE(net->ipv4.sysctl_ip_fwd_update_priority);
11272
11273         mlxsw_reg_rgcr_pack(rgcr_pl, true, true);
11274         mlxsw_reg_rgcr_max_router_interfaces_set(rgcr_pl, max_rifs);
11275         mlxsw_reg_rgcr_usp_set(rgcr_pl, usp);
11276         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rgcr), rgcr_pl);
11277 }
11278
11279 static void __mlxsw_sp_router_fini(struct mlxsw_sp *mlxsw_sp)
11280 {
11281         char rgcr_pl[MLXSW_REG_RGCR_LEN];
11282
11283         mlxsw_reg_rgcr_pack(rgcr_pl, false, false);
11284         mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rgcr), rgcr_pl);
11285 }
11286
11287 static int mlxsw_sp_lb_rif_init(struct mlxsw_sp *mlxsw_sp,
11288                                 struct netlink_ext_ack *extack)
11289 {
11290         struct mlxsw_sp_router *router = mlxsw_sp->router;
11291         struct mlxsw_sp_rif *lb_rif;
11292         int err;
11293
11294         router->lb_crif = mlxsw_sp_crif_alloc(NULL);
11295         if (!router->lb_crif)
11296                 return -ENOMEM;
11297
11298         /* Create a generic loopback RIF associated with the main table
11299          * (default VRF). Any table can be used, but the main table exists
11300          * anyway, so we do not waste resources. Loopback RIFs are usually
11301          * created with a NULL CRIF, but this RIF is used as a fallback RIF
11302          * for blackhole nexthops, and nexthops expect to have a valid CRIF.
11303          */
11304         lb_rif = mlxsw_sp_ul_rif_get(mlxsw_sp, RT_TABLE_MAIN, router->lb_crif,
11305                                      extack);
11306         if (IS_ERR(lb_rif)) {
11307                 err = PTR_ERR(lb_rif);
11308                 goto err_ul_rif_get;
11309         }
11310
11311         return 0;
11312
11313 err_ul_rif_get:
11314         mlxsw_sp_crif_free(router->lb_crif);
11315         return err;
11316 }
11317
11318 static void mlxsw_sp_lb_rif_fini(struct mlxsw_sp *mlxsw_sp)
11319 {
11320         mlxsw_sp_ul_rif_put(mlxsw_sp->router->lb_crif->rif);
11321         mlxsw_sp_crif_free(mlxsw_sp->router->lb_crif);
11322 }
11323
11324 static int mlxsw_sp1_router_init(struct mlxsw_sp *mlxsw_sp)
11325 {
11326         size_t size_ranges_count = ARRAY_SIZE(mlxsw_sp1_adj_grp_size_ranges);
11327
11328         mlxsw_sp->router->rif_ops_arr = mlxsw_sp1_rif_ops_arr;
11329         mlxsw_sp->router->adj_grp_size_ranges = mlxsw_sp1_adj_grp_size_ranges;
11330         mlxsw_sp->router->adj_grp_size_ranges_count = size_ranges_count;
11331
11332         return 0;
11333 }
11334
11335 const struct mlxsw_sp_router_ops mlxsw_sp1_router_ops = {
11336         .init = mlxsw_sp1_router_init,
11337         .ipips_init = mlxsw_sp1_ipips_init,
11338 };
11339
11340 static int mlxsw_sp2_router_init(struct mlxsw_sp *mlxsw_sp)
11341 {
11342         size_t size_ranges_count = ARRAY_SIZE(mlxsw_sp2_adj_grp_size_ranges);
11343
11344         mlxsw_sp->router->rif_ops_arr = mlxsw_sp2_rif_ops_arr;
11345         mlxsw_sp->router->adj_grp_size_ranges = mlxsw_sp2_adj_grp_size_ranges;
11346         mlxsw_sp->router->adj_grp_size_ranges_count = size_ranges_count;
11347
11348         return 0;
11349 }
11350
11351 const struct mlxsw_sp_router_ops mlxsw_sp2_router_ops = {
11352         .init = mlxsw_sp2_router_init,
11353         .ipips_init = mlxsw_sp2_ipips_init,
11354 };
11355
11356 int mlxsw_sp_router_init(struct mlxsw_sp *mlxsw_sp,
11357                          struct netlink_ext_ack *extack)
11358 {
11359         struct mlxsw_sp_router *router;
11360         struct notifier_block *nb;
11361         int err;
11362
11363         router = kzalloc(sizeof(*mlxsw_sp->router), GFP_KERNEL);
11364         if (!router)
11365                 return -ENOMEM;
11366         mutex_init(&router->lock);
11367         mlxsw_sp->router = router;
11368         router->mlxsw_sp = mlxsw_sp;
11369
11370         err = mlxsw_sp->router_ops->init(mlxsw_sp);
11371         if (err)
11372                 goto err_router_ops_init;
11373
11374         INIT_LIST_HEAD(&mlxsw_sp->router->nh_res_grp_list);
11375         INIT_DELAYED_WORK(&mlxsw_sp->router->nh_grp_activity_dw,
11376                           mlxsw_sp_nh_grp_activity_work);
11377         INIT_LIST_HEAD(&mlxsw_sp->router->nexthop_neighs_list);
11378         err = __mlxsw_sp_router_init(mlxsw_sp);
11379         if (err)
11380                 goto err_router_init;
11381
11382         err = mlxsw_sp->router_ops->ipips_init(mlxsw_sp);
11383         if (err)
11384                 goto err_ipips_init;
11385
11386         err = rhashtable_init(&mlxsw_sp->router->crif_ht,
11387                               &mlxsw_sp_crif_ht_params);
11388         if (err)
11389                 goto err_crif_ht_init;
11390
11391         err = mlxsw_sp_rifs_init(mlxsw_sp);
11392         if (err)
11393                 goto err_rifs_init;
11394
11395         err = rhashtable_init(&mlxsw_sp->router->nexthop_ht,
11396                               &mlxsw_sp_nexthop_ht_params);
11397         if (err)
11398                 goto err_nexthop_ht_init;
11399
11400         err = rhashtable_init(&mlxsw_sp->router->nexthop_group_ht,
11401                               &mlxsw_sp_nexthop_group_ht_params);
11402         if (err)
11403                 goto err_nexthop_group_ht_init;
11404
11405         INIT_LIST_HEAD(&mlxsw_sp->router->nexthop_list);
11406         err = mlxsw_sp_lpm_init(mlxsw_sp);
11407         if (err)
11408                 goto err_lpm_init;
11409
11410         err = mlxsw_sp_mr_init(mlxsw_sp, &mlxsw_sp_mr_tcam_ops);
11411         if (err)
11412                 goto err_mr_init;
11413
11414         err = mlxsw_sp_vrs_init(mlxsw_sp);
11415         if (err)
11416                 goto err_vrs_init;
11417
11418         err = mlxsw_sp_lb_rif_init(mlxsw_sp, extack);
11419         if (err)
11420                 goto err_lb_rif_init;
11421
11422         err = mlxsw_sp_neigh_init(mlxsw_sp);
11423         if (err)
11424                 goto err_neigh_init;
11425
11426         err = mlxsw_sp_mp_hash_init(mlxsw_sp);
11427         if (err)
11428                 goto err_mp_hash_init;
11429
11430         err = mlxsw_sp_dscp_init(mlxsw_sp);
11431         if (err)
11432                 goto err_dscp_init;
11433
11434         router->inetaddr_nb.notifier_call = mlxsw_sp_inetaddr_event;
11435         err = register_inetaddr_notifier(&router->inetaddr_nb);
11436         if (err)
11437                 goto err_register_inetaddr_notifier;
11438
11439         router->inet6addr_nb.notifier_call = mlxsw_sp_inet6addr_event;
11440         err = register_inet6addr_notifier(&router->inet6addr_nb);
11441         if (err)
11442                 goto err_register_inet6addr_notifier;
11443
11444         router->inetaddr_valid_nb.notifier_call = mlxsw_sp_inetaddr_valid_event;
11445         err = register_inetaddr_validator_notifier(&router->inetaddr_valid_nb);
11446         if (err)
11447                 goto err_register_inetaddr_valid_notifier;
11448
11449         nb = &router->inet6addr_valid_nb;
11450         nb->notifier_call = mlxsw_sp_inet6addr_valid_event;
11451         err = register_inet6addr_validator_notifier(nb);
11452         if (err)
11453                 goto err_register_inet6addr_valid_notifier;
11454
11455         mlxsw_sp->router->netevent_nb.notifier_call =
11456                 mlxsw_sp_router_netevent_event;
11457         err = register_netevent_notifier(&mlxsw_sp->router->netevent_nb);
11458         if (err)
11459                 goto err_register_netevent_notifier;
11460
11461         mlxsw_sp->router->nexthop_nb.notifier_call =
11462                 mlxsw_sp_nexthop_obj_event;
11463         err = register_nexthop_notifier(mlxsw_sp_net(mlxsw_sp),
11464                                         &mlxsw_sp->router->nexthop_nb,
11465                                         extack);
11466         if (err)
11467                 goto err_register_nexthop_notifier;
11468
11469         mlxsw_sp->router->fib_nb.notifier_call = mlxsw_sp_router_fib_event;
11470         err = register_fib_notifier(mlxsw_sp_net(mlxsw_sp),
11471                                     &mlxsw_sp->router->fib_nb,
11472                                     mlxsw_sp_router_fib_dump_flush, extack);
11473         if (err)
11474                 goto err_register_fib_notifier;
11475
11476         mlxsw_sp->router->netdevice_nb.notifier_call =
11477                 mlxsw_sp_router_netdevice_event;
11478         err = register_netdevice_notifier_net(mlxsw_sp_net(mlxsw_sp),
11479                                               &mlxsw_sp->router->netdevice_nb);
11480         if (err)
11481                 goto err_register_netdev_notifier;
11482
11483         return 0;
11484
11485 err_register_netdev_notifier:
11486         unregister_fib_notifier(mlxsw_sp_net(mlxsw_sp),
11487                                 &mlxsw_sp->router->fib_nb);
11488 err_register_fib_notifier:
11489         unregister_nexthop_notifier(mlxsw_sp_net(mlxsw_sp),
11490                                     &mlxsw_sp->router->nexthop_nb);
11491 err_register_nexthop_notifier:
11492         unregister_netevent_notifier(&mlxsw_sp->router->netevent_nb);
11493 err_register_netevent_notifier:
11494         unregister_inet6addr_validator_notifier(&router->inet6addr_valid_nb);
11495 err_register_inet6addr_valid_notifier:
11496         unregister_inetaddr_validator_notifier(&router->inetaddr_valid_nb);
11497 err_register_inetaddr_valid_notifier:
11498         unregister_inet6addr_notifier(&router->inet6addr_nb);
11499 err_register_inet6addr_notifier:
11500         unregister_inetaddr_notifier(&router->inetaddr_nb);
11501 err_register_inetaddr_notifier:
11502         mlxsw_core_flush_owq();
11503 err_dscp_init:
11504         mlxsw_sp_mp_hash_fini(mlxsw_sp);
11505 err_mp_hash_init:
11506         mlxsw_sp_neigh_fini(mlxsw_sp);
11507 err_neigh_init:
11508         mlxsw_sp_lb_rif_fini(mlxsw_sp);
11509 err_lb_rif_init:
11510         mlxsw_sp_vrs_fini(mlxsw_sp);
11511 err_vrs_init:
11512         mlxsw_sp_mr_fini(mlxsw_sp);
11513 err_mr_init:
11514         mlxsw_sp_lpm_fini(mlxsw_sp);
11515 err_lpm_init:
11516         rhashtable_destroy(&mlxsw_sp->router->nexthop_group_ht);
11517 err_nexthop_group_ht_init:
11518         rhashtable_destroy(&mlxsw_sp->router->nexthop_ht);
11519 err_nexthop_ht_init:
11520         mlxsw_sp_rifs_fini(mlxsw_sp);
11521 err_rifs_init:
11522         rhashtable_destroy(&mlxsw_sp->router->crif_ht);
11523 err_crif_ht_init:
11524         mlxsw_sp_ipips_fini(mlxsw_sp);
11525 err_ipips_init:
11526         __mlxsw_sp_router_fini(mlxsw_sp);
11527 err_router_init:
11528         cancel_delayed_work_sync(&mlxsw_sp->router->nh_grp_activity_dw);
11529 err_router_ops_init:
11530         mutex_destroy(&mlxsw_sp->router->lock);
11531         kfree(mlxsw_sp->router);
11532         return err;
11533 }
11534
11535 void mlxsw_sp_router_fini(struct mlxsw_sp *mlxsw_sp)
11536 {
11537         struct mlxsw_sp_router *router = mlxsw_sp->router;
11538
11539         unregister_netdevice_notifier_net(mlxsw_sp_net(mlxsw_sp),
11540                                           &router->netdevice_nb);
11541         unregister_fib_notifier(mlxsw_sp_net(mlxsw_sp), &router->fib_nb);
11542         unregister_nexthop_notifier(mlxsw_sp_net(mlxsw_sp),
11543                                     &router->nexthop_nb);
11544         unregister_netevent_notifier(&router->netevent_nb);
11545         unregister_inet6addr_validator_notifier(&router->inet6addr_valid_nb);
11546         unregister_inetaddr_validator_notifier(&router->inetaddr_valid_nb);
11547         unregister_inet6addr_notifier(&router->inet6addr_nb);
11548         unregister_inetaddr_notifier(&router->inetaddr_nb);
11549         mlxsw_core_flush_owq();
11550         mlxsw_sp_mp_hash_fini(mlxsw_sp);
11551         mlxsw_sp_neigh_fini(mlxsw_sp);
11552         mlxsw_sp_lb_rif_fini(mlxsw_sp);
11553         mlxsw_sp_vrs_fini(mlxsw_sp);
11554         mlxsw_sp_mr_fini(mlxsw_sp);
11555         mlxsw_sp_lpm_fini(mlxsw_sp);
11556         rhashtable_destroy(&router->nexthop_group_ht);
11557         rhashtable_destroy(&router->nexthop_ht);
11558         mlxsw_sp_rifs_fini(mlxsw_sp);
11559         rhashtable_destroy(&mlxsw_sp->router->crif_ht);
11560         mlxsw_sp_ipips_fini(mlxsw_sp);
11561         __mlxsw_sp_router_fini(mlxsw_sp);
11562         cancel_delayed_work_sync(&router->nh_grp_activity_dw);
11563         mutex_destroy(&router->lock);
11564         kfree(router);
11565 }