spectrum_router.c 96.4 KB
Newer Older
1 2 3 4 5
/*
 * drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
 * Copyright (c) 2016 Mellanox Technologies. All rights reserved.
 * Copyright (c) 2016 Jiri Pirko <jiri@mellanox.com>
 * Copyright (c) 2016 Ido Schimmel <idosch@mellanox.com>
6
 * Copyright (c) 2016 Yotam Gigi <yotamg@mellanox.com>
7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38
 *
 * Redistribution and use in source and binary forms, with or without
 * modification, are permitted provided that the following conditions are met:
 *
 * 1. Redistributions of source code must retain the above copyright
 *    notice, this list of conditions and the following disclaimer.
 * 2. Redistributions in binary form must reproduce the above copyright
 *    notice, this list of conditions and the following disclaimer in the
 *    documentation and/or other materials provided with the distribution.
 * 3. Neither the names of the copyright holders nor the names of its
 *    contributors may be used to endorse or promote products derived from
 *    this software without specific prior written permission.
 *
 * Alternatively, this software may be distributed under the terms of the
 * GNU General Public License ("GPL") version 2 as published by the Free
 * Software Foundation.
 *
 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
 * POSSIBILITY OF SUCH DAMAGE.
 */

#include <linux/kernel.h>
#include <linux/types.h>
39 40 41
#include <linux/rhashtable.h>
#include <linux/bitops.h>
#include <linux/in6.h>
42
#include <linux/notifier.h>
43
#include <linux/inetdevice.h>
44
#include <linux/netdevice.h>
45
#include <net/netevent.h>
46 47
#include <net/neighbour.h>
#include <net/arp.h>
48
#include <net/ip_fib.h>
49
#include <net/fib_rules.h>
50
#include <net/l3mdev.h>
51 52 53 54

#include "spectrum.h"
#include "core.h"
#include "reg.h"
55 56 57
#include "spectrum_cnt.h"
#include "spectrum_dpipe.h"
#include "spectrum_router.h"
58

59 60 61 62 63
struct mlxsw_sp_vr;
struct mlxsw_sp_lpm_tree;

struct mlxsw_sp_router {
	struct mlxsw_sp *mlxsw_sp;
64
	struct mlxsw_sp_rif **rifs;
65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82
	struct mlxsw_sp_vr *vrs;
	struct rhashtable neigh_ht;
	struct rhashtable nexthop_group_ht;
	struct rhashtable nexthop_ht;
	struct {
		struct mlxsw_sp_lpm_tree *trees;
		unsigned int tree_count;
	} lpm;
	struct {
		struct delayed_work dw;
		unsigned long interval;	/* ms */
	} neighs_update;
	struct delayed_work nexthop_probe_dw;
#define MLXSW_SP_UNRESOLVED_NH_PROBE_INTERVAL 5000 /* ms */
	struct list_head nexthop_neighs_list;
	bool aborted;
};

83 84 85 86 87 88 89
struct mlxsw_sp_rif {
	struct list_head nexthop_list;
	struct list_head neigh_list;
	struct net_device *dev;
	struct mlxsw_sp_fid *f;
	unsigned char addr[ETH_ALEN];
	int mtu;
90
	u16 rif_index;
91
	u16 vr_id;
92 93 94 95
	unsigned int counter_ingress;
	bool counter_ingress_valid;
	unsigned int counter_egress;
	bool counter_egress_valid;
96 97
};

98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242
static unsigned int *
mlxsw_sp_rif_p_counter_get(struct mlxsw_sp_rif *rif,
			   enum mlxsw_sp_rif_counter_dir dir)
{
	switch (dir) {
	case MLXSW_SP_RIF_COUNTER_EGRESS:
		return &rif->counter_egress;
	case MLXSW_SP_RIF_COUNTER_INGRESS:
		return &rif->counter_ingress;
	}
	return NULL;
}

static bool
mlxsw_sp_rif_counter_valid_get(struct mlxsw_sp_rif *rif,
			       enum mlxsw_sp_rif_counter_dir dir)
{
	switch (dir) {
	case MLXSW_SP_RIF_COUNTER_EGRESS:
		return rif->counter_egress_valid;
	case MLXSW_SP_RIF_COUNTER_INGRESS:
		return rif->counter_ingress_valid;
	}
	return false;
}

static void
mlxsw_sp_rif_counter_valid_set(struct mlxsw_sp_rif *rif,
			       enum mlxsw_sp_rif_counter_dir dir,
			       bool valid)
{
	switch (dir) {
	case MLXSW_SP_RIF_COUNTER_EGRESS:
		rif->counter_egress_valid = valid;
		break;
	case MLXSW_SP_RIF_COUNTER_INGRESS:
		rif->counter_ingress_valid = valid;
		break;
	}
}

static int mlxsw_sp_rif_counter_edit(struct mlxsw_sp *mlxsw_sp, u16 rif_index,
				     unsigned int counter_index, bool enable,
				     enum mlxsw_sp_rif_counter_dir dir)
{
	char ritr_pl[MLXSW_REG_RITR_LEN];
	bool is_egress = false;
	int err;

	if (dir == MLXSW_SP_RIF_COUNTER_EGRESS)
		is_egress = true;
	mlxsw_reg_ritr_rif_pack(ritr_pl, rif_index);
	err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
	if (err)
		return err;

	mlxsw_reg_ritr_counter_pack(ritr_pl, counter_index, enable,
				    is_egress);
	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
}

int mlxsw_sp_rif_counter_value_get(struct mlxsw_sp *mlxsw_sp,
				   struct mlxsw_sp_rif *rif,
				   enum mlxsw_sp_rif_counter_dir dir, u64 *cnt)
{
	char ricnt_pl[MLXSW_REG_RICNT_LEN];
	unsigned int *p_counter_index;
	bool valid;
	int err;

	valid = mlxsw_sp_rif_counter_valid_get(rif, dir);
	if (!valid)
		return -EINVAL;

	p_counter_index = mlxsw_sp_rif_p_counter_get(rif, dir);
	if (!p_counter_index)
		return -EINVAL;
	mlxsw_reg_ricnt_pack(ricnt_pl, *p_counter_index,
			     MLXSW_REG_RICNT_OPCODE_NOP);
	err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(ricnt), ricnt_pl);
	if (err)
		return err;
	*cnt = mlxsw_reg_ricnt_good_unicast_packets_get(ricnt_pl);
	return 0;
}

static int mlxsw_sp_rif_counter_clear(struct mlxsw_sp *mlxsw_sp,
				      unsigned int counter_index)
{
	char ricnt_pl[MLXSW_REG_RICNT_LEN];

	mlxsw_reg_ricnt_pack(ricnt_pl, counter_index,
			     MLXSW_REG_RICNT_OPCODE_CLEAR);
	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ricnt), ricnt_pl);
}

int mlxsw_sp_rif_counter_alloc(struct mlxsw_sp *mlxsw_sp,
			       struct mlxsw_sp_rif *rif,
			       enum mlxsw_sp_rif_counter_dir dir)
{
	unsigned int *p_counter_index;
	int err;

	p_counter_index = mlxsw_sp_rif_p_counter_get(rif, dir);
	if (!p_counter_index)
		return -EINVAL;
	err = mlxsw_sp_counter_alloc(mlxsw_sp, MLXSW_SP_COUNTER_SUB_POOL_RIF,
				     p_counter_index);
	if (err)
		return err;

	err = mlxsw_sp_rif_counter_clear(mlxsw_sp, *p_counter_index);
	if (err)
		goto err_counter_clear;

	err = mlxsw_sp_rif_counter_edit(mlxsw_sp, rif->rif_index,
					*p_counter_index, true, dir);
	if (err)
		goto err_counter_edit;
	mlxsw_sp_rif_counter_valid_set(rif, dir, true);
	return 0;

err_counter_edit:
err_counter_clear:
	mlxsw_sp_counter_free(mlxsw_sp, MLXSW_SP_COUNTER_SUB_POOL_RIF,
			      *p_counter_index);
	return err;
}

void mlxsw_sp_rif_counter_free(struct mlxsw_sp *mlxsw_sp,
			       struct mlxsw_sp_rif *rif,
			       enum mlxsw_sp_rif_counter_dir dir)
{
	unsigned int *p_counter_index;

	p_counter_index = mlxsw_sp_rif_p_counter_get(rif, dir);
	if (WARN_ON(!p_counter_index))
		return;
	mlxsw_sp_rif_counter_edit(mlxsw_sp, rif->rif_index,
				  *p_counter_index, false, dir);
	mlxsw_sp_counter_free(mlxsw_sp, MLXSW_SP_COUNTER_SUB_POOL_RIF,
			      *p_counter_index);
	mlxsw_sp_rif_counter_valid_set(rif, dir, false);
}

243 244 245 246
static struct mlxsw_sp_rif *
mlxsw_sp_rif_find_by_dev(const struct mlxsw_sp *mlxsw_sp,
			 const struct net_device *dev);

247 248 249 250 251 252
#define MLXSW_SP_PREFIX_COUNT (sizeof(struct in6_addr) * BITS_PER_BYTE)

struct mlxsw_sp_prefix_usage {
	DECLARE_BITMAP(b, MLXSW_SP_PREFIX_COUNT);
};

253 254 255
#define mlxsw_sp_prefix_usage_for_each(prefix, prefix_usage) \
	for_each_set_bit(prefix, (prefix_usage)->b, MLXSW_SP_PREFIX_COUNT)

256 257 258 259 260 261 262 263 264 265 266 267 268
static bool
mlxsw_sp_prefix_usage_subset(struct mlxsw_sp_prefix_usage *prefix_usage1,
			     struct mlxsw_sp_prefix_usage *prefix_usage2)
{
	unsigned char prefix;

	mlxsw_sp_prefix_usage_for_each(prefix, prefix_usage1) {
		if (!test_bit(prefix, prefix_usage2->b))
			return false;
	}
	return true;
}

269 270 271 272 273 274 275
static bool
mlxsw_sp_prefix_usage_eq(struct mlxsw_sp_prefix_usage *prefix_usage1,
			 struct mlxsw_sp_prefix_usage *prefix_usage2)
{
	return !memcmp(prefix_usage1, prefix_usage2, sizeof(*prefix_usage1));
}

276 277 278 279 280 281 282 283 284 285 286 287 288 289 290
static bool
mlxsw_sp_prefix_usage_none(struct mlxsw_sp_prefix_usage *prefix_usage)
{
	struct mlxsw_sp_prefix_usage prefix_usage_none = {{ 0 } };

	return mlxsw_sp_prefix_usage_eq(prefix_usage, &prefix_usage_none);
}

static void
mlxsw_sp_prefix_usage_cpy(struct mlxsw_sp_prefix_usage *prefix_usage1,
			  struct mlxsw_sp_prefix_usage *prefix_usage2)
{
	memcpy(prefix_usage1, prefix_usage2, sizeof(*prefix_usage1));
}

291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309
static void
mlxsw_sp_prefix_usage_set(struct mlxsw_sp_prefix_usage *prefix_usage,
			  unsigned char prefix_len)
{
	set_bit(prefix_len, prefix_usage->b);
}

static void
mlxsw_sp_prefix_usage_clear(struct mlxsw_sp_prefix_usage *prefix_usage,
			    unsigned char prefix_len)
{
	clear_bit(prefix_len, prefix_usage->b);
}

struct mlxsw_sp_fib_key {
	unsigned char addr[sizeof(struct in6_addr)];
	unsigned char prefix_len;
};

310 311 312 313 314 315
enum mlxsw_sp_fib_entry_type {
	MLXSW_SP_FIB_ENTRY_TYPE_REMOTE,
	MLXSW_SP_FIB_ENTRY_TYPE_LOCAL,
	MLXSW_SP_FIB_ENTRY_TYPE_TRAP,
};

316
struct mlxsw_sp_nexthop_group;
317
struct mlxsw_sp_fib;
318

319 320
struct mlxsw_sp_fib_node {
	struct list_head entry_list;
321
	struct list_head list;
322
	struct rhash_head ht_node;
323
	struct mlxsw_sp_fib *fib;
324
	struct mlxsw_sp_fib_key key;
325 326 327 328 329 330 331 332 333 334 335 336
};

struct mlxsw_sp_fib_entry_params {
	u32 tb_id;
	u32 prio;
	u8 tos;
	u8 type;
};

struct mlxsw_sp_fib_entry {
	struct list_head list;
	struct mlxsw_sp_fib_node *fib_node;
337
	enum mlxsw_sp_fib_entry_type type;
338 339
	struct list_head nexthop_group_node;
	struct mlxsw_sp_nexthop_group *nh_group;
340
	struct mlxsw_sp_fib_entry_params params;
341
	bool offloaded;
342 343
};

344 345 346 347 348 349 350 351 352 353 354 355
enum mlxsw_sp_l3proto {
	MLXSW_SP_L3_PROTO_IPV4,
	MLXSW_SP_L3_PROTO_IPV6,
};

struct mlxsw_sp_lpm_tree {
	u8 id; /* tree ID */
	unsigned int ref_count;
	enum mlxsw_sp_l3proto proto;
	struct mlxsw_sp_prefix_usage prefix_usage;
};

356 357
struct mlxsw_sp_fib {
	struct rhashtable ht;
358
	struct list_head node_list;
359 360
	struct mlxsw_sp_vr *vr;
	struct mlxsw_sp_lpm_tree *lpm_tree;
361 362
	unsigned long prefix_ref_count[MLXSW_SP_PREFIX_COUNT];
	struct mlxsw_sp_prefix_usage prefix_usage;
363
	enum mlxsw_sp_l3proto proto;
364 365
};

366 367 368 369 370 371 372
struct mlxsw_sp_vr {
	u16 id; /* virtual router ID */
	u32 tb_id; /* kernel fib table id */
	unsigned int rif_count;
	struct mlxsw_sp_fib *fib4;
};

373
static const struct rhashtable_params mlxsw_sp_fib_ht_params;
374

375 376
static struct mlxsw_sp_fib *mlxsw_sp_fib_create(struct mlxsw_sp_vr *vr,
						enum mlxsw_sp_l3proto proto)
377 378 379 380 381 382 383 384 385 386
{
	struct mlxsw_sp_fib *fib;
	int err;

	fib = kzalloc(sizeof(*fib), GFP_KERNEL);
	if (!fib)
		return ERR_PTR(-ENOMEM);
	err = rhashtable_init(&fib->ht, &mlxsw_sp_fib_ht_params);
	if (err)
		goto err_rhashtable_init;
387
	INIT_LIST_HEAD(&fib->node_list);
388 389
	fib->proto = proto;
	fib->vr = vr;
390 391 392 393 394 395 396 397 398
	return fib;

err_rhashtable_init:
	kfree(fib);
	return ERR_PTR(err);
}

static void mlxsw_sp_fib_destroy(struct mlxsw_sp_fib *fib)
{
399
	WARN_ON(!list_empty(&fib->node_list));
400
	WARN_ON(fib->lpm_tree);
401 402 403 404
	rhashtable_destroy(&fib->ht);
	kfree(fib);
}

405
static struct mlxsw_sp_lpm_tree *
406
mlxsw_sp_lpm_tree_find_unused(struct mlxsw_sp *mlxsw_sp)
407 408 409 410
{
	static struct mlxsw_sp_lpm_tree *lpm_tree;
	int i;

411 412
	for (i = 0; i < mlxsw_sp->router->lpm.tree_count; i++) {
		lpm_tree = &mlxsw_sp->router->lpm.trees[i];
413 414
		if (lpm_tree->ref_count == 0)
			return lpm_tree;
415 416 417 418 419 420 421 422 423
	}
	return NULL;
}

static int mlxsw_sp_lpm_tree_alloc(struct mlxsw_sp *mlxsw_sp,
				   struct mlxsw_sp_lpm_tree *lpm_tree)
{
	char ralta_pl[MLXSW_REG_RALTA_LEN];

424 425 426
	mlxsw_reg_ralta_pack(ralta_pl, true,
			     (enum mlxsw_reg_ralxx_protocol) lpm_tree->proto,
			     lpm_tree->id);
427 428 429 430 431 432 433 434
	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralta), ralta_pl);
}

static int mlxsw_sp_lpm_tree_free(struct mlxsw_sp *mlxsw_sp,
				  struct mlxsw_sp_lpm_tree *lpm_tree)
{
	char ralta_pl[MLXSW_REG_RALTA_LEN];

435 436 437
	mlxsw_reg_ralta_pack(ralta_pl, false,
			     (enum mlxsw_reg_ralxx_protocol) lpm_tree->proto,
			     lpm_tree->id);
438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467
	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralta), ralta_pl);
}

static int
mlxsw_sp_lpm_tree_left_struct_set(struct mlxsw_sp *mlxsw_sp,
				  struct mlxsw_sp_prefix_usage *prefix_usage,
				  struct mlxsw_sp_lpm_tree *lpm_tree)
{
	char ralst_pl[MLXSW_REG_RALST_LEN];
	u8 root_bin = 0;
	u8 prefix;
	u8 last_prefix = MLXSW_REG_RALST_BIN_NO_CHILD;

	mlxsw_sp_prefix_usage_for_each(prefix, prefix_usage)
		root_bin = prefix;

	mlxsw_reg_ralst_pack(ralst_pl, root_bin, lpm_tree->id);
	mlxsw_sp_prefix_usage_for_each(prefix, prefix_usage) {
		if (prefix == 0)
			continue;
		mlxsw_reg_ralst_bin_pack(ralst_pl, prefix, last_prefix,
					 MLXSW_REG_RALST_BIN_NO_CHILD);
		last_prefix = prefix;
	}
	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralst), ralst_pl);
}

static struct mlxsw_sp_lpm_tree *
mlxsw_sp_lpm_tree_create(struct mlxsw_sp *mlxsw_sp,
			 struct mlxsw_sp_prefix_usage *prefix_usage,
468
			 enum mlxsw_sp_l3proto proto)
469 470 471 472
{
	struct mlxsw_sp_lpm_tree *lpm_tree;
	int err;

473
	lpm_tree = mlxsw_sp_lpm_tree_find_unused(mlxsw_sp);
474 475 476 477 478 479 480 481 482 483 484
	if (!lpm_tree)
		return ERR_PTR(-EBUSY);
	lpm_tree->proto = proto;
	err = mlxsw_sp_lpm_tree_alloc(mlxsw_sp, lpm_tree);
	if (err)
		return ERR_PTR(err);

	err = mlxsw_sp_lpm_tree_left_struct_set(mlxsw_sp, prefix_usage,
						lpm_tree);
	if (err)
		goto err_left_struct_set;
485 486
	memcpy(&lpm_tree->prefix_usage, prefix_usage,
	       sizeof(lpm_tree->prefix_usage));
487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502
	return lpm_tree;

err_left_struct_set:
	mlxsw_sp_lpm_tree_free(mlxsw_sp, lpm_tree);
	return ERR_PTR(err);
}

static int mlxsw_sp_lpm_tree_destroy(struct mlxsw_sp *mlxsw_sp,
				     struct mlxsw_sp_lpm_tree *lpm_tree)
{
	return mlxsw_sp_lpm_tree_free(mlxsw_sp, lpm_tree);
}

static struct mlxsw_sp_lpm_tree *
mlxsw_sp_lpm_tree_get(struct mlxsw_sp *mlxsw_sp,
		      struct mlxsw_sp_prefix_usage *prefix_usage,
503
		      enum mlxsw_sp_l3proto proto)
504 505 506 507
{
	struct mlxsw_sp_lpm_tree *lpm_tree;
	int i;

508 509
	for (i = 0; i < mlxsw_sp->router->lpm.tree_count; i++) {
		lpm_tree = &mlxsw_sp->router->lpm.trees[i];
510 511
		if (lpm_tree->ref_count != 0 &&
		    lpm_tree->proto == proto &&
512 513 514 515 516
		    mlxsw_sp_prefix_usage_eq(&lpm_tree->prefix_usage,
					     prefix_usage))
			goto inc_ref_count;
	}
	lpm_tree = mlxsw_sp_lpm_tree_create(mlxsw_sp, prefix_usage,
517
					    proto);
518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533
	if (IS_ERR(lpm_tree))
		return lpm_tree;

inc_ref_count:
	lpm_tree->ref_count++;
	return lpm_tree;
}

static int mlxsw_sp_lpm_tree_put(struct mlxsw_sp *mlxsw_sp,
				 struct mlxsw_sp_lpm_tree *lpm_tree)
{
	if (--lpm_tree->ref_count == 0)
		return mlxsw_sp_lpm_tree_destroy(mlxsw_sp, lpm_tree);
	return 0;
}

534 535 536
#define MLXSW_SP_LPM_TREE_MIN 2 /* trees 0 and 1 are reserved */

static int mlxsw_sp_lpm_init(struct mlxsw_sp *mlxsw_sp)
537 538
{
	struct mlxsw_sp_lpm_tree *lpm_tree;
539
	u64 max_trees;
540 541
	int i;

542 543 544 545
	if (!MLXSW_CORE_RES_VALID(mlxsw_sp->core, MAX_LPM_TREES))
		return -EIO;

	max_trees = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_LPM_TREES);
546 547
	mlxsw_sp->router->lpm.tree_count = max_trees - MLXSW_SP_LPM_TREE_MIN;
	mlxsw_sp->router->lpm.trees = kcalloc(mlxsw_sp->router->lpm.tree_count,
548 549
					     sizeof(struct mlxsw_sp_lpm_tree),
					     GFP_KERNEL);
550
	if (!mlxsw_sp->router->lpm.trees)
551 552
		return -ENOMEM;

553 554
	for (i = 0; i < mlxsw_sp->router->lpm.tree_count; i++) {
		lpm_tree = &mlxsw_sp->router->lpm.trees[i];
555 556
		lpm_tree->id = i + MLXSW_SP_LPM_TREE_MIN;
	}
557 558 559 560 561 562

	return 0;
}

static void mlxsw_sp_lpm_fini(struct mlxsw_sp *mlxsw_sp)
{
563
	kfree(mlxsw_sp->router->lpm.trees);
564 565
}

566 567 568 569 570
static bool mlxsw_sp_vr_is_used(const struct mlxsw_sp_vr *vr)
{
	return !!vr->fib4;
}

571 572 573 574 575
static struct mlxsw_sp_vr *mlxsw_sp_vr_find_unused(struct mlxsw_sp *mlxsw_sp)
{
	struct mlxsw_sp_vr *vr;
	int i;

J
Jiri Pirko 已提交
576
	for (i = 0; i < MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_VRS); i++) {
577
		vr = &mlxsw_sp->router->vrs[i];
578
		if (!mlxsw_sp_vr_is_used(vr))
579 580 581 582 583 584
			return vr;
	}
	return NULL;
}

static int mlxsw_sp_vr_lpm_tree_bind(struct mlxsw_sp *mlxsw_sp,
585
				     const struct mlxsw_sp_fib *fib)
586 587 588
{
	char raltb_pl[MLXSW_REG_RALTB_LEN];

589 590 591
	mlxsw_reg_raltb_pack(raltb_pl, fib->vr->id,
			     (enum mlxsw_reg_ralxx_protocol) fib->proto,
			     fib->lpm_tree->id);
592 593 594 595
	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(raltb), raltb_pl);
}

static int mlxsw_sp_vr_lpm_tree_unbind(struct mlxsw_sp *mlxsw_sp,
596
				       const struct mlxsw_sp_fib *fib)
597 598 599 600
{
	char raltb_pl[MLXSW_REG_RALTB_LEN];

	/* Bind to tree 0 which is default */
601 602
	mlxsw_reg_raltb_pack(raltb_pl, fib->vr->id,
			     (enum mlxsw_reg_ralxx_protocol) fib->proto, 0);
603 604 605 606 607 608 609 610 611 612 613 614
	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(raltb), raltb_pl);
}

static u32 mlxsw_sp_fix_tb_id(u32 tb_id)
{
	/* For our purpose, squash main and local table into one */
	if (tb_id == RT_TABLE_LOCAL)
		tb_id = RT_TABLE_MAIN;
	return tb_id;
}

static struct mlxsw_sp_vr *mlxsw_sp_vr_find(struct mlxsw_sp *mlxsw_sp,
615
					    u32 tb_id)
616 617 618 619 620
{
	struct mlxsw_sp_vr *vr;
	int i;

	tb_id = mlxsw_sp_fix_tb_id(tb_id);
621

J
Jiri Pirko 已提交
622
	for (i = 0; i < MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_VRS); i++) {
623
		vr = &mlxsw_sp->router->vrs[i];
624
		if (mlxsw_sp_vr_is_used(vr) && vr->tb_id == tb_id)
625 626 627 628 629
			return vr;
	}
	return NULL;
}

630 631 632 633 634 635 636 637 638 639 640 641
static struct mlxsw_sp_fib *mlxsw_sp_vr_fib(const struct mlxsw_sp_vr *vr,
					    enum mlxsw_sp_l3proto proto)
{
	switch (proto) {
	case MLXSW_SP_L3_PROTO_IPV4:
		return vr->fib4;
	case MLXSW_SP_L3_PROTO_IPV6:
		BUG_ON(1);
	}
	return NULL;
}

642
static struct mlxsw_sp_vr *mlxsw_sp_vr_create(struct mlxsw_sp *mlxsw_sp,
643
					      u32 tb_id)
644 645 646 647 648 649
{
	struct mlxsw_sp_vr *vr;

	vr = mlxsw_sp_vr_find_unused(mlxsw_sp);
	if (!vr)
		return ERR_PTR(-EBUSY);
650 651 652
	vr->fib4 = mlxsw_sp_fib_create(vr, MLXSW_SP_L3_PROTO_IPV4);
	if (IS_ERR(vr->fib4))
		return ERR_CAST(vr->fib4);
653 654 655 656
	vr->tb_id = tb_id;
	return vr;
}

657
static void mlxsw_sp_vr_destroy(struct mlxsw_sp_vr *vr)
658
{
659 660
	mlxsw_sp_fib_destroy(vr->fib4);
	vr->fib4 = NULL;
661 662 663
}

static int
664
mlxsw_sp_vr_lpm_tree_check(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_fib *fib,
665 666
			   struct mlxsw_sp_prefix_usage *req_prefix_usage)
{
667
	struct mlxsw_sp_lpm_tree *lpm_tree = fib->lpm_tree;
668 669
	struct mlxsw_sp_lpm_tree *new_tree;
	int err;
670

671
	if (mlxsw_sp_prefix_usage_eq(req_prefix_usage, &lpm_tree->prefix_usage))
672 673
		return 0;

674
	new_tree = mlxsw_sp_lpm_tree_get(mlxsw_sp, req_prefix_usage,
675
					 fib->proto);
676
	if (IS_ERR(new_tree)) {
677 678 679 680 681 682
		/* We failed to get a tree according to the required
		 * prefix usage. However, the current tree might be still good
		 * for us if our requirement is subset of the prefixes used
		 * in the tree.
		 */
		if (mlxsw_sp_prefix_usage_subset(req_prefix_usage,
683
						 &lpm_tree->prefix_usage))
684
			return 0;
685
		return PTR_ERR(new_tree);
686 687
	}

688
	/* Prevent packet loss by overwriting existing binding */
689 690
	fib->lpm_tree = new_tree;
	err = mlxsw_sp_vr_lpm_tree_bind(mlxsw_sp, fib);
691 692 693 694 695 696 697
	if (err)
		goto err_tree_bind;
	mlxsw_sp_lpm_tree_put(mlxsw_sp, lpm_tree);

	return 0;

err_tree_bind:
698
	fib->lpm_tree = lpm_tree;
699 700
	mlxsw_sp_lpm_tree_put(mlxsw_sp, new_tree);
	return err;
701 702
}

703
static struct mlxsw_sp_vr *mlxsw_sp_vr_get(struct mlxsw_sp *mlxsw_sp, u32 tb_id)
704 705 706 707
{
	struct mlxsw_sp_vr *vr;

	tb_id = mlxsw_sp_fix_tb_id(tb_id);
708 709 710
	vr = mlxsw_sp_vr_find(mlxsw_sp, tb_id);
	if (!vr)
		vr = mlxsw_sp_vr_create(mlxsw_sp, tb_id);
711 712 713
	return vr;
}

714
static void mlxsw_sp_vr_put(struct mlxsw_sp_vr *vr)
715
{
716
	if (!vr->rif_count && list_empty(&vr->fib4->node_list))
717
		mlxsw_sp_vr_destroy(vr);
718 719
}

720
static int mlxsw_sp_vrs_init(struct mlxsw_sp *mlxsw_sp)
721 722
{
	struct mlxsw_sp_vr *vr;
J
Jiri Pirko 已提交
723
	u64 max_vrs;
724 725
	int i;

J
Jiri Pirko 已提交
726
	if (!MLXSW_CORE_RES_VALID(mlxsw_sp->core, MAX_VRS))
727 728
		return -EIO;

J
Jiri Pirko 已提交
729
	max_vrs = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_VRS);
730 731 732
	mlxsw_sp->router->vrs = kcalloc(max_vrs, sizeof(struct mlxsw_sp_vr),
					GFP_KERNEL);
	if (!mlxsw_sp->router->vrs)
733 734
		return -ENOMEM;

J
Jiri Pirko 已提交
735
	for (i = 0; i < max_vrs; i++) {
736
		vr = &mlxsw_sp->router->vrs[i];
737 738
		vr->id = i;
	}
739 740 741 742

	return 0;
}

743 744
static void mlxsw_sp_router_fib_flush(struct mlxsw_sp *mlxsw_sp);

745 746
static void mlxsw_sp_vrs_fini(struct mlxsw_sp *mlxsw_sp)
{
747 748 749 750 751 752 753 754
	/* At this stage we're guaranteed not to have new incoming
	 * FIB notifications and the work queue is free from FIBs
	 * sitting on top of mlxsw netdevs. However, we can still
	 * have other FIBs queued. Flush the queue before flushing
	 * the device's tables. No need for locks, as we're the only
	 * writer.
	 */
	mlxsw_core_flush_owq();
755
	mlxsw_sp_router_fib_flush(mlxsw_sp);
756
	kfree(mlxsw_sp->router->vrs);
757 758
}

759
struct mlxsw_sp_neigh_key {
760
	struct neighbour *n;
761 762 763
};

struct mlxsw_sp_neigh_entry {
764
	struct list_head rif_list_node;
765 766 767
	struct rhash_head ht_node;
	struct mlxsw_sp_neigh_key key;
	u16 rif;
768
	bool connected;
769
	unsigned char ha[ETH_ALEN];
770 771 772
	struct list_head nexthop_list; /* list of nexthops using
					* this neigh entry
					*/
773
	struct list_head nexthop_neighs_list_node;
774 775 776 777 778 779 780 781 782
};

static const struct rhashtable_params mlxsw_sp_neigh_ht_params = {
	.key_offset = offsetof(struct mlxsw_sp_neigh_entry, key),
	.head_offset = offsetof(struct mlxsw_sp_neigh_entry, ht_node),
	.key_len = sizeof(struct mlxsw_sp_neigh_key),
};

static struct mlxsw_sp_neigh_entry *
783 784
mlxsw_sp_neigh_entry_alloc(struct mlxsw_sp *mlxsw_sp, struct neighbour *n,
			   u16 rif)
785 786 787
{
	struct mlxsw_sp_neigh_entry *neigh_entry;

788
	neigh_entry = kzalloc(sizeof(*neigh_entry), GFP_KERNEL);
789 790
	if (!neigh_entry)
		return NULL;
791

792
	neigh_entry->key.n = n;
793
	neigh_entry->rif = rif;
794
	INIT_LIST_HEAD(&neigh_entry->nexthop_list);
795

796 797 798
	return neigh_entry;
}

799
static void mlxsw_sp_neigh_entry_free(struct mlxsw_sp_neigh_entry *neigh_entry)
800 801 802 803
{
	kfree(neigh_entry);
}

804 805 806
static int
mlxsw_sp_neigh_entry_insert(struct mlxsw_sp *mlxsw_sp,
			    struct mlxsw_sp_neigh_entry *neigh_entry)
807
{
808
	return rhashtable_insert_fast(&mlxsw_sp->router->neigh_ht,
809 810 811
				      &neigh_entry->ht_node,
				      mlxsw_sp_neigh_ht_params);
}
812

813 814 815 816
static void
mlxsw_sp_neigh_entry_remove(struct mlxsw_sp *mlxsw_sp,
			    struct mlxsw_sp_neigh_entry *neigh_entry)
{
817
	rhashtable_remove_fast(&mlxsw_sp->router->neigh_ht,
818 819
			       &neigh_entry->ht_node,
			       mlxsw_sp_neigh_ht_params);
820 821
}

822 823
static struct mlxsw_sp_neigh_entry *
mlxsw_sp_neigh_entry_create(struct mlxsw_sp *mlxsw_sp, struct neighbour *n)
824 825
{
	struct mlxsw_sp_neigh_entry *neigh_entry;
826
	struct mlxsw_sp_rif *rif;
827 828
	int err;

829 830
	rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, n->dev);
	if (!rif)
831
		return ERR_PTR(-EINVAL);
832

833
	neigh_entry = mlxsw_sp_neigh_entry_alloc(mlxsw_sp, n, rif->rif_index);
834
	if (!neigh_entry)
835 836
		return ERR_PTR(-ENOMEM);

837 838 839
	err = mlxsw_sp_neigh_entry_insert(mlxsw_sp, neigh_entry);
	if (err)
		goto err_neigh_entry_insert;
840

841
	list_add(&neigh_entry->rif_list_node, &rif->neigh_list);
842

843
	return neigh_entry;
844 845

err_neigh_entry_insert:
846 847
	mlxsw_sp_neigh_entry_free(neigh_entry);
	return ERR_PTR(err);
848 849
}

850 851 852
static void
mlxsw_sp_neigh_entry_destroy(struct mlxsw_sp *mlxsw_sp,
			     struct mlxsw_sp_neigh_entry *neigh_entry)
853
{
854
	list_del(&neigh_entry->rif_list_node);
855 856 857
	mlxsw_sp_neigh_entry_remove(mlxsw_sp, neigh_entry);
	mlxsw_sp_neigh_entry_free(neigh_entry);
}
858

859 860 861 862
static struct mlxsw_sp_neigh_entry *
mlxsw_sp_neigh_entry_lookup(struct mlxsw_sp *mlxsw_sp, struct neighbour *n)
{
	struct mlxsw_sp_neigh_key key;
863

864
	key.n = n;
865
	return rhashtable_lookup_fast(&mlxsw_sp->router->neigh_ht,
866
				      &key, mlxsw_sp_neigh_ht_params);
867 868
}

869 870 871 872 873
static void
mlxsw_sp_router_neighs_update_interval_init(struct mlxsw_sp *mlxsw_sp)
{
	unsigned long interval = NEIGH_VAR(&arp_tbl.parms, DELAY_PROBE_TIME);

874
	mlxsw_sp->router->neighs_update.interval = jiffies_to_msecs(interval);
875 876 877 878 879 880 881 882 883 884 885 886 887 888
}

static void mlxsw_sp_router_neigh_ent_ipv4_process(struct mlxsw_sp *mlxsw_sp,
						   char *rauhtd_pl,
						   int ent_index)
{
	struct net_device *dev;
	struct neighbour *n;
	__be32 dipn;
	u32 dip;
	u16 rif;

	mlxsw_reg_rauhtd_ent_ipv4_unpack(rauhtd_pl, ent_index, &rif, &dip);

889
	if (!mlxsw_sp->router->rifs[rif]) {
890 891 892 893 894
		dev_err_ratelimited(mlxsw_sp->bus_info->dev, "Incorrect RIF in neighbour entry\n");
		return;
	}

	dipn = htonl(dip);
895
	dev = mlxsw_sp->router->rifs[rif]->dev;
896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944
	n = neigh_lookup(&arp_tbl, &dipn, dev);
	if (!n) {
		netdev_err(dev, "Failed to find matching neighbour for IP=%pI4h\n",
			   &dip);
		return;
	}

	netdev_dbg(dev, "Updating neighbour with IP=%pI4h\n", &dip);
	neigh_event_send(n, NULL);
	neigh_release(n);
}

static void mlxsw_sp_router_neigh_rec_ipv4_process(struct mlxsw_sp *mlxsw_sp,
						   char *rauhtd_pl,
						   int rec_index)
{
	u8 num_entries;
	int i;

	num_entries = mlxsw_reg_rauhtd_ipv4_rec_num_entries_get(rauhtd_pl,
								rec_index);
	/* Hardware starts counting at 0, so add 1. */
	num_entries++;

	/* Each record consists of several neighbour entries. */
	for (i = 0; i < num_entries; i++) {
		int ent_index;

		ent_index = rec_index * MLXSW_REG_RAUHTD_IPV4_ENT_PER_REC + i;
		mlxsw_sp_router_neigh_ent_ipv4_process(mlxsw_sp, rauhtd_pl,
						       ent_index);
	}

}

static void mlxsw_sp_router_neigh_rec_process(struct mlxsw_sp *mlxsw_sp,
					      char *rauhtd_pl, int rec_index)
{
	switch (mlxsw_reg_rauhtd_rec_type_get(rauhtd_pl, rec_index)) {
	case MLXSW_REG_RAUHTD_TYPE_IPV4:
		mlxsw_sp_router_neigh_rec_ipv4_process(mlxsw_sp, rauhtd_pl,
						       rec_index);
		break;
	case MLXSW_REG_RAUHTD_TYPE_IPV6:
		WARN_ON_ONCE(1);
		break;
	}
}

945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964
static bool mlxsw_sp_router_rauhtd_is_full(char *rauhtd_pl)
{
	u8 num_rec, last_rec_index, num_entries;

	num_rec = mlxsw_reg_rauhtd_num_rec_get(rauhtd_pl);
	last_rec_index = num_rec - 1;

	if (num_rec < MLXSW_REG_RAUHTD_REC_MAX_NUM)
		return false;
	if (mlxsw_reg_rauhtd_rec_type_get(rauhtd_pl, last_rec_index) ==
	    MLXSW_REG_RAUHTD_TYPE_IPV6)
		return true;

	num_entries = mlxsw_reg_rauhtd_ipv4_rec_num_entries_get(rauhtd_pl,
								last_rec_index);
	if (++num_entries == MLXSW_REG_RAUHTD_IPV4_ENT_PER_REC)
		return true;
	return false;
}

965
static int mlxsw_sp_router_neighs_update_rauhtd(struct mlxsw_sp *mlxsw_sp)
966 967 968 969 970 971 972
{
	char *rauhtd_pl;
	u8 num_rec;
	int i, err;

	rauhtd_pl = kmalloc(MLXSW_REG_RAUHTD_LEN, GFP_KERNEL);
	if (!rauhtd_pl)
973
		return -ENOMEM;
974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990

	/* Make sure the neighbour's netdev isn't removed in the
	 * process.
	 */
	rtnl_lock();
	do {
		mlxsw_reg_rauhtd_pack(rauhtd_pl, MLXSW_REG_RAUHTD_TYPE_IPV4);
		err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(rauhtd),
				      rauhtd_pl);
		if (err) {
			dev_err_ratelimited(mlxsw_sp->bus_info->dev, "Failed to dump neighbour talbe\n");
			break;
		}
		num_rec = mlxsw_reg_rauhtd_num_rec_get(rauhtd_pl);
		for (i = 0; i < num_rec; i++)
			mlxsw_sp_router_neigh_rec_process(mlxsw_sp, rauhtd_pl,
							  i);
991
	} while (mlxsw_sp_router_rauhtd_is_full(rauhtd_pl));
992 993 994
	rtnl_unlock();

	kfree(rauhtd_pl);
995 996 997 998 999 1000 1001 1002 1003
	return err;
}

static void mlxsw_sp_router_neighs_update_nh(struct mlxsw_sp *mlxsw_sp)
{
	struct mlxsw_sp_neigh_entry *neigh_entry;

	/* Take RTNL mutex here to prevent lists from changes */
	rtnl_lock();
1004
	list_for_each_entry(neigh_entry, &mlxsw_sp->router->nexthop_neighs_list,
1005
			    nexthop_neighs_list_node)
1006 1007 1008
		/* If this neigh have nexthops, make the kernel think this neigh
		 * is active regardless of the traffic.
		 */
1009
		neigh_event_send(neigh_entry->key.n, NULL);
1010 1011 1012 1013 1014 1015
	rtnl_unlock();
}

static void
mlxsw_sp_router_neighs_update_work_schedule(struct mlxsw_sp *mlxsw_sp)
{
1016
	unsigned long interval = mlxsw_sp->router->neighs_update.interval;
1017

1018
	mlxsw_core_schedule_dw(&mlxsw_sp->router->neighs_update.dw,
1019 1020 1021 1022 1023
			       msecs_to_jiffies(interval));
}

static void mlxsw_sp_router_neighs_update_work(struct work_struct *work)
{
1024
	struct mlxsw_sp_router *router;
1025 1026
	int err;

1027 1028 1029
	router = container_of(work, struct mlxsw_sp_router,
			      neighs_update.dw.work);
	err = mlxsw_sp_router_neighs_update_rauhtd(router->mlxsw_sp);
1030
	if (err)
1031
		dev_err(router->mlxsw_sp->bus_info->dev, "Could not update kernel for neigh activity");
1032

1033
	mlxsw_sp_router_neighs_update_nh(router->mlxsw_sp);
1034

1035
	mlxsw_sp_router_neighs_update_work_schedule(router->mlxsw_sp);
1036 1037
}

1038 1039 1040
static void mlxsw_sp_router_probe_unresolved_nexthops(struct work_struct *work)
{
	struct mlxsw_sp_neigh_entry *neigh_entry;
1041
	struct mlxsw_sp_router *router;
1042

1043 1044
	router = container_of(work, struct mlxsw_sp_router,
			      nexthop_probe_dw.work);
1045 1046 1047 1048 1049 1050 1051 1052 1053
	/* Iterate over nexthop neighbours, find those who are unresolved and
	 * send arp on them. This solves the chicken-egg problem when
	 * the nexthop wouldn't get offloaded until the neighbor is resolved
	 * but it wouldn't get resolved ever in case traffic is flowing in HW
	 * using different nexthop.
	 *
	 * Take RTNL mutex here to prevent lists from changes.
	 */
	rtnl_lock();
1054
	list_for_each_entry(neigh_entry, &router->nexthop_neighs_list,
1055
			    nexthop_neighs_list_node)
1056
		if (!neigh_entry->connected)
1057
			neigh_event_send(neigh_entry->key.n, NULL);
1058 1059
	rtnl_unlock();

1060
	mlxsw_core_schedule_dw(&router->nexthop_probe_dw,
1061 1062 1063
			       MLXSW_SP_UNRESOLVED_NH_PROBE_INTERVAL);
}

1064 1065 1066 1067 1068
static void
mlxsw_sp_nexthop_neigh_update(struct mlxsw_sp *mlxsw_sp,
			      struct mlxsw_sp_neigh_entry *neigh_entry,
			      bool removing);

1069 1070 1071 1072 1073 1074 1075 1076 1077 1078
static enum mlxsw_reg_rauht_op mlxsw_sp_rauht_op(bool adding)
{
	return adding ? MLXSW_REG_RAUHT_OP_WRITE_ADD :
			MLXSW_REG_RAUHT_OP_WRITE_DELETE;
}

static void
mlxsw_sp_router_neigh_entry_op4(struct mlxsw_sp *mlxsw_sp,
				struct mlxsw_sp_neigh_entry *neigh_entry,
				enum mlxsw_reg_rauht_op op)
1079
{
1080
	struct neighbour *n = neigh_entry->key.n;
1081
	u32 dip = ntohl(*((__be32 *) n->primary_key));
1082
	char rauht_pl[MLXSW_REG_RAUHT_LEN];
1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117

	mlxsw_reg_rauht_pack4(rauht_pl, op, neigh_entry->rif, neigh_entry->ha,
			      dip);
	mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rauht), rauht_pl);
}

static void
mlxsw_sp_neigh_entry_update(struct mlxsw_sp *mlxsw_sp,
			    struct mlxsw_sp_neigh_entry *neigh_entry,
			    bool adding)
{
	if (!adding && !neigh_entry->connected)
		return;
	neigh_entry->connected = adding;
	if (neigh_entry->key.n->tbl == &arp_tbl)
		mlxsw_sp_router_neigh_entry_op4(mlxsw_sp, neigh_entry,
						mlxsw_sp_rauht_op(adding));
	else
		WARN_ON_ONCE(1);
}

struct mlxsw_sp_neigh_event_work {
	struct work_struct work;
	struct mlxsw_sp *mlxsw_sp;
	struct neighbour *n;
};

static void mlxsw_sp_router_neigh_event_work(struct work_struct *work)
{
	struct mlxsw_sp_neigh_event_work *neigh_work =
		container_of(work, struct mlxsw_sp_neigh_event_work, work);
	struct mlxsw_sp *mlxsw_sp = neigh_work->mlxsw_sp;
	struct mlxsw_sp_neigh_entry *neigh_entry;
	struct neighbour *n = neigh_work->n;
	unsigned char ha[ETH_ALEN];
1118
	bool entry_connected;
1119
	u8 nud_state, dead;
1120

1121 1122 1123 1124
	/* If these parameters are changed after we release the lock,
	 * then we are guaranteed to receive another event letting us
	 * know about it.
	 */
1125
	read_lock_bh(&n->lock);
1126
	memcpy(ha, n->ha, ETH_ALEN);
1127
	nud_state = n->nud_state;
1128
	dead = n->dead;
1129 1130
	read_unlock_bh(&n->lock);

1131
	rtnl_lock();
1132
	entry_connected = nud_state & NUD_VALID && !dead;
1133 1134 1135 1136 1137 1138 1139
	neigh_entry = mlxsw_sp_neigh_entry_lookup(mlxsw_sp, n);
	if (!entry_connected && !neigh_entry)
		goto out;
	if (!neigh_entry) {
		neigh_entry = mlxsw_sp_neigh_entry_create(mlxsw_sp, n);
		if (IS_ERR(neigh_entry))
			goto out;
1140 1141
	}

1142 1143 1144 1145 1146 1147 1148 1149 1150
	memcpy(neigh_entry->ha, ha, ETH_ALEN);
	mlxsw_sp_neigh_entry_update(mlxsw_sp, neigh_entry, entry_connected);
	mlxsw_sp_nexthop_neigh_update(mlxsw_sp, neigh_entry, !entry_connected);

	if (!neigh_entry->connected && list_empty(&neigh_entry->nexthop_list))
		mlxsw_sp_neigh_entry_destroy(mlxsw_sp, neigh_entry);

out:
	rtnl_unlock();
1151
	neigh_release(n);
1152
	kfree(neigh_work);
1153 1154
}

1155 1156
int mlxsw_sp_router_netevent_event(struct notifier_block *unused,
				   unsigned long event, void *ptr)
1157
{
1158
	struct mlxsw_sp_neigh_event_work *neigh_work;
1159 1160 1161 1162
	struct mlxsw_sp_port *mlxsw_sp_port;
	struct mlxsw_sp *mlxsw_sp;
	unsigned long interval;
	struct neigh_parms *p;
1163
	struct neighbour *n;
1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181

	switch (event) {
	case NETEVENT_DELAY_PROBE_TIME_UPDATE:
		p = ptr;

		/* We don't care about changes in the default table. */
		if (!p->dev || p->tbl != &arp_tbl)
			return NOTIFY_DONE;

		/* We are in atomic context and can't take RTNL mutex,
		 * so use RCU variant to walk the device chain.
		 */
		mlxsw_sp_port = mlxsw_sp_port_lower_dev_hold(p->dev);
		if (!mlxsw_sp_port)
			return NOTIFY_DONE;

		mlxsw_sp = mlxsw_sp_port->mlxsw_sp;
		interval = jiffies_to_msecs(NEIGH_VAR(p, DELAY_PROBE_TIME));
1182
		mlxsw_sp->router->neighs_update.interval = interval;
1183 1184 1185

		mlxsw_sp_port_dev_put(mlxsw_sp_port);
		break;
1186 1187 1188 1189 1190 1191
	case NETEVENT_NEIGH_UPDATE:
		n = ptr;

		if (n->tbl != &arp_tbl)
			return NOTIFY_DONE;

1192
		mlxsw_sp_port = mlxsw_sp_port_lower_dev_hold(n->dev);
1193 1194 1195
		if (!mlxsw_sp_port)
			return NOTIFY_DONE;

1196 1197
		neigh_work = kzalloc(sizeof(*neigh_work), GFP_ATOMIC);
		if (!neigh_work) {
1198
			mlxsw_sp_port_dev_put(mlxsw_sp_port);
1199
			return NOTIFY_BAD;
1200
		}
1201 1202 1203 1204

		INIT_WORK(&neigh_work->work, mlxsw_sp_router_neigh_event_work);
		neigh_work->mlxsw_sp = mlxsw_sp_port->mlxsw_sp;
		neigh_work->n = n;
1205 1206 1207 1208 1209 1210

		/* Take a reference to ensure the neighbour won't be
		 * destructed until we drop the reference in delayed
		 * work.
		 */
		neigh_clone(n);
1211 1212
		mlxsw_core_schedule_work(&neigh_work->work);
		mlxsw_sp_port_dev_put(mlxsw_sp_port);
1213
		break;
1214 1215 1216 1217 1218
	}

	return NOTIFY_DONE;
}

1219 1220
static int mlxsw_sp_neigh_init(struct mlxsw_sp *mlxsw_sp)
{
1221 1222
	int err;

1223
	err = rhashtable_init(&mlxsw_sp->router->neigh_ht,
1224 1225 1226 1227 1228 1229 1230 1231 1232
			      &mlxsw_sp_neigh_ht_params);
	if (err)
		return err;

	/* Initialize the polling interval according to the default
	 * table.
	 */
	mlxsw_sp_router_neighs_update_interval_init(mlxsw_sp);

1233
	/* Create the delayed works for the activity_update */
1234
	INIT_DELAYED_WORK(&mlxsw_sp->router->neighs_update.dw,
1235
			  mlxsw_sp_router_neighs_update_work);
1236
	INIT_DELAYED_WORK(&mlxsw_sp->router->nexthop_probe_dw,
1237
			  mlxsw_sp_router_probe_unresolved_nexthops);
1238 1239
	mlxsw_core_schedule_dw(&mlxsw_sp->router->neighs_update.dw, 0);
	mlxsw_core_schedule_dw(&mlxsw_sp->router->nexthop_probe_dw, 0);
1240
	return 0;
1241 1242 1243 1244
}

static void mlxsw_sp_neigh_fini(struct mlxsw_sp *mlxsw_sp)
{
1245 1246 1247
	cancel_delayed_work_sync(&mlxsw_sp->router->neighs_update.dw);
	cancel_delayed_work_sync(&mlxsw_sp->router->nexthop_probe_dw);
	rhashtable_destroy(&mlxsw_sp->router->neigh_ht);
1248 1249
}

1250
static int mlxsw_sp_neigh_rif_flush(struct mlxsw_sp *mlxsw_sp,
1251
				    const struct mlxsw_sp_rif *rif)
1252 1253 1254 1255
{
	char rauht_pl[MLXSW_REG_RAUHT_LEN];

	mlxsw_reg_rauht_pack(rauht_pl, MLXSW_REG_RAUHT_OP_WRITE_DELETE_ALL,
1256
			     rif->rif_index, rif->addr);
1257 1258 1259 1260
	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rauht), rauht_pl);
}

static void mlxsw_sp_neigh_rif_gone_sync(struct mlxsw_sp *mlxsw_sp,
1261
					 struct mlxsw_sp_rif *rif)
1262 1263 1264
{
	struct mlxsw_sp_neigh_entry *neigh_entry, *tmp;

1265 1266
	mlxsw_sp_neigh_rif_flush(mlxsw_sp, rif);
	list_for_each_entry_safe(neigh_entry, tmp, &rif->neigh_list,
1267 1268 1269 1270
				 rif_list_node)
		mlxsw_sp_neigh_entry_destroy(mlxsw_sp, neigh_entry);
}

1271 1272 1273 1274
struct mlxsw_sp_nexthop_key {
	struct fib_nh *fib_nh;
};

1275 1276
struct mlxsw_sp_nexthop {
	struct list_head neigh_list_node; /* member of neigh entry list */
1277
	struct list_head rif_list_node;
1278 1279 1280
	struct mlxsw_sp_nexthop_group *nh_grp; /* pointer back to the group
						* this belongs to
						*/
1281 1282
	struct rhash_head ht_node;
	struct mlxsw_sp_nexthop_key key;
1283
	struct mlxsw_sp_rif *rif;
1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295
	u8 should_offload:1, /* set indicates this neigh is connected and
			      * should be put to KVD linear area of this group.
			      */
	   offloaded:1, /* set in case the neigh is actually put into
			 * KVD linear area of this group.
			 */
	   update:1; /* set indicates that MAC of this neigh should be
		      * updated in HW
		      */
	struct mlxsw_sp_neigh_entry *neigh_entry;
};

1296 1297 1298 1299
struct mlxsw_sp_nexthop_group_key {
	struct fib_info *fi;
};

1300
struct mlxsw_sp_nexthop_group {
1301
	struct rhash_head ht_node;
1302
	struct list_head fib_list; /* list of fib entries that use this group */
1303
	struct mlxsw_sp_nexthop_group_key key;
1304 1305
	u8 adj_index_valid:1,
	   gateway:1; /* routes using the group use a gateway */
1306 1307 1308 1309
	u32 adj_index;
	u16 ecmp_size;
	u16 count;
	struct mlxsw_sp_nexthop nexthops[0];
1310
#define nh_rif	nexthops[0].rif
1311 1312
};

1313 1314 1315 1316 1317 1318 1319 1320 1321
static const struct rhashtable_params mlxsw_sp_nexthop_group_ht_params = {
	.key_offset = offsetof(struct mlxsw_sp_nexthop_group, key),
	.head_offset = offsetof(struct mlxsw_sp_nexthop_group, ht_node),
	.key_len = sizeof(struct mlxsw_sp_nexthop_group_key),
};

static int mlxsw_sp_nexthop_group_insert(struct mlxsw_sp *mlxsw_sp,
					 struct mlxsw_sp_nexthop_group *nh_grp)
{
1322
	return rhashtable_insert_fast(&mlxsw_sp->router->nexthop_group_ht,
1323 1324 1325 1326 1327 1328 1329
				      &nh_grp->ht_node,
				      mlxsw_sp_nexthop_group_ht_params);
}

static void mlxsw_sp_nexthop_group_remove(struct mlxsw_sp *mlxsw_sp,
					  struct mlxsw_sp_nexthop_group *nh_grp)
{
1330
	rhashtable_remove_fast(&mlxsw_sp->router->nexthop_group_ht,
1331 1332 1333 1334 1335 1336 1337 1338
			       &nh_grp->ht_node,
			       mlxsw_sp_nexthop_group_ht_params);
}

static struct mlxsw_sp_nexthop_group *
mlxsw_sp_nexthop_group_lookup(struct mlxsw_sp *mlxsw_sp,
			      struct mlxsw_sp_nexthop_group_key key)
{
1339
	return rhashtable_lookup_fast(&mlxsw_sp->router->nexthop_group_ht, &key,
1340 1341 1342
				      mlxsw_sp_nexthop_group_ht_params);
}

1343 1344 1345 1346 1347 1348 1349 1350 1351
static const struct rhashtable_params mlxsw_sp_nexthop_ht_params = {
	.key_offset = offsetof(struct mlxsw_sp_nexthop, key),
	.head_offset = offsetof(struct mlxsw_sp_nexthop, ht_node),
	.key_len = sizeof(struct mlxsw_sp_nexthop_key),
};

static int mlxsw_sp_nexthop_insert(struct mlxsw_sp *mlxsw_sp,
				   struct mlxsw_sp_nexthop *nh)
{
1352
	return rhashtable_insert_fast(&mlxsw_sp->router->nexthop_ht,
1353 1354 1355 1356 1357 1358
				      &nh->ht_node, mlxsw_sp_nexthop_ht_params);
}

static void mlxsw_sp_nexthop_remove(struct mlxsw_sp *mlxsw_sp,
				    struct mlxsw_sp_nexthop *nh)
{
1359
	rhashtable_remove_fast(&mlxsw_sp->router->nexthop_ht, &nh->ht_node,
1360 1361 1362
			       mlxsw_sp_nexthop_ht_params);
}

1363 1364 1365 1366
static struct mlxsw_sp_nexthop *
mlxsw_sp_nexthop_lookup(struct mlxsw_sp *mlxsw_sp,
			struct mlxsw_sp_nexthop_key key)
{
1367
	return rhashtable_lookup_fast(&mlxsw_sp->router->nexthop_ht, &key,
1368 1369 1370
				      mlxsw_sp_nexthop_ht_params);
}

1371
static int mlxsw_sp_adj_index_mass_update_vr(struct mlxsw_sp *mlxsw_sp,
1372
					     const struct mlxsw_sp_fib *fib,
1373 1374 1375 1376 1377 1378
					     u32 adj_index, u16 ecmp_size,
					     u32 new_adj_index,
					     u16 new_ecmp_size)
{
	char raleu_pl[MLXSW_REG_RALEU_LEN];

1379
	mlxsw_reg_raleu_pack(raleu_pl,
1380 1381
			     (enum mlxsw_reg_ralxx_protocol) fib->proto,
			     fib->vr->id, adj_index, ecmp_size, new_adj_index,
1382
			     new_ecmp_size);
1383 1384 1385 1386 1387 1388 1389 1390
	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(raleu), raleu_pl);
}

static int mlxsw_sp_adj_index_mass_update(struct mlxsw_sp *mlxsw_sp,
					  struct mlxsw_sp_nexthop_group *nh_grp,
					  u32 old_adj_index, u16 old_ecmp_size)
{
	struct mlxsw_sp_fib_entry *fib_entry;
1391
	struct mlxsw_sp_fib *fib = NULL;
1392 1393 1394
	int err;

	list_for_each_entry(fib_entry, &nh_grp->fib_list, nexthop_group_node) {
1395
		if (fib == fib_entry->fib_node->fib)
1396
			continue;
1397 1398
		fib = fib_entry->fib_node->fib;
		err = mlxsw_sp_adj_index_mass_update_vr(mlxsw_sp, fib,
1399 1400 1401 1402 1403 1404 1405 1406 1407 1408 1409 1410 1411 1412 1413 1414 1415 1416 1417 1418 1419 1420 1421 1422
							old_adj_index,
							old_ecmp_size,
							nh_grp->adj_index,
							nh_grp->ecmp_size);
		if (err)
			return err;
	}
	return 0;
}

static int mlxsw_sp_nexthop_mac_update(struct mlxsw_sp *mlxsw_sp, u32 adj_index,
				       struct mlxsw_sp_nexthop *nh)
{
	struct mlxsw_sp_neigh_entry *neigh_entry = nh->neigh_entry;
	char ratr_pl[MLXSW_REG_RATR_LEN];

	mlxsw_reg_ratr_pack(ratr_pl, MLXSW_REG_RATR_OP_WRITE_WRITE_ENTRY,
			    true, adj_index, neigh_entry->rif);
	mlxsw_reg_ratr_eth_entry_pack(ratr_pl, neigh_entry->ha);
	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ratr), ratr_pl);
}

static int
mlxsw_sp_nexthop_group_mac_update(struct mlxsw_sp *mlxsw_sp,
1423 1424
				  struct mlxsw_sp_nexthop_group *nh_grp,
				  bool reallocate)
1425 1426 1427 1428 1429 1430 1431 1432 1433 1434 1435 1436 1437 1438
{
	u32 adj_index = nh_grp->adj_index; /* base */
	struct mlxsw_sp_nexthop *nh;
	int i;
	int err;

	for (i = 0; i < nh_grp->count; i++) {
		nh = &nh_grp->nexthops[i];

		if (!nh->should_offload) {
			nh->offloaded = 0;
			continue;
		}

1439
		if (nh->update || reallocate) {
1440 1441 1442 1443 1444 1445 1446 1447 1448 1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467 1468 1469 1470 1471 1472 1473 1474 1475 1476 1477 1478 1479 1480 1481 1482 1483
			err = mlxsw_sp_nexthop_mac_update(mlxsw_sp,
							  adj_index, nh);
			if (err)
				return err;
			nh->update = 0;
			nh->offloaded = 1;
		}
		adj_index++;
	}
	return 0;
}

static int mlxsw_sp_fib_entry_update(struct mlxsw_sp *mlxsw_sp,
				     struct mlxsw_sp_fib_entry *fib_entry);

static int
mlxsw_sp_nexthop_fib_entries_update(struct mlxsw_sp *mlxsw_sp,
				    struct mlxsw_sp_nexthop_group *nh_grp)
{
	struct mlxsw_sp_fib_entry *fib_entry;
	int err;

	list_for_each_entry(fib_entry, &nh_grp->fib_list, nexthop_group_node) {
		err = mlxsw_sp_fib_entry_update(mlxsw_sp, fib_entry);
		if (err)
			return err;
	}
	return 0;
}

static void
mlxsw_sp_nexthop_group_refresh(struct mlxsw_sp *mlxsw_sp,
			       struct mlxsw_sp_nexthop_group *nh_grp)
{
	struct mlxsw_sp_nexthop *nh;
	bool offload_change = false;
	u32 adj_index;
	u16 ecmp_size = 0;
	bool old_adj_index_valid;
	u32 old_adj_index;
	u16 old_ecmp_size;
	int i;
	int err;

1484 1485 1486 1487 1488
	if (!nh_grp->gateway) {
		mlxsw_sp_nexthop_fib_entries_update(mlxsw_sp, nh_grp);
		return;
	}

1489 1490 1491 1492 1493 1494 1495 1496 1497 1498 1499 1500 1501 1502 1503
	for (i = 0; i < nh_grp->count; i++) {
		nh = &nh_grp->nexthops[i];

		if (nh->should_offload ^ nh->offloaded) {
			offload_change = true;
			if (nh->should_offload)
				nh->update = 1;
		}
		if (nh->should_offload)
			ecmp_size++;
	}
	if (!offload_change) {
		/* Nothing was added or removed, so no need to reallocate. Just
		 * update MAC on existing adjacency indexes.
		 */
1504 1505
		err = mlxsw_sp_nexthop_group_mac_update(mlxsw_sp, nh_grp,
							false);
1506 1507 1508 1509 1510 1511 1512 1513 1514 1515 1516 1517
		if (err) {
			dev_warn(mlxsw_sp->bus_info->dev, "Failed to update neigh MAC in adjacency table.\n");
			goto set_trap;
		}
		return;
	}
	if (!ecmp_size)
		/* No neigh of this group is connected so we just set
		 * the trap and let everthing flow through kernel.
		 */
		goto set_trap;

1518 1519
	err = mlxsw_sp_kvdl_alloc(mlxsw_sp, ecmp_size, &adj_index);
	if (err) {
1520 1521 1522 1523 1524 1525 1526 1527 1528 1529 1530 1531
		/* We ran out of KVD linear space, just set the
		 * trap and let everything flow through kernel.
		 */
		dev_warn(mlxsw_sp->bus_info->dev, "Failed to allocate KVD linear area for nexthop group.\n");
		goto set_trap;
	}
	old_adj_index_valid = nh_grp->adj_index_valid;
	old_adj_index = nh_grp->adj_index;
	old_ecmp_size = nh_grp->ecmp_size;
	nh_grp->adj_index_valid = 1;
	nh_grp->adj_index = adj_index;
	nh_grp->ecmp_size = ecmp_size;
1532
	err = mlxsw_sp_nexthop_group_mac_update(mlxsw_sp, nh_grp, true);
1533 1534 1535 1536 1537 1538 1539 1540 1541 1542 1543 1544 1545 1546 1547 1548 1549 1550 1551 1552 1553 1554 1555 1556 1557 1558 1559 1560 1561 1562 1563 1564 1565 1566 1567 1568 1569 1570 1571 1572 1573 1574 1575 1576 1577 1578 1579 1580 1581 1582 1583 1584 1585 1586 1587 1588 1589 1590 1591 1592 1593 1594 1595 1596
	if (err) {
		dev_warn(mlxsw_sp->bus_info->dev, "Failed to update neigh MAC in adjacency table.\n");
		goto set_trap;
	}

	if (!old_adj_index_valid) {
		/* The trap was set for fib entries, so we have to call
		 * fib entry update to unset it and use adjacency index.
		 */
		err = mlxsw_sp_nexthop_fib_entries_update(mlxsw_sp, nh_grp);
		if (err) {
			dev_warn(mlxsw_sp->bus_info->dev, "Failed to add adjacency index to fib entries.\n");
			goto set_trap;
		}
		return;
	}

	err = mlxsw_sp_adj_index_mass_update(mlxsw_sp, nh_grp,
					     old_adj_index, old_ecmp_size);
	mlxsw_sp_kvdl_free(mlxsw_sp, old_adj_index);
	if (err) {
		dev_warn(mlxsw_sp->bus_info->dev, "Failed to mass-update adjacency index for nexthop group.\n");
		goto set_trap;
	}
	return;

set_trap:
	old_adj_index_valid = nh_grp->adj_index_valid;
	nh_grp->adj_index_valid = 0;
	for (i = 0; i < nh_grp->count; i++) {
		nh = &nh_grp->nexthops[i];
		nh->offloaded = 0;
	}
	err = mlxsw_sp_nexthop_fib_entries_update(mlxsw_sp, nh_grp);
	if (err)
		dev_warn(mlxsw_sp->bus_info->dev, "Failed to set traps for fib entries.\n");
	if (old_adj_index_valid)
		mlxsw_sp_kvdl_free(mlxsw_sp, nh_grp->adj_index);
}

static void __mlxsw_sp_nexthop_neigh_update(struct mlxsw_sp_nexthop *nh,
					    bool removing)
{
	if (!removing && !nh->should_offload)
		nh->should_offload = 1;
	else if (removing && nh->offloaded)
		nh->should_offload = 0;
	nh->update = 1;
}

static void
mlxsw_sp_nexthop_neigh_update(struct mlxsw_sp *mlxsw_sp,
			      struct mlxsw_sp_neigh_entry *neigh_entry,
			      bool removing)
{
	struct mlxsw_sp_nexthop *nh;

	list_for_each_entry(nh, &neigh_entry->nexthop_list,
			    neigh_list_node) {
		__mlxsw_sp_nexthop_neigh_update(nh, removing);
		mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh->nh_grp);
	}
}

1597
static void mlxsw_sp_nexthop_rif_init(struct mlxsw_sp_nexthop *nh,
1598
				      struct mlxsw_sp_rif *rif)
1599
{
1600
	if (nh->rif)
1601 1602
		return;

1603 1604
	nh->rif = rif;
	list_add(&nh->rif_list_node, &rif->nexthop_list);
1605 1606 1607 1608
}

static void mlxsw_sp_nexthop_rif_fini(struct mlxsw_sp_nexthop *nh)
{
1609
	if (!nh->rif)
1610 1611 1612
		return;

	list_del(&nh->rif_list_node);
1613
	nh->rif = NULL;
1614 1615
}

1616 1617
static int mlxsw_sp_nexthop_neigh_init(struct mlxsw_sp *mlxsw_sp,
				       struct mlxsw_sp_nexthop *nh)
1618 1619
{
	struct mlxsw_sp_neigh_entry *neigh_entry;
1620
	struct fib_nh *fib_nh = nh->key.fib_nh;
1621
	struct neighbour *n;
1622
	u8 nud_state, dead;
1623 1624
	int err;

1625
	if (!nh->nh_grp->gateway || nh->neigh_entry)
1626 1627
		return 0;

1628 1629 1630
	/* Take a reference of neigh here ensuring that neigh would
	 * not be detructed before the nexthop entry is finished.
	 * The reference is taken either in neigh_lookup() or
1631
	 * in neigh_create() in case n is not found.
1632
	 */
1633
	n = neigh_lookup(&arp_tbl, &fib_nh->nh_gw, fib_nh->nh_dev);
1634
	if (!n) {
1635 1636 1637
		n = neigh_create(&arp_tbl, &fib_nh->nh_gw, fib_nh->nh_dev);
		if (IS_ERR(n))
			return PTR_ERR(n);
1638
		neigh_event_send(n, NULL);
1639 1640 1641
	}
	neigh_entry = mlxsw_sp_neigh_entry_lookup(mlxsw_sp, n);
	if (!neigh_entry) {
1642 1643
		neigh_entry = mlxsw_sp_neigh_entry_create(mlxsw_sp, n);
		if (IS_ERR(neigh_entry)) {
1644 1645
			err = -EINVAL;
			goto err_neigh_entry_create;
1646
		}
1647
	}
1648 1649 1650 1651 1652 1653

	/* If that is the first nexthop connected to that neigh, add to
	 * nexthop_neighs_list
	 */
	if (list_empty(&neigh_entry->nexthop_list))
		list_add_tail(&neigh_entry->nexthop_neighs_list_node,
1654
			      &mlxsw_sp->router->nexthop_neighs_list);
1655

1656 1657 1658 1659
	nh->neigh_entry = neigh_entry;
	list_add_tail(&nh->neigh_list_node, &neigh_entry->nexthop_list);
	read_lock_bh(&n->lock);
	nud_state = n->nud_state;
1660
	dead = n->dead;
1661
	read_unlock_bh(&n->lock);
1662
	__mlxsw_sp_nexthop_neigh_update(nh, !(nud_state & NUD_VALID && !dead));
1663 1664

	return 0;
1665 1666 1667 1668

err_neigh_entry_create:
	neigh_release(n);
	return err;
1669 1670
}

1671 1672
static void mlxsw_sp_nexthop_neigh_fini(struct mlxsw_sp *mlxsw_sp,
					struct mlxsw_sp_nexthop *nh)
1673 1674
{
	struct mlxsw_sp_neigh_entry *neigh_entry = nh->neigh_entry;
1675
	struct neighbour *n;
1676

1677
	if (!neigh_entry)
1678 1679
		return;
	n = neigh_entry->key.n;
1680

1681
	__mlxsw_sp_nexthop_neigh_update(nh, true);
1682
	list_del(&nh->neigh_list_node);
1683
	nh->neigh_entry = NULL;
1684 1685 1686 1687

	/* If that is the last nexthop connected to that neigh, remove from
	 * nexthop_neighs_list
	 */
1688 1689
	if (list_empty(&neigh_entry->nexthop_list))
		list_del(&neigh_entry->nexthop_neighs_list_node);
1690

1691 1692 1693 1694
	if (!neigh_entry->connected && list_empty(&neigh_entry->nexthop_list))
		mlxsw_sp_neigh_entry_destroy(mlxsw_sp, neigh_entry);

	neigh_release(n);
1695
}
1696

1697 1698 1699 1700 1701 1702
static int mlxsw_sp_nexthop_init(struct mlxsw_sp *mlxsw_sp,
				 struct mlxsw_sp_nexthop_group *nh_grp,
				 struct mlxsw_sp_nexthop *nh,
				 struct fib_nh *fib_nh)
{
	struct net_device *dev = fib_nh->nh_dev;
1703
	struct in_device *in_dev;
1704
	struct mlxsw_sp_rif *rif;
1705 1706 1707 1708 1709 1710 1711 1712
	int err;

	nh->nh_grp = nh_grp;
	nh->key.fib_nh = fib_nh;
	err = mlxsw_sp_nexthop_insert(mlxsw_sp, nh);
	if (err)
		return err;

1713 1714 1715
	if (!dev)
		return 0;

1716 1717 1718 1719 1720
	in_dev = __in_dev_get_rtnl(dev);
	if (in_dev && IN_DEV_IGNORE_ROUTES_WITH_LINKDOWN(in_dev) &&
	    fib_nh->nh_flags & RTNH_F_LINKDOWN)
		return 0;

1721 1722
	rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
	if (!rif)
1723
		return 0;
1724
	mlxsw_sp_nexthop_rif_init(nh, rif);
1725 1726 1727 1728 1729 1730 1731 1732 1733 1734 1735 1736 1737 1738 1739 1740

	err = mlxsw_sp_nexthop_neigh_init(mlxsw_sp, nh);
	if (err)
		goto err_nexthop_neigh_init;

	return 0;

err_nexthop_neigh_init:
	mlxsw_sp_nexthop_remove(mlxsw_sp, nh);
	return err;
}

static void mlxsw_sp_nexthop_fini(struct mlxsw_sp *mlxsw_sp,
				  struct mlxsw_sp_nexthop *nh)
{
	mlxsw_sp_nexthop_neigh_fini(mlxsw_sp, nh);
1741
	mlxsw_sp_nexthop_rif_fini(nh);
1742
	mlxsw_sp_nexthop_remove(mlxsw_sp, nh);
1743 1744
}

1745 1746 1747 1748 1749
static void mlxsw_sp_nexthop_event(struct mlxsw_sp *mlxsw_sp,
				   unsigned long event, struct fib_nh *fib_nh)
{
	struct mlxsw_sp_nexthop_key key;
	struct mlxsw_sp_nexthop *nh;
1750
	struct mlxsw_sp_rif *rif;
1751

1752
	if (mlxsw_sp->router->aborted)
1753 1754 1755 1756 1757 1758 1759
		return;

	key.fib_nh = fib_nh;
	nh = mlxsw_sp_nexthop_lookup(mlxsw_sp, key);
	if (WARN_ON_ONCE(!nh))
		return;

1760 1761
	rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, fib_nh->nh_dev);
	if (!rif)
1762 1763 1764 1765
		return;

	switch (event) {
	case FIB_EVENT_NH_ADD:
1766
		mlxsw_sp_nexthop_rif_init(nh, rif);
1767 1768 1769 1770
		mlxsw_sp_nexthop_neigh_init(mlxsw_sp, nh);
		break;
	case FIB_EVENT_NH_DEL:
		mlxsw_sp_nexthop_neigh_fini(mlxsw_sp, nh);
1771
		mlxsw_sp_nexthop_rif_fini(nh);
1772 1773 1774 1775 1776 1777
		break;
	}

	mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh->nh_grp);
}

1778
static void mlxsw_sp_nexthop_rif_gone_sync(struct mlxsw_sp *mlxsw_sp,
1779
					   struct mlxsw_sp_rif *rif)
1780 1781 1782
{
	struct mlxsw_sp_nexthop *nh, *tmp;

1783
	list_for_each_entry_safe(nh, tmp, &rif->nexthop_list, rif_list_node) {
1784 1785 1786 1787 1788 1789
		mlxsw_sp_nexthop_neigh_fini(mlxsw_sp, nh);
		mlxsw_sp_nexthop_rif_fini(nh);
		mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh->nh_grp);
	}
}

1790 1791 1792 1793 1794 1795 1796 1797 1798 1799 1800 1801 1802 1803 1804 1805
static struct mlxsw_sp_nexthop_group *
mlxsw_sp_nexthop_group_create(struct mlxsw_sp *mlxsw_sp, struct fib_info *fi)
{
	struct mlxsw_sp_nexthop_group *nh_grp;
	struct mlxsw_sp_nexthop *nh;
	struct fib_nh *fib_nh;
	size_t alloc_size;
	int i;
	int err;

	alloc_size = sizeof(*nh_grp) +
		     fi->fib_nhs * sizeof(struct mlxsw_sp_nexthop);
	nh_grp = kzalloc(alloc_size, GFP_KERNEL);
	if (!nh_grp)
		return ERR_PTR(-ENOMEM);
	INIT_LIST_HEAD(&nh_grp->fib_list);
1806
	nh_grp->gateway = fi->fib_nh->nh_scope == RT_SCOPE_LINK;
1807
	nh_grp->count = fi->fib_nhs;
1808
	nh_grp->key.fi = fi;
1809 1810 1811 1812 1813 1814 1815
	for (i = 0; i < nh_grp->count; i++) {
		nh = &nh_grp->nexthops[i];
		fib_nh = &fi->fib_nh[i];
		err = mlxsw_sp_nexthop_init(mlxsw_sp, nh_grp, nh, fib_nh);
		if (err)
			goto err_nexthop_init;
	}
1816 1817 1818
	err = mlxsw_sp_nexthop_group_insert(mlxsw_sp, nh_grp);
	if (err)
		goto err_nexthop_group_insert;
1819 1820 1821
	mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh_grp);
	return nh_grp;

1822
err_nexthop_group_insert:
1823
err_nexthop_init:
1824 1825
	for (i--; i >= 0; i--) {
		nh = &nh_grp->nexthops[i];
1826
		mlxsw_sp_nexthop_fini(mlxsw_sp, nh);
1827
	}
1828 1829 1830 1831 1832 1833 1834 1835 1836 1837 1838
	kfree(nh_grp);
	return ERR_PTR(err);
}

static void
mlxsw_sp_nexthop_group_destroy(struct mlxsw_sp *mlxsw_sp,
			       struct mlxsw_sp_nexthop_group *nh_grp)
{
	struct mlxsw_sp_nexthop *nh;
	int i;

1839
	mlxsw_sp_nexthop_group_remove(mlxsw_sp, nh_grp);
1840 1841 1842 1843
	for (i = 0; i < nh_grp->count; i++) {
		nh = &nh_grp->nexthops[i];
		mlxsw_sp_nexthop_fini(mlxsw_sp, nh);
	}
1844 1845
	mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh_grp);
	WARN_ON_ONCE(nh_grp->adj_index_valid);
1846 1847 1848 1849 1850 1851 1852
	kfree(nh_grp);
}

static int mlxsw_sp_nexthop_group_get(struct mlxsw_sp *mlxsw_sp,
				      struct mlxsw_sp_fib_entry *fib_entry,
				      struct fib_info *fi)
{
1853
	struct mlxsw_sp_nexthop_group_key key;
1854 1855
	struct mlxsw_sp_nexthop_group *nh_grp;

1856 1857
	key.fi = fi;
	nh_grp = mlxsw_sp_nexthop_group_lookup(mlxsw_sp, key);
1858 1859 1860 1861 1862 1863 1864 1865 1866 1867 1868 1869 1870 1871 1872 1873 1874 1875 1876 1877 1878
	if (!nh_grp) {
		nh_grp = mlxsw_sp_nexthop_group_create(mlxsw_sp, fi);
		if (IS_ERR(nh_grp))
			return PTR_ERR(nh_grp);
	}
	list_add_tail(&fib_entry->nexthop_group_node, &nh_grp->fib_list);
	fib_entry->nh_group = nh_grp;
	return 0;
}

static void mlxsw_sp_nexthop_group_put(struct mlxsw_sp *mlxsw_sp,
				       struct mlxsw_sp_fib_entry *fib_entry)
{
	struct mlxsw_sp_nexthop_group *nh_grp = fib_entry->nh_group;

	list_del(&fib_entry->nexthop_group_node);
	if (!list_empty(&nh_grp->fib_list))
		return;
	mlxsw_sp_nexthop_group_destroy(mlxsw_sp, nh_grp);
}

1879 1880 1881 1882 1883
static bool
mlxsw_sp_fib_entry_should_offload(const struct mlxsw_sp_fib_entry *fib_entry)
{
	struct mlxsw_sp_nexthop_group *nh_group = fib_entry->nh_group;

1884 1885 1886
	if (fib_entry->params.tos)
		return false;

1887 1888 1889 1890
	switch (fib_entry->type) {
	case MLXSW_SP_FIB_ENTRY_TYPE_REMOTE:
		return !!nh_group->adj_index_valid;
	case MLXSW_SP_FIB_ENTRY_TYPE_LOCAL:
1891
		return !!nh_group->nh_rif;
1892 1893 1894 1895 1896 1897 1898 1899 1900
	default:
		return false;
	}
}

static void mlxsw_sp_fib_entry_offload_set(struct mlxsw_sp_fib_entry *fib_entry)
{
	fib_entry->offloaded = true;

1901
	switch (fib_entry->fib_node->fib->proto) {
1902 1903 1904 1905 1906 1907 1908 1909 1910 1911 1912
	case MLXSW_SP_L3_PROTO_IPV4:
		fib_info_offload_inc(fib_entry->nh_group->key.fi);
		break;
	case MLXSW_SP_L3_PROTO_IPV6:
		WARN_ON_ONCE(1);
	}
}

static void
mlxsw_sp_fib_entry_offload_unset(struct mlxsw_sp_fib_entry *fib_entry)
{
1913
	switch (fib_entry->fib_node->fib->proto) {
1914 1915 1916 1917 1918 1919 1920 1921 1922 1923 1924 1925 1926 1927 1928 1929 1930 1931 1932 1933 1934 1935 1936 1937 1938 1939 1940 1941 1942 1943 1944 1945 1946 1947
	case MLXSW_SP_L3_PROTO_IPV4:
		fib_info_offload_dec(fib_entry->nh_group->key.fi);
		break;
	case MLXSW_SP_L3_PROTO_IPV6:
		WARN_ON_ONCE(1);
	}

	fib_entry->offloaded = false;
}

static void
mlxsw_sp_fib_entry_offload_refresh(struct mlxsw_sp_fib_entry *fib_entry,
				   enum mlxsw_reg_ralue_op op, int err)
{
	switch (op) {
	case MLXSW_REG_RALUE_OP_WRITE_DELETE:
		if (!fib_entry->offloaded)
			return;
		return mlxsw_sp_fib_entry_offload_unset(fib_entry);
	case MLXSW_REG_RALUE_OP_WRITE_WRITE:
		if (err)
			return;
		if (mlxsw_sp_fib_entry_should_offload(fib_entry) &&
		    !fib_entry->offloaded)
			mlxsw_sp_fib_entry_offload_set(fib_entry);
		else if (!mlxsw_sp_fib_entry_should_offload(fib_entry) &&
			 fib_entry->offloaded)
			mlxsw_sp_fib_entry_offload_unset(fib_entry);
		return;
	default:
		return;
	}
}

1948 1949 1950 1951 1952
static int mlxsw_sp_fib_entry_op4_remote(struct mlxsw_sp *mlxsw_sp,
					 struct mlxsw_sp_fib_entry *fib_entry,
					 enum mlxsw_reg_ralue_op op)
{
	char ralue_pl[MLXSW_REG_RALUE_LEN];
1953
	struct mlxsw_sp_fib *fib = fib_entry->fib_node->fib;
1954
	u32 *p_dip = (u32 *) fib_entry->fib_node->key.addr;
1955 1956 1957 1958 1959 1960 1961 1962 1963
	enum mlxsw_reg_ralue_trap_action trap_action;
	u16 trap_id = 0;
	u32 adjacency_index = 0;
	u16 ecmp_size = 0;

	/* In case the nexthop group adjacency index is valid, use it
	 * with provided ECMP size. Otherwise, setup trap and pass
	 * traffic to kernel.
	 */
1964
	if (mlxsw_sp_fib_entry_should_offload(fib_entry)) {
1965 1966 1967 1968 1969 1970 1971 1972
		trap_action = MLXSW_REG_RALUE_TRAP_ACTION_NOP;
		adjacency_index = fib_entry->nh_group->adj_index;
		ecmp_size = fib_entry->nh_group->ecmp_size;
	} else {
		trap_action = MLXSW_REG_RALUE_TRAP_ACTION_TRAP;
		trap_id = MLXSW_TRAP_ID_RTR_INGRESS0;
	}

1973
	mlxsw_reg_ralue_pack4(ralue_pl,
1974 1975
			      (enum mlxsw_reg_ralxx_protocol) fib->proto, op,
			      fib->vr->id, fib_entry->fib_node->key.prefix_len,
1976
			      *p_dip);
1977 1978 1979 1980 1981
	mlxsw_reg_ralue_act_remote_pack(ralue_pl, trap_action, trap_id,
					adjacency_index, ecmp_size);
	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralue), ralue_pl);
}

1982 1983 1984 1985
static int mlxsw_sp_fib_entry_op4_local(struct mlxsw_sp *mlxsw_sp,
					struct mlxsw_sp_fib_entry *fib_entry,
					enum mlxsw_reg_ralue_op op)
{
1986
	struct mlxsw_sp_rif *rif = fib_entry->nh_group->nh_rif;
1987
	struct mlxsw_sp_fib *fib = fib_entry->fib_node->fib;
1988
	enum mlxsw_reg_ralue_trap_action trap_action;
1989
	char ralue_pl[MLXSW_REG_RALUE_LEN];
1990
	u32 *p_dip = (u32 *) fib_entry->fib_node->key.addr;
1991
	u16 trap_id = 0;
1992
	u16 rif_index = 0;
1993 1994 1995

	if (mlxsw_sp_fib_entry_should_offload(fib_entry)) {
		trap_action = MLXSW_REG_RALUE_TRAP_ACTION_NOP;
1996
		rif_index = rif->rif_index;
1997 1998 1999 2000
	} else {
		trap_action = MLXSW_REG_RALUE_TRAP_ACTION_TRAP;
		trap_id = MLXSW_TRAP_ID_RTR_INGRESS0;
	}
2001

2002
	mlxsw_reg_ralue_pack4(ralue_pl,
2003 2004
			      (enum mlxsw_reg_ralxx_protocol) fib->proto, op,
			      fib->vr->id, fib_entry->fib_node->key.prefix_len,
2005
			      *p_dip);
2006 2007
	mlxsw_reg_ralue_act_local_pack(ralue_pl, trap_action, trap_id,
				       rif_index);
2008 2009 2010 2011 2012 2013 2014
	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralue), ralue_pl);
}

static int mlxsw_sp_fib_entry_op4_trap(struct mlxsw_sp *mlxsw_sp,
				       struct mlxsw_sp_fib_entry *fib_entry,
				       enum mlxsw_reg_ralue_op op)
{
2015
	struct mlxsw_sp_fib *fib = fib_entry->fib_node->fib;
2016
	char ralue_pl[MLXSW_REG_RALUE_LEN];
2017
	u32 *p_dip = (u32 *) fib_entry->fib_node->key.addr;
2018

2019
	mlxsw_reg_ralue_pack4(ralue_pl,
2020 2021
			      (enum mlxsw_reg_ralxx_protocol) fib->proto, op,
			      fib->vr->id, fib_entry->fib_node->key.prefix_len,
2022
			      *p_dip);
2023 2024 2025 2026 2027 2028 2029 2030 2031 2032
	mlxsw_reg_ralue_act_ip2me_pack(ralue_pl);
	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralue), ralue_pl);
}

static int mlxsw_sp_fib_entry_op4(struct mlxsw_sp *mlxsw_sp,
				  struct mlxsw_sp_fib_entry *fib_entry,
				  enum mlxsw_reg_ralue_op op)
{
	switch (fib_entry->type) {
	case MLXSW_SP_FIB_ENTRY_TYPE_REMOTE:
2033
		return mlxsw_sp_fib_entry_op4_remote(mlxsw_sp, fib_entry, op);
2034 2035 2036 2037 2038 2039 2040 2041 2042 2043 2044 2045
	case MLXSW_SP_FIB_ENTRY_TYPE_LOCAL:
		return mlxsw_sp_fib_entry_op4_local(mlxsw_sp, fib_entry, op);
	case MLXSW_SP_FIB_ENTRY_TYPE_TRAP:
		return mlxsw_sp_fib_entry_op4_trap(mlxsw_sp, fib_entry, op);
	}
	return -EINVAL;
}

static int mlxsw_sp_fib_entry_op(struct mlxsw_sp *mlxsw_sp,
				 struct mlxsw_sp_fib_entry *fib_entry,
				 enum mlxsw_reg_ralue_op op)
{
2046 2047
	int err = -EINVAL;

2048
	switch (fib_entry->fib_node->fib->proto) {
2049
	case MLXSW_SP_L3_PROTO_IPV4:
2050 2051
		err = mlxsw_sp_fib_entry_op4(mlxsw_sp, fib_entry, op);
		break;
2052
	case MLXSW_SP_L3_PROTO_IPV6:
2053
		return err;
2054
	}
2055 2056
	mlxsw_sp_fib_entry_offload_refresh(fib_entry, op, err);
	return err;
2057 2058 2059 2060 2061
}

static int mlxsw_sp_fib_entry_update(struct mlxsw_sp *mlxsw_sp,
				     struct mlxsw_sp_fib_entry *fib_entry)
{
2062 2063
	return mlxsw_sp_fib_entry_op(mlxsw_sp, fib_entry,
				     MLXSW_REG_RALUE_OP_WRITE_WRITE);
2064 2065 2066 2067 2068 2069 2070 2071 2072 2073
}

static int mlxsw_sp_fib_entry_del(struct mlxsw_sp *mlxsw_sp,
				  struct mlxsw_sp_fib_entry *fib_entry)
{
	return mlxsw_sp_fib_entry_op(mlxsw_sp, fib_entry,
				     MLXSW_REG_RALUE_OP_WRITE_DELETE);
}

static int
2074 2075 2076
mlxsw_sp_fib4_entry_type_set(struct mlxsw_sp *mlxsw_sp,
			     const struct fib_entry_notifier_info *fen_info,
			     struct mlxsw_sp_fib_entry *fib_entry)
2077
{
2078
	struct fib_info *fi = fen_info->fi;
2079

2080 2081 2082
	switch (fen_info->type) {
	case RTN_BROADCAST: /* fall through */
	case RTN_LOCAL:
2083 2084
		fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_TRAP;
		return 0;
2085 2086 2087 2088 2089 2090 2091
	case RTN_UNREACHABLE: /* fall through */
	case RTN_BLACKHOLE: /* fall through */
	case RTN_PROHIBIT:
		/* Packets hitting these routes need to be trapped, but
		 * can do so with a lower priority than packets directed
		 * at the host, so use action type local instead of trap.
		 */
2092
		fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_LOCAL;
2093 2094 2095 2096 2097 2098 2099 2100 2101 2102
		return 0;
	case RTN_UNICAST:
		if (fi->fib_nh->nh_scope != RT_SCOPE_LINK)
			fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_LOCAL;
		else
			fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_REMOTE;
		return 0;
	default:
		return -EINVAL;
	}
2103 2104
}

2105
static struct mlxsw_sp_fib_entry *
2106 2107 2108
mlxsw_sp_fib4_entry_create(struct mlxsw_sp *mlxsw_sp,
			   struct mlxsw_sp_fib_node *fib_node,
			   const struct fib_entry_notifier_info *fen_info)
2109 2110 2111 2112
{
	struct mlxsw_sp_fib_entry *fib_entry;
	int err;

2113
	fib_entry = kzalloc(sizeof(*fib_entry), GFP_KERNEL);
2114 2115
	if (!fib_entry) {
		err = -ENOMEM;
2116
		goto err_fib_entry_alloc;
2117 2118
	}

2119
	err = mlxsw_sp_fib4_entry_type_set(mlxsw_sp, fen_info, fib_entry);
2120
	if (err)
2121
		goto err_fib4_entry_type_set;
2122

2123
	err = mlxsw_sp_nexthop_group_get(mlxsw_sp, fib_entry, fen_info->fi);
2124 2125 2126
	if (err)
		goto err_nexthop_group_get;

2127 2128 2129 2130 2131 2132 2133
	fib_entry->params.prio = fen_info->fi->fib_priority;
	fib_entry->params.tb_id = fen_info->tb_id;
	fib_entry->params.type = fen_info->type;
	fib_entry->params.tos = fen_info->tos;

	fib_entry->fib_node = fib_node;

2134 2135
	return fib_entry;

2136
err_nexthop_group_get:
2137
err_fib4_entry_type_set:
2138 2139
	kfree(fib_entry);
err_fib_entry_alloc:
2140 2141 2142
	return ERR_PTR(err);
}

2143 2144 2145 2146 2147 2148 2149 2150 2151 2152 2153
static void mlxsw_sp_fib4_entry_destroy(struct mlxsw_sp *mlxsw_sp,
					struct mlxsw_sp_fib_entry *fib_entry)
{
	mlxsw_sp_nexthop_group_put(mlxsw_sp, fib_entry);
	kfree(fib_entry);
}

static struct mlxsw_sp_fib_node *
mlxsw_sp_fib4_node_get(struct mlxsw_sp *mlxsw_sp,
		       const struct fib_entry_notifier_info *fen_info);

2154
static struct mlxsw_sp_fib_entry *
2155 2156
mlxsw_sp_fib4_entry_lookup(struct mlxsw_sp *mlxsw_sp,
			   const struct fib_entry_notifier_info *fen_info)
2157
{
2158 2159
	struct mlxsw_sp_fib_entry *fib_entry;
	struct mlxsw_sp_fib_node *fib_node;
2160

2161 2162 2163 2164 2165 2166 2167 2168 2169 2170 2171 2172 2173 2174 2175 2176 2177 2178 2179 2180 2181 2182 2183 2184 2185 2186 2187 2188 2189 2190 2191 2192 2193 2194 2195 2196 2197 2198 2199 2200 2201 2202 2203 2204 2205 2206 2207 2208 2209 2210
	fib_node = mlxsw_sp_fib4_node_get(mlxsw_sp, fen_info);
	if (IS_ERR(fib_node))
		return NULL;

	list_for_each_entry(fib_entry, &fib_node->entry_list, list) {
		if (fib_entry->params.tb_id == fen_info->tb_id &&
		    fib_entry->params.tos == fen_info->tos &&
		    fib_entry->params.type == fen_info->type &&
		    fib_entry->nh_group->key.fi == fen_info->fi) {
			return fib_entry;
		}
	}

	return NULL;
}

static const struct rhashtable_params mlxsw_sp_fib_ht_params = {
	.key_offset = offsetof(struct mlxsw_sp_fib_node, key),
	.head_offset = offsetof(struct mlxsw_sp_fib_node, ht_node),
	.key_len = sizeof(struct mlxsw_sp_fib_key),
	.automatic_shrinking = true,
};

static int mlxsw_sp_fib_node_insert(struct mlxsw_sp_fib *fib,
				    struct mlxsw_sp_fib_node *fib_node)
{
	return rhashtable_insert_fast(&fib->ht, &fib_node->ht_node,
				      mlxsw_sp_fib_ht_params);
}

static void mlxsw_sp_fib_node_remove(struct mlxsw_sp_fib *fib,
				     struct mlxsw_sp_fib_node *fib_node)
{
	rhashtable_remove_fast(&fib->ht, &fib_node->ht_node,
			       mlxsw_sp_fib_ht_params);
}

static struct mlxsw_sp_fib_node *
mlxsw_sp_fib_node_lookup(struct mlxsw_sp_fib *fib, const void *addr,
			 size_t addr_len, unsigned char prefix_len)
{
	struct mlxsw_sp_fib_key key;

	memset(&key, 0, sizeof(key));
	memcpy(key.addr, addr, addr_len);
	key.prefix_len = prefix_len;
	return rhashtable_lookup_fast(&fib->ht, &key, mlxsw_sp_fib_ht_params);
}

static struct mlxsw_sp_fib_node *
2211
mlxsw_sp_fib_node_create(struct mlxsw_sp_fib *fib, const void *addr,
2212 2213 2214 2215 2216 2217
			 size_t addr_len, unsigned char prefix_len)
{
	struct mlxsw_sp_fib_node *fib_node;

	fib_node = kzalloc(sizeof(*fib_node), GFP_KERNEL);
	if (!fib_node)
2218 2219
		return NULL;

2220
	INIT_LIST_HEAD(&fib_node->entry_list);
2221
	list_add(&fib_node->list, &fib->node_list);
2222 2223 2224 2225 2226 2227 2228 2229 2230 2231 2232 2233 2234 2235 2236 2237 2238 2239 2240 2241 2242 2243 2244 2245
	memcpy(fib_node->key.addr, addr, addr_len);
	fib_node->key.prefix_len = prefix_len;

	return fib_node;
}

static void mlxsw_sp_fib_node_destroy(struct mlxsw_sp_fib_node *fib_node)
{
	list_del(&fib_node->list);
	WARN_ON(!list_empty(&fib_node->entry_list));
	kfree(fib_node);
}

static bool
mlxsw_sp_fib_node_entry_is_first(const struct mlxsw_sp_fib_node *fib_node,
				 const struct mlxsw_sp_fib_entry *fib_entry)
{
	return list_first_entry(&fib_node->entry_list,
				struct mlxsw_sp_fib_entry, list) == fib_entry;
}

static void mlxsw_sp_fib_node_prefix_inc(struct mlxsw_sp_fib_node *fib_node)
{
	unsigned char prefix_len = fib_node->key.prefix_len;
2246
	struct mlxsw_sp_fib *fib = fib_node->fib;
2247 2248 2249 2250 2251 2252 2253 2254

	if (fib->prefix_ref_count[prefix_len]++ == 0)
		mlxsw_sp_prefix_usage_set(&fib->prefix_usage, prefix_len);
}

static void mlxsw_sp_fib_node_prefix_dec(struct mlxsw_sp_fib_node *fib_node)
{
	unsigned char prefix_len = fib_node->key.prefix_len;
2255
	struct mlxsw_sp_fib *fib = fib_node->fib;
2256 2257 2258

	if (--fib->prefix_ref_count[prefix_len] == 0)
		mlxsw_sp_prefix_usage_clear(&fib->prefix_usage, prefix_len);
2259 2260
}

2261 2262 2263 2264 2265 2266 2267 2268 2269 2270 2271 2272 2273 2274 2275 2276 2277 2278 2279 2280 2281 2282 2283 2284 2285 2286 2287 2288 2289 2290 2291 2292 2293 2294 2295 2296 2297 2298 2299 2300 2301 2302 2303 2304 2305 2306 2307 2308 2309 2310 2311 2312 2313 2314 2315 2316 2317 2318 2319 2320 2321 2322 2323 2324 2325
static int mlxsw_sp_fib_node_init(struct mlxsw_sp *mlxsw_sp,
				  struct mlxsw_sp_fib_node *fib_node,
				  struct mlxsw_sp_fib *fib)
{
	struct mlxsw_sp_prefix_usage req_prefix_usage;
	struct mlxsw_sp_lpm_tree *lpm_tree;
	int err;

	err = mlxsw_sp_fib_node_insert(fib, fib_node);
	if (err)
		return err;
	fib_node->fib = fib;

	mlxsw_sp_prefix_usage_cpy(&req_prefix_usage, &fib->prefix_usage);
	mlxsw_sp_prefix_usage_set(&req_prefix_usage, fib_node->key.prefix_len);

	if (!mlxsw_sp_prefix_usage_none(&fib->prefix_usage)) {
		err = mlxsw_sp_vr_lpm_tree_check(mlxsw_sp, fib,
						 &req_prefix_usage);
		if (err)
			goto err_tree_check;
	} else {
		lpm_tree = mlxsw_sp_lpm_tree_get(mlxsw_sp, &req_prefix_usage,
						 fib->proto);
		if (IS_ERR(lpm_tree))
			return PTR_ERR(lpm_tree);
		fib->lpm_tree = lpm_tree;
		err = mlxsw_sp_vr_lpm_tree_bind(mlxsw_sp, fib);
		if (err)
			goto err_tree_bind;
	}

	mlxsw_sp_fib_node_prefix_inc(fib_node);

	return 0;

err_tree_bind:
	fib->lpm_tree = NULL;
	mlxsw_sp_lpm_tree_put(mlxsw_sp, lpm_tree);
err_tree_check:
	fib_node->fib = NULL;
	mlxsw_sp_fib_node_remove(fib, fib_node);
	return err;
}

static void mlxsw_sp_fib_node_fini(struct mlxsw_sp *mlxsw_sp,
				   struct mlxsw_sp_fib_node *fib_node)
{
	struct mlxsw_sp_lpm_tree *lpm_tree = fib_node->fib->lpm_tree;
	struct mlxsw_sp_fib *fib = fib_node->fib;

	mlxsw_sp_fib_node_prefix_dec(fib_node);

	if (mlxsw_sp_prefix_usage_none(&fib->prefix_usage)) {
		mlxsw_sp_vr_lpm_tree_unbind(mlxsw_sp, fib);
		fib->lpm_tree = NULL;
		mlxsw_sp_lpm_tree_put(mlxsw_sp, lpm_tree);
	} else {
		mlxsw_sp_vr_lpm_tree_check(mlxsw_sp, fib, &fib->prefix_usage);
	}

	fib_node->fib = NULL;
	mlxsw_sp_fib_node_remove(fib, fib_node);
}

2326 2327 2328
static struct mlxsw_sp_fib_node *
mlxsw_sp_fib4_node_get(struct mlxsw_sp *mlxsw_sp,
		       const struct fib_entry_notifier_info *fen_info)
2329
{
2330
	struct mlxsw_sp_fib_node *fib_node;
2331
	struct mlxsw_sp_fib *fib;
2332 2333 2334
	struct mlxsw_sp_vr *vr;
	int err;

2335
	vr = mlxsw_sp_vr_get(mlxsw_sp, fen_info->tb_id);
2336 2337
	if (IS_ERR(vr))
		return ERR_CAST(vr);
2338
	fib = mlxsw_sp_vr_fib(vr, MLXSW_SP_L3_PROTO_IPV4);
2339

2340
	fib_node = mlxsw_sp_fib_node_lookup(fib, &fen_info->dst,
2341 2342 2343 2344
					    sizeof(fen_info->dst),
					    fen_info->dst_len);
	if (fib_node)
		return fib_node;
2345

2346
	fib_node = mlxsw_sp_fib_node_create(fib, &fen_info->dst,
2347 2348 2349 2350 2351
					    sizeof(fen_info->dst),
					    fen_info->dst_len);
	if (!fib_node) {
		err = -ENOMEM;
		goto err_fib_node_create;
2352
	}
2353

2354 2355 2356 2357
	err = mlxsw_sp_fib_node_init(mlxsw_sp, fib_node, fib);
	if (err)
		goto err_fib_node_init;

2358 2359
	return fib_node;

2360 2361
err_fib_node_init:
	mlxsw_sp_fib_node_destroy(fib_node);
2362
err_fib_node_create:
2363
	mlxsw_sp_vr_put(vr);
2364
	return ERR_PTR(err);
2365 2366
}

2367 2368
static void mlxsw_sp_fib4_node_put(struct mlxsw_sp *mlxsw_sp,
				   struct mlxsw_sp_fib_node *fib_node)
2369
{
2370
	struct mlxsw_sp_vr *vr = fib_node->fib->vr;
2371

2372 2373
	if (!list_empty(&fib_node->entry_list))
		return;
2374
	mlxsw_sp_fib_node_fini(mlxsw_sp, fib_node);
2375
	mlxsw_sp_fib_node_destroy(fib_node);
2376
	mlxsw_sp_vr_put(vr);
2377 2378
}

2379 2380 2381
static struct mlxsw_sp_fib_entry *
mlxsw_sp_fib4_node_entry_find(const struct mlxsw_sp_fib_node *fib_node,
			      const struct mlxsw_sp_fib_entry_params *params)
2382 2383
{
	struct mlxsw_sp_fib_entry *fib_entry;
2384 2385 2386 2387 2388 2389 2390 2391 2392 2393 2394 2395 2396 2397 2398 2399

	list_for_each_entry(fib_entry, &fib_node->entry_list, list) {
		if (fib_entry->params.tb_id > params->tb_id)
			continue;
		if (fib_entry->params.tb_id != params->tb_id)
			break;
		if (fib_entry->params.tos > params->tos)
			continue;
		if (fib_entry->params.prio >= params->prio ||
		    fib_entry->params.tos < params->tos)
			return fib_entry;
	}

	return NULL;
}

2400 2401 2402 2403 2404 2405 2406 2407 2408 2409 2410 2411 2412 2413 2414 2415 2416 2417 2418 2419
static int mlxsw_sp_fib4_node_list_append(struct mlxsw_sp_fib_entry *fib_entry,
					  struct mlxsw_sp_fib_entry *new_entry)
{
	struct mlxsw_sp_fib_node *fib_node;

	if (WARN_ON(!fib_entry))
		return -EINVAL;

	fib_node = fib_entry->fib_node;
	list_for_each_entry_from(fib_entry, &fib_node->entry_list, list) {
		if (fib_entry->params.tb_id != new_entry->params.tb_id ||
		    fib_entry->params.tos != new_entry->params.tos ||
		    fib_entry->params.prio != new_entry->params.prio)
			break;
	}

	list_add_tail(&new_entry->list, &fib_entry->list);
	return 0;
}

2420 2421
static int
mlxsw_sp_fib4_node_list_insert(struct mlxsw_sp_fib_node *fib_node,
2422
			       struct mlxsw_sp_fib_entry *new_entry,
2423
			       bool replace, bool append)
2424 2425 2426 2427 2428
{
	struct mlxsw_sp_fib_entry *fib_entry;

	fib_entry = mlxsw_sp_fib4_node_entry_find(fib_node, &new_entry->params);

2429 2430
	if (append)
		return mlxsw_sp_fib4_node_list_append(fib_entry, new_entry);
2431 2432
	if (replace && WARN_ON(!fib_entry))
		return -EINVAL;
2433

2434 2435 2436
	/* Insert new entry before replaced one, so that we can later
	 * remove the second.
	 */
2437 2438 2439 2440 2441 2442 2443 2444 2445 2446 2447 2448 2449 2450 2451 2452 2453 2454 2455 2456 2457 2458 2459 2460 2461 2462 2463 2464 2465 2466 2467 2468 2469 2470 2471 2472 2473 2474 2475 2476 2477 2478 2479 2480 2481 2482 2483 2484 2485 2486 2487 2488 2489 2490 2491 2492 2493 2494 2495 2496 2497 2498 2499 2500 2501 2502 2503 2504 2505
	if (fib_entry) {
		list_add_tail(&new_entry->list, &fib_entry->list);
	} else {
		struct mlxsw_sp_fib_entry *last;

		list_for_each_entry(last, &fib_node->entry_list, list) {
			if (new_entry->params.tb_id > last->params.tb_id)
				break;
			fib_entry = last;
		}

		if (fib_entry)
			list_add(&new_entry->list, &fib_entry->list);
		else
			list_add(&new_entry->list, &fib_node->entry_list);
	}

	return 0;
}

static void
mlxsw_sp_fib4_node_list_remove(struct mlxsw_sp_fib_entry *fib_entry)
{
	list_del(&fib_entry->list);
}

static int
mlxsw_sp_fib4_node_entry_add(struct mlxsw_sp *mlxsw_sp,
			     const struct mlxsw_sp_fib_node *fib_node,
			     struct mlxsw_sp_fib_entry *fib_entry)
{
	if (!mlxsw_sp_fib_node_entry_is_first(fib_node, fib_entry))
		return 0;

	/* To prevent packet loss, overwrite the previously offloaded
	 * entry.
	 */
	if (!list_is_singular(&fib_node->entry_list)) {
		enum mlxsw_reg_ralue_op op = MLXSW_REG_RALUE_OP_WRITE_DELETE;
		struct mlxsw_sp_fib_entry *n = list_next_entry(fib_entry, list);

		mlxsw_sp_fib_entry_offload_refresh(n, op, 0);
	}

	return mlxsw_sp_fib_entry_update(mlxsw_sp, fib_entry);
}

static void
mlxsw_sp_fib4_node_entry_del(struct mlxsw_sp *mlxsw_sp,
			     const struct mlxsw_sp_fib_node *fib_node,
			     struct mlxsw_sp_fib_entry *fib_entry)
{
	if (!mlxsw_sp_fib_node_entry_is_first(fib_node, fib_entry))
		return;

	/* Promote the next entry by overwriting the deleted entry */
	if (!list_is_singular(&fib_node->entry_list)) {
		struct mlxsw_sp_fib_entry *n = list_next_entry(fib_entry, list);
		enum mlxsw_reg_ralue_op op = MLXSW_REG_RALUE_OP_WRITE_DELETE;

		mlxsw_sp_fib_entry_update(mlxsw_sp, n);
		mlxsw_sp_fib_entry_offload_refresh(fib_entry, op, 0);
		return;
	}

	mlxsw_sp_fib_entry_del(mlxsw_sp, fib_entry);
}

static int mlxsw_sp_fib4_node_entry_link(struct mlxsw_sp *mlxsw_sp,
2506
					 struct mlxsw_sp_fib_entry *fib_entry,
2507
					 bool replace, bool append)
2508 2509 2510 2511
{
	struct mlxsw_sp_fib_node *fib_node = fib_entry->fib_node;
	int err;

2512 2513
	err = mlxsw_sp_fib4_node_list_insert(fib_node, fib_entry, replace,
					     append);
2514 2515 2516 2517 2518 2519 2520 2521 2522 2523 2524 2525 2526 2527 2528 2529 2530 2531 2532 2533 2534 2535 2536 2537
	if (err)
		return err;

	err = mlxsw_sp_fib4_node_entry_add(mlxsw_sp, fib_node, fib_entry);
	if (err)
		goto err_fib4_node_entry_add;

	return 0;

err_fib4_node_entry_add:
	mlxsw_sp_fib4_node_list_remove(fib_entry);
	return err;
}

static void
mlxsw_sp_fib4_node_entry_unlink(struct mlxsw_sp *mlxsw_sp,
				struct mlxsw_sp_fib_entry *fib_entry)
{
	struct mlxsw_sp_fib_node *fib_node = fib_entry->fib_node;

	mlxsw_sp_fib4_node_entry_del(mlxsw_sp, fib_node, fib_entry);
	mlxsw_sp_fib4_node_list_remove(fib_entry);
}

2538 2539 2540 2541 2542 2543 2544 2545 2546 2547 2548 2549 2550 2551 2552 2553 2554 2555
static void mlxsw_sp_fib4_entry_replace(struct mlxsw_sp *mlxsw_sp,
					struct mlxsw_sp_fib_entry *fib_entry,
					bool replace)
{
	struct mlxsw_sp_fib_node *fib_node = fib_entry->fib_node;
	struct mlxsw_sp_fib_entry *replaced;

	if (!replace)
		return;

	/* We inserted the new entry before replaced one */
	replaced = list_next_entry(fib_entry, list);

	mlxsw_sp_fib4_node_entry_unlink(mlxsw_sp, replaced);
	mlxsw_sp_fib4_entry_destroy(mlxsw_sp, replaced);
	mlxsw_sp_fib4_node_put(mlxsw_sp, fib_node);
}

2556 2557
static int
mlxsw_sp_router_fib4_add(struct mlxsw_sp *mlxsw_sp,
2558
			 const struct fib_entry_notifier_info *fen_info,
2559
			 bool replace, bool append)
2560 2561 2562
{
	struct mlxsw_sp_fib_entry *fib_entry;
	struct mlxsw_sp_fib_node *fib_node;
2563 2564
	int err;

2565
	if (mlxsw_sp->router->aborted)
2566 2567
		return 0;

2568 2569 2570 2571
	fib_node = mlxsw_sp_fib4_node_get(mlxsw_sp, fen_info);
	if (IS_ERR(fib_node)) {
		dev_warn(mlxsw_sp->bus_info->dev, "Failed to get FIB node\n");
		return PTR_ERR(fib_node);
2572
	}
2573

2574 2575 2576 2577 2578 2579
	fib_entry = mlxsw_sp_fib4_entry_create(mlxsw_sp, fib_node, fen_info);
	if (IS_ERR(fib_entry)) {
		dev_warn(mlxsw_sp->bus_info->dev, "Failed to create FIB entry\n");
		err = PTR_ERR(fib_entry);
		goto err_fib4_entry_create;
	}
2580

2581 2582
	err = mlxsw_sp_fib4_node_entry_link(mlxsw_sp, fib_entry, replace,
					    append);
2583
	if (err) {
2584 2585
		dev_warn(mlxsw_sp->bus_info->dev, "Failed to link FIB entry to node\n");
		goto err_fib4_node_entry_link;
2586
	}
2587

2588 2589
	mlxsw_sp_fib4_entry_replace(mlxsw_sp, fib_entry, replace);

2590 2591
	return 0;

2592 2593 2594 2595
err_fib4_node_entry_link:
	mlxsw_sp_fib4_entry_destroy(mlxsw_sp, fib_entry);
err_fib4_entry_create:
	mlxsw_sp_fib4_node_put(mlxsw_sp, fib_node);
2596 2597 2598
	return err;
}

2599 2600
static void mlxsw_sp_router_fib4_del(struct mlxsw_sp *mlxsw_sp,
				     struct fib_entry_notifier_info *fen_info)
2601 2602
{
	struct mlxsw_sp_fib_entry *fib_entry;
2603
	struct mlxsw_sp_fib_node *fib_node;
2604

2605
	if (mlxsw_sp->router->aborted)
2606
		return;
2607

2608 2609
	fib_entry = mlxsw_sp_fib4_entry_lookup(mlxsw_sp, fen_info);
	if (WARN_ON(!fib_entry))
2610
		return;
2611
	fib_node = fib_entry->fib_node;
2612

2613 2614 2615
	mlxsw_sp_fib4_node_entry_unlink(mlxsw_sp, fib_entry);
	mlxsw_sp_fib4_entry_destroy(mlxsw_sp, fib_entry);
	mlxsw_sp_fib4_node_put(mlxsw_sp, fib_node);
2616
}
2617 2618 2619 2620 2621

static int mlxsw_sp_router_set_abort_trap(struct mlxsw_sp *mlxsw_sp)
{
	char ralta_pl[MLXSW_REG_RALTA_LEN];
	char ralst_pl[MLXSW_REG_RALST_LEN];
2622
	int i, err;
2623 2624 2625 2626 2627 2628 2629 2630 2631 2632 2633 2634

	mlxsw_reg_ralta_pack(ralta_pl, true, MLXSW_REG_RALXX_PROTOCOL_IPV4,
			     MLXSW_SP_LPM_TREE_MIN);
	err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralta), ralta_pl);
	if (err)
		return err;

	mlxsw_reg_ralst_pack(ralst_pl, 0xff, MLXSW_SP_LPM_TREE_MIN);
	err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralst), ralst_pl);
	if (err)
		return err;

2635
	for (i = 0; i < MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_VRS); i++) {
2636
		struct mlxsw_sp_vr *vr = &mlxsw_sp->router->vrs[i];
2637 2638
		char raltb_pl[MLXSW_REG_RALTB_LEN];
		char ralue_pl[MLXSW_REG_RALUE_LEN];
2639

2640 2641 2642 2643 2644 2645 2646 2647 2648 2649 2650 2651 2652 2653 2654 2655 2656 2657 2658 2659 2660 2661
		if (!mlxsw_sp_vr_is_used(vr))
			continue;

		mlxsw_reg_raltb_pack(raltb_pl, vr->id,
				     MLXSW_REG_RALXX_PROTOCOL_IPV4,
				     MLXSW_SP_LPM_TREE_MIN);
		err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(raltb),
				      raltb_pl);
		if (err)
			return err;

		mlxsw_reg_ralue_pack4(ralue_pl, MLXSW_SP_L3_PROTO_IPV4,
				      MLXSW_REG_RALUE_OP_WRITE_WRITE, vr->id, 0,
				      0);
		mlxsw_reg_ralue_act_ip2me_pack(ralue_pl);
		err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralue),
				      ralue_pl);
		if (err)
			return err;
	}

	return 0;
2662 2663
}

2664 2665 2666 2667 2668 2669 2670 2671 2672 2673 2674 2675 2676 2677 2678 2679 2680 2681 2682 2683 2684 2685 2686
static void mlxsw_sp_fib4_node_flush(struct mlxsw_sp *mlxsw_sp,
				     struct mlxsw_sp_fib_node *fib_node)
{
	struct mlxsw_sp_fib_entry *fib_entry, *tmp;

	list_for_each_entry_safe(fib_entry, tmp, &fib_node->entry_list, list) {
		bool do_break = &tmp->list == &fib_node->entry_list;

		mlxsw_sp_fib4_node_entry_unlink(mlxsw_sp, fib_entry);
		mlxsw_sp_fib4_entry_destroy(mlxsw_sp, fib_entry);
		mlxsw_sp_fib4_node_put(mlxsw_sp, fib_node);
		/* Break when entry list is empty and node was freed.
		 * Otherwise, we'll access freed memory in the next
		 * iteration.
		 */
		if (do_break)
			break;
	}
}

static void mlxsw_sp_fib_node_flush(struct mlxsw_sp *mlxsw_sp,
				    struct mlxsw_sp_fib_node *fib_node)
{
2687
	switch (fib_node->fib->proto) {
2688 2689 2690 2691 2692 2693 2694 2695 2696
	case MLXSW_SP_L3_PROTO_IPV4:
		mlxsw_sp_fib4_node_flush(mlxsw_sp, fib_node);
		break;
	case MLXSW_SP_L3_PROTO_IPV6:
		WARN_ON_ONCE(1);
		break;
	}
}

2697 2698 2699
static void mlxsw_sp_vr_fib_flush(struct mlxsw_sp *mlxsw_sp,
				  struct mlxsw_sp_vr *vr,
				  enum mlxsw_sp_l3proto proto)
2700
{
2701
	struct mlxsw_sp_fib *fib = mlxsw_sp_vr_fib(vr, proto);
2702
	struct mlxsw_sp_fib_node *fib_node, *tmp;
2703 2704 2705 2706 2707 2708 2709 2710 2711 2712 2713 2714

	list_for_each_entry_safe(fib_node, tmp, &fib->node_list, list) {
		bool do_break = &tmp->list == &fib->node_list;

		mlxsw_sp_fib_node_flush(mlxsw_sp, fib_node);
		if (do_break)
			break;
	}
}

static void mlxsw_sp_router_fib_flush(struct mlxsw_sp *mlxsw_sp)
{
2715 2716
	int i;

J
Jiri Pirko 已提交
2717
	for (i = 0; i < MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_VRS); i++) {
2718
		struct mlxsw_sp_vr *vr = &mlxsw_sp->router->vrs[i];
2719

2720
		if (!mlxsw_sp_vr_is_used(vr))
2721
			continue;
2722
		mlxsw_sp_vr_fib_flush(mlxsw_sp, vr, MLXSW_SP_L3_PROTO_IPV4);
2723
	}
2724 2725 2726 2727 2728 2729
}

static void mlxsw_sp_router_fib4_abort(struct mlxsw_sp *mlxsw_sp)
{
	int err;

2730
	if (mlxsw_sp->router->aborted)
2731 2732
		return;
	dev_warn(mlxsw_sp->bus_info->dev, "FIB abort triggered. Note that FIB entries are no longer being offloaded to this device.\n");
2733
	mlxsw_sp_router_fib_flush(mlxsw_sp);
2734
	mlxsw_sp->router->aborted = true;
2735 2736 2737 2738 2739
	err = mlxsw_sp_router_set_abort_trap(mlxsw_sp);
	if (err)
		dev_warn(mlxsw_sp->bus_info->dev, "Failed to set abort trap.\n");
}

2740
struct mlxsw_sp_fib_event_work {
2741
	struct work_struct work;
2742 2743
	union {
		struct fib_entry_notifier_info fen_info;
2744
		struct fib_rule_notifier_info fr_info;
2745 2746
		struct fib_nh_notifier_info fnh_info;
	};
2747 2748 2749 2750 2751
	struct mlxsw_sp *mlxsw_sp;
	unsigned long event;
};

static void mlxsw_sp_router_fib_event_work(struct work_struct *work)
2752
{
2753
	struct mlxsw_sp_fib_event_work *fib_work =
2754
		container_of(work, struct mlxsw_sp_fib_event_work, work);
2755
	struct mlxsw_sp *mlxsw_sp = fib_work->mlxsw_sp;
2756
	struct fib_rule *rule;
2757
	bool replace, append;
2758 2759
	int err;

2760 2761 2762
	/* Protect internal structures from changes */
	rtnl_lock();
	switch (fib_work->event) {
2763
	case FIB_EVENT_ENTRY_REPLACE: /* fall through */
2764
	case FIB_EVENT_ENTRY_APPEND: /* fall through */
2765
	case FIB_EVENT_ENTRY_ADD:
2766
		replace = fib_work->event == FIB_EVENT_ENTRY_REPLACE;
2767 2768
		append = fib_work->event == FIB_EVENT_ENTRY_APPEND;
		err = mlxsw_sp_router_fib4_add(mlxsw_sp, &fib_work->fen_info,
2769
					       replace, append);
2770 2771
		if (err)
			mlxsw_sp_router_fib4_abort(mlxsw_sp);
2772
		fib_info_put(fib_work->fen_info.fi);
2773 2774
		break;
	case FIB_EVENT_ENTRY_DEL:
2775 2776
		mlxsw_sp_router_fib4_del(mlxsw_sp, &fib_work->fen_info);
		fib_info_put(fib_work->fen_info.fi);
2777 2778 2779
		break;
	case FIB_EVENT_RULE_ADD: /* fall through */
	case FIB_EVENT_RULE_DEL:
2780
		rule = fib_work->fr_info.rule;
2781
		if (!fib4_rule_default(rule) && !rule->l3mdev)
2782 2783
			mlxsw_sp_router_fib4_abort(mlxsw_sp);
		fib_rule_put(rule);
2784
		break;
2785 2786 2787 2788 2789 2790
	case FIB_EVENT_NH_ADD: /* fall through */
	case FIB_EVENT_NH_DEL:
		mlxsw_sp_nexthop_event(mlxsw_sp, fib_work->event,
				       fib_work->fnh_info.fib_nh);
		fib_info_put(fib_work->fnh_info.fib_nh->nh_parent);
		break;
2791
	}
2792 2793 2794 2795 2796 2797 2798 2799 2800 2801 2802 2803 2804 2805 2806 2807 2808 2809 2810
	rtnl_unlock();
	kfree(fib_work);
}

/* Called with rcu_read_lock() */
static int mlxsw_sp_router_fib_event(struct notifier_block *nb,
				     unsigned long event, void *ptr)
{
	struct mlxsw_sp *mlxsw_sp = container_of(nb, struct mlxsw_sp, fib_nb);
	struct mlxsw_sp_fib_event_work *fib_work;
	struct fib_notifier_info *info = ptr;

	if (!net_eq(info->net, &init_net))
		return NOTIFY_DONE;

	fib_work = kzalloc(sizeof(*fib_work), GFP_ATOMIC);
	if (WARN_ON(!fib_work))
		return NOTIFY_BAD;

2811
	INIT_WORK(&fib_work->work, mlxsw_sp_router_fib_event_work);
2812 2813 2814 2815
	fib_work->mlxsw_sp = mlxsw_sp;
	fib_work->event = event;

	switch (event) {
2816
	case FIB_EVENT_ENTRY_REPLACE: /* fall through */
2817
	case FIB_EVENT_ENTRY_APPEND: /* fall through */
2818 2819 2820 2821 2822 2823 2824 2825
	case FIB_EVENT_ENTRY_ADD: /* fall through */
	case FIB_EVENT_ENTRY_DEL:
		memcpy(&fib_work->fen_info, ptr, sizeof(fib_work->fen_info));
		/* Take referece on fib_info to prevent it from being
		 * freed while work is queued. Release it afterwards.
		 */
		fib_info_hold(fib_work->fen_info.fi);
		break;
2826 2827 2828 2829 2830
	case FIB_EVENT_RULE_ADD: /* fall through */
	case FIB_EVENT_RULE_DEL:
		memcpy(&fib_work->fr_info, ptr, sizeof(fib_work->fr_info));
		fib_rule_get(fib_work->fr_info.rule);
		break;
2831 2832 2833 2834 2835
	case FIB_EVENT_NH_ADD: /* fall through */
	case FIB_EVENT_NH_DEL:
		memcpy(&fib_work->fnh_info, ptr, sizeof(fib_work->fnh_info));
		fib_info_hold(fib_work->fnh_info.fib_nh->nh_parent);
		break;
2836 2837
	}

2838
	mlxsw_core_schedule_work(&fib_work->work);
2839

2840 2841 2842
	return NOTIFY_DONE;
}

2843 2844 2845 2846 2847 2848 2849
static struct mlxsw_sp_rif *
mlxsw_sp_rif_find_by_dev(const struct mlxsw_sp *mlxsw_sp,
			 const struct net_device *dev)
{
	int i;

	for (i = 0; i < MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_RIFS); i++)
2850 2851 2852
		if (mlxsw_sp->router->rifs[i] &&
		    mlxsw_sp->router->rifs[i]->dev == dev)
			return mlxsw_sp->router->rifs[i];
2853 2854 2855 2856 2857 2858 2859 2860 2861 2862 2863 2864 2865 2866 2867 2868 2869 2870 2871

	return NULL;
}

static int mlxsw_sp_router_rif_disable(struct mlxsw_sp *mlxsw_sp, u16 rif)
{
	char ritr_pl[MLXSW_REG_RITR_LEN];
	int err;

	mlxsw_reg_ritr_rif_pack(ritr_pl, rif);
	err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
	if (WARN_ON_ONCE(err))
		return err;

	mlxsw_reg_ritr_enable_set(ritr_pl, false);
	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
}

static void mlxsw_sp_router_rif_gone_sync(struct mlxsw_sp *mlxsw_sp,
2872
					  struct mlxsw_sp_rif *rif)
2873
{
2874 2875 2876
	mlxsw_sp_router_rif_disable(mlxsw_sp, rif->rif_index);
	mlxsw_sp_nexthop_rif_gone_sync(mlxsw_sp, rif);
	mlxsw_sp_neigh_rif_gone_sync(mlxsw_sp, rif);
2877 2878
}

2879
static bool mlxsw_sp_rif_should_config(struct mlxsw_sp_rif *rif,
2880 2881 2882 2883 2884
				       const struct in_device *in_dev,
				       unsigned long event)
{
	switch (event) {
	case NETDEV_UP:
2885
		if (!rif)
2886 2887 2888
			return true;
		return false;
	case NETDEV_DOWN:
2889 2890
		if (rif && !in_dev->ifa_list &&
		    !netif_is_l3_slave(rif->dev))
2891 2892 2893 2894 2895 2896 2897 2898 2899 2900 2901
			return true;
		/* It is possible we already removed the RIF ourselves
		 * if it was assigned to a netdev that is now a bridge
		 * or LAG slave.
		 */
		return false;
	}

	return false;
}

2902
#define MLXSW_SP_INVALID_INDEX_RIF 0xffff
2903 2904 2905 2906 2907
static int mlxsw_sp_avail_rif_get(struct mlxsw_sp *mlxsw_sp)
{
	int i;

	for (i = 0; i < MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_RIFS); i++)
2908
		if (!mlxsw_sp->router->rifs[i])
2909 2910
			return i;

2911
	return MLXSW_SP_INVALID_INDEX_RIF;
2912 2913 2914 2915 2916 2917 2918 2919 2920 2921 2922 2923
}

static void mlxsw_sp_vport_rif_sp_attr_get(struct mlxsw_sp_port *mlxsw_sp_vport,
					   bool *p_lagged, u16 *p_system_port)
{
	u8 local_port = mlxsw_sp_vport->local_port;

	*p_lagged = mlxsw_sp_vport->lagged;
	*p_system_port = *p_lagged ? mlxsw_sp_vport->lag_id : local_port;
}

static int mlxsw_sp_vport_rif_sp_op(struct mlxsw_sp_port *mlxsw_sp_vport,
2924
				    u16 vr_id, struct net_device *l3_dev,
2925
				    u16 rif_index, bool create)
2926 2927 2928 2929 2930 2931
{
	struct mlxsw_sp *mlxsw_sp = mlxsw_sp_vport->mlxsw_sp;
	bool lagged = mlxsw_sp_vport->lagged;
	char ritr_pl[MLXSW_REG_RITR_LEN];
	u16 system_port;

2932 2933
	mlxsw_reg_ritr_pack(ritr_pl, create, MLXSW_REG_RITR_SP_IF, rif_index,
			    vr_id, l3_dev->mtu, l3_dev->dev_addr);
2934 2935 2936 2937 2938 2939 2940 2941 2942 2943

	mlxsw_sp_vport_rif_sp_attr_get(mlxsw_sp_vport, &lagged, &system_port);
	mlxsw_reg_ritr_sp_if_pack(ritr_pl, lagged, system_port,
				  mlxsw_sp_vport_vid_get(mlxsw_sp_vport));

	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
}

static void mlxsw_sp_vport_rif_sp_leave(struct mlxsw_sp_port *mlxsw_sp_vport);

2944
static u16 mlxsw_sp_rif_sp_to_fid(u16 rif_index)
2945
{
2946
	return MLXSW_SP_RFID_BASE + rif_index;
2947 2948 2949 2950 2951 2952 2953 2954 2955 2956 2957 2958 2959 2960 2961 2962 2963 2964 2965 2966
}

static struct mlxsw_sp_fid *
mlxsw_sp_rfid_alloc(u16 fid, struct net_device *l3_dev)
{
	struct mlxsw_sp_fid *f;

	f = kzalloc(sizeof(*f), GFP_KERNEL);
	if (!f)
		return NULL;

	f->leave = mlxsw_sp_vport_rif_sp_leave;
	f->ref_count = 0;
	f->dev = l3_dev;
	f->fid = fid;

	return f;
}

static struct mlxsw_sp_rif *
2967
mlxsw_sp_rif_alloc(u16 rif_index, u16 vr_id, struct net_device *l3_dev,
2968
		   struct mlxsw_sp_fid *f)
2969
{
2970
	struct mlxsw_sp_rif *rif;
2971

2972 2973
	rif = kzalloc(sizeof(*rif), GFP_KERNEL);
	if (!rif)
2974 2975
		return NULL;

2976 2977 2978 2979 2980 2981 2982 2983
	INIT_LIST_HEAD(&rif->nexthop_list);
	INIT_LIST_HEAD(&rif->neigh_list);
	ether_addr_copy(rif->addr, l3_dev->dev_addr);
	rif->mtu = l3_dev->mtu;
	rif->vr_id = vr_id;
	rif->dev = l3_dev;
	rif->rif_index = rif_index;
	rif->f = f;
2984

2985
	return rif;
2986 2987
}

2988 2989 2990 2991 2992 2993
struct mlxsw_sp_rif *mlxsw_sp_rif_by_index(const struct mlxsw_sp *mlxsw_sp,
					   u16 rif_index)
{
	return mlxsw_sp->router->rifs[rif_index];
}

2994 2995 2996 2997 2998 2999 3000 3001 3002 3003
u16 mlxsw_sp_rif_index(const struct mlxsw_sp_rif *rif)
{
	return rif->rif_index;
}

int mlxsw_sp_rif_dev_ifindex(const struct mlxsw_sp_rif *rif)
{
	return rif->dev->ifindex;
}

3004 3005 3006 3007 3008
static struct mlxsw_sp_rif *
mlxsw_sp_vport_rif_sp_create(struct mlxsw_sp_port *mlxsw_sp_vport,
			     struct net_device *l3_dev)
{
	struct mlxsw_sp *mlxsw_sp = mlxsw_sp_vport->mlxsw_sp;
3009
	u32 tb_id = l3mdev_fib_table(l3_dev);
3010
	struct mlxsw_sp_vr *vr;
3011
	struct mlxsw_sp_fid *f;
3012 3013
	struct mlxsw_sp_rif *rif;
	u16 fid, rif_index;
3014 3015
	int err;

3016 3017
	rif_index = mlxsw_sp_avail_rif_get(mlxsw_sp);
	if (rif_index == MLXSW_SP_INVALID_INDEX_RIF)
3018 3019
		return ERR_PTR(-ERANGE);

3020
	vr = mlxsw_sp_vr_get(mlxsw_sp, tb_id ? : RT_TABLE_MAIN);
3021 3022 3023
	if (IS_ERR(vr))
		return ERR_CAST(vr);

3024 3025
	err = mlxsw_sp_vport_rif_sp_op(mlxsw_sp_vport, vr->id, l3_dev,
				       rif_index, true);
3026
	if (err)
3027
		goto err_vport_rif_sp_op;
3028

3029
	fid = mlxsw_sp_rif_sp_to_fid(rif_index);
3030 3031 3032 3033 3034 3035 3036 3037 3038 3039
	err = mlxsw_sp_rif_fdb_op(mlxsw_sp, l3_dev->dev_addr, fid, true);
	if (err)
		goto err_rif_fdb_op;

	f = mlxsw_sp_rfid_alloc(fid, l3_dev);
	if (!f) {
		err = -ENOMEM;
		goto err_rfid_alloc;
	}

3040 3041
	rif = mlxsw_sp_rif_alloc(rif_index, vr->id, l3_dev, f);
	if (!rif) {
3042 3043 3044 3045
		err = -ENOMEM;
		goto err_rif_alloc;
	}

3046 3047 3048 3049 3050 3051 3052 3053 3054
	if (devlink_dpipe_table_counter_enabled(priv_to_devlink(mlxsw_sp->core),
						MLXSW_SP_DPIPE_TABLE_NAME_ERIF)) {
		err = mlxsw_sp_rif_counter_alloc(mlxsw_sp, rif,
						 MLXSW_SP_RIF_COUNTER_EGRESS);
		if (err)
			netdev_dbg(mlxsw_sp_vport->dev,
				   "Counter alloc Failed err=%d\n", err);
	}

3055
	f->rif = rif;
3056
	mlxsw_sp->router->rifs[rif_index] = rif;
3057
	vr->rif_count++;
3058

3059
	return rif;
3060 3061 3062 3063 3064 3065

err_rif_alloc:
	kfree(f);
err_rfid_alloc:
	mlxsw_sp_rif_fdb_op(mlxsw_sp, l3_dev->dev_addr, fid, false);
err_rif_fdb_op:
3066 3067
	mlxsw_sp_vport_rif_sp_op(mlxsw_sp_vport, vr->id, l3_dev, rif_index,
				 false);
3068 3069
err_vport_rif_sp_op:
	mlxsw_sp_vr_put(vr);
3070 3071 3072 3073
	return ERR_PTR(err);
}

static void mlxsw_sp_vport_rif_sp_destroy(struct mlxsw_sp_port *mlxsw_sp_vport,
3074
					  struct mlxsw_sp_rif *rif)
3075 3076
{
	struct mlxsw_sp *mlxsw_sp = mlxsw_sp_vport->mlxsw_sp;
3077
	struct mlxsw_sp_vr *vr = &mlxsw_sp->router->vrs[rif->vr_id];
3078 3079 3080
	struct net_device *l3_dev = rif->dev;
	struct mlxsw_sp_fid *f = rif->f;
	u16 rif_index = rif->rif_index;
3081 3082
	u16 fid = f->fid;

3083
	mlxsw_sp_router_rif_gone_sync(mlxsw_sp, rif);
3084

3085 3086 3087
	mlxsw_sp_rif_counter_free(mlxsw_sp, rif, MLXSW_SP_RIF_COUNTER_EGRESS);
	mlxsw_sp_rif_counter_free(mlxsw_sp, rif, MLXSW_SP_RIF_COUNTER_INGRESS);

3088
	vr->rif_count--;
3089
	mlxsw_sp->router->rifs[rif_index] = NULL;
3090
	f->rif = NULL;
3091

3092
	kfree(rif);
3093 3094 3095 3096 3097

	kfree(f);

	mlxsw_sp_rif_fdb_op(mlxsw_sp, l3_dev->dev_addr, fid, false);

3098 3099
	mlxsw_sp_vport_rif_sp_op(mlxsw_sp_vport, vr->id, l3_dev, rif_index,
				 false);
3100
	mlxsw_sp_vr_put(vr);
3101 3102 3103 3104 3105 3106
}

static int mlxsw_sp_vport_rif_sp_join(struct mlxsw_sp_port *mlxsw_sp_vport,
				      struct net_device *l3_dev)
{
	struct mlxsw_sp *mlxsw_sp = mlxsw_sp_vport->mlxsw_sp;
3107
	struct mlxsw_sp_rif *rif;
3108

3109 3110 3111 3112 3113
	rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, l3_dev);
	if (!rif) {
		rif = mlxsw_sp_vport_rif_sp_create(mlxsw_sp_vport, l3_dev);
		if (IS_ERR(rif))
			return PTR_ERR(rif);
3114 3115
	}

3116 3117
	mlxsw_sp_vport_fid_set(mlxsw_sp_vport, rif->f);
	rif->f->ref_count++;
3118

3119
	netdev_dbg(mlxsw_sp_vport->dev, "Joined FID=%d\n", rif->f->fid);
3120 3121 3122 3123 3124 3125 3126 3127 3128 3129 3130 3131

	return 0;
}

static void mlxsw_sp_vport_rif_sp_leave(struct mlxsw_sp_port *mlxsw_sp_vport)
{
	struct mlxsw_sp_fid *f = mlxsw_sp_vport_fid_get(mlxsw_sp_vport);

	netdev_dbg(mlxsw_sp_vport->dev, "Left FID=%d\n", f->fid);

	mlxsw_sp_vport_fid_set(mlxsw_sp_vport, NULL);
	if (--f->ref_count == 0)
3132
		mlxsw_sp_vport_rif_sp_destroy(mlxsw_sp_vport, f->rif);
3133 3134 3135 3136 3137 3138 3139 3140 3141 3142 3143 3144 3145 3146 3147 3148 3149 3150 3151 3152 3153 3154 3155 3156 3157 3158 3159
}

static int mlxsw_sp_inetaddr_vport_event(struct net_device *l3_dev,
					 struct net_device *port_dev,
					 unsigned long event, u16 vid)
{
	struct mlxsw_sp_port *mlxsw_sp_port = netdev_priv(port_dev);
	struct mlxsw_sp_port *mlxsw_sp_vport;

	mlxsw_sp_vport = mlxsw_sp_port_vport_find(mlxsw_sp_port, vid);
	if (WARN_ON(!mlxsw_sp_vport))
		return -EINVAL;

	switch (event) {
	case NETDEV_UP:
		return mlxsw_sp_vport_rif_sp_join(mlxsw_sp_vport, l3_dev);
	case NETDEV_DOWN:
		mlxsw_sp_vport_rif_sp_leave(mlxsw_sp_vport);
		break;
	}

	return 0;
}

static int mlxsw_sp_inetaddr_port_event(struct net_device *port_dev,
					unsigned long event)
{
3160 3161 3162
	if (netif_is_bridge_port(port_dev) ||
	    netif_is_lag_port(port_dev) ||
	    netif_is_ovs_port(port_dev))
3163 3164 3165 3166 3167 3168 3169 3170 3171 3172 3173 3174 3175 3176 3177 3178 3179 3180 3181 3182 3183 3184 3185 3186 3187 3188 3189 3190 3191 3192 3193 3194 3195 3196 3197 3198 3199 3200 3201 3202 3203
		return 0;

	return mlxsw_sp_inetaddr_vport_event(port_dev, port_dev, event, 1);
}

static int __mlxsw_sp_inetaddr_lag_event(struct net_device *l3_dev,
					 struct net_device *lag_dev,
					 unsigned long event, u16 vid)
{
	struct net_device *port_dev;
	struct list_head *iter;
	int err;

	netdev_for_each_lower_dev(lag_dev, port_dev, iter) {
		if (mlxsw_sp_port_dev_check(port_dev)) {
			err = mlxsw_sp_inetaddr_vport_event(l3_dev, port_dev,
							    event, vid);
			if (err)
				return err;
		}
	}

	return 0;
}

static int mlxsw_sp_inetaddr_lag_event(struct net_device *lag_dev,
				       unsigned long event)
{
	if (netif_is_bridge_port(lag_dev))
		return 0;

	return __mlxsw_sp_inetaddr_lag_event(lag_dev, lag_dev, event, 1);
}

static struct mlxsw_sp_fid *mlxsw_sp_bridge_fid_get(struct mlxsw_sp *mlxsw_sp,
						    struct net_device *l3_dev)
{
	u16 fid;

	if (is_vlan_dev(l3_dev))
		fid = vlan_dev_vlan_id(l3_dev);
3204
	else if (mlxsw_sp_master_bridge(mlxsw_sp)->dev == l3_dev)
3205 3206 3207 3208 3209 3210 3211
		fid = 1;
	else
		return mlxsw_sp_vfid_find(mlxsw_sp, l3_dev);

	return mlxsw_sp_fid_find(mlxsw_sp, fid);
}

3212 3213 3214 3215 3216
static u8 mlxsw_sp_router_port(const struct mlxsw_sp *mlxsw_sp)
{
	return mlxsw_core_max_ports(mlxsw_sp->core) + 1;
}

3217 3218 3219 3220 3221 3222 3223 3224 3225 3226 3227 3228 3229 3230
static enum mlxsw_flood_table_type mlxsw_sp_flood_table_type_get(u16 fid)
{
	return mlxsw_sp_fid_is_vfid(fid) ? MLXSW_REG_SFGC_TABLE_TYPE_FID :
	       MLXSW_REG_SFGC_TABLE_TYPE_FID_OFFEST;
}

static u16 mlxsw_sp_flood_table_index_get(u16 fid)
{
	return mlxsw_sp_fid_is_vfid(fid) ? mlxsw_sp_fid_to_vfid(fid) : fid;
}

static int mlxsw_sp_router_port_flood_set(struct mlxsw_sp *mlxsw_sp, u16 fid,
					  bool set)
{
3231
	u8 router_port = mlxsw_sp_router_port(mlxsw_sp);
3232 3233 3234 3235 3236 3237 3238 3239 3240 3241 3242 3243
	enum mlxsw_flood_table_type table_type;
	char *sftr_pl;
	u16 index;
	int err;

	sftr_pl = kmalloc(MLXSW_REG_SFTR_LEN, GFP_KERNEL);
	if (!sftr_pl)
		return -ENOMEM;

	table_type = mlxsw_sp_flood_table_type_get(fid);
	index = mlxsw_sp_flood_table_index_get(fid);
	mlxsw_reg_sftr_pack(sftr_pl, MLXSW_SP_FLOOD_TABLE_BC, index, table_type,
3244
			    1, router_port, set);
3245 3246 3247 3248 3249 3250 3251 3252 3253 3254 3255 3256 3257 3258
	err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(sftr), sftr_pl);

	kfree(sftr_pl);
	return err;
}

static enum mlxsw_reg_ritr_if_type mlxsw_sp_rif_type_get(u16 fid)
{
	if (mlxsw_sp_fid_is_vfid(fid))
		return MLXSW_REG_RITR_FID_IF;
	else
		return MLXSW_REG_RITR_VLAN_IF;
}

3259
static int mlxsw_sp_rif_bridge_op(struct mlxsw_sp *mlxsw_sp, u16 vr_id,
3260 3261 3262 3263 3264 3265 3266 3267
				  struct net_device *l3_dev,
				  u16 fid, u16 rif,
				  bool create)
{
	enum mlxsw_reg_ritr_if_type rif_type;
	char ritr_pl[MLXSW_REG_RITR_LEN];

	rif_type = mlxsw_sp_rif_type_get(fid);
3268
	mlxsw_reg_ritr_pack(ritr_pl, create, rif_type, rif, vr_id, l3_dev->mtu,
3269 3270 3271 3272 3273 3274 3275 3276 3277 3278
			    l3_dev->dev_addr);
	mlxsw_reg_ritr_fid_set(ritr_pl, rif_type, fid);

	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
}

static int mlxsw_sp_rif_bridge_create(struct mlxsw_sp *mlxsw_sp,
				      struct net_device *l3_dev,
				      struct mlxsw_sp_fid *f)
{
3279
	u32 tb_id = l3mdev_fib_table(l3_dev);
3280
	struct mlxsw_sp_rif *rif;
3281
	struct mlxsw_sp_vr *vr;
3282
	u16 rif_index;
3283 3284
	int err;

3285 3286
	rif_index = mlxsw_sp_avail_rif_get(mlxsw_sp);
	if (rif_index == MLXSW_SP_INVALID_INDEX_RIF)
3287 3288
		return -ERANGE;

3289
	vr = mlxsw_sp_vr_get(mlxsw_sp, tb_id ? : RT_TABLE_MAIN);
3290 3291 3292
	if (IS_ERR(vr))
		return PTR_ERR(vr);

3293 3294
	err = mlxsw_sp_router_port_flood_set(mlxsw_sp, f->fid, true);
	if (err)
3295
		goto err_port_flood_set;
3296

3297 3298
	err = mlxsw_sp_rif_bridge_op(mlxsw_sp, vr->id, l3_dev, f->fid,
				     rif_index, true);
3299 3300 3301 3302 3303 3304 3305
	if (err)
		goto err_rif_bridge_op;

	err = mlxsw_sp_rif_fdb_op(mlxsw_sp, l3_dev->dev_addr, f->fid, true);
	if (err)
		goto err_rif_fdb_op;

3306 3307
	rif = mlxsw_sp_rif_alloc(rif_index, vr->id, l3_dev, f);
	if (!rif) {
3308 3309 3310 3311
		err = -ENOMEM;
		goto err_rif_alloc;
	}

3312
	f->rif = rif;
3313
	mlxsw_sp->router->rifs[rif_index] = rif;
3314
	vr->rif_count++;
3315

3316
	netdev_dbg(l3_dev, "RIF=%d created\n", rif_index);
3317 3318 3319 3320 3321 3322

	return 0;

err_rif_alloc:
	mlxsw_sp_rif_fdb_op(mlxsw_sp, l3_dev->dev_addr, f->fid, false);
err_rif_fdb_op:
3323 3324
	mlxsw_sp_rif_bridge_op(mlxsw_sp, vr->id, l3_dev, f->fid, rif_index,
			       false);
3325 3326
err_rif_bridge_op:
	mlxsw_sp_router_port_flood_set(mlxsw_sp, f->fid, false);
3327 3328
err_port_flood_set:
	mlxsw_sp_vr_put(vr);
3329 3330 3331 3332
	return err;
}

void mlxsw_sp_rif_bridge_destroy(struct mlxsw_sp *mlxsw_sp,
3333
				 struct mlxsw_sp_rif *rif)
3334
{
3335
	struct mlxsw_sp_vr *vr = &mlxsw_sp->router->vrs[rif->vr_id];
3336 3337 3338
	struct net_device *l3_dev = rif->dev;
	struct mlxsw_sp_fid *f = rif->f;
	u16 rif_index = rif->rif_index;
3339

3340
	mlxsw_sp_router_rif_gone_sync(mlxsw_sp, rif);
3341

3342
	vr->rif_count--;
3343
	mlxsw_sp->router->rifs[rif_index] = NULL;
3344
	f->rif = NULL;
3345

3346
	kfree(rif);
3347 3348 3349

	mlxsw_sp_rif_fdb_op(mlxsw_sp, l3_dev->dev_addr, f->fid, false);

3350 3351
	mlxsw_sp_rif_bridge_op(mlxsw_sp, vr->id, l3_dev, f->fid, rif_index,
			       false);
3352 3353 3354

	mlxsw_sp_router_port_flood_set(mlxsw_sp, f->fid, false);

3355 3356
	mlxsw_sp_vr_put(vr);

3357
	netdev_dbg(l3_dev, "RIF=%d destroyed\n", rif_index);
3358 3359 3360 3361 3362 3363 3364 3365 3366 3367 3368 3369 3370 3371 3372 3373 3374 3375 3376 3377 3378
}

static int mlxsw_sp_inetaddr_bridge_event(struct net_device *l3_dev,
					  struct net_device *br_dev,
					  unsigned long event)
{
	struct mlxsw_sp *mlxsw_sp = mlxsw_sp_lower_get(l3_dev);
	struct mlxsw_sp_fid *f;

	/* FID can either be an actual FID if the L3 device is the
	 * VLAN-aware bridge or a VLAN device on top. Otherwise, the
	 * L3 device is a VLAN-unaware bridge and we get a vFID.
	 */
	f = mlxsw_sp_bridge_fid_get(mlxsw_sp, l3_dev);
	if (WARN_ON(!f))
		return -EINVAL;

	switch (event) {
	case NETDEV_UP:
		return mlxsw_sp_rif_bridge_create(mlxsw_sp, l3_dev, f);
	case NETDEV_DOWN:
3379
		mlxsw_sp_rif_bridge_destroy(mlxsw_sp, f->rif);
3380 3381 3382 3383 3384 3385 3386 3387 3388 3389 3390 3391 3392 3393 3394 3395 3396 3397 3398 3399
		break;
	}

	return 0;
}

static int mlxsw_sp_inetaddr_vlan_event(struct net_device *vlan_dev,
					unsigned long event)
{
	struct net_device *real_dev = vlan_dev_real_dev(vlan_dev);
	struct mlxsw_sp *mlxsw_sp = mlxsw_sp_lower_get(vlan_dev);
	u16 vid = vlan_dev_vlan_id(vlan_dev);

	if (mlxsw_sp_port_dev_check(real_dev))
		return mlxsw_sp_inetaddr_vport_event(vlan_dev, real_dev, event,
						     vid);
	else if (netif_is_lag_master(real_dev))
		return __mlxsw_sp_inetaddr_lag_event(vlan_dev, real_dev, event,
						     vid);
	else if (netif_is_bridge_master(real_dev) &&
3400
		 mlxsw_sp_master_bridge(mlxsw_sp)->dev == real_dev)
3401 3402 3403 3404 3405 3406
		return mlxsw_sp_inetaddr_bridge_event(vlan_dev, real_dev,
						      event);

	return 0;
}

3407 3408 3409 3410 3411 3412 3413 3414 3415 3416 3417 3418 3419 3420 3421
static int __mlxsw_sp_inetaddr_event(struct net_device *dev,
				     unsigned long event)
{
	if (mlxsw_sp_port_dev_check(dev))
		return mlxsw_sp_inetaddr_port_event(dev, event);
	else if (netif_is_lag_master(dev))
		return mlxsw_sp_inetaddr_lag_event(dev, event);
	else if (netif_is_bridge_master(dev))
		return mlxsw_sp_inetaddr_bridge_event(dev, dev, event);
	else if (is_vlan_dev(dev))
		return mlxsw_sp_inetaddr_vlan_event(dev, event);
	else
		return 0;
}

3422 3423 3424 3425 3426 3427
int mlxsw_sp_inetaddr_event(struct notifier_block *unused,
			    unsigned long event, void *ptr)
{
	struct in_ifaddr *ifa = (struct in_ifaddr *) ptr;
	struct net_device *dev = ifa->ifa_dev->dev;
	struct mlxsw_sp *mlxsw_sp;
3428
	struct mlxsw_sp_rif *rif;
3429 3430 3431 3432 3433 3434
	int err = 0;

	mlxsw_sp = mlxsw_sp_lower_get(dev);
	if (!mlxsw_sp)
		goto out;

3435 3436
	rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
	if (!mlxsw_sp_rif_should_config(rif, ifa->ifa_dev, event))
3437 3438
		goto out;

3439
	err = __mlxsw_sp_inetaddr_event(dev, event);
3440 3441 3442 3443
out:
	return notifier_from_errno(err);
}

3444
static int mlxsw_sp_rif_edit(struct mlxsw_sp *mlxsw_sp, u16 rif_index,
3445 3446 3447 3448 3449
			     const char *mac, int mtu)
{
	char ritr_pl[MLXSW_REG_RITR_LEN];
	int err;

3450
	mlxsw_reg_ritr_rif_pack(ritr_pl, rif_index);
3451 3452 3453 3454 3455 3456 3457 3458 3459 3460 3461 3462 3463
	err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
	if (err)
		return err;

	mlxsw_reg_ritr_mtu_set(ritr_pl, mtu);
	mlxsw_reg_ritr_if_mac_memcpy_to(ritr_pl, mac);
	mlxsw_reg_ritr_op_set(ritr_pl, MLXSW_REG_RITR_RIF_CREATE);
	return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
}

int mlxsw_sp_netdevice_router_port_event(struct net_device *dev)
{
	struct mlxsw_sp *mlxsw_sp;
3464
	struct mlxsw_sp_rif *rif;
3465 3466 3467 3468 3469 3470
	int err;

	mlxsw_sp = mlxsw_sp_lower_get(dev);
	if (!mlxsw_sp)
		return 0;

3471 3472
	rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
	if (!rif)
3473 3474
		return 0;

3475
	err = mlxsw_sp_rif_fdb_op(mlxsw_sp, rif->addr, rif->f->fid, false);
3476 3477 3478
	if (err)
		return err;

3479 3480
	err = mlxsw_sp_rif_edit(mlxsw_sp, rif->rif_index, dev->dev_addr,
				dev->mtu);
3481 3482 3483
	if (err)
		goto err_rif_edit;

3484
	err = mlxsw_sp_rif_fdb_op(mlxsw_sp, dev->dev_addr, rif->f->fid, true);
3485 3486 3487
	if (err)
		goto err_rif_fdb_op;

3488 3489
	ether_addr_copy(rif->addr, dev->dev_addr);
	rif->mtu = dev->mtu;
3490

3491
	netdev_dbg(dev, "Updated RIF=%d\n", rif->rif_index);
3492 3493 3494 3495

	return 0;

err_rif_fdb_op:
3496
	mlxsw_sp_rif_edit(mlxsw_sp, rif->rif_index, rif->addr, rif->mtu);
3497
err_rif_edit:
3498
	mlxsw_sp_rif_fdb_op(mlxsw_sp, rif->addr, rif->f->fid, true);
3499 3500 3501
	return err;
}

3502 3503
static int mlxsw_sp_port_vrf_join(struct mlxsw_sp *mlxsw_sp,
				  struct net_device *l3_dev)
3504
{
3505
	struct mlxsw_sp_rif *rif;
3506

3507 3508
	/* If netdev is already associated with a RIF, then we need to
	 * destroy it and create a new one with the new virtual router ID.
3509
	 */
3510 3511 3512
	rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, l3_dev);
	if (rif)
		__mlxsw_sp_inetaddr_event(l3_dev, NETDEV_DOWN);
3513

3514
	return __mlxsw_sp_inetaddr_event(l3_dev, NETDEV_UP);
3515 3516
}

3517 3518
static void mlxsw_sp_port_vrf_leave(struct mlxsw_sp *mlxsw_sp,
				    struct net_device *l3_dev)
3519
{
3520
	struct mlxsw_sp_rif *rif;
3521

3522 3523
	rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, l3_dev);
	if (!rif)
3524
		return;
3525
	__mlxsw_sp_inetaddr_event(l3_dev, NETDEV_DOWN);
3526 3527
}

3528 3529
int mlxsw_sp_netdevice_vrf_event(struct net_device *l3_dev, unsigned long event,
				 struct netdev_notifier_changeupper_info *info)
3530
{
3531 3532
	struct mlxsw_sp *mlxsw_sp = mlxsw_sp_lower_get(l3_dev);
	int err = 0;
3533

3534 3535
	if (!mlxsw_sp)
		return 0;
3536

3537 3538 3539 3540 3541 3542 3543 3544 3545 3546
	switch (event) {
	case NETDEV_PRECHANGEUPPER:
		return 0;
	case NETDEV_CHANGEUPPER:
		if (info->linking)
			err = mlxsw_sp_port_vrf_join(mlxsw_sp, l3_dev);
		else
			mlxsw_sp_port_vrf_leave(mlxsw_sp, l3_dev);
		break;
	}
3547

3548
	return err;
3549 3550
}

3551 3552 3553 3554 3555 3556 3557 3558 3559 3560 3561 3562
static void mlxsw_sp_router_fib_dump_flush(struct notifier_block *nb)
{
	struct mlxsw_sp *mlxsw_sp = container_of(nb, struct mlxsw_sp, fib_nb);

	/* Flush pending FIB notifications and then flush the device's
	 * table before requesting another dump. The FIB notification
	 * block is unregistered, so no need to take RTNL.
	 */
	mlxsw_core_flush_owq();
	mlxsw_sp_router_fib_flush(mlxsw_sp);
}

3563 3564 3565 3566 3567 3568 3569 3570 3571 3572
static int __mlxsw_sp_router_init(struct mlxsw_sp *mlxsw_sp)
{
	char rgcr_pl[MLXSW_REG_RGCR_LEN];
	u64 max_rifs;
	int err;

	if (!MLXSW_CORE_RES_VALID(mlxsw_sp->core, MAX_RIFS))
		return -EIO;

	max_rifs = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_RIFS);
3573 3574 3575 3576
	mlxsw_sp->router->rifs = kcalloc(max_rifs,
					 sizeof(struct mlxsw_sp_rif *),
					 GFP_KERNEL);
	if (!mlxsw_sp->router->rifs)
3577 3578 3579 3580 3581 3582 3583 3584 3585 3586 3587
		return -ENOMEM;

	mlxsw_reg_rgcr_pack(rgcr_pl, true);
	mlxsw_reg_rgcr_max_router_interfaces_set(rgcr_pl, max_rifs);
	err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rgcr), rgcr_pl);
	if (err)
		goto err_rgcr_fail;

	return 0;

err_rgcr_fail:
3588
	kfree(mlxsw_sp->router->rifs);
3589 3590 3591 3592 3593 3594 3595 3596 3597 3598 3599 3600
	return err;
}

static void __mlxsw_sp_router_fini(struct mlxsw_sp *mlxsw_sp)
{
	char rgcr_pl[MLXSW_REG_RGCR_LEN];
	int i;

	mlxsw_reg_rgcr_pack(rgcr_pl, false);
	mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rgcr), rgcr_pl);

	for (i = 0; i < MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_RIFS); i++)
3601
		WARN_ON_ONCE(mlxsw_sp->router->rifs[i]);
3602

3603
	kfree(mlxsw_sp->router->rifs);
3604 3605
}

3606 3607
int mlxsw_sp_router_init(struct mlxsw_sp *mlxsw_sp)
{
3608
	struct mlxsw_sp_router *router;
3609 3610
	int err;

3611 3612 3613 3614 3615 3616 3617
	router = kzalloc(sizeof(*mlxsw_sp->router), GFP_KERNEL);
	if (!router)
		return -ENOMEM;
	mlxsw_sp->router = router;
	router->mlxsw_sp = mlxsw_sp;

	INIT_LIST_HEAD(&mlxsw_sp->router->nexthop_neighs_list);
3618 3619
	err = __mlxsw_sp_router_init(mlxsw_sp);
	if (err)
3620
		goto err_router_init;
3621

3622
	err = rhashtable_init(&mlxsw_sp->router->nexthop_ht,
3623 3624 3625 3626
			      &mlxsw_sp_nexthop_ht_params);
	if (err)
		goto err_nexthop_ht_init;

3627
	err = rhashtable_init(&mlxsw_sp->router->nexthop_group_ht,
3628 3629 3630 3631
			      &mlxsw_sp_nexthop_group_ht_params);
	if (err)
		goto err_nexthop_group_ht_init;

3632 3633 3634 3635
	err = mlxsw_sp_lpm_init(mlxsw_sp);
	if (err)
		goto err_lpm_init;

3636 3637 3638 3639
	err = mlxsw_sp_vrs_init(mlxsw_sp);
	if (err)
		goto err_vrs_init;

3640
	err = mlxsw_sp_neigh_init(mlxsw_sp);
3641 3642 3643 3644
	if (err)
		goto err_neigh_init;

	mlxsw_sp->fib_nb.notifier_call = mlxsw_sp_router_fib_event;
3645 3646 3647 3648 3649
	err = register_fib_notifier(&mlxsw_sp->fib_nb,
				    mlxsw_sp_router_fib_dump_flush);
	if (err)
		goto err_register_fib_notifier;

3650 3651
	return 0;

3652 3653
err_register_fib_notifier:
	mlxsw_sp_neigh_fini(mlxsw_sp);
3654 3655 3656
err_neigh_init:
	mlxsw_sp_vrs_fini(mlxsw_sp);
err_vrs_init:
3657 3658
	mlxsw_sp_lpm_fini(mlxsw_sp);
err_lpm_init:
3659
	rhashtable_destroy(&mlxsw_sp->router->nexthop_group_ht);
3660
err_nexthop_group_ht_init:
3661
	rhashtable_destroy(&mlxsw_sp->router->nexthop_ht);
3662
err_nexthop_ht_init:
3663
	__mlxsw_sp_router_fini(mlxsw_sp);
3664 3665
err_router_init:
	kfree(mlxsw_sp->router);
3666 3667 3668 3669 3670 3671 3672 3673
	return err;
}

void mlxsw_sp_router_fini(struct mlxsw_sp *mlxsw_sp)
{
	unregister_fib_notifier(&mlxsw_sp->fib_nb);
	mlxsw_sp_neigh_fini(mlxsw_sp);
	mlxsw_sp_vrs_fini(mlxsw_sp);
3674
	mlxsw_sp_lpm_fini(mlxsw_sp);
3675 3676
	rhashtable_destroy(&mlxsw_sp->router->nexthop_group_ht);
	rhashtable_destroy(&mlxsw_sp->router->nexthop_ht);
3677
	__mlxsw_sp_router_fini(mlxsw_sp);
3678
	kfree(mlxsw_sp->router);
3679
}