ksz_common.c 11.2 KB
Newer Older
1
// SPDX-License-Identifier: GPL-2.0
2 3 4
/*
 * Microchip switch driver main logic
 *
5
 * Copyright (C) 2017-2019 Microchip Technology Inc.
6 7 8 9
 */

#include <linux/delay.h>
#include <linux/export.h>
10
#include <linux/gpio/consumer.h>
11 12 13 14 15 16
#include <linux/kernel.h>
#include <linux/module.h>
#include <linux/platform_data/microchip-ksz.h>
#include <linux/phy.h>
#include <linux/etherdevice.h>
#include <linux/if_bridge.h>
17
#include <linux/of_net.h>
18 19 20
#include <net/dsa.h>
#include <net/switchdev.h>

21
#include "ksz_common.h"
22

23
void ksz_update_port_member(struct ksz_device *dev, int port)
24
{
25 26 27 28
	struct ksz_port *p = &dev->ports[port];
	struct dsa_switch *ds = dev->ds;
	u8 port_member = 0, cpu_port;
	const struct dsa_port *dp;
29 30
	int i;

31 32 33 34 35 36 37 38 39 40 41 42
	if (!dsa_is_user_port(ds, port))
		return;

	dp = dsa_to_port(ds, port);
	cpu_port = BIT(dsa_upstream_port(ds, port));

	for (i = 0; i < ds->num_ports; i++) {
		const struct dsa_port *other_dp = dsa_to_port(ds, i);
		struct ksz_port *other_p = &dev->ports[i];
		u8 val = 0;

		if (!dsa_is_user_port(ds, i))
43
			continue;
44 45
		if (port == i)
			continue;
46
		if (!dsa_port_bridge_same(dp, other_dp))
47 48
			continue;

49 50 51 52 53 54 55
		if (other_p->stp_state == BR_STATE_FORWARDING &&
		    p->stp_state == BR_STATE_FORWARDING) {
			val |= BIT(port);
			port_member |= BIT(i);
		}

		dev->dev_ops->cfg_port_member(dev, i, val | cpu_port);
56
	}
57 58

	dev->dev_ops->cfg_port_member(dev, port, port_member | cpu_port);
59
}
60
EXPORT_SYMBOL_GPL(ksz_update_port_member);
61

62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88
static void port_r_cnt(struct ksz_device *dev, int port)
{
	struct ksz_port_mib *mib = &dev->ports[port].mib;
	u64 *dropped;

	/* Some ports may not have MIB counters before SWITCH_COUNTER_NUM. */
	while (mib->cnt_ptr < dev->reg_mib_cnt) {
		dev->dev_ops->r_mib_cnt(dev, port, mib->cnt_ptr,
					&mib->counters[mib->cnt_ptr]);
		++mib->cnt_ptr;
	}

	/* last one in storage */
	dropped = &mib->counters[dev->mib_cnt];

	/* Some ports may not have MIB counters after SWITCH_COUNTER_NUM. */
	while (mib->cnt_ptr < dev->mib_cnt) {
		dev->dev_ops->r_mib_pkt(dev, port, mib->cnt_ptr,
					dropped, &mib->counters[mib->cnt_ptr]);
		++mib->cnt_ptr;
	}
	mib->cnt_ptr = 0;
}

static void ksz_mib_read_work(struct work_struct *work)
{
	struct ksz_device *dev = container_of(work, struct ksz_device,
89
					      mib_read.work);
90 91 92 93
	struct ksz_port_mib *mib;
	struct ksz_port *p;
	int i;

94
	for (i = 0; i < dev->port_cnt; i++) {
95 96 97
		if (dsa_is_unused_port(dev->ds, i))
			continue;

98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115
		p = &dev->ports[i];
		mib = &p->mib;
		mutex_lock(&mib->cnt_mutex);

		/* Only read MIB counters when the port is told to do.
		 * If not, read only dropped counters when link is not up.
		 */
		if (!p->read) {
			const struct dsa_port *dp = dsa_to_port(dev->ds, i);

			if (!netif_carrier_ok(dp->slave))
				mib->cnt_ptr = dev->reg_mib_cnt;
		}
		port_r_cnt(dev, i);
		p->read = false;
		mutex_unlock(&mib->cnt_mutex);
	}

116
	schedule_delayed_work(&dev->mib_read, dev->mib_read_interval);
117 118 119 120 121 122
}

void ksz_init_mib_timer(struct ksz_device *dev)
{
	int i;

123 124
	INIT_DELAYED_WORK(&dev->mib_read, ksz_mib_read_work);

125
	for (i = 0; i < dev->port_cnt; i++)
126 127 128 129
		dev->dev_ops->port_init_cnt(dev, i);
}
EXPORT_SYMBOL_GPL(ksz_init_mib_timer);

130
int ksz_phy_read16(struct dsa_switch *ds, int addr, int reg)
131 132
{
	struct ksz_device *dev = ds->priv;
133
	u16 val = 0xffff;
134

135
	dev->dev_ops->r_phy(dev, addr, reg, &val);
136 137 138

	return val;
}
139
EXPORT_SYMBOL_GPL(ksz_phy_read16);
140

141
int ksz_phy_write16(struct dsa_switch *ds, int addr, int reg, u16 val)
142 143 144
{
	struct ksz_device *dev = ds->priv;

145
	dev->dev_ops->w_phy(dev, addr, reg, val);
146 147 148

	return 0;
}
149
EXPORT_SYMBOL_GPL(ksz_phy_write16);
150

151 152
void ksz_mac_link_down(struct dsa_switch *ds, int port, unsigned int mode,
		       phy_interface_t interface)
153 154 155 156 157
{
	struct ksz_device *dev = ds->priv;
	struct ksz_port *p = &dev->ports[port];

	/* Read all MIB counters when the link is going down. */
158
	p->read = true;
159 160 161
	/* timer started */
	if (dev->mib_read_interval)
		schedule_delayed_work(&dev->mib_read, 0);
162 163 164
}
EXPORT_SYMBOL_GPL(ksz_mac_link_down);

165
int ksz_sset_count(struct dsa_switch *ds, int port, int sset)
166 167 168
{
	struct ksz_device *dev = ds->priv;

169 170 171
	if (sset != ETH_SS_STATS)
		return 0;

172
	return dev->mib_cnt;
173
}
174
EXPORT_SYMBOL_GPL(ksz_sset_count);
175

176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193
void ksz_get_ethtool_stats(struct dsa_switch *ds, int port, uint64_t *buf)
{
	const struct dsa_port *dp = dsa_to_port(ds, port);
	struct ksz_device *dev = ds->priv;
	struct ksz_port_mib *mib;

	mib = &dev->ports[port].mib;
	mutex_lock(&mib->cnt_mutex);

	/* Only read dropped counters if no link. */
	if (!netif_carrier_ok(dp->slave))
		mib->cnt_ptr = dev->reg_mib_cnt;
	port_r_cnt(dev, port);
	memcpy(buf, mib->counters, dev->mib_cnt * sizeof(u64));
	mutex_unlock(&mib->cnt_mutex);
}
EXPORT_SYMBOL_GPL(ksz_get_ethtool_stats);

194
int ksz_port_bridge_join(struct dsa_switch *ds, int port,
195 196
			 struct dsa_bridge bridge,
			 bool *tx_fwd_offload)
197
{
198 199 200
	/* port_stp_state_set() will be called after to put the port in
	 * appropriate state so there is no need to do anything.
	 */
201

202
	return 0;
203
}
204
EXPORT_SYMBOL_GPL(ksz_port_bridge_join);
205

206
void ksz_port_bridge_leave(struct dsa_switch *ds, int port,
207
			   struct dsa_bridge bridge)
208
{
209 210 211
	/* port_stp_state_set() will be called after to put the port in
	 * forwarding state so there is no need to do anything.
	 */
212
}
213
EXPORT_SYMBOL_GPL(ksz_port_bridge_leave);
214

215
void ksz_port_fast_age(struct dsa_switch *ds, int port)
216 217 218
{
	struct ksz_device *dev = ds->priv;

219
	dev->dev_ops->flush_dyn_mac_table(dev, port);
220
}
221
EXPORT_SYMBOL_GPL(ksz_port_fast_age);
222

223 224
int ksz_port_fdb_dump(struct dsa_switch *ds, int port, dsa_fdb_dump_cb_t *cb,
		      void *data)
225 226
{
	struct ksz_device *dev = ds->priv;
227
	int ret = 0;
228 229 230 231 232
	u16 i = 0;
	u16 entries = 0;
	u8 timestamp = 0;
	u8 fid;
	u8 member;
233 234 235
	struct alu_struct alu;

	do {
236 237 238 239 240
		alu.is_static = false;
		ret = dev->dev_ops->r_dyn_mac_table(dev, i, alu.mac, &fid,
						    &member, &timestamp,
						    &entries);
		if (!ret && (member & BIT(port))) {
241
			ret = cb(alu.mac, alu.fid, alu.is_static, data);
242
			if (ret)
243
				break;
244
		}
245 246 247 248
		i++;
	} while (i < entries);
	if (i >= entries)
		ret = 0;
249 250 251

	return ret;
}
252
EXPORT_SYMBOL_GPL(ksz_port_fdb_dump);
253

254 255
int ksz_port_mdb_add(struct dsa_switch *ds, int port,
		     const struct switchdev_obj_port_mdb *mdb)
256 257
{
	struct ksz_device *dev = ds->priv;
258
	struct alu_struct alu;
259
	int index;
260
	int empty = 0;
261

262
	alu.port_forward = 0;
263
	for (index = 0; index < dev->num_statics; index++) {
264 265 266 267
		if (!dev->dev_ops->r_sta_mac_table(dev, index, &alu)) {
			/* Found one already in static MAC table. */
			if (!memcmp(alu.mac, mdb->addr, ETH_ALEN) &&
			    alu.fid == mdb->vid)
268
				break;
269 270 271
		/* Remember the first empty entry. */
		} else if (!empty) {
			empty = index + 1;
272 273 274 275
		}
	}

	/* no available entry */
276
	if (index == dev->num_statics && !empty)
277
		return -ENOSPC;
278 279

	/* add entry */
280 281 282 283 284 285 286 287 288
	if (index == dev->num_statics) {
		index = empty - 1;
		memset(&alu, 0, sizeof(alu));
		memcpy(alu.mac, mdb->addr, ETH_ALEN);
		alu.is_static = true;
	}
	alu.port_forward |= BIT(port);
	if (mdb->vid) {
		alu.is_use_fid = true;
289

290 291 292 293
		/* Need a way to map VID to FID. */
		alu.fid = mdb->vid;
	}
	dev->dev_ops->w_sta_mac_table(dev, index, &alu);
294 295

	return 0;
296
}
297
EXPORT_SYMBOL_GPL(ksz_port_mdb_add);
298

299 300
int ksz_port_mdb_del(struct dsa_switch *ds, int port,
		     const struct switchdev_obj_port_mdb *mdb)
301 302
{
	struct ksz_device *dev = ds->priv;
303
	struct alu_struct alu;
304 305 306
	int index;

	for (index = 0; index < dev->num_statics; index++) {
307 308 309 310
		if (!dev->dev_ops->r_sta_mac_table(dev, index, &alu)) {
			/* Found one already in static MAC table. */
			if (!memcmp(alu.mac, mdb->addr, ETH_ALEN) &&
			    alu.fid == mdb->vid)
311 312 313 314 315
				break;
		}
	}

	/* no available entry */
316
	if (index == dev->num_statics)
317 318 319
		goto exit;

	/* clear port */
320 321 322 323
	alu.port_forward &= ~BIT(port);
	if (!alu.port_forward)
		alu.is_static = false;
	dev->dev_ops->w_sta_mac_table(dev, index, &alu);
324 325

exit:
326
	return 0;
327
}
328
EXPORT_SYMBOL_GPL(ksz_port_mdb_del);
329

330
int ksz_enable_port(struct dsa_switch *ds, int port, struct phy_device *phy)
331 332 333
{
	struct ksz_device *dev = ds->priv;

334 335 336
	if (!dsa_is_user_port(ds, port))
		return 0;

337 338
	/* setup slave port */
	dev->dev_ops->port_setup(dev, port, false);
339

340 341 342
	/* port_stp_state_set() will be called after to enable the port so
	 * there is no need to do anything.
	 */
343 344 345

	return 0;
}
346
EXPORT_SYMBOL_GPL(ksz_enable_port);
347

348
struct ksz_device *ksz_switch_alloc(struct device *base, void *priv)
349 350 351 352
{
	struct dsa_switch *ds;
	struct ksz_device *swdev;

353
	ds = devm_kzalloc(base, sizeof(*ds), GFP_KERNEL);
354 355 356
	if (!ds)
		return NULL;

357 358 359
	ds->dev = base;
	ds->num_ports = DSA_MAX_PORTS;

360 361 362 363 364 365 366 367 368 369 370 371 372 373
	swdev = devm_kzalloc(base, sizeof(*swdev), GFP_KERNEL);
	if (!swdev)
		return NULL;

	ds->priv = swdev;
	swdev->dev = base;

	swdev->ds = ds;
	swdev->priv = priv;

	return swdev;
}
EXPORT_SYMBOL(ksz_switch_alloc);

374 375
int ksz_switch_register(struct ksz_device *dev,
			const struct ksz_dev_ops *ops)
376
{
377
	struct device_node *port, *ports;
378
	phy_interface_t interface;
379
	unsigned int port_num;
380 381 382 383 384
	int ret;

	if (dev->pdata)
		dev->chip_id = dev->pdata->chip_id;

385 386 387 388 389 390
	dev->reset_gpio = devm_gpiod_get_optional(dev->dev, "reset",
						  GPIOD_OUT_LOW);
	if (IS_ERR(dev->reset_gpio))
		return PTR_ERR(dev->reset_gpio);

	if (dev->reset_gpio) {
391
		gpiod_set_value_cansleep(dev->reset_gpio, 1);
392
		usleep_range(10000, 12000);
393
		gpiod_set_value_cansleep(dev->reset_gpio, 0);
394
		msleep(100);
395 396
	}

397
	mutex_init(&dev->dev_mutex);
398
	mutex_init(&dev->regmap_mutex);
399 400 401
	mutex_init(&dev->alu_mutex);
	mutex_init(&dev->vlan_mutex);

402 403 404
	dev->dev_ops = ops;

	if (dev->dev_ops->detect(dev))
405 406
		return -EINVAL;

407
	ret = dev->dev_ops->init(dev);
408 409 410
	if (ret)
		return ret;

411 412 413
	/* Host port interface will be self detected, or specifically set in
	 * device tree.
	 */
414 415
	for (port_num = 0; port_num < dev->port_cnt; ++port_num)
		dev->ports[port_num].interface = PHY_INTERFACE_MODE_NA;
416
	if (dev->dev->of_node) {
417 418
		ret = of_get_phy_mode(dev->dev->of_node, &interface);
		if (ret == 0)
419
			dev->compat_interface = interface;
420 421 422
		ports = of_get_child_by_name(dev->dev->of_node, "ethernet-ports");
		if (!ports)
			ports = of_get_child_by_name(dev->dev->of_node, "ports");
423 424 425 426 427
		if (ports)
			for_each_available_child_of_node(ports, port) {
				if (of_property_read_u32(port, "reg",
							 &port_num))
					continue;
428 429
				if (!(dev->port_mask & BIT(port_num))) {
					of_node_put(port);
430
					return -EINVAL;
431
				}
432 433 434
				of_get_phy_mode(port,
						&dev->ports[port_num].interface);
			}
435 436
		dev->synclko_125 = of_property_read_bool(dev->dev->of_node,
							 "microchip,synclko-125");
437 438 439 440 441 442
		dev->synclko_disable = of_property_read_bool(dev->dev->of_node,
							     "microchip,synclko-disable");
		if (dev->synclko_125 && dev->synclko_disable) {
			dev_err(dev->dev, "inconsistent synclko settings\n");
			return -EINVAL;
		}
443 444 445 446 447 448 449 450
	}

	ret = dsa_register_switch(dev->ds);
	if (ret) {
		dev->dev_ops->exit(dev);
		return ret;
	}

451 452 453 454 455 456
	/* Read MIB counters every 30 seconds to avoid overflow. */
	dev->mib_read_interval = msecs_to_jiffies(30000);

	/* Start the MIB timer. */
	schedule_delayed_work(&dev->mib_read, 0);

457
	return 0;
458 459 460 461 462
}
EXPORT_SYMBOL(ksz_switch_register);

void ksz_switch_remove(struct ksz_device *dev)
{
463
	/* timer started */
464 465
	if (dev->mib_read_interval) {
		dev->mib_read_interval = 0;
466
		cancel_delayed_work_sync(&dev->mib_read);
467
	}
468

469
	dev->dev_ops->exit(dev);
470
	dsa_unregister_switch(dev->ds);
471 472

	if (dev->reset_gpio)
473
		gpiod_set_value_cansleep(dev->reset_gpio, 1);
474

475 476 477 478 479 480
}
EXPORT_SYMBOL(ksz_switch_remove);

MODULE_AUTHOR("Woojung Huh <Woojung.Huh@microchip.com>");
MODULE_DESCRIPTION("Microchip KSZ Series Switch DSA Driver");
MODULE_LICENSE("GPL");