spu_priv1_mmio.c 13.3 KB
Newer Older
1
/*
2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
 * spu hypervisor abstraction for direct hardware access.
 *
 *  (C) Copyright IBM Deutschland Entwicklung GmbH 2005
 *  Copyright 2006 Sony Corp.
 *
 *  This program is free software; you can redistribute it and/or modify
 *  it under the terms of the GNU General Public License as published by
 *  the Free Software Foundation; version 2 of the License.
 *
 *  This program is distributed in the hope that it will be useful,
 *  but WITHOUT ANY WARRANTY; without even the implied warranty of
 *  MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 *  GNU General Public License for more details.
 *
 *  You should have received a copy of the GNU General Public License
 *  along with this program; if not, write to the Free Software
 *  Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
19
 */
20

21 22
#include <linux/interrupt.h>
#include <linux/list.h>
23
#include <linux/module.h>
24 25 26 27 28 29 30
#include <linux/ptrace.h>
#include <linux/slab.h>
#include <linux/wait.h>
#include <linux/mm.h>
#include <linux/io.h>
#include <linux/mutex.h>
#include <linux/device.h>
31 32

#include <asm/spu.h>
33
#include <asm/spu_priv1.h>
34 35
#include <asm/firmware.h>
#include <asm/prom.h>
36

37
#include "interrupt.h"
38 39
#include "spu_priv1_mmio.h"

40 41
static DEFINE_MUTEX(add_spumem_mutex);

42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239
struct spu_pdata {
	int nid;
	struct device_node *devnode;
	struct spu_priv1 __iomem *priv1;
};

static struct spu_pdata *spu_get_pdata(struct spu *spu)
{
	BUG_ON(!spu->pdata);
	return spu->pdata;
}

struct device_node *spu_devnode(struct spu *spu)
{
	return spu_get_pdata(spu)->devnode;
}

EXPORT_SYMBOL_GPL(spu_devnode);

static int __init find_spu_node_id(struct device_node *spe)
{
	const unsigned int *id;
	struct device_node *cpu;
	cpu = spe->parent->parent;
	id = get_property(cpu, "node-id", NULL);
	return id ? *id : 0;
}

static int __init cell_spuprop_present(struct spu *spu, struct device_node *spe,
		const char *prop)
{
	const struct address_prop {
		unsigned long address;
		unsigned int len;
	} __attribute__((packed)) *p;
	int proplen;

	unsigned long start_pfn, nr_pages;
	struct pglist_data *pgdata;
	struct zone *zone;
	int ret;

	p = get_property(spe, prop, &proplen);
	WARN_ON(proplen != sizeof (*p));

	start_pfn = p->address >> PAGE_SHIFT;
	nr_pages = ((unsigned long)p->len + PAGE_SIZE - 1) >> PAGE_SHIFT;

	pgdata = NODE_DATA(spu_get_pdata(spu)->nid);
	zone = pgdata->node_zones;

	/* XXX rethink locking here */
	mutex_lock(&add_spumem_mutex);
	ret = __add_pages(zone, start_pfn, nr_pages);
	mutex_unlock(&add_spumem_mutex);

	return ret;
}

static void __iomem * __init map_spe_prop(struct spu *spu,
		struct device_node *n, const char *name)
{
	const struct address_prop {
		unsigned long address;
		unsigned int len;
	} __attribute__((packed)) *prop;

	const void *p;
	int proplen;
	void __iomem *ret = NULL;
	int err = 0;

	p = get_property(n, name, &proplen);
	if (proplen != sizeof (struct address_prop))
		return NULL;

	prop = p;

	err = cell_spuprop_present(spu, n, name);
	if (err && (err != -EEXIST))
		goto out;

	ret = ioremap(prop->address, prop->len);

 out:
	return ret;
}

static void spu_unmap(struct spu *spu)
{
	iounmap(spu->priv2);
	iounmap(spu_get_pdata(spu)->priv1);
	iounmap(spu->problem);
	iounmap((__force u8 __iomem *)spu->local_store);
}

static int __init spu_map_interrupts_old(struct spu *spu,
	struct device_node *np)
{
	unsigned int isrc;
	const u32 *tmp;

	/* Get the interrupt source unit from the device-tree */
	tmp = get_property(np, "isrc", NULL);
	if (!tmp)
		return -ENODEV;
	isrc = tmp[0];

	/* Add the node number */
	isrc |= spu->node << IIC_IRQ_NODE_SHIFT;

	/* Now map interrupts of all 3 classes */
	spu->irqs[0] = irq_create_mapping(NULL, IIC_IRQ_CLASS_0 | isrc);
	spu->irqs[1] = irq_create_mapping(NULL, IIC_IRQ_CLASS_1 | isrc);
	spu->irqs[2] = irq_create_mapping(NULL, IIC_IRQ_CLASS_2 | isrc);

	/* Right now, we only fail if class 2 failed */
	return spu->irqs[2] == NO_IRQ ? -EINVAL : 0;
}

static int __init spu_map_device_old(struct spu *spu, struct device_node *node)
{
	const char *prop;
	int ret;

	ret = -ENODEV;
	spu->name = get_property(node, "name", NULL);
	if (!spu->name)
		goto out;

	prop = get_property(node, "local-store", NULL);
	if (!prop)
		goto out;
	spu->local_store_phys = *(unsigned long *)prop;

	/* we use local store as ram, not io memory */
	spu->local_store = (void __force *)
		map_spe_prop(spu, node, "local-store");
	if (!spu->local_store)
		goto out;

	prop = get_property(node, "problem", NULL);
	if (!prop)
		goto out_unmap;
	spu->problem_phys = *(unsigned long *)prop;

	spu->problem= map_spe_prop(spu, node, "problem");
	if (!spu->problem)
		goto out_unmap;

	spu_get_pdata(spu)->priv1= map_spe_prop(spu, node, "priv1");

	spu->priv2= map_spe_prop(spu, node, "priv2");
	if (!spu->priv2)
		goto out_unmap;
	ret = 0;
	goto out;

out_unmap:
	spu_unmap(spu);
out:
	return ret;
}

static int __init spu_map_interrupts(struct spu *spu, struct device_node *np)
{
	struct of_irq oirq;
	int ret;
	int i;

	for (i=0; i < 3; i++) {
		ret = of_irq_map_one(np, i, &oirq);
		if (ret) {
			pr_debug("spu_new: failed to get irq %d\n", i);
			goto err;
		}
		ret = -EINVAL;
		pr_debug("  irq %d no 0x%x on %s\n", i, oirq.specifier[0],
			 oirq.controller->full_name);
		spu->irqs[i] = irq_create_of_mapping(oirq.controller,
					oirq.specifier, oirq.size);
		if (spu->irqs[i] == NO_IRQ) {
			pr_debug("spu_new: failed to map it !\n");
			goto err;
		}
	}
	return 0;

err:
	pr_debug("failed to map irq %x for spu %s\n", *oirq.specifier,
		spu->name);
	for (; i >= 0; i--) {
		if (spu->irqs[i] != NO_IRQ)
			irq_dispose_mapping(spu->irqs[i]);
	}
	return ret;
}

240 241
static int spu_map_resource(struct spu *spu, int nr,
			    void __iomem** virt, unsigned long *phys)
242
{
243 244 245 246
	struct device_node *np = spu_get_pdata(spu)->devnode;
	unsigned long start_pfn, nr_pages;
	struct pglist_data *pgdata;
	struct zone *zone;
247
	struct resource resource = { };
248
	unsigned long len;
249 250
	int ret;

251
	ret = of_address_to_resource(np, nr, &resource);
252 253 254 255 256
	if (ret)
		goto out;

	if (phys)
		*phys = resource.start;
257 258
	len = resource.end - resource.start + 1;
	*virt = ioremap(resource.start, len);
259 260 261
	if (!*virt)
		ret = -EINVAL;

262 263 264 265 266 267 268 269 270 271 272
	start_pfn = resource.start >> PAGE_SHIFT;
	nr_pages = (len + PAGE_SIZE - 1) >> PAGE_SHIFT;

	pgdata = NODE_DATA(spu_get_pdata(spu)->nid);
	zone = pgdata->node_zones;

	/* XXX rethink locking here */
	mutex_lock(&add_spumem_mutex);
	ret = __add_pages(zone, start_pfn, nr_pages);
	mutex_unlock(&add_spumem_mutex);

273 274 275 276
out:
	return ret;
}

277
static int __init spu_map_device(struct spu *spu)
278
{
279
	struct device_node *np = spu_get_pdata(spu)->devnode;
280
	int ret = -ENODEV;
281 282

	spu->name = get_property(np, "name", NULL);
283 284 285
	if (!spu->name)
		goto out;

286 287
	ret = spu_map_resource(spu, 0, (void __iomem**)&spu->local_store,
			       &spu->local_store_phys);
288 289
	if (ret) {
		pr_debug("spu_new: failed to map %s resource 0\n",
290
			 np->full_name);
291 292
		goto out;
	}
293 294
	ret = spu_map_resource(spu, 1, (void __iomem**)&spu->problem,
			       &spu->problem_phys);
295 296
	if (ret) {
		pr_debug("spu_new: failed to map %s resource 1\n",
297
			 np->full_name);
298 299
		goto out_unmap;
	}
300
	ret = spu_map_resource(spu, 2, (void __iomem**)&spu->priv2, NULL);
301 302
	if (ret) {
		pr_debug("spu_new: failed to map %s resource 2\n",
303
			 np->full_name);
304 305 306
		goto out_unmap;
	}
	if (!firmware_has_feature(FW_FEATURE_LPAR))
307 308
		ret = spu_map_resource(spu, 3,
			       (void __iomem**)&spu_get_pdata(spu)->priv1, NULL);
309 310
	if (ret) {
		pr_debug("spu_new: failed to map %s resource 3\n",
311
			 np->full_name);
312 313
		goto out_unmap;
	}
314
	pr_debug("spu_new: %s maps:\n", np->full_name);
315 316 317 318 319 320
	pr_debug("  local store   : 0x%016lx -> 0x%p\n",
		 spu->local_store_phys, spu->local_store);
	pr_debug("  problem state : 0x%016lx -> 0x%p\n",
		 spu->problem_phys, spu->problem);
	pr_debug("  priv2         :                       0x%p\n", spu->priv2);
	pr_debug("  priv1         :                       0x%p\n",
321
		 spu_get_pdata(spu)->priv1);
322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360

	return 0;

out_unmap:
	spu_unmap(spu);
out:
	pr_debug("failed to map spe %s: %d\n", spu->name, ret);
	return ret;
}

static int __init of_enumerate_spus(int (*fn)(void *data))
{
	int ret;
	struct device_node *node;

	ret = -ENODEV;
	for (node = of_find_node_by_type(NULL, "spe");
			node; node = of_find_node_by_type(node, "spe")) {
		ret = fn(node);
		if (ret) {
			printk(KERN_WARNING "%s: Error initializing %s\n",
				__FUNCTION__, node->name);
			break;
		}
	}
	return ret;
}

static int __init of_create_spu(struct spu *spu, void *data)
{
	int ret;
	struct device_node *spe = (struct device_node *)data;

	spu->pdata = kzalloc(sizeof(struct spu_pdata),
		GFP_KERNEL);
	if (!spu->pdata) {
		ret = -ENOMEM;
		goto out;
	}
361
	spu_get_pdata(spu)->devnode = of_node_get(spe);
362 363 364 365 366 367 368 369 370 371 372 373 374 375

	spu->node = find_spu_node_id(spe);
	if (spu->node >= MAX_NUMNODES) {
		printk(KERN_WARNING "SPE %s on node %d ignored,"
		       " node number too big\n", spe->full_name, spu->node);
		printk(KERN_WARNING "Check if CONFIG_NUMA is enabled.\n");
		ret = -ENODEV;
		goto out_free;
	}

	spu_get_pdata(spu)->nid = of_node_to_nid(spe);
	if (spu_get_pdata(spu)->nid == -1)
		spu_get_pdata(spu)->nid = 0;

376
	ret = spu_map_device(spu);
377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416
	/* try old method */
	if (ret)
		ret = spu_map_device_old(spu, spe);
	if (ret)
		goto out_free;

	ret = spu_map_interrupts(spu, spe);
	if (ret)
		ret = spu_map_interrupts_old(spu, spe);
	if (ret)
		goto out_unmap;

	pr_debug(KERN_DEBUG "Using SPE %s %p %p %p %p %d\n", spu->name,
		spu->local_store, spu->problem, spu_get_pdata(spu)->priv1,
		spu->priv2, spu->number);
	goto out;

out_unmap:
	spu_unmap(spu);
out_free:
	kfree(spu->pdata);
	spu->pdata = NULL;
out:
	return ret;
}

static int of_destroy_spu(struct spu *spu)
{
	spu_unmap(spu);
	of_node_put(spu_get_pdata(spu)->devnode);
	kfree(spu->pdata);
	spu->pdata = NULL;
	return 0;
}

const struct spu_management_ops spu_management_of_ops = {
	.enumerate_spus = of_enumerate_spus,
	.create_spu = of_create_spu,
	.destroy_spu = of_destroy_spu,
};
417

418
static void int_mask_and(struct spu *spu, int class, u64 mask)
419 420 421
{
	u64 old_mask;

422 423 424
	old_mask = in_be64(&spu_get_pdata(spu)->priv1->int_mask_RW[class]);
	out_be64(&spu_get_pdata(spu)->priv1->int_mask_RW[class],
		old_mask & mask);
425 426
}

427
static void int_mask_or(struct spu *spu, int class, u64 mask)
428 429 430
{
	u64 old_mask;

431 432 433
	old_mask = in_be64(&spu_get_pdata(spu)->priv1->int_mask_RW[class]);
	out_be64(&spu_get_pdata(spu)->priv1->int_mask_RW[class],
		old_mask | mask);
434 435
}

436
static void int_mask_set(struct spu *spu, int class, u64 mask)
437
{
438
	out_be64(&spu_get_pdata(spu)->priv1->int_mask_RW[class], mask);
439 440
}

441
static u64 int_mask_get(struct spu *spu, int class)
442
{
443
	return in_be64(&spu_get_pdata(spu)->priv1->int_mask_RW[class]);
444 445
}

446
static void int_stat_clear(struct spu *spu, int class, u64 stat)
447
{
448
	out_be64(&spu_get_pdata(spu)->priv1->int_stat_RW[class], stat);
449 450
}

451
static u64 int_stat_get(struct spu *spu, int class)
452
{
453
	return in_be64(&spu_get_pdata(spu)->priv1->int_stat_RW[class]);
454 455
}

456
static void cpu_affinity_set(struct spu *spu, int cpu)
457
{
458 459
	u64 target = iic_get_target_id(cpu);
	u64 route = target << 48 | target << 32 | target << 16;
460
	out_be64(&spu_get_pdata(spu)->priv1->int_route_RW, route);
461 462
}

463
static u64 mfc_dar_get(struct spu *spu)
464
{
465
	return in_be64(&spu_get_pdata(spu)->priv1->mfc_dar_RW);
466 467
}

468
static u64 mfc_dsisr_get(struct spu *spu)
469
{
470
	return in_be64(&spu_get_pdata(spu)->priv1->mfc_dsisr_RW);
471 472
}

473
static void mfc_dsisr_set(struct spu *spu, u64 dsisr)
474
{
475
	out_be64(&spu_get_pdata(spu)->priv1->mfc_dsisr_RW, dsisr);
476 477
}

478
static void mfc_sdr_setup(struct spu *spu)
479
{
480
	out_be64(&spu_get_pdata(spu)->priv1->mfc_sdr_RW, mfspr(SPRN_SDR1));
481 482
}

483
static void mfc_sr1_set(struct spu *spu, u64 sr1)
484
{
485
	out_be64(&spu_get_pdata(spu)->priv1->mfc_sr1_RW, sr1);
486 487
}

488
static u64 mfc_sr1_get(struct spu *spu)
489
{
490
	return in_be64(&spu_get_pdata(spu)->priv1->mfc_sr1_RW);
491 492
}

493
static void mfc_tclass_id_set(struct spu *spu, u64 tclass_id)
494
{
495
	out_be64(&spu_get_pdata(spu)->priv1->mfc_tclass_id_RW, tclass_id);
496 497
}

498
static u64 mfc_tclass_id_get(struct spu *spu)
499
{
500
	return in_be64(&spu_get_pdata(spu)->priv1->mfc_tclass_id_RW);
501 502
}

503
static void tlb_invalidate(struct spu *spu)
504
{
505
	out_be64(&spu_get_pdata(spu)->priv1->tlb_invalidate_entry_W, 0ul);
506 507
}

508
static void resource_allocation_groupID_set(struct spu *spu, u64 id)
509
{
510 511
	out_be64(&spu_get_pdata(spu)->priv1->resource_allocation_groupID_RW,
		id);
512 513
}

514
static u64 resource_allocation_groupID_get(struct spu *spu)
515
{
516 517
	return in_be64(
		&spu_get_pdata(spu)->priv1->resource_allocation_groupID_RW);
518 519
}

520
static void resource_allocation_enable_set(struct spu *spu, u64 enable)
521
{
522 523
	out_be64(&spu_get_pdata(spu)->priv1->resource_allocation_enable_RW,
		enable);
524 525
}

526
static u64 resource_allocation_enable_get(struct spu *spu)
527
{
528 529
	return in_be64(
		&spu_get_pdata(spu)->priv1->resource_allocation_enable_RW);
530
}
531 532 533 534 535 536 537 538 539

const struct spu_priv1_ops spu_priv1_mmio_ops =
{
	.int_mask_and = int_mask_and,
	.int_mask_or = int_mask_or,
	.int_mask_set = int_mask_set,
	.int_mask_get = int_mask_get,
	.int_stat_clear = int_stat_clear,
	.int_stat_get = int_stat_get,
540
	.cpu_affinity_set = cpu_affinity_set,
541 542 543
	.mfc_dar_get = mfc_dar_get,
	.mfc_dsisr_get = mfc_dsisr_get,
	.mfc_dsisr_set = mfc_dsisr_set,
544
	.mfc_sdr_setup = mfc_sdr_setup,
545 546 547 548 549 550 551 552 553 554
	.mfc_sr1_set = mfc_sr1_set,
	.mfc_sr1_get = mfc_sr1_get,
	.mfc_tclass_id_set = mfc_tclass_id_set,
	.mfc_tclass_id_get = mfc_tclass_id_get,
	.tlb_invalidate = tlb_invalidate,
	.resource_allocation_groupID_set = resource_allocation_groupID_set,
	.resource_allocation_groupID_get = resource_allocation_groupID_get,
	.resource_allocation_enable_set = resource_allocation_enable_set,
	.resource_allocation_enable_get = resource_allocation_enable_get,
};