ioport.c 19.7 KB
Newer Older
A
Adrian Bunk 已提交
1
/*
L
Linus Torvalds 已提交
2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27
 * ioport.c:  Simple io mapping allocator.
 *
 * Copyright (C) 1995 David S. Miller (davem@caip.rutgers.edu)
 * Copyright (C) 1995 Miguel de Icaza (miguel@nuclecu.unam.mx)
 *
 * 1996: sparc_free_io, 1999: ioremap()/iounmap() by Pete Zaitcev.
 *
 * 2000/01/29
 * <rth> zait: as long as pci_alloc_consistent produces something addressable, 
 *	things are ok.
 * <zaitcev> rth: no, it is relevant, because get_free_pages returns you a
 *	pointer into the big page mapping
 * <rth> zait: so what?
 * <rth> zait: remap_it_my_way(virt_to_phys(get_free_page()))
 * <zaitcev> Hmm
 * <zaitcev> Suppose I did this remap_it_my_way(virt_to_phys(get_free_page())).
 *	So far so good.
 * <zaitcev> Now, driver calls pci_free_consistent(with result of
 *	remap_it_my_way()).
 * <zaitcev> How do you find the address to pass to free_pages()?
 * <rth> zait: walk the page tables?  It's only two or three level after all.
 * <rth> zait: you have to walk them anyway to remove the mapping.
 * <zaitcev> Hmm
 * <zaitcev> Sounds reasonable
 */

28
#include <linux/module.h>
L
Linus Torvalds 已提交
29 30 31 32 33 34 35 36 37
#include <linux/sched.h>
#include <linux/kernel.h>
#include <linux/errno.h>
#include <linux/types.h>
#include <linux/ioport.h>
#include <linux/mm.h>
#include <linux/slab.h>
#include <linux/pci.h>		/* struct pci_dev */
#include <linux/proc_fs.h>
38
#include <linux/seq_file.h>
J
Jens Axboe 已提交
39
#include <linux/scatterlist.h>
40
#include <linux/of_device.h>
L
Linus Torvalds 已提交
41 42 43 44

#include <asm/io.h>
#include <asm/vaddrs.h>
#include <asm/oplib.h>
45
#include <asm/prom.h>
L
Linus Torvalds 已提交
46 47 48
#include <asm/page.h>
#include <asm/pgalloc.h>
#include <asm/dma.h>
49 50
#include <asm/iommu.h>
#include <asm/io-unit.h>
K
Konrad Eisele 已提交
51
#include <asm/leon.h>
L
Linus Torvalds 已提交
52

K
Konrad Eisele 已提交
53 54 55
#ifdef CONFIG_SPARC_LEON
#define mmu_inval_dma_area(p, l) leon_flush_dcache_all()
#else
L
Linus Torvalds 已提交
56
#define mmu_inval_dma_area(p, l)	/* Anton pulled it out for 2.4.0-xx */
K
Konrad Eisele 已提交
57
#endif
L
Linus Torvalds 已提交
58

A
Adrian Bunk 已提交
59 60
static struct resource *_sparc_find_resource(struct resource *r,
					     unsigned long);
L
Linus Torvalds 已提交
61 62 63 64 65 66

static void __iomem *_sparc_ioremap(struct resource *res, u32 bus, u32 pa, int sz);
static void __iomem *_sparc_alloc_io(unsigned int busno, unsigned long phys,
    unsigned long size, char *name);
static void _sparc_free_io(struct resource *res);

A
Adrian Bunk 已提交
67 68
static void register_proc_sparc_ioport(void);

L
Linus Torvalds 已提交
69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126
/* This points to the next to use virtual memory for DVMA mappings */
static struct resource _sparc_dvma = {
	.name = "sparc_dvma", .start = DVMA_VADDR, .end = DVMA_END - 1
};
/* This points to the start of I/O mappings, cluable from outside. */
/*ext*/ struct resource sparc_iomap = {
	.name = "sparc_iomap", .start = IOBASE_VADDR, .end = IOBASE_END - 1
};

/*
 * Our mini-allocator...
 * Boy this is gross! We need it because we must map I/O for
 * timers and interrupt controller before the kmalloc is available.
 */

#define XNMLN  15
#define XNRES  10	/* SS-10 uses 8 */

struct xresource {
	struct resource xres;	/* Must be first */
	int xflag;		/* 1 == used */
	char xname[XNMLN+1];
};

static struct xresource xresv[XNRES];

static struct xresource *xres_alloc(void) {
	struct xresource *xrp;
	int n;

	xrp = xresv;
	for (n = 0; n < XNRES; n++) {
		if (xrp->xflag == 0) {
			xrp->xflag = 1;
			return xrp;
		}
		xrp++;
	}
	return NULL;
}

static void xres_free(struct xresource *xrp) {
	xrp->xflag = 0;
}

/*
 * These are typically used in PCI drivers
 * which are trying to be cross-platform.
 *
 * Bus type is always zero on IIep.
 */
void __iomem *ioremap(unsigned long offset, unsigned long size)
{
	char name[14];

	sprintf(name, "phys_%08x", (u32)offset);
	return _sparc_alloc_io(0, offset, size, name);
}
127
EXPORT_SYMBOL(ioremap);
L
Linus Torvalds 已提交
128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148

/*
 * Comlimentary to ioremap().
 */
void iounmap(volatile void __iomem *virtual)
{
	unsigned long vaddr = (unsigned long) virtual & PAGE_MASK;
	struct resource *res;

	if ((res = _sparc_find_resource(&sparc_iomap, vaddr)) == NULL) {
		printk("free_io/iounmap: cannot free %lx\n", vaddr);
		return;
	}
	_sparc_free_io(res);

	if ((char *)res >= (char*)xresv && (char *)res < (char *)&xresv[XNRES]) {
		xres_free((struct xresource *)res);
	} else {
		kfree(res);
	}
}
149
EXPORT_SYMBOL(iounmap);
L
Linus Torvalds 已提交
150

151 152 153 154 155 156 157 158 159
void __iomem *of_ioremap(struct resource *res, unsigned long offset,
			 unsigned long size, char *name)
{
	return _sparc_alloc_io(res->flags & 0xF,
			       res->start + offset,
			       size, name);
}
EXPORT_SYMBOL(of_ioremap);

160
void of_iounmap(struct resource *res, void __iomem *base, unsigned long size)
161 162 163 164 165
{
	iounmap(base);
}
EXPORT_SYMBOL(of_iounmap);

L
Linus Torvalds 已提交
166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223
/*
 * Meat of mapping
 */
static void __iomem *_sparc_alloc_io(unsigned int busno, unsigned long phys,
    unsigned long size, char *name)
{
	static int printed_full;
	struct xresource *xres;
	struct resource *res;
	char *tack;
	int tlen;
	void __iomem *va;	/* P3 diag */

	if (name == NULL) name = "???";

	if ((xres = xres_alloc()) != 0) {
		tack = xres->xname;
		res = &xres->xres;
	} else {
		if (!printed_full) {
			printk("ioremap: done with statics, switching to malloc\n");
			printed_full = 1;
		}
		tlen = strlen(name);
		tack = kmalloc(sizeof (struct resource) + tlen + 1, GFP_KERNEL);
		if (tack == NULL) return NULL;
		memset(tack, 0, sizeof(struct resource));
		res = (struct resource *) tack;
		tack += sizeof (struct resource);
	}

	strlcpy(tack, name, XNMLN+1);
	res->name = tack;

	va = _sparc_ioremap(res, busno, phys, size);
	/* printk("ioremap(0x%x:%08lx[0x%lx])=%p\n", busno, phys, size, va); */ /* P3 diag */
	return va;
}

/*
 */
static void __iomem *
_sparc_ioremap(struct resource *res, u32 bus, u32 pa, int sz)
{
	unsigned long offset = ((unsigned long) pa) & (~PAGE_MASK);

	if (allocate_resource(&sparc_iomap, res,
	    (offset + sz + PAGE_SIZE-1) & PAGE_MASK,
	    sparc_iomap.start, sparc_iomap.end, PAGE_SIZE, NULL, NULL) != 0) {
		/* Usually we cannot see printks in this case. */
		prom_printf("alloc_io_res(%s): cannot occupy\n",
		    (res->name != NULL)? res->name: "???");
		prom_halt();
	}

	pa &= PAGE_MASK;
	sparc_mapiorange(bus, pa, res->start, res->end - res->start + 1);

224
	return (void __iomem *)(unsigned long)(res->start + offset);
L
Linus Torvalds 已提交
225 226 227 228 229 230 231 232 233 234
}

/*
 * Comlimentary to _sparc_ioremap().
 */
static void _sparc_free_io(struct resource *res)
{
	unsigned long plen;

	plen = res->end - res->start + 1;
235
	BUG_ON((plen & (PAGE_SIZE-1)) != 0);
L
Linus Torvalds 已提交
236 237 238 239 240 241
	sparc_unmapiorange(res->start, plen);
	release_resource(res);
}

#ifdef CONFIG_SBUS

242
void sbus_set_sbus64(struct device *dev, int x)
243
{
L
Linus Torvalds 已提交
244 245
	printk("sbus_set_sbus64: unsupported\n");
}
246
EXPORT_SYMBOL(sbus_set_sbus64);
L
Linus Torvalds 已提交
247 248 249 250 251 252

/*
 * Allocate a chunk of memory suitable for DMA.
 * Typically devices use them for control blocks.
 * CPU may access them without any explicit flushing.
 */
253 254
static void *sbus_alloc_coherent(struct device *dev, size_t len,
				 dma_addr_t *dma_addrp, gfp_t gfp)
L
Linus Torvalds 已提交
255
{
256
	struct of_device *op = to_of_device(dev);
L
Linus Torvalds 已提交
257 258 259 260 261
	unsigned long len_total = (len + PAGE_SIZE-1) & PAGE_MASK;
	unsigned long va;
	struct resource *res;
	int order;

262
	/* XXX why are some lengths signed, others unsigned? */
L
Linus Torvalds 已提交
263 264 265 266 267 268 269 270 271
	if (len <= 0) {
		return NULL;
	}
	/* XXX So what is maxphys for us and how do drivers know it? */
	if (len > 256*1024) {			/* __get_free_pages() limit */
		return NULL;
	}

	order = get_order(len_total);
272
	if ((va = __get_free_pages(GFP_KERNEL|__GFP_COMP, order)) == 0)
L
Linus Torvalds 已提交
273 274
		goto err_nopages;

275
	if ((res = kzalloc(sizeof(struct resource), GFP_KERNEL)) == NULL)
L
Linus Torvalds 已提交
276 277 278 279 280 281 282 283 284 285 286 287 288 289
		goto err_nomem;

	if (allocate_resource(&_sparc_dvma, res, len_total,
	    _sparc_dvma.start, _sparc_dvma.end, PAGE_SIZE, NULL, NULL) != 0) {
		printk("sbus_alloc_consistent: cannot occupy 0x%lx", len_total);
		goto err_nova;
	}
	mmu_inval_dma_area(va, len_total);
	// XXX The mmu_map_dma_area does this for us below, see comments.
	// sparc_mapiorange(0, virt_to_phys(va), res->start, len_total);
	/*
	 * XXX That's where sdev would be used. Currently we load
	 * all iommu tables with the same translations.
	 */
290
	if (mmu_map_dma_area(dev, dma_addrp, va, res->start, len_total) != 0)
L
Linus Torvalds 已提交
291 292
		goto err_noiommu;

293
	res->name = op->node->name;
294

295
	return (void *)(unsigned long)res->start;
L
Linus Torvalds 已提交
296 297 298 299 300 301 302 303 304 305 306

err_noiommu:
	release_resource(res);
err_nova:
	free_pages(va, order);
err_nomem:
	kfree(res);
err_nopages:
	return NULL;
}

307 308
static void sbus_free_coherent(struct device *dev, size_t n, void *p,
			       dma_addr_t ba)
L
Linus Torvalds 已提交
309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325
{
	struct resource *res;
	struct page *pgv;

	if ((res = _sparc_find_resource(&_sparc_dvma,
	    (unsigned long)p)) == NULL) {
		printk("sbus_free_consistent: cannot free %p\n", p);
		return;
	}

	if (((unsigned long)p & (PAGE_SIZE-1)) != 0) {
		printk("sbus_free_consistent: unaligned va %p\n", p);
		return;
	}

	n = (n + PAGE_SIZE-1) & PAGE_MASK;
	if ((res->end-res->start)+1 != n) {
326
		printk("sbus_free_consistent: region 0x%lx asked 0x%zx\n",
L
Linus Torvalds 已提交
327 328 329 330 331 332 333 334
		    (long)((res->end-res->start)+1), n);
		return;
	}

	release_resource(res);
	kfree(res);

	/* mmu_inval_dma_area(va, n); */ /* it's consistent, isn't it */
335
	pgv = virt_to_page(p);
336
	mmu_unmap_dma_area(dev, ba, n);
L
Linus Torvalds 已提交
337 338 339 340 341 342 343 344 345

	__free_pages(pgv, get_order(n));
}

/*
 * Map a chunk of memory so that devices can see it.
 * CPU view of this memory may be inconsistent with
 * a device view and explicit flushing is necessary.
 */
346 347 348 349
static dma_addr_t sbus_map_page(struct device *dev, struct page *page,
				unsigned long offset, size_t len,
				enum dma_data_direction dir,
				struct dma_attrs *attrs)
L
Linus Torvalds 已提交
350
{
351 352
	void *va = page_address(page) + offset;

353
	/* XXX why are some lengths signed, others unsigned? */
L
Linus Torvalds 已提交
354 355 356 357 358 359 360
	if (len <= 0) {
		return 0;
	}
	/* XXX So what is maxphys for us and how do drivers know it? */
	if (len > 256*1024) {			/* __get_free_pages() limit */
		return 0;
	}
361
	return mmu_get_scsi_one(dev, va, len);
L
Linus Torvalds 已提交
362 363
}

364 365
static void sbus_unmap_page(struct device *dev, dma_addr_t ba, size_t n,
			    enum dma_data_direction dir, struct dma_attrs *attrs)
L
Linus Torvalds 已提交
366
{
367
	mmu_release_scsi_one(dev, ba, n);
L
Linus Torvalds 已提交
368 369
}

370 371
static int sbus_map_sg(struct device *dev, struct scatterlist *sg, int n,
		       enum dma_data_direction dir, struct dma_attrs *attrs)
L
Linus Torvalds 已提交
372
{
373
	mmu_get_scsi_sgl(dev, sg, n);
L
Linus Torvalds 已提交
374 375 376 377 378 379 380 381

	/*
	 * XXX sparc64 can return a partial length here. sun4c should do this
	 * but it currently panics if it can't fulfill the request - Anton
	 */
	return n;
}

382 383
static void sbus_unmap_sg(struct device *dev, struct scatterlist *sg, int n,
			  enum dma_data_direction dir, struct dma_attrs *attrs)
L
Linus Torvalds 已提交
384
{
385
	mmu_release_scsi_sgl(dev, sg, n);
L
Linus Torvalds 已提交
386 387
}

388 389
static void sbus_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg,
				 int n,	enum dma_data_direction dir)
L
Linus Torvalds 已提交
390
{
391
	BUG();
L
Linus Torvalds 已提交
392 393
}

394 395
static void sbus_sync_sg_for_device(struct device *dev, struct scatterlist *sg,
				    int n, enum dma_data_direction dir)
L
Linus Torvalds 已提交
396
{
397
	BUG();
L
Linus Torvalds 已提交
398 399
}

400 401 402 403 404 405 406 407 408 409 410 411 412
struct dma_map_ops sbus_dma_ops = {
	.alloc_coherent		= sbus_alloc_coherent,
	.free_coherent		= sbus_free_coherent,
	.map_page		= sbus_map_page,
	.unmap_page		= sbus_unmap_page,
	.map_sg			= sbus_map_sg,
	.unmap_sg		= sbus_unmap_sg,
	.sync_sg_for_cpu	= sbus_sync_sg_for_cpu,
	.sync_sg_for_device	= sbus_sync_sg_for_device,
};

struct dma_map_ops *dma_ops = &sbus_dma_ops;
EXPORT_SYMBOL(dma_ops);
L
Linus Torvalds 已提交
413

414
static int __init sparc_register_ioport(void)
415 416 417 418 419 420
{
	register_proc_sparc_ioport();

	return 0;
}

421 422
arch_initcall(sparc_register_ioport);

L
Linus Torvalds 已提交
423 424 425 426 427 428 429
#endif /* CONFIG_SBUS */

#ifdef CONFIG_PCI

/* Allocate and map kernel buffer using consistent mode DMA for a device.
 * hwdev should be valid struct pci_dev pointer for PCI devices.
 */
430 431
static void *pci32_alloc_coherent(struct device *dev, size_t len,
				  dma_addr_t *pba, gfp_t gfp)
L
Linus Torvalds 已提交
432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451
{
	unsigned long len_total = (len + PAGE_SIZE-1) & PAGE_MASK;
	unsigned long va;
	struct resource *res;
	int order;

	if (len == 0) {
		return NULL;
	}
	if (len > 256*1024) {			/* __get_free_pages() limit */
		return NULL;
	}

	order = get_order(len_total);
	va = __get_free_pages(GFP_KERNEL, order);
	if (va == 0) {
		printk("pci_alloc_consistent: no %ld pages\n", len_total>>PAGE_SHIFT);
		return NULL;
	}

452
	if ((res = kzalloc(sizeof(struct resource), GFP_KERNEL)) == NULL) {
L
Linus Torvalds 已提交
453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480
		free_pages(va, order);
		printk("pci_alloc_consistent: no core\n");
		return NULL;
	}

	if (allocate_resource(&_sparc_dvma, res, len_total,
	    _sparc_dvma.start, _sparc_dvma.end, PAGE_SIZE, NULL, NULL) != 0) {
		printk("pci_alloc_consistent: cannot occupy 0x%lx", len_total);
		free_pages(va, order);
		kfree(res);
		return NULL;
	}
	mmu_inval_dma_area(va, len_total);
#if 0
/* P3 */ printk("pci_alloc_consistent: kva %lx uncva %lx phys %lx size %lx\n",
  (long)va, (long)res->start, (long)virt_to_phys(va), len_total);
#endif
	sparc_mapiorange(0, virt_to_phys(va), res->start, len_total);

	*pba = virt_to_phys(va); /* equals virt_to_bus (R.I.P.) for us. */
	return (void *) res->start;
}

/* Free and unmap a consistent DMA buffer.
 * cpu_addr is what was returned from pci_alloc_consistent,
 * size must be the same as what as passed into pci_alloc_consistent,
 * and likewise dma_addr must be the same as what *dma_addrp was set to.
 *
S
Simon Arlott 已提交
481
 * References to the memory and mappings associated with cpu_addr/dma_addr
L
Linus Torvalds 已提交
482 483
 * past this call are illegal.
 */
484 485
static void pci32_free_coherent(struct device *dev, size_t n, void *p,
				dma_addr_t ba)
L
Linus Torvalds 已提交
486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520
{
	struct resource *res;
	unsigned long pgp;

	if ((res = _sparc_find_resource(&_sparc_dvma,
	    (unsigned long)p)) == NULL) {
		printk("pci_free_consistent: cannot free %p\n", p);
		return;
	}

	if (((unsigned long)p & (PAGE_SIZE-1)) != 0) {
		printk("pci_free_consistent: unaligned va %p\n", p);
		return;
	}

	n = (n + PAGE_SIZE-1) & PAGE_MASK;
	if ((res->end-res->start)+1 != n) {
		printk("pci_free_consistent: region 0x%lx asked 0x%lx\n",
		    (long)((res->end-res->start)+1), (long)n);
		return;
	}

	pgp = (unsigned long) phys_to_virt(ba);	/* bus_to_virt actually */
	mmu_inval_dma_area(pgp, n);
	sparc_unmapiorange((unsigned long)p, n);

	release_resource(res);
	kfree(res);

	free_pages(pgp, get_order(n));
}

/*
 * Same as pci_map_single, but with pages.
 */
521 522 523 524
static dma_addr_t pci32_map_page(struct device *dev, struct page *page,
				 unsigned long offset, size_t size,
				 enum dma_data_direction dir,
				 struct dma_attrs *attrs)
L
Linus Torvalds 已提交
525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544
{
	/* IIep is write-through, not flushing. */
	return page_to_phys(page) + offset;
}

/* Map a set of buffers described by scatterlist in streaming
 * mode for DMA.  This is the scather-gather version of the
 * above pci_map_single interface.  Here the scatter gather list
 * elements are each tagged with the appropriate dma address
 * and length.  They are obtained via sg_dma_{address,length}(SG).
 *
 * NOTE: An implementation may be able to use a smaller number of
 *       DMA address/length pairs than there are SG table elements.
 *       (for example via virtual mapping capabilities)
 *       The routine returns the number of addr/length pairs actually
 *       used, at most nents.
 *
 * Device ownership issues as mentioned above for pci_map_single are
 * the same here.
 */
545 546 547
static int pci32_map_sg(struct device *device, struct scatterlist *sgl,
			int nents, enum dma_data_direction dir,
			struct dma_attrs *attrs)
L
Linus Torvalds 已提交
548
{
J
Jens Axboe 已提交
549
	struct scatterlist *sg;
L
Linus Torvalds 已提交
550 551 552
	int n;

	/* IIep is write-through, not flushing. */
J
Jens Axboe 已提交
553
	for_each_sg(sgl, sg, nents, n) {
J
Jens Axboe 已提交
554
		BUG_ON(page_address(sg_page(sg)) == NULL);
555 556
		sg->dma_address = virt_to_phys(sg_virt(sg));
		sg->dma_length = sg->length;
L
Linus Torvalds 已提交
557 558 559 560 561 562 563 564
	}
	return nents;
}

/* Unmap a set of streaming mode DMA translations.
 * Again, cpu read rules concerning calls here are the same as for
 * pci_unmap_single() above.
 */
565 566 567
static void pci32_unmap_sg(struct device *dev, struct scatterlist *sgl,
			   int nents, enum dma_data_direction dir,
			   struct dma_attrs *attrs)
L
Linus Torvalds 已提交
568
{
J
Jens Axboe 已提交
569
	struct scatterlist *sg;
L
Linus Torvalds 已提交
570 571
	int n;

572
	if (dir != PCI_DMA_TODEVICE) {
J
Jens Axboe 已提交
573
		for_each_sg(sgl, sg, nents, n) {
J
Jens Axboe 已提交
574
			BUG_ON(page_address(sg_page(sg)) == NULL);
L
Linus Torvalds 已提交
575
			mmu_inval_dma_area(
J
Jens Axboe 已提交
576
			    (unsigned long) page_address(sg_page(sg)),
L
Linus Torvalds 已提交
577 578 579 580 581 582 583 584 585 586 587 588 589 590 591
			    (sg->length + PAGE_SIZE-1) & PAGE_MASK);
		}
	}
}

/* Make physical memory consistent for a single
 * streaming mode DMA translation before or after a transfer.
 *
 * If you perform a pci_map_single() but wish to interrogate the
 * buffer using the cpu, yet do not wish to teardown the PCI dma
 * mapping, you must call this function before doing so.  At the
 * next point you give the PCI dma address back to the card, you
 * must first perform a pci_dma_sync_for_device, and then the
 * device again owns the buffer.
 */
592 593
static void pci32_sync_single_for_cpu(struct device *dev, dma_addr_t ba,
				      size_t size, enum dma_data_direction dir)
L
Linus Torvalds 已提交
594
{
595
	if (dir != PCI_DMA_TODEVICE) {
L
Linus Torvalds 已提交
596 597 598 599 600
		mmu_inval_dma_area((unsigned long)phys_to_virt(ba),
		    (size + PAGE_SIZE-1) & PAGE_MASK);
	}
}

601 602
static void pci32_sync_single_for_device(struct device *dev, dma_addr_t ba,
					 size_t size, enum dma_data_direction dir)
L
Linus Torvalds 已提交
603
{
604
	if (dir != PCI_DMA_TODEVICE) {
L
Linus Torvalds 已提交
605 606 607 608 609 610 611 612 613 614 615
		mmu_inval_dma_area((unsigned long)phys_to_virt(ba),
		    (size + PAGE_SIZE-1) & PAGE_MASK);
	}
}

/* Make physical memory consistent for a set of streaming
 * mode DMA translations after a transfer.
 *
 * The same as pci_dma_sync_single_* but for a scatter-gather list,
 * same rules and usage.
 */
616 617
static void pci32_sync_sg_for_cpu(struct device *dev, struct scatterlist *sgl,
				  int nents, enum dma_data_direction dir)
L
Linus Torvalds 已提交
618
{
J
Jens Axboe 已提交
619
	struct scatterlist *sg;
L
Linus Torvalds 已提交
620 621
	int n;

622
	if (dir != PCI_DMA_TODEVICE) {
J
Jens Axboe 已提交
623
		for_each_sg(sgl, sg, nents, n) {
J
Jens Axboe 已提交
624
			BUG_ON(page_address(sg_page(sg)) == NULL);
L
Linus Torvalds 已提交
625
			mmu_inval_dma_area(
J
Jens Axboe 已提交
626
			    (unsigned long) page_address(sg_page(sg)),
L
Linus Torvalds 已提交
627 628 629 630 631
			    (sg->length + PAGE_SIZE-1) & PAGE_MASK);
		}
	}
}

632 633
static void pci32_sync_sg_for_device(struct device *device, struct scatterlist *sgl,
				     int nents, enum dma_data_direction dir)
L
Linus Torvalds 已提交
634
{
J
Jens Axboe 已提交
635
	struct scatterlist *sg;
L
Linus Torvalds 已提交
636 637
	int n;

638
	if (dir != PCI_DMA_TODEVICE) {
J
Jens Axboe 已提交
639
		for_each_sg(sgl, sg, nents, n) {
J
Jens Axboe 已提交
640
			BUG_ON(page_address(sg_page(sg)) == NULL);
L
Linus Torvalds 已提交
641
			mmu_inval_dma_area(
J
Jens Axboe 已提交
642
			    (unsigned long) page_address(sg_page(sg)),
L
Linus Torvalds 已提交
643 644 645 646
			    (sg->length + PAGE_SIZE-1) & PAGE_MASK);
		}
	}
}
647 648 649 650 651 652 653 654 655 656 657 658 659 660

struct dma_map_ops pci32_dma_ops = {
	.alloc_coherent		= pci32_alloc_coherent,
	.free_coherent		= pci32_free_coherent,
	.map_page		= pci32_map_page,
	.map_sg			= pci32_map_sg,
	.unmap_sg		= pci32_unmap_sg,
	.sync_single_for_cpu	= pci32_sync_single_for_cpu,
	.sync_single_for_device	= pci32_sync_single_for_device,
	.sync_sg_for_cpu	= pci32_sync_sg_for_cpu,
	.sync_sg_for_device	= pci32_sync_sg_for_device,
};
EXPORT_SYMBOL(pci32_dma_ops);

L
Linus Torvalds 已提交
661 662
#endif /* CONFIG_PCI */

663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689
/*
 * Return whether the given PCI device DMA address mask can be
 * supported properly.  For example, if your device can only drive the
 * low 24-bits during PCI bus mastering, then you would pass
 * 0x00ffffff as the mask to this function.
 */
int dma_supported(struct device *dev, u64 mask)
{
#ifdef CONFIG_PCI
	if (dev->bus == &pci_bus_type)
		return 1;
#endif
	return 0;
}
EXPORT_SYMBOL(dma_supported);

int dma_set_mask(struct device *dev, u64 dma_mask)
{
#ifdef CONFIG_PCI
	if (dev->bus == &pci_bus_type)
		return pci_set_dma_mask(to_pci_dev(dev), dma_mask);
#endif
	return -EOPNOTSUPP;
}
EXPORT_SYMBOL(dma_set_mask);


L
Linus Torvalds 已提交
690 691
#ifdef CONFIG_PROC_FS

692
static int sparc_io_proc_show(struct seq_file *m, void *v)
L
Linus Torvalds 已提交
693
{
694
	struct resource *root = m->private, *r;
L
Linus Torvalds 已提交
695 696
	const char *nm;

697
	for (r = root->child; r != NULL; r = r->sibling) {
L
Linus Torvalds 已提交
698
		if ((nm = r->name) == 0) nm = "???";
699
		seq_printf(m, "%016llx-%016llx: %s\n",
700 701
				(unsigned long long)r->start,
				(unsigned long long)r->end, nm);
L
Linus Torvalds 已提交
702 703
	}

704
	return 0;
L
Linus Torvalds 已提交
705 706
}

707 708 709 710 711 712 713 714 715 716 717 718
static int sparc_io_proc_open(struct inode *inode, struct file *file)
{
	return single_open(file, sparc_io_proc_show, PDE(inode)->data);
}

static const struct file_operations sparc_io_proc_fops = {
	.owner		= THIS_MODULE,
	.open		= sparc_io_proc_open,
	.read		= seq_read,
	.llseek		= seq_lseek,
	.release	= single_release,
};
L
Linus Torvalds 已提交
719 720 721 722 723 724 725 726 727
#endif /* CONFIG_PROC_FS */

/*
 * This is a version of find_resource and it belongs to kernel/resource.c.
 * Until we have agreement with Linus and Martin, it lingers here.
 *
 * XXX Too slow. Can have 8192 DVMA pages on sun4m in the worst case.
 * This probably warrants some sort of hashing.
 */
A
Adrian Bunk 已提交
728 729
static struct resource *_sparc_find_resource(struct resource *root,
					     unsigned long hit)
L
Linus Torvalds 已提交
730 731 732 733 734 735 736 737 738 739
{
        struct resource *tmp;

	for (tmp = root->child; tmp != 0; tmp = tmp->sibling) {
		if (tmp->start <= hit && tmp->end >= hit)
			return tmp;
	}
	return NULL;
}

A
Adrian Bunk 已提交
740
static void register_proc_sparc_ioport(void)
L
Linus Torvalds 已提交
741 742
{
#ifdef CONFIG_PROC_FS
743 744
	proc_create_data("io_map", 0, NULL, &sparc_io_proc_fops, &sparc_iomap);
	proc_create_data("dvma_map", 0, NULL, &sparc_io_proc_fops, &_sparc_dvma);
L
Linus Torvalds 已提交
745 746
#endif
}