inode.c 17.3 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28
/*
 * SPU file system
 *
 * (C) Copyright IBM Deutschland Entwicklung GmbH 2005
 *
 * Author: Arnd Bergmann <arndb@de.ibm.com>
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License as published by
 * the Free Software Foundation; either version 2, or (at your option)
 * any later version.
 *
 * This program is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 * GNU General Public License for more details.
 *
 * You should have received a copy of the GNU General Public License
 * along with this program; if not, write to the Free Software
 * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
 */

#include <linux/file.h>
#include <linux/fs.h>
#include <linux/backing-dev.h>
#include <linux/init.h>
#include <linux/ioctl.h>
#include <linux/module.h>
29
#include <linux/mount.h>
30 31 32 33 34 35
#include <linux/namei.h>
#include <linux/pagemap.h>
#include <linux/poll.h>
#include <linux/slab.h>
#include <linux/parser.h>

36
#include <asm/prom.h>
37 38
#include <asm/semaphore.h>
#include <asm/spu.h>
39
#include <asm/spu_priv1.h>
40 41 42 43
#include <asm/uaccess.h>

#include "spufs.h"

44
static struct kmem_cache *spufs_inode_cache;
45
char *isolated_loader;
46
static int isolated_loader_size;
47 48 49 50 51 52

static struct inode *
spufs_alloc_inode(struct super_block *sb)
{
	struct spufs_inode_info *ei;

53
	ei = kmem_cache_alloc(spufs_inode_cache, GFP_KERNEL);
54 55
	if (!ei)
		return NULL;
56 57 58

	ei->i_gang = NULL;
	ei->i_ctx = NULL;
59
	ei->i_openers = 0;
60

61 62 63 64 65 66 67 68 69 70
	return &ei->vfs_inode;
}

static void
spufs_destroy_inode(struct inode *inode)
{
	kmem_cache_free(spufs_inode_cache, SPUFS_I(inode));
}

static void
71
spufs_init_once(struct kmem_cache *cachep, void *p)
72 73 74
{
	struct spufs_inode_info *ei = p;

C
Christoph Lameter 已提交
75
	inode_init_once(&ei->vfs_inode);
76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109
}

static struct inode *
spufs_new_inode(struct super_block *sb, int mode)
{
	struct inode *inode;

	inode = new_inode(sb);
	if (!inode)
		goto out;

	inode->i_mode = mode;
	inode->i_uid = current->fsuid;
	inode->i_gid = current->fsgid;
	inode->i_blocks = 0;
	inode->i_atime = inode->i_mtime = inode->i_ctime = CURRENT_TIME;
out:
	return inode;
}

static int
spufs_setattr(struct dentry *dentry, struct iattr *attr)
{
	struct inode *inode = dentry->d_inode;

	if ((attr->ia_valid & ATTR_SIZE) &&
	    (attr->ia_size != inode->i_size))
		return -EINVAL;
	return inode_setattr(inode, attr);
}


static int
spufs_new_file(struct super_block *sb, struct dentry *dentry,
110
		const struct file_operations *fops, int mode,
111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126
		struct spu_context *ctx)
{
	static struct inode_operations spufs_file_iops = {
		.setattr = spufs_setattr,
	};
	struct inode *inode;
	int ret;

	ret = -ENOSPC;
	inode = spufs_new_inode(sb, S_IFREG | mode);
	if (!inode)
		goto out;

	ret = 0;
	inode->i_op = &spufs_file_iops;
	inode->i_fop = fops;
127
	inode->i_private = SPUFS_I(inode)->i_ctx = get_spu_context(ctx);
128 129 130 131 132 133 134 135
	d_add(dentry, inode);
out:
	return ret;
}

static void
spufs_delete_inode(struct inode *inode)
{
136 137 138 139 140 141
	struct spufs_inode_info *ei = SPUFS_I(inode);

	if (ei->i_ctx)
		put_spu_context(ei->i_ctx);
	if (ei->i_gang)
		put_spu_gang(ei->i_gang);
142 143 144
	clear_inode(inode);
}

145
static void spufs_prune_dir(struct dentry *dir)
146
{
147
	struct dentry *dentry, *tmp;
148

149
	mutex_lock(&dir->d_inode->i_mutex);
150
	list_for_each_entry_safe(dentry, tmp, &dir->d_subdirs, d_u.d_child) {
151
		spin_lock(&dcache_lock);
152
		spin_lock(&dentry->d_lock);
153 154 155 156
		if (!(d_unhashed(dentry)) && dentry->d_inode) {
			dget_locked(dentry);
			__d_drop(dentry);
			spin_unlock(&dentry->d_lock);
157
			simple_unlink(dir->d_inode, dentry);
158 159 160 161 162 163
			spin_unlock(&dcache_lock);
			dput(dentry);
		} else {
			spin_unlock(&dentry->d_lock);
			spin_unlock(&dcache_lock);
		}
164
	}
165
	shrink_dcache_parent(dir);
166
	mutex_unlock(&dir->d_inode->i_mutex);
167 168
}

169 170
/* Caller must hold parent->i_mutex */
static int spufs_rmdir(struct inode *parent, struct dentry *dir)
171 172
{
	/* remove all entries */
173
	spufs_prune_dir(dir);
174
	d_drop(dir);
175

176
	return simple_rmdir(parent, dir);
177 178
}

179 180 181
static int spufs_fill_dir(struct dentry *dir, struct tree_descr *files,
			  int mode, struct spu_context *ctx)
{
182
	struct dentry *dentry, *tmp;
183 184 185 186 187 188 189 190 191 192 193 194 195 196 197
	int ret;

	while (files->name && files->name[0]) {
		ret = -ENOMEM;
		dentry = d_alloc_name(dir, files->name);
		if (!dentry)
			goto out;
		ret = spufs_new_file(dir->d_sb, dentry, files->ops,
					files->mode & mode, ctx);
		if (ret)
			goto out;
		files++;
	}
	return 0;
out:
198 199 200 201 202 203 204 205 206 207 208 209 210 211
	/*
	 * remove all children from dir. dir->inode is not set so don't
	 * just simply use spufs_prune_dir() and panic afterwards :)
	 * dput() looks like it will do the right thing:
	 * - dec parent's ref counter
	 * - remove child from parent's child list
	 * - free child's inode if possible
	 * - free child
	 */
	list_for_each_entry_safe(dentry, tmp, &dir->d_subdirs, d_u.d_child) {
		dput(dentry);
	}

	shrink_dcache_parent(dir);
212 213 214
	return ret;
}

215 216
static int spufs_dir_close(struct inode *inode, struct file *file)
{
217
	struct spu_context *ctx;
218 219
	struct inode *parent;
	struct dentry *dir;
220 221
	int ret;

J
Josef Sipek 已提交
222
	dir = file->f_path.dentry;
223 224
	parent = dir->d_parent->d_inode;
	ctx = SPUFS_I(dir->d_inode)->i_ctx;
225

226 227 228
	mutex_lock(&parent->i_mutex);
	ret = spufs_rmdir(parent, dir);
	mutex_unlock(&parent->i_mutex);
229
	WARN_ON(ret);
230

231 232 233
	/* We have to give up the mm_struct */
	spu_forget(ctx);

234 235 236
	return dcache_dir_close(inode, file);
}

237
const struct file_operations spufs_context_fops = {
238 239 240 241 242 243 244
	.open		= dcache_dir_open,
	.release	= spufs_dir_close,
	.llseek		= dcache_dir_lseek,
	.read		= generic_read_dir,
	.readdir	= dcache_readdir,
	.fsync		= simple_sync_file,
};
245
EXPORT_SYMBOL_GPL(spufs_context_fops);
246 247

static int
248 249
spufs_mkdir(struct inode *dir, struct dentry *dentry, unsigned int flags,
		int mode)
250 251 252 253 254 255 256 257 258 259 260 261 262 263
{
	int ret;
	struct inode *inode;
	struct spu_context *ctx;

	ret = -ENOSPC;
	inode = spufs_new_inode(dir->i_sb, mode | S_IFDIR);
	if (!inode)
		goto out;

	if (dir->i_mode & S_ISGID) {
		inode->i_gid = dir->i_gid;
		inode->i_mode &= S_ISGID;
	}
264
	ctx = alloc_spu_context(SPUFS_I(dir)->i_gang); /* XXX gang */
265 266 267 268
	SPUFS_I(inode)->i_ctx = ctx;
	if (!ctx)
		goto out_iput;

269
	ctx->flags = flags;
270
	inode->i_op = &simple_dir_inode_operations;
271
	inode->i_fop = &simple_dir_operations;
272 273 274 275 276 277
	if (flags & SPU_CREATE_NOSCHED)
		ret = spufs_fill_dir(dentry, spufs_dir_nosched_contents,
					 mode, ctx);
	else
		ret = spufs_fill_dir(dentry, spufs_dir_contents, mode, ctx);

278 279 280 281 282 283
	if (ret)
		goto out_free_ctx;

	d_instantiate(dentry, inode);
	dget(dentry);
	dir->i_nlink++;
284
	dentry->d_inode->i_nlink++;
285 286 287
	goto out;

out_free_ctx:
288
	spu_forget(ctx);
289 290 291 292 293 294 295
	put_spu_context(ctx);
out_iput:
	iput(inode);
out:
	return ret;
}

296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320
static int spufs_context_open(struct dentry *dentry, struct vfsmount *mnt)
{
	int ret;
	struct file *filp;

	ret = get_unused_fd();
	if (ret < 0) {
		dput(dentry);
		mntput(mnt);
		goto out;
	}

	filp = dentry_open(dentry, mnt, O_RDONLY);
	if (IS_ERR(filp)) {
		put_unused_fd(ret);
		ret = PTR_ERR(filp);
		goto out;
	}

	filp->f_op = &spufs_context_fops;
	fd_install(ret, filp);
out:
	return ret;
}

321 322 323 324
static struct spu_context *
spufs_assert_affinity(unsigned int flags, struct spu_gang *gang,
						struct file *filp)
{
325
	struct spu_context *tmp, *neighbor, *err;
326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356
	int count, node;
	int aff_supp;

	aff_supp = !list_empty(&(list_entry(cbe_spu_info[0].spus.next,
					struct spu, cbe_list))->aff_list);

	if (!aff_supp)
		return ERR_PTR(-EINVAL);

	if (flags & SPU_CREATE_GANG)
		return ERR_PTR(-EINVAL);

	if (flags & SPU_CREATE_AFFINITY_MEM &&
	    gang->aff_ref_ctx &&
	    gang->aff_ref_ctx->flags & SPU_CREATE_AFFINITY_MEM)
		return ERR_PTR(-EEXIST);

	if (gang->aff_flags & AFF_MERGED)
		return ERR_PTR(-EBUSY);

	neighbor = NULL;
	if (flags & SPU_CREATE_AFFINITY_SPU) {
		if (!filp || filp->f_op != &spufs_context_fops)
			return ERR_PTR(-EINVAL);

		neighbor = get_spu_context(
				SPUFS_I(filp->f_dentry->d_inode)->i_ctx);

		if (!list_empty(&neighbor->aff_list) && !(neighbor->aff_head) &&
		    !list_is_last(&neighbor->aff_list, &gang->aff_list_head) &&
		    !list_entry(neighbor->aff_list.next, struct spu_context,
357 358 359 360
		    aff_list)->aff_head) {
			err = ERR_PTR(-EEXIST);
			goto out_put_neighbor;
		}
361

362 363 364 365
		if (gang != neighbor->gang) {
			err = ERR_PTR(-EINVAL);
			goto out_put_neighbor;
		}
366 367 368 369 370 371 372 373 374 375 376 377 378

		count = 1;
		list_for_each_entry(tmp, &gang->aff_list_head, aff_list)
			count++;
		if (list_empty(&neighbor->aff_list))
			count++;

		for (node = 0; node < MAX_NUMNODES; node++) {
			if ((cbe_spu_info[node].n_spus - atomic_read(
				&cbe_spu_info[node].reserved_spus)) >= count)
				break;
		}

379 380 381 382
		if (node == MAX_NUMNODES) {
			err = ERR_PTR(-EEXIST);
			goto out_put_neighbor;
		}
383 384 385
	}

	return neighbor;
386 387 388 389

out_put_neighbor:
	put_spu_context(neighbor);
	return err;
390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426
}

static void
spufs_set_affinity(unsigned int flags, struct spu_context *ctx,
					struct spu_context *neighbor)
{
	if (flags & SPU_CREATE_AFFINITY_MEM)
		ctx->gang->aff_ref_ctx = ctx;

	if (flags & SPU_CREATE_AFFINITY_SPU) {
		if (list_empty(&neighbor->aff_list)) {
			list_add_tail(&neighbor->aff_list,
				&ctx->gang->aff_list_head);
			neighbor->aff_head = 1;
		}

		if (list_is_last(&neighbor->aff_list, &ctx->gang->aff_list_head)
		    || list_entry(neighbor->aff_list.next, struct spu_context,
							aff_list)->aff_head) {
			list_add(&ctx->aff_list, &neighbor->aff_list);
		} else  {
			list_add_tail(&ctx->aff_list, &neighbor->aff_list);
			if (neighbor->aff_head) {
				neighbor->aff_head = 0;
				ctx->aff_head = 1;
			}
		}

		if (!ctx->gang->aff_ref_ctx)
			ctx->gang->aff_ref_ctx = ctx;
	}
}

static int
spufs_create_context(struct inode *inode, struct dentry *dentry,
			struct vfsmount *mnt, int flags, int mode,
			struct file *aff_filp)
427 428
{
	int ret;
429 430 431
	int affinity;
	struct spu_gang *gang;
	struct spu_context *neighbor;
432

433 434 435 436 437 438 439 440 441 442
	ret = -EPERM;
	if ((flags & SPU_CREATE_NOSCHED) &&
	    !capable(CAP_SYS_NICE))
		goto out_unlock;

	ret = -EINVAL;
	if ((flags & (SPU_CREATE_NOSCHED | SPU_CREATE_ISOLATE))
	    == SPU_CREATE_ISOLATE)
		goto out_unlock;

443 444 445 446
	ret = -ENODEV;
	if ((flags & SPU_CREATE_ISOLATE) && !isolated_loader)
		goto out_unlock;

447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462
	gang = NULL;
	neighbor = NULL;
	affinity = flags & (SPU_CREATE_AFFINITY_MEM | SPU_CREATE_AFFINITY_SPU);
	if (affinity) {
		gang = SPUFS_I(inode)->i_gang;
		ret = -EINVAL;
		if (!gang)
			goto out_unlock;
		mutex_lock(&gang->aff_mutex);
		neighbor = spufs_assert_affinity(flags, gang, aff_filp);
		if (IS_ERR(neighbor)) {
			ret = PTR_ERR(neighbor);
			goto out_aff_unlock;
		}
	}

463 464
	ret = spufs_mkdir(inode, dentry, flags, mode & S_IRWXUGO);
	if (ret)
465 466
		goto out_aff_unlock;

467
	if (affinity) {
468 469
		spufs_set_affinity(flags, SPUFS_I(dentry->d_inode)->i_ctx,
								neighbor);
470 471 472
		if (neighbor)
			put_spu_context(neighbor);
	}
473 474 475 476 477 478 479 480 481 482 483 484 485

	/*
	 * get references for dget and mntget, will be released
	 * in error path of *_open().
	 */
	ret = spufs_context_open(dget(dentry), mntget(mnt));
	if (ret < 0) {
		WARN_ON(spufs_rmdir(inode, dentry));
		mutex_unlock(&inode->i_mutex);
		spu_forget(SPUFS_I(dentry->d_inode)->i_ctx);
		goto out;
	}

486 487 488
out_aff_unlock:
	if (affinity)
		mutex_unlock(&gang->aff_mutex);
489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518
out_unlock:
	mutex_unlock(&inode->i_mutex);
out:
	dput(dentry);
	return ret;
}

static int
spufs_mkgang(struct inode *dir, struct dentry *dentry, int mode)
{
	int ret;
	struct inode *inode;
	struct spu_gang *gang;

	ret = -ENOSPC;
	inode = spufs_new_inode(dir->i_sb, mode | S_IFDIR);
	if (!inode)
		goto out;

	ret = 0;
	if (dir->i_mode & S_ISGID) {
		inode->i_gid = dir->i_gid;
		inode->i_mode &= S_ISGID;
	}
	gang = alloc_spu_gang();
	SPUFS_I(inode)->i_ctx = NULL;
	SPUFS_I(inode)->i_gang = gang;
	if (!gang)
		goto out_iput;

519
	inode->i_op = &simple_dir_inode_operations;
520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551
	inode->i_fop = &simple_dir_operations;

	d_instantiate(dentry, inode);
	dir->i_nlink++;
	dentry->d_inode->i_nlink++;
	return ret;

out_iput:
	iput(inode);
out:
	return ret;
}

static int spufs_gang_open(struct dentry *dentry, struct vfsmount *mnt)
{
	int ret;
	struct file *filp;

	ret = get_unused_fd();
	if (ret < 0) {
		dput(dentry);
		mntput(mnt);
		goto out;
	}

	filp = dentry_open(dentry, mnt, O_RDONLY);
	if (IS_ERR(filp)) {
		put_unused_fd(ret);
		ret = PTR_ERR(filp);
		goto out;
	}

552
	filp->f_op = &simple_dir_operations;
553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572
	fd_install(ret, filp);
out:
	return ret;
}

static int spufs_create_gang(struct inode *inode,
			struct dentry *dentry,
			struct vfsmount *mnt, int mode)
{
	int ret;

	ret = spufs_mkgang(inode, dentry, mode & S_IRWXUGO);
	if (ret)
		goto out;

	/*
	 * get references for dget and mntget, will be released
	 * in error path of *_open().
	 */
	ret = spufs_gang_open(dget(dentry), mntget(mnt));
573 574 575 576
	if (ret < 0) {
		int err = simple_rmdir(inode, dentry);
		WARN_ON(err);
	}
577 578 579 580 581 582 583 584

out:
	mutex_unlock(&inode->i_mutex);
	dput(dentry);
	return ret;
}


585 586
static struct file_system_type spufs_type;

587 588
long spufs_create(struct nameidata *nd, unsigned int flags, mode_t mode,
							struct file *filp)
589 590 591 592 593
{
	struct dentry *dentry;
	int ret;

	ret = -EINVAL;
594 595
	/* check if we are on spufs */
	if (nd->dentry->d_sb->s_type != &spufs_type)
596 597
		goto out;

598
	/* don't accept undefined flags */
599
	if (flags & (~SPU_CREATE_FLAG_ALL))
600 601
		goto out;

602 603 604 605 606 607 608
	/* only threads can be underneath a gang */
	if (nd->dentry != nd->dentry->d_sb->s_root) {
		if ((flags & SPU_CREATE_GANG) ||
		    !SPUFS_I(nd->dentry->d_inode)->i_gang)
			goto out;
	}

609 610 611 612 613 614 615 616 617 618 619
	dentry = lookup_create(nd, 1);
	ret = PTR_ERR(dentry);
	if (IS_ERR(dentry))
		goto out_dir;

	ret = -EEXIST;
	if (dentry->d_inode)
		goto out_dput;

	mode &= ~current->fs->umask;

620 621 622 623 624
	if (flags & SPU_CREATE_GANG)
		return spufs_create_gang(nd->dentry->d_inode,
					dentry, nd->mnt, mode);
	else
		return spufs_create_context(nd->dentry->d_inode,
625
					dentry, nd->mnt, flags, mode, filp);
626 627 628 629

out_dput:
	dput(dentry);
out_dir:
630
	mutex_unlock(&nd->dentry->d_inode->i_mutex);
631 632 633 634 635 636
out:
	return ret;
}

/* File system initialization */
enum {
637
	Opt_uid, Opt_gid, Opt_mode, Opt_err,
638 639 640
};

static match_table_t spufs_tokens = {
641 642 643 644
	{ Opt_uid,  "uid=%d" },
	{ Opt_gid,  "gid=%d" },
	{ Opt_mode, "mode=%o" },
	{ Opt_err,   NULL  },
645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670
};

static int
spufs_parse_options(char *options, struct inode *root)
{
	char *p;
	substring_t args[MAX_OPT_ARGS];

	while ((p = strsep(&options, ",")) != NULL) {
		int token, option;

		if (!*p)
			continue;

		token = match_token(p, spufs_tokens, args);
		switch (token) {
		case Opt_uid:
			if (match_int(&args[0], &option))
				return 0;
			root->i_uid = option;
			break;
		case Opt_gid:
			if (match_int(&args[0], &option))
				return 0;
			root->i_gid = option;
			break;
671 672 673 674 675
		case Opt_mode:
			if (match_octal(&args[0], &option))
				return 0;
			root->i_mode = option | S_IFDIR;
			break;
676 677 678 679 680 681 682
		default:
			return 0;
		}
	}
	return 1;
}

683 684
static void spufs_exit_isolated_loader(void)
{
685 686
	free_pages((unsigned long) isolated_loader,
			get_order(isolated_loader_size));
687 688
}

689 690 691 692 693 694 695 696 697 698 699
static void
spufs_init_isolated_loader(void)
{
	struct device_node *dn;
	const char *loader;
	int size;

	dn = of_find_node_by_path("/spu-isolation");
	if (!dn)
		return;

700
	loader = of_get_property(dn, "loader", &size);
701 702 703
	if (!loader)
		return;

704 705
	/* the loader must be align on a 16 byte boundary */
	isolated_loader = (char *)__get_free_pages(GFP_KERNEL, get_order(size));
706 707 708
	if (!isolated_loader)
		return;

709
	isolated_loader_size = size;
710 711 712 713
	memcpy(isolated_loader, loader, size);
	printk(KERN_INFO "spufs: SPU isolation mode enabled\n");
}

714
static int
715 716
spufs_create_root(struct super_block *sb, void *data)
{
717 718 719
	struct inode *inode;
	int ret;

720 721 722 723
	ret = -ENODEV;
	if (!spu_management_ops)
		goto out;

724 725 726 727 728
	ret = -ENOMEM;
	inode = spufs_new_inode(sb, S_IFDIR | 0775);
	if (!inode)
		goto out;

729
	inode->i_op = &simple_dir_inode_operations;
730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768
	inode->i_fop = &simple_dir_operations;
	SPUFS_I(inode)->i_ctx = NULL;

	ret = -EINVAL;
	if (!spufs_parse_options(data, inode))
		goto out_iput;

	ret = -ENOMEM;
	sb->s_root = d_alloc_root(inode);
	if (!sb->s_root)
		goto out_iput;

	return 0;
out_iput:
	iput(inode);
out:
	return ret;
}

static int
spufs_fill_super(struct super_block *sb, void *data, int silent)
{
	static struct super_operations s_ops = {
		.alloc_inode = spufs_alloc_inode,
		.destroy_inode = spufs_destroy_inode,
		.statfs = simple_statfs,
		.delete_inode = spufs_delete_inode,
		.drop_inode = generic_delete_inode,
	};

	sb->s_maxbytes = MAX_LFS_FILESIZE;
	sb->s_blocksize = PAGE_CACHE_SIZE;
	sb->s_blocksize_bits = PAGE_CACHE_SHIFT;
	sb->s_magic = SPUFS_MAGIC;
	sb->s_op = &s_ops;

	return spufs_create_root(sb, data);
}

769
static int
770
spufs_get_sb(struct file_system_type *fstype, int flags,
771
		const char *name, void *data, struct vfsmount *mnt)
772
{
773
	return get_sb_single(fstype, flags, data, spufs_fill_super, mnt);
774 775 776 777 778 779 780 781 782
}

static struct file_system_type spufs_type = {
	.owner = THIS_MODULE,
	.name = "spufs",
	.get_sb = spufs_get_sb,
	.kill_sb = kill_litter_super,
};

783
static int __init spufs_init(void)
784 785
{
	int ret;
786

787 788 789 790
	ret = -ENODEV;
	if (!spu_management_ops)
		goto out;

791 792 793
	ret = -ENOMEM;
	spufs_inode_cache = kmem_cache_create("spufs_inode_cache",
			sizeof(struct spufs_inode_info), 0,
794
			SLAB_HWCACHE_ALIGN, spufs_init_once);
795 796 797

	if (!spufs_inode_cache)
		goto out;
798
	ret = spu_sched_init();
799 800
	if (ret)
		goto out_cache;
801 802 803
	ret = register_filesystem(&spufs_type);
	if (ret)
		goto out_sched;
804
	ret = register_spu_syscalls(&spufs_calls);
805 806
	if (ret)
		goto out_fs;
807 808

	spufs_init_isolated_loader();
809

810
	return 0;
811

812 813
out_fs:
	unregister_filesystem(&spufs_type);
814 815
out_sched:
	spu_sched_exit();
816 817 818 819 820 821 822
out_cache:
	kmem_cache_destroy(spufs_inode_cache);
out:
	return ret;
}
module_init(spufs_init);

823
static void __exit spufs_exit(void)
824
{
825
	spu_sched_exit();
826
	spufs_exit_isolated_loader();
827 828 829 830 831 832 833 834 835
	unregister_spu_syscalls(&spufs_calls);
	unregister_filesystem(&spufs_type);
	kmem_cache_destroy(spufs_inode_cache);
}
module_exit(spufs_exit);

MODULE_LICENSE("GPL");
MODULE_AUTHOR("Arnd Bergmann <arndb@de.ibm.com>");