algapi.c 29.4 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12
/*
 * Cryptographic API for algorithms (i.e., low-level API).
 *
 * Copyright (c) 2006 Herbert Xu <herbert@gondor.apana.org.au>
 *
 * This program is free software; you can redistribute it and/or modify it
 * under the terms of the GNU General Public License as published by the Free
 * Software Foundation; either version 2 of the License, or (at your option)
 * any later version.
 *
 */

13
#include <crypto/algapi.h>
H
Herbert Xu 已提交
14
#include <linux/err.h>
15
#include <linux/errno.h>
H
Herbert Xu 已提交
16
#include <linux/fips.h>
17 18
#include <linux/init.h>
#include <linux/kernel.h>
19
#include <linux/list.h>
20
#include <linux/module.h>
21
#include <linux/rtnetlink.h>
22
#include <linux/slab.h>
23 24 25 26
#include <linux/string.h>

#include "internal.h"

27 28
static LIST_HEAD(crypto_template_list);

29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45
static inline int crypto_set_driver_name(struct crypto_alg *alg)
{
	static const char suffix[] = "-generic";
	char *driver_name = alg->cra_driver_name;
	int len;

	if (*driver_name)
		return 0;

	len = strlcpy(driver_name, alg->cra_name, CRYPTO_MAX_ALG_NAME);
	if (len + sizeof(suffix) > CRYPTO_MAX_ALG_NAME)
		return -ENAMETOOLONG;

	memcpy(driver_name + len, suffix, sizeof(suffix));
	return 0;
}

46 47
static inline void crypto_check_module_sig(struct module *mod)
{
48
	if (fips_enabled && mod && !module_sig_ok(mod))
49
		panic("Module %s signature verification failed in FIPS mode\n",
50
		      module_name(mod));
51 52
}

53
static int crypto_check_alg(struct crypto_alg *alg)
54
{
55 56
	crypto_check_module_sig(alg->cra_module);

57 58 59
	if (alg->cra_alignmask & (alg->cra_alignmask + 1))
		return -EINVAL;

60 61
	/* General maximums for all algs. */
	if (alg->cra_alignmask > MAX_ALGAPI_ALIGNMASK)
62 63
		return -EINVAL;

64 65 66 67
	if (alg->cra_blocksize > MAX_ALGAPI_BLOCKSIZE)
		return -EINVAL;

	/* Lower maximums for specific alg types. */
68 69 70 71 72 73 74 75 76
	if (!alg->cra_type && (alg->cra_flags & CRYPTO_ALG_TYPE_MASK) ==
			       CRYPTO_ALG_TYPE_CIPHER) {
		if (alg->cra_alignmask > MAX_CIPHER_ALIGNMASK)
			return -EINVAL;

		if (alg->cra_blocksize > MAX_CIPHER_BLOCKSIZE)
			return -EINVAL;
	}

77 78 79
	if (alg->cra_priority < 0)
		return -EINVAL;

80
	refcount_set(&alg->cra_refcnt, 1);
81

82 83 84
	return crypto_set_driver_name(alg);
}

85 86 87 88 89 90 91 92 93 94
static void crypto_free_instance(struct crypto_instance *inst)
{
	if (!inst->alg.cra_type->free) {
		inst->tmpl->free(inst);
		return;
	}

	inst->alg.cra_type->free(inst);
}

H
Herbert Xu 已提交
95 96 97 98 99
static void crypto_destroy_instance(struct crypto_alg *alg)
{
	struct crypto_instance *inst = (void *)alg;
	struct crypto_template *tmpl = inst->tmpl;

100
	crypto_free_instance(inst);
H
Herbert Xu 已提交
101 102 103
	crypto_tmpl_put(tmpl);
}

104 105 106 107 108 109 110
static struct list_head *crypto_more_spawns(struct crypto_alg *alg,
					    struct list_head *stack,
					    struct list_head *top,
					    struct list_head *secondary_spawns)
{
	struct crypto_spawn *spawn, *n;

111 112
	spawn = list_first_entry_or_null(stack, struct crypto_spawn, list);
	if (!spawn)
113 114
		return NULL;

115
	n = list_next_entry(spawn, list);
116 117 118

	if (spawn->alg && &n->list != stack && !n->alg)
		n->alg = (n->list.next == stack) ? alg :
119
			 &list_next_entry(n, list)->inst->alg;
120 121 122 123 124 125

	list_move(&spawn->list, secondary_spawns);

	return &n->list == stack ? top : &n->inst->alg.cra_users;
}

126 127
static void crypto_remove_instance(struct crypto_instance *inst,
				   struct list_head *list)
H
Herbert Xu 已提交
128
{
129
	struct crypto_template *tmpl = inst->tmpl;
H
Herbert Xu 已提交
130

131 132
	if (crypto_is_dead(&inst->alg))
		return;
H
Herbert Xu 已提交
133

134
	inst->alg.cra_flags |= CRYPTO_ALG_DEAD;
135 136 137
	if (hlist_unhashed(&inst->list))
		return;

138 139 140 141 142 143 144
	if (!tmpl || !crypto_tmpl_get(tmpl))
		return;

	list_move(&inst->alg.cra_list, list);
	hlist_del(&inst->list);
	inst->alg.cra_destroy = crypto_destroy_instance;

145
	BUG_ON(!list_empty(&inst->alg.cra_users));
146 147
}

148 149
void crypto_remove_spawns(struct crypto_alg *alg, struct list_head *list,
			  struct crypto_alg *nalg)
150
{
151
	u32 new_type = (nalg ?: alg)->cra_flags;
152 153
	struct crypto_spawn *spawn, *n;
	LIST_HEAD(secondary_spawns);
154 155 156
	struct list_head *spawns;
	LIST_HEAD(stack);
	LIST_HEAD(top);
H
Herbert Xu 已提交
157

158
	spawns = &alg->cra_users;
159 160
	list_for_each_entry_safe(spawn, n, spawns, list) {
		if ((spawn->alg->cra_flags ^ new_type) & spawn->mask)
H
Herbert Xu 已提交
161 162
			continue;

163
		list_move(&spawn->list, &top);
164
	}
H
Herbert Xu 已提交
165

166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183
	spawns = &top;
	do {
		while (!list_empty(spawns)) {
			struct crypto_instance *inst;

			spawn = list_first_entry(spawns, struct crypto_spawn,
						 list);
			inst = spawn->inst;

			BUG_ON(&inst->alg == alg);

			list_move(&spawn->list, &stack);

			if (&inst->alg == nalg)
				break;

			spawn->alg = NULL;
			spawns = &inst->alg.cra_users;
184 185 186 187 188 189 190 191 192 193 194 195

			/*
			 * We may encounter an unregistered instance here, since
			 * an instance's spawns are set up prior to the instance
			 * being registered.  An unregistered instance will have
			 * NULL ->cra_users.next, since ->cra_users isn't
			 * properly initialized until registration.  But an
			 * unregistered instance cannot have any users, so treat
			 * it the same as ->cra_users being empty.
			 */
			if (spawns->next == NULL)
				break;
196 197 198 199 200 201 202 203
		}
	} while ((spawns = crypto_more_spawns(alg, &stack, &top,
					      &secondary_spawns)));

	list_for_each_entry_safe(spawn, n, &secondary_spawns, list) {
		if (spawn->alg)
			list_move(&spawn->list, &spawn->alg->cra_users);
		else
204
			crypto_remove_instance(spawn->inst, list);
H
Herbert Xu 已提交
205 206
	}
}
207
EXPORT_SYMBOL_GPL(crypto_remove_spawns);
H
Herbert Xu 已提交
208

209
static struct crypto_larval *__crypto_register_alg(struct crypto_alg *alg)
210 211
{
	struct crypto_alg *q;
212
	struct crypto_larval *larval;
H
Herbert Xu 已提交
213 214 215
	int ret = -EAGAIN;

	if (crypto_is_dead(alg))
216
		goto err;
H
Herbert Xu 已提交
217 218 219

	INIT_LIST_HEAD(&alg->cra_users);

220 221 222
	/* No cheating! */
	alg->cra_flags &= ~CRYPTO_ALG_TESTED;

H
Herbert Xu 已提交
223
	ret = -EEXIST;
224

225
	list_for_each_entry(q, &crypto_alg_list, cra_list) {
226
		if (q == alg)
227 228
			goto err;

229 230 231
		if (crypto_is_moribund(q))
			continue;

232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252
		if (crypto_is_larval(q)) {
			if (!strcmp(alg->cra_driver_name, q->cra_driver_name))
				goto err;
			continue;
		}

		if (!strcmp(q->cra_driver_name, alg->cra_name) ||
		    !strcmp(q->cra_name, alg->cra_driver_name))
			goto err;
	}

	larval = crypto_larval_alloc(alg->cra_name,
				     alg->cra_flags | CRYPTO_ALG_TESTED, 0);
	if (IS_ERR(larval))
		goto out;

	ret = -ENOENT;
	larval->adult = crypto_mod_get(alg);
	if (!larval->adult)
		goto free_larval;

253
	refcount_set(&larval->alg.cra_refcnt, 1);
254 255 256 257 258 259 260
	memcpy(larval->alg.cra_driver_name, alg->cra_driver_name,
	       CRYPTO_MAX_ALG_NAME);
	larval->alg.cra_priority = alg->cra_priority;

	list_add(&alg->cra_list, &crypto_alg_list);
	list_add(&larval->alg.cra_list, &crypto_alg_list);

261
#ifdef CONFIG_CRYPTO_STATS
262
	memset(&alg->stats, 0, sizeof(alg->stats));
263
#endif
264

265
out:
266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283
	return larval;

free_larval:
	kfree(larval);
err:
	larval = ERR_PTR(ret);
	goto out;
}

void crypto_alg_tested(const char *name, int err)
{
	struct crypto_larval *test;
	struct crypto_alg *alg;
	struct crypto_alg *q;
	LIST_HEAD(list);

	down_write(&crypto_alg_sem);
	list_for_each_entry(q, &crypto_alg_list, cra_list) {
284
		if (crypto_is_moribund(q) || !crypto_is_larval(q))
285 286 287 288 289 290 291 292
			continue;

		test = (struct crypto_larval *)q;

		if (!strcmp(q->cra_driver_name, name))
			goto found;
	}

293
	pr_err("alg: Unexpected test result for %s: %d\n", name, err);
294 295 296
	goto unlock;

found:
297
	q->cra_flags |= CRYPTO_ALG_DEAD;
298 299 300 301 302 303 304 305 306
	alg = test->adult;
	if (err || list_empty(&alg->cra_list))
		goto complete;

	alg->cra_flags |= CRYPTO_ALG_TESTED;

	list_for_each_entry(q, &crypto_alg_list, cra_list) {
		if (q == alg)
			continue;
H
Herbert Xu 已提交
307 308 309 310 311

		if (crypto_is_moribund(q))
			continue;

		if (crypto_is_larval(q)) {
312 313
			struct crypto_larval *larval = (void *)q;

314 315 316 317 318
			/*
			 * Check to see if either our generic name or
			 * specific name can satisfy the name requested
			 * by the larval entry q.
			 */
H
Herbert Xu 已提交
319 320 321 322 323 324
			if (strcmp(alg->cra_name, q->cra_name) &&
			    strcmp(alg->cra_driver_name, q->cra_name))
				continue;

			if (larval->adult)
				continue;
325 326
			if ((q->cra_flags ^ alg->cra_flags) & larval->mask)
				continue;
327 328
			if (!crypto_mod_get(alg))
				continue;
H
Herbert Xu 已提交
329

330
			larval->adult = alg;
H
Herbert Xu 已提交
331
			continue;
332
		}
H
Herbert Xu 已提交
333 334 335 336 337 338 339 340

		if (strcmp(alg->cra_name, q->cra_name))
			continue;

		if (strcmp(alg->cra_driver_name, q->cra_driver_name) &&
		    q->cra_priority > alg->cra_priority)
			continue;

341
		crypto_remove_spawns(q, &list, alg);
342
	}
343

344 345
complete:
	complete_all(&test->completion);
346

347 348 349 350
unlock:
	up_write(&crypto_alg_sem);

	crypto_remove_final(&list);
351
}
352
EXPORT_SYMBOL_GPL(crypto_alg_tested);
353

354
void crypto_remove_final(struct list_head *list)
H
Herbert Xu 已提交
355 356 357 358 359 360 361 362 363
{
	struct crypto_alg *alg;
	struct crypto_alg *n;

	list_for_each_entry_safe(alg, n, list, cra_list) {
		list_del_init(&alg->cra_list);
		crypto_alg_put(alg);
	}
}
364
EXPORT_SYMBOL_GPL(crypto_remove_final);
H
Herbert Xu 已提交
365

366 367 368 369 370 371 372 373 374 375 376
static void crypto_wait_for_test(struct crypto_larval *larval)
{
	int err;

	err = crypto_probing_notify(CRYPTO_MSG_ALG_REGISTER, larval->adult);
	if (err != NOTIFY_STOP) {
		if (WARN_ON(err != NOTIFY_DONE))
			goto out;
		crypto_alg_tested(larval->alg.cra_driver_name, 0);
	}

377
	err = wait_for_completion_killable(&larval->completion);
378
	WARN_ON(err);
379 380
	if (!err)
		crypto_probing_notify(CRYPTO_MSG_ALG_LOADED, larval);
381 382 383 384 385

out:
	crypto_larval_kill(&larval->alg);
}

386 387
int crypto_register_alg(struct crypto_alg *alg)
{
388
	struct crypto_larval *larval;
389 390
	int err;

391
	alg->cra_flags &= ~CRYPTO_ALG_DEAD;
392 393 394 395 396
	err = crypto_check_alg(alg);
	if (err)
		return err;

	down_write(&crypto_alg_sem);
397
	larval = __crypto_register_alg(alg);
398 399
	up_write(&crypto_alg_sem);

400 401 402 403 404
	if (IS_ERR(larval))
		return PTR_ERR(larval);

	crypto_wait_for_test(larval);
	return 0;
405
}
406 407
EXPORT_SYMBOL_GPL(crypto_register_alg);

H
Herbert Xu 已提交
408 409 410 411 412 413 414 415
static int crypto_remove_alg(struct crypto_alg *alg, struct list_head *list)
{
	if (unlikely(list_empty(&alg->cra_list)))
		return -ENOENT;

	alg->cra_flags |= CRYPTO_ALG_DEAD;

	list_del_init(&alg->cra_list);
416
	crypto_remove_spawns(alg, list, NULL);
H
Herbert Xu 已提交
417 418 419 420

	return 0;
}

421 422
int crypto_unregister_alg(struct crypto_alg *alg)
{
H
Herbert Xu 已提交
423 424
	int ret;
	LIST_HEAD(list);
425

426
	down_write(&crypto_alg_sem);
H
Herbert Xu 已提交
427
	ret = crypto_remove_alg(alg, &list);
428 429 430 431 432
	up_write(&crypto_alg_sem);

	if (ret)
		return ret;

433
	BUG_ON(refcount_read(&alg->cra_refcnt) != 1);
434 435 436
	if (alg->cra_destroy)
		alg->cra_destroy(alg);

H
Herbert Xu 已提交
437
	crypto_remove_final(&list);
438 439 440 441
	return 0;
}
EXPORT_SYMBOL_GPL(crypto_unregister_alg);

442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476
int crypto_register_algs(struct crypto_alg *algs, int count)
{
	int i, ret;

	for (i = 0; i < count; i++) {
		ret = crypto_register_alg(&algs[i]);
		if (ret)
			goto err;
	}

	return 0;

err:
	for (--i; i >= 0; --i)
		crypto_unregister_alg(&algs[i]);

	return ret;
}
EXPORT_SYMBOL_GPL(crypto_register_algs);

int crypto_unregister_algs(struct crypto_alg *algs, int count)
{
	int i, ret;

	for (i = 0; i < count; i++) {
		ret = crypto_unregister_alg(&algs[i]);
		if (ret)
			pr_err("Failed to unregister %s %s: %d\n",
			       algs[i].cra_driver_name, algs[i].cra_name, ret);
	}

	return 0;
}
EXPORT_SYMBOL_GPL(crypto_unregister_algs);

477 478 479 480 481 482 483
int crypto_register_template(struct crypto_template *tmpl)
{
	struct crypto_template *q;
	int err = -EEXIST;

	down_write(&crypto_alg_sem);

484 485
	crypto_check_module_sig(tmpl->module);

486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501
	list_for_each_entry(q, &crypto_template_list, list) {
		if (q == tmpl)
			goto out;
	}

	list_add(&tmpl->list, &crypto_template_list);
	err = 0;
out:
	up_write(&crypto_alg_sem);
	return err;
}
EXPORT_SYMBOL_GPL(crypto_register_template);

void crypto_unregister_template(struct crypto_template *tmpl)
{
	struct crypto_instance *inst;
502
	struct hlist_node *n;
503
	struct hlist_head *list;
H
Herbert Xu 已提交
504
	LIST_HEAD(users);
505 506 507 508 509 510 511

	down_write(&crypto_alg_sem);

	BUG_ON(list_empty(&tmpl->list));
	list_del_init(&tmpl->list);

	list = &tmpl->instances;
512
	hlist_for_each_entry(inst, list, list) {
H
Herbert Xu 已提交
513
		int err = crypto_remove_alg(&inst->alg, &users);
514

H
Herbert Xu 已提交
515
		BUG_ON(err);
516 517 518 519
	}

	up_write(&crypto_alg_sem);

520
	hlist_for_each_entry_safe(inst, n, list, list) {
521
		BUG_ON(refcount_read(&inst->alg.cra_refcnt) != 1);
522
		crypto_free_instance(inst);
523
	}
H
Herbert Xu 已提交
524
	crypto_remove_final(&users);
525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548
}
EXPORT_SYMBOL_GPL(crypto_unregister_template);

static struct crypto_template *__crypto_lookup_template(const char *name)
{
	struct crypto_template *q, *tmpl = NULL;

	down_read(&crypto_alg_sem);
	list_for_each_entry(q, &crypto_template_list, list) {
		if (strcmp(q->name, name))
			continue;
		if (unlikely(!crypto_tmpl_get(q)))
			continue;

		tmpl = q;
		break;
	}
	up_read(&crypto_alg_sem);

	return tmpl;
}

struct crypto_template *crypto_lookup_template(const char *name)
{
549 550
	return try_then_request_module(__crypto_lookup_template(name),
				       "crypto-%s", name);
551 552 553 554 555 556
}
EXPORT_SYMBOL_GPL(crypto_lookup_template);

int crypto_register_instance(struct crypto_template *tmpl,
			     struct crypto_instance *inst)
{
557 558
	struct crypto_larval *larval;
	int err;
559 560 561

	err = crypto_check_alg(&inst->alg);
	if (err)
562 563
		return err;

564
	inst->alg.cra_module = tmpl->module;
565
	inst->alg.cra_flags |= CRYPTO_ALG_INSTANCE;
566 567 568

	down_write(&crypto_alg_sem);

569 570
	larval = __crypto_register_alg(&inst->alg);
	if (IS_ERR(larval))
571 572 573 574 575 576 577 578
		goto unlock;

	hlist_add_head(&inst->list, &tmpl->instances);
	inst->tmpl = tmpl;

unlock:
	up_write(&crypto_alg_sem);

579 580 581 582 583 584
	err = PTR_ERR(larval);
	if (IS_ERR(larval))
		goto err;

	crypto_wait_for_test(larval);
	err = 0;
H
Herbert Xu 已提交
585

586 587 588 589
err:
	return err;
}
EXPORT_SYMBOL_GPL(crypto_register_instance);
590

591
int crypto_unregister_instance(struct crypto_instance *inst)
592
{
593
	LIST_HEAD(list);
594 595 596

	down_write(&crypto_alg_sem);

597
	crypto_remove_spawns(&inst->alg, &list, NULL);
598
	crypto_remove_instance(inst, &list);
599 600 601

	up_write(&crypto_alg_sem);

602
	crypto_remove_final(&list);
603 604 605 606

	return 0;
}
EXPORT_SYMBOL_GPL(crypto_unregister_instance);
607

H
Herbert Xu 已提交
608
int crypto_init_spawn(struct crypto_spawn *spawn, struct crypto_alg *alg,
609
		      struct crypto_instance *inst, u32 mask)
H
Herbert Xu 已提交
610 611 612 613
{
	int err = -EAGAIN;

	spawn->inst = inst;
614
	spawn->mask = mask;
H
Herbert Xu 已提交
615 616 617 618 619 620 621 622 623 624 625 626 627

	down_write(&crypto_alg_sem);
	if (!crypto_is_moribund(alg)) {
		list_add(&spawn->list, &alg->cra_users);
		spawn->alg = alg;
		err = 0;
	}
	up_write(&crypto_alg_sem);

	return err;
}
EXPORT_SYMBOL_GPL(crypto_init_spawn);

628 629 630 631 632 633
int crypto_init_spawn2(struct crypto_spawn *spawn, struct crypto_alg *alg,
		       struct crypto_instance *inst,
		       const struct crypto_type *frontend)
{
	int err = -EINVAL;

634
	if ((alg->cra_flags ^ frontend->type) & frontend->maskset)
635 636 637 638 639 640 641 642 643 644
		goto out;

	spawn->frontend = frontend;
	err = crypto_init_spawn(spawn, alg, inst, frontend->maskset);

out:
	return err;
}
EXPORT_SYMBOL_GPL(crypto_init_spawn2);

645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660
int crypto_grab_spawn(struct crypto_spawn *spawn, const char *name,
		      u32 type, u32 mask)
{
	struct crypto_alg *alg;
	int err;

	alg = crypto_find_alg(name, spawn->frontend, type, mask);
	if (IS_ERR(alg))
		return PTR_ERR(alg);

	err = crypto_init_spawn(spawn, alg, spawn->inst, mask);
	crypto_mod_put(alg);
	return err;
}
EXPORT_SYMBOL_GPL(crypto_grab_spawn);

H
Herbert Xu 已提交
661 662
void crypto_drop_spawn(struct crypto_spawn *spawn)
{
663 664 665
	if (!spawn->alg)
		return;

H
Herbert Xu 已提交
666 667 668 669 670 671
	down_write(&crypto_alg_sem);
	list_del(&spawn->list);
	up_write(&crypto_alg_sem);
}
EXPORT_SYMBOL_GPL(crypto_drop_spawn);

672
static struct crypto_alg *crypto_spawn_alg(struct crypto_spawn *spawn)
H
Herbert Xu 已提交
673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689
{
	struct crypto_alg *alg;
	struct crypto_alg *alg2;

	down_read(&crypto_alg_sem);
	alg = spawn->alg;
	alg2 = alg;
	if (alg2)
		alg2 = crypto_mod_get(alg2);
	up_read(&crypto_alg_sem);

	if (!alg2) {
		if (alg)
			crypto_shoot_alg(alg);
		return ERR_PTR(-EAGAIN);
	}

690 691 692 693 694 695 696 697 698 699 700 701 702
	return alg;
}

struct crypto_tfm *crypto_spawn_tfm(struct crypto_spawn *spawn, u32 type,
				    u32 mask)
{
	struct crypto_alg *alg;
	struct crypto_tfm *tfm;

	alg = crypto_spawn_alg(spawn);
	if (IS_ERR(alg))
		return ERR_CAST(alg);

H
Herbert Xu 已提交
703 704 705 706
	tfm = ERR_PTR(-EINVAL);
	if (unlikely((alg->cra_flags ^ type) & mask))
		goto out_put_alg;

707
	tfm = __crypto_alloc_tfm(alg, type, mask);
H
Herbert Xu 已提交
708
	if (IS_ERR(tfm))
H
Herbert Xu 已提交
709 710 711
		goto out_put_alg;

	return tfm;
H
Herbert Xu 已提交
712

H
Herbert Xu 已提交
713 714
out_put_alg:
	crypto_mod_put(alg);
H
Herbert Xu 已提交
715 716 717 718
	return tfm;
}
EXPORT_SYMBOL_GPL(crypto_spawn_tfm);

719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739
void *crypto_spawn_tfm2(struct crypto_spawn *spawn)
{
	struct crypto_alg *alg;
	struct crypto_tfm *tfm;

	alg = crypto_spawn_alg(spawn);
	if (IS_ERR(alg))
		return ERR_CAST(alg);

	tfm = crypto_create_tfm(alg, spawn->frontend);
	if (IS_ERR(tfm))
		goto out_put_alg;

	return tfm;

out_put_alg:
	crypto_mod_put(alg);
	return tfm;
}
EXPORT_SYMBOL_GPL(crypto_spawn_tfm2);

740 741 742 743 744 745 746 747 748 749 750 751
int crypto_register_notifier(struct notifier_block *nb)
{
	return blocking_notifier_chain_register(&crypto_chain, nb);
}
EXPORT_SYMBOL_GPL(crypto_register_notifier);

int crypto_unregister_notifier(struct notifier_block *nb)
{
	return blocking_notifier_chain_unregister(&crypto_chain, nb);
}
EXPORT_SYMBOL_GPL(crypto_unregister_notifier);

752
struct crypto_attr_type *crypto_get_attr_type(struct rtattr **tb)
753
{
754
	struct rtattr *rta = tb[0];
755 756 757 758 759 760
	struct crypto_attr_type *algt;

	if (!rta)
		return ERR_PTR(-ENOENT);
	if (RTA_PAYLOAD(rta) < sizeof(*algt))
		return ERR_PTR(-EINVAL);
761 762
	if (rta->rta_type != CRYPTOA_TYPE)
		return ERR_PTR(-EINVAL);
763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784

	algt = RTA_DATA(rta);

	return algt;
}
EXPORT_SYMBOL_GPL(crypto_get_attr_type);

int crypto_check_attr_type(struct rtattr **tb, u32 type)
{
	struct crypto_attr_type *algt;

	algt = crypto_get_attr_type(tb);
	if (IS_ERR(algt))
		return PTR_ERR(algt);

	if ((algt->type ^ type) & algt->mask)
		return -EINVAL;

	return 0;
}
EXPORT_SYMBOL_GPL(crypto_check_attr_type);

785
const char *crypto_attr_alg_name(struct rtattr *rta)
786
{
787 788
	struct crypto_attr_alg *alga;

789 790 791
	if (!rta)
		return ERR_PTR(-ENOENT);
	if (RTA_PAYLOAD(rta) < sizeof(*alga))
792
		return ERR_PTR(-EINVAL);
793 794
	if (rta->rta_type != CRYPTOA_ALG)
		return ERR_PTR(-EINVAL);
795 796 797 798

	alga = RTA_DATA(rta);
	alga->name[CRYPTO_MAX_ALG_NAME - 1] = 0;

799 800 801 802
	return alga->name;
}
EXPORT_SYMBOL_GPL(crypto_attr_alg_name);

803 804 805
struct crypto_alg *crypto_attr_alg2(struct rtattr *rta,
				    const struct crypto_type *frontend,
				    u32 type, u32 mask)
806 807 808 809 810
{
	const char *name;

	name = crypto_attr_alg_name(rta);
	if (IS_ERR(name))
J
Julia Lawall 已提交
811
		return ERR_CAST(name);
812

813
	return crypto_find_alg(name, frontend, type, mask);
814
}
815
EXPORT_SYMBOL_GPL(crypto_attr_alg2);
H
Herbert Xu 已提交
816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833

int crypto_attr_u32(struct rtattr *rta, u32 *num)
{
	struct crypto_attr_u32 *nu32;

	if (!rta)
		return -ENOENT;
	if (RTA_PAYLOAD(rta) < sizeof(*nu32))
		return -EINVAL;
	if (rta->rta_type != CRYPTOA_U32)
		return -EINVAL;

	nu32 = RTA_DATA(rta);
	*num = nu32->num;

	return 0;
}
EXPORT_SYMBOL_GPL(crypto_attr_u32);
834

835 836 837 838 839 840 841 842 843 844 845 846 847 848 849
int crypto_inst_setname(struct crypto_instance *inst, const char *name,
			struct crypto_alg *alg)
{
	if (snprintf(inst->alg.cra_name, CRYPTO_MAX_ALG_NAME, "%s(%s)", name,
		     alg->cra_name) >= CRYPTO_MAX_ALG_NAME)
		return -ENAMETOOLONG;

	if (snprintf(inst->alg.cra_driver_name, CRYPTO_MAX_ALG_NAME, "%s(%s)",
		     name, alg->cra_driver_name) >= CRYPTO_MAX_ALG_NAME)
		return -ENAMETOOLONG;

	return 0;
}
EXPORT_SYMBOL_GPL(crypto_inst_setname);

850 851
void *crypto_alloc_instance2(const char *name, struct crypto_alg *alg,
			     unsigned int head)
852 853
{
	struct crypto_instance *inst;
854
	char *p;
855 856
	int err;

857 858 859
	p = kzalloc(head + sizeof(*inst) + sizeof(struct crypto_spawn),
		    GFP_KERNEL);
	if (!p)
860 861
		return ERR_PTR(-ENOMEM);

862 863
	inst = (void *)(p + head);

864 865
	err = crypto_inst_setname(inst, name, alg);
	if (err)
866 867
		goto err_free_inst;

868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886
	return p;

err_free_inst:
	kfree(p);
	return ERR_PTR(err);
}
EXPORT_SYMBOL_GPL(crypto_alloc_instance2);

struct crypto_instance *crypto_alloc_instance(const char *name,
					      struct crypto_alg *alg)
{
	struct crypto_instance *inst;
	struct crypto_spawn *spawn;
	int err;

	inst = crypto_alloc_instance2(name, alg, 0);
	if (IS_ERR(inst))
		goto out;

887
	spawn = crypto_instance_ctx(inst);
888 889
	err = crypto_init_spawn(spawn, alg, inst,
				CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_ASYNC);
890 891 892 893 894 895 896 897

	if (err)
		goto err_free_inst;

	return inst;

err_free_inst:
	kfree(inst);
898 899 900 901
	inst = ERR_PTR(err);

out:
	return inst;
902 903 904
}
EXPORT_SYMBOL_GPL(crypto_alloc_instance);

905 906 907 908 909 910 911 912 913 914 915 916 917 918 919
void crypto_init_queue(struct crypto_queue *queue, unsigned int max_qlen)
{
	INIT_LIST_HEAD(&queue->list);
	queue->backlog = &queue->list;
	queue->qlen = 0;
	queue->max_qlen = max_qlen;
}
EXPORT_SYMBOL_GPL(crypto_init_queue);

int crypto_enqueue_request(struct crypto_queue *queue,
			   struct crypto_async_request *request)
{
	int err = -EINPROGRESS;

	if (unlikely(queue->qlen >= queue->max_qlen)) {
920 921
		if (!(request->flags & CRYPTO_TFM_REQ_MAY_BACKLOG)) {
			err = -ENOSPC;
922
			goto out;
923 924
		}
		err = -EBUSY;
925 926 927 928 929 930 931 932 933 934 935 936
		if (queue->backlog == &queue->list)
			queue->backlog = &request->list;
	}

	queue->qlen++;
	list_add_tail(&request->list, &queue->list);

out:
	return err;
}
EXPORT_SYMBOL_GPL(crypto_enqueue_request);

937
struct crypto_async_request *crypto_dequeue_request(struct crypto_queue *queue)
938 939 940 941 942 943 944 945 946 947 948 949 950 951
{
	struct list_head *request;

	if (unlikely(!queue->qlen))
		return NULL;

	queue->qlen--;

	if (queue->backlog != &queue->list)
		queue->backlog = queue->backlog->next;

	request = queue->list.next;
	list_del(request);

952
	return list_entry(request, struct crypto_async_request, list);
953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968
}
EXPORT_SYMBOL_GPL(crypto_dequeue_request);

int crypto_tfm_in_queue(struct crypto_queue *queue, struct crypto_tfm *tfm)
{
	struct crypto_async_request *req;

	list_for_each_entry(req, &queue->list, list) {
		if (req->tfm == tfm)
			return 1;
	}

	return 0;
}
EXPORT_SYMBOL_GPL(crypto_tfm_in_queue);

969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986
static inline void crypto_inc_byte(u8 *a, unsigned int size)
{
	u8 *b = (a + size);
	u8 c;

	for (; size; size--) {
		c = *--b + 1;
		*b = c;
		if (c)
			break;
	}
}

void crypto_inc(u8 *a, unsigned int size)
{
	__be32 *b = (__be32 *)(a + size);
	u32 c;

987
	if (IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) ||
988
	    IS_ALIGNED((unsigned long)b, __alignof__(*b)))
989 990 991
		for (; size >= 4; size -= 4) {
			c = be32_to_cpu(*--b) + 1;
			*b = cpu_to_be32(c);
992
			if (likely(c))
993 994
				return;
		}
995 996 997 998 999

	crypto_inc_byte(a, size);
}
EXPORT_SYMBOL_GPL(crypto_inc);

1000
void __crypto_xor(u8 *dst, const u8 *src1, const u8 *src2, unsigned int len)
1001
{
1002 1003 1004 1005
	int relalign = 0;

	if (!IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS)) {
		int size = sizeof(unsigned long);
1006 1007 1008
		int d = (((unsigned long)dst ^ (unsigned long)src1) |
			 ((unsigned long)dst ^ (unsigned long)src2)) &
			(size - 1);
1009 1010 1011 1012 1013 1014 1015 1016 1017 1018

		relalign = d ? 1 << __ffs(d) : size;

		/*
		 * If we care about alignment, process as many bytes as
		 * needed to advance dst and src to values whose alignments
		 * equal their relative alignment. This will allow us to
		 * process the remainder of the input using optimal strides.
		 */
		while (((unsigned long)dst & (relalign - 1)) && len > 0) {
1019
			*dst++ = *src1++ ^ *src2++;
1020 1021 1022
			len--;
		}
	}
1023

1024
	while (IS_ENABLED(CONFIG_64BIT) && len >= 8 && !(relalign & 7)) {
1025
		*(u64 *)dst = *(u64 *)src1 ^  *(u64 *)src2;
1026
		dst += 8;
1027 1028
		src1 += 8;
		src2 += 8;
1029 1030
		len -= 8;
	}
1031

1032
	while (len >= 4 && !(relalign & 3)) {
1033
		*(u32 *)dst = *(u32 *)src1 ^ *(u32 *)src2;
1034
		dst += 4;
1035 1036
		src1 += 4;
		src2 += 4;
1037 1038 1039 1040
		len -= 4;
	}

	while (len >= 2 && !(relalign & 1)) {
1041
		*(u16 *)dst = *(u16 *)src1 ^ *(u16 *)src2;
1042
		dst += 2;
1043 1044
		src1 += 2;
		src2 += 2;
1045 1046
		len -= 2;
	}
1047

1048
	while (len--)
1049
		*dst++ = *src1++ ^ *src2++;
1050
}
1051
EXPORT_SYMBOL_GPL(__crypto_xor);
1052

1053 1054
unsigned int crypto_alg_extsize(struct crypto_alg *alg)
{
1055 1056
	return alg->cra_ctxsize +
	       (alg->cra_alignmask & ~(crypto_tfm_ctx_alignment() - 1));
1057 1058 1059
}
EXPORT_SYMBOL_GPL(crypto_alg_extsize);

1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074
int crypto_type_has_alg(const char *name, const struct crypto_type *frontend,
			u32 type, u32 mask)
{
	int ret = 0;
	struct crypto_alg *alg = crypto_find_alg(name, frontend, type, mask);

	if (!IS_ERR(alg)) {
		crypto_mod_put(alg);
		ret = 1;
	}

	return ret;
}
EXPORT_SYMBOL_GPL(crypto_type_has_alg);

1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085
#ifdef CONFIG_CRYPTO_STATS
void crypto_stats_get(struct crypto_alg *alg)
{
	crypto_alg_get(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_get);

void crypto_stats_ablkcipher_encrypt(unsigned int nbytes, int ret,
				     struct crypto_alg *alg)
{
	if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
1086
		atomic64_inc(&alg->stats.cipher.cipher_err_cnt);
1087
	} else {
1088 1089
		atomic64_inc(&alg->stats.cipher.encrypt_cnt);
		atomic64_add(nbytes, &alg->stats.cipher.encrypt_tlen);
1090 1091 1092 1093 1094 1095 1096 1097 1098
	}
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_ablkcipher_encrypt);

void crypto_stats_ablkcipher_decrypt(unsigned int nbytes, int ret,
				     struct crypto_alg *alg)
{
	if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
1099
		atomic64_inc(&alg->stats.cipher.cipher_err_cnt);
1100
	} else {
1101 1102
		atomic64_inc(&alg->stats.cipher.decrypt_cnt);
		atomic64_add(nbytes, &alg->stats.cipher.decrypt_tlen);
1103 1104 1105 1106 1107 1108 1109 1110 1111
	}
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_ablkcipher_decrypt);

void crypto_stats_aead_encrypt(unsigned int cryptlen, struct crypto_alg *alg,
			       int ret)
{
	if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
1112
		atomic64_inc(&alg->stats.aead.aead_err_cnt);
1113
	} else {
1114 1115
		atomic64_inc(&alg->stats.aead.encrypt_cnt);
		atomic64_add(cryptlen, &alg->stats.aead.encrypt_tlen);
1116 1117 1118 1119 1120 1121 1122 1123 1124
	}
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_aead_encrypt);

void crypto_stats_aead_decrypt(unsigned int cryptlen, struct crypto_alg *alg,
			       int ret)
{
	if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
1125
		atomic64_inc(&alg->stats.aead.aead_err_cnt);
1126
	} else {
1127 1128
		atomic64_inc(&alg->stats.aead.decrypt_cnt);
		atomic64_add(cryptlen, &alg->stats.aead.decrypt_tlen);
1129 1130 1131 1132 1133 1134 1135 1136 1137
	}
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_aead_decrypt);

void crypto_stats_akcipher_encrypt(unsigned int src_len, int ret,
				   struct crypto_alg *alg)
{
	if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
1138
		atomic64_inc(&alg->stats.akcipher.akcipher_err_cnt);
1139
	} else {
1140 1141
		atomic64_inc(&alg->stats.akcipher.encrypt_cnt);
		atomic64_add(src_len, &alg->stats.akcipher.encrypt_tlen);
1142 1143 1144 1145 1146 1147 1148 1149 1150
	}
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_akcipher_encrypt);

void crypto_stats_akcipher_decrypt(unsigned int src_len, int ret,
				   struct crypto_alg *alg)
{
	if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
1151
		atomic64_inc(&alg->stats.akcipher.akcipher_err_cnt);
1152
	} else {
1153 1154
		atomic64_inc(&alg->stats.akcipher.decrypt_cnt);
		atomic64_add(src_len, &alg->stats.akcipher.decrypt_tlen);
1155 1156 1157 1158 1159 1160 1161 1162
	}
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_akcipher_decrypt);

void crypto_stats_akcipher_sign(int ret, struct crypto_alg *alg)
{
	if (ret && ret != -EINPROGRESS && ret != -EBUSY)
1163
		atomic64_inc(&alg->stats.akcipher.akcipher_err_cnt);
1164
	else
1165
		atomic64_inc(&alg->stats.akcipher.sign_cnt);
1166 1167 1168 1169 1170 1171 1172
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_akcipher_sign);

void crypto_stats_akcipher_verify(int ret, struct crypto_alg *alg)
{
	if (ret && ret != -EINPROGRESS && ret != -EBUSY)
1173
		atomic64_inc(&alg->stats.akcipher.akcipher_err_cnt);
1174
	else
1175
		atomic64_inc(&alg->stats.akcipher.verify_cnt);
1176 1177 1178 1179 1180 1181 1182
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_akcipher_verify);

void crypto_stats_compress(unsigned int slen, int ret, struct crypto_alg *alg)
{
	if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
1183
		atomic64_inc(&alg->stats.compress.compress_err_cnt);
1184
	} else {
1185 1186
		atomic64_inc(&alg->stats.compress.compress_cnt);
		atomic64_add(slen, &alg->stats.compress.compress_tlen);
1187 1188 1189 1190 1191 1192 1193 1194
	}
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_compress);

void crypto_stats_decompress(unsigned int slen, int ret, struct crypto_alg *alg)
{
	if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
1195
		atomic64_inc(&alg->stats.compress.compress_err_cnt);
1196
	} else {
1197 1198
		atomic64_inc(&alg->stats.compress.decompress_cnt);
		atomic64_add(slen, &alg->stats.compress.decompress_tlen);
1199 1200 1201 1202 1203 1204 1205 1206 1207
	}
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_decompress);

void crypto_stats_ahash_update(unsigned int nbytes, int ret,
			       struct crypto_alg *alg)
{
	if (ret && ret != -EINPROGRESS && ret != -EBUSY)
1208
		atomic64_inc(&alg->stats.hash.hash_err_cnt);
1209
	else
1210
		atomic64_add(nbytes, &alg->stats.hash.hash_tlen);
1211 1212 1213 1214 1215 1216 1217 1218
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_ahash_update);

void crypto_stats_ahash_final(unsigned int nbytes, int ret,
			      struct crypto_alg *alg)
{
	if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
1219
		atomic64_inc(&alg->stats.hash.hash_err_cnt);
1220
	} else {
1221 1222
		atomic64_inc(&alg->stats.hash.hash_cnt);
		atomic64_add(nbytes, &alg->stats.hash.hash_tlen);
1223 1224 1225 1226 1227 1228 1229 1230
	}
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_ahash_final);

void crypto_stats_kpp_set_secret(struct crypto_alg *alg, int ret)
{
	if (ret)
1231
		atomic64_inc(&alg->stats.kpp.kpp_err_cnt);
1232
	else
1233
		atomic64_inc(&alg->stats.kpp.setsecret_cnt);
1234 1235 1236 1237 1238 1239 1240
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_kpp_set_secret);

void crypto_stats_kpp_generate_public_key(struct crypto_alg *alg, int ret)
{
	if (ret)
1241
		atomic64_inc(&alg->stats.kpp.kpp_err_cnt);
1242
	else
1243
		atomic64_inc(&alg->stats.kpp.generate_public_key_cnt);
1244 1245 1246 1247 1248 1249 1250
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_kpp_generate_public_key);

void crypto_stats_kpp_compute_shared_secret(struct crypto_alg *alg, int ret)
{
	if (ret)
1251
		atomic64_inc(&alg->stats.kpp.kpp_err_cnt);
1252
	else
1253
		atomic64_inc(&alg->stats.kpp.compute_shared_secret_cnt);
1254 1255 1256 1257 1258 1259 1260
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_kpp_compute_shared_secret);

void crypto_stats_rng_seed(struct crypto_alg *alg, int ret)
{
	if (ret && ret != -EINPROGRESS && ret != -EBUSY)
1261
		atomic64_inc(&alg->stats.rng.rng_err_cnt);
1262
	else
1263
		atomic64_inc(&alg->stats.rng.seed_cnt);
1264 1265 1266 1267 1268 1269 1270 1271
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_rng_seed);

void crypto_stats_rng_generate(struct crypto_alg *alg, unsigned int dlen,
			       int ret)
{
	if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
1272
		atomic64_inc(&alg->stats.rng.rng_err_cnt);
1273
	} else {
1274 1275
		atomic64_inc(&alg->stats.rng.generate_cnt);
		atomic64_add(dlen, &alg->stats.rng.generate_tlen);
1276 1277 1278 1279 1280 1281 1282 1283 1284
	}
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_rng_generate);

void crypto_stats_skcipher_encrypt(unsigned int cryptlen, int ret,
				   struct crypto_alg *alg)
{
	if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
1285
		atomic64_inc(&alg->stats.cipher.cipher_err_cnt);
1286
	} else {
1287 1288
		atomic64_inc(&alg->stats.cipher.encrypt_cnt);
		atomic64_add(cryptlen, &alg->stats.cipher.encrypt_tlen);
1289 1290 1291 1292 1293 1294 1295 1296 1297
	}
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_skcipher_encrypt);

void crypto_stats_skcipher_decrypt(unsigned int cryptlen, int ret,
				   struct crypto_alg *alg)
{
	if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
1298
		atomic64_inc(&alg->stats.cipher.cipher_err_cnt);
1299
	} else {
1300 1301
		atomic64_inc(&alg->stats.cipher.decrypt_cnt);
		atomic64_add(cryptlen, &alg->stats.cipher.decrypt_tlen);
1302 1303 1304 1305 1306 1307
	}
	crypto_alg_put(alg);
}
EXPORT_SYMBOL_GPL(crypto_stats_skcipher_decrypt);
#endif

1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323
static int __init crypto_algapi_init(void)
{
	crypto_init_proc();
	return 0;
}

static void __exit crypto_algapi_exit(void)
{
	crypto_exit_proc();
}

module_init(crypto_algapi_init);
module_exit(crypto_algapi_exit);

MODULE_LICENSE("GPL");
MODULE_DESCRIPTION("Cryptographic algorithms API");