Revert "locking/lockdep: Add debug_locks check in __lock_downgrade()"
[linux-stable.git] / crypto / algapi.c
blob50eb828db767e8256082f7c4b42d0935b8b5606c
1 /*
2 * Cryptographic API for algorithms (i.e., low-level API).
4 * Copyright (c) 2006 Herbert Xu <herbert@gondor.apana.org.au>
6 * This program is free software; you can redistribute it and/or modify it
7 * under the terms of the GNU General Public License as published by the Free
8 * Software Foundation; either version 2 of the License, or (at your option)
9 * any later version.
13 #include <linux/err.h>
14 #include <linux/errno.h>
15 #include <linux/fips.h>
16 #include <linux/init.h>
17 #include <linux/kernel.h>
18 #include <linux/list.h>
19 #include <linux/module.h>
20 #include <linux/rtnetlink.h>
21 #include <linux/slab.h>
22 #include <linux/string.h>
24 #include "internal.h"
26 static LIST_HEAD(crypto_template_list);
28 static inline int crypto_set_driver_name(struct crypto_alg *alg)
30 static const char suffix[] = "-generic";
31 char *driver_name = alg->cra_driver_name;
32 int len;
34 if (*driver_name)
35 return 0;
37 len = strlcpy(driver_name, alg->cra_name, CRYPTO_MAX_ALG_NAME);
38 if (len + sizeof(suffix) > CRYPTO_MAX_ALG_NAME)
39 return -ENAMETOOLONG;
41 memcpy(driver_name + len, suffix, sizeof(suffix));
42 return 0;
45 static inline void crypto_check_module_sig(struct module *mod)
47 if (fips_enabled && mod && !module_sig_ok(mod))
48 panic("Module %s signature verification failed in FIPS mode\n",
49 module_name(mod));
52 static int crypto_check_alg(struct crypto_alg *alg)
54 crypto_check_module_sig(alg->cra_module);
56 if (alg->cra_alignmask & (alg->cra_alignmask + 1))
57 return -EINVAL;
59 if (alg->cra_blocksize > PAGE_SIZE / 8)
60 return -EINVAL;
62 if (alg->cra_priority < 0)
63 return -EINVAL;
65 atomic_set(&alg->cra_refcnt, 1);
67 return crypto_set_driver_name(alg);
70 static void crypto_free_instance(struct crypto_instance *inst)
72 if (!inst->alg.cra_type->free) {
73 inst->tmpl->free(inst);
74 return;
77 inst->alg.cra_type->free(inst);
80 static void crypto_destroy_instance(struct crypto_alg *alg)
82 struct crypto_instance *inst = (void *)alg;
83 struct crypto_template *tmpl = inst->tmpl;
85 crypto_free_instance(inst);
86 crypto_tmpl_put(tmpl);
89 static struct list_head *crypto_more_spawns(struct crypto_alg *alg,
90 struct list_head *stack,
91 struct list_head *top,
92 struct list_head *secondary_spawns)
94 struct crypto_spawn *spawn, *n;
96 spawn = list_first_entry_or_null(stack, struct crypto_spawn, list);
97 if (!spawn)
98 return NULL;
100 n = list_next_entry(spawn, list);
102 if (spawn->alg && &n->list != stack && !n->alg)
103 n->alg = (n->list.next == stack) ? alg :
104 &list_next_entry(n, list)->inst->alg;
106 list_move(&spawn->list, secondary_spawns);
108 return &n->list == stack ? top : &n->inst->alg.cra_users;
111 static void crypto_remove_instance(struct crypto_instance *inst,
112 struct list_head *list)
114 struct crypto_template *tmpl = inst->tmpl;
116 if (crypto_is_dead(&inst->alg))
117 return;
119 inst->alg.cra_flags |= CRYPTO_ALG_DEAD;
120 if (hlist_unhashed(&inst->list))
121 return;
123 if (!tmpl || !crypto_tmpl_get(tmpl))
124 return;
126 crypto_notify(CRYPTO_MSG_ALG_UNREGISTER, &inst->alg);
127 list_move(&inst->alg.cra_list, list);
128 hlist_del(&inst->list);
129 inst->alg.cra_destroy = crypto_destroy_instance;
131 BUG_ON(!list_empty(&inst->alg.cra_users));
134 void crypto_remove_spawns(struct crypto_alg *alg, struct list_head *list,
135 struct crypto_alg *nalg)
137 u32 new_type = (nalg ?: alg)->cra_flags;
138 struct crypto_spawn *spawn, *n;
139 LIST_HEAD(secondary_spawns);
140 struct list_head *spawns;
141 LIST_HEAD(stack);
142 LIST_HEAD(top);
144 spawns = &alg->cra_users;
145 list_for_each_entry_safe(spawn, n, spawns, list) {
146 if ((spawn->alg->cra_flags ^ new_type) & spawn->mask)
147 continue;
149 list_move(&spawn->list, &top);
152 spawns = &top;
153 do {
154 while (!list_empty(spawns)) {
155 struct crypto_instance *inst;
157 spawn = list_first_entry(spawns, struct crypto_spawn,
158 list);
159 inst = spawn->inst;
161 BUG_ON(&inst->alg == alg);
163 list_move(&spawn->list, &stack);
165 if (&inst->alg == nalg)
166 break;
168 spawn->alg = NULL;
169 spawns = &inst->alg.cra_users;
172 * We may encounter an unregistered instance here, since
173 * an instance's spawns are set up prior to the instance
174 * being registered. An unregistered instance will have
175 * NULL ->cra_users.next, since ->cra_users isn't
176 * properly initialized until registration. But an
177 * unregistered instance cannot have any users, so treat
178 * it the same as ->cra_users being empty.
180 if (spawns->next == NULL)
181 break;
183 } while ((spawns = crypto_more_spawns(alg, &stack, &top,
184 &secondary_spawns)));
186 list_for_each_entry_safe(spawn, n, &secondary_spawns, list) {
187 if (spawn->alg)
188 list_move(&spawn->list, &spawn->alg->cra_users);
189 else
190 crypto_remove_instance(spawn->inst, list);
193 EXPORT_SYMBOL_GPL(crypto_remove_spawns);
195 static struct crypto_larval *__crypto_register_alg(struct crypto_alg *alg)
197 struct crypto_alg *q;
198 struct crypto_larval *larval;
199 int ret = -EAGAIN;
201 if (crypto_is_dead(alg))
202 goto err;
204 INIT_LIST_HEAD(&alg->cra_users);
206 /* No cheating! */
207 alg->cra_flags &= ~CRYPTO_ALG_TESTED;
209 ret = -EEXIST;
211 list_for_each_entry(q, &crypto_alg_list, cra_list) {
212 if (q == alg)
213 goto err;
215 if (crypto_is_moribund(q))
216 continue;
218 if (crypto_is_larval(q)) {
219 if (!strcmp(alg->cra_driver_name, q->cra_driver_name))
220 goto err;
221 continue;
224 if (!strcmp(q->cra_driver_name, alg->cra_name) ||
225 !strcmp(q->cra_name, alg->cra_driver_name))
226 goto err;
229 larval = crypto_larval_alloc(alg->cra_name,
230 alg->cra_flags | CRYPTO_ALG_TESTED, 0);
231 if (IS_ERR(larval))
232 goto out;
234 ret = -ENOENT;
235 larval->adult = crypto_mod_get(alg);
236 if (!larval->adult)
237 goto free_larval;
239 atomic_set(&larval->alg.cra_refcnt, 1);
240 memcpy(larval->alg.cra_driver_name, alg->cra_driver_name,
241 CRYPTO_MAX_ALG_NAME);
242 larval->alg.cra_priority = alg->cra_priority;
244 list_add(&alg->cra_list, &crypto_alg_list);
245 list_add(&larval->alg.cra_list, &crypto_alg_list);
247 out:
248 return larval;
250 free_larval:
251 kfree(larval);
252 err:
253 larval = ERR_PTR(ret);
254 goto out;
257 void crypto_alg_tested(const char *name, int err)
259 struct crypto_larval *test;
260 struct crypto_alg *alg;
261 struct crypto_alg *q;
262 LIST_HEAD(list);
264 down_write(&crypto_alg_sem);
265 list_for_each_entry(q, &crypto_alg_list, cra_list) {
266 if (crypto_is_moribund(q) || !crypto_is_larval(q))
267 continue;
269 test = (struct crypto_larval *)q;
271 if (!strcmp(q->cra_driver_name, name))
272 goto found;
275 pr_err("alg: Unexpected test result for %s: %d\n", name, err);
276 goto unlock;
278 found:
279 q->cra_flags |= CRYPTO_ALG_DEAD;
280 alg = test->adult;
281 if (err || list_empty(&alg->cra_list))
282 goto complete;
284 alg->cra_flags |= CRYPTO_ALG_TESTED;
286 list_for_each_entry(q, &crypto_alg_list, cra_list) {
287 if (q == alg)
288 continue;
290 if (crypto_is_moribund(q))
291 continue;
293 if (crypto_is_larval(q)) {
294 struct crypto_larval *larval = (void *)q;
297 * Check to see if either our generic name or
298 * specific name can satisfy the name requested
299 * by the larval entry q.
301 if (strcmp(alg->cra_name, q->cra_name) &&
302 strcmp(alg->cra_driver_name, q->cra_name))
303 continue;
305 if (larval->adult)
306 continue;
307 if ((q->cra_flags ^ alg->cra_flags) & larval->mask)
308 continue;
309 if (!crypto_mod_get(alg))
310 continue;
312 larval->adult = alg;
313 continue;
316 if (strcmp(alg->cra_name, q->cra_name))
317 continue;
319 if (strcmp(alg->cra_driver_name, q->cra_driver_name) &&
320 q->cra_priority > alg->cra_priority)
321 continue;
323 crypto_remove_spawns(q, &list, alg);
326 complete:
327 complete_all(&test->completion);
329 unlock:
330 up_write(&crypto_alg_sem);
332 crypto_remove_final(&list);
334 EXPORT_SYMBOL_GPL(crypto_alg_tested);
336 void crypto_remove_final(struct list_head *list)
338 struct crypto_alg *alg;
339 struct crypto_alg *n;
341 list_for_each_entry_safe(alg, n, list, cra_list) {
342 list_del_init(&alg->cra_list);
343 crypto_alg_put(alg);
346 EXPORT_SYMBOL_GPL(crypto_remove_final);
348 static void crypto_wait_for_test(struct crypto_larval *larval)
350 int err;
352 err = crypto_probing_notify(CRYPTO_MSG_ALG_REGISTER, larval->adult);
353 if (err != NOTIFY_STOP) {
354 if (WARN_ON(err != NOTIFY_DONE))
355 goto out;
356 crypto_alg_tested(larval->alg.cra_driver_name, 0);
359 err = wait_for_completion_killable(&larval->completion);
360 WARN_ON(err);
362 out:
363 crypto_larval_kill(&larval->alg);
366 int crypto_register_alg(struct crypto_alg *alg)
368 struct crypto_larval *larval;
369 int err;
371 alg->cra_flags &= ~CRYPTO_ALG_DEAD;
372 err = crypto_check_alg(alg);
373 if (err)
374 return err;
376 down_write(&crypto_alg_sem);
377 larval = __crypto_register_alg(alg);
378 up_write(&crypto_alg_sem);
380 if (IS_ERR(larval))
381 return PTR_ERR(larval);
383 crypto_wait_for_test(larval);
384 return 0;
386 EXPORT_SYMBOL_GPL(crypto_register_alg);
388 static int crypto_remove_alg(struct crypto_alg *alg, struct list_head *list)
390 if (unlikely(list_empty(&alg->cra_list)))
391 return -ENOENT;
393 alg->cra_flags |= CRYPTO_ALG_DEAD;
395 crypto_notify(CRYPTO_MSG_ALG_UNREGISTER, alg);
396 list_del_init(&alg->cra_list);
397 crypto_remove_spawns(alg, list, NULL);
399 return 0;
402 int crypto_unregister_alg(struct crypto_alg *alg)
404 int ret;
405 LIST_HEAD(list);
407 down_write(&crypto_alg_sem);
408 ret = crypto_remove_alg(alg, &list);
409 up_write(&crypto_alg_sem);
411 if (ret)
412 return ret;
414 BUG_ON(atomic_read(&alg->cra_refcnt) != 1);
415 if (alg->cra_destroy)
416 alg->cra_destroy(alg);
418 crypto_remove_final(&list);
419 return 0;
421 EXPORT_SYMBOL_GPL(crypto_unregister_alg);
423 int crypto_register_algs(struct crypto_alg *algs, int count)
425 int i, ret;
427 for (i = 0; i < count; i++) {
428 ret = crypto_register_alg(&algs[i]);
429 if (ret)
430 goto err;
433 return 0;
435 err:
436 for (--i; i >= 0; --i)
437 crypto_unregister_alg(&algs[i]);
439 return ret;
441 EXPORT_SYMBOL_GPL(crypto_register_algs);
443 int crypto_unregister_algs(struct crypto_alg *algs, int count)
445 int i, ret;
447 for (i = 0; i < count; i++) {
448 ret = crypto_unregister_alg(&algs[i]);
449 if (ret)
450 pr_err("Failed to unregister %s %s: %d\n",
451 algs[i].cra_driver_name, algs[i].cra_name, ret);
454 return 0;
456 EXPORT_SYMBOL_GPL(crypto_unregister_algs);
458 int crypto_register_template(struct crypto_template *tmpl)
460 struct crypto_template *q;
461 int err = -EEXIST;
463 down_write(&crypto_alg_sem);
465 crypto_check_module_sig(tmpl->module);
467 list_for_each_entry(q, &crypto_template_list, list) {
468 if (q == tmpl)
469 goto out;
472 list_add(&tmpl->list, &crypto_template_list);
473 crypto_notify(CRYPTO_MSG_TMPL_REGISTER, tmpl);
474 err = 0;
475 out:
476 up_write(&crypto_alg_sem);
477 return err;
479 EXPORT_SYMBOL_GPL(crypto_register_template);
481 void crypto_unregister_template(struct crypto_template *tmpl)
483 struct crypto_instance *inst;
484 struct hlist_node *n;
485 struct hlist_head *list;
486 LIST_HEAD(users);
488 down_write(&crypto_alg_sem);
490 BUG_ON(list_empty(&tmpl->list));
491 list_del_init(&tmpl->list);
493 list = &tmpl->instances;
494 hlist_for_each_entry(inst, list, list) {
495 int err = crypto_remove_alg(&inst->alg, &users);
497 BUG_ON(err);
500 crypto_notify(CRYPTO_MSG_TMPL_UNREGISTER, tmpl);
502 up_write(&crypto_alg_sem);
504 hlist_for_each_entry_safe(inst, n, list, list) {
505 BUG_ON(atomic_read(&inst->alg.cra_refcnt) != 1);
506 crypto_free_instance(inst);
508 crypto_remove_final(&users);
510 EXPORT_SYMBOL_GPL(crypto_unregister_template);
512 static struct crypto_template *__crypto_lookup_template(const char *name)
514 struct crypto_template *q, *tmpl = NULL;
516 down_read(&crypto_alg_sem);
517 list_for_each_entry(q, &crypto_template_list, list) {
518 if (strcmp(q->name, name))
519 continue;
520 if (unlikely(!crypto_tmpl_get(q)))
521 continue;
523 tmpl = q;
524 break;
526 up_read(&crypto_alg_sem);
528 return tmpl;
531 struct crypto_template *crypto_lookup_template(const char *name)
533 return try_then_request_module(__crypto_lookup_template(name),
534 "crypto-%s", name);
536 EXPORT_SYMBOL_GPL(crypto_lookup_template);
538 int crypto_register_instance(struct crypto_template *tmpl,
539 struct crypto_instance *inst)
541 struct crypto_larval *larval;
542 int err;
544 err = crypto_check_alg(&inst->alg);
545 if (err)
546 return err;
548 inst->alg.cra_module = tmpl->module;
549 inst->alg.cra_flags |= CRYPTO_ALG_INSTANCE;
551 if (unlikely(!crypto_mod_get(&inst->alg)))
552 return -EAGAIN;
554 down_write(&crypto_alg_sem);
556 larval = __crypto_register_alg(&inst->alg);
557 if (IS_ERR(larval))
558 goto unlock;
560 hlist_add_head(&inst->list, &tmpl->instances);
561 inst->tmpl = tmpl;
563 unlock:
564 up_write(&crypto_alg_sem);
566 err = PTR_ERR(larval);
567 if (IS_ERR(larval))
568 goto err;
570 crypto_wait_for_test(larval);
572 /* Remove instance if test failed */
573 if (!(inst->alg.cra_flags & CRYPTO_ALG_TESTED))
574 crypto_unregister_instance(inst);
575 err = 0;
577 err:
578 crypto_mod_put(&inst->alg);
579 return err;
581 EXPORT_SYMBOL_GPL(crypto_register_instance);
583 int crypto_unregister_instance(struct crypto_instance *inst)
585 LIST_HEAD(list);
587 down_write(&crypto_alg_sem);
589 crypto_remove_spawns(&inst->alg, &list, NULL);
590 crypto_remove_instance(inst, &list);
592 up_write(&crypto_alg_sem);
594 crypto_remove_final(&list);
596 return 0;
598 EXPORT_SYMBOL_GPL(crypto_unregister_instance);
600 int crypto_init_spawn(struct crypto_spawn *spawn, struct crypto_alg *alg,
601 struct crypto_instance *inst, u32 mask)
603 int err = -EAGAIN;
605 spawn->inst = inst;
606 spawn->mask = mask;
608 down_write(&crypto_alg_sem);
609 if (!crypto_is_moribund(alg)) {
610 list_add(&spawn->list, &alg->cra_users);
611 spawn->alg = alg;
612 err = 0;
614 up_write(&crypto_alg_sem);
616 return err;
618 EXPORT_SYMBOL_GPL(crypto_init_spawn);
620 int crypto_init_spawn2(struct crypto_spawn *spawn, struct crypto_alg *alg,
621 struct crypto_instance *inst,
622 const struct crypto_type *frontend)
624 int err = -EINVAL;
626 if ((alg->cra_flags ^ frontend->type) & frontend->maskset)
627 goto out;
629 spawn->frontend = frontend;
630 err = crypto_init_spawn(spawn, alg, inst, frontend->maskset);
632 out:
633 return err;
635 EXPORT_SYMBOL_GPL(crypto_init_spawn2);
637 int crypto_grab_spawn(struct crypto_spawn *spawn, const char *name,
638 u32 type, u32 mask)
640 struct crypto_alg *alg;
641 int err;
643 alg = crypto_find_alg(name, spawn->frontend, type, mask);
644 if (IS_ERR(alg))
645 return PTR_ERR(alg);
647 err = crypto_init_spawn(spawn, alg, spawn->inst, mask);
648 crypto_mod_put(alg);
649 return err;
651 EXPORT_SYMBOL_GPL(crypto_grab_spawn);
653 void crypto_drop_spawn(struct crypto_spawn *spawn)
655 if (!spawn->alg)
656 return;
658 down_write(&crypto_alg_sem);
659 list_del(&spawn->list);
660 up_write(&crypto_alg_sem);
662 EXPORT_SYMBOL_GPL(crypto_drop_spawn);
664 static struct crypto_alg *crypto_spawn_alg(struct crypto_spawn *spawn)
666 struct crypto_alg *alg;
667 struct crypto_alg *alg2;
669 down_read(&crypto_alg_sem);
670 alg = spawn->alg;
671 alg2 = alg;
672 if (alg2)
673 alg2 = crypto_mod_get(alg2);
674 up_read(&crypto_alg_sem);
676 if (!alg2) {
677 if (alg)
678 crypto_shoot_alg(alg);
679 return ERR_PTR(-EAGAIN);
682 return alg;
685 struct crypto_tfm *crypto_spawn_tfm(struct crypto_spawn *spawn, u32 type,
686 u32 mask)
688 struct crypto_alg *alg;
689 struct crypto_tfm *tfm;
691 alg = crypto_spawn_alg(spawn);
692 if (IS_ERR(alg))
693 return ERR_CAST(alg);
695 tfm = ERR_PTR(-EINVAL);
696 if (unlikely((alg->cra_flags ^ type) & mask))
697 goto out_put_alg;
699 tfm = __crypto_alloc_tfm(alg, type, mask);
700 if (IS_ERR(tfm))
701 goto out_put_alg;
703 return tfm;
705 out_put_alg:
706 crypto_mod_put(alg);
707 return tfm;
709 EXPORT_SYMBOL_GPL(crypto_spawn_tfm);
711 void *crypto_spawn_tfm2(struct crypto_spawn *spawn)
713 struct crypto_alg *alg;
714 struct crypto_tfm *tfm;
716 alg = crypto_spawn_alg(spawn);
717 if (IS_ERR(alg))
718 return ERR_CAST(alg);
720 tfm = crypto_create_tfm(alg, spawn->frontend);
721 if (IS_ERR(tfm))
722 goto out_put_alg;
724 return tfm;
726 out_put_alg:
727 crypto_mod_put(alg);
728 return tfm;
730 EXPORT_SYMBOL_GPL(crypto_spawn_tfm2);
732 int crypto_register_notifier(struct notifier_block *nb)
734 return blocking_notifier_chain_register(&crypto_chain, nb);
736 EXPORT_SYMBOL_GPL(crypto_register_notifier);
738 int crypto_unregister_notifier(struct notifier_block *nb)
740 return blocking_notifier_chain_unregister(&crypto_chain, nb);
742 EXPORT_SYMBOL_GPL(crypto_unregister_notifier);
744 struct crypto_attr_type *crypto_get_attr_type(struct rtattr **tb)
746 struct rtattr *rta = tb[0];
747 struct crypto_attr_type *algt;
749 if (!rta)
750 return ERR_PTR(-ENOENT);
751 if (RTA_PAYLOAD(rta) < sizeof(*algt))
752 return ERR_PTR(-EINVAL);
753 if (rta->rta_type != CRYPTOA_TYPE)
754 return ERR_PTR(-EINVAL);
756 algt = RTA_DATA(rta);
758 return algt;
760 EXPORT_SYMBOL_GPL(crypto_get_attr_type);
762 int crypto_check_attr_type(struct rtattr **tb, u32 type)
764 struct crypto_attr_type *algt;
766 algt = crypto_get_attr_type(tb);
767 if (IS_ERR(algt))
768 return PTR_ERR(algt);
770 if ((algt->type ^ type) & algt->mask)
771 return -EINVAL;
773 return 0;
775 EXPORT_SYMBOL_GPL(crypto_check_attr_type);
777 const char *crypto_attr_alg_name(struct rtattr *rta)
779 struct crypto_attr_alg *alga;
781 if (!rta)
782 return ERR_PTR(-ENOENT);
783 if (RTA_PAYLOAD(rta) < sizeof(*alga))
784 return ERR_PTR(-EINVAL);
785 if (rta->rta_type != CRYPTOA_ALG)
786 return ERR_PTR(-EINVAL);
788 alga = RTA_DATA(rta);
789 alga->name[CRYPTO_MAX_ALG_NAME - 1] = 0;
791 return alga->name;
793 EXPORT_SYMBOL_GPL(crypto_attr_alg_name);
795 struct crypto_alg *crypto_attr_alg2(struct rtattr *rta,
796 const struct crypto_type *frontend,
797 u32 type, u32 mask)
799 const char *name;
801 name = crypto_attr_alg_name(rta);
802 if (IS_ERR(name))
803 return ERR_CAST(name);
805 return crypto_find_alg(name, frontend, type, mask);
807 EXPORT_SYMBOL_GPL(crypto_attr_alg2);
809 int crypto_attr_u32(struct rtattr *rta, u32 *num)
811 struct crypto_attr_u32 *nu32;
813 if (!rta)
814 return -ENOENT;
815 if (RTA_PAYLOAD(rta) < sizeof(*nu32))
816 return -EINVAL;
817 if (rta->rta_type != CRYPTOA_U32)
818 return -EINVAL;
820 nu32 = RTA_DATA(rta);
821 *num = nu32->num;
823 return 0;
825 EXPORT_SYMBOL_GPL(crypto_attr_u32);
827 int crypto_inst_setname(struct crypto_instance *inst, const char *name,
828 struct crypto_alg *alg)
830 if (snprintf(inst->alg.cra_name, CRYPTO_MAX_ALG_NAME, "%s(%s)", name,
831 alg->cra_name) >= CRYPTO_MAX_ALG_NAME)
832 return -ENAMETOOLONG;
834 if (snprintf(inst->alg.cra_driver_name, CRYPTO_MAX_ALG_NAME, "%s(%s)",
835 name, alg->cra_driver_name) >= CRYPTO_MAX_ALG_NAME)
836 return -ENAMETOOLONG;
838 return 0;
840 EXPORT_SYMBOL_GPL(crypto_inst_setname);
842 void *crypto_alloc_instance2(const char *name, struct crypto_alg *alg,
843 unsigned int head)
845 struct crypto_instance *inst;
846 char *p;
847 int err;
849 p = kzalloc(head + sizeof(*inst) + sizeof(struct crypto_spawn),
850 GFP_KERNEL);
851 if (!p)
852 return ERR_PTR(-ENOMEM);
854 inst = (void *)(p + head);
856 err = crypto_inst_setname(inst, name, alg);
857 if (err)
858 goto err_free_inst;
860 return p;
862 err_free_inst:
863 kfree(p);
864 return ERR_PTR(err);
866 EXPORT_SYMBOL_GPL(crypto_alloc_instance2);
868 struct crypto_instance *crypto_alloc_instance(const char *name,
869 struct crypto_alg *alg)
871 struct crypto_instance *inst;
872 struct crypto_spawn *spawn;
873 int err;
875 inst = crypto_alloc_instance2(name, alg, 0);
876 if (IS_ERR(inst))
877 goto out;
879 spawn = crypto_instance_ctx(inst);
880 err = crypto_init_spawn(spawn, alg, inst,
881 CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_ASYNC);
883 if (err)
884 goto err_free_inst;
886 return inst;
888 err_free_inst:
889 kfree(inst);
890 inst = ERR_PTR(err);
892 out:
893 return inst;
895 EXPORT_SYMBOL_GPL(crypto_alloc_instance);
897 void crypto_init_queue(struct crypto_queue *queue, unsigned int max_qlen)
899 INIT_LIST_HEAD(&queue->list);
900 queue->backlog = &queue->list;
901 queue->qlen = 0;
902 queue->max_qlen = max_qlen;
904 EXPORT_SYMBOL_GPL(crypto_init_queue);
906 int crypto_enqueue_request(struct crypto_queue *queue,
907 struct crypto_async_request *request)
909 int err = -EINPROGRESS;
911 if (unlikely(queue->qlen >= queue->max_qlen)) {
912 err = -EBUSY;
913 if (!(request->flags & CRYPTO_TFM_REQ_MAY_BACKLOG))
914 goto out;
915 if (queue->backlog == &queue->list)
916 queue->backlog = &request->list;
919 queue->qlen++;
920 list_add_tail(&request->list, &queue->list);
922 out:
923 return err;
925 EXPORT_SYMBOL_GPL(crypto_enqueue_request);
927 struct crypto_async_request *crypto_dequeue_request(struct crypto_queue *queue)
929 struct list_head *request;
931 if (unlikely(!queue->qlen))
932 return NULL;
934 queue->qlen--;
936 if (queue->backlog != &queue->list)
937 queue->backlog = queue->backlog->next;
939 request = queue->list.next;
940 list_del(request);
942 return list_entry(request, struct crypto_async_request, list);
944 EXPORT_SYMBOL_GPL(crypto_dequeue_request);
946 int crypto_tfm_in_queue(struct crypto_queue *queue, struct crypto_tfm *tfm)
948 struct crypto_async_request *req;
950 list_for_each_entry(req, &queue->list, list) {
951 if (req->tfm == tfm)
952 return 1;
955 return 0;
957 EXPORT_SYMBOL_GPL(crypto_tfm_in_queue);
959 static inline void crypto_inc_byte(u8 *a, unsigned int size)
961 u8 *b = (a + size);
962 u8 c;
964 for (; size; size--) {
965 c = *--b + 1;
966 *b = c;
967 if (c)
968 break;
972 void crypto_inc(u8 *a, unsigned int size)
974 __be32 *b = (__be32 *)(a + size);
975 u32 c;
977 if (IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) ||
978 IS_ALIGNED((unsigned long)b, __alignof__(*b)))
979 for (; size >= 4; size -= 4) {
980 c = be32_to_cpu(*--b) + 1;
981 *b = cpu_to_be32(c);
982 if (likely(c))
983 return;
986 crypto_inc_byte(a, size);
988 EXPORT_SYMBOL_GPL(crypto_inc);
990 void __crypto_xor(u8 *dst, const u8 *src1, const u8 *src2, unsigned int len)
992 int relalign = 0;
994 if (!IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS)) {
995 int size = sizeof(unsigned long);
996 int d = (((unsigned long)dst ^ (unsigned long)src1) |
997 ((unsigned long)dst ^ (unsigned long)src2)) &
998 (size - 1);
1000 relalign = d ? 1 << __ffs(d) : size;
1003 * If we care about alignment, process as many bytes as
1004 * needed to advance dst and src to values whose alignments
1005 * equal their relative alignment. This will allow us to
1006 * process the remainder of the input using optimal strides.
1008 while (((unsigned long)dst & (relalign - 1)) && len > 0) {
1009 *dst++ = *src1++ ^ *src2++;
1010 len--;
1014 while (IS_ENABLED(CONFIG_64BIT) && len >= 8 && !(relalign & 7)) {
1015 *(u64 *)dst = *(u64 *)src1 ^ *(u64 *)src2;
1016 dst += 8;
1017 src1 += 8;
1018 src2 += 8;
1019 len -= 8;
1022 while (len >= 4 && !(relalign & 3)) {
1023 *(u32 *)dst = *(u32 *)src1 ^ *(u32 *)src2;
1024 dst += 4;
1025 src1 += 4;
1026 src2 += 4;
1027 len -= 4;
1030 while (len >= 2 && !(relalign & 1)) {
1031 *(u16 *)dst = *(u16 *)src1 ^ *(u16 *)src2;
1032 dst += 2;
1033 src1 += 2;
1034 src2 += 2;
1035 len -= 2;
1038 while (len--)
1039 *dst++ = *src1++ ^ *src2++;
1041 EXPORT_SYMBOL_GPL(__crypto_xor);
1043 unsigned int crypto_alg_extsize(struct crypto_alg *alg)
1045 return alg->cra_ctxsize +
1046 (alg->cra_alignmask & ~(crypto_tfm_ctx_alignment() - 1));
1048 EXPORT_SYMBOL_GPL(crypto_alg_extsize);
1050 int crypto_type_has_alg(const char *name, const struct crypto_type *frontend,
1051 u32 type, u32 mask)
1053 int ret = 0;
1054 struct crypto_alg *alg = crypto_find_alg(name, frontend, type, mask);
1056 if (!IS_ERR(alg)) {
1057 crypto_mod_put(alg);
1058 ret = 1;
1061 return ret;
1063 EXPORT_SYMBOL_GPL(crypto_type_has_alg);
1065 static int __init crypto_algapi_init(void)
1067 crypto_init_proc();
1068 return 0;
1071 static void __exit crypto_algapi_exit(void)
1073 crypto_exit_proc();
1076 module_init(crypto_algapi_init);
1077 module_exit(crypto_algapi_exit);
1079 MODULE_LICENSE("GPL");
1080 MODULE_DESCRIPTION("Cryptographic algorithms API");