2 * Algorithm testing framework and tests.
4 * Copyright (c) 2002 James Morris <jmorris@intercode.com.au>
5 * Copyright (c) 2002 Jean-Francois Dive <jef@linuxbe.org>
6 * Copyright (c) 2007 Nokia Siemens Networks
7 * Copyright (c) 2008 Herbert Xu <herbert@gondor.apana.org.au>
9 * Updated RFC4106 AES-GCM testing.
10 * Authors: Aidan O'Mahony (aidan.o.mahony@intel.com)
11 * Adrian Hoban <adrian.hoban@intel.com>
12 * Gabriele Paoloni <gabriele.paoloni@intel.com>
13 * Tadeusz Struk (tadeusz.struk@intel.com)
14 * Copyright (c) 2010, Intel Corporation.
16 * This program is free software; you can redistribute it and/or modify it
17 * under the terms of the GNU General Public License as published by the Free
18 * Software Foundation; either version 2 of the License, or (at your option)
23 #include <crypto/hash.h>
24 #include <linux/err.h>
25 #include <linux/module.h>
26 #include <linux/scatterlist.h>
27 #include <linux/slab.h>
28 #include <linux/string.h>
29 #include <crypto/rng.h>
33 #ifdef CONFIG_CRYPTO_MANAGER_DISABLE_TESTS
36 int alg_test(const char *driver
, const char *alg
, u32 type
, u32 mask
)
46 * Need slab memory for testing (size in number of pages).
51 * Indexes into the xbuf to simulate cross-page access.
63 * Used by test_cipher()
68 struct tcrypt_result
{
69 struct completion completion
;
73 struct aead_test_suite
{
75 struct aead_testvec
*vecs
;
80 struct cipher_test_suite
{
82 struct cipher_testvec
*vecs
;
87 struct comp_test_suite
{
89 struct comp_testvec
*vecs
;
94 struct pcomp_test_suite
{
96 struct pcomp_testvec
*vecs
;
101 struct hash_test_suite
{
102 struct hash_testvec
*vecs
;
106 struct cprng_test_suite
{
107 struct cprng_testvec
*vecs
;
111 struct alg_test_desc
{
113 int (*test
)(const struct alg_test_desc
*desc
, const char *driver
,
115 int fips_allowed
; /* set if alg is allowed in fips mode */
118 struct aead_test_suite aead
;
119 struct cipher_test_suite cipher
;
120 struct comp_test_suite comp
;
121 struct pcomp_test_suite pcomp
;
122 struct hash_test_suite hash
;
123 struct cprng_test_suite cprng
;
127 static unsigned int IDX
[8] = { IDX1
, IDX2
, IDX3
, IDX4
, IDX5
, IDX6
, IDX7
, IDX8
};
129 static void hexdump(unsigned char *buf
, unsigned int len
)
131 print_hex_dump(KERN_CONT
, "", DUMP_PREFIX_OFFSET
,
136 static void tcrypt_complete(struct crypto_async_request
*req
, int err
)
138 struct tcrypt_result
*res
= req
->data
;
140 if (err
== -EINPROGRESS
)
144 complete(&res
->completion
);
147 static int testmgr_alloc_buf(char *buf
[XBUFSIZE
])
151 for (i
= 0; i
< XBUFSIZE
; i
++) {
152 buf
[i
] = (void *)__get_free_page(GFP_KERNEL
);
161 free_page((unsigned long)buf
[i
]);
166 static void testmgr_free_buf(char *buf
[XBUFSIZE
])
170 for (i
= 0; i
< XBUFSIZE
; i
++)
171 free_page((unsigned long)buf
[i
]);
174 static int do_one_async_hash_op(struct ahash_request
*req
,
175 struct tcrypt_result
*tr
,
178 if (ret
== -EINPROGRESS
|| ret
== -EBUSY
) {
179 ret
= wait_for_completion_interruptible(&tr
->completion
);
182 INIT_COMPLETION(tr
->completion
);
187 static int test_hash(struct crypto_ahash
*tfm
, struct hash_testvec
*template,
188 unsigned int tcount
, bool use_digest
)
190 const char *algo
= crypto_tfm_alg_driver_name(crypto_ahash_tfm(tfm
));
191 unsigned int i
, j
, k
, temp
;
192 struct scatterlist sg
[8];
194 struct ahash_request
*req
;
195 struct tcrypt_result tresult
;
197 char *xbuf
[XBUFSIZE
];
200 if (testmgr_alloc_buf(xbuf
))
203 init_completion(&tresult
.completion
);
205 req
= ahash_request_alloc(tfm
, GFP_KERNEL
);
207 printk(KERN_ERR
"alg: hash: Failed to allocate request for "
211 ahash_request_set_callback(req
, CRYPTO_TFM_REQ_MAY_BACKLOG
,
212 tcrypt_complete
, &tresult
);
215 for (i
= 0; i
< tcount
; i
++) {
220 memset(result
, 0, 64);
224 memcpy(hash_buff
, template[i
].plaintext
, template[i
].psize
);
225 sg_init_one(&sg
[0], hash_buff
, template[i
].psize
);
227 if (template[i
].ksize
) {
228 crypto_ahash_clear_flags(tfm
, ~0);
229 ret
= crypto_ahash_setkey(tfm
, template[i
].key
,
232 printk(KERN_ERR
"alg: hash: setkey failed on "
233 "test %d for %s: ret=%d\n", j
, algo
,
239 ahash_request_set_crypt(req
, sg
, result
, template[i
].psize
);
241 ret
= do_one_async_hash_op(req
, &tresult
,
242 crypto_ahash_digest(req
));
244 pr_err("alg: hash: digest failed on test %d "
245 "for %s: ret=%d\n", j
, algo
, -ret
);
249 ret
= do_one_async_hash_op(req
, &tresult
,
250 crypto_ahash_init(req
));
252 pr_err("alt: hash: init failed on test %d "
253 "for %s: ret=%d\n", j
, algo
, -ret
);
256 ret
= do_one_async_hash_op(req
, &tresult
,
257 crypto_ahash_update(req
));
259 pr_err("alt: hash: update failed on test %d "
260 "for %s: ret=%d\n", j
, algo
, -ret
);
263 ret
= do_one_async_hash_op(req
, &tresult
,
264 crypto_ahash_final(req
));
266 pr_err("alt: hash: final failed on test %d "
267 "for %s: ret=%d\n", j
, algo
, -ret
);
272 if (memcmp(result
, template[i
].digest
,
273 crypto_ahash_digestsize(tfm
))) {
274 printk(KERN_ERR
"alg: hash: Test %d failed for %s\n",
276 hexdump(result
, crypto_ahash_digestsize(tfm
));
283 for (i
= 0; i
< tcount
; i
++) {
284 if (template[i
].np
) {
286 memset(result
, 0, 64);
289 sg_init_table(sg
, template[i
].np
);
291 for (k
= 0; k
< template[i
].np
; k
++) {
292 if (WARN_ON(offset_in_page(IDX
[k
]) +
293 template[i
].tap
[k
] > PAGE_SIZE
))
296 memcpy(xbuf
[IDX
[k
] >> PAGE_SHIFT
] +
297 offset_in_page(IDX
[k
]),
298 template[i
].plaintext
+ temp
,
301 temp
+= template[i
].tap
[k
];
304 if (template[i
].ksize
) {
305 crypto_ahash_clear_flags(tfm
, ~0);
306 ret
= crypto_ahash_setkey(tfm
, template[i
].key
,
310 printk(KERN_ERR
"alg: hash: setkey "
311 "failed on chunking test %d "
312 "for %s: ret=%d\n", j
, algo
,
318 ahash_request_set_crypt(req
, sg
, result
,
320 ret
= crypto_ahash_digest(req
);
326 ret
= wait_for_completion_interruptible(
327 &tresult
.completion
);
328 if (!ret
&& !(ret
= tresult
.err
)) {
329 INIT_COMPLETION(tresult
.completion
);
334 printk(KERN_ERR
"alg: hash: digest failed "
335 "on chunking test %d for %s: "
336 "ret=%d\n", j
, algo
, -ret
);
340 if (memcmp(result
, template[i
].digest
,
341 crypto_ahash_digestsize(tfm
))) {
342 printk(KERN_ERR
"alg: hash: Chunking test %d "
343 "failed for %s\n", j
, algo
);
344 hexdump(result
, crypto_ahash_digestsize(tfm
));
354 ahash_request_free(req
);
356 testmgr_free_buf(xbuf
);
361 static int test_aead(struct crypto_aead
*tfm
, int enc
,
362 struct aead_testvec
*template, unsigned int tcount
)
364 const char *algo
= crypto_tfm_alg_driver_name(crypto_aead_tfm(tfm
));
365 unsigned int i
, j
, k
, n
, temp
;
369 struct aead_request
*req
;
370 struct scatterlist sg
[8];
371 struct scatterlist asg
[8];
373 struct tcrypt_result result
;
374 unsigned int authsize
;
378 char *xbuf
[XBUFSIZE
];
379 char *axbuf
[XBUFSIZE
];
381 if (testmgr_alloc_buf(xbuf
))
383 if (testmgr_alloc_buf(axbuf
))
391 init_completion(&result
.completion
);
393 req
= aead_request_alloc(tfm
, GFP_KERNEL
);
395 printk(KERN_ERR
"alg: aead: Failed to allocate request for "
400 aead_request_set_callback(req
, CRYPTO_TFM_REQ_MAY_BACKLOG
,
401 tcrypt_complete
, &result
);
403 for (i
= 0, j
= 0; i
< tcount
; i
++) {
404 if (!template[i
].np
) {
407 /* some tepmplates have no input data but they will
414 if (WARN_ON(template[i
].ilen
> PAGE_SIZE
||
415 template[i
].alen
> PAGE_SIZE
))
418 memcpy(input
, template[i
].input
, template[i
].ilen
);
419 memcpy(assoc
, template[i
].assoc
, template[i
].alen
);
421 memcpy(iv
, template[i
].iv
, MAX_IVLEN
);
423 memset(iv
, 0, MAX_IVLEN
);
425 crypto_aead_clear_flags(tfm
, ~0);
427 crypto_aead_set_flags(
428 tfm
, CRYPTO_TFM_REQ_WEAK_KEY
);
430 key
= template[i
].key
;
432 ret
= crypto_aead_setkey(tfm
, key
,
434 if (!ret
== template[i
].fail
) {
435 printk(KERN_ERR
"alg: aead: setkey failed on "
436 "test %d for %s: flags=%x\n", j
, algo
,
437 crypto_aead_get_flags(tfm
));
442 authsize
= abs(template[i
].rlen
- template[i
].ilen
);
443 ret
= crypto_aead_setauthsize(tfm
, authsize
);
445 printk(KERN_ERR
"alg: aead: Failed to set "
446 "authsize to %u on test %d for %s\n",
451 sg_init_one(&sg
[0], input
,
452 template[i
].ilen
+ (enc
? authsize
: 0));
454 sg_init_one(&asg
[0], assoc
, template[i
].alen
);
456 aead_request_set_crypt(req
, sg
, sg
,
457 template[i
].ilen
, iv
);
459 aead_request_set_assoc(req
, asg
, template[i
].alen
);
462 crypto_aead_encrypt(req
) :
463 crypto_aead_decrypt(req
);
467 if (template[i
].novrfy
) {
468 /* verification was supposed to fail */
469 printk(KERN_ERR
"alg: aead: %s failed "
470 "on test %d for %s: ret was 0, "
471 "expected -EBADMSG\n",
473 /* so really, we got a bad message */
480 ret
= wait_for_completion_interruptible(
482 if (!ret
&& !(ret
= result
.err
)) {
483 INIT_COMPLETION(result
.completion
);
487 if (template[i
].novrfy
)
488 /* verification failure was expected */
492 printk(KERN_ERR
"alg: aead: %s failed on test "
493 "%d for %s: ret=%d\n", e
, j
, algo
, -ret
);
498 if (memcmp(q
, template[i
].result
, template[i
].rlen
)) {
499 printk(KERN_ERR
"alg: aead: Test %d failed on "
500 "%s for %s\n", j
, e
, algo
);
501 hexdump(q
, template[i
].rlen
);
508 for (i
= 0, j
= 0; i
< tcount
; i
++) {
509 if (template[i
].np
) {
513 memcpy(iv
, template[i
].iv
, MAX_IVLEN
);
515 memset(iv
, 0, MAX_IVLEN
);
517 crypto_aead_clear_flags(tfm
, ~0);
519 crypto_aead_set_flags(
520 tfm
, CRYPTO_TFM_REQ_WEAK_KEY
);
521 key
= template[i
].key
;
523 ret
= crypto_aead_setkey(tfm
, key
, template[i
].klen
);
524 if (!ret
== template[i
].fail
) {
525 printk(KERN_ERR
"alg: aead: setkey failed on "
526 "chunk test %d for %s: flags=%x\n", j
,
527 algo
, crypto_aead_get_flags(tfm
));
532 authsize
= abs(template[i
].rlen
- template[i
].ilen
);
535 sg_init_table(sg
, template[i
].np
);
536 for (k
= 0, temp
= 0; k
< template[i
].np
; k
++) {
537 if (WARN_ON(offset_in_page(IDX
[k
]) +
538 template[i
].tap
[k
] > PAGE_SIZE
))
541 q
= xbuf
[IDX
[k
] >> PAGE_SHIFT
] +
542 offset_in_page(IDX
[k
]);
544 memcpy(q
, template[i
].input
+ temp
,
547 n
= template[i
].tap
[k
];
548 if (k
== template[i
].np
- 1 && enc
)
550 if (offset_in_page(q
) + n
< PAGE_SIZE
)
553 sg_set_buf(&sg
[k
], q
, template[i
].tap
[k
]);
554 temp
+= template[i
].tap
[k
];
557 ret
= crypto_aead_setauthsize(tfm
, authsize
);
559 printk(KERN_ERR
"alg: aead: Failed to set "
560 "authsize to %u on chunk test %d for "
561 "%s\n", authsize
, j
, algo
);
566 if (WARN_ON(sg
[k
- 1].offset
+
567 sg
[k
- 1].length
+ authsize
>
573 sg
[k
- 1].length
+= authsize
;
576 sg_init_table(asg
, template[i
].anp
);
578 for (k
= 0, temp
= 0; k
< template[i
].anp
; k
++) {
579 if (WARN_ON(offset_in_page(IDX
[k
]) +
580 template[i
].atap
[k
] > PAGE_SIZE
))
583 memcpy(axbuf
[IDX
[k
] >> PAGE_SHIFT
] +
584 offset_in_page(IDX
[k
]),
585 template[i
].assoc
+ temp
,
586 template[i
].atap
[k
]),
587 template[i
].atap
[k
]);
588 temp
+= template[i
].atap
[k
];
591 aead_request_set_crypt(req
, sg
, sg
,
595 aead_request_set_assoc(req
, asg
, template[i
].alen
);
598 crypto_aead_encrypt(req
) :
599 crypto_aead_decrypt(req
);
603 if (template[i
].novrfy
) {
604 /* verification was supposed to fail */
605 printk(KERN_ERR
"alg: aead: %s failed "
606 "on chunk test %d for %s: ret "
607 "was 0, expected -EBADMSG\n",
609 /* so really, we got a bad message */
616 ret
= wait_for_completion_interruptible(
618 if (!ret
&& !(ret
= result
.err
)) {
619 INIT_COMPLETION(result
.completion
);
623 if (template[i
].novrfy
)
624 /* verification failure was expected */
628 printk(KERN_ERR
"alg: aead: %s failed on "
629 "chunk test %d for %s: ret=%d\n", e
, j
,
635 for (k
= 0, temp
= 0; k
< template[i
].np
; k
++) {
636 q
= xbuf
[IDX
[k
] >> PAGE_SHIFT
] +
637 offset_in_page(IDX
[k
]);
639 n
= template[i
].tap
[k
];
640 if (k
== template[i
].np
- 1)
641 n
+= enc
? authsize
: -authsize
;
643 if (memcmp(q
, template[i
].result
+ temp
, n
)) {
644 printk(KERN_ERR
"alg: aead: Chunk "
645 "test %d failed on %s at page "
646 "%u for %s\n", j
, e
, k
, algo
);
652 if (k
== template[i
].np
- 1 && !enc
) {
653 if (memcmp(q
, template[i
].input
+
659 for (n
= 0; offset_in_page(q
+ n
) &&
664 printk(KERN_ERR
"alg: aead: Result "
665 "buffer corruption in chunk "
666 "test %d on %s at page %u for "
667 "%s: %u bytes:\n", j
, e
, k
,
673 temp
+= template[i
].tap
[k
];
681 aead_request_free(req
);
682 testmgr_free_buf(axbuf
);
684 testmgr_free_buf(xbuf
);
689 static int test_cipher(struct crypto_cipher
*tfm
, int enc
,
690 struct cipher_testvec
*template, unsigned int tcount
)
692 const char *algo
= crypto_tfm_alg_driver_name(crypto_cipher_tfm(tfm
));
693 unsigned int i
, j
, k
;
697 char *xbuf
[XBUFSIZE
];
700 if (testmgr_alloc_buf(xbuf
))
709 for (i
= 0; i
< tcount
; i
++) {
716 if (WARN_ON(template[i
].ilen
> PAGE_SIZE
))
720 memcpy(data
, template[i
].input
, template[i
].ilen
);
722 crypto_cipher_clear_flags(tfm
, ~0);
724 crypto_cipher_set_flags(tfm
, CRYPTO_TFM_REQ_WEAK_KEY
);
726 ret
= crypto_cipher_setkey(tfm
, template[i
].key
,
728 if (!ret
== template[i
].fail
) {
729 printk(KERN_ERR
"alg: cipher: setkey failed "
730 "on test %d for %s: flags=%x\n", j
,
731 algo
, crypto_cipher_get_flags(tfm
));
736 for (k
= 0; k
< template[i
].ilen
;
737 k
+= crypto_cipher_blocksize(tfm
)) {
739 crypto_cipher_encrypt_one(tfm
, data
+ k
,
742 crypto_cipher_decrypt_one(tfm
, data
+ k
,
747 if (memcmp(q
, template[i
].result
, template[i
].rlen
)) {
748 printk(KERN_ERR
"alg: cipher: Test %d failed "
749 "on %s for %s\n", j
, e
, algo
);
750 hexdump(q
, template[i
].rlen
);
759 testmgr_free_buf(xbuf
);
764 static int test_skcipher(struct crypto_ablkcipher
*tfm
, int enc
,
765 struct cipher_testvec
*template, unsigned int tcount
)
768 crypto_tfm_alg_driver_name(crypto_ablkcipher_tfm(tfm
));
769 unsigned int i
, j
, k
, n
, temp
;
771 struct ablkcipher_request
*req
;
772 struct scatterlist sg
[8];
774 struct tcrypt_result result
;
777 char *xbuf
[XBUFSIZE
];
780 if (testmgr_alloc_buf(xbuf
))
788 init_completion(&result
.completion
);
790 req
= ablkcipher_request_alloc(tfm
, GFP_KERNEL
);
792 printk(KERN_ERR
"alg: skcipher: Failed to allocate request "
797 ablkcipher_request_set_callback(req
, CRYPTO_TFM_REQ_MAY_BACKLOG
,
798 tcrypt_complete
, &result
);
801 for (i
= 0; i
< tcount
; i
++) {
803 memcpy(iv
, template[i
].iv
, MAX_IVLEN
);
805 memset(iv
, 0, MAX_IVLEN
);
807 if (!(template[i
].np
)) {
811 if (WARN_ON(template[i
].ilen
> PAGE_SIZE
))
815 memcpy(data
, template[i
].input
, template[i
].ilen
);
817 crypto_ablkcipher_clear_flags(tfm
, ~0);
819 crypto_ablkcipher_set_flags(
820 tfm
, CRYPTO_TFM_REQ_WEAK_KEY
);
822 ret
= crypto_ablkcipher_setkey(tfm
, template[i
].key
,
824 if (!ret
== template[i
].fail
) {
825 printk(KERN_ERR
"alg: skcipher: setkey failed "
826 "on test %d for %s: flags=%x\n", j
,
827 algo
, crypto_ablkcipher_get_flags(tfm
));
832 sg_init_one(&sg
[0], data
, template[i
].ilen
);
834 ablkcipher_request_set_crypt(req
, sg
, sg
,
835 template[i
].ilen
, iv
);
837 crypto_ablkcipher_encrypt(req
) :
838 crypto_ablkcipher_decrypt(req
);
845 ret
= wait_for_completion_interruptible(
847 if (!ret
&& !((ret
= result
.err
))) {
848 INIT_COMPLETION(result
.completion
);
853 printk(KERN_ERR
"alg: skcipher: %s failed on "
854 "test %d for %s: ret=%d\n", e
, j
, algo
,
860 if (memcmp(q
, template[i
].result
, template[i
].rlen
)) {
861 printk(KERN_ERR
"alg: skcipher: Test %d "
862 "failed on %s for %s\n", j
, e
, algo
);
863 hexdump(q
, template[i
].rlen
);
871 for (i
= 0; i
< tcount
; i
++) {
874 memcpy(iv
, template[i
].iv
, MAX_IVLEN
);
876 memset(iv
, 0, MAX_IVLEN
);
878 if (template[i
].np
) {
881 crypto_ablkcipher_clear_flags(tfm
, ~0);
883 crypto_ablkcipher_set_flags(
884 tfm
, CRYPTO_TFM_REQ_WEAK_KEY
);
886 ret
= crypto_ablkcipher_setkey(tfm
, template[i
].key
,
888 if (!ret
== template[i
].fail
) {
889 printk(KERN_ERR
"alg: skcipher: setkey failed "
890 "on chunk test %d for %s: flags=%x\n",
892 crypto_ablkcipher_get_flags(tfm
));
899 sg_init_table(sg
, template[i
].np
);
900 for (k
= 0; k
< template[i
].np
; k
++) {
901 if (WARN_ON(offset_in_page(IDX
[k
]) +
902 template[i
].tap
[k
] > PAGE_SIZE
))
905 q
= xbuf
[IDX
[k
] >> PAGE_SHIFT
] +
906 offset_in_page(IDX
[k
]);
908 memcpy(q
, template[i
].input
+ temp
,
911 if (offset_in_page(q
) + template[i
].tap
[k
] <
913 q
[template[i
].tap
[k
]] = 0;
915 sg_set_buf(&sg
[k
], q
, template[i
].tap
[k
]);
917 temp
+= template[i
].tap
[k
];
920 ablkcipher_request_set_crypt(req
, sg
, sg
,
921 template[i
].ilen
, iv
);
924 crypto_ablkcipher_encrypt(req
) :
925 crypto_ablkcipher_decrypt(req
);
932 ret
= wait_for_completion_interruptible(
934 if (!ret
&& !((ret
= result
.err
))) {
935 INIT_COMPLETION(result
.completion
);
940 printk(KERN_ERR
"alg: skcipher: %s failed on "
941 "chunk test %d for %s: ret=%d\n", e
, j
,
948 for (k
= 0; k
< template[i
].np
; k
++) {
949 q
= xbuf
[IDX
[k
] >> PAGE_SHIFT
] +
950 offset_in_page(IDX
[k
]);
952 if (memcmp(q
, template[i
].result
+ temp
,
953 template[i
].tap
[k
])) {
954 printk(KERN_ERR
"alg: skcipher: Chunk "
955 "test %d failed on %s at page "
956 "%u for %s\n", j
, e
, k
, algo
);
957 hexdump(q
, template[i
].tap
[k
]);
961 q
+= template[i
].tap
[k
];
962 for (n
= 0; offset_in_page(q
+ n
) && q
[n
]; n
++)
965 printk(KERN_ERR
"alg: skcipher: "
966 "Result buffer corruption in "
967 "chunk test %d on %s at page "
968 "%u for %s: %u bytes:\n", j
, e
,
973 temp
+= template[i
].tap
[k
];
981 ablkcipher_request_free(req
);
982 testmgr_free_buf(xbuf
);
987 static int test_comp(struct crypto_comp
*tfm
, struct comp_testvec
*ctemplate
,
988 struct comp_testvec
*dtemplate
, int ctcount
, int dtcount
)
990 const char *algo
= crypto_tfm_alg_driver_name(crypto_comp_tfm(tfm
));
992 char result
[COMP_BUF_SIZE
];
995 for (i
= 0; i
< ctcount
; i
++) {
997 unsigned int dlen
= COMP_BUF_SIZE
;
999 memset(result
, 0, sizeof (result
));
1001 ilen
= ctemplate
[i
].inlen
;
1002 ret
= crypto_comp_compress(tfm
, ctemplate
[i
].input
,
1003 ilen
, result
, &dlen
);
1005 printk(KERN_ERR
"alg: comp: compression failed "
1006 "on test %d for %s: ret=%d\n", i
+ 1, algo
,
1011 if (dlen
!= ctemplate
[i
].outlen
) {
1012 printk(KERN_ERR
"alg: comp: Compression test %d "
1013 "failed for %s: output len = %d\n", i
+ 1, algo
,
1019 if (memcmp(result
, ctemplate
[i
].output
, dlen
)) {
1020 printk(KERN_ERR
"alg: comp: Compression test %d "
1021 "failed for %s\n", i
+ 1, algo
);
1022 hexdump(result
, dlen
);
1028 for (i
= 0; i
< dtcount
; i
++) {
1030 unsigned int dlen
= COMP_BUF_SIZE
;
1032 memset(result
, 0, sizeof (result
));
1034 ilen
= dtemplate
[i
].inlen
;
1035 ret
= crypto_comp_decompress(tfm
, dtemplate
[i
].input
,
1036 ilen
, result
, &dlen
);
1038 printk(KERN_ERR
"alg: comp: decompression failed "
1039 "on test %d for %s: ret=%d\n", i
+ 1, algo
,
1044 if (dlen
!= dtemplate
[i
].outlen
) {
1045 printk(KERN_ERR
"alg: comp: Decompression test %d "
1046 "failed for %s: output len = %d\n", i
+ 1, algo
,
1052 if (memcmp(result
, dtemplate
[i
].output
, dlen
)) {
1053 printk(KERN_ERR
"alg: comp: Decompression test %d "
1054 "failed for %s\n", i
+ 1, algo
);
1055 hexdump(result
, dlen
);
1067 static int test_pcomp(struct crypto_pcomp
*tfm
,
1068 struct pcomp_testvec
*ctemplate
,
1069 struct pcomp_testvec
*dtemplate
, int ctcount
,
1072 const char *algo
= crypto_tfm_alg_driver_name(crypto_pcomp_tfm(tfm
));
1074 char result
[COMP_BUF_SIZE
];
1077 for (i
= 0; i
< ctcount
; i
++) {
1078 struct comp_request req
;
1079 unsigned int produced
= 0;
1081 res
= crypto_compress_setup(tfm
, ctemplate
[i
].params
,
1082 ctemplate
[i
].paramsize
);
1084 pr_err("alg: pcomp: compression setup failed on test "
1085 "%d for %s: error=%d\n", i
+ 1, algo
, res
);
1089 res
= crypto_compress_init(tfm
);
1091 pr_err("alg: pcomp: compression init failed on test "
1092 "%d for %s: error=%d\n", i
+ 1, algo
, res
);
1096 memset(result
, 0, sizeof(result
));
1098 req
.next_in
= ctemplate
[i
].input
;
1099 req
.avail_in
= ctemplate
[i
].inlen
/ 2;
1100 req
.next_out
= result
;
1101 req
.avail_out
= ctemplate
[i
].outlen
/ 2;
1103 res
= crypto_compress_update(tfm
, &req
);
1104 if (res
< 0 && (res
!= -EAGAIN
|| req
.avail_in
)) {
1105 pr_err("alg: pcomp: compression update failed on test "
1106 "%d for %s: error=%d\n", i
+ 1, algo
, res
);
1112 /* Add remaining input data */
1113 req
.avail_in
+= (ctemplate
[i
].inlen
+ 1) / 2;
1115 res
= crypto_compress_update(tfm
, &req
);
1116 if (res
< 0 && (res
!= -EAGAIN
|| req
.avail_in
)) {
1117 pr_err("alg: pcomp: compression update failed on test "
1118 "%d for %s: error=%d\n", i
+ 1, algo
, res
);
1124 /* Provide remaining output space */
1125 req
.avail_out
+= COMP_BUF_SIZE
- ctemplate
[i
].outlen
/ 2;
1127 res
= crypto_compress_final(tfm
, &req
);
1129 pr_err("alg: pcomp: compression final failed on test "
1130 "%d for %s: error=%d\n", i
+ 1, algo
, res
);
1135 if (COMP_BUF_SIZE
- req
.avail_out
!= ctemplate
[i
].outlen
) {
1136 pr_err("alg: comp: Compression test %d failed for %s: "
1137 "output len = %d (expected %d)\n", i
+ 1, algo
,
1138 COMP_BUF_SIZE
- req
.avail_out
,
1139 ctemplate
[i
].outlen
);
1143 if (produced
!= ctemplate
[i
].outlen
) {
1144 pr_err("alg: comp: Compression test %d failed for %s: "
1145 "returned len = %u (expected %d)\n", i
+ 1,
1146 algo
, produced
, ctemplate
[i
].outlen
);
1150 if (memcmp(result
, ctemplate
[i
].output
, ctemplate
[i
].outlen
)) {
1151 pr_err("alg: pcomp: Compression test %d failed for "
1152 "%s\n", i
+ 1, algo
);
1153 hexdump(result
, ctemplate
[i
].outlen
);
1158 for (i
= 0; i
< dtcount
; i
++) {
1159 struct comp_request req
;
1160 unsigned int produced
= 0;
1162 res
= crypto_decompress_setup(tfm
, dtemplate
[i
].params
,
1163 dtemplate
[i
].paramsize
);
1165 pr_err("alg: pcomp: decompression setup failed on "
1166 "test %d for %s: error=%d\n", i
+ 1, algo
, res
);
1170 res
= crypto_decompress_init(tfm
);
1172 pr_err("alg: pcomp: decompression init failed on test "
1173 "%d for %s: error=%d\n", i
+ 1, algo
, res
);
1177 memset(result
, 0, sizeof(result
));
1179 req
.next_in
= dtemplate
[i
].input
;
1180 req
.avail_in
= dtemplate
[i
].inlen
/ 2;
1181 req
.next_out
= result
;
1182 req
.avail_out
= dtemplate
[i
].outlen
/ 2;
1184 res
= crypto_decompress_update(tfm
, &req
);
1185 if (res
< 0 && (res
!= -EAGAIN
|| req
.avail_in
)) {
1186 pr_err("alg: pcomp: decompression update failed on "
1187 "test %d for %s: error=%d\n", i
+ 1, algo
, res
);
1193 /* Add remaining input data */
1194 req
.avail_in
+= (dtemplate
[i
].inlen
+ 1) / 2;
1196 res
= crypto_decompress_update(tfm
, &req
);
1197 if (res
< 0 && (res
!= -EAGAIN
|| req
.avail_in
)) {
1198 pr_err("alg: pcomp: decompression update failed on "
1199 "test %d for %s: error=%d\n", i
+ 1, algo
, res
);
1205 /* Provide remaining output space */
1206 req
.avail_out
+= COMP_BUF_SIZE
- dtemplate
[i
].outlen
/ 2;
1208 res
= crypto_decompress_final(tfm
, &req
);
1209 if (res
< 0 && (res
!= -EAGAIN
|| req
.avail_in
)) {
1210 pr_err("alg: pcomp: decompression final failed on "
1211 "test %d for %s: error=%d\n", i
+ 1, algo
, res
);
1217 if (COMP_BUF_SIZE
- req
.avail_out
!= dtemplate
[i
].outlen
) {
1218 pr_err("alg: comp: Decompression test %d failed for "
1219 "%s: output len = %d (expected %d)\n", i
+ 1,
1220 algo
, COMP_BUF_SIZE
- req
.avail_out
,
1221 dtemplate
[i
].outlen
);
1225 if (produced
!= dtemplate
[i
].outlen
) {
1226 pr_err("alg: comp: Decompression test %d failed for "
1227 "%s: returned len = %u (expected %d)\n", i
+ 1,
1228 algo
, produced
, dtemplate
[i
].outlen
);
1232 if (memcmp(result
, dtemplate
[i
].output
, dtemplate
[i
].outlen
)) {
1233 pr_err("alg: pcomp: Decompression test %d failed for "
1234 "%s\n", i
+ 1, algo
);
1235 hexdump(result
, dtemplate
[i
].outlen
);
1244 static int test_cprng(struct crypto_rng
*tfm
, struct cprng_testvec
*template,
1245 unsigned int tcount
)
1247 const char *algo
= crypto_tfm_alg_driver_name(crypto_rng_tfm(tfm
));
1248 int err
= 0, i
, j
, seedsize
;
1252 seedsize
= crypto_rng_seedsize(tfm
);
1254 seed
= kmalloc(seedsize
, GFP_KERNEL
);
1256 printk(KERN_ERR
"alg: cprng: Failed to allocate seed space "
1261 for (i
= 0; i
< tcount
; i
++) {
1262 memset(result
, 0, 32);
1264 memcpy(seed
, template[i
].v
, template[i
].vlen
);
1265 memcpy(seed
+ template[i
].vlen
, template[i
].key
,
1267 memcpy(seed
+ template[i
].vlen
+ template[i
].klen
,
1268 template[i
].dt
, template[i
].dtlen
);
1270 err
= crypto_rng_reset(tfm
, seed
, seedsize
);
1272 printk(KERN_ERR
"alg: cprng: Failed to reset rng "
1277 for (j
= 0; j
< template[i
].loops
; j
++) {
1278 err
= crypto_rng_get_bytes(tfm
, result
,
1280 if (err
!= template[i
].rlen
) {
1281 printk(KERN_ERR
"alg: cprng: Failed to obtain "
1282 "the correct amount of random data for "
1283 "%s (requested %d, got %d)\n", algo
,
1284 template[i
].rlen
, err
);
1289 err
= memcmp(result
, template[i
].result
,
1292 printk(KERN_ERR
"alg: cprng: Test %d failed for %s\n",
1294 hexdump(result
, template[i
].rlen
);
1305 static int alg_test_aead(const struct alg_test_desc
*desc
, const char *driver
,
1308 struct crypto_aead
*tfm
;
1311 tfm
= crypto_alloc_aead(driver
, type
, mask
);
1313 printk(KERN_ERR
"alg: aead: Failed to load transform for %s: "
1314 "%ld\n", driver
, PTR_ERR(tfm
));
1315 return PTR_ERR(tfm
);
1318 if (desc
->suite
.aead
.enc
.vecs
) {
1319 err
= test_aead(tfm
, ENCRYPT
, desc
->suite
.aead
.enc
.vecs
,
1320 desc
->suite
.aead
.enc
.count
);
1325 if (!err
&& desc
->suite
.aead
.dec
.vecs
)
1326 err
= test_aead(tfm
, DECRYPT
, desc
->suite
.aead
.dec
.vecs
,
1327 desc
->suite
.aead
.dec
.count
);
1330 crypto_free_aead(tfm
);
1334 static int alg_test_cipher(const struct alg_test_desc
*desc
,
1335 const char *driver
, u32 type
, u32 mask
)
1337 struct crypto_cipher
*tfm
;
1340 tfm
= crypto_alloc_cipher(driver
, type
, mask
);
1342 printk(KERN_ERR
"alg: cipher: Failed to load transform for "
1343 "%s: %ld\n", driver
, PTR_ERR(tfm
));
1344 return PTR_ERR(tfm
);
1347 if (desc
->suite
.cipher
.enc
.vecs
) {
1348 err
= test_cipher(tfm
, ENCRYPT
, desc
->suite
.cipher
.enc
.vecs
,
1349 desc
->suite
.cipher
.enc
.count
);
1354 if (desc
->suite
.cipher
.dec
.vecs
)
1355 err
= test_cipher(tfm
, DECRYPT
, desc
->suite
.cipher
.dec
.vecs
,
1356 desc
->suite
.cipher
.dec
.count
);
1359 crypto_free_cipher(tfm
);
1363 static int alg_test_skcipher(const struct alg_test_desc
*desc
,
1364 const char *driver
, u32 type
, u32 mask
)
1366 struct crypto_ablkcipher
*tfm
;
1369 tfm
= crypto_alloc_ablkcipher(driver
, type
, mask
);
1371 printk(KERN_ERR
"alg: skcipher: Failed to load transform for "
1372 "%s: %ld\n", driver
, PTR_ERR(tfm
));
1373 return PTR_ERR(tfm
);
1376 if (desc
->suite
.cipher
.enc
.vecs
) {
1377 err
= test_skcipher(tfm
, ENCRYPT
, desc
->suite
.cipher
.enc
.vecs
,
1378 desc
->suite
.cipher
.enc
.count
);
1383 if (desc
->suite
.cipher
.dec
.vecs
)
1384 err
= test_skcipher(tfm
, DECRYPT
, desc
->suite
.cipher
.dec
.vecs
,
1385 desc
->suite
.cipher
.dec
.count
);
1388 crypto_free_ablkcipher(tfm
);
1392 static int alg_test_comp(const struct alg_test_desc
*desc
, const char *driver
,
1395 struct crypto_comp
*tfm
;
1398 tfm
= crypto_alloc_comp(driver
, type
, mask
);
1400 printk(KERN_ERR
"alg: comp: Failed to load transform for %s: "
1401 "%ld\n", driver
, PTR_ERR(tfm
));
1402 return PTR_ERR(tfm
);
1405 err
= test_comp(tfm
, desc
->suite
.comp
.comp
.vecs
,
1406 desc
->suite
.comp
.decomp
.vecs
,
1407 desc
->suite
.comp
.comp
.count
,
1408 desc
->suite
.comp
.decomp
.count
);
1410 crypto_free_comp(tfm
);
1414 static int alg_test_pcomp(const struct alg_test_desc
*desc
, const char *driver
,
1417 struct crypto_pcomp
*tfm
;
1420 tfm
= crypto_alloc_pcomp(driver
, type
, mask
);
1422 pr_err("alg: pcomp: Failed to load transform for %s: %ld\n",
1423 driver
, PTR_ERR(tfm
));
1424 return PTR_ERR(tfm
);
1427 err
= test_pcomp(tfm
, desc
->suite
.pcomp
.comp
.vecs
,
1428 desc
->suite
.pcomp
.decomp
.vecs
,
1429 desc
->suite
.pcomp
.comp
.count
,
1430 desc
->suite
.pcomp
.decomp
.count
);
1432 crypto_free_pcomp(tfm
);
1436 static int alg_test_hash(const struct alg_test_desc
*desc
, const char *driver
,
1439 struct crypto_ahash
*tfm
;
1442 tfm
= crypto_alloc_ahash(driver
, type
, mask
);
1444 printk(KERN_ERR
"alg: hash: Failed to load transform for %s: "
1445 "%ld\n", driver
, PTR_ERR(tfm
));
1446 return PTR_ERR(tfm
);
1449 err
= test_hash(tfm
, desc
->suite
.hash
.vecs
,
1450 desc
->suite
.hash
.count
, true);
1452 err
= test_hash(tfm
, desc
->suite
.hash
.vecs
,
1453 desc
->suite
.hash
.count
, false);
1455 crypto_free_ahash(tfm
);
1459 static int alg_test_crc32c(const struct alg_test_desc
*desc
,
1460 const char *driver
, u32 type
, u32 mask
)
1462 struct crypto_shash
*tfm
;
1466 err
= alg_test_hash(desc
, driver
, type
, mask
);
1470 tfm
= crypto_alloc_shash(driver
, type
, mask
);
1472 printk(KERN_ERR
"alg: crc32c: Failed to load transform for %s: "
1473 "%ld\n", driver
, PTR_ERR(tfm
));
1480 struct shash_desc shash
;
1481 char ctx
[crypto_shash_descsize(tfm
)];
1484 sdesc
.shash
.tfm
= tfm
;
1485 sdesc
.shash
.flags
= 0;
1487 *(u32
*)sdesc
.ctx
= le32_to_cpu(420553207);
1488 err
= crypto_shash_final(&sdesc
.shash
, (u8
*)&val
);
1490 printk(KERN_ERR
"alg: crc32c: Operation failed for "
1491 "%s: %d\n", driver
, err
);
1495 if (val
!= ~420553207) {
1496 printk(KERN_ERR
"alg: crc32c: Test failed for %s: "
1497 "%d\n", driver
, val
);
1502 crypto_free_shash(tfm
);
1508 static int alg_test_cprng(const struct alg_test_desc
*desc
, const char *driver
,
1511 struct crypto_rng
*rng
;
1514 rng
= crypto_alloc_rng(driver
, type
, mask
);
1516 printk(KERN_ERR
"alg: cprng: Failed to load transform for %s: "
1517 "%ld\n", driver
, PTR_ERR(rng
));
1518 return PTR_ERR(rng
);
1521 err
= test_cprng(rng
, desc
->suite
.cprng
.vecs
, desc
->suite
.cprng
.count
);
1523 crypto_free_rng(rng
);
1528 static int alg_test_null(const struct alg_test_desc
*desc
,
1529 const char *driver
, u32 type
, u32 mask
)
1534 /* Please keep this list sorted by algorithm name. */
1535 static const struct alg_test_desc alg_test_descs
[] = {
1537 .alg
= "__driver-cbc-aes-aesni",
1538 .test
= alg_test_null
,
1552 .alg
= "__driver-ecb-aes-aesni",
1553 .test
= alg_test_null
,
1567 .alg
= "__ghash-pclmulqdqni",
1568 .test
= alg_test_null
,
1576 .alg
= "ansi_cprng",
1577 .test
= alg_test_cprng
,
1581 .vecs
= ansi_cprng_aes_tv_template
,
1582 .count
= ANSI_CPRNG_AES_TEST_VECTORS
1587 .test
= alg_test_skcipher
,
1592 .vecs
= aes_cbc_enc_tv_template
,
1593 .count
= AES_CBC_ENC_TEST_VECTORS
1596 .vecs
= aes_cbc_dec_tv_template
,
1597 .count
= AES_CBC_DEC_TEST_VECTORS
1602 .alg
= "cbc(anubis)",
1603 .test
= alg_test_skcipher
,
1607 .vecs
= anubis_cbc_enc_tv_template
,
1608 .count
= ANUBIS_CBC_ENC_TEST_VECTORS
1611 .vecs
= anubis_cbc_dec_tv_template
,
1612 .count
= ANUBIS_CBC_DEC_TEST_VECTORS
1617 .alg
= "cbc(blowfish)",
1618 .test
= alg_test_skcipher
,
1622 .vecs
= bf_cbc_enc_tv_template
,
1623 .count
= BF_CBC_ENC_TEST_VECTORS
1626 .vecs
= bf_cbc_dec_tv_template
,
1627 .count
= BF_CBC_DEC_TEST_VECTORS
1632 .alg
= "cbc(camellia)",
1633 .test
= alg_test_skcipher
,
1637 .vecs
= camellia_cbc_enc_tv_template
,
1638 .count
= CAMELLIA_CBC_ENC_TEST_VECTORS
1641 .vecs
= camellia_cbc_dec_tv_template
,
1642 .count
= CAMELLIA_CBC_DEC_TEST_VECTORS
1648 .test
= alg_test_skcipher
,
1652 .vecs
= des_cbc_enc_tv_template
,
1653 .count
= DES_CBC_ENC_TEST_VECTORS
1656 .vecs
= des_cbc_dec_tv_template
,
1657 .count
= DES_CBC_DEC_TEST_VECTORS
1662 .alg
= "cbc(des3_ede)",
1663 .test
= alg_test_skcipher
,
1668 .vecs
= des3_ede_cbc_enc_tv_template
,
1669 .count
= DES3_EDE_CBC_ENC_TEST_VECTORS
1672 .vecs
= des3_ede_cbc_dec_tv_template
,
1673 .count
= DES3_EDE_CBC_DEC_TEST_VECTORS
1678 .alg
= "cbc(twofish)",
1679 .test
= alg_test_skcipher
,
1683 .vecs
= tf_cbc_enc_tv_template
,
1684 .count
= TF_CBC_ENC_TEST_VECTORS
1687 .vecs
= tf_cbc_dec_tv_template
,
1688 .count
= TF_CBC_DEC_TEST_VECTORS
1694 .test
= alg_test_aead
,
1699 .vecs
= aes_ccm_enc_tv_template
,
1700 .count
= AES_CCM_ENC_TEST_VECTORS
1703 .vecs
= aes_ccm_dec_tv_template
,
1704 .count
= AES_CCM_DEC_TEST_VECTORS
1710 .test
= alg_test_crc32c
,
1714 .vecs
= crc32c_tv_template
,
1715 .count
= CRC32C_TEST_VECTORS
1719 .alg
= "cryptd(__driver-ecb-aes-aesni)",
1720 .test
= alg_test_null
,
1734 .alg
= "cryptd(__ghash-pclmulqdqni)",
1735 .test
= alg_test_null
,
1744 .test
= alg_test_skcipher
,
1749 .vecs
= aes_ctr_enc_tv_template
,
1750 .count
= AES_CTR_ENC_TEST_VECTORS
1753 .vecs
= aes_ctr_dec_tv_template
,
1754 .count
= AES_CTR_DEC_TEST_VECTORS
1759 .alg
= "cts(cbc(aes))",
1760 .test
= alg_test_skcipher
,
1764 .vecs
= cts_mode_enc_tv_template
,
1765 .count
= CTS_MODE_ENC_TEST_VECTORS
1768 .vecs
= cts_mode_dec_tv_template
,
1769 .count
= CTS_MODE_DEC_TEST_VECTORS
1775 .test
= alg_test_comp
,
1779 .vecs
= deflate_comp_tv_template
,
1780 .count
= DEFLATE_COMP_TEST_VECTORS
1783 .vecs
= deflate_decomp_tv_template
,
1784 .count
= DEFLATE_DECOMP_TEST_VECTORS
1789 .alg
= "ecb(__aes-aesni)",
1790 .test
= alg_test_null
,
1805 .test
= alg_test_skcipher
,
1810 .vecs
= aes_enc_tv_template
,
1811 .count
= AES_ENC_TEST_VECTORS
1814 .vecs
= aes_dec_tv_template
,
1815 .count
= AES_DEC_TEST_VECTORS
1820 .alg
= "ecb(anubis)",
1821 .test
= alg_test_skcipher
,
1825 .vecs
= anubis_enc_tv_template
,
1826 .count
= ANUBIS_ENC_TEST_VECTORS
1829 .vecs
= anubis_dec_tv_template
,
1830 .count
= ANUBIS_DEC_TEST_VECTORS
1836 .test
= alg_test_skcipher
,
1840 .vecs
= arc4_enc_tv_template
,
1841 .count
= ARC4_ENC_TEST_VECTORS
1844 .vecs
= arc4_dec_tv_template
,
1845 .count
= ARC4_DEC_TEST_VECTORS
1850 .alg
= "ecb(blowfish)",
1851 .test
= alg_test_skcipher
,
1855 .vecs
= bf_enc_tv_template
,
1856 .count
= BF_ENC_TEST_VECTORS
1859 .vecs
= bf_dec_tv_template
,
1860 .count
= BF_DEC_TEST_VECTORS
1865 .alg
= "ecb(camellia)",
1866 .test
= alg_test_skcipher
,
1870 .vecs
= camellia_enc_tv_template
,
1871 .count
= CAMELLIA_ENC_TEST_VECTORS
1874 .vecs
= camellia_dec_tv_template
,
1875 .count
= CAMELLIA_DEC_TEST_VECTORS
1880 .alg
= "ecb(cast5)",
1881 .test
= alg_test_skcipher
,
1885 .vecs
= cast5_enc_tv_template
,
1886 .count
= CAST5_ENC_TEST_VECTORS
1889 .vecs
= cast5_dec_tv_template
,
1890 .count
= CAST5_DEC_TEST_VECTORS
1895 .alg
= "ecb(cast6)",
1896 .test
= alg_test_skcipher
,
1900 .vecs
= cast6_enc_tv_template
,
1901 .count
= CAST6_ENC_TEST_VECTORS
1904 .vecs
= cast6_dec_tv_template
,
1905 .count
= CAST6_DEC_TEST_VECTORS
1911 .test
= alg_test_skcipher
,
1916 .vecs
= des_enc_tv_template
,
1917 .count
= DES_ENC_TEST_VECTORS
1920 .vecs
= des_dec_tv_template
,
1921 .count
= DES_DEC_TEST_VECTORS
1926 .alg
= "ecb(des3_ede)",
1927 .test
= alg_test_skcipher
,
1932 .vecs
= des3_ede_enc_tv_template
,
1933 .count
= DES3_EDE_ENC_TEST_VECTORS
1936 .vecs
= des3_ede_dec_tv_template
,
1937 .count
= DES3_EDE_DEC_TEST_VECTORS
1942 .alg
= "ecb(khazad)",
1943 .test
= alg_test_skcipher
,
1947 .vecs
= khazad_enc_tv_template
,
1948 .count
= KHAZAD_ENC_TEST_VECTORS
1951 .vecs
= khazad_dec_tv_template
,
1952 .count
= KHAZAD_DEC_TEST_VECTORS
1958 .test
= alg_test_skcipher
,
1962 .vecs
= seed_enc_tv_template
,
1963 .count
= SEED_ENC_TEST_VECTORS
1966 .vecs
= seed_dec_tv_template
,
1967 .count
= SEED_DEC_TEST_VECTORS
1972 .alg
= "ecb(serpent)",
1973 .test
= alg_test_skcipher
,
1977 .vecs
= serpent_enc_tv_template
,
1978 .count
= SERPENT_ENC_TEST_VECTORS
1981 .vecs
= serpent_dec_tv_template
,
1982 .count
= SERPENT_DEC_TEST_VECTORS
1988 .test
= alg_test_skcipher
,
1992 .vecs
= tea_enc_tv_template
,
1993 .count
= TEA_ENC_TEST_VECTORS
1996 .vecs
= tea_dec_tv_template
,
1997 .count
= TEA_DEC_TEST_VECTORS
2002 .alg
= "ecb(tnepres)",
2003 .test
= alg_test_skcipher
,
2007 .vecs
= tnepres_enc_tv_template
,
2008 .count
= TNEPRES_ENC_TEST_VECTORS
2011 .vecs
= tnepres_dec_tv_template
,
2012 .count
= TNEPRES_DEC_TEST_VECTORS
2017 .alg
= "ecb(twofish)",
2018 .test
= alg_test_skcipher
,
2022 .vecs
= tf_enc_tv_template
,
2023 .count
= TF_ENC_TEST_VECTORS
2026 .vecs
= tf_dec_tv_template
,
2027 .count
= TF_DEC_TEST_VECTORS
2033 .test
= alg_test_skcipher
,
2037 .vecs
= xeta_enc_tv_template
,
2038 .count
= XETA_ENC_TEST_VECTORS
2041 .vecs
= xeta_dec_tv_template
,
2042 .count
= XETA_DEC_TEST_VECTORS
2048 .test
= alg_test_skcipher
,
2052 .vecs
= xtea_enc_tv_template
,
2053 .count
= XTEA_ENC_TEST_VECTORS
2056 .vecs
= xtea_dec_tv_template
,
2057 .count
= XTEA_DEC_TEST_VECTORS
2063 .test
= alg_test_aead
,
2068 .vecs
= aes_gcm_enc_tv_template
,
2069 .count
= AES_GCM_ENC_TEST_VECTORS
2072 .vecs
= aes_gcm_dec_tv_template
,
2073 .count
= AES_GCM_DEC_TEST_VECTORS
2079 .test
= alg_test_hash
,
2083 .vecs
= ghash_tv_template
,
2084 .count
= GHASH_TEST_VECTORS
2089 .test
= alg_test_hash
,
2092 .vecs
= hmac_md5_tv_template
,
2093 .count
= HMAC_MD5_TEST_VECTORS
2097 .alg
= "hmac(rmd128)",
2098 .test
= alg_test_hash
,
2101 .vecs
= hmac_rmd128_tv_template
,
2102 .count
= HMAC_RMD128_TEST_VECTORS
2106 .alg
= "hmac(rmd160)",
2107 .test
= alg_test_hash
,
2110 .vecs
= hmac_rmd160_tv_template
,
2111 .count
= HMAC_RMD160_TEST_VECTORS
2115 .alg
= "hmac(sha1)",
2116 .test
= alg_test_hash
,
2120 .vecs
= hmac_sha1_tv_template
,
2121 .count
= HMAC_SHA1_TEST_VECTORS
2125 .alg
= "hmac(sha224)",
2126 .test
= alg_test_hash
,
2130 .vecs
= hmac_sha224_tv_template
,
2131 .count
= HMAC_SHA224_TEST_VECTORS
2135 .alg
= "hmac(sha256)",
2136 .test
= alg_test_hash
,
2140 .vecs
= hmac_sha256_tv_template
,
2141 .count
= HMAC_SHA256_TEST_VECTORS
2145 .alg
= "hmac(sha384)",
2146 .test
= alg_test_hash
,
2150 .vecs
= hmac_sha384_tv_template
,
2151 .count
= HMAC_SHA384_TEST_VECTORS
2155 .alg
= "hmac(sha512)",
2156 .test
= alg_test_hash
,
2160 .vecs
= hmac_sha512_tv_template
,
2161 .count
= HMAC_SHA512_TEST_VECTORS
2166 .test
= alg_test_skcipher
,
2170 .vecs
= aes_lrw_enc_tv_template
,
2171 .count
= AES_LRW_ENC_TEST_VECTORS
2174 .vecs
= aes_lrw_dec_tv_template
,
2175 .count
= AES_LRW_DEC_TEST_VECTORS
2181 .test
= alg_test_comp
,
2185 .vecs
= lzo_comp_tv_template
,
2186 .count
= LZO_COMP_TEST_VECTORS
2189 .vecs
= lzo_decomp_tv_template
,
2190 .count
= LZO_DECOMP_TEST_VECTORS
2196 .test
= alg_test_hash
,
2199 .vecs
= md4_tv_template
,
2200 .count
= MD4_TEST_VECTORS
2205 .test
= alg_test_hash
,
2208 .vecs
= md5_tv_template
,
2209 .count
= MD5_TEST_VECTORS
2213 .alg
= "michael_mic",
2214 .test
= alg_test_hash
,
2217 .vecs
= michael_mic_tv_template
,
2218 .count
= MICHAEL_MIC_TEST_VECTORS
2222 .alg
= "pcbc(fcrypt)",
2223 .test
= alg_test_skcipher
,
2227 .vecs
= fcrypt_pcbc_enc_tv_template
,
2228 .count
= FCRYPT_ENC_TEST_VECTORS
2231 .vecs
= fcrypt_pcbc_dec_tv_template
,
2232 .count
= FCRYPT_DEC_TEST_VECTORS
2237 .alg
= "rfc3686(ctr(aes))",
2238 .test
= alg_test_skcipher
,
2243 .vecs
= aes_ctr_rfc3686_enc_tv_template
,
2244 .count
= AES_CTR_3686_ENC_TEST_VECTORS
2247 .vecs
= aes_ctr_rfc3686_dec_tv_template
,
2248 .count
= AES_CTR_3686_DEC_TEST_VECTORS
2253 .alg
= "rfc4106(gcm(aes))",
2254 .test
= alg_test_aead
,
2258 .vecs
= aes_gcm_rfc4106_enc_tv_template
,
2259 .count
= AES_GCM_4106_ENC_TEST_VECTORS
2262 .vecs
= aes_gcm_rfc4106_dec_tv_template
,
2263 .count
= AES_GCM_4106_DEC_TEST_VECTORS
2270 .alg
= "rfc4309(ccm(aes))",
2271 .test
= alg_test_aead
,
2276 .vecs
= aes_ccm_rfc4309_enc_tv_template
,
2277 .count
= AES_CCM_4309_ENC_TEST_VECTORS
2280 .vecs
= aes_ccm_rfc4309_dec_tv_template
,
2281 .count
= AES_CCM_4309_DEC_TEST_VECTORS
2287 .test
= alg_test_hash
,
2290 .vecs
= rmd128_tv_template
,
2291 .count
= RMD128_TEST_VECTORS
2296 .test
= alg_test_hash
,
2299 .vecs
= rmd160_tv_template
,
2300 .count
= RMD160_TEST_VECTORS
2305 .test
= alg_test_hash
,
2308 .vecs
= rmd256_tv_template
,
2309 .count
= RMD256_TEST_VECTORS
2314 .test
= alg_test_hash
,
2317 .vecs
= rmd320_tv_template
,
2318 .count
= RMD320_TEST_VECTORS
2323 .test
= alg_test_skcipher
,
2327 .vecs
= salsa20_stream_enc_tv_template
,
2328 .count
= SALSA20_STREAM_ENC_TEST_VECTORS
2334 .test
= alg_test_hash
,
2338 .vecs
= sha1_tv_template
,
2339 .count
= SHA1_TEST_VECTORS
2344 .test
= alg_test_hash
,
2348 .vecs
= sha224_tv_template
,
2349 .count
= SHA224_TEST_VECTORS
2354 .test
= alg_test_hash
,
2358 .vecs
= sha256_tv_template
,
2359 .count
= SHA256_TEST_VECTORS
2364 .test
= alg_test_hash
,
2368 .vecs
= sha384_tv_template
,
2369 .count
= SHA384_TEST_VECTORS
2374 .test
= alg_test_hash
,
2378 .vecs
= sha512_tv_template
,
2379 .count
= SHA512_TEST_VECTORS
2384 .test
= alg_test_hash
,
2387 .vecs
= tgr128_tv_template
,
2388 .count
= TGR128_TEST_VECTORS
2393 .test
= alg_test_hash
,
2396 .vecs
= tgr160_tv_template
,
2397 .count
= TGR160_TEST_VECTORS
2402 .test
= alg_test_hash
,
2405 .vecs
= tgr192_tv_template
,
2406 .count
= TGR192_TEST_VECTORS
2411 .test
= alg_test_hash
,
2414 .vecs
= aes_vmac128_tv_template
,
2415 .count
= VMAC_AES_TEST_VECTORS
2420 .test
= alg_test_hash
,
2423 .vecs
= wp256_tv_template
,
2424 .count
= WP256_TEST_VECTORS
2429 .test
= alg_test_hash
,
2432 .vecs
= wp384_tv_template
,
2433 .count
= WP384_TEST_VECTORS
2438 .test
= alg_test_hash
,
2441 .vecs
= wp512_tv_template
,
2442 .count
= WP512_TEST_VECTORS
2447 .test
= alg_test_hash
,
2450 .vecs
= aes_xcbc128_tv_template
,
2451 .count
= XCBC_AES_TEST_VECTORS
2456 .test
= alg_test_skcipher
,
2461 .vecs
= aes_xts_enc_tv_template
,
2462 .count
= AES_XTS_ENC_TEST_VECTORS
2465 .vecs
= aes_xts_dec_tv_template
,
2466 .count
= AES_XTS_DEC_TEST_VECTORS
2472 .test
= alg_test_pcomp
,
2476 .vecs
= zlib_comp_tv_template
,
2477 .count
= ZLIB_COMP_TEST_VECTORS
2480 .vecs
= zlib_decomp_tv_template
,
2481 .count
= ZLIB_DECOMP_TEST_VECTORS
2488 static int alg_find_test(const char *alg
)
2491 int end
= ARRAY_SIZE(alg_test_descs
);
2493 while (start
< end
) {
2494 int i
= (start
+ end
) / 2;
2495 int diff
= strcmp(alg_test_descs
[i
].alg
, alg
);
2513 int alg_test(const char *driver
, const char *alg
, u32 type
, u32 mask
)
2519 if ((type
& CRYPTO_ALG_TYPE_MASK
) == CRYPTO_ALG_TYPE_CIPHER
) {
2520 char nalg
[CRYPTO_MAX_ALG_NAME
];
2522 if (snprintf(nalg
, sizeof(nalg
), "ecb(%s)", alg
) >=
2524 return -ENAMETOOLONG
;
2526 i
= alg_find_test(nalg
);
2530 if (fips_enabled
&& !alg_test_descs
[i
].fips_allowed
)
2533 rc
= alg_test_cipher(alg_test_descs
+ i
, driver
, type
, mask
);
2537 i
= alg_find_test(alg
);
2538 j
= alg_find_test(driver
);
2542 if (fips_enabled
&& ((i
>= 0 && !alg_test_descs
[i
].fips_allowed
) ||
2543 (j
>= 0 && !alg_test_descs
[j
].fips_allowed
)))
2548 rc
|= alg_test_descs
[i
].test(alg_test_descs
+ i
, driver
,
2551 rc
|= alg_test_descs
[j
].test(alg_test_descs
+ j
, driver
,
2555 if (fips_enabled
&& rc
)
2556 panic("%s: %s alg self test failed in fips mode!\n", driver
, alg
);
2558 if (fips_enabled
&& !rc
)
2559 printk(KERN_INFO
"alg: self-tests for %s (%s) passed\n",
2565 printk(KERN_INFO
"alg: No test for %s (%s)\n", alg
, driver
);
2571 #endif /* CONFIG_CRYPTO_MANAGER_DISABLE_TESTS */
2573 EXPORT_SYMBOL_GPL(alg_test
);