Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Gilad Ben-Yossef | 4923 | 32.37% | 11 | 5.58% |
Jussi Kivilinna | 2163 | 14.22% | 22 | 11.17% |
Herbert Xu | 2144 | 14.10% | 30 | 15.23% |
David S. Miller | 883 | 5.81% | 1 | 0.51% |
Robert Elliott | 744 | 4.89% | 1 | 0.51% |
Adrian-Ken Rueegsegger | 668 | 4.39% | 2 | 1.02% |
Tianjia Zhang | 344 | 2.26% | 5 | 2.54% |
James Morris | 292 | 1.92% | 17 | 8.63% |
Ard Biesheuvel | 285 | 1.87% | 8 | 4.06% |
Tim Chen | 265 | 1.74% | 2 | 1.02% |
Johannes Goetzfried | 244 | 1.60% | 3 | 1.52% |
Harald Welte | 218 | 1.43% | 1 | 0.51% |
Michal Ludvig | 205 | 1.35% | 1 | 0.51% |
Kees Cook | 171 | 1.12% | 1 | 0.51% |
Taehee Yoo | 156 | 1.03% | 2 | 1.02% |
Mikko Herranen | 126 | 0.83% | 1 | 0.51% |
Megha Dey | 116 | 0.76% | 2 | 1.02% |
Horia Geantă | 101 | 0.66% | 9 | 4.57% |
Loc Ho | 99 | 0.65% | 1 | 0.51% |
Vitaly Chikunov | 86 | 0.57% | 1 | 0.51% |
Raveendra Padasalagi | 79 | 0.52% | 2 | 1.02% |
Gustavo A. R. Silva | 78 | 0.51% | 1 | 0.51% |
Kartikey Mahendra Bhatt | 71 | 0.47% | 3 | 1.52% |
Luca Clementi | 67 | 0.44% | 1 | 0.51% |
Jarod Wilson | 65 | 0.43% | 4 | 2.03% |
Nitesh Lal | 50 | 0.33% | 1 | 0.51% |
Rik Snel | 41 | 0.27% | 2 | 1.02% |
Nicolas Royer | 40 | 0.26% | 1 | 0.51% |
Aaron Grothe | 39 | 0.26% | 7 | 3.55% |
Sebastian Andrzej Siewior | 37 | 0.24% | 3 | 1.52% |
Noriaki Takamiya | 34 | 0.22% | 1 | 0.51% |
Lokesh Vutla | 32 | 0.21% | 1 | 0.51% |
Eric Biggers | 29 | 0.19% | 2 | 1.02% |
Steffen Klassert | 29 | 0.19% | 1 | 0.51% |
Mark D Rustad | 25 | 0.16% | 1 | 0.51% |
Nathan Huckleberry | 23 | 0.15% | 3 | 1.52% |
Rusty Russell | 18 | 0.12% | 1 | 0.51% |
Robert Baronescu | 18 | 0.12% | 2 | 1.02% |
Tudor-Dan Ambarus | 17 | 0.11% | 1 | 0.51% |
Vladis Dronov | 16 | 0.11% | 1 | 0.51% |
Rabin Vincent | 15 | 0.10% | 1 | 0.51% |
Martin Willi | 14 | 0.09% | 1 | 0.51% |
Anirudh Venkataramanan | 13 | 0.09% | 4 | 2.03% |
Jonathan Lynch | 13 | 0.09% | 1 | 0.51% |
Tan Swee Heng | 11 | 0.07% | 2 | 1.02% |
Joy Latten | 10 | 0.07% | 2 | 1.02% |
Andrew Donofrio | 9 | 0.06% | 1 | 0.51% |
Christian Engelmayer | 8 | 0.05% | 1 | 0.51% |
Nicolai Stange | 8 | 0.05% | 1 | 0.51% |
Zoltan Sogor | 7 | 0.05% | 2 | 1.02% |
Jan Glauber | 6 | 0.04% | 1 | 0.51% |
Jon Oberheide | 6 | 0.04% | 1 | 0.51% |
Sonic Zhang | 5 | 0.03% | 1 | 0.51% |
Cristian Stoica | 5 | 0.03% | 2 | 1.02% |
David Howells | 4 | 0.03% | 1 | 0.51% |
Kamalesh Babulal | 4 | 0.03% | 1 | 0.51% |
Clay Haapala | 3 | 0.02% | 1 | 0.51% |
Geert Uytterhoeven | 3 | 0.02% | 1 | 0.51% |
Adrian Hoban | 3 | 0.02% | 1 | 0.51% |
Kevin Coffman | 3 | 0.02% | 1 | 0.51% |
Colin Ian King | 3 | 0.02% | 2 | 1.02% |
Shane Wang | 3 | 0.02% | 1 | 0.51% |
Thomas Gleixner | 2 | 0.01% | 1 | 0.51% |
Neil Horman | 2 | 0.01% | 1 | 0.51% |
Randy Dunlap | 2 | 0.01% | 1 | 0.51% |
David Sterba | 1 | 0.01% | 1 | 0.51% |
Patrick McHardy | 1 | 0.01% | 1 | 0.51% |
Kazunori Miyazawa | 1 | 0.01% | 1 | 0.51% |
David Härdeman | 1 | 0.01% | 1 | 0.51% |
Total | 15207 | 197 |
// SPDX-License-Identifier: GPL-2.0-or-later /* * Quick & dirty crypto testing module. * * This will only exist until we have a better testing mechanism * (e.g. a char device). * * Copyright (c) 2002 James Morris <jmorris@intercode.com.au> * Copyright (c) 2002 Jean-Francois Dive <jef@linuxbe.org> * Copyright (c) 2007 Nokia Siemens Networks * * Updated RFC4106 AES-GCM testing. * Authors: Aidan O'Mahony (aidan.o.mahony@intel.com) * Adrian Hoban <adrian.hoban@intel.com> * Gabriele Paoloni <gabriele.paoloni@intel.com> * Tadeusz Struk (tadeusz.struk@intel.com) * Copyright (c) 2010, Intel Corporation. */ #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt #include <crypto/aead.h> #include <crypto/hash.h> #include <crypto/skcipher.h> #include <linux/err.h> #include <linux/fips.h> #include <linux/init.h> #include <linux/interrupt.h> #include <linux/jiffies.h> #include <linux/kernel.h> #include <linux/module.h> #include <linux/moduleparam.h> #include <linux/scatterlist.h> #include <linux/slab.h> #include <linux/string.h> #include <linux/timex.h> #include "internal.h" #include "tcrypt.h" /* * Need slab memory for testing (size in number of pages). */ #define TVMEMSIZE 4 /* * Used by test_cipher_speed() */ #define ENCRYPT 1 #define DECRYPT 0 #define MAX_DIGEST_SIZE 64 /* * return a string with the driver name */ #define get_driver_name(tfm_type, tfm) crypto_tfm_alg_driver_name(tfm_type ## _tfm(tfm)) /* * Used by test_cipher_speed() */ static unsigned int sec; static char *alg; static u32 type; static u32 mask; static int mode; static u32 num_mb = 8; static unsigned int klen; static char *tvmem[TVMEMSIZE]; static const int block_sizes[] = { 16, 64, 128, 256, 1024, 1420, 4096, 0 }; static const int aead_sizes[] = { 16, 64, 256, 512, 1024, 1420, 4096, 8192, 0 }; #define XBUFSIZE 8 #define MAX_IVLEN 32 static int testmgr_alloc_buf(char *buf[XBUFSIZE]) { int i; for (i = 0; i < XBUFSIZE; i++) { buf[i] = (void *)__get_free_page(GFP_KERNEL); if (!buf[i]) goto err_free_buf; } return 0; err_free_buf: while (i-- > 0) free_page((unsigned long)buf[i]); return -ENOMEM; } static void testmgr_free_buf(char *buf[XBUFSIZE]) { int i; for (i = 0; i < XBUFSIZE; i++) free_page((unsigned long)buf[i]); } static void sg_init_aead(struct scatterlist *sg, char *xbuf[XBUFSIZE], unsigned int buflen, const void *assoc, unsigned int aad_size) { int np = (buflen + PAGE_SIZE - 1)/PAGE_SIZE; int k, rem; if (np > XBUFSIZE) { rem = PAGE_SIZE; np = XBUFSIZE; } else { rem = buflen % PAGE_SIZE; } sg_init_table(sg, np + 1); sg_set_buf(&sg[0], assoc, aad_size); if (rem) np--; for (k = 0; k < np; k++) sg_set_buf(&sg[k + 1], xbuf[k], PAGE_SIZE); if (rem) sg_set_buf(&sg[k + 1], xbuf[k], rem); } static inline int do_one_aead_op(struct aead_request *req, int ret) { struct crypto_wait *wait = req->base.data; return crypto_wait_req(ret, wait); } struct test_mb_aead_data { struct scatterlist sg[XBUFSIZE]; struct scatterlist sgout[XBUFSIZE]; struct aead_request *req; struct crypto_wait wait; char *xbuf[XBUFSIZE]; char *xoutbuf[XBUFSIZE]; char *axbuf[XBUFSIZE]; }; static int do_mult_aead_op(struct test_mb_aead_data *data, int enc, u32 num_mb, int *rc) { int i, err = 0; /* Fire up a bunch of concurrent requests */ for (i = 0; i < num_mb; i++) { if (enc == ENCRYPT) rc[i] = crypto_aead_encrypt(data[i].req); else rc[i] = crypto_aead_decrypt(data[i].req); } /* Wait for all requests to finish */ for (i = 0; i < num_mb; i++) { rc[i] = crypto_wait_req(rc[i], &data[i].wait); if (rc[i]) { pr_info("concurrent request %d error %d\n", i, rc[i]); err = rc[i]; } } return err; } static int test_mb_aead_jiffies(struct test_mb_aead_data *data, int enc, int blen, int secs, u32 num_mb) { unsigned long start, end; int bcount; int ret = 0; int *rc; rc = kcalloc(num_mb, sizeof(*rc), GFP_KERNEL); if (!rc) return -ENOMEM; for (start = jiffies, end = start + secs * HZ, bcount = 0; time_before(jiffies, end); bcount++) { ret = do_mult_aead_op(data, enc, num_mb, rc); if (ret) goto out; } pr_cont("%d operations in %d seconds (%llu bytes)\n", bcount * num_mb, secs, (u64)bcount * blen * num_mb); out: kfree(rc); return ret; } static int test_mb_aead_cycles(struct test_mb_aead_data *data, int enc, int blen, u32 num_mb) { unsigned long cycles = 0; int ret = 0; int i; int *rc; rc = kcalloc(num_mb, sizeof(*rc), GFP_KERNEL); if (!rc) return -ENOMEM; /* Warm-up run. */ for (i = 0; i < 4; i++) { ret = do_mult_aead_op(data, enc, num_mb, rc); if (ret) goto out; } /* The real thing. */ for (i = 0; i < 8; i++) { cycles_t start, end; start = get_cycles(); ret = do_mult_aead_op(data, enc, num_mb, rc); end = get_cycles(); if (ret) goto out; cycles += end - start; } pr_cont("1 operation in %lu cycles (%d bytes)\n", (cycles + 4) / (8 * num_mb), blen); out: kfree(rc); return ret; } static void test_mb_aead_speed(const char *algo, int enc, int secs, struct aead_speed_template *template, unsigned int tcount, u8 authsize, unsigned int aad_size, u8 *keysize, u32 num_mb) { struct test_mb_aead_data *data; struct crypto_aead *tfm; unsigned int i, j, iv_len; const int *b_size; const char *key; const char *e; void *assoc; char *iv; int ret; if (aad_size >= PAGE_SIZE) { pr_err("associate data length (%u) too big\n", aad_size); return; } iv = kzalloc(MAX_IVLEN, GFP_KERNEL); if (!iv) return; if (enc == ENCRYPT) e = "encryption"; else e = "decryption"; data = kcalloc(num_mb, sizeof(*data), GFP_KERNEL); if (!data) goto out_free_iv; tfm = crypto_alloc_aead(algo, 0, 0); if (IS_ERR(tfm)) { pr_err("failed to load transform for %s: %ld\n", algo, PTR_ERR(tfm)); goto out_free_data; } ret = crypto_aead_setauthsize(tfm, authsize); if (ret) { pr_err("alg: aead: Failed to setauthsize for %s: %d\n", algo, ret); goto out_free_tfm; } for (i = 0; i < num_mb; ++i) if (testmgr_alloc_buf(data[i].xbuf)) { while (i--) testmgr_free_buf(data[i].xbuf); goto out_free_tfm; } for (i = 0; i < num_mb; ++i) if (testmgr_alloc_buf(data[i].axbuf)) { while (i--) testmgr_free_buf(data[i].axbuf); goto out_free_xbuf; } for (i = 0; i < num_mb; ++i) if (testmgr_alloc_buf(data[i].xoutbuf)) { while (i--) testmgr_free_buf(data[i].xoutbuf); goto out_free_axbuf; } for (i = 0; i < num_mb; ++i) { data[i].req = aead_request_alloc(tfm, GFP_KERNEL); if (!data[i].req) { pr_err("alg: aead: Failed to allocate request for %s\n", algo); while (i--) aead_request_free(data[i].req); goto out_free_xoutbuf; } } for (i = 0; i < num_mb; ++i) { crypto_init_wait(&data[i].wait); aead_request_set_callback(data[i].req, CRYPTO_TFM_REQ_MAY_BACKLOG, crypto_req_done, &data[i].wait); } pr_info("testing speed of multibuffer %s (%s) %s\n", algo, get_driver_name(crypto_aead, tfm), e); i = 0; do { b_size = aead_sizes; do { int bs = round_up(*b_size, crypto_aead_blocksize(tfm)); if (bs + authsize > XBUFSIZE * PAGE_SIZE) { pr_err("template (%u) too big for buffer (%lu)\n", authsize + bs, XBUFSIZE * PAGE_SIZE); goto out; } pr_info("test %u (%d bit key, %d byte blocks): ", i, *keysize * 8, bs); /* Set up tfm global state, i.e. the key */ memset(tvmem[0], 0xff, PAGE_SIZE); key = tvmem[0]; for (j = 0; j < tcount; j++) { if (template[j].klen == *keysize) { key = template[j].key; break; } } crypto_aead_clear_flags(tfm, ~0); ret = crypto_aead_setkey(tfm, key, *keysize); if (ret) { pr_err("setkey() failed flags=%x\n", crypto_aead_get_flags(tfm)); goto out; } iv_len = crypto_aead_ivsize(tfm); if (iv_len) memset(iv, 0xff, iv_len); /* Now setup per request stuff, i.e. buffers */ for (j = 0; j < num_mb; ++j) { struct test_mb_aead_data *cur = &data[j]; assoc = cur->axbuf[0]; memset(assoc, 0xff, aad_size); sg_init_aead(cur->sg, cur->xbuf, bs + (enc ? 0 : authsize), assoc, aad_size); sg_init_aead(cur->sgout, cur->xoutbuf, bs + (enc ? authsize : 0), assoc, aad_size); aead_request_set_ad(cur->req, aad_size); if (!enc) { aead_request_set_crypt(cur->req, cur->sgout, cur->sg, bs, iv); ret = crypto_aead_encrypt(cur->req); ret = do_one_aead_op(cur->req, ret); if (ret) { pr_err("calculating auth failed (%d)\n", ret); break; } } aead_request_set_crypt(cur->req, cur->sg, cur->sgout, bs + (enc ? 0 : authsize), iv); } if (secs) { ret = test_mb_aead_jiffies(data, enc, bs, secs, num_mb); cond_resched(); } else { ret = test_mb_aead_cycles(data, enc, bs, num_mb); } if (ret) { pr_err("%s() failed return code=%d\n", e, ret); break; } b_size++; i++; } while (*b_size); keysize++; } while (*keysize); out: for (i = 0; i < num_mb; ++i) aead_request_free(data[i].req); out_free_xoutbuf: for (i = 0; i < num_mb; ++i) testmgr_free_buf(data[i].xoutbuf); out_free_axbuf: for (i = 0; i < num_mb; ++i) testmgr_free_buf(data[i].axbuf); out_free_xbuf: for (i = 0; i < num_mb; ++i) testmgr_free_buf(data[i].xbuf); out_free_tfm: crypto_free_aead(tfm); out_free_data: kfree(data); out_free_iv: kfree(iv); } static int test_aead_jiffies(struct aead_request *req, int enc, int blen, int secs) { unsigned long start, end; int bcount; int ret; for (start = jiffies, end = start + secs * HZ, bcount = 0; time_before(jiffies, end); bcount++) { if (enc) ret = do_one_aead_op(req, crypto_aead_encrypt(req)); else ret = do_one_aead_op(req, crypto_aead_decrypt(req)); if (ret) return ret; } pr_cont("%d operations in %d seconds (%llu bytes)\n", bcount, secs, (u64)bcount * blen); return 0; } static int test_aead_cycles(struct aead_request *req, int enc, int blen) { unsigned long cycles = 0; int ret = 0; int i; /* Warm-up run. */ for (i = 0; i < 4; i++) { if (enc) ret = do_one_aead_op(req, crypto_aead_encrypt(req)); else ret = do_one_aead_op(req, crypto_aead_decrypt(req)); if (ret) goto out; } /* The real thing. */ for (i = 0; i < 8; i++) { cycles_t start, end; start = get_cycles(); if (enc) ret = do_one_aead_op(req, crypto_aead_encrypt(req)); else ret = do_one_aead_op(req, crypto_aead_decrypt(req)); end = get_cycles(); if (ret) goto out; cycles += end - start; } out: if (ret == 0) pr_cont("1 operation in %lu cycles (%d bytes)\n", (cycles + 4) / 8, blen); return ret; } static void test_aead_speed(const char *algo, int enc, unsigned int secs, struct aead_speed_template *template, unsigned int tcount, u8 authsize, unsigned int aad_size, u8 *keysize) { unsigned int i, j; struct crypto_aead *tfm; int ret = -ENOMEM; const char *key; struct aead_request *req; struct scatterlist *sg; struct scatterlist *sgout; const char *e; void *assoc; char *iv; char *xbuf[XBUFSIZE]; char *xoutbuf[XBUFSIZE]; char *axbuf[XBUFSIZE]; const int *b_size; unsigned int iv_len; struct crypto_wait wait; iv = kzalloc(MAX_IVLEN, GFP_KERNEL); if (!iv) return; if (aad_size >= PAGE_SIZE) { pr_err("associate data length (%u) too big\n", aad_size); goto out_noxbuf; } if (enc == ENCRYPT) e = "encryption"; else e = "decryption"; if (testmgr_alloc_buf(xbuf)) goto out_noxbuf; if (testmgr_alloc_buf(axbuf)) goto out_noaxbuf; if (testmgr_alloc_buf(xoutbuf)) goto out_nooutbuf; sg = kmalloc(sizeof(*sg) * 9 * 2, GFP_KERNEL); if (!sg) goto out_nosg; sgout = &sg[9]; tfm = crypto_alloc_aead(algo, 0, 0); if (IS_ERR(tfm)) { pr_err("alg: aead: Failed to load transform for %s: %ld\n", algo, PTR_ERR(tfm)); goto out_notfm; } ret = crypto_aead_setauthsize(tfm, authsize); if (ret) { pr_err("alg: aead: Failed to setauthsize for %s: %d\n", algo, ret); goto out_noreq; } crypto_init_wait(&wait); pr_info("testing speed of %s (%s) %s\n", algo, get_driver_name(crypto_aead, tfm), e); req = aead_request_alloc(tfm, GFP_KERNEL); if (!req) { pr_err("alg: aead: Failed to allocate request for %s\n", algo); goto out_noreq; } aead_request_set_callback(req, CRYPTO_TFM_REQ_MAY_BACKLOG, crypto_req_done, &wait); i = 0; do { b_size = aead_sizes; do { u32 bs = round_up(*b_size, crypto_aead_blocksize(tfm)); assoc = axbuf[0]; memset(assoc, 0xff, aad_size); if ((*keysize + bs) > TVMEMSIZE * PAGE_SIZE) { pr_err("template (%u) too big for tvmem (%lu)\n", *keysize + bs, TVMEMSIZE * PAGE_SIZE); goto out; } key = tvmem[0]; for (j = 0; j < tcount; j++) { if (template[j].klen == *keysize) { key = template[j].key; break; } } ret = crypto_aead_setkey(tfm, key, *keysize); if (ret) { pr_err("setkey() failed flags=%x: %d\n", crypto_aead_get_flags(tfm), ret); goto out; } iv_len = crypto_aead_ivsize(tfm); if (iv_len) memset(iv, 0xff, iv_len); crypto_aead_clear_flags(tfm, ~0); pr_info("test %u (%d bit key, %d byte blocks): ", i, *keysize * 8, bs); memset(tvmem[0], 0xff, PAGE_SIZE); sg_init_aead(sg, xbuf, bs + (enc ? 0 : authsize), assoc, aad_size); sg_init_aead(sgout, xoutbuf, bs + (enc ? authsize : 0), assoc, aad_size); aead_request_set_ad(req, aad_size); if (!enc) { /* * For decryption we need a proper auth so * we do the encryption path once with buffers * reversed (input <-> output) to calculate it */ aead_request_set_crypt(req, sgout, sg, bs, iv); ret = do_one_aead_op(req, crypto_aead_encrypt(req)); if (ret) { pr_err("calculating auth failed (%d)\n", ret); break; } } aead_request_set_crypt(req, sg, sgout, bs + (enc ? 0 : authsize), iv); if (secs) { ret = test_aead_jiffies(req, enc, bs, secs); cond_resched(); } else { ret = test_aead_cycles(req, enc, bs); } if (ret) { pr_err("%s() failed return code=%d\n", e, ret); break; } b_size++; i++; } while (*b_size); keysize++; } while (*keysize); out: aead_request_free(req); out_noreq: crypto_free_aead(tfm); out_notfm: kfree(sg); out_nosg: testmgr_free_buf(xoutbuf); out_nooutbuf: testmgr_free_buf(axbuf); out_noaxbuf: testmgr_free_buf(xbuf); out_noxbuf: kfree(iv); } static void test_hash_sg_init(struct scatterlist *sg) { int i; sg_init_table(sg, TVMEMSIZE); for (i = 0; i < TVMEMSIZE; i++) { sg_set_buf(sg + i, tvmem[i], PAGE_SIZE); memset(tvmem[i], 0xff, PAGE_SIZE); } } static inline int do_one_ahash_op(struct ahash_request *req, int ret) { struct crypto_wait *wait = req->base.data; return crypto_wait_req(ret, wait); } static int test_ahash_jiffies_digest(struct ahash_request *req, int blen, char *out, int secs) { unsigned long start, end; int bcount; int ret; for (start = jiffies, end = start + secs * HZ, bcount = 0; time_before(jiffies, end); bcount++) { ret = do_one_ahash_op(req, crypto_ahash_digest(req)); if (ret) return ret; } pr_cont("%6u opers/sec, %9lu bytes/sec\n", bcount / secs, ((long)bcount * blen) / secs); return 0; } static int test_ahash_jiffies(struct ahash_request *req, int blen, int plen, char *out, int secs) { unsigned long start, end; int bcount, pcount; int ret; if (plen == blen) return test_ahash_jiffies_digest(req, blen, out, secs); for (start = jiffies, end = start + secs * HZ, bcount = 0; time_before(jiffies, end); bcount++) { ret = do_one_ahash_op(req, crypto_ahash_init(req)); if (ret) return ret; for (pcount = 0; pcount < blen; pcount += plen) { ret = do_one_ahash_op(req, crypto_ahash_update(req)); if (ret) return ret; } /* we assume there is enough space in 'out' for the result */ ret = do_one_ahash_op(req, crypto_ahash_final(req)); if (ret) return ret; } pr_cont("%6u opers/sec, %9lu bytes/sec\n", bcount / secs, ((long)bcount * blen) / secs); return 0; } static int test_ahash_cycles_digest(struct ahash_request *req, int blen, char *out) { unsigned long cycles = 0; int ret, i; /* Warm-up run. */ for (i = 0; i < 4; i++) { ret = do_one_ahash_op(req, crypto_ahash_digest(req)); if (ret) goto out; } /* The real thing. */ for (i = 0; i < 8; i++) { cycles_t start, end; start = get_cycles(); ret = do_one_ahash_op(req, crypto_ahash_digest(req)); if (ret) goto out; end = get_cycles(); cycles += end - start; } out: if (ret) return ret; pr_cont("%6lu cycles/operation, %4lu cycles/byte\n", cycles / 8, cycles / (8 * blen)); return 0; } static int test_ahash_cycles(struct ahash_request *req, int blen, int plen, char *out) { unsigned long cycles = 0; int i, pcount, ret; if (plen == blen) return test_ahash_cycles_digest(req, blen, out); /* Warm-up run. */ for (i = 0; i < 4; i++) { ret = do_one_ahash_op(req, crypto_ahash_init(req)); if (ret) goto out; for (pcount = 0; pcount < blen; pcount += plen) { ret = do_one_ahash_op(req, crypto_ahash_update(req)); if (ret) goto out; } ret = do_one_ahash_op(req, crypto_ahash_final(req)); if (ret) goto out; } /* The real thing. */ for (i = 0; i < 8; i++) { cycles_t start, end; start = get_cycles(); ret = do_one_ahash_op(req, crypto_ahash_init(req)); if (ret) goto out; for (pcount = 0; pcount < blen; pcount += plen) { ret = do_one_ahash_op(req, crypto_ahash_update(req)); if (ret) goto out; } ret = do_one_ahash_op(req, crypto_ahash_final(req)); if (ret) goto out; end = get_cycles(); cycles += end - start; } out: if (ret) return ret; pr_cont("%6lu cycles/operation, %4lu cycles/byte\n", cycles / 8, cycles / (8 * blen)); return 0; } static void test_ahash_speed_common(const char *algo, unsigned int secs, struct hash_speed *speed, unsigned mask) { struct scatterlist sg[TVMEMSIZE]; struct crypto_wait wait; struct ahash_request *req; struct crypto_ahash *tfm; char *output; int i, ret; tfm = crypto_alloc_ahash(algo, 0, mask); if (IS_ERR(tfm)) { pr_err("failed to load transform for %s: %ld\n", algo, PTR_ERR(tfm)); return; } pr_info("testing speed of async %s (%s)\n", algo, get_driver_name(crypto_ahash, tfm)); if (crypto_ahash_digestsize(tfm) > MAX_DIGEST_SIZE) { pr_err("digestsize(%u) > %d\n", crypto_ahash_digestsize(tfm), MAX_DIGEST_SIZE); goto out; } test_hash_sg_init(sg); req = ahash_request_alloc(tfm, GFP_KERNEL); if (!req) { pr_err("ahash request allocation failure\n"); goto out; } crypto_init_wait(&wait); ahash_request_set_callback(req, CRYPTO_TFM_REQ_MAY_BACKLOG, crypto_req_done, &wait); output = kmalloc(MAX_DIGEST_SIZE, GFP_KERNEL); if (!output) goto out_nomem; for (i = 0; speed[i].blen != 0; i++) { if (speed[i].blen > TVMEMSIZE * PAGE_SIZE) { pr_err("template (%u) too big for tvmem (%lu)\n", speed[i].blen, TVMEMSIZE * PAGE_SIZE); break; } if (klen) crypto_ahash_setkey(tfm, tvmem[0], klen); pr_info("test%3u " "(%5u byte blocks,%5u bytes per update,%4u updates): ", i, speed[i].blen, speed[i].plen, speed[i].blen / speed[i].plen); ahash_request_set_crypt(req, sg, output, speed[i].plen); if (secs) { ret = test_ahash_jiffies(req, speed[i].blen, speed[i].plen, output, secs); cond_resched(); } else { ret = test_ahash_cycles(req, speed[i].blen, speed[i].plen, output); } if (ret) { pr_err("hashing failed ret=%d\n", ret); break; } } kfree(output); out_nomem: ahash_request_free(req); out: crypto_free_ahash(tfm); } static void test_ahash_speed(const char *algo, unsigned int secs, struct hash_speed *speed) { return test_ahash_speed_common(algo, secs, speed, 0); } static void test_hash_speed(const char *algo, unsigned int secs, struct hash_speed *speed) { return test_ahash_speed_common(algo, secs, speed, CRYPTO_ALG_ASYNC); } struct test_mb_skcipher_data { struct scatterlist sg[XBUFSIZE]; struct skcipher_request *req; struct crypto_wait wait; char *xbuf[XBUFSIZE]; }; static int do_mult_acipher_op(struct test_mb_skcipher_data *data, int enc, u32 num_mb, int *rc) { int i, err = 0; /* Fire up a bunch of concurrent requests */ for (i = 0; i < num_mb; i++) { if (enc == ENCRYPT) rc[i] = crypto_skcipher_encrypt(data[i].req); else rc[i] = crypto_skcipher_decrypt(data[i].req); } /* Wait for all requests to finish */ for (i = 0; i < num_mb; i++) { rc[i] = crypto_wait_req(rc[i], &data[i].wait); if (rc[i]) { pr_info("concurrent request %d error %d\n", i, rc[i]); err = rc[i]; } } return err; } static int test_mb_acipher_jiffies(struct test_mb_skcipher_data *data, int enc, int blen, int secs, u32 num_mb) { unsigned long start, end; int bcount; int ret = 0; int *rc; rc = kcalloc(num_mb, sizeof(*rc), GFP_KERNEL); if (!rc) return -ENOMEM; for (start = jiffies, end = start + secs * HZ, bcount = 0; time_before(jiffies, end); bcount++) { ret = do_mult_acipher_op(data, enc, num_mb, rc); if (ret) goto out; } pr_cont("%d operations in %d seconds (%llu bytes)\n", bcount * num_mb, secs, (u64)bcount * blen * num_mb); out: kfree(rc); return ret; } static int test_mb_acipher_cycles(struct test_mb_skcipher_data *data, int enc, int blen, u32 num_mb) { unsigned long cycles = 0; int ret = 0; int i; int *rc; rc = kcalloc(num_mb, sizeof(*rc), GFP_KERNEL); if (!rc) return -ENOMEM; /* Warm-up run. */ for (i = 0; i < 4; i++) { ret = do_mult_acipher_op(data, enc, num_mb, rc); if (ret) goto out; } /* The real thing. */ for (i = 0; i < 8; i++) { cycles_t start, end; start = get_cycles(); ret = do_mult_acipher_op(data, enc, num_mb, rc); end = get_cycles(); if (ret) goto out; cycles += end - start; } pr_cont("1 operation in %lu cycles (%d bytes)\n", (cycles + 4) / (8 * num_mb), blen); out: kfree(rc); return ret; } static void test_mb_skcipher_speed(const char *algo, int enc, int secs, struct cipher_speed_template *template, unsigned int tcount, u8 *keysize, u32 num_mb) { struct test_mb_skcipher_data *data; struct crypto_skcipher *tfm; unsigned int i, j, iv_len; const int *b_size; const char *key; const char *e; char iv[128]; int ret; if (enc == ENCRYPT) e = "encryption"; else e = "decryption"; data = kcalloc(num_mb, sizeof(*data), GFP_KERNEL); if (!data) return; tfm = crypto_alloc_skcipher(algo, 0, 0); if (IS_ERR(tfm)) { pr_err("failed to load transform for %s: %ld\n", algo, PTR_ERR(tfm)); goto out_free_data; } for (i = 0; i < num_mb; ++i) if (testmgr_alloc_buf(data[i].xbuf)) { while (i--) testmgr_free_buf(data[i].xbuf); goto out_free_tfm; } for (i = 0; i < num_mb; ++i) { data[i].req = skcipher_request_alloc(tfm, GFP_KERNEL); if (!data[i].req) { pr_err("alg: skcipher: Failed to allocate request for %s\n", algo); while (i--) skcipher_request_free(data[i].req); goto out_free_xbuf; } } for (i = 0; i < num_mb; ++i) { skcipher_request_set_callback(data[i].req, CRYPTO_TFM_REQ_MAY_BACKLOG, crypto_req_done, &data[i].wait); crypto_init_wait(&data[i].wait); } pr_info("testing speed of multibuffer %s (%s) %s\n", algo, get_driver_name(crypto_skcipher, tfm), e); i = 0; do { b_size = block_sizes; do { u32 bs = round_up(*b_size, crypto_skcipher_blocksize(tfm)); if (bs > XBUFSIZE * PAGE_SIZE) { pr_err("template (%u) too big for buffer (%lu)\n", bs, XBUFSIZE * PAGE_SIZE); goto out; } pr_info("test %u (%d bit key, %d byte blocks): ", i, *keysize * 8, bs); /* Set up tfm global state, i.e. the key */ memset(tvmem[0], 0xff, PAGE_SIZE); key = tvmem[0]; for (j = 0; j < tcount; j++) { if (template[j].klen == *keysize) { key = template[j].key; break; } } crypto_skcipher_clear_flags(tfm, ~0); ret = crypto_skcipher_setkey(tfm, key, *keysize); if (ret) { pr_err("setkey() failed flags=%x\n", crypto_skcipher_get_flags(tfm)); goto out; } iv_len = crypto_skcipher_ivsize(tfm); if (iv_len) memset(&iv, 0xff, iv_len); /* Now setup per request stuff, i.e. buffers */ for (j = 0; j < num_mb; ++j) { struct test_mb_skcipher_data *cur = &data[j]; unsigned int k = bs; unsigned int pages = DIV_ROUND_UP(k, PAGE_SIZE); unsigned int p = 0; sg_init_table(cur->sg, pages); while (k > PAGE_SIZE) { sg_set_buf(cur->sg + p, cur->xbuf[p], PAGE_SIZE); memset(cur->xbuf[p], 0xff, PAGE_SIZE); p++; k -= PAGE_SIZE; } sg_set_buf(cur->sg + p, cur->xbuf[p], k); memset(cur->xbuf[p], 0xff, k); skcipher_request_set_crypt(cur->req, cur->sg, cur->sg, bs, iv); } if (secs) { ret = test_mb_acipher_jiffies(data, enc, bs, secs, num_mb); cond_resched(); } else { ret = test_mb_acipher_cycles(data, enc, bs, num_mb); } if (ret) { pr_err("%s() failed flags=%x\n", e, crypto_skcipher_get_flags(tfm)); break; } b_size++; i++; } while (*b_size); keysize++; } while (*keysize); out: for (i = 0; i < num_mb; ++i) skcipher_request_free(data[i].req); out_free_xbuf: for (i = 0; i < num_mb; ++i) testmgr_free_buf(data[i].xbuf); out_free_tfm: crypto_free_skcipher(tfm); out_free_data: kfree(data); } static inline int do_one_acipher_op(struct skcipher_request *req, int ret) { struct crypto_wait *wait = req->base.data; return crypto_wait_req(ret, wait); } static int test_acipher_jiffies(struct skcipher_request *req, int enc, int blen, int secs) { unsigned long start, end; int bcount; int ret; for (start = jiffies, end = start + secs * HZ, bcount = 0; time_before(jiffies, end); bcount++) { if (enc) ret = do_one_acipher_op(req, crypto_skcipher_encrypt(req)); else ret = do_one_acipher_op(req, crypto_skcipher_decrypt(req)); if (ret) return ret; } pr_cont("%d operations in %d seconds (%llu bytes)\n", bcount, secs, (u64)bcount * blen); return 0; } static int test_acipher_cycles(struct skcipher_request *req, int enc, int blen) { unsigned long cycles = 0; int ret = 0; int i; /* Warm-up run. */ for (i = 0; i < 4; i++) { if (enc) ret = do_one_acipher_op(req, crypto_skcipher_encrypt(req)); else ret = do_one_acipher_op(req, crypto_skcipher_decrypt(req)); if (ret) goto out; } /* The real thing. */ for (i = 0; i < 8; i++) { cycles_t start, end; start = get_cycles(); if (enc) ret = do_one_acipher_op(req, crypto_skcipher_encrypt(req)); else ret = do_one_acipher_op(req, crypto_skcipher_decrypt(req)); end = get_cycles(); if (ret) goto out; cycles += end - start; } out: if (ret == 0) pr_cont("1 operation in %lu cycles (%d bytes)\n", (cycles + 4) / 8, blen); return ret; } static void test_skcipher_speed(const char *algo, int enc, unsigned int secs, struct cipher_speed_template *template, unsigned int tcount, u8 *keysize, bool async) { unsigned int ret, i, j, k, iv_len; struct crypto_wait wait; const char *key; char iv[128]; struct skcipher_request *req; struct crypto_skcipher *tfm; const int *b_size; const char *e; if (enc == ENCRYPT) e = "encryption"; else e = "decryption"; crypto_init_wait(&wait); tfm = crypto_alloc_skcipher(algo, 0, async ? 0 : CRYPTO_ALG_ASYNC); if (IS_ERR(tfm)) { pr_err("failed to load transform for %s: %ld\n", algo, PTR_ERR(tfm)); return; } pr_info("testing speed of %s %s (%s) %s\n", async ? "async" : "sync", algo, get_driver_name(crypto_skcipher, tfm), e); req = skcipher_request_alloc(tfm, GFP_KERNEL); if (!req) { pr_err("skcipher: Failed to allocate request for %s\n", algo); goto out; } skcipher_request_set_callback(req, CRYPTO_TFM_REQ_MAY_BACKLOG, crypto_req_done, &wait); i = 0; do { b_size = block_sizes; do { u32 bs = round_up(*b_size, crypto_skcipher_blocksize(tfm)); struct scatterlist sg[TVMEMSIZE]; if ((*keysize + bs) > TVMEMSIZE * PAGE_SIZE) { pr_err("template (%u) too big for " "tvmem (%lu)\n", *keysize + bs, TVMEMSIZE * PAGE_SIZE); goto out_free_req; } pr_info("test %u (%d bit key, %d byte blocks): ", i, *keysize * 8, bs); memset(tvmem[0], 0xff, PAGE_SIZE); /* set key, plain text and IV */ key = tvmem[0]; for (j = 0; j < tcount; j++) { if (template[j].klen == *keysize) { key = template[j].key; break; } } crypto_skcipher_clear_flags(tfm, ~0); ret = crypto_skcipher_setkey(tfm, key, *keysize); if (ret) { pr_err("setkey() failed flags=%x\n", crypto_skcipher_get_flags(tfm)); goto out_free_req; } k = *keysize + bs; sg_init_table(sg, DIV_ROUND_UP(k, PAGE_SIZE)); if (k > PAGE_SIZE) { sg_set_buf(sg, tvmem[0] + *keysize, PAGE_SIZE - *keysize); k -= PAGE_SIZE; j = 1; while (k > PAGE_SIZE) { sg_set_buf(sg + j, tvmem[j], PAGE_SIZE); memset(tvmem[j], 0xff, PAGE_SIZE); j++; k -= PAGE_SIZE; } sg_set_buf(sg + j, tvmem[j], k); memset(tvmem[j], 0xff, k); } else { sg_set_buf(sg, tvmem[0] + *keysize, bs); } iv_len = crypto_skcipher_ivsize(tfm); if (iv_len) memset(&iv, 0xff, iv_len); skcipher_request_set_crypt(req, sg, sg, bs, iv); if (secs) { ret = test_acipher_jiffies(req, enc, bs, secs); cond_resched(); } else { ret = test_acipher_cycles(req, enc, bs); } if (ret) { pr_err("%s() failed flags=%x\n", e, crypto_skcipher_get_flags(tfm)); break; } b_size++; i++; } while (*b_size); keysize++; } while (*keysize); out_free_req: skcipher_request_free(req); out: crypto_free_skcipher(tfm); } static void test_acipher_speed(const char *algo, int enc, unsigned int secs, struct cipher_speed_template *template, unsigned int tcount, u8 *keysize) { return test_skcipher_speed(algo, enc, secs, template, tcount, keysize, true); } static void test_cipher_speed(const char *algo, int enc, unsigned int secs, struct cipher_speed_template *template, unsigned int tcount, u8 *keysize) { return test_skcipher_speed(algo, enc, secs, template, tcount, keysize, false); } static inline int tcrypt_test(const char *alg) { int ret; pr_debug("testing %s\n", alg); ret = alg_test(alg, alg, 0, 0); /* non-fips algs return -EINVAL or -ECANCELED in fips mode */ if (fips_enabled && (ret == -EINVAL || ret == -ECANCELED)) ret = 0; return ret; } static int do_test(const char *alg, u32 type, u32 mask, int m, u32 num_mb) { int i; int ret = 0; switch (m) { case 0: if (alg) { if (!crypto_has_alg(alg, type, mask ?: CRYPTO_ALG_TYPE_MASK)) ret = -ENOENT; break; } for (i = 1; i < 200; i++) ret = min(ret, do_test(NULL, 0, 0, i, num_mb)); break; case 1: ret = min(ret, tcrypt_test("md5")); break; case 2: ret = min(ret, tcrypt_test("sha1")); break; case 3: ret = min(ret, tcrypt_test("ecb(des)")); ret = min(ret, tcrypt_test("cbc(des)")); ret = min(ret, tcrypt_test("ctr(des)")); break; case 4: ret = min(ret, tcrypt_test("ecb(des3_ede)")); ret = min(ret, tcrypt_test("cbc(des3_ede)")); ret = min(ret, tcrypt_test("ctr(des3_ede)")); break; case 5: ret = min(ret, tcrypt_test("md4")); break; case 6: ret = min(ret, tcrypt_test("sha256")); break; case 7: ret = min(ret, tcrypt_test("ecb(blowfish)")); ret = min(ret, tcrypt_test("cbc(blowfish)")); ret = min(ret, tcrypt_test("ctr(blowfish)")); break; case 8: ret = min(ret, tcrypt_test("ecb(twofish)")); ret = min(ret, tcrypt_test("cbc(twofish)")); ret = min(ret, tcrypt_test("ctr(twofish)")); ret = min(ret, tcrypt_test("lrw(twofish)")); ret = min(ret, tcrypt_test("xts(twofish)")); break; case 9: ret = min(ret, tcrypt_test("ecb(serpent)")); ret = min(ret, tcrypt_test("cbc(serpent)")); ret = min(ret, tcrypt_test("ctr(serpent)")); ret = min(ret, tcrypt_test("lrw(serpent)")); ret = min(ret, tcrypt_test("xts(serpent)")); break; case 10: ret = min(ret, tcrypt_test("ecb(aes)")); ret = min(ret, tcrypt_test("cbc(aes)")); ret = min(ret, tcrypt_test("lrw(aes)")); ret = min(ret, tcrypt_test("xts(aes)")); ret = min(ret, tcrypt_test("ctr(aes)")); ret = min(ret, tcrypt_test("rfc3686(ctr(aes))")); ret = min(ret, tcrypt_test("xctr(aes)")); break; case 11: ret = min(ret, tcrypt_test("sha384")); break; case 12: ret = min(ret, tcrypt_test("sha512")); break; case 13: ret = min(ret, tcrypt_test("deflate")); break; case 14: ret = min(ret, tcrypt_test("ecb(cast5)")); ret = min(ret, tcrypt_test("cbc(cast5)")); ret = min(ret, tcrypt_test("ctr(cast5)")); break; case 15: ret = min(ret, tcrypt_test("ecb(cast6)")); ret = min(ret, tcrypt_test("cbc(cast6)")); ret = min(ret, tcrypt_test("ctr(cast6)")); ret = min(ret, tcrypt_test("lrw(cast6)")); ret = min(ret, tcrypt_test("xts(cast6)")); break; case 16: ret = min(ret, tcrypt_test("ecb(arc4)")); break; case 17: ret = min(ret, tcrypt_test("michael_mic")); break; case 18: ret = min(ret, tcrypt_test("crc32c")); break; case 19: ret = min(ret, tcrypt_test("ecb(tea)")); break; case 20: ret = min(ret, tcrypt_test("ecb(xtea)")); break; case 21: ret = min(ret, tcrypt_test("ecb(khazad)")); break; case 22: ret = min(ret, tcrypt_test("wp512")); break; case 23: ret = min(ret, tcrypt_test("wp384")); break; case 24: ret = min(ret, tcrypt_test("wp256")); break; case 26: ret = min(ret, tcrypt_test("ecb(anubis)")); ret = min(ret, tcrypt_test("cbc(anubis)")); break; case 30: ret = min(ret, tcrypt_test("ecb(xeta)")); break; case 31: ret = min(ret, tcrypt_test("pcbc(fcrypt)")); break; case 32: ret = min(ret, tcrypt_test("ecb(camellia)")); ret = min(ret, tcrypt_test("cbc(camellia)")); ret = min(ret, tcrypt_test("ctr(camellia)")); ret = min(ret, tcrypt_test("lrw(camellia)")); ret = min(ret, tcrypt_test("xts(camellia)")); break; case 33: ret = min(ret, tcrypt_test("sha224")); break; case 35: ret = min(ret, tcrypt_test("gcm(aes)")); break; case 36: ret = min(ret, tcrypt_test("lzo")); break; case 37: ret = min(ret, tcrypt_test("ccm(aes)")); break; case 38: ret = min(ret, tcrypt_test("cts(cbc(aes))")); break; case 39: ret = min(ret, tcrypt_test("xxhash64")); break; case 40: ret = min(ret, tcrypt_test("rmd160")); break; case 42: ret = min(ret, tcrypt_test("blake2b-512")); break; case 43: ret = min(ret, tcrypt_test("ecb(seed)")); break; case 45: ret = min(ret, tcrypt_test("rfc4309(ccm(aes))")); break; case 46: ret = min(ret, tcrypt_test("ghash")); break; case 47: ret = min(ret, tcrypt_test("crct10dif")); break; case 48: ret = min(ret, tcrypt_test("sha3-224")); break; case 49: ret = min(ret, tcrypt_test("sha3-256")); break; case 50: ret = min(ret, tcrypt_test("sha3-384")); break; case 51: ret = min(ret, tcrypt_test("sha3-512")); break; case 52: ret = min(ret, tcrypt_test("sm3")); break; case 53: ret = min(ret, tcrypt_test("streebog256")); break; case 54: ret = min(ret, tcrypt_test("streebog512")); break; case 55: ret = min(ret, tcrypt_test("gcm(sm4)")); break; case 56: ret = min(ret, tcrypt_test("ccm(sm4)")); break; case 57: ret = min(ret, tcrypt_test("polyval")); break; case 58: ret = min(ret, tcrypt_test("gcm(aria)")); break; case 59: ret = min(ret, tcrypt_test("cts(cbc(sm4))")); break; case 100: ret = min(ret, tcrypt_test("hmac(md5)")); break; case 101: ret = min(ret, tcrypt_test("hmac(sha1)")); break; case 102: ret = min(ret, tcrypt_test("hmac(sha256)")); break; case 103: ret = min(ret, tcrypt_test("hmac(sha384)")); break; case 104: ret = min(ret, tcrypt_test("hmac(sha512)")); break; case 105: ret = min(ret, tcrypt_test("hmac(sha224)")); break; case 106: ret = min(ret, tcrypt_test("xcbc(aes)")); break; case 108: ret = min(ret, tcrypt_test("hmac(rmd160)")); break; case 109: ret = min(ret, tcrypt_test("vmac64(aes)")); break; case 111: ret = min(ret, tcrypt_test("hmac(sha3-224)")); break; case 112: ret = min(ret, tcrypt_test("hmac(sha3-256)")); break; case 113: ret = min(ret, tcrypt_test("hmac(sha3-384)")); break; case 114: ret = min(ret, tcrypt_test("hmac(sha3-512)")); break; case 115: ret = min(ret, tcrypt_test("hmac(streebog256)")); break; case 116: ret = min(ret, tcrypt_test("hmac(streebog512)")); break; case 150: ret = min(ret, tcrypt_test("ansi_cprng")); break; case 151: ret = min(ret, tcrypt_test("rfc4106(gcm(aes))")); break; case 152: ret = min(ret, tcrypt_test("rfc4543(gcm(aes))")); break; case 153: ret = min(ret, tcrypt_test("cmac(aes)")); break; case 154: ret = min(ret, tcrypt_test("cmac(des3_ede)")); break; case 155: ret = min(ret, tcrypt_test("authenc(hmac(sha1),cbc(aes))")); break; case 156: ret = min(ret, tcrypt_test("authenc(hmac(md5),ecb(cipher_null))")); break; case 157: ret = min(ret, tcrypt_test("authenc(hmac(sha1),ecb(cipher_null))")); break; case 158: ret = min(ret, tcrypt_test("cbcmac(sm4)")); break; case 159: ret = min(ret, tcrypt_test("cmac(sm4)")); break; case 160: ret = min(ret, tcrypt_test("xcbc(sm4)")); break; case 181: ret = min(ret, tcrypt_test("authenc(hmac(sha1),cbc(des))")); break; case 182: ret = min(ret, tcrypt_test("authenc(hmac(sha1),cbc(des3_ede))")); break; case 183: ret = min(ret, tcrypt_test("authenc(hmac(sha224),cbc(des))")); break; case 184: ret = min(ret, tcrypt_test("authenc(hmac(sha224),cbc(des3_ede))")); break; case 185: ret = min(ret, tcrypt_test("authenc(hmac(sha256),cbc(des))")); break; case 186: ret = min(ret, tcrypt_test("authenc(hmac(sha256),cbc(des3_ede))")); break; case 187: ret = min(ret, tcrypt_test("authenc(hmac(sha384),cbc(des))")); break; case 188: ret = min(ret, tcrypt_test("authenc(hmac(sha384),cbc(des3_ede))")); break; case 189: ret = min(ret, tcrypt_test("authenc(hmac(sha512),cbc(des))")); break; case 190: ret = min(ret, tcrypt_test("authenc(hmac(sha512),cbc(des3_ede))")); break; case 191: ret = min(ret, tcrypt_test("ecb(sm4)")); ret = min(ret, tcrypt_test("cbc(sm4)")); ret = min(ret, tcrypt_test("ctr(sm4)")); ret = min(ret, tcrypt_test("xts(sm4)")); break; case 192: ret = min(ret, tcrypt_test("ecb(aria)")); ret = min(ret, tcrypt_test("cbc(aria)")); ret = min(ret, tcrypt_test("ctr(aria)")); break; case 193: ret = min(ret, tcrypt_test("ffdhe2048(dh)")); break; case 200: test_cipher_speed("ecb(aes)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("ecb(aes)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("cbc(aes)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("cbc(aes)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("lrw(aes)", ENCRYPT, sec, NULL, 0, speed_template_32_40_48); test_cipher_speed("lrw(aes)", DECRYPT, sec, NULL, 0, speed_template_32_40_48); test_cipher_speed("xts(aes)", ENCRYPT, sec, NULL, 0, speed_template_32_64); test_cipher_speed("xts(aes)", DECRYPT, sec, NULL, 0, speed_template_32_64); test_cipher_speed("cts(cbc(aes))", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("cts(cbc(aes))", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("ctr(aes)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("ctr(aes)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); break; case 201: test_cipher_speed("ecb(des3_ede)", ENCRYPT, sec, des3_speed_template, DES3_SPEED_VECTORS, speed_template_24); test_cipher_speed("ecb(des3_ede)", DECRYPT, sec, des3_speed_template, DES3_SPEED_VECTORS, speed_template_24); test_cipher_speed("cbc(des3_ede)", ENCRYPT, sec, des3_speed_template, DES3_SPEED_VECTORS, speed_template_24); test_cipher_speed("cbc(des3_ede)", DECRYPT, sec, des3_speed_template, DES3_SPEED_VECTORS, speed_template_24); test_cipher_speed("ctr(des3_ede)", ENCRYPT, sec, des3_speed_template, DES3_SPEED_VECTORS, speed_template_24); test_cipher_speed("ctr(des3_ede)", DECRYPT, sec, des3_speed_template, DES3_SPEED_VECTORS, speed_template_24); break; case 202: test_cipher_speed("ecb(twofish)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("ecb(twofish)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("cbc(twofish)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("cbc(twofish)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("ctr(twofish)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("ctr(twofish)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("lrw(twofish)", ENCRYPT, sec, NULL, 0, speed_template_32_40_48); test_cipher_speed("lrw(twofish)", DECRYPT, sec, NULL, 0, speed_template_32_40_48); test_cipher_speed("xts(twofish)", ENCRYPT, sec, NULL, 0, speed_template_32_48_64); test_cipher_speed("xts(twofish)", DECRYPT, sec, NULL, 0, speed_template_32_48_64); break; case 203: test_cipher_speed("ecb(blowfish)", ENCRYPT, sec, NULL, 0, speed_template_8_32); test_cipher_speed("ecb(blowfish)", DECRYPT, sec, NULL, 0, speed_template_8_32); test_cipher_speed("cbc(blowfish)", ENCRYPT, sec, NULL, 0, speed_template_8_32); test_cipher_speed("cbc(blowfish)", DECRYPT, sec, NULL, 0, speed_template_8_32); test_cipher_speed("ctr(blowfish)", ENCRYPT, sec, NULL, 0, speed_template_8_32); test_cipher_speed("ctr(blowfish)", DECRYPT, sec, NULL, 0, speed_template_8_32); break; case 204: test_cipher_speed("ecb(des)", ENCRYPT, sec, NULL, 0, speed_template_8); test_cipher_speed("ecb(des)", DECRYPT, sec, NULL, 0, speed_template_8); test_cipher_speed("cbc(des)", ENCRYPT, sec, NULL, 0, speed_template_8); test_cipher_speed("cbc(des)", DECRYPT, sec, NULL, 0, speed_template_8); break; case 205: test_cipher_speed("ecb(camellia)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("ecb(camellia)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("cbc(camellia)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("cbc(camellia)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("ctr(camellia)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("ctr(camellia)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("lrw(camellia)", ENCRYPT, sec, NULL, 0, speed_template_32_40_48); test_cipher_speed("lrw(camellia)", DECRYPT, sec, NULL, 0, speed_template_32_40_48); test_cipher_speed("xts(camellia)", ENCRYPT, sec, NULL, 0, speed_template_32_48_64); test_cipher_speed("xts(camellia)", DECRYPT, sec, NULL, 0, speed_template_32_48_64); break; case 207: test_cipher_speed("ecb(serpent)", ENCRYPT, sec, NULL, 0, speed_template_16_32); test_cipher_speed("ecb(serpent)", DECRYPT, sec, NULL, 0, speed_template_16_32); test_cipher_speed("cbc(serpent)", ENCRYPT, sec, NULL, 0, speed_template_16_32); test_cipher_speed("cbc(serpent)", DECRYPT, sec, NULL, 0, speed_template_16_32); test_cipher_speed("ctr(serpent)", ENCRYPT, sec, NULL, 0, speed_template_16_32); test_cipher_speed("ctr(serpent)", DECRYPT, sec, NULL, 0, speed_template_16_32); test_cipher_speed("lrw(serpent)", ENCRYPT, sec, NULL, 0, speed_template_32_48); test_cipher_speed("lrw(serpent)", DECRYPT, sec, NULL, 0, speed_template_32_48); test_cipher_speed("xts(serpent)", ENCRYPT, sec, NULL, 0, speed_template_32_64); test_cipher_speed("xts(serpent)", DECRYPT, sec, NULL, 0, speed_template_32_64); break; case 208: test_cipher_speed("ecb(arc4)", ENCRYPT, sec, NULL, 0, speed_template_8); break; case 209: test_cipher_speed("ecb(cast5)", ENCRYPT, sec, NULL, 0, speed_template_8_16); test_cipher_speed("ecb(cast5)", DECRYPT, sec, NULL, 0, speed_template_8_16); test_cipher_speed("cbc(cast5)", ENCRYPT, sec, NULL, 0, speed_template_8_16); test_cipher_speed("cbc(cast5)", DECRYPT, sec, NULL, 0, speed_template_8_16); test_cipher_speed("ctr(cast5)", ENCRYPT, sec, NULL, 0, speed_template_8_16); test_cipher_speed("ctr(cast5)", DECRYPT, sec, NULL, 0, speed_template_8_16); break; case 210: test_cipher_speed("ecb(cast6)", ENCRYPT, sec, NULL, 0, speed_template_16_32); test_cipher_speed("ecb(cast6)", DECRYPT, sec, NULL, 0, speed_template_16_32); test_cipher_speed("cbc(cast6)", ENCRYPT, sec, NULL, 0, speed_template_16_32); test_cipher_speed("cbc(cast6)", DECRYPT, sec, NULL, 0, speed_template_16_32); test_cipher_speed("ctr(cast6)", ENCRYPT, sec, NULL, 0, speed_template_16_32); test_cipher_speed("ctr(cast6)", DECRYPT, sec, NULL, 0, speed_template_16_32); test_cipher_speed("lrw(cast6)", ENCRYPT, sec, NULL, 0, speed_template_32_48); test_cipher_speed("lrw(cast6)", DECRYPT, sec, NULL, 0, speed_template_32_48); test_cipher_speed("xts(cast6)", ENCRYPT, sec, NULL, 0, speed_template_32_64); test_cipher_speed("xts(cast6)", DECRYPT, sec, NULL, 0, speed_template_32_64); break; case 211: test_aead_speed("rfc4106(gcm(aes))", ENCRYPT, sec, NULL, 0, 16, 16, aead_speed_template_20_28_36); test_aead_speed("gcm(aes)", ENCRYPT, sec, NULL, 0, 16, 8, speed_template_16_24_32); test_aead_speed("rfc4106(gcm(aes))", DECRYPT, sec, NULL, 0, 16, 16, aead_speed_template_20_28_36); test_aead_speed("gcm(aes)", DECRYPT, sec, NULL, 0, 16, 8, speed_template_16_24_32); break; case 212: test_aead_speed("rfc4309(ccm(aes))", ENCRYPT, sec, NULL, 0, 16, 16, aead_speed_template_19); test_aead_speed("rfc4309(ccm(aes))", DECRYPT, sec, NULL, 0, 16, 16, aead_speed_template_19); break; case 213: test_aead_speed("rfc7539esp(chacha20,poly1305)", ENCRYPT, sec, NULL, 0, 16, 8, aead_speed_template_36); test_aead_speed("rfc7539esp(chacha20,poly1305)", DECRYPT, sec, NULL, 0, 16, 8, aead_speed_template_36); break; case 214: test_cipher_speed("chacha20", ENCRYPT, sec, NULL, 0, speed_template_32); break; case 215: test_mb_aead_speed("rfc4106(gcm(aes))", ENCRYPT, sec, NULL, 0, 16, 16, aead_speed_template_20_28_36, num_mb); test_mb_aead_speed("gcm(aes)", ENCRYPT, sec, NULL, 0, 16, 8, speed_template_16_24_32, num_mb); test_mb_aead_speed("rfc4106(gcm(aes))", DECRYPT, sec, NULL, 0, 16, 16, aead_speed_template_20_28_36, num_mb); test_mb_aead_speed("gcm(aes)", DECRYPT, sec, NULL, 0, 16, 8, speed_template_16_24_32, num_mb); break; case 216: test_mb_aead_speed("rfc4309(ccm(aes))", ENCRYPT, sec, NULL, 0, 16, 16, aead_speed_template_19, num_mb); test_mb_aead_speed("rfc4309(ccm(aes))", DECRYPT, sec, NULL, 0, 16, 16, aead_speed_template_19, num_mb); break; case 217: test_mb_aead_speed("rfc7539esp(chacha20,poly1305)", ENCRYPT, sec, NULL, 0, 16, 8, aead_speed_template_36, num_mb); test_mb_aead_speed("rfc7539esp(chacha20,poly1305)", DECRYPT, sec, NULL, 0, 16, 8, aead_speed_template_36, num_mb); break; case 218: test_cipher_speed("ecb(sm4)", ENCRYPT, sec, NULL, 0, speed_template_16); test_cipher_speed("ecb(sm4)", DECRYPT, sec, NULL, 0, speed_template_16); test_cipher_speed("cbc(sm4)", ENCRYPT, sec, NULL, 0, speed_template_16); test_cipher_speed("cbc(sm4)", DECRYPT, sec, NULL, 0, speed_template_16); test_cipher_speed("cts(cbc(sm4))", ENCRYPT, sec, NULL, 0, speed_template_16); test_cipher_speed("cts(cbc(sm4))", DECRYPT, sec, NULL, 0, speed_template_16); test_cipher_speed("ctr(sm4)", ENCRYPT, sec, NULL, 0, speed_template_16); test_cipher_speed("ctr(sm4)", DECRYPT, sec, NULL, 0, speed_template_16); test_cipher_speed("xts(sm4)", ENCRYPT, sec, NULL, 0, speed_template_32); test_cipher_speed("xts(sm4)", DECRYPT, sec, NULL, 0, speed_template_32); break; case 219: test_cipher_speed("adiantum(xchacha12,aes)", ENCRYPT, sec, NULL, 0, speed_template_32); test_cipher_speed("adiantum(xchacha12,aes)", DECRYPT, sec, NULL, 0, speed_template_32); test_cipher_speed("adiantum(xchacha20,aes)", ENCRYPT, sec, NULL, 0, speed_template_32); test_cipher_speed("adiantum(xchacha20,aes)", DECRYPT, sec, NULL, 0, speed_template_32); break; case 220: test_acipher_speed("essiv(cbc(aes),sha256)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("essiv(cbc(aes),sha256)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); break; case 221: test_aead_speed("aegis128", ENCRYPT, sec, NULL, 0, 16, 8, speed_template_16); test_aead_speed("aegis128", DECRYPT, sec, NULL, 0, 16, 8, speed_template_16); break; case 222: test_aead_speed("gcm(sm4)", ENCRYPT, sec, NULL, 0, 16, 8, speed_template_16); test_aead_speed("gcm(sm4)", DECRYPT, sec, NULL, 0, 16, 8, speed_template_16); break; case 223: test_aead_speed("rfc4309(ccm(sm4))", ENCRYPT, sec, NULL, 0, 16, 16, aead_speed_template_19); test_aead_speed("rfc4309(ccm(sm4))", DECRYPT, sec, NULL, 0, 16, 16, aead_speed_template_19); break; case 224: test_mb_aead_speed("gcm(sm4)", ENCRYPT, sec, NULL, 0, 16, 8, speed_template_16, num_mb); test_mb_aead_speed("gcm(sm4)", DECRYPT, sec, NULL, 0, 16, 8, speed_template_16, num_mb); break; case 225: test_mb_aead_speed("rfc4309(ccm(sm4))", ENCRYPT, sec, NULL, 0, 16, 16, aead_speed_template_19, num_mb); test_mb_aead_speed("rfc4309(ccm(sm4))", DECRYPT, sec, NULL, 0, 16, 16, aead_speed_template_19, num_mb); break; case 226: test_cipher_speed("hctr2(aes)", ENCRYPT, sec, NULL, 0, speed_template_32); break; case 227: test_cipher_speed("ecb(aria)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("ecb(aria)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("cbc(aria)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("cbc(aria)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("ctr(aria)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_cipher_speed("ctr(aria)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); break; case 228: test_aead_speed("gcm(aria)", ENCRYPT, sec, NULL, 0, 16, 8, speed_template_16_24_32); test_aead_speed("gcm(aria)", DECRYPT, sec, NULL, 0, 16, 8, speed_template_16_24_32); break; case 229: test_mb_aead_speed("gcm(aria)", ENCRYPT, sec, NULL, 0, 16, 8, speed_template_16, num_mb); test_mb_aead_speed("gcm(aria)", DECRYPT, sec, NULL, 0, 16, 8, speed_template_16, num_mb); break; case 300: if (alg) { test_hash_speed(alg, sec, generic_hash_speed_template); break; } fallthrough; case 301: test_hash_speed("md4", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 302: test_hash_speed("md5", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 303: test_hash_speed("sha1", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 304: test_hash_speed("sha256", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 305: test_hash_speed("sha384", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 306: test_hash_speed("sha512", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 307: test_hash_speed("wp256", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 308: test_hash_speed("wp384", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 309: test_hash_speed("wp512", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 313: test_hash_speed("sha224", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 314: test_hash_speed("xxhash64", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 315: test_hash_speed("rmd160", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 317: test_hash_speed("blake2b-512", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 318: klen = 16; test_hash_speed("ghash", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 319: test_hash_speed("crc32c", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 320: test_hash_speed("crct10dif", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 321: test_hash_speed("poly1305", sec, poly1305_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 322: test_hash_speed("sha3-224", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 323: test_hash_speed("sha3-256", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 324: test_hash_speed("sha3-384", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 325: test_hash_speed("sha3-512", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 326: test_hash_speed("sm3", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 327: test_hash_speed("streebog256", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 328: test_hash_speed("streebog512", sec, generic_hash_speed_template); if (mode > 300 && mode < 400) break; fallthrough; case 399: break; case 400: if (alg) { test_ahash_speed(alg, sec, generic_hash_speed_template); break; } fallthrough; case 401: test_ahash_speed("md4", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 402: test_ahash_speed("md5", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 403: test_ahash_speed("sha1", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 404: test_ahash_speed("sha256", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 405: test_ahash_speed("sha384", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 406: test_ahash_speed("sha512", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 407: test_ahash_speed("wp256", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 408: test_ahash_speed("wp384", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 409: test_ahash_speed("wp512", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 413: test_ahash_speed("sha224", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 414: test_ahash_speed("xxhash64", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 415: test_ahash_speed("rmd160", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 417: test_ahash_speed("blake2b-512", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 418: test_ahash_speed("sha3-224", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 419: test_ahash_speed("sha3-256", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 420: test_ahash_speed("sha3-384", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 421: test_ahash_speed("sha3-512", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 422: test_ahash_speed("sm3", sec, generic_hash_speed_template); if (mode > 400 && mode < 500) break; fallthrough; case 499: break; case 500: test_acipher_speed("ecb(aes)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("ecb(aes)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("cbc(aes)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("cbc(aes)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("lrw(aes)", ENCRYPT, sec, NULL, 0, speed_template_32_40_48); test_acipher_speed("lrw(aes)", DECRYPT, sec, NULL, 0, speed_template_32_40_48); test_acipher_speed("xts(aes)", ENCRYPT, sec, NULL, 0, speed_template_32_64); test_acipher_speed("xts(aes)", DECRYPT, sec, NULL, 0, speed_template_32_64); test_acipher_speed("cts(cbc(aes))", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("cts(cbc(aes))", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("ctr(aes)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("ctr(aes)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("rfc3686(ctr(aes))", ENCRYPT, sec, NULL, 0, speed_template_20_28_36); test_acipher_speed("rfc3686(ctr(aes))", DECRYPT, sec, NULL, 0, speed_template_20_28_36); break; case 501: test_acipher_speed("ecb(des3_ede)", ENCRYPT, sec, des3_speed_template, DES3_SPEED_VECTORS, speed_template_24); test_acipher_speed("ecb(des3_ede)", DECRYPT, sec, des3_speed_template, DES3_SPEED_VECTORS, speed_template_24); test_acipher_speed("cbc(des3_ede)", ENCRYPT, sec, des3_speed_template, DES3_SPEED_VECTORS, speed_template_24); test_acipher_speed("cbc(des3_ede)", DECRYPT, sec, des3_speed_template, DES3_SPEED_VECTORS, speed_template_24); break; case 502: test_acipher_speed("ecb(des)", ENCRYPT, sec, NULL, 0, speed_template_8); test_acipher_speed("ecb(des)", DECRYPT, sec, NULL, 0, speed_template_8); test_acipher_speed("cbc(des)", ENCRYPT, sec, NULL, 0, speed_template_8); test_acipher_speed("cbc(des)", DECRYPT, sec, NULL, 0, speed_template_8); break; case 503: test_acipher_speed("ecb(serpent)", ENCRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("ecb(serpent)", DECRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("cbc(serpent)", ENCRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("cbc(serpent)", DECRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("ctr(serpent)", ENCRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("ctr(serpent)", DECRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("lrw(serpent)", ENCRYPT, sec, NULL, 0, speed_template_32_48); test_acipher_speed("lrw(serpent)", DECRYPT, sec, NULL, 0, speed_template_32_48); test_acipher_speed("xts(serpent)", ENCRYPT, sec, NULL, 0, speed_template_32_64); test_acipher_speed("xts(serpent)", DECRYPT, sec, NULL, 0, speed_template_32_64); break; case 504: test_acipher_speed("ecb(twofish)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("ecb(twofish)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("cbc(twofish)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("cbc(twofish)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("ctr(twofish)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("ctr(twofish)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("lrw(twofish)", ENCRYPT, sec, NULL, 0, speed_template_32_40_48); test_acipher_speed("lrw(twofish)", DECRYPT, sec, NULL, 0, speed_template_32_40_48); test_acipher_speed("xts(twofish)", ENCRYPT, sec, NULL, 0, speed_template_32_48_64); test_acipher_speed("xts(twofish)", DECRYPT, sec, NULL, 0, speed_template_32_48_64); break; case 505: test_acipher_speed("ecb(arc4)", ENCRYPT, sec, NULL, 0, speed_template_8); break; case 506: test_acipher_speed("ecb(cast5)", ENCRYPT, sec, NULL, 0, speed_template_8_16); test_acipher_speed("ecb(cast5)", DECRYPT, sec, NULL, 0, speed_template_8_16); test_acipher_speed("cbc(cast5)", ENCRYPT, sec, NULL, 0, speed_template_8_16); test_acipher_speed("cbc(cast5)", DECRYPT, sec, NULL, 0, speed_template_8_16); test_acipher_speed("ctr(cast5)", ENCRYPT, sec, NULL, 0, speed_template_8_16); test_acipher_speed("ctr(cast5)", DECRYPT, sec, NULL, 0, speed_template_8_16); break; case 507: test_acipher_speed("ecb(cast6)", ENCRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("ecb(cast6)", DECRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("cbc(cast6)", ENCRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("cbc(cast6)", DECRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("ctr(cast6)", ENCRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("ctr(cast6)", DECRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("lrw(cast6)", ENCRYPT, sec, NULL, 0, speed_template_32_48); test_acipher_speed("lrw(cast6)", DECRYPT, sec, NULL, 0, speed_template_32_48); test_acipher_speed("xts(cast6)", ENCRYPT, sec, NULL, 0, speed_template_32_64); test_acipher_speed("xts(cast6)", DECRYPT, sec, NULL, 0, speed_template_32_64); break; case 508: test_acipher_speed("ecb(camellia)", ENCRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("ecb(camellia)", DECRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("cbc(camellia)", ENCRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("cbc(camellia)", DECRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("ctr(camellia)", ENCRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("ctr(camellia)", DECRYPT, sec, NULL, 0, speed_template_16_32); test_acipher_speed("lrw(camellia)", ENCRYPT, sec, NULL, 0, speed_template_32_48); test_acipher_speed("lrw(camellia)", DECRYPT, sec, NULL, 0, speed_template_32_48); test_acipher_speed("xts(camellia)", ENCRYPT, sec, NULL, 0, speed_template_32_64); test_acipher_speed("xts(camellia)", DECRYPT, sec, NULL, 0, speed_template_32_64); break; case 509: test_acipher_speed("ecb(blowfish)", ENCRYPT, sec, NULL, 0, speed_template_8_32); test_acipher_speed("ecb(blowfish)", DECRYPT, sec, NULL, 0, speed_template_8_32); test_acipher_speed("cbc(blowfish)", ENCRYPT, sec, NULL, 0, speed_template_8_32); test_acipher_speed("cbc(blowfish)", DECRYPT, sec, NULL, 0, speed_template_8_32); test_acipher_speed("ctr(blowfish)", ENCRYPT, sec, NULL, 0, speed_template_8_32); test_acipher_speed("ctr(blowfish)", DECRYPT, sec, NULL, 0, speed_template_8_32); break; case 518: test_acipher_speed("ecb(sm4)", ENCRYPT, sec, NULL, 0, speed_template_16); test_acipher_speed("ecb(sm4)", DECRYPT, sec, NULL, 0, speed_template_16); test_acipher_speed("cbc(sm4)", ENCRYPT, sec, NULL, 0, speed_template_16); test_acipher_speed("cbc(sm4)", DECRYPT, sec, NULL, 0, speed_template_16); test_acipher_speed("ctr(sm4)", ENCRYPT, sec, NULL, 0, speed_template_16); test_acipher_speed("ctr(sm4)", DECRYPT, sec, NULL, 0, speed_template_16); test_acipher_speed("xts(sm4)", ENCRYPT, sec, NULL, 0, speed_template_32); test_acipher_speed("xts(sm4)", DECRYPT, sec, NULL, 0, speed_template_32); break; case 519: test_acipher_speed("ecb(aria)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("ecb(aria)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("ctr(aria)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32); test_acipher_speed("ctr(aria)", DECRYPT, sec, NULL, 0, speed_template_16_24_32); break; case 600: test_mb_skcipher_speed("ecb(aes)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32, num_mb); test_mb_skcipher_speed("ecb(aes)", DECRYPT, sec, NULL, 0, speed_template_16_24_32, num_mb); test_mb_skcipher_speed("cbc(aes)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32, num_mb); test_mb_skcipher_speed("cbc(aes)", DECRYPT, sec, NULL, 0, speed_template_16_24_32, num_mb); test_mb_skcipher_speed("lrw(aes)", ENCRYPT, sec, NULL, 0, speed_template_32_40_48, num_mb); test_mb_skcipher_speed("lrw(aes)", DECRYPT, sec, NULL, 0, speed_template_32_40_48, num_mb); test_mb_skcipher_speed("xts(aes)", ENCRYPT, sec, NULL, 0, speed_template_32_64, num_mb); test_mb_skcipher_speed("xts(aes)", DECRYPT, sec, NULL, 0, speed_template_32_64, num_mb); test_mb_skcipher_speed("cts(cbc(aes))", ENCRYPT, sec, NULL, 0, speed_template_16_24_32, num_mb); test_mb_skcipher_speed("cts(cbc(aes))", DECRYPT, sec, NULL, 0, speed_template_16_24_32, num_mb); test_mb_skcipher_speed("ctr(aes)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32, num_mb); test_mb_skcipher_speed("ctr(aes)", DECRYPT, sec, NULL, 0, speed_template_16_24_32, num_mb); test_mb_skcipher_speed("rfc3686(ctr(aes))", ENCRYPT, sec, NULL, 0, speed_template_20_28_36, num_mb); test_mb_skcipher_speed("rfc3686(ctr(aes))", DECRYPT, sec, NULL, 0, speed_template_20_28_36, num_mb); break; case 601: test_mb_skcipher_speed("ecb(des3_ede)", ENCRYPT, sec, des3_speed_template, DES3_SPEED_VECTORS, speed_template_24, num_mb); test_mb_skcipher_speed("ecb(des3_ede)", DECRYPT, sec, des3_speed_template, DES3_SPEED_VECTORS, speed_template_24, num_mb); test_mb_skcipher_speed("cbc(des3_ede)", ENCRYPT, sec, des3_speed_template, DES3_SPEED_VECTORS, speed_template_24, num_mb); test_mb_skcipher_speed("cbc(des3_ede)", DECRYPT, sec, des3_speed_template, DES3_SPEED_VECTORS, speed_template_24, num_mb); break; case 602: test_mb_skcipher_speed("ecb(des)", ENCRYPT, sec, NULL, 0, speed_template_8, num_mb); test_mb_skcipher_speed("ecb(des)", DECRYPT, sec, NULL, 0, speed_template_8, num_mb); test_mb_skcipher_speed("cbc(des)", ENCRYPT, sec, NULL, 0, speed_template_8, num_mb); test_mb_skcipher_speed("cbc(des)", DECRYPT, sec, NULL, 0, speed_template_8, num_mb); break; case 603: test_mb_skcipher_speed("ecb(serpent)", ENCRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("ecb(serpent)", DECRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("cbc(serpent)", ENCRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("cbc(serpent)", DECRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("ctr(serpent)", ENCRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("ctr(serpent)", DECRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("lrw(serpent)", ENCRYPT, sec, NULL, 0, speed_template_32_48, num_mb); test_mb_skcipher_speed("lrw(serpent)", DECRYPT, sec, NULL, 0, speed_template_32_48, num_mb); test_mb_skcipher_speed("xts(serpent)", ENCRYPT, sec, NULL, 0, speed_template_32_64, num_mb); test_mb_skcipher_speed("xts(serpent)", DECRYPT, sec, NULL, 0, speed_template_32_64, num_mb); break; case 604: test_mb_skcipher_speed("ecb(twofish)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32, num_mb); test_mb_skcipher_speed("ecb(twofish)", DECRYPT, sec, NULL, 0, speed_template_16_24_32, num_mb); test_mb_skcipher_speed("cbc(twofish)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32, num_mb); test_mb_skcipher_speed("cbc(twofish)", DECRYPT, sec, NULL, 0, speed_template_16_24_32, num_mb); test_mb_skcipher_speed("ctr(twofish)", ENCRYPT, sec, NULL, 0, speed_template_16_24_32, num_mb); test_mb_skcipher_speed("ctr(twofish)", DECRYPT, sec, NULL, 0, speed_template_16_24_32, num_mb); test_mb_skcipher_speed("lrw(twofish)", ENCRYPT, sec, NULL, 0, speed_template_32_40_48, num_mb); test_mb_skcipher_speed("lrw(twofish)", DECRYPT, sec, NULL, 0, speed_template_32_40_48, num_mb); test_mb_skcipher_speed("xts(twofish)", ENCRYPT, sec, NULL, 0, speed_template_32_48_64, num_mb); test_mb_skcipher_speed("xts(twofish)", DECRYPT, sec, NULL, 0, speed_template_32_48_64, num_mb); break; case 605: test_mb_skcipher_speed("ecb(arc4)", ENCRYPT, sec, NULL, 0, speed_template_8, num_mb); break; case 606: test_mb_skcipher_speed("ecb(cast5)", ENCRYPT, sec, NULL, 0, speed_template_8_16, num_mb); test_mb_skcipher_speed("ecb(cast5)", DECRYPT, sec, NULL, 0, speed_template_8_16, num_mb); test_mb_skcipher_speed("cbc(cast5)", ENCRYPT, sec, NULL, 0, speed_template_8_16, num_mb); test_mb_skcipher_speed("cbc(cast5)", DECRYPT, sec, NULL, 0, speed_template_8_16, num_mb); test_mb_skcipher_speed("ctr(cast5)", ENCRYPT, sec, NULL, 0, speed_template_8_16, num_mb); test_mb_skcipher_speed("ctr(cast5)", DECRYPT, sec, NULL, 0, speed_template_8_16, num_mb); break; case 607: test_mb_skcipher_speed("ecb(cast6)", ENCRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("ecb(cast6)", DECRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("cbc(cast6)", ENCRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("cbc(cast6)", DECRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("ctr(cast6)", ENCRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("ctr(cast6)", DECRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("lrw(cast6)", ENCRYPT, sec, NULL, 0, speed_template_32_48, num_mb); test_mb_skcipher_speed("lrw(cast6)", DECRYPT, sec, NULL, 0, speed_template_32_48, num_mb); test_mb_skcipher_speed("xts(cast6)", ENCRYPT, sec, NULL, 0, speed_template_32_64, num_mb); test_mb_skcipher_speed("xts(cast6)", DECRYPT, sec, NULL, 0, speed_template_32_64, num_mb); break; case 608: test_mb_skcipher_speed("ecb(camellia)", ENCRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("ecb(camellia)", DECRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("cbc(camellia)", ENCRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("cbc(camellia)", DECRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("ctr(camellia)", ENCRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("ctr(camellia)", DECRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("lrw(camellia)", ENCRYPT, sec, NULL, 0, speed_template_32_48, num_mb); test_mb_skcipher_speed("lrw(camellia)", DECRYPT, sec, NULL, 0, speed_template_32_48, num_mb); test_mb_skcipher_speed("xts(camellia)", ENCRYPT, sec, NULL, 0, speed_template_32_64, num_mb); test_mb_skcipher_speed("xts(camellia)", DECRYPT, sec, NULL, 0, speed_template_32_64, num_mb); break; case 609: test_mb_skcipher_speed("ecb(blowfish)", ENCRYPT, sec, NULL, 0, speed_template_8_32, num_mb); test_mb_skcipher_speed("ecb(blowfish)", DECRYPT, sec, NULL, 0, speed_template_8_32, num_mb); test_mb_skcipher_speed("cbc(blowfish)", ENCRYPT, sec, NULL, 0, speed_template_8_32, num_mb); test_mb_skcipher_speed("cbc(blowfish)", DECRYPT, sec, NULL, 0, speed_template_8_32, num_mb); test_mb_skcipher_speed("ctr(blowfish)", ENCRYPT, sec, NULL, 0, speed_template_8_32, num_mb); test_mb_skcipher_speed("ctr(blowfish)", DECRYPT, sec, NULL, 0, speed_template_8_32, num_mb); break; case 610: test_mb_skcipher_speed("ecb(aria)", ENCRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("ecb(aria)", DECRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("ctr(aria)", ENCRYPT, sec, NULL, 0, speed_template_16_32, num_mb); test_mb_skcipher_speed("ctr(aria)", DECRYPT, sec, NULL, 0, speed_template_16_32, num_mb); break; } return ret; } static int __init tcrypt_mod_init(void) { int err = -ENOMEM; int i; for (i = 0; i < TVMEMSIZE; i++) { tvmem[i] = (void *)__get_free_page(GFP_KERNEL); if (!tvmem[i]) goto err_free_tv; } err = do_test(alg, type, mask, mode, num_mb); if (err) { pr_err("one or more tests failed!\n"); goto err_free_tv; } else { pr_debug("all tests passed\n"); } /* We intentionaly return -EAGAIN to prevent keeping the module, * unless we're running in fips mode. It does all its work from * init() and doesn't offer any runtime functionality, but in * the fips case, checking for a successful load is helpful. * => we don't need it in the memory, do we? * -- mludvig */ if (!fips_enabled) err = -EAGAIN; err_free_tv: for (i = 0; i < TVMEMSIZE && tvmem[i]; i++) free_page((unsigned long)tvmem[i]); return err; } /* * If an init function is provided, an exit function must also be provided * to allow module unload. */ static void __exit tcrypt_mod_fini(void) { } late_initcall(tcrypt_mod_init); module_exit(tcrypt_mod_fini); module_param(alg, charp, 0); module_param(type, uint, 0); module_param(mask, uint, 0); module_param(mode, int, 0); module_param(sec, uint, 0); MODULE_PARM_DESC(sec, "Length in seconds of speed tests " "(defaults to zero which uses CPU cycles instead)"); module_param(num_mb, uint, 0000); MODULE_PARM_DESC(num_mb, "Number of concurrent requests to be used in mb speed tests (defaults to 8)"); module_param(klen, uint, 0); MODULE_PARM_DESC(klen, "Key length (defaults to 0)"); MODULE_LICENSE("GPL"); MODULE_DESCRIPTION("Quick & dirty crypto testing module"); MODULE_AUTHOR("James Morris <jmorris@intercode.com.au>");
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1