|
@@ -67,6 +67,7 @@ static char *alg = NULL;
|
|
|
static u32 type;
|
|
|
static u32 mask;
|
|
|
static int mode;
|
|
|
+static u32 num_mb = 8;
|
|
|
static char *tvmem[TVMEMSIZE];
|
|
|
|
|
|
static char *check[] = {
|
|
@@ -79,6 +80,66 @@ static char *check[] = {
|
|
|
NULL
|
|
|
};
|
|
|
|
|
|
+static u32 block_sizes[] = { 16, 64, 256, 1024, 8192, 0 };
|
|
|
+static u32 aead_sizes[] = { 16, 64, 256, 512, 1024, 2048, 4096, 8192, 0 };
|
|
|
+
|
|
|
+#define XBUFSIZE 8
|
|
|
+#define MAX_IVLEN 32
|
|
|
+
|
|
|
+static int testmgr_alloc_buf(char *buf[XBUFSIZE])
|
|
|
+{
|
|
|
+ int i;
|
|
|
+
|
|
|
+ for (i = 0; i < XBUFSIZE; i++) {
|
|
|
+ buf[i] = (void *)__get_free_page(GFP_KERNEL);
|
|
|
+ if (!buf[i])
|
|
|
+ goto err_free_buf;
|
|
|
+ }
|
|
|
+
|
|
|
+ return 0;
|
|
|
+
|
|
|
+err_free_buf:
|
|
|
+ while (i-- > 0)
|
|
|
+ free_page((unsigned long)buf[i]);
|
|
|
+
|
|
|
+ return -ENOMEM;
|
|
|
+}
|
|
|
+
|
|
|
+static void testmgr_free_buf(char *buf[XBUFSIZE])
|
|
|
+{
|
|
|
+ int i;
|
|
|
+
|
|
|
+ for (i = 0; i < XBUFSIZE; i++)
|
|
|
+ free_page((unsigned long)buf[i]);
|
|
|
+}
|
|
|
+
|
|
|
+static void sg_init_aead(struct scatterlist *sg, char *xbuf[XBUFSIZE],
|
|
|
+ unsigned int buflen, const void *assoc,
|
|
|
+ unsigned int aad_size)
|
|
|
+{
|
|
|
+ int np = (buflen + PAGE_SIZE - 1)/PAGE_SIZE;
|
|
|
+ int k, rem;
|
|
|
+
|
|
|
+ if (np > XBUFSIZE) {
|
|
|
+ rem = PAGE_SIZE;
|
|
|
+ np = XBUFSIZE;
|
|
|
+ } else {
|
|
|
+ rem = buflen % PAGE_SIZE;
|
|
|
+ }
|
|
|
+
|
|
|
+ sg_init_table(sg, np + 1);
|
|
|
+
|
|
|
+ sg_set_buf(&sg[0], assoc, aad_size);
|
|
|
+
|
|
|
+ if (rem)
|
|
|
+ np--;
|
|
|
+ for (k = 0; k < np; k++)
|
|
|
+ sg_set_buf(&sg[k + 1], xbuf[k], PAGE_SIZE);
|
|
|
+
|
|
|
+ if (rem)
|
|
|
+ sg_set_buf(&sg[k + 1], xbuf[k], rem);
|
|
|
+}
|
|
|
+
|
|
|
static inline int do_one_aead_op(struct aead_request *req, int ret)
|
|
|
{
|
|
|
struct crypto_wait *wait = req->base.data;
|
|
@@ -86,8 +147,44 @@ static inline int do_one_aead_op(struct aead_request *req, int ret)
|
|
|
return crypto_wait_req(ret, wait);
|
|
|
}
|
|
|
|
|
|
-static int test_aead_jiffies(struct aead_request *req, int enc,
|
|
|
- int blen, int secs)
|
|
|
+struct test_mb_aead_data {
|
|
|
+ struct scatterlist sg[XBUFSIZE];
|
|
|
+ struct scatterlist sgout[XBUFSIZE];
|
|
|
+ struct aead_request *req;
|
|
|
+ struct crypto_wait wait;
|
|
|
+ char *xbuf[XBUFSIZE];
|
|
|
+ char *xoutbuf[XBUFSIZE];
|
|
|
+ char *axbuf[XBUFSIZE];
|
|
|
+};
|
|
|
+
|
|
|
+static int do_mult_aead_op(struct test_mb_aead_data *data, int enc,
|
|
|
+ u32 num_mb)
|
|
|
+{
|
|
|
+ int i, rc[num_mb], err = 0;
|
|
|
+
|
|
|
+ /* Fire up a bunch of concurrent requests */
|
|
|
+ for (i = 0; i < num_mb; i++) {
|
|
|
+ if (enc == ENCRYPT)
|
|
|
+ rc[i] = crypto_aead_encrypt(data[i].req);
|
|
|
+ else
|
|
|
+ rc[i] = crypto_aead_decrypt(data[i].req);
|
|
|
+ }
|
|
|
+
|
|
|
+ /* Wait for all requests to finish */
|
|
|
+ for (i = 0; i < num_mb; i++) {
|
|
|
+ rc[i] = crypto_wait_req(rc[i], &data[i].wait);
|
|
|
+
|
|
|
+ if (rc[i]) {
|
|
|
+ pr_info("concurrent request %d error %d\n", i, rc[i]);
|
|
|
+ err = rc[i];
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ return err;
|
|
|
+}
|
|
|
+
|
|
|
+static int test_mb_aead_jiffies(struct test_mb_aead_data *data, int enc,
|
|
|
+ int blen, int secs, u32 num_mb)
|
|
|
{
|
|
|
unsigned long start, end;
|
|
|
int bcount;
|
|
@@ -95,21 +192,18 @@ static int test_aead_jiffies(struct aead_request *req, int enc,
|
|
|
|
|
|
for (start = jiffies, end = start + secs * HZ, bcount = 0;
|
|
|
time_before(jiffies, end); bcount++) {
|
|
|
- if (enc)
|
|
|
- ret = do_one_aead_op(req, crypto_aead_encrypt(req));
|
|
|
- else
|
|
|
- ret = do_one_aead_op(req, crypto_aead_decrypt(req));
|
|
|
-
|
|
|
+ ret = do_mult_aead_op(data, enc, num_mb);
|
|
|
if (ret)
|
|
|
return ret;
|
|
|
}
|
|
|
|
|
|
- printk("%d operations in %d seconds (%ld bytes)\n",
|
|
|
- bcount, secs, (long)bcount * blen);
|
|
|
+ pr_cont("%d operations in %d seconds (%ld bytes)\n",
|
|
|
+ bcount * num_mb, secs, (long)bcount * blen * num_mb);
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
-static int test_aead_cycles(struct aead_request *req, int enc, int blen)
|
|
|
+static int test_mb_aead_cycles(struct test_mb_aead_data *data, int enc,
|
|
|
+ int blen, u32 num_mb)
|
|
|
{
|
|
|
unsigned long cycles = 0;
|
|
|
int ret = 0;
|
|
@@ -117,11 +211,7 @@ static int test_aead_cycles(struct aead_request *req, int enc, int blen)
|
|
|
|
|
|
/* Warm-up run. */
|
|
|
for (i = 0; i < 4; i++) {
|
|
|
- if (enc)
|
|
|
- ret = do_one_aead_op(req, crypto_aead_encrypt(req));
|
|
|
- else
|
|
|
- ret = do_one_aead_op(req, crypto_aead_decrypt(req));
|
|
|
-
|
|
|
+ ret = do_mult_aead_op(data, enc, num_mb);
|
|
|
if (ret)
|
|
|
goto out;
|
|
|
}
|
|
@@ -131,10 +221,7 @@ static int test_aead_cycles(struct aead_request *req, int enc, int blen)
|
|
|
cycles_t start, end;
|
|
|
|
|
|
start = get_cycles();
|
|
|
- if (enc)
|
|
|
- ret = do_one_aead_op(req, crypto_aead_encrypt(req));
|
|
|
- else
|
|
|
- ret = do_one_aead_op(req, crypto_aead_decrypt(req));
|
|
|
+ ret = do_mult_aead_op(data, enc, num_mb);
|
|
|
end = get_cycles();
|
|
|
|
|
|
if (ret)
|
|
@@ -145,64 +232,276 @@ static int test_aead_cycles(struct aead_request *req, int enc, int blen)
|
|
|
|
|
|
out:
|
|
|
if (ret == 0)
|
|
|
- printk("1 operation in %lu cycles (%d bytes)\n",
|
|
|
- (cycles + 4) / 8, blen);
|
|
|
+ pr_cont("1 operation in %lu cycles (%d bytes)\n",
|
|
|
+ (cycles + 4) / (8 * num_mb), blen);
|
|
|
|
|
|
return ret;
|
|
|
}
|
|
|
|
|
|
-static u32 block_sizes[] = { 16, 64, 256, 1024, 8192, 0 };
|
|
|
-static u32 aead_sizes[] = { 16, 64, 256, 512, 1024, 2048, 4096, 8192, 0 };
|
|
|
+static void test_mb_aead_speed(const char *algo, int enc, int secs,
|
|
|
+ struct aead_speed_template *template,
|
|
|
+ unsigned int tcount, u8 authsize,
|
|
|
+ unsigned int aad_size, u8 *keysize, u32 num_mb)
|
|
|
+{
|
|
|
+ struct test_mb_aead_data *data;
|
|
|
+ struct crypto_aead *tfm;
|
|
|
+ unsigned int i, j, iv_len;
|
|
|
+ const char *key;
|
|
|
+ const char *e;
|
|
|
+ void *assoc;
|
|
|
+ u32 *b_size;
|
|
|
+ char *iv;
|
|
|
+ int ret;
|
|
|
|
|
|
-#define XBUFSIZE 8
|
|
|
-#define MAX_IVLEN 32
|
|
|
|
|
|
-static int testmgr_alloc_buf(char *buf[XBUFSIZE])
|
|
|
-{
|
|
|
- int i;
|
|
|
+ if (aad_size >= PAGE_SIZE) {
|
|
|
+ pr_err("associate data length (%u) too big\n", aad_size);
|
|
|
+ return;
|
|
|
+ }
|
|
|
|
|
|
- for (i = 0; i < XBUFSIZE; i++) {
|
|
|
- buf[i] = (void *)__get_free_page(GFP_KERNEL);
|
|
|
- if (!buf[i])
|
|
|
- goto err_free_buf;
|
|
|
+ iv = kzalloc(MAX_IVLEN, GFP_KERNEL);
|
|
|
+ if (!iv)
|
|
|
+ return;
|
|
|
+
|
|
|
+ if (enc == ENCRYPT)
|
|
|
+ e = "encryption";
|
|
|
+ else
|
|
|
+ e = "decryption";
|
|
|
+
|
|
|
+ data = kcalloc(num_mb, sizeof(*data), GFP_KERNEL);
|
|
|
+ if (!data)
|
|
|
+ goto out_free_iv;
|
|
|
+
|
|
|
+ tfm = crypto_alloc_aead(algo, 0, 0);
|
|
|
+ if (IS_ERR(tfm)) {
|
|
|
+ pr_err("failed to load transform for %s: %ld\n",
|
|
|
+ algo, PTR_ERR(tfm));
|
|
|
+ goto out_free_data;
|
|
|
}
|
|
|
|
|
|
- return 0;
|
|
|
+ ret = crypto_aead_setauthsize(tfm, authsize);
|
|
|
|
|
|
-err_free_buf:
|
|
|
- while (i-- > 0)
|
|
|
- free_page((unsigned long)buf[i]);
|
|
|
+ for (i = 0; i < num_mb; ++i)
|
|
|
+ if (testmgr_alloc_buf(data[i].xbuf)) {
|
|
|
+ while (i--)
|
|
|
+ testmgr_free_buf(data[i].xbuf);
|
|
|
+ goto out_free_tfm;
|
|
|
+ }
|
|
|
|
|
|
- return -ENOMEM;
|
|
|
+ for (i = 0; i < num_mb; ++i)
|
|
|
+ if (testmgr_alloc_buf(data[i].axbuf)) {
|
|
|
+ while (i--)
|
|
|
+ testmgr_free_buf(data[i].axbuf);
|
|
|
+ goto out_free_xbuf;
|
|
|
+ }
|
|
|
+
|
|
|
+ for (i = 0; i < num_mb; ++i)
|
|
|
+ if (testmgr_alloc_buf(data[i].xoutbuf)) {
|
|
|
+ while (i--)
|
|
|
+ testmgr_free_buf(data[i].xoutbuf);
|
|
|
+ goto out_free_axbuf;
|
|
|
+ }
|
|
|
+
|
|
|
+ for (i = 0; i < num_mb; ++i) {
|
|
|
+ data[i].req = aead_request_alloc(tfm, GFP_KERNEL);
|
|
|
+ if (!data[i].req) {
|
|
|
+ pr_err("alg: skcipher: Failed to allocate request for %s\n",
|
|
|
+ algo);
|
|
|
+ while (i--)
|
|
|
+ aead_request_free(data[i].req);
|
|
|
+ goto out_free_xoutbuf;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ for (i = 0; i < num_mb; ++i) {
|
|
|
+ crypto_init_wait(&data[i].wait);
|
|
|
+ aead_request_set_callback(data[i].req,
|
|
|
+ CRYPTO_TFM_REQ_MAY_BACKLOG,
|
|
|
+ crypto_req_done, &data[i].wait);
|
|
|
+ }
|
|
|
+
|
|
|
+ pr_info("\ntesting speed of multibuffer %s (%s) %s\n", algo,
|
|
|
+ get_driver_name(crypto_aead, tfm), e);
|
|
|
+
|
|
|
+ i = 0;
|
|
|
+ do {
|
|
|
+ b_size = aead_sizes;
|
|
|
+ do {
|
|
|
+ if (*b_size + authsize > XBUFSIZE * PAGE_SIZE) {
|
|
|
+ pr_err("template (%u) too big for buffer (%lu)\n",
|
|
|
+ authsize + *b_size,
|
|
|
+ XBUFSIZE * PAGE_SIZE);
|
|
|
+ goto out;
|
|
|
+ }
|
|
|
+
|
|
|
+ pr_info("test %u (%d bit key, %d byte blocks): ", i,
|
|
|
+ *keysize * 8, *b_size);
|
|
|
+
|
|
|
+ /* Set up tfm global state, i.e. the key */
|
|
|
+
|
|
|
+ memset(tvmem[0], 0xff, PAGE_SIZE);
|
|
|
+ key = tvmem[0];
|
|
|
+ for (j = 0; j < tcount; j++) {
|
|
|
+ if (template[j].klen == *keysize) {
|
|
|
+ key = template[j].key;
|
|
|
+ break;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ crypto_aead_clear_flags(tfm, ~0);
|
|
|
+
|
|
|
+ ret = crypto_aead_setkey(tfm, key, *keysize);
|
|
|
+ if (ret) {
|
|
|
+ pr_err("setkey() failed flags=%x\n",
|
|
|
+ crypto_aead_get_flags(tfm));
|
|
|
+ goto out;
|
|
|
+ }
|
|
|
+
|
|
|
+ iv_len = crypto_aead_ivsize(tfm);
|
|
|
+ if (iv_len)
|
|
|
+ memset(iv, 0xff, iv_len);
|
|
|
+
|
|
|
+ /* Now setup per request stuff, i.e. buffers */
|
|
|
+
|
|
|
+ for (j = 0; j < num_mb; ++j) {
|
|
|
+ struct test_mb_aead_data *cur = &data[j];
|
|
|
+
|
|
|
+ assoc = cur->axbuf[0];
|
|
|
+ memset(assoc, 0xff, aad_size);
|
|
|
+
|
|
|
+ sg_init_aead(cur->sg, cur->xbuf,
|
|
|
+ *b_size + (enc ? 0 : authsize),
|
|
|
+ assoc, aad_size);
|
|
|
+
|
|
|
+ sg_init_aead(cur->sgout, cur->xoutbuf,
|
|
|
+ *b_size + (enc ? authsize : 0),
|
|
|
+ assoc, aad_size);
|
|
|
+
|
|
|
+ aead_request_set_ad(cur->req, aad_size);
|
|
|
+
|
|
|
+ if (!enc) {
|
|
|
+
|
|
|
+ aead_request_set_crypt(cur->req,
|
|
|
+ cur->sgout,
|
|
|
+ cur->sg,
|
|
|
+ *b_size, iv);
|
|
|
+ ret = crypto_aead_encrypt(cur->req);
|
|
|
+ ret = do_one_aead_op(cur->req, ret);
|
|
|
+
|
|
|
+ if (ret) {
|
|
|
+ pr_err("calculating auth failed failed (%d)\n",
|
|
|
+ ret);
|
|
|
+ break;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ aead_request_set_crypt(cur->req, cur->sg,
|
|
|
+ cur->sgout, *b_size +
|
|
|
+ (enc ? 0 : authsize),
|
|
|
+ iv);
|
|
|
+
|
|
|
+ }
|
|
|
+
|
|
|
+ if (secs)
|
|
|
+ ret = test_mb_aead_jiffies(data, enc, *b_size,
|
|
|
+ secs, num_mb);
|
|
|
+ else
|
|
|
+ ret = test_mb_aead_cycles(data, enc, *b_size,
|
|
|
+ num_mb);
|
|
|
+
|
|
|
+ if (ret) {
|
|
|
+ pr_err("%s() failed return code=%d\n", e, ret);
|
|
|
+ break;
|
|
|
+ }
|
|
|
+ b_size++;
|
|
|
+ i++;
|
|
|
+ } while (*b_size);
|
|
|
+ keysize++;
|
|
|
+ } while (*keysize);
|
|
|
+
|
|
|
+out:
|
|
|
+ for (i = 0; i < num_mb; ++i)
|
|
|
+ aead_request_free(data[i].req);
|
|
|
+out_free_xoutbuf:
|
|
|
+ for (i = 0; i < num_mb; ++i)
|
|
|
+ testmgr_free_buf(data[i].xoutbuf);
|
|
|
+out_free_axbuf:
|
|
|
+ for (i = 0; i < num_mb; ++i)
|
|
|
+ testmgr_free_buf(data[i].axbuf);
|
|
|
+out_free_xbuf:
|
|
|
+ for (i = 0; i < num_mb; ++i)
|
|
|
+ testmgr_free_buf(data[i].xbuf);
|
|
|
+out_free_tfm:
|
|
|
+ crypto_free_aead(tfm);
|
|
|
+out_free_data:
|
|
|
+ kfree(data);
|
|
|
+out_free_iv:
|
|
|
+ kfree(iv);
|
|
|
}
|
|
|
|
|
|
-static void testmgr_free_buf(char *buf[XBUFSIZE])
|
|
|
+static int test_aead_jiffies(struct aead_request *req, int enc,
|
|
|
+ int blen, int secs)
|
|
|
{
|
|
|
- int i;
|
|
|
+ unsigned long start, end;
|
|
|
+ int bcount;
|
|
|
+ int ret;
|
|
|
|
|
|
- for (i = 0; i < XBUFSIZE; i++)
|
|
|
- free_page((unsigned long)buf[i]);
|
|
|
+ for (start = jiffies, end = start + secs * HZ, bcount = 0;
|
|
|
+ time_before(jiffies, end); bcount++) {
|
|
|
+ if (enc)
|
|
|
+ ret = do_one_aead_op(req, crypto_aead_encrypt(req));
|
|
|
+ else
|
|
|
+ ret = do_one_aead_op(req, crypto_aead_decrypt(req));
|
|
|
+
|
|
|
+ if (ret)
|
|
|
+ return ret;
|
|
|
+ }
|
|
|
+
|
|
|
+ printk("%d operations in %d seconds (%ld bytes)\n",
|
|
|
+ bcount, secs, (long)bcount * blen);
|
|
|
+ return 0;
|
|
|
}
|
|
|
|
|
|
-static void sg_init_aead(struct scatterlist *sg, char *xbuf[XBUFSIZE],
|
|
|
- unsigned int buflen)
|
|
|
+static int test_aead_cycles(struct aead_request *req, int enc, int blen)
|
|
|
{
|
|
|
- int np = (buflen + PAGE_SIZE - 1)/PAGE_SIZE;
|
|
|
- int k, rem;
|
|
|
+ unsigned long cycles = 0;
|
|
|
+ int ret = 0;
|
|
|
+ int i;
|
|
|
|
|
|
- if (np > XBUFSIZE) {
|
|
|
- rem = PAGE_SIZE;
|
|
|
- np = XBUFSIZE;
|
|
|
- } else {
|
|
|
- rem = buflen % PAGE_SIZE;
|
|
|
+ /* Warm-up run. */
|
|
|
+ for (i = 0; i < 4; i++) {
|
|
|
+ if (enc)
|
|
|
+ ret = do_one_aead_op(req, crypto_aead_encrypt(req));
|
|
|
+ else
|
|
|
+ ret = do_one_aead_op(req, crypto_aead_decrypt(req));
|
|
|
+
|
|
|
+ if (ret)
|
|
|
+ goto out;
|
|
|
}
|
|
|
|
|
|
- sg_init_table(sg, np + 1);
|
|
|
- np--;
|
|
|
- for (k = 0; k < np; k++)
|
|
|
- sg_set_buf(&sg[k + 1], xbuf[k], PAGE_SIZE);
|
|
|
+ /* The real thing. */
|
|
|
+ for (i = 0; i < 8; i++) {
|
|
|
+ cycles_t start, end;
|
|
|
+
|
|
|
+ start = get_cycles();
|
|
|
+ if (enc)
|
|
|
+ ret = do_one_aead_op(req, crypto_aead_encrypt(req));
|
|
|
+ else
|
|
|
+ ret = do_one_aead_op(req, crypto_aead_decrypt(req));
|
|
|
+ end = get_cycles();
|
|
|
+
|
|
|
+ if (ret)
|
|
|
+ goto out;
|
|
|
+
|
|
|
+ cycles += end - start;
|
|
|
+ }
|
|
|
+
|
|
|
+out:
|
|
|
+ if (ret == 0)
|
|
|
+ printk("1 operation in %lu cycles (%d bytes)\n",
|
|
|
+ (cycles + 4) / 8, blen);
|
|
|
|
|
|
- sg_set_buf(&sg[k + 1], xbuf[k], rem);
|
|
|
+ return ret;
|
|
|
}
|
|
|
|
|
|
static void test_aead_speed(const char *algo, int enc, unsigned int secs,
|
|
@@ -316,19 +615,37 @@ static void test_aead_speed(const char *algo, int enc, unsigned int secs,
|
|
|
goto out;
|
|
|
}
|
|
|
|
|
|
- sg_init_aead(sg, xbuf,
|
|
|
- *b_size + (enc ? 0 : authsize));
|
|
|
+ sg_init_aead(sg, xbuf, *b_size + (enc ? 0 : authsize),
|
|
|
+ assoc, aad_size);
|
|
|
|
|
|
sg_init_aead(sgout, xoutbuf,
|
|
|
- *b_size + (enc ? authsize : 0));
|
|
|
+ *b_size + (enc ? authsize : 0), assoc,
|
|
|
+ aad_size);
|
|
|
+
|
|
|
+ aead_request_set_ad(req, aad_size);
|
|
|
|
|
|
- sg_set_buf(&sg[0], assoc, aad_size);
|
|
|
- sg_set_buf(&sgout[0], assoc, aad_size);
|
|
|
+ if (!enc) {
|
|
|
+
|
|
|
+ /*
|
|
|
+ * For decryption we need a proper auth so
|
|
|
+ * we do the encryption path once with buffers
|
|
|
+ * reversed (input <-> output) to calculate it
|
|
|
+ */
|
|
|
+ aead_request_set_crypt(req, sgout, sg,
|
|
|
+ *b_size, iv);
|
|
|
+ ret = do_one_aead_op(req,
|
|
|
+ crypto_aead_encrypt(req));
|
|
|
+
|
|
|
+ if (ret) {
|
|
|
+ pr_err("calculating auth failed failed (%d)\n",
|
|
|
+ ret);
|
|
|
+ break;
|
|
|
+ }
|
|
|
+ }
|
|
|
|
|
|
aead_request_set_crypt(req, sg, sgout,
|
|
|
*b_size + (enc ? 0 : authsize),
|
|
|
iv);
|
|
|
- aead_request_set_ad(req, aad_size);
|
|
|
|
|
|
if (secs)
|
|
|
ret = test_aead_jiffies(req, enc, *b_size,
|
|
@@ -381,24 +698,98 @@ static inline int do_one_ahash_op(struct ahash_request *req, int ret)
|
|
|
}
|
|
|
|
|
|
struct test_mb_ahash_data {
|
|
|
- struct scatterlist sg[TVMEMSIZE];
|
|
|
+ struct scatterlist sg[XBUFSIZE];
|
|
|
char result[64];
|
|
|
struct ahash_request *req;
|
|
|
struct crypto_wait wait;
|
|
|
char *xbuf[XBUFSIZE];
|
|
|
};
|
|
|
|
|
|
-static void test_mb_ahash_speed(const char *algo, unsigned int sec,
|
|
|
- struct hash_speed *speed)
|
|
|
+static inline int do_mult_ahash_op(struct test_mb_ahash_data *data, u32 num_mb)
|
|
|
+{
|
|
|
+ int i, rc[num_mb], err = 0;
|
|
|
+
|
|
|
+ /* Fire up a bunch of concurrent requests */
|
|
|
+ for (i = 0; i < num_mb; i++)
|
|
|
+ rc[i] = crypto_ahash_digest(data[i].req);
|
|
|
+
|
|
|
+ /* Wait for all requests to finish */
|
|
|
+ for (i = 0; i < num_mb; i++) {
|
|
|
+ rc[i] = crypto_wait_req(rc[i], &data[i].wait);
|
|
|
+
|
|
|
+ if (rc[i]) {
|
|
|
+ pr_info("concurrent request %d error %d\n", i, rc[i]);
|
|
|
+ err = rc[i];
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ return err;
|
|
|
+}
|
|
|
+
|
|
|
+static int test_mb_ahash_jiffies(struct test_mb_ahash_data *data, int blen,
|
|
|
+ int secs, u32 num_mb)
|
|
|
+{
|
|
|
+ unsigned long start, end;
|
|
|
+ int bcount;
|
|
|
+ int ret;
|
|
|
+
|
|
|
+ for (start = jiffies, end = start + secs * HZ, bcount = 0;
|
|
|
+ time_before(jiffies, end); bcount++) {
|
|
|
+ ret = do_mult_ahash_op(data, num_mb);
|
|
|
+ if (ret)
|
|
|
+ return ret;
|
|
|
+ }
|
|
|
+
|
|
|
+ pr_cont("%d operations in %d seconds (%ld bytes)\n",
|
|
|
+ bcount * num_mb, secs, (long)bcount * blen * num_mb);
|
|
|
+ return 0;
|
|
|
+}
|
|
|
+
|
|
|
+static int test_mb_ahash_cycles(struct test_mb_ahash_data *data, int blen,
|
|
|
+ u32 num_mb)
|
|
|
+{
|
|
|
+ unsigned long cycles = 0;
|
|
|
+ int ret = 0;
|
|
|
+ int i;
|
|
|
+
|
|
|
+ /* Warm-up run. */
|
|
|
+ for (i = 0; i < 4; i++) {
|
|
|
+ ret = do_mult_ahash_op(data, num_mb);
|
|
|
+ if (ret)
|
|
|
+ goto out;
|
|
|
+ }
|
|
|
+
|
|
|
+ /* The real thing. */
|
|
|
+ for (i = 0; i < 8; i++) {
|
|
|
+ cycles_t start, end;
|
|
|
+
|
|
|
+ start = get_cycles();
|
|
|
+ ret = do_mult_ahash_op(data, num_mb);
|
|
|
+ end = get_cycles();
|
|
|
+
|
|
|
+ if (ret)
|
|
|
+ goto out;
|
|
|
+
|
|
|
+ cycles += end - start;
|
|
|
+ }
|
|
|
+
|
|
|
+out:
|
|
|
+ if (ret == 0)
|
|
|
+ pr_cont("1 operation in %lu cycles (%d bytes)\n",
|
|
|
+ (cycles + 4) / (8 * num_mb), blen);
|
|
|
+
|
|
|
+ return ret;
|
|
|
+}
|
|
|
+
|
|
|
+static void test_mb_ahash_speed(const char *algo, unsigned int secs,
|
|
|
+ struct hash_speed *speed, u32 num_mb)
|
|
|
{
|
|
|
struct test_mb_ahash_data *data;
|
|
|
struct crypto_ahash *tfm;
|
|
|
- unsigned long start, end;
|
|
|
- unsigned long cycles;
|
|
|
unsigned int i, j, k;
|
|
|
int ret;
|
|
|
|
|
|
- data = kzalloc(sizeof(*data) * 8, GFP_KERNEL);
|
|
|
+ data = kcalloc(num_mb, sizeof(*data), GFP_KERNEL);
|
|
|
if (!data)
|
|
|
return;
|
|
|
|
|
@@ -409,7 +800,7 @@ static void test_mb_ahash_speed(const char *algo, unsigned int sec,
|
|
|
goto free_data;
|
|
|
}
|
|
|
|
|
|
- for (i = 0; i < 8; ++i) {
|
|
|
+ for (i = 0; i < num_mb; ++i) {
|
|
|
if (testmgr_alloc_buf(data[i].xbuf))
|
|
|
goto out;
|
|
|
|
|
@@ -424,7 +815,12 @@ static void test_mb_ahash_speed(const char *algo, unsigned int sec,
|
|
|
|
|
|
ahash_request_set_callback(data[i].req, 0, crypto_req_done,
|
|
|
&data[i].wait);
|
|
|
- test_hash_sg_init(data[i].sg);
|
|
|
+
|
|
|
+ sg_init_table(data[i].sg, XBUFSIZE);
|
|
|
+ for (j = 0; j < XBUFSIZE; j++) {
|
|
|
+ sg_set_buf(data[i].sg + j, data[i].xbuf[j], PAGE_SIZE);
|
|
|
+ memset(data[i].xbuf[j], 0xff, PAGE_SIZE);
|
|
|
+ }
|
|
|
}
|
|
|
|
|
|
pr_info("\ntesting speed of multibuffer %s (%s)\n", algo,
|
|
@@ -435,16 +831,16 @@ static void test_mb_ahash_speed(const char *algo, unsigned int sec,
|
|
|
if (speed[i].blen != speed[i].plen)
|
|
|
continue;
|
|
|
|
|
|
- if (speed[i].blen > TVMEMSIZE * PAGE_SIZE) {
|
|
|
+ if (speed[i].blen > XBUFSIZE * PAGE_SIZE) {
|
|
|
pr_err("template (%u) too big for tvmem (%lu)\n",
|
|
|
- speed[i].blen, TVMEMSIZE * PAGE_SIZE);
|
|
|
+ speed[i].blen, XBUFSIZE * PAGE_SIZE);
|
|
|
goto out;
|
|
|
}
|
|
|
|
|
|
if (speed[i].klen)
|
|
|
crypto_ahash_setkey(tfm, tvmem[0], speed[i].klen);
|
|
|
|
|
|
- for (k = 0; k < 8; k++)
|
|
|
+ for (k = 0; k < num_mb; k++)
|
|
|
ahash_request_set_crypt(data[k].req, data[k].sg,
|
|
|
data[k].result, speed[i].blen);
|
|
|
|
|
@@ -453,34 +849,12 @@ static void test_mb_ahash_speed(const char *algo, unsigned int sec,
|
|
|
i, speed[i].blen, speed[i].plen,
|
|
|
speed[i].blen / speed[i].plen);
|
|
|
|
|
|
- start = get_cycles();
|
|
|
-
|
|
|
- for (k = 0; k < 8; k++) {
|
|
|
- ret = crypto_ahash_digest(data[k].req);
|
|
|
- if (ret == -EINPROGRESS) {
|
|
|
- ret = 0;
|
|
|
- continue;
|
|
|
- }
|
|
|
-
|
|
|
- if (ret)
|
|
|
- break;
|
|
|
-
|
|
|
- crypto_req_done(&data[k].req->base, 0);
|
|
|
- }
|
|
|
-
|
|
|
- for (j = 0; j < k; j++) {
|
|
|
- struct crypto_wait *wait = &data[j].wait;
|
|
|
- int wait_ret;
|
|
|
-
|
|
|
- wait_ret = crypto_wait_req(-EINPROGRESS, wait);
|
|
|
- if (wait_ret)
|
|
|
- ret = wait_ret;
|
|
|
- }
|
|
|
+ if (secs)
|
|
|
+ ret = test_mb_ahash_jiffies(data, speed[i].blen, secs,
|
|
|
+ num_mb);
|
|
|
+ else
|
|
|
+ ret = test_mb_ahash_cycles(data, speed[i].blen, num_mb);
|
|
|
|
|
|
- end = get_cycles();
|
|
|
- cycles = end - start;
|
|
|
- pr_cont("%6lu cycles/operation, %4lu cycles/byte\n",
|
|
|
- cycles, cycles / (8 * speed[i].blen));
|
|
|
|
|
|
if (ret) {
|
|
|
pr_err("At least one hashing failed ret=%d\n", ret);
|
|
@@ -489,10 +863,10 @@ static void test_mb_ahash_speed(const char *algo, unsigned int sec,
|
|
|
}
|
|
|
|
|
|
out:
|
|
|
- for (k = 0; k < 8; ++k)
|
|
|
+ for (k = 0; k < num_mb; ++k)
|
|
|
ahash_request_free(data[k].req);
|
|
|
|
|
|
- for (k = 0; k < 8; ++k)
|
|
|
+ for (k = 0; k < num_mb; ++k)
|
|
|
testmgr_free_buf(data[k].xbuf);
|
|
|
|
|
|
crypto_free_ahash(tfm);
|
|
@@ -736,6 +1110,254 @@ static void test_hash_speed(const char *algo, unsigned int secs,
|
|
|
return test_ahash_speed_common(algo, secs, speed, CRYPTO_ALG_ASYNC);
|
|
|
}
|
|
|
|
|
|
+struct test_mb_skcipher_data {
|
|
|
+ struct scatterlist sg[XBUFSIZE];
|
|
|
+ struct skcipher_request *req;
|
|
|
+ struct crypto_wait wait;
|
|
|
+ char *xbuf[XBUFSIZE];
|
|
|
+};
|
|
|
+
|
|
|
+static int do_mult_acipher_op(struct test_mb_skcipher_data *data, int enc,
|
|
|
+ u32 num_mb)
|
|
|
+{
|
|
|
+ int i, rc[num_mb], err = 0;
|
|
|
+
|
|
|
+ /* Fire up a bunch of concurrent requests */
|
|
|
+ for (i = 0; i < num_mb; i++) {
|
|
|
+ if (enc == ENCRYPT)
|
|
|
+ rc[i] = crypto_skcipher_encrypt(data[i].req);
|
|
|
+ else
|
|
|
+ rc[i] = crypto_skcipher_decrypt(data[i].req);
|
|
|
+ }
|
|
|
+
|
|
|
+ /* Wait for all requests to finish */
|
|
|
+ for (i = 0; i < num_mb; i++) {
|
|
|
+ rc[i] = crypto_wait_req(rc[i], &data[i].wait);
|
|
|
+
|
|
|
+ if (rc[i]) {
|
|
|
+ pr_info("concurrent request %d error %d\n", i, rc[i]);
|
|
|
+ err = rc[i];
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ return err;
|
|
|
+}
|
|
|
+
|
|
|
+static int test_mb_acipher_jiffies(struct test_mb_skcipher_data *data, int enc,
|
|
|
+ int blen, int secs, u32 num_mb)
|
|
|
+{
|
|
|
+ unsigned long start, end;
|
|
|
+ int bcount;
|
|
|
+ int ret;
|
|
|
+
|
|
|
+ for (start = jiffies, end = start + secs * HZ, bcount = 0;
|
|
|
+ time_before(jiffies, end); bcount++) {
|
|
|
+ ret = do_mult_acipher_op(data, enc, num_mb);
|
|
|
+ if (ret)
|
|
|
+ return ret;
|
|
|
+ }
|
|
|
+
|
|
|
+ pr_cont("%d operations in %d seconds (%ld bytes)\n",
|
|
|
+ bcount * num_mb, secs, (long)bcount * blen * num_mb);
|
|
|
+ return 0;
|
|
|
+}
|
|
|
+
|
|
|
+static int test_mb_acipher_cycles(struct test_mb_skcipher_data *data, int enc,
|
|
|
+ int blen, u32 num_mb)
|
|
|
+{
|
|
|
+ unsigned long cycles = 0;
|
|
|
+ int ret = 0;
|
|
|
+ int i;
|
|
|
+
|
|
|
+ /* Warm-up run. */
|
|
|
+ for (i = 0; i < 4; i++) {
|
|
|
+ ret = do_mult_acipher_op(data, enc, num_mb);
|
|
|
+ if (ret)
|
|
|
+ goto out;
|
|
|
+ }
|
|
|
+
|
|
|
+ /* The real thing. */
|
|
|
+ for (i = 0; i < 8; i++) {
|
|
|
+ cycles_t start, end;
|
|
|
+
|
|
|
+ start = get_cycles();
|
|
|
+ ret = do_mult_acipher_op(data, enc, num_mb);
|
|
|
+ end = get_cycles();
|
|
|
+
|
|
|
+ if (ret)
|
|
|
+ goto out;
|
|
|
+
|
|
|
+ cycles += end - start;
|
|
|
+ }
|
|
|
+
|
|
|
+out:
|
|
|
+ if (ret == 0)
|
|
|
+ pr_cont("1 operation in %lu cycles (%d bytes)\n",
|
|
|
+ (cycles + 4) / (8 * num_mb), blen);
|
|
|
+
|
|
|
+ return ret;
|
|
|
+}
|
|
|
+
|
|
|
+static void test_mb_skcipher_speed(const char *algo, int enc, int secs,
|
|
|
+ struct cipher_speed_template *template,
|
|
|
+ unsigned int tcount, u8 *keysize, u32 num_mb)
|
|
|
+{
|
|
|
+ struct test_mb_skcipher_data *data;
|
|
|
+ struct crypto_skcipher *tfm;
|
|
|
+ unsigned int i, j, iv_len;
|
|
|
+ const char *key;
|
|
|
+ const char *e;
|
|
|
+ u32 *b_size;
|
|
|
+ char iv[128];
|
|
|
+ int ret;
|
|
|
+
|
|
|
+ if (enc == ENCRYPT)
|
|
|
+ e = "encryption";
|
|
|
+ else
|
|
|
+ e = "decryption";
|
|
|
+
|
|
|
+ data = kcalloc(num_mb, sizeof(*data), GFP_KERNEL);
|
|
|
+ if (!data)
|
|
|
+ return;
|
|
|
+
|
|
|
+ tfm = crypto_alloc_skcipher(algo, 0, 0);
|
|
|
+ if (IS_ERR(tfm)) {
|
|
|
+ pr_err("failed to load transform for %s: %ld\n",
|
|
|
+ algo, PTR_ERR(tfm));
|
|
|
+ goto out_free_data;
|
|
|
+ }
|
|
|
+
|
|
|
+ for (i = 0; i < num_mb; ++i)
|
|
|
+ if (testmgr_alloc_buf(data[i].xbuf)) {
|
|
|
+ while (i--)
|
|
|
+ testmgr_free_buf(data[i].xbuf);
|
|
|
+ goto out_free_tfm;
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+ for (i = 0; i < num_mb; ++i)
|
|
|
+ if (testmgr_alloc_buf(data[i].xbuf)) {
|
|
|
+ while (i--)
|
|
|
+ testmgr_free_buf(data[i].xbuf);
|
|
|
+ goto out_free_tfm;
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+ for (i = 0; i < num_mb; ++i) {
|
|
|
+ data[i].req = skcipher_request_alloc(tfm, GFP_KERNEL);
|
|
|
+ if (!data[i].req) {
|
|
|
+ pr_err("alg: skcipher: Failed to allocate request for %s\n",
|
|
|
+ algo);
|
|
|
+ while (i--)
|
|
|
+ skcipher_request_free(data[i].req);
|
|
|
+ goto out_free_xbuf;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ for (i = 0; i < num_mb; ++i) {
|
|
|
+ skcipher_request_set_callback(data[i].req,
|
|
|
+ CRYPTO_TFM_REQ_MAY_BACKLOG,
|
|
|
+ crypto_req_done, &data[i].wait);
|
|
|
+ crypto_init_wait(&data[i].wait);
|
|
|
+ }
|
|
|
+
|
|
|
+ pr_info("\ntesting speed of multibuffer %s (%s) %s\n", algo,
|
|
|
+ get_driver_name(crypto_skcipher, tfm), e);
|
|
|
+
|
|
|
+ i = 0;
|
|
|
+ do {
|
|
|
+ b_size = block_sizes;
|
|
|
+ do {
|
|
|
+ if (*b_size > XBUFSIZE * PAGE_SIZE) {
|
|
|
+ pr_err("template (%u) too big for buffer (%lu)\n",
|
|
|
+ *b_size, XBUFSIZE * PAGE_SIZE);
|
|
|
+ goto out;
|
|
|
+ }
|
|
|
+
|
|
|
+ pr_info("test %u (%d bit key, %d byte blocks): ", i,
|
|
|
+ *keysize * 8, *b_size);
|
|
|
+
|
|
|
+ /* Set up tfm global state, i.e. the key */
|
|
|
+
|
|
|
+ memset(tvmem[0], 0xff, PAGE_SIZE);
|
|
|
+ key = tvmem[0];
|
|
|
+ for (j = 0; j < tcount; j++) {
|
|
|
+ if (template[j].klen == *keysize) {
|
|
|
+ key = template[j].key;
|
|
|
+ break;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ crypto_skcipher_clear_flags(tfm, ~0);
|
|
|
+
|
|
|
+ ret = crypto_skcipher_setkey(tfm, key, *keysize);
|
|
|
+ if (ret) {
|
|
|
+ pr_err("setkey() failed flags=%x\n",
|
|
|
+ crypto_skcipher_get_flags(tfm));
|
|
|
+ goto out;
|
|
|
+ }
|
|
|
+
|
|
|
+ iv_len = crypto_skcipher_ivsize(tfm);
|
|
|
+ if (iv_len)
|
|
|
+ memset(&iv, 0xff, iv_len);
|
|
|
+
|
|
|
+ /* Now setup per request stuff, i.e. buffers */
|
|
|
+
|
|
|
+ for (j = 0; j < num_mb; ++j) {
|
|
|
+ struct test_mb_skcipher_data *cur = &data[j];
|
|
|
+ unsigned int k = *b_size;
|
|
|
+ unsigned int pages = DIV_ROUND_UP(k, PAGE_SIZE);
|
|
|
+ unsigned int p = 0;
|
|
|
+
|
|
|
+ sg_init_table(cur->sg, pages);
|
|
|
+
|
|
|
+ while (k > PAGE_SIZE) {
|
|
|
+ sg_set_buf(cur->sg + p, cur->xbuf[p],
|
|
|
+ PAGE_SIZE);
|
|
|
+ memset(cur->xbuf[p], 0xff, PAGE_SIZE);
|
|
|
+ p++;
|
|
|
+ k -= PAGE_SIZE;
|
|
|
+ }
|
|
|
+
|
|
|
+ sg_set_buf(cur->sg + p, cur->xbuf[p], k);
|
|
|
+ memset(cur->xbuf[p], 0xff, k);
|
|
|
+
|
|
|
+ skcipher_request_set_crypt(cur->req, cur->sg,
|
|
|
+ cur->sg, *b_size,
|
|
|
+ iv);
|
|
|
+ }
|
|
|
+
|
|
|
+ if (secs)
|
|
|
+ ret = test_mb_acipher_jiffies(data, enc,
|
|
|
+ *b_size, secs,
|
|
|
+ num_mb);
|
|
|
+ else
|
|
|
+ ret = test_mb_acipher_cycles(data, enc,
|
|
|
+ *b_size, num_mb);
|
|
|
+
|
|
|
+ if (ret) {
|
|
|
+ pr_err("%s() failed flags=%x\n", e,
|
|
|
+ crypto_skcipher_get_flags(tfm));
|
|
|
+ break;
|
|
|
+ }
|
|
|
+ b_size++;
|
|
|
+ i++;
|
|
|
+ } while (*b_size);
|
|
|
+ keysize++;
|
|
|
+ } while (*keysize);
|
|
|
+
|
|
|
+out:
|
|
|
+ for (i = 0; i < num_mb; ++i)
|
|
|
+ skcipher_request_free(data[i].req);
|
|
|
+out_free_xbuf:
|
|
|
+ for (i = 0; i < num_mb; ++i)
|
|
|
+ testmgr_free_buf(data[i].xbuf);
|
|
|
+out_free_tfm:
|
|
|
+ crypto_free_skcipher(tfm);
|
|
|
+out_free_data:
|
|
|
+ kfree(data);
|
|
|
+}
|
|
|
+
|
|
|
static inline int do_one_acipher_op(struct skcipher_request *req, int ret)
|
|
|
{
|
|
|
struct crypto_wait *wait = req->base.data;
|
|
@@ -1557,16 +2179,24 @@ static int do_test(const char *alg, u32 type, u32 mask, int m)
|
|
|
NULL, 0, 16, 16, aead_speed_template_20);
|
|
|
test_aead_speed("gcm(aes)", ENCRYPT, sec,
|
|
|
NULL, 0, 16, 8, speed_template_16_24_32);
|
|
|
+ test_aead_speed("rfc4106(gcm(aes))", DECRYPT, sec,
|
|
|
+ NULL, 0, 16, 16, aead_speed_template_20);
|
|
|
+ test_aead_speed("gcm(aes)", DECRYPT, sec,
|
|
|
+ NULL, 0, 16, 8, speed_template_16_24_32);
|
|
|
break;
|
|
|
|
|
|
case 212:
|
|
|
test_aead_speed("rfc4309(ccm(aes))", ENCRYPT, sec,
|
|
|
NULL, 0, 16, 16, aead_speed_template_19);
|
|
|
+ test_aead_speed("rfc4309(ccm(aes))", DECRYPT, sec,
|
|
|
+ NULL, 0, 16, 16, aead_speed_template_19);
|
|
|
break;
|
|
|
|
|
|
case 213:
|
|
|
test_aead_speed("rfc7539esp(chacha20,poly1305)", ENCRYPT, sec,
|
|
|
NULL, 0, 16, 8, aead_speed_template_36);
|
|
|
+ test_aead_speed("rfc7539esp(chacha20,poly1305)", DECRYPT, sec,
|
|
|
+ NULL, 0, 16, 8, aead_speed_template_36);
|
|
|
break;
|
|
|
|
|
|
case 214:
|
|
@@ -1574,6 +2204,33 @@ static int do_test(const char *alg, u32 type, u32 mask, int m)
|
|
|
speed_template_32);
|
|
|
break;
|
|
|
|
|
|
+ case 215:
|
|
|
+ test_mb_aead_speed("rfc4106(gcm(aes))", ENCRYPT, sec, NULL,
|
|
|
+ 0, 16, 16, aead_speed_template_20, num_mb);
|
|
|
+ test_mb_aead_speed("gcm(aes)", ENCRYPT, sec, NULL, 0, 16, 8,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_aead_speed("rfc4106(gcm(aes))", DECRYPT, sec, NULL,
|
|
|
+ 0, 16, 16, aead_speed_template_20, num_mb);
|
|
|
+ test_mb_aead_speed("gcm(aes)", DECRYPT, sec, NULL, 0, 16, 8,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ break;
|
|
|
+
|
|
|
+ case 216:
|
|
|
+ test_mb_aead_speed("rfc4309(ccm(aes))", ENCRYPT, sec, NULL, 0,
|
|
|
+ 16, 16, aead_speed_template_19, num_mb);
|
|
|
+ test_mb_aead_speed("rfc4309(ccm(aes))", DECRYPT, sec, NULL, 0,
|
|
|
+ 16, 16, aead_speed_template_19, num_mb);
|
|
|
+ break;
|
|
|
+
|
|
|
+ case 217:
|
|
|
+ test_mb_aead_speed("rfc7539esp(chacha20,poly1305)", ENCRYPT,
|
|
|
+ sec, NULL, 0, 16, 8, aead_speed_template_36,
|
|
|
+ num_mb);
|
|
|
+ test_mb_aead_speed("rfc7539esp(chacha20,poly1305)", DECRYPT,
|
|
|
+ sec, NULL, 0, 16, 8, aead_speed_template_36,
|
|
|
+ num_mb);
|
|
|
+ break;
|
|
|
+
|
|
|
case 300:
|
|
|
if (alg) {
|
|
|
test_hash_speed(alg, sec, generic_hash_speed_template);
|
|
@@ -1778,19 +2435,23 @@ static int do_test(const char *alg, u32 type, u32 mask, int m)
|
|
|
if (mode > 400 && mode < 500) break;
|
|
|
/* fall through */
|
|
|
case 422:
|
|
|
- test_mb_ahash_speed("sha1", sec, generic_hash_speed_template);
|
|
|
+ test_mb_ahash_speed("sha1", sec, generic_hash_speed_template,
|
|
|
+ num_mb);
|
|
|
if (mode > 400 && mode < 500) break;
|
|
|
/* fall through */
|
|
|
case 423:
|
|
|
- test_mb_ahash_speed("sha256", sec, generic_hash_speed_template);
|
|
|
+ test_mb_ahash_speed("sha256", sec, generic_hash_speed_template,
|
|
|
+ num_mb);
|
|
|
if (mode > 400 && mode < 500) break;
|
|
|
/* fall through */
|
|
|
case 424:
|
|
|
- test_mb_ahash_speed("sha512", sec, generic_hash_speed_template);
|
|
|
+ test_mb_ahash_speed("sha512", sec, generic_hash_speed_template,
|
|
|
+ num_mb);
|
|
|
if (mode > 400 && mode < 500) break;
|
|
|
/* fall through */
|
|
|
case 425:
|
|
|
- test_mb_ahash_speed("sm3", sec, generic_hash_speed_template);
|
|
|
+ test_mb_ahash_speed("sm3", sec, generic_hash_speed_template,
|
|
|
+ num_mb);
|
|
|
if (mode > 400 && mode < 500) break;
|
|
|
/* fall through */
|
|
|
case 499:
|
|
@@ -2008,6 +2669,218 @@ static int do_test(const char *alg, u32 type, u32 mask, int m)
|
|
|
speed_template_8_32);
|
|
|
break;
|
|
|
|
|
|
+ case 600:
|
|
|
+ test_mb_skcipher_speed("ecb(aes)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ecb(aes)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(aes)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(aes)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("lrw(aes)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_40_48, num_mb);
|
|
|
+ test_mb_skcipher_speed("lrw(aes)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_40_48, num_mb);
|
|
|
+ test_mb_skcipher_speed("xts(aes)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_64, num_mb);
|
|
|
+ test_mb_skcipher_speed("xts(aes)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_64, num_mb);
|
|
|
+ test_mb_skcipher_speed("cts(cbc(aes))", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("cts(cbc(aes))", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ctr(aes)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ctr(aes)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("cfb(aes)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("cfb(aes)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ofb(aes)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ofb(aes)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("rfc3686(ctr(aes))", ENCRYPT, sec, NULL,
|
|
|
+ 0, speed_template_20_28_36, num_mb);
|
|
|
+ test_mb_skcipher_speed("rfc3686(ctr(aes))", DECRYPT, sec, NULL,
|
|
|
+ 0, speed_template_20_28_36, num_mb);
|
|
|
+ break;
|
|
|
+
|
|
|
+ case 601:
|
|
|
+ test_mb_skcipher_speed("ecb(des3_ede)", ENCRYPT, sec,
|
|
|
+ des3_speed_template, DES3_SPEED_VECTORS,
|
|
|
+ speed_template_24, num_mb);
|
|
|
+ test_mb_skcipher_speed("ecb(des3_ede)", DECRYPT, sec,
|
|
|
+ des3_speed_template, DES3_SPEED_VECTORS,
|
|
|
+ speed_template_24, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(des3_ede)", ENCRYPT, sec,
|
|
|
+ des3_speed_template, DES3_SPEED_VECTORS,
|
|
|
+ speed_template_24, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(des3_ede)", DECRYPT, sec,
|
|
|
+ des3_speed_template, DES3_SPEED_VECTORS,
|
|
|
+ speed_template_24, num_mb);
|
|
|
+ test_mb_skcipher_speed("cfb(des3_ede)", ENCRYPT, sec,
|
|
|
+ des3_speed_template, DES3_SPEED_VECTORS,
|
|
|
+ speed_template_24, num_mb);
|
|
|
+ test_mb_skcipher_speed("cfb(des3_ede)", DECRYPT, sec,
|
|
|
+ des3_speed_template, DES3_SPEED_VECTORS,
|
|
|
+ speed_template_24, num_mb);
|
|
|
+ test_mb_skcipher_speed("ofb(des3_ede)", ENCRYPT, sec,
|
|
|
+ des3_speed_template, DES3_SPEED_VECTORS,
|
|
|
+ speed_template_24, num_mb);
|
|
|
+ test_mb_skcipher_speed("ofb(des3_ede)", DECRYPT, sec,
|
|
|
+ des3_speed_template, DES3_SPEED_VECTORS,
|
|
|
+ speed_template_24, num_mb);
|
|
|
+ break;
|
|
|
+
|
|
|
+ case 602:
|
|
|
+ test_mb_skcipher_speed("ecb(des)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8, num_mb);
|
|
|
+ test_mb_skcipher_speed("ecb(des)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(des)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(des)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8, num_mb);
|
|
|
+ test_mb_skcipher_speed("cfb(des)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8, num_mb);
|
|
|
+ test_mb_skcipher_speed("cfb(des)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8, num_mb);
|
|
|
+ test_mb_skcipher_speed("ofb(des)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8, num_mb);
|
|
|
+ test_mb_skcipher_speed("ofb(des)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8, num_mb);
|
|
|
+ break;
|
|
|
+
|
|
|
+ case 603:
|
|
|
+ test_mb_skcipher_speed("ecb(serpent)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ecb(serpent)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(serpent)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(serpent)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ctr(serpent)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ctr(serpent)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("lrw(serpent)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_48, num_mb);
|
|
|
+ test_mb_skcipher_speed("lrw(serpent)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_48, num_mb);
|
|
|
+ test_mb_skcipher_speed("xts(serpent)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_64, num_mb);
|
|
|
+ test_mb_skcipher_speed("xts(serpent)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_64, num_mb);
|
|
|
+ break;
|
|
|
+
|
|
|
+ case 604:
|
|
|
+ test_mb_skcipher_speed("ecb(twofish)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ecb(twofish)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(twofish)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(twofish)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ctr(twofish)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ctr(twofish)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_24_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("lrw(twofish)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_40_48, num_mb);
|
|
|
+ test_mb_skcipher_speed("lrw(twofish)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_40_48, num_mb);
|
|
|
+ test_mb_skcipher_speed("xts(twofish)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_48_64, num_mb);
|
|
|
+ test_mb_skcipher_speed("xts(twofish)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_48_64, num_mb);
|
|
|
+ break;
|
|
|
+
|
|
|
+ case 605:
|
|
|
+ test_mb_skcipher_speed("ecb(arc4)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8, num_mb);
|
|
|
+ break;
|
|
|
+
|
|
|
+ case 606:
|
|
|
+ test_mb_skcipher_speed("ecb(cast5)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8_16, num_mb);
|
|
|
+ test_mb_skcipher_speed("ecb(cast5)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8_16, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(cast5)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8_16, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(cast5)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8_16, num_mb);
|
|
|
+ test_mb_skcipher_speed("ctr(cast5)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8_16, num_mb);
|
|
|
+ test_mb_skcipher_speed("ctr(cast5)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8_16, num_mb);
|
|
|
+ break;
|
|
|
+
|
|
|
+ case 607:
|
|
|
+ test_mb_skcipher_speed("ecb(cast6)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ecb(cast6)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(cast6)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(cast6)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ctr(cast6)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ctr(cast6)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("lrw(cast6)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_48, num_mb);
|
|
|
+ test_mb_skcipher_speed("lrw(cast6)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_48, num_mb);
|
|
|
+ test_mb_skcipher_speed("xts(cast6)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_64, num_mb);
|
|
|
+ test_mb_skcipher_speed("xts(cast6)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_64, num_mb);
|
|
|
+ break;
|
|
|
+
|
|
|
+ case 608:
|
|
|
+ test_mb_skcipher_speed("ecb(camellia)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ecb(camellia)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(camellia)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(camellia)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ctr(camellia)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ctr(camellia)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_16_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("lrw(camellia)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_48, num_mb);
|
|
|
+ test_mb_skcipher_speed("lrw(camellia)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_48, num_mb);
|
|
|
+ test_mb_skcipher_speed("xts(camellia)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_64, num_mb);
|
|
|
+ test_mb_skcipher_speed("xts(camellia)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_32_64, num_mb);
|
|
|
+ break;
|
|
|
+
|
|
|
+ case 609:
|
|
|
+ test_mb_skcipher_speed("ecb(blowfish)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ecb(blowfish)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(blowfish)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("cbc(blowfish)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ctr(blowfish)", ENCRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8_32, num_mb);
|
|
|
+ test_mb_skcipher_speed("ctr(blowfish)", DECRYPT, sec, NULL, 0,
|
|
|
+ speed_template_8_32, num_mb);
|
|
|
+ break;
|
|
|
+
|
|
|
case 1000:
|
|
|
test_available();
|
|
|
break;
|
|
@@ -2069,6 +2942,8 @@ module_param(mode, int, 0);
|
|
|
module_param(sec, uint, 0);
|
|
|
MODULE_PARM_DESC(sec, "Length in seconds of speed tests "
|
|
|
"(defaults to zero which uses CPU cycles instead)");
|
|
|
+module_param(num_mb, uint, 0000);
|
|
|
+MODULE_PARM_DESC(num_mb, "Number of concurrent requests to be used in mb speed tests (defaults to 8)");
|
|
|
|
|
|
MODULE_LICENSE("GPL");
|
|
|
MODULE_DESCRIPTION("Quick & dirty crypto testing module");
|