serpent_avx2_glue.c 16 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567
  1. /*
  2. * Glue Code for x86_64/AVX2 assembler optimized version of Serpent
  3. *
  4. * Copyright © 2012-2013 Jussi Kivilinna <jussi.kivilinna@mbnet.fi>
  5. *
  6. * This program is free software; you can redistribute it and/or modify
  7. * it under the terms of the GNU General Public License as published by
  8. * the Free Software Foundation; either version 2 of the License, or
  9. * (at your option) any later version.
  10. *
  11. */
  12. #include <linux/module.h>
  13. #include <linux/types.h>
  14. #include <linux/crypto.h>
  15. #include <linux/err.h>
  16. #include <crypto/ablk_helper.h>
  17. #include <crypto/algapi.h>
  18. #include <crypto/ctr.h>
  19. #include <crypto/lrw.h>
  20. #include <crypto/xts.h>
  21. #include <crypto/serpent.h>
  22. #include <asm/xcr.h>
  23. #include <asm/xsave.h>
  24. #include <asm/crypto/serpent-avx.h>
  25. #include <asm/crypto/glue_helper.h>
  26. #define SERPENT_AVX2_PARALLEL_BLOCKS 16
  27. /* 16-way AVX2 parallel cipher functions */
  28. asmlinkage void serpent_ecb_enc_16way(struct serpent_ctx *ctx, u8 *dst,
  29. const u8 *src);
  30. asmlinkage void serpent_ecb_dec_16way(struct serpent_ctx *ctx, u8 *dst,
  31. const u8 *src);
  32. asmlinkage void serpent_cbc_dec_16way(void *ctx, u128 *dst, const u128 *src);
  33. asmlinkage void serpent_ctr_16way(void *ctx, u128 *dst, const u128 *src,
  34. le128 *iv);
  35. asmlinkage void serpent_xts_enc_16way(struct serpent_ctx *ctx, u8 *dst,
  36. const u8 *src, le128 *iv);
  37. asmlinkage void serpent_xts_dec_16way(struct serpent_ctx *ctx, u8 *dst,
  38. const u8 *src, le128 *iv);
  39. static const struct common_glue_ctx serpent_enc = {
  40. .num_funcs = 3,
  41. .fpu_blocks_limit = 8,
  42. .funcs = { {
  43. .num_blocks = 16,
  44. .fn_u = { .ecb = GLUE_FUNC_CAST(serpent_ecb_enc_16way) }
  45. }, {
  46. .num_blocks = 8,
  47. .fn_u = { .ecb = GLUE_FUNC_CAST(serpent_ecb_enc_8way_avx) }
  48. }, {
  49. .num_blocks = 1,
  50. .fn_u = { .ecb = GLUE_FUNC_CAST(__serpent_encrypt) }
  51. } }
  52. };
  53. static const struct common_glue_ctx serpent_ctr = {
  54. .num_funcs = 3,
  55. .fpu_blocks_limit = 8,
  56. .funcs = { {
  57. .num_blocks = 16,
  58. .fn_u = { .ctr = GLUE_CTR_FUNC_CAST(serpent_ctr_16way) }
  59. }, {
  60. .num_blocks = 8,
  61. .fn_u = { .ctr = GLUE_CTR_FUNC_CAST(serpent_ctr_8way_avx) }
  62. }, {
  63. .num_blocks = 1,
  64. .fn_u = { .ctr = GLUE_CTR_FUNC_CAST(__serpent_crypt_ctr) }
  65. } }
  66. };
  67. static const struct common_glue_ctx serpent_enc_xts = {
  68. .num_funcs = 3,
  69. .fpu_blocks_limit = 8,
  70. .funcs = { {
  71. .num_blocks = 16,
  72. .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_enc_16way) }
  73. }, {
  74. .num_blocks = 8,
  75. .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_enc_8way_avx) }
  76. }, {
  77. .num_blocks = 1,
  78. .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_enc) }
  79. } }
  80. };
  81. static const struct common_glue_ctx serpent_dec = {
  82. .num_funcs = 3,
  83. .fpu_blocks_limit = 8,
  84. .funcs = { {
  85. .num_blocks = 16,
  86. .fn_u = { .ecb = GLUE_FUNC_CAST(serpent_ecb_dec_16way) }
  87. }, {
  88. .num_blocks = 8,
  89. .fn_u = { .ecb = GLUE_FUNC_CAST(serpent_ecb_dec_8way_avx) }
  90. }, {
  91. .num_blocks = 1,
  92. .fn_u = { .ecb = GLUE_FUNC_CAST(__serpent_decrypt) }
  93. } }
  94. };
  95. static const struct common_glue_ctx serpent_dec_cbc = {
  96. .num_funcs = 3,
  97. .fpu_blocks_limit = 8,
  98. .funcs = { {
  99. .num_blocks = 16,
  100. .fn_u = { .cbc = GLUE_CBC_FUNC_CAST(serpent_cbc_dec_16way) }
  101. }, {
  102. .num_blocks = 8,
  103. .fn_u = { .cbc = GLUE_CBC_FUNC_CAST(serpent_cbc_dec_8way_avx) }
  104. }, {
  105. .num_blocks = 1,
  106. .fn_u = { .cbc = GLUE_CBC_FUNC_CAST(__serpent_decrypt) }
  107. } }
  108. };
  109. static const struct common_glue_ctx serpent_dec_xts = {
  110. .num_funcs = 3,
  111. .fpu_blocks_limit = 8,
  112. .funcs = { {
  113. .num_blocks = 16,
  114. .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_dec_16way) }
  115. }, {
  116. .num_blocks = 8,
  117. .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_dec_8way_avx) }
  118. }, {
  119. .num_blocks = 1,
  120. .fn_u = { .xts = GLUE_XTS_FUNC_CAST(serpent_xts_dec) }
  121. } }
  122. };
  123. static int ecb_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  124. struct scatterlist *src, unsigned int nbytes)
  125. {
  126. return glue_ecb_crypt_128bit(&serpent_enc, desc, dst, src, nbytes);
  127. }
  128. static int ecb_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  129. struct scatterlist *src, unsigned int nbytes)
  130. {
  131. return glue_ecb_crypt_128bit(&serpent_dec, desc, dst, src, nbytes);
  132. }
  133. static int cbc_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  134. struct scatterlist *src, unsigned int nbytes)
  135. {
  136. return glue_cbc_encrypt_128bit(GLUE_FUNC_CAST(__serpent_encrypt), desc,
  137. dst, src, nbytes);
  138. }
  139. static int cbc_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  140. struct scatterlist *src, unsigned int nbytes)
  141. {
  142. return glue_cbc_decrypt_128bit(&serpent_dec_cbc, desc, dst, src,
  143. nbytes);
  144. }
  145. static int ctr_crypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  146. struct scatterlist *src, unsigned int nbytes)
  147. {
  148. return glue_ctr_crypt_128bit(&serpent_ctr, desc, dst, src, nbytes);
  149. }
  150. static inline bool serpent_fpu_begin(bool fpu_enabled, unsigned int nbytes)
  151. {
  152. /* since reusing AVX functions, starts using FPU at 8 parallel blocks */
  153. return glue_fpu_begin(SERPENT_BLOCK_SIZE, 8, NULL, fpu_enabled, nbytes);
  154. }
  155. static inline void serpent_fpu_end(bool fpu_enabled)
  156. {
  157. glue_fpu_end(fpu_enabled);
  158. }
  159. struct crypt_priv {
  160. struct serpent_ctx *ctx;
  161. bool fpu_enabled;
  162. };
  163. static void encrypt_callback(void *priv, u8 *srcdst, unsigned int nbytes)
  164. {
  165. const unsigned int bsize = SERPENT_BLOCK_SIZE;
  166. struct crypt_priv *ctx = priv;
  167. int i;
  168. ctx->fpu_enabled = serpent_fpu_begin(ctx->fpu_enabled, nbytes);
  169. if (nbytes >= SERPENT_AVX2_PARALLEL_BLOCKS * bsize) {
  170. serpent_ecb_enc_16way(ctx->ctx, srcdst, srcdst);
  171. srcdst += bsize * SERPENT_AVX2_PARALLEL_BLOCKS;
  172. nbytes -= bsize * SERPENT_AVX2_PARALLEL_BLOCKS;
  173. }
  174. while (nbytes >= SERPENT_PARALLEL_BLOCKS * bsize) {
  175. serpent_ecb_enc_8way_avx(ctx->ctx, srcdst, srcdst);
  176. srcdst += bsize * SERPENT_PARALLEL_BLOCKS;
  177. nbytes -= bsize * SERPENT_PARALLEL_BLOCKS;
  178. }
  179. for (i = 0; i < nbytes / bsize; i++, srcdst += bsize)
  180. __serpent_encrypt(ctx->ctx, srcdst, srcdst);
  181. }
  182. static void decrypt_callback(void *priv, u8 *srcdst, unsigned int nbytes)
  183. {
  184. const unsigned int bsize = SERPENT_BLOCK_SIZE;
  185. struct crypt_priv *ctx = priv;
  186. int i;
  187. ctx->fpu_enabled = serpent_fpu_begin(ctx->fpu_enabled, nbytes);
  188. if (nbytes >= SERPENT_AVX2_PARALLEL_BLOCKS * bsize) {
  189. serpent_ecb_dec_16way(ctx->ctx, srcdst, srcdst);
  190. srcdst += bsize * SERPENT_AVX2_PARALLEL_BLOCKS;
  191. nbytes -= bsize * SERPENT_AVX2_PARALLEL_BLOCKS;
  192. }
  193. while (nbytes >= SERPENT_PARALLEL_BLOCKS * bsize) {
  194. serpent_ecb_dec_8way_avx(ctx->ctx, srcdst, srcdst);
  195. srcdst += bsize * SERPENT_PARALLEL_BLOCKS;
  196. nbytes -= bsize * SERPENT_PARALLEL_BLOCKS;
  197. }
  198. for (i = 0; i < nbytes / bsize; i++, srcdst += bsize)
  199. __serpent_decrypt(ctx->ctx, srcdst, srcdst);
  200. }
  201. static int lrw_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  202. struct scatterlist *src, unsigned int nbytes)
  203. {
  204. struct serpent_lrw_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
  205. be128 buf[SERPENT_AVX2_PARALLEL_BLOCKS];
  206. struct crypt_priv crypt_ctx = {
  207. .ctx = &ctx->serpent_ctx,
  208. .fpu_enabled = false,
  209. };
  210. struct lrw_crypt_req req = {
  211. .tbuf = buf,
  212. .tbuflen = sizeof(buf),
  213. .table_ctx = &ctx->lrw_table,
  214. .crypt_ctx = &crypt_ctx,
  215. .crypt_fn = encrypt_callback,
  216. };
  217. int ret;
  218. desc->flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP;
  219. ret = lrw_crypt(desc, dst, src, nbytes, &req);
  220. serpent_fpu_end(crypt_ctx.fpu_enabled);
  221. return ret;
  222. }
  223. static int lrw_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  224. struct scatterlist *src, unsigned int nbytes)
  225. {
  226. struct serpent_lrw_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
  227. be128 buf[SERPENT_AVX2_PARALLEL_BLOCKS];
  228. struct crypt_priv crypt_ctx = {
  229. .ctx = &ctx->serpent_ctx,
  230. .fpu_enabled = false,
  231. };
  232. struct lrw_crypt_req req = {
  233. .tbuf = buf,
  234. .tbuflen = sizeof(buf),
  235. .table_ctx = &ctx->lrw_table,
  236. .crypt_ctx = &crypt_ctx,
  237. .crypt_fn = decrypt_callback,
  238. };
  239. int ret;
  240. desc->flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP;
  241. ret = lrw_crypt(desc, dst, src, nbytes, &req);
  242. serpent_fpu_end(crypt_ctx.fpu_enabled);
  243. return ret;
  244. }
  245. static int xts_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  246. struct scatterlist *src, unsigned int nbytes)
  247. {
  248. struct serpent_xts_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
  249. return glue_xts_crypt_128bit(&serpent_enc_xts, desc, dst, src, nbytes,
  250. XTS_TWEAK_CAST(__serpent_encrypt),
  251. &ctx->tweak_ctx, &ctx->crypt_ctx);
  252. }
  253. static int xts_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  254. struct scatterlist *src, unsigned int nbytes)
  255. {
  256. struct serpent_xts_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
  257. return glue_xts_crypt_128bit(&serpent_dec_xts, desc, dst, src, nbytes,
  258. XTS_TWEAK_CAST(__serpent_encrypt),
  259. &ctx->tweak_ctx, &ctx->crypt_ctx);
  260. }
  261. static struct crypto_alg srp_algs[10] = { {
  262. .cra_name = "__ecb-serpent-avx2",
  263. .cra_driver_name = "__driver-ecb-serpent-avx2",
  264. .cra_priority = 0,
  265. .cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER |
  266. CRYPTO_ALG_INTERNAL,
  267. .cra_blocksize = SERPENT_BLOCK_SIZE,
  268. .cra_ctxsize = sizeof(struct serpent_ctx),
  269. .cra_alignmask = 0,
  270. .cra_type = &crypto_blkcipher_type,
  271. .cra_module = THIS_MODULE,
  272. .cra_list = LIST_HEAD_INIT(srp_algs[0].cra_list),
  273. .cra_u = {
  274. .blkcipher = {
  275. .min_keysize = SERPENT_MIN_KEY_SIZE,
  276. .max_keysize = SERPENT_MAX_KEY_SIZE,
  277. .setkey = serpent_setkey,
  278. .encrypt = ecb_encrypt,
  279. .decrypt = ecb_decrypt,
  280. },
  281. },
  282. }, {
  283. .cra_name = "__cbc-serpent-avx2",
  284. .cra_driver_name = "__driver-cbc-serpent-avx2",
  285. .cra_priority = 0,
  286. .cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER |
  287. CRYPTO_ALG_INTERNAL,
  288. .cra_blocksize = SERPENT_BLOCK_SIZE,
  289. .cra_ctxsize = sizeof(struct serpent_ctx),
  290. .cra_alignmask = 0,
  291. .cra_type = &crypto_blkcipher_type,
  292. .cra_module = THIS_MODULE,
  293. .cra_list = LIST_HEAD_INIT(srp_algs[1].cra_list),
  294. .cra_u = {
  295. .blkcipher = {
  296. .min_keysize = SERPENT_MIN_KEY_SIZE,
  297. .max_keysize = SERPENT_MAX_KEY_SIZE,
  298. .setkey = serpent_setkey,
  299. .encrypt = cbc_encrypt,
  300. .decrypt = cbc_decrypt,
  301. },
  302. },
  303. }, {
  304. .cra_name = "__ctr-serpent-avx2",
  305. .cra_driver_name = "__driver-ctr-serpent-avx2",
  306. .cra_priority = 0,
  307. .cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER |
  308. CRYPTO_ALG_INTERNAL,
  309. .cra_blocksize = 1,
  310. .cra_ctxsize = sizeof(struct serpent_ctx),
  311. .cra_alignmask = 0,
  312. .cra_type = &crypto_blkcipher_type,
  313. .cra_module = THIS_MODULE,
  314. .cra_list = LIST_HEAD_INIT(srp_algs[2].cra_list),
  315. .cra_u = {
  316. .blkcipher = {
  317. .min_keysize = SERPENT_MIN_KEY_SIZE,
  318. .max_keysize = SERPENT_MAX_KEY_SIZE,
  319. .ivsize = SERPENT_BLOCK_SIZE,
  320. .setkey = serpent_setkey,
  321. .encrypt = ctr_crypt,
  322. .decrypt = ctr_crypt,
  323. },
  324. },
  325. }, {
  326. .cra_name = "__lrw-serpent-avx2",
  327. .cra_driver_name = "__driver-lrw-serpent-avx2",
  328. .cra_priority = 0,
  329. .cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER |
  330. CRYPTO_ALG_INTERNAL,
  331. .cra_blocksize = SERPENT_BLOCK_SIZE,
  332. .cra_ctxsize = sizeof(struct serpent_lrw_ctx),
  333. .cra_alignmask = 0,
  334. .cra_type = &crypto_blkcipher_type,
  335. .cra_module = THIS_MODULE,
  336. .cra_list = LIST_HEAD_INIT(srp_algs[3].cra_list),
  337. .cra_exit = lrw_serpent_exit_tfm,
  338. .cra_u = {
  339. .blkcipher = {
  340. .min_keysize = SERPENT_MIN_KEY_SIZE +
  341. SERPENT_BLOCK_SIZE,
  342. .max_keysize = SERPENT_MAX_KEY_SIZE +
  343. SERPENT_BLOCK_SIZE,
  344. .ivsize = SERPENT_BLOCK_SIZE,
  345. .setkey = lrw_serpent_setkey,
  346. .encrypt = lrw_encrypt,
  347. .decrypt = lrw_decrypt,
  348. },
  349. },
  350. }, {
  351. .cra_name = "__xts-serpent-avx2",
  352. .cra_driver_name = "__driver-xts-serpent-avx2",
  353. .cra_priority = 0,
  354. .cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER |
  355. CRYPTO_ALG_INTERNAL,
  356. .cra_blocksize = SERPENT_BLOCK_SIZE,
  357. .cra_ctxsize = sizeof(struct serpent_xts_ctx),
  358. .cra_alignmask = 0,
  359. .cra_type = &crypto_blkcipher_type,
  360. .cra_module = THIS_MODULE,
  361. .cra_list = LIST_HEAD_INIT(srp_algs[4].cra_list),
  362. .cra_u = {
  363. .blkcipher = {
  364. .min_keysize = SERPENT_MIN_KEY_SIZE * 2,
  365. .max_keysize = SERPENT_MAX_KEY_SIZE * 2,
  366. .ivsize = SERPENT_BLOCK_SIZE,
  367. .setkey = xts_serpent_setkey,
  368. .encrypt = xts_encrypt,
  369. .decrypt = xts_decrypt,
  370. },
  371. },
  372. }, {
  373. .cra_name = "ecb(serpent)",
  374. .cra_driver_name = "ecb-serpent-avx2",
  375. .cra_priority = 600,
  376. .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
  377. .cra_blocksize = SERPENT_BLOCK_SIZE,
  378. .cra_ctxsize = sizeof(struct async_helper_ctx),
  379. .cra_alignmask = 0,
  380. .cra_type = &crypto_ablkcipher_type,
  381. .cra_module = THIS_MODULE,
  382. .cra_list = LIST_HEAD_INIT(srp_algs[5].cra_list),
  383. .cra_init = ablk_init,
  384. .cra_exit = ablk_exit,
  385. .cra_u = {
  386. .ablkcipher = {
  387. .min_keysize = SERPENT_MIN_KEY_SIZE,
  388. .max_keysize = SERPENT_MAX_KEY_SIZE,
  389. .setkey = ablk_set_key,
  390. .encrypt = ablk_encrypt,
  391. .decrypt = ablk_decrypt,
  392. },
  393. },
  394. }, {
  395. .cra_name = "cbc(serpent)",
  396. .cra_driver_name = "cbc-serpent-avx2",
  397. .cra_priority = 600,
  398. .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
  399. .cra_blocksize = SERPENT_BLOCK_SIZE,
  400. .cra_ctxsize = sizeof(struct async_helper_ctx),
  401. .cra_alignmask = 0,
  402. .cra_type = &crypto_ablkcipher_type,
  403. .cra_module = THIS_MODULE,
  404. .cra_list = LIST_HEAD_INIT(srp_algs[6].cra_list),
  405. .cra_init = ablk_init,
  406. .cra_exit = ablk_exit,
  407. .cra_u = {
  408. .ablkcipher = {
  409. .min_keysize = SERPENT_MIN_KEY_SIZE,
  410. .max_keysize = SERPENT_MAX_KEY_SIZE,
  411. .ivsize = SERPENT_BLOCK_SIZE,
  412. .setkey = ablk_set_key,
  413. .encrypt = __ablk_encrypt,
  414. .decrypt = ablk_decrypt,
  415. },
  416. },
  417. }, {
  418. .cra_name = "ctr(serpent)",
  419. .cra_driver_name = "ctr-serpent-avx2",
  420. .cra_priority = 600,
  421. .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
  422. .cra_blocksize = 1,
  423. .cra_ctxsize = sizeof(struct async_helper_ctx),
  424. .cra_alignmask = 0,
  425. .cra_type = &crypto_ablkcipher_type,
  426. .cra_module = THIS_MODULE,
  427. .cra_list = LIST_HEAD_INIT(srp_algs[7].cra_list),
  428. .cra_init = ablk_init,
  429. .cra_exit = ablk_exit,
  430. .cra_u = {
  431. .ablkcipher = {
  432. .min_keysize = SERPENT_MIN_KEY_SIZE,
  433. .max_keysize = SERPENT_MAX_KEY_SIZE,
  434. .ivsize = SERPENT_BLOCK_SIZE,
  435. .setkey = ablk_set_key,
  436. .encrypt = ablk_encrypt,
  437. .decrypt = ablk_encrypt,
  438. .geniv = "chainiv",
  439. },
  440. },
  441. }, {
  442. .cra_name = "lrw(serpent)",
  443. .cra_driver_name = "lrw-serpent-avx2",
  444. .cra_priority = 600,
  445. .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
  446. .cra_blocksize = SERPENT_BLOCK_SIZE,
  447. .cra_ctxsize = sizeof(struct async_helper_ctx),
  448. .cra_alignmask = 0,
  449. .cra_type = &crypto_ablkcipher_type,
  450. .cra_module = THIS_MODULE,
  451. .cra_list = LIST_HEAD_INIT(srp_algs[8].cra_list),
  452. .cra_init = ablk_init,
  453. .cra_exit = ablk_exit,
  454. .cra_u = {
  455. .ablkcipher = {
  456. .min_keysize = SERPENT_MIN_KEY_SIZE +
  457. SERPENT_BLOCK_SIZE,
  458. .max_keysize = SERPENT_MAX_KEY_SIZE +
  459. SERPENT_BLOCK_SIZE,
  460. .ivsize = SERPENT_BLOCK_SIZE,
  461. .setkey = ablk_set_key,
  462. .encrypt = ablk_encrypt,
  463. .decrypt = ablk_decrypt,
  464. },
  465. },
  466. }, {
  467. .cra_name = "xts(serpent)",
  468. .cra_driver_name = "xts-serpent-avx2",
  469. .cra_priority = 600,
  470. .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
  471. .cra_blocksize = SERPENT_BLOCK_SIZE,
  472. .cra_ctxsize = sizeof(struct async_helper_ctx),
  473. .cra_alignmask = 0,
  474. .cra_type = &crypto_ablkcipher_type,
  475. .cra_module = THIS_MODULE,
  476. .cra_list = LIST_HEAD_INIT(srp_algs[9].cra_list),
  477. .cra_init = ablk_init,
  478. .cra_exit = ablk_exit,
  479. .cra_u = {
  480. .ablkcipher = {
  481. .min_keysize = SERPENT_MIN_KEY_SIZE * 2,
  482. .max_keysize = SERPENT_MAX_KEY_SIZE * 2,
  483. .ivsize = SERPENT_BLOCK_SIZE,
  484. .setkey = ablk_set_key,
  485. .encrypt = ablk_encrypt,
  486. .decrypt = ablk_decrypt,
  487. },
  488. },
  489. } };
  490. static int __init init(void)
  491. {
  492. u64 xcr0;
  493. if (!cpu_has_avx2 || !cpu_has_osxsave) {
  494. pr_info("AVX2 instructions are not detected.\n");
  495. return -ENODEV;
  496. }
  497. xcr0 = xgetbv(XCR_XFEATURE_ENABLED_MASK);
  498. if ((xcr0 & (XSTATE_SSE | XSTATE_YMM)) != (XSTATE_SSE | XSTATE_YMM)) {
  499. pr_info("AVX detected but unusable.\n");
  500. return -ENODEV;
  501. }
  502. return crypto_register_algs(srp_algs, ARRAY_SIZE(srp_algs));
  503. }
  504. static void __exit fini(void)
  505. {
  506. crypto_unregister_algs(srp_algs, ARRAY_SIZE(srp_algs));
  507. }
  508. module_init(init);
  509. module_exit(fini);
  510. MODULE_LICENSE("GPL");
  511. MODULE_DESCRIPTION("Serpent Cipher Algorithm, AVX2 optimized");
  512. MODULE_ALIAS_CRYPTO("serpent");
  513. MODULE_ALIAS_CRYPTO("serpent-asm");