|
@@ -46,9 +46,9 @@ static int set_secret(struct ceph_crypto_key *key, void *buf)
|
|
|
goto fail;
|
|
|
}
|
|
|
|
|
|
- /* crypto_alloc_skcipher() allocates with GFP_KERNEL */
|
|
|
+ /* crypto_alloc_sync_skcipher() allocates with GFP_KERNEL */
|
|
|
noio_flag = memalloc_noio_save();
|
|
|
- key->tfm = crypto_alloc_skcipher("cbc(aes)", 0, CRYPTO_ALG_ASYNC);
|
|
|
+ key->tfm = crypto_alloc_sync_skcipher("cbc(aes)", 0, 0);
|
|
|
memalloc_noio_restore(noio_flag);
|
|
|
if (IS_ERR(key->tfm)) {
|
|
|
ret = PTR_ERR(key->tfm);
|
|
@@ -56,7 +56,7 @@ static int set_secret(struct ceph_crypto_key *key, void *buf)
|
|
|
goto fail;
|
|
|
}
|
|
|
|
|
|
- ret = crypto_skcipher_setkey(key->tfm, key->key, key->len);
|
|
|
+ ret = crypto_sync_skcipher_setkey(key->tfm, key->key, key->len);
|
|
|
if (ret)
|
|
|
goto fail;
|
|
|
|
|
@@ -136,7 +136,7 @@ void ceph_crypto_key_destroy(struct ceph_crypto_key *key)
|
|
|
if (key) {
|
|
|
kfree(key->key);
|
|
|
key->key = NULL;
|
|
|
- crypto_free_skcipher(key->tfm);
|
|
|
+ crypto_free_sync_skcipher(key->tfm);
|
|
|
key->tfm = NULL;
|
|
|
}
|
|
|
}
|
|
@@ -216,7 +216,7 @@ static void teardown_sgtable(struct sg_table *sgt)
|
|
|
static int ceph_aes_crypt(const struct ceph_crypto_key *key, bool encrypt,
|
|
|
void *buf, int buf_len, int in_len, int *pout_len)
|
|
|
{
|
|
|
- SKCIPHER_REQUEST_ON_STACK(req, key->tfm);
|
|
|
+ SYNC_SKCIPHER_REQUEST_ON_STACK(req, key->tfm);
|
|
|
struct sg_table sgt;
|
|
|
struct scatterlist prealloc_sg;
|
|
|
char iv[AES_BLOCK_SIZE] __aligned(8);
|
|
@@ -232,7 +232,7 @@ static int ceph_aes_crypt(const struct ceph_crypto_key *key, bool encrypt,
|
|
|
return ret;
|
|
|
|
|
|
memcpy(iv, aes_iv, AES_BLOCK_SIZE);
|
|
|
- skcipher_request_set_tfm(req, key->tfm);
|
|
|
+ skcipher_request_set_sync_tfm(req, key->tfm);
|
|
|
skcipher_request_set_callback(req, 0, NULL, NULL);
|
|
|
skcipher_request_set_crypt(req, sgt.sgl, sgt.sgl, crypt_len, iv);
|
|
|
|