|
@@ -15,12 +15,49 @@
|
|
|
#include <linux/err.h>
|
|
|
#include <linux/delay.h>
|
|
|
#include <crypto/engine.h>
|
|
|
-#include <crypto/internal/hash.h>
|
|
|
#include <uapi/linux/sched/types.h>
|
|
|
#include "internal.h"
|
|
|
|
|
|
#define CRYPTO_ENGINE_MAX_QLEN 10
|
|
|
|
|
|
+/**
|
|
|
+ * crypto_finalize_request - finalize one request if the request is done
|
|
|
+ * @engine: the hardware engine
|
|
|
+ * @req: the request need to be finalized
|
|
|
+ * @err: error number
|
|
|
+ */
|
|
|
+static void crypto_finalize_request(struct crypto_engine *engine,
|
|
|
+ struct crypto_async_request *req, int err)
|
|
|
+{
|
|
|
+ unsigned long flags;
|
|
|
+ bool finalize_cur_req = false;
|
|
|
+ int ret;
|
|
|
+ struct crypto_engine_ctx *enginectx;
|
|
|
+
|
|
|
+ spin_lock_irqsave(&engine->queue_lock, flags);
|
|
|
+ if (engine->cur_req == req)
|
|
|
+ finalize_cur_req = true;
|
|
|
+ spin_unlock_irqrestore(&engine->queue_lock, flags);
|
|
|
+
|
|
|
+ if (finalize_cur_req) {
|
|
|
+ enginectx = crypto_tfm_ctx(req->tfm);
|
|
|
+ if (engine->cur_req_prepared &&
|
|
|
+ enginectx->op.unprepare_request) {
|
|
|
+ ret = enginectx->op.unprepare_request(engine, req);
|
|
|
+ if (ret)
|
|
|
+ dev_err(engine->dev, "failed to unprepare request\n");
|
|
|
+ }
|
|
|
+ spin_lock_irqsave(&engine->queue_lock, flags);
|
|
|
+ engine->cur_req = NULL;
|
|
|
+ engine->cur_req_prepared = false;
|
|
|
+ spin_unlock_irqrestore(&engine->queue_lock, flags);
|
|
|
+ }
|
|
|
+
|
|
|
+ req->complete(req, err);
|
|
|
+
|
|
|
+ kthread_queue_work(engine->kworker, &engine->pump_requests);
|
|
|
+}
|
|
|
+
|
|
|
/**
|
|
|
* crypto_pump_requests - dequeue one request from engine queue to process
|
|
|
* @engine: the hardware engine
|
|
@@ -34,11 +71,10 @@ static void crypto_pump_requests(struct crypto_engine *engine,
|
|
|
bool in_kthread)
|
|
|
{
|
|
|
struct crypto_async_request *async_req, *backlog;
|
|
|
- struct ahash_request *hreq;
|
|
|
- struct ablkcipher_request *breq;
|
|
|
unsigned long flags;
|
|
|
bool was_busy = false;
|
|
|
- int ret, rtype;
|
|
|
+ int ret;
|
|
|
+ struct crypto_engine_ctx *enginectx;
|
|
|
|
|
|
spin_lock_irqsave(&engine->queue_lock, flags);
|
|
|
|
|
@@ -94,7 +130,6 @@ static void crypto_pump_requests(struct crypto_engine *engine,
|
|
|
|
|
|
spin_unlock_irqrestore(&engine->queue_lock, flags);
|
|
|
|
|
|
- rtype = crypto_tfm_alg_type(engine->cur_req->tfm);
|
|
|
/* Until here we get the request need to be encrypted successfully */
|
|
|
if (!was_busy && engine->prepare_crypt_hardware) {
|
|
|
ret = engine->prepare_crypt_hardware(engine);
|
|
@@ -104,57 +139,31 @@ static void crypto_pump_requests(struct crypto_engine *engine,
|
|
|
}
|
|
|
}
|
|
|
|
|
|
- switch (rtype) {
|
|
|
- case CRYPTO_ALG_TYPE_AHASH:
|
|
|
- hreq = ahash_request_cast(engine->cur_req);
|
|
|
- if (engine->prepare_hash_request) {
|
|
|
- ret = engine->prepare_hash_request(engine, hreq);
|
|
|
- if (ret) {
|
|
|
- dev_err(engine->dev, "failed to prepare request: %d\n",
|
|
|
- ret);
|
|
|
- goto req_err;
|
|
|
- }
|
|
|
- engine->cur_req_prepared = true;
|
|
|
- }
|
|
|
- ret = engine->hash_one_request(engine, hreq);
|
|
|
- if (ret) {
|
|
|
- dev_err(engine->dev, "failed to hash one request from queue\n");
|
|
|
- goto req_err;
|
|
|
- }
|
|
|
- return;
|
|
|
- case CRYPTO_ALG_TYPE_ABLKCIPHER:
|
|
|
- breq = ablkcipher_request_cast(engine->cur_req);
|
|
|
- if (engine->prepare_cipher_request) {
|
|
|
- ret = engine->prepare_cipher_request(engine, breq);
|
|
|
- if (ret) {
|
|
|
- dev_err(engine->dev, "failed to prepare request: %d\n",
|
|
|
- ret);
|
|
|
- goto req_err;
|
|
|
- }
|
|
|
- engine->cur_req_prepared = true;
|
|
|
- }
|
|
|
- ret = engine->cipher_one_request(engine, breq);
|
|
|
+ enginectx = crypto_tfm_ctx(async_req->tfm);
|
|
|
+
|
|
|
+ if (enginectx->op.prepare_request) {
|
|
|
+ ret = enginectx->op.prepare_request(engine, async_req);
|
|
|
if (ret) {
|
|
|
- dev_err(engine->dev, "failed to cipher one request from queue\n");
|
|
|
+ dev_err(engine->dev, "failed to prepare request: %d\n",
|
|
|
+ ret);
|
|
|
goto req_err;
|
|
|
}
|
|
|
- return;
|
|
|
- default:
|
|
|
- dev_err(engine->dev, "failed to prepare request of unknown type\n");
|
|
|
- return;
|
|
|
+ engine->cur_req_prepared = true;
|
|
|
+ }
|
|
|
+ if (!enginectx->op.do_one_request) {
|
|
|
+ dev_err(engine->dev, "failed to do request\n");
|
|
|
+ ret = -EINVAL;
|
|
|
+ goto req_err;
|
|
|
}
|
|
|
+ ret = enginectx->op.do_one_request(engine, async_req);
|
|
|
+ if (ret) {
|
|
|
+ dev_err(engine->dev, "Failed to do one request from queue: %d\n", ret);
|
|
|
+ goto req_err;
|
|
|
+ }
|
|
|
+ return;
|
|
|
|
|
|
req_err:
|
|
|
- switch (rtype) {
|
|
|
- case CRYPTO_ALG_TYPE_AHASH:
|
|
|
- hreq = ahash_request_cast(engine->cur_req);
|
|
|
- crypto_finalize_hash_request(engine, hreq, ret);
|
|
|
- break;
|
|
|
- case CRYPTO_ALG_TYPE_ABLKCIPHER:
|
|
|
- breq = ablkcipher_request_cast(engine->cur_req);
|
|
|
- crypto_finalize_cipher_request(engine, breq, ret);
|
|
|
- break;
|
|
|
- }
|
|
|
+ crypto_finalize_request(engine, async_req, ret);
|
|
|
return;
|
|
|
|
|
|
out:
|
|
@@ -170,13 +179,12 @@ static void crypto_pump_work(struct kthread_work *work)
|
|
|
}
|
|
|
|
|
|
/**
|
|
|
- * crypto_transfer_cipher_request - transfer the new request into the
|
|
|
- * enginequeue
|
|
|
+ * crypto_transfer_request - transfer the new request into the engine queue
|
|
|
* @engine: the hardware engine
|
|
|
* @req: the request need to be listed into the engine queue
|
|
|
*/
|
|
|
-int crypto_transfer_cipher_request(struct crypto_engine *engine,
|
|
|
- struct ablkcipher_request *req,
|
|
|
+static int crypto_transfer_request(struct crypto_engine *engine,
|
|
|
+ struct crypto_async_request *req,
|
|
|
bool need_pump)
|
|
|
{
|
|
|
unsigned long flags;
|
|
@@ -189,7 +197,7 @@ int crypto_transfer_cipher_request(struct crypto_engine *engine,
|
|
|
return -ESHUTDOWN;
|
|
|
}
|
|
|
|
|
|
- ret = ablkcipher_enqueue_request(&engine->queue, req);
|
|
|
+ ret = crypto_enqueue_request(&engine->queue, req);
|
|
|
|
|
|
if (!engine->busy && need_pump)
|
|
|
kthread_queue_work(engine->kworker, &engine->pump_requests);
|
|
@@ -197,102 +205,131 @@ int crypto_transfer_cipher_request(struct crypto_engine *engine,
|
|
|
spin_unlock_irqrestore(&engine->queue_lock, flags);
|
|
|
return ret;
|
|
|
}
|
|
|
-EXPORT_SYMBOL_GPL(crypto_transfer_cipher_request);
|
|
|
|
|
|
/**
|
|
|
- * crypto_transfer_cipher_request_to_engine - transfer one request to list
|
|
|
+ * crypto_transfer_request_to_engine - transfer one request to list
|
|
|
* into the engine queue
|
|
|
* @engine: the hardware engine
|
|
|
* @req: the request need to be listed into the engine queue
|
|
|
*/
|
|
|
-int crypto_transfer_cipher_request_to_engine(struct crypto_engine *engine,
|
|
|
- struct ablkcipher_request *req)
|
|
|
+static int crypto_transfer_request_to_engine(struct crypto_engine *engine,
|
|
|
+ struct crypto_async_request *req)
|
|
|
{
|
|
|
- return crypto_transfer_cipher_request(engine, req, true);
|
|
|
+ return crypto_transfer_request(engine, req, true);
|
|
|
}
|
|
|
-EXPORT_SYMBOL_GPL(crypto_transfer_cipher_request_to_engine);
|
|
|
|
|
|
/**
|
|
|
- * crypto_transfer_hash_request - transfer the new request into the
|
|
|
- * enginequeue
|
|
|
+ * crypto_transfer_ablkcipher_request_to_engine - transfer one ablkcipher_request
|
|
|
+ * to list into the engine queue
|
|
|
* @engine: the hardware engine
|
|
|
* @req: the request need to be listed into the engine queue
|
|
|
+ * TODO: Remove this function when skcipher conversion is finished
|
|
|
*/
|
|
|
-int crypto_transfer_hash_request(struct crypto_engine *engine,
|
|
|
- struct ahash_request *req, bool need_pump)
|
|
|
+int crypto_transfer_ablkcipher_request_to_engine(struct crypto_engine *engine,
|
|
|
+ struct ablkcipher_request *req)
|
|
|
{
|
|
|
- unsigned long flags;
|
|
|
- int ret;
|
|
|
-
|
|
|
- spin_lock_irqsave(&engine->queue_lock, flags);
|
|
|
-
|
|
|
- if (!engine->running) {
|
|
|
- spin_unlock_irqrestore(&engine->queue_lock, flags);
|
|
|
- return -ESHUTDOWN;
|
|
|
- }
|
|
|
-
|
|
|
- ret = ahash_enqueue_request(&engine->queue, req);
|
|
|
+ return crypto_transfer_request_to_engine(engine, &req->base);
|
|
|
+}
|
|
|
+EXPORT_SYMBOL_GPL(crypto_transfer_ablkcipher_request_to_engine);
|
|
|
|
|
|
- if (!engine->busy && need_pump)
|
|
|
- kthread_queue_work(engine->kworker, &engine->pump_requests);
|
|
|
+/**
|
|
|
+ * crypto_transfer_aead_request_to_engine - transfer one aead_request
|
|
|
+ * to list into the engine queue
|
|
|
+ * @engine: the hardware engine
|
|
|
+ * @req: the request need to be listed into the engine queue
|
|
|
+ */
|
|
|
+int crypto_transfer_aead_request_to_engine(struct crypto_engine *engine,
|
|
|
+ struct aead_request *req)
|
|
|
+{
|
|
|
+ return crypto_transfer_request_to_engine(engine, &req->base);
|
|
|
+}
|
|
|
+EXPORT_SYMBOL_GPL(crypto_transfer_aead_request_to_engine);
|
|
|
|
|
|
- spin_unlock_irqrestore(&engine->queue_lock, flags);
|
|
|
- return ret;
|
|
|
+/**
|
|
|
+ * crypto_transfer_akcipher_request_to_engine - transfer one akcipher_request
|
|
|
+ * to list into the engine queue
|
|
|
+ * @engine: the hardware engine
|
|
|
+ * @req: the request need to be listed into the engine queue
|
|
|
+ */
|
|
|
+int crypto_transfer_akcipher_request_to_engine(struct crypto_engine *engine,
|
|
|
+ struct akcipher_request *req)
|
|
|
+{
|
|
|
+ return crypto_transfer_request_to_engine(engine, &req->base);
|
|
|
}
|
|
|
-EXPORT_SYMBOL_GPL(crypto_transfer_hash_request);
|
|
|
+EXPORT_SYMBOL_GPL(crypto_transfer_akcipher_request_to_engine);
|
|
|
|
|
|
/**
|
|
|
- * crypto_transfer_hash_request_to_engine - transfer one request to list
|
|
|
- * into the engine queue
|
|
|
+ * crypto_transfer_hash_request_to_engine - transfer one ahash_request
|
|
|
+ * to list into the engine queue
|
|
|
* @engine: the hardware engine
|
|
|
* @req: the request need to be listed into the engine queue
|
|
|
*/
|
|
|
int crypto_transfer_hash_request_to_engine(struct crypto_engine *engine,
|
|
|
struct ahash_request *req)
|
|
|
{
|
|
|
- return crypto_transfer_hash_request(engine, req, true);
|
|
|
+ return crypto_transfer_request_to_engine(engine, &req->base);
|
|
|
}
|
|
|
EXPORT_SYMBOL_GPL(crypto_transfer_hash_request_to_engine);
|
|
|
|
|
|
/**
|
|
|
- * crypto_finalize_cipher_request - finalize one request if the request is done
|
|
|
+ * crypto_transfer_skcipher_request_to_engine - transfer one skcipher_request
|
|
|
+ * to list into the engine queue
|
|
|
+ * @engine: the hardware engine
|
|
|
+ * @req: the request need to be listed into the engine queue
|
|
|
+ */
|
|
|
+int crypto_transfer_skcipher_request_to_engine(struct crypto_engine *engine,
|
|
|
+ struct skcipher_request *req)
|
|
|
+{
|
|
|
+ return crypto_transfer_request_to_engine(engine, &req->base);
|
|
|
+}
|
|
|
+EXPORT_SYMBOL_GPL(crypto_transfer_skcipher_request_to_engine);
|
|
|
+
|
|
|
+/**
|
|
|
+ * crypto_finalize_ablkcipher_request - finalize one ablkcipher_request if
|
|
|
+ * the request is done
|
|
|
* @engine: the hardware engine
|
|
|
* @req: the request need to be finalized
|
|
|
* @err: error number
|
|
|
+ * TODO: Remove this function when skcipher conversion is finished
|
|
|
*/
|
|
|
-void crypto_finalize_cipher_request(struct crypto_engine *engine,
|
|
|
- struct ablkcipher_request *req, int err)
|
|
|
+void crypto_finalize_ablkcipher_request(struct crypto_engine *engine,
|
|
|
+ struct ablkcipher_request *req, int err)
|
|
|
{
|
|
|
- unsigned long flags;
|
|
|
- bool finalize_cur_req = false;
|
|
|
- int ret;
|
|
|
-
|
|
|
- spin_lock_irqsave(&engine->queue_lock, flags);
|
|
|
- if (engine->cur_req == &req->base)
|
|
|
- finalize_cur_req = true;
|
|
|
- spin_unlock_irqrestore(&engine->queue_lock, flags);
|
|
|
-
|
|
|
- if (finalize_cur_req) {
|
|
|
- if (engine->cur_req_prepared &&
|
|
|
- engine->unprepare_cipher_request) {
|
|
|
- ret = engine->unprepare_cipher_request(engine, req);
|
|
|
- if (ret)
|
|
|
- dev_err(engine->dev, "failed to unprepare request\n");
|
|
|
- }
|
|
|
- spin_lock_irqsave(&engine->queue_lock, flags);
|
|
|
- engine->cur_req = NULL;
|
|
|
- engine->cur_req_prepared = false;
|
|
|
- spin_unlock_irqrestore(&engine->queue_lock, flags);
|
|
|
- }
|
|
|
+ return crypto_finalize_request(engine, &req->base, err);
|
|
|
+}
|
|
|
+EXPORT_SYMBOL_GPL(crypto_finalize_ablkcipher_request);
|
|
|
|
|
|
- req->base.complete(&req->base, err);
|
|
|
+/**
|
|
|
+ * crypto_finalize_aead_request - finalize one aead_request if
|
|
|
+ * the request is done
|
|
|
+ * @engine: the hardware engine
|
|
|
+ * @req: the request need to be finalized
|
|
|
+ * @err: error number
|
|
|
+ */
|
|
|
+void crypto_finalize_aead_request(struct crypto_engine *engine,
|
|
|
+ struct aead_request *req, int err)
|
|
|
+{
|
|
|
+ return crypto_finalize_request(engine, &req->base, err);
|
|
|
+}
|
|
|
+EXPORT_SYMBOL_GPL(crypto_finalize_aead_request);
|
|
|
|
|
|
- kthread_queue_work(engine->kworker, &engine->pump_requests);
|
|
|
+/**
|
|
|
+ * crypto_finalize_akcipher_request - finalize one akcipher_request if
|
|
|
+ * the request is done
|
|
|
+ * @engine: the hardware engine
|
|
|
+ * @req: the request need to be finalized
|
|
|
+ * @err: error number
|
|
|
+ */
|
|
|
+void crypto_finalize_akcipher_request(struct crypto_engine *engine,
|
|
|
+ struct akcipher_request *req, int err)
|
|
|
+{
|
|
|
+ return crypto_finalize_request(engine, &req->base, err);
|
|
|
}
|
|
|
-EXPORT_SYMBOL_GPL(crypto_finalize_cipher_request);
|
|
|
+EXPORT_SYMBOL_GPL(crypto_finalize_akcipher_request);
|
|
|
|
|
|
/**
|
|
|
- * crypto_finalize_hash_request - finalize one request if the request is done
|
|
|
+ * crypto_finalize_hash_request - finalize one ahash_request if
|
|
|
+ * the request is done
|
|
|
* @engine: the hardware engine
|
|
|
* @req: the request need to be finalized
|
|
|
* @err: error number
|
|
@@ -300,34 +337,24 @@ EXPORT_SYMBOL_GPL(crypto_finalize_cipher_request);
|
|
|
void crypto_finalize_hash_request(struct crypto_engine *engine,
|
|
|
struct ahash_request *req, int err)
|
|
|
{
|
|
|
- unsigned long flags;
|
|
|
- bool finalize_cur_req = false;
|
|
|
- int ret;
|
|
|
-
|
|
|
- spin_lock_irqsave(&engine->queue_lock, flags);
|
|
|
- if (engine->cur_req == &req->base)
|
|
|
- finalize_cur_req = true;
|
|
|
- spin_unlock_irqrestore(&engine->queue_lock, flags);
|
|
|
-
|
|
|
- if (finalize_cur_req) {
|
|
|
- if (engine->cur_req_prepared &&
|
|
|
- engine->unprepare_hash_request) {
|
|
|
- ret = engine->unprepare_hash_request(engine, req);
|
|
|
- if (ret)
|
|
|
- dev_err(engine->dev, "failed to unprepare request\n");
|
|
|
- }
|
|
|
- spin_lock_irqsave(&engine->queue_lock, flags);
|
|
|
- engine->cur_req = NULL;
|
|
|
- engine->cur_req_prepared = false;
|
|
|
- spin_unlock_irqrestore(&engine->queue_lock, flags);
|
|
|
- }
|
|
|
-
|
|
|
- req->base.complete(&req->base, err);
|
|
|
-
|
|
|
- kthread_queue_work(engine->kworker, &engine->pump_requests);
|
|
|
+ return crypto_finalize_request(engine, &req->base, err);
|
|
|
}
|
|
|
EXPORT_SYMBOL_GPL(crypto_finalize_hash_request);
|
|
|
|
|
|
+/**
|
|
|
+ * crypto_finalize_skcipher_request - finalize one skcipher_request if
|
|
|
+ * the request is done
|
|
|
+ * @engine: the hardware engine
|
|
|
+ * @req: the request need to be finalized
|
|
|
+ * @err: error number
|
|
|
+ */
|
|
|
+void crypto_finalize_skcipher_request(struct crypto_engine *engine,
|
|
|
+ struct skcipher_request *req, int err)
|
|
|
+{
|
|
|
+ return crypto_finalize_request(engine, &req->base, err);
|
|
|
+}
|
|
|
+EXPORT_SYMBOL_GPL(crypto_finalize_skcipher_request);
|
|
|
+
|
|
|
/**
|
|
|
* crypto_engine_start - start the hardware engine
|
|
|
* @engine: the hardware engine need to be started
|