Loading...
1/* SPDX-License-Identifier: GPL-2.0-or-later */
2/*
3 * Crypto engine API
4 *
5 * Copyright (c) 2016 Baolin Wang <baolin.wang@linaro.org>
6 */
7#ifndef _CRYPTO_ENGINE_H
8#define _CRYPTO_ENGINE_H
9
10#include <linux/crypto.h>
11#include <linux/list.h>
12#include <linux/kthread.h>
13#include <linux/spinlock.h>
14#include <linux/types.h>
15
16#include <crypto/algapi.h>
17#include <crypto/aead.h>
18#include <crypto/akcipher.h>
19#include <crypto/hash.h>
20#include <crypto/skcipher.h>
21#include <crypto/kpp.h>
22
23struct device;
24
25#define ENGINE_NAME_LEN 30
26/*
27 * struct crypto_engine - crypto hardware engine
28 * @name: the engine name
29 * @idling: the engine is entering idle state
30 * @busy: request pump is busy
31 * @running: the engine is on working
32 * @retry_support: indication that the hardware allows re-execution
33 * of a failed backlog request
34 * crypto-engine, in head position to keep order
35 * @list: link with the global crypto engine list
36 * @queue_lock: spinlock to synchronise access to request queue
37 * @queue: the crypto queue of the engine
38 * @rt: whether this queue is set to run as a realtime task
39 * @prepare_crypt_hardware: a request will soon arrive from the queue
40 * so the subsystem requests the driver to prepare the hardware
41 * by issuing this call
42 * @unprepare_crypt_hardware: there are currently no more requests on the
43 * queue so the subsystem notifies the driver that it may relax the
44 * hardware by issuing this call
45 * @do_batch_requests: execute a batch of requests. Depends on multiple
46 * requests support.
47 * @kworker: kthread worker struct for request pump
48 * @pump_requests: work struct for scheduling work to the request pump
49 * @priv_data: the engine private data
50 * @cur_req: the current request which is on processing
51 */
52struct crypto_engine {
53 char name[ENGINE_NAME_LEN];
54 bool idling;
55 bool busy;
56 bool running;
57
58 bool retry_support;
59
60 struct list_head list;
61 spinlock_t queue_lock;
62 struct crypto_queue queue;
63 struct device *dev;
64
65 bool rt;
66
67 int (*prepare_crypt_hardware)(struct crypto_engine *engine);
68 int (*unprepare_crypt_hardware)(struct crypto_engine *engine);
69 int (*do_batch_requests)(struct crypto_engine *engine);
70
71
72 struct kthread_worker *kworker;
73 struct kthread_work pump_requests;
74
75 void *priv_data;
76 struct crypto_async_request *cur_req;
77};
78
79/*
80 * struct crypto_engine_op - crypto hardware engine operations
81 * @prepare__request: do some prepare if need before handle the current request
82 * @unprepare_request: undo any work done by prepare_request()
83 * @do_one_request: do encryption for current request
84 */
85struct crypto_engine_op {
86 int (*prepare_request)(struct crypto_engine *engine,
87 void *areq);
88 int (*unprepare_request)(struct crypto_engine *engine,
89 void *areq);
90 int (*do_one_request)(struct crypto_engine *engine,
91 void *areq);
92};
93
94struct crypto_engine_ctx {
95 struct crypto_engine_op op;
96};
97
98int crypto_transfer_aead_request_to_engine(struct crypto_engine *engine,
99 struct aead_request *req);
100int crypto_transfer_akcipher_request_to_engine(struct crypto_engine *engine,
101 struct akcipher_request *req);
102int crypto_transfer_hash_request_to_engine(struct crypto_engine *engine,
103 struct ahash_request *req);
104int crypto_transfer_kpp_request_to_engine(struct crypto_engine *engine,
105 struct kpp_request *req);
106int crypto_transfer_skcipher_request_to_engine(struct crypto_engine *engine,
107 struct skcipher_request *req);
108void crypto_finalize_aead_request(struct crypto_engine *engine,
109 struct aead_request *req, int err);
110void crypto_finalize_akcipher_request(struct crypto_engine *engine,
111 struct akcipher_request *req, int err);
112void crypto_finalize_hash_request(struct crypto_engine *engine,
113 struct ahash_request *req, int err);
114void crypto_finalize_kpp_request(struct crypto_engine *engine,
115 struct kpp_request *req, int err);
116void crypto_finalize_skcipher_request(struct crypto_engine *engine,
117 struct skcipher_request *req, int err);
118int crypto_engine_start(struct crypto_engine *engine);
119int crypto_engine_stop(struct crypto_engine *engine);
120struct crypto_engine *crypto_engine_alloc_init(struct device *dev, bool rt);
121struct crypto_engine *crypto_engine_alloc_init_and_set(struct device *dev,
122 bool retry_support,
123 int (*cbk_do_batch)(struct crypto_engine *engine),
124 bool rt, int qlen);
125int crypto_engine_exit(struct crypto_engine *engine);
126
127#endif /* _CRYPTO_ENGINE_H */
1/*
2 * Crypto engine API
3 *
4 * Copyright (c) 2016 Baolin Wang <baolin.wang@linaro.org>
5 *
6 * This program is free software; you can redistribute it and/or modify it
7 * under the terms of the GNU General Public License as published by the Free
8 * Software Foundation; either version 2 of the License, or (at your option)
9 * any later version.
10 *
11 */
12#ifndef _CRYPTO_ENGINE_H
13#define _CRYPTO_ENGINE_H
14
15#include <linux/crypto.h>
16#include <linux/list.h>
17#include <linux/kernel.h>
18#include <linux/kthread.h>
19#include <crypto/algapi.h>
20#include <crypto/aead.h>
21#include <crypto/akcipher.h>
22#include <crypto/hash.h>
23#include <crypto/skcipher.h>
24
25#define ENGINE_NAME_LEN 30
26/*
27 * struct crypto_engine - crypto hardware engine
28 * @name: the engine name
29 * @idling: the engine is entering idle state
30 * @busy: request pump is busy
31 * @running: the engine is on working
32 * @cur_req_prepared: current request is prepared
33 * @list: link with the global crypto engine list
34 * @queue_lock: spinlock to syncronise access to request queue
35 * @queue: the crypto queue of the engine
36 * @rt: whether this queue is set to run as a realtime task
37 * @prepare_crypt_hardware: a request will soon arrive from the queue
38 * so the subsystem requests the driver to prepare the hardware
39 * by issuing this call
40 * @unprepare_crypt_hardware: there are currently no more requests on the
41 * queue so the subsystem notifies the driver that it may relax the
42 * hardware by issuing this call
43 * @kworker: kthread worker struct for request pump
44 * @pump_requests: work struct for scheduling work to the request pump
45 * @priv_data: the engine private data
46 * @cur_req: the current request which is on processing
47 */
48struct crypto_engine {
49 char name[ENGINE_NAME_LEN];
50 bool idling;
51 bool busy;
52 bool running;
53 bool cur_req_prepared;
54
55 struct list_head list;
56 spinlock_t queue_lock;
57 struct crypto_queue queue;
58 struct device *dev;
59
60 bool rt;
61
62 int (*prepare_crypt_hardware)(struct crypto_engine *engine);
63 int (*unprepare_crypt_hardware)(struct crypto_engine *engine);
64
65 struct kthread_worker *kworker;
66 struct kthread_work pump_requests;
67
68 void *priv_data;
69 struct crypto_async_request *cur_req;
70};
71
72/*
73 * struct crypto_engine_op - crypto hardware engine operations
74 * @prepare__request: do some prepare if need before handle the current request
75 * @unprepare_request: undo any work done by prepare_request()
76 * @do_one_request: do encryption for current request
77 */
78struct crypto_engine_op {
79 int (*prepare_request)(struct crypto_engine *engine,
80 void *areq);
81 int (*unprepare_request)(struct crypto_engine *engine,
82 void *areq);
83 int (*do_one_request)(struct crypto_engine *engine,
84 void *areq);
85};
86
87struct crypto_engine_ctx {
88 struct crypto_engine_op op;
89};
90
91int crypto_transfer_ablkcipher_request_to_engine(struct crypto_engine *engine,
92 struct ablkcipher_request *req);
93int crypto_transfer_aead_request_to_engine(struct crypto_engine *engine,
94 struct aead_request *req);
95int crypto_transfer_akcipher_request_to_engine(struct crypto_engine *engine,
96 struct akcipher_request *req);
97int crypto_transfer_hash_request_to_engine(struct crypto_engine *engine,
98 struct ahash_request *req);
99int crypto_transfer_skcipher_request_to_engine(struct crypto_engine *engine,
100 struct skcipher_request *req);
101void crypto_finalize_ablkcipher_request(struct crypto_engine *engine,
102 struct ablkcipher_request *req, int err);
103void crypto_finalize_aead_request(struct crypto_engine *engine,
104 struct aead_request *req, int err);
105void crypto_finalize_akcipher_request(struct crypto_engine *engine,
106 struct akcipher_request *req, int err);
107void crypto_finalize_hash_request(struct crypto_engine *engine,
108 struct ahash_request *req, int err);
109void crypto_finalize_skcipher_request(struct crypto_engine *engine,
110 struct skcipher_request *req, int err);
111int crypto_engine_start(struct crypto_engine *engine);
112int crypto_engine_stop(struct crypto_engine *engine);
113struct crypto_engine *crypto_engine_alloc_init(struct device *dev, bool rt);
114int crypto_engine_exit(struct crypto_engine *engine);
115
116#endif /* _CRYPTO_ENGINE_H */