Loading...
1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * Cryptographic API.
4 *
5 * Serpent Cipher Algorithm.
6 *
7 * Copyright (C) 2002 Dag Arne Osvik <osvik@ii.uib.no>
8 * 2003 Herbert Valerio Riedel <hvr@gnu.org>
9 *
10 * Added tnepres support:
11 * Ruben Jesus Garcia Hernandez <ruben@ugr.es>, 18.10.2004
12 * Based on code by hvr
13 */
14
15#include <linux/init.h>
16#include <linux/module.h>
17#include <linux/errno.h>
18#include <asm/byteorder.h>
19#include <linux/crypto.h>
20#include <linux/types.h>
21#include <crypto/serpent.h>
22
23/* Key is padded to the maximum of 256 bits before round key generation.
24 * Any key length <= 256 bits (32 bytes) is allowed by the algorithm.
25 */
26
27#define PHI 0x9e3779b9UL
28
29#define keyiter(a, b, c, d, i, j) \
30 ({ b ^= d; b ^= c; b ^= a; b ^= PHI ^ i; b = rol32(b, 11); k[j] = b; })
31
32#define loadkeys(x0, x1, x2, x3, i) \
33 ({ x0 = k[i]; x1 = k[i+1]; x2 = k[i+2]; x3 = k[i+3]; })
34
35#define storekeys(x0, x1, x2, x3, i) \
36 ({ k[i] = x0; k[i+1] = x1; k[i+2] = x2; k[i+3] = x3; })
37
38#define store_and_load_keys(x0, x1, x2, x3, s, l) \
39 ({ storekeys(x0, x1, x2, x3, s); loadkeys(x0, x1, x2, x3, l); })
40
41#define K(x0, x1, x2, x3, i) ({ \
42 x3 ^= k[4*(i)+3]; x2 ^= k[4*(i)+2]; \
43 x1 ^= k[4*(i)+1]; x0 ^= k[4*(i)+0]; \
44 })
45
46#define LK(x0, x1, x2, x3, x4, i) ({ \
47 x0 = rol32(x0, 13);\
48 x2 = rol32(x2, 3); x1 ^= x0; x4 = x0 << 3; \
49 x3 ^= x2; x1 ^= x2; \
50 x1 = rol32(x1, 1); x3 ^= x4; \
51 x3 = rol32(x3, 7); x4 = x1; \
52 x0 ^= x1; x4 <<= 7; x2 ^= x3; \
53 x0 ^= x3; x2 ^= x4; x3 ^= k[4*i+3]; \
54 x1 ^= k[4*i+1]; x0 = rol32(x0, 5); x2 = rol32(x2, 22);\
55 x0 ^= k[4*i+0]; x2 ^= k[4*i+2]; \
56 })
57
58#define KL(x0, x1, x2, x3, x4, i) ({ \
59 x0 ^= k[4*i+0]; x1 ^= k[4*i+1]; x2 ^= k[4*i+2]; \
60 x3 ^= k[4*i+3]; x0 = ror32(x0, 5); x2 = ror32(x2, 22);\
61 x4 = x1; x2 ^= x3; x0 ^= x3; \
62 x4 <<= 7; x0 ^= x1; x1 = ror32(x1, 1); \
63 x2 ^= x4; x3 = ror32(x3, 7); x4 = x0 << 3; \
64 x1 ^= x0; x3 ^= x4; x0 = ror32(x0, 13);\
65 x1 ^= x2; x3 ^= x2; x2 = ror32(x2, 3); \
66 })
67
68#define S0(x0, x1, x2, x3, x4) ({ \
69 x4 = x3; \
70 x3 |= x0; x0 ^= x4; x4 ^= x2; \
71 x4 = ~x4; x3 ^= x1; x1 &= x0; \
72 x1 ^= x4; x2 ^= x0; x0 ^= x3; \
73 x4 |= x0; x0 ^= x2; x2 &= x1; \
74 x3 ^= x2; x1 = ~x1; x2 ^= x4; \
75 x1 ^= x2; \
76 })
77
78#define S1(x0, x1, x2, x3, x4) ({ \
79 x4 = x1; \
80 x1 ^= x0; x0 ^= x3; x3 = ~x3; \
81 x4 &= x1; x0 |= x1; x3 ^= x2; \
82 x0 ^= x3; x1 ^= x3; x3 ^= x4; \
83 x1 |= x4; x4 ^= x2; x2 &= x0; \
84 x2 ^= x1; x1 |= x0; x0 = ~x0; \
85 x0 ^= x2; x4 ^= x1; \
86 })
87
88#define S2(x0, x1, x2, x3, x4) ({ \
89 x3 = ~x3; \
90 x1 ^= x0; x4 = x0; x0 &= x2; \
91 x0 ^= x3; x3 |= x4; x2 ^= x1; \
92 x3 ^= x1; x1 &= x0; x0 ^= x2; \
93 x2 &= x3; x3 |= x1; x0 = ~x0; \
94 x3 ^= x0; x4 ^= x0; x0 ^= x2; \
95 x1 |= x2; \
96 })
97
98#define S3(x0, x1, x2, x3, x4) ({ \
99 x4 = x1; \
100 x1 ^= x3; x3 |= x0; x4 &= x0; \
101 x0 ^= x2; x2 ^= x1; x1 &= x3; \
102 x2 ^= x3; x0 |= x4; x4 ^= x3; \
103 x1 ^= x0; x0 &= x3; x3 &= x4; \
104 x3 ^= x2; x4 |= x1; x2 &= x1; \
105 x4 ^= x3; x0 ^= x3; x3 ^= x2; \
106 })
107
108#define S4(x0, x1, x2, x3, x4) ({ \
109 x4 = x3; \
110 x3 &= x0; x0 ^= x4; \
111 x3 ^= x2; x2 |= x4; x0 ^= x1; \
112 x4 ^= x3; x2 |= x0; \
113 x2 ^= x1; x1 &= x0; \
114 x1 ^= x4; x4 &= x2; x2 ^= x3; \
115 x4 ^= x0; x3 |= x1; x1 = ~x1; \
116 x3 ^= x0; \
117 })
118
119#define S5(x0, x1, x2, x3, x4) ({ \
120 x4 = x1; x1 |= x0; \
121 x2 ^= x1; x3 = ~x3; x4 ^= x0; \
122 x0 ^= x2; x1 &= x4; x4 |= x3; \
123 x4 ^= x0; x0 &= x3; x1 ^= x3; \
124 x3 ^= x2; x0 ^= x1; x2 &= x4; \
125 x1 ^= x2; x2 &= x0; \
126 x3 ^= x2; \
127 })
128
129#define S6(x0, x1, x2, x3, x4) ({ \
130 x4 = x1; \
131 x3 ^= x0; x1 ^= x2; x2 ^= x0; \
132 x0 &= x3; x1 |= x3; x4 = ~x4; \
133 x0 ^= x1; x1 ^= x2; \
134 x3 ^= x4; x4 ^= x0; x2 &= x0; \
135 x4 ^= x1; x2 ^= x3; x3 &= x1; \
136 x3 ^= x0; x1 ^= x2; \
137 })
138
139#define S7(x0, x1, x2, x3, x4) ({ \
140 x1 = ~x1; \
141 x4 = x1; x0 = ~x0; x1 &= x2; \
142 x1 ^= x3; x3 |= x4; x4 ^= x2; \
143 x2 ^= x3; x3 ^= x0; x0 |= x1; \
144 x2 &= x0; x0 ^= x4; x4 ^= x3; \
145 x3 &= x0; x4 ^= x1; \
146 x2 ^= x4; x3 ^= x1; x4 |= x0; \
147 x4 ^= x1; \
148 })
149
150#define SI0(x0, x1, x2, x3, x4) ({ \
151 x4 = x3; x1 ^= x0; \
152 x3 |= x1; x4 ^= x1; x0 = ~x0; \
153 x2 ^= x3; x3 ^= x0; x0 &= x1; \
154 x0 ^= x2; x2 &= x3; x3 ^= x4; \
155 x2 ^= x3; x1 ^= x3; x3 &= x0; \
156 x1 ^= x0; x0 ^= x2; x4 ^= x3; \
157 })
158
159#define SI1(x0, x1, x2, x3, x4) ({ \
160 x1 ^= x3; x4 = x0; \
161 x0 ^= x2; x2 = ~x2; x4 |= x1; \
162 x4 ^= x3; x3 &= x1; x1 ^= x2; \
163 x2 &= x4; x4 ^= x1; x1 |= x3; \
164 x3 ^= x0; x2 ^= x0; x0 |= x4; \
165 x2 ^= x4; x1 ^= x0; \
166 x4 ^= x1; \
167 })
168
169#define SI2(x0, x1, x2, x3, x4) ({ \
170 x2 ^= x1; x4 = x3; x3 = ~x3; \
171 x3 |= x2; x2 ^= x4; x4 ^= x0; \
172 x3 ^= x1; x1 |= x2; x2 ^= x0; \
173 x1 ^= x4; x4 |= x3; x2 ^= x3; \
174 x4 ^= x2; x2 &= x1; \
175 x2 ^= x3; x3 ^= x4; x4 ^= x0; \
176 })
177
178#define SI3(x0, x1, x2, x3, x4) ({ \
179 x2 ^= x1; \
180 x4 = x1; x1 &= x2; \
181 x1 ^= x0; x0 |= x4; x4 ^= x3; \
182 x0 ^= x3; x3 |= x1; x1 ^= x2; \
183 x1 ^= x3; x0 ^= x2; x2 ^= x3; \
184 x3 &= x1; x1 ^= x0; x0 &= x2; \
185 x4 ^= x3; x3 ^= x0; x0 ^= x1; \
186 })
187
188#define SI4(x0, x1, x2, x3, x4) ({ \
189 x2 ^= x3; x4 = x0; x0 &= x1; \
190 x0 ^= x2; x2 |= x3; x4 = ~x4; \
191 x1 ^= x0; x0 ^= x2; x2 &= x4; \
192 x2 ^= x0; x0 |= x4; \
193 x0 ^= x3; x3 &= x2; \
194 x4 ^= x3; x3 ^= x1; x1 &= x0; \
195 x4 ^= x1; x0 ^= x3; \
196 })
197
198#define SI5(x0, x1, x2, x3, x4) ({ \
199 x4 = x1; x1 |= x2; \
200 x2 ^= x4; x1 ^= x3; x3 &= x4; \
201 x2 ^= x3; x3 |= x0; x0 = ~x0; \
202 x3 ^= x2; x2 |= x0; x4 ^= x1; \
203 x2 ^= x4; x4 &= x0; x0 ^= x1; \
204 x1 ^= x3; x0 &= x2; x2 ^= x3; \
205 x0 ^= x2; x2 ^= x4; x4 ^= x3; \
206 })
207
208#define SI6(x0, x1, x2, x3, x4) ({ \
209 x0 ^= x2; \
210 x4 = x0; x0 &= x3; x2 ^= x3; \
211 x0 ^= x2; x3 ^= x1; x2 |= x4; \
212 x2 ^= x3; x3 &= x0; x0 = ~x0; \
213 x3 ^= x1; x1 &= x2; x4 ^= x0; \
214 x3 ^= x4; x4 ^= x2; x0 ^= x1; \
215 x2 ^= x0; \
216 })
217
218#define SI7(x0, x1, x2, x3, x4) ({ \
219 x4 = x3; x3 &= x0; x0 ^= x2; \
220 x2 |= x4; x4 ^= x1; x0 = ~x0; \
221 x1 |= x3; x4 ^= x0; x0 &= x2; \
222 x0 ^= x1; x1 &= x2; x3 ^= x2; \
223 x4 ^= x3; x2 &= x3; x3 |= x0; \
224 x1 ^= x4; x3 ^= x4; x4 &= x0; \
225 x4 ^= x2; \
226 })
227
228/*
229 * both gcc and clang have misoptimized this function in the past,
230 * producing horrible object code from spilling temporary variables
231 * on the stack. Forcing this part out of line avoids that.
232 */
233static noinline void __serpent_setkey_sbox(u32 r0, u32 r1, u32 r2,
234 u32 r3, u32 r4, u32 *k)
235{
236 k += 100;
237 S3(r3, r4, r0, r1, r2); store_and_load_keys(r1, r2, r4, r3, 28, 24);
238 S4(r1, r2, r4, r3, r0); store_and_load_keys(r2, r4, r3, r0, 24, 20);
239 S5(r2, r4, r3, r0, r1); store_and_load_keys(r1, r2, r4, r0, 20, 16);
240 S6(r1, r2, r4, r0, r3); store_and_load_keys(r4, r3, r2, r0, 16, 12);
241 S7(r4, r3, r2, r0, r1); store_and_load_keys(r1, r2, r0, r4, 12, 8);
242 S0(r1, r2, r0, r4, r3); store_and_load_keys(r0, r2, r4, r1, 8, 4);
243 S1(r0, r2, r4, r1, r3); store_and_load_keys(r3, r4, r1, r0, 4, 0);
244 S2(r3, r4, r1, r0, r2); store_and_load_keys(r2, r4, r3, r0, 0, -4);
245 S3(r2, r4, r3, r0, r1); store_and_load_keys(r0, r1, r4, r2, -4, -8);
246 S4(r0, r1, r4, r2, r3); store_and_load_keys(r1, r4, r2, r3, -8, -12);
247 S5(r1, r4, r2, r3, r0); store_and_load_keys(r0, r1, r4, r3, -12, -16);
248 S6(r0, r1, r4, r3, r2); store_and_load_keys(r4, r2, r1, r3, -16, -20);
249 S7(r4, r2, r1, r3, r0); store_and_load_keys(r0, r1, r3, r4, -20, -24);
250 S0(r0, r1, r3, r4, r2); store_and_load_keys(r3, r1, r4, r0, -24, -28);
251 k -= 50;
252 S1(r3, r1, r4, r0, r2); store_and_load_keys(r2, r4, r0, r3, 22, 18);
253 S2(r2, r4, r0, r3, r1); store_and_load_keys(r1, r4, r2, r3, 18, 14);
254 S3(r1, r4, r2, r3, r0); store_and_load_keys(r3, r0, r4, r1, 14, 10);
255 S4(r3, r0, r4, r1, r2); store_and_load_keys(r0, r4, r1, r2, 10, 6);
256 S5(r0, r4, r1, r2, r3); store_and_load_keys(r3, r0, r4, r2, 6, 2);
257 S6(r3, r0, r4, r2, r1); store_and_load_keys(r4, r1, r0, r2, 2, -2);
258 S7(r4, r1, r0, r2, r3); store_and_load_keys(r3, r0, r2, r4, -2, -6);
259 S0(r3, r0, r2, r4, r1); store_and_load_keys(r2, r0, r4, r3, -6, -10);
260 S1(r2, r0, r4, r3, r1); store_and_load_keys(r1, r4, r3, r2, -10, -14);
261 S2(r1, r4, r3, r2, r0); store_and_load_keys(r0, r4, r1, r2, -14, -18);
262 S3(r0, r4, r1, r2, r3); store_and_load_keys(r2, r3, r4, r0, -18, -22);
263 k -= 50;
264 S4(r2, r3, r4, r0, r1); store_and_load_keys(r3, r4, r0, r1, 28, 24);
265 S5(r3, r4, r0, r1, r2); store_and_load_keys(r2, r3, r4, r1, 24, 20);
266 S6(r2, r3, r4, r1, r0); store_and_load_keys(r4, r0, r3, r1, 20, 16);
267 S7(r4, r0, r3, r1, r2); store_and_load_keys(r2, r3, r1, r4, 16, 12);
268 S0(r2, r3, r1, r4, r0); store_and_load_keys(r1, r3, r4, r2, 12, 8);
269 S1(r1, r3, r4, r2, r0); store_and_load_keys(r0, r4, r2, r1, 8, 4);
270 S2(r0, r4, r2, r1, r3); store_and_load_keys(r3, r4, r0, r1, 4, 0);
271 S3(r3, r4, r0, r1, r2); storekeys(r1, r2, r4, r3, 0);
272}
273
274int __serpent_setkey(struct serpent_ctx *ctx, const u8 *key,
275 unsigned int keylen)
276{
277 u32 *k = ctx->expkey;
278 u8 *k8 = (u8 *)k;
279 u32 r0, r1, r2, r3, r4;
280 int i;
281
282 /* Copy key, add padding */
283
284 for (i = 0; i < keylen; ++i)
285 k8[i] = key[i];
286 if (i < SERPENT_MAX_KEY_SIZE)
287 k8[i++] = 1;
288 while (i < SERPENT_MAX_KEY_SIZE)
289 k8[i++] = 0;
290
291 /* Expand key using polynomial */
292
293 r0 = le32_to_cpu(k[3]);
294 r1 = le32_to_cpu(k[4]);
295 r2 = le32_to_cpu(k[5]);
296 r3 = le32_to_cpu(k[6]);
297 r4 = le32_to_cpu(k[7]);
298
299 keyiter(le32_to_cpu(k[0]), r0, r4, r2, 0, 0);
300 keyiter(le32_to_cpu(k[1]), r1, r0, r3, 1, 1);
301 keyiter(le32_to_cpu(k[2]), r2, r1, r4, 2, 2);
302 keyiter(le32_to_cpu(k[3]), r3, r2, r0, 3, 3);
303 keyiter(le32_to_cpu(k[4]), r4, r3, r1, 4, 4);
304 keyiter(le32_to_cpu(k[5]), r0, r4, r2, 5, 5);
305 keyiter(le32_to_cpu(k[6]), r1, r0, r3, 6, 6);
306 keyiter(le32_to_cpu(k[7]), r2, r1, r4, 7, 7);
307
308 keyiter(k[0], r3, r2, r0, 8, 8);
309 keyiter(k[1], r4, r3, r1, 9, 9);
310 keyiter(k[2], r0, r4, r2, 10, 10);
311 keyiter(k[3], r1, r0, r3, 11, 11);
312 keyiter(k[4], r2, r1, r4, 12, 12);
313 keyiter(k[5], r3, r2, r0, 13, 13);
314 keyiter(k[6], r4, r3, r1, 14, 14);
315 keyiter(k[7], r0, r4, r2, 15, 15);
316 keyiter(k[8], r1, r0, r3, 16, 16);
317 keyiter(k[9], r2, r1, r4, 17, 17);
318 keyiter(k[10], r3, r2, r0, 18, 18);
319 keyiter(k[11], r4, r3, r1, 19, 19);
320 keyiter(k[12], r0, r4, r2, 20, 20);
321 keyiter(k[13], r1, r0, r3, 21, 21);
322 keyiter(k[14], r2, r1, r4, 22, 22);
323 keyiter(k[15], r3, r2, r0, 23, 23);
324 keyiter(k[16], r4, r3, r1, 24, 24);
325 keyiter(k[17], r0, r4, r2, 25, 25);
326 keyiter(k[18], r1, r0, r3, 26, 26);
327 keyiter(k[19], r2, r1, r4, 27, 27);
328 keyiter(k[20], r3, r2, r0, 28, 28);
329 keyiter(k[21], r4, r3, r1, 29, 29);
330 keyiter(k[22], r0, r4, r2, 30, 30);
331 keyiter(k[23], r1, r0, r3, 31, 31);
332
333 k += 50;
334
335 keyiter(k[-26], r2, r1, r4, 32, -18);
336 keyiter(k[-25], r3, r2, r0, 33, -17);
337 keyiter(k[-24], r4, r3, r1, 34, -16);
338 keyiter(k[-23], r0, r4, r2, 35, -15);
339 keyiter(k[-22], r1, r0, r3, 36, -14);
340 keyiter(k[-21], r2, r1, r4, 37, -13);
341 keyiter(k[-20], r3, r2, r0, 38, -12);
342 keyiter(k[-19], r4, r3, r1, 39, -11);
343 keyiter(k[-18], r0, r4, r2, 40, -10);
344 keyiter(k[-17], r1, r0, r3, 41, -9);
345 keyiter(k[-16], r2, r1, r4, 42, -8);
346 keyiter(k[-15], r3, r2, r0, 43, -7);
347 keyiter(k[-14], r4, r3, r1, 44, -6);
348 keyiter(k[-13], r0, r4, r2, 45, -5);
349 keyiter(k[-12], r1, r0, r3, 46, -4);
350 keyiter(k[-11], r2, r1, r4, 47, -3);
351 keyiter(k[-10], r3, r2, r0, 48, -2);
352 keyiter(k[-9], r4, r3, r1, 49, -1);
353 keyiter(k[-8], r0, r4, r2, 50, 0);
354 keyiter(k[-7], r1, r0, r3, 51, 1);
355 keyiter(k[-6], r2, r1, r4, 52, 2);
356 keyiter(k[-5], r3, r2, r0, 53, 3);
357 keyiter(k[-4], r4, r3, r1, 54, 4);
358 keyiter(k[-3], r0, r4, r2, 55, 5);
359 keyiter(k[-2], r1, r0, r3, 56, 6);
360 keyiter(k[-1], r2, r1, r4, 57, 7);
361 keyiter(k[0], r3, r2, r0, 58, 8);
362 keyiter(k[1], r4, r3, r1, 59, 9);
363 keyiter(k[2], r0, r4, r2, 60, 10);
364 keyiter(k[3], r1, r0, r3, 61, 11);
365 keyiter(k[4], r2, r1, r4, 62, 12);
366 keyiter(k[5], r3, r2, r0, 63, 13);
367 keyiter(k[6], r4, r3, r1, 64, 14);
368 keyiter(k[7], r0, r4, r2, 65, 15);
369 keyiter(k[8], r1, r0, r3, 66, 16);
370 keyiter(k[9], r2, r1, r4, 67, 17);
371 keyiter(k[10], r3, r2, r0, 68, 18);
372 keyiter(k[11], r4, r3, r1, 69, 19);
373 keyiter(k[12], r0, r4, r2, 70, 20);
374 keyiter(k[13], r1, r0, r3, 71, 21);
375 keyiter(k[14], r2, r1, r4, 72, 22);
376 keyiter(k[15], r3, r2, r0, 73, 23);
377 keyiter(k[16], r4, r3, r1, 74, 24);
378 keyiter(k[17], r0, r4, r2, 75, 25);
379 keyiter(k[18], r1, r0, r3, 76, 26);
380 keyiter(k[19], r2, r1, r4, 77, 27);
381 keyiter(k[20], r3, r2, r0, 78, 28);
382 keyiter(k[21], r4, r3, r1, 79, 29);
383 keyiter(k[22], r0, r4, r2, 80, 30);
384 keyiter(k[23], r1, r0, r3, 81, 31);
385
386 k += 50;
387
388 keyiter(k[-26], r2, r1, r4, 82, -18);
389 keyiter(k[-25], r3, r2, r0, 83, -17);
390 keyiter(k[-24], r4, r3, r1, 84, -16);
391 keyiter(k[-23], r0, r4, r2, 85, -15);
392 keyiter(k[-22], r1, r0, r3, 86, -14);
393 keyiter(k[-21], r2, r1, r4, 87, -13);
394 keyiter(k[-20], r3, r2, r0, 88, -12);
395 keyiter(k[-19], r4, r3, r1, 89, -11);
396 keyiter(k[-18], r0, r4, r2, 90, -10);
397 keyiter(k[-17], r1, r0, r3, 91, -9);
398 keyiter(k[-16], r2, r1, r4, 92, -8);
399 keyiter(k[-15], r3, r2, r0, 93, -7);
400 keyiter(k[-14], r4, r3, r1, 94, -6);
401 keyiter(k[-13], r0, r4, r2, 95, -5);
402 keyiter(k[-12], r1, r0, r3, 96, -4);
403 keyiter(k[-11], r2, r1, r4, 97, -3);
404 keyiter(k[-10], r3, r2, r0, 98, -2);
405 keyiter(k[-9], r4, r3, r1, 99, -1);
406 keyiter(k[-8], r0, r4, r2, 100, 0);
407 keyiter(k[-7], r1, r0, r3, 101, 1);
408 keyiter(k[-6], r2, r1, r4, 102, 2);
409 keyiter(k[-5], r3, r2, r0, 103, 3);
410 keyiter(k[-4], r4, r3, r1, 104, 4);
411 keyiter(k[-3], r0, r4, r2, 105, 5);
412 keyiter(k[-2], r1, r0, r3, 106, 6);
413 keyiter(k[-1], r2, r1, r4, 107, 7);
414 keyiter(k[0], r3, r2, r0, 108, 8);
415 keyiter(k[1], r4, r3, r1, 109, 9);
416 keyiter(k[2], r0, r4, r2, 110, 10);
417 keyiter(k[3], r1, r0, r3, 111, 11);
418 keyiter(k[4], r2, r1, r4, 112, 12);
419 keyiter(k[5], r3, r2, r0, 113, 13);
420 keyiter(k[6], r4, r3, r1, 114, 14);
421 keyiter(k[7], r0, r4, r2, 115, 15);
422 keyiter(k[8], r1, r0, r3, 116, 16);
423 keyiter(k[9], r2, r1, r4, 117, 17);
424 keyiter(k[10], r3, r2, r0, 118, 18);
425 keyiter(k[11], r4, r3, r1, 119, 19);
426 keyiter(k[12], r0, r4, r2, 120, 20);
427 keyiter(k[13], r1, r0, r3, 121, 21);
428 keyiter(k[14], r2, r1, r4, 122, 22);
429 keyiter(k[15], r3, r2, r0, 123, 23);
430 keyiter(k[16], r4, r3, r1, 124, 24);
431 keyiter(k[17], r0, r4, r2, 125, 25);
432 keyiter(k[18], r1, r0, r3, 126, 26);
433 keyiter(k[19], r2, r1, r4, 127, 27);
434 keyiter(k[20], r3, r2, r0, 128, 28);
435 keyiter(k[21], r4, r3, r1, 129, 29);
436 keyiter(k[22], r0, r4, r2, 130, 30);
437 keyiter(k[23], r1, r0, r3, 131, 31);
438
439 /* Apply S-boxes */
440 __serpent_setkey_sbox(r0, r1, r2, r3, r4, ctx->expkey);
441
442 return 0;
443}
444EXPORT_SYMBOL_GPL(__serpent_setkey);
445
446int serpent_setkey(struct crypto_tfm *tfm, const u8 *key, unsigned int keylen)
447{
448 return __serpent_setkey(crypto_tfm_ctx(tfm), key, keylen);
449}
450EXPORT_SYMBOL_GPL(serpent_setkey);
451
452void __serpent_encrypt(const void *c, u8 *dst, const u8 *src)
453{
454 const struct serpent_ctx *ctx = c;
455 const u32 *k = ctx->expkey;
456 const __le32 *s = (const __le32 *)src;
457 __le32 *d = (__le32 *)dst;
458 u32 r0, r1, r2, r3, r4;
459
460/*
461 * Note: The conversions between u8* and u32* might cause trouble
462 * on architectures with stricter alignment rules than x86
463 */
464
465 r0 = le32_to_cpu(s[0]);
466 r1 = le32_to_cpu(s[1]);
467 r2 = le32_to_cpu(s[2]);
468 r3 = le32_to_cpu(s[3]);
469
470 K(r0, r1, r2, r3, 0);
471 S0(r0, r1, r2, r3, r4); LK(r2, r1, r3, r0, r4, 1);
472 S1(r2, r1, r3, r0, r4); LK(r4, r3, r0, r2, r1, 2);
473 S2(r4, r3, r0, r2, r1); LK(r1, r3, r4, r2, r0, 3);
474 S3(r1, r3, r4, r2, r0); LK(r2, r0, r3, r1, r4, 4);
475 S4(r2, r0, r3, r1, r4); LK(r0, r3, r1, r4, r2, 5);
476 S5(r0, r3, r1, r4, r2); LK(r2, r0, r3, r4, r1, 6);
477 S6(r2, r0, r3, r4, r1); LK(r3, r1, r0, r4, r2, 7);
478 S7(r3, r1, r0, r4, r2); LK(r2, r0, r4, r3, r1, 8);
479 S0(r2, r0, r4, r3, r1); LK(r4, r0, r3, r2, r1, 9);
480 S1(r4, r0, r3, r2, r1); LK(r1, r3, r2, r4, r0, 10);
481 S2(r1, r3, r2, r4, r0); LK(r0, r3, r1, r4, r2, 11);
482 S3(r0, r3, r1, r4, r2); LK(r4, r2, r3, r0, r1, 12);
483 S4(r4, r2, r3, r0, r1); LK(r2, r3, r0, r1, r4, 13);
484 S5(r2, r3, r0, r1, r4); LK(r4, r2, r3, r1, r0, 14);
485 S6(r4, r2, r3, r1, r0); LK(r3, r0, r2, r1, r4, 15);
486 S7(r3, r0, r2, r1, r4); LK(r4, r2, r1, r3, r0, 16);
487 S0(r4, r2, r1, r3, r0); LK(r1, r2, r3, r4, r0, 17);
488 S1(r1, r2, r3, r4, r0); LK(r0, r3, r4, r1, r2, 18);
489 S2(r0, r3, r4, r1, r2); LK(r2, r3, r0, r1, r4, 19);
490 S3(r2, r3, r0, r1, r4); LK(r1, r4, r3, r2, r0, 20);
491 S4(r1, r4, r3, r2, r0); LK(r4, r3, r2, r0, r1, 21);
492 S5(r4, r3, r2, r0, r1); LK(r1, r4, r3, r0, r2, 22);
493 S6(r1, r4, r3, r0, r2); LK(r3, r2, r4, r0, r1, 23);
494 S7(r3, r2, r4, r0, r1); LK(r1, r4, r0, r3, r2, 24);
495 S0(r1, r4, r0, r3, r2); LK(r0, r4, r3, r1, r2, 25);
496 S1(r0, r4, r3, r1, r2); LK(r2, r3, r1, r0, r4, 26);
497 S2(r2, r3, r1, r0, r4); LK(r4, r3, r2, r0, r1, 27);
498 S3(r4, r3, r2, r0, r1); LK(r0, r1, r3, r4, r2, 28);
499 S4(r0, r1, r3, r4, r2); LK(r1, r3, r4, r2, r0, 29);
500 S5(r1, r3, r4, r2, r0); LK(r0, r1, r3, r2, r4, 30);
501 S6(r0, r1, r3, r2, r4); LK(r3, r4, r1, r2, r0, 31);
502 S7(r3, r4, r1, r2, r0); K(r0, r1, r2, r3, 32);
503
504 d[0] = cpu_to_le32(r0);
505 d[1] = cpu_to_le32(r1);
506 d[2] = cpu_to_le32(r2);
507 d[3] = cpu_to_le32(r3);
508}
509EXPORT_SYMBOL_GPL(__serpent_encrypt);
510
511static void serpent_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
512{
513 struct serpent_ctx *ctx = crypto_tfm_ctx(tfm);
514
515 __serpent_encrypt(ctx, dst, src);
516}
517
518void __serpent_decrypt(const void *c, u8 *dst, const u8 *src)
519{
520 const struct serpent_ctx *ctx = c;
521 const u32 *k = ctx->expkey;
522 const __le32 *s = (const __le32 *)src;
523 __le32 *d = (__le32 *)dst;
524 u32 r0, r1, r2, r3, r4;
525
526 r0 = le32_to_cpu(s[0]);
527 r1 = le32_to_cpu(s[1]);
528 r2 = le32_to_cpu(s[2]);
529 r3 = le32_to_cpu(s[3]);
530
531 K(r0, r1, r2, r3, 32);
532 SI7(r0, r1, r2, r3, r4); KL(r1, r3, r0, r4, r2, 31);
533 SI6(r1, r3, r0, r4, r2); KL(r0, r2, r4, r1, r3, 30);
534 SI5(r0, r2, r4, r1, r3); KL(r2, r3, r0, r4, r1, 29);
535 SI4(r2, r3, r0, r4, r1); KL(r2, r0, r1, r4, r3, 28);
536 SI3(r2, r0, r1, r4, r3); KL(r1, r2, r3, r4, r0, 27);
537 SI2(r1, r2, r3, r4, r0); KL(r2, r0, r4, r3, r1, 26);
538 SI1(r2, r0, r4, r3, r1); KL(r1, r0, r4, r3, r2, 25);
539 SI0(r1, r0, r4, r3, r2); KL(r4, r2, r0, r1, r3, 24);
540 SI7(r4, r2, r0, r1, r3); KL(r2, r1, r4, r3, r0, 23);
541 SI6(r2, r1, r4, r3, r0); KL(r4, r0, r3, r2, r1, 22);
542 SI5(r4, r0, r3, r2, r1); KL(r0, r1, r4, r3, r2, 21);
543 SI4(r0, r1, r4, r3, r2); KL(r0, r4, r2, r3, r1, 20);
544 SI3(r0, r4, r2, r3, r1); KL(r2, r0, r1, r3, r4, 19);
545 SI2(r2, r0, r1, r3, r4); KL(r0, r4, r3, r1, r2, 18);
546 SI1(r0, r4, r3, r1, r2); KL(r2, r4, r3, r1, r0, 17);
547 SI0(r2, r4, r3, r1, r0); KL(r3, r0, r4, r2, r1, 16);
548 SI7(r3, r0, r4, r2, r1); KL(r0, r2, r3, r1, r4, 15);
549 SI6(r0, r2, r3, r1, r4); KL(r3, r4, r1, r0, r2, 14);
550 SI5(r3, r4, r1, r0, r2); KL(r4, r2, r3, r1, r0, 13);
551 SI4(r4, r2, r3, r1, r0); KL(r4, r3, r0, r1, r2, 12);
552 SI3(r4, r3, r0, r1, r2); KL(r0, r4, r2, r1, r3, 11);
553 SI2(r0, r4, r2, r1, r3); KL(r4, r3, r1, r2, r0, 10);
554 SI1(r4, r3, r1, r2, r0); KL(r0, r3, r1, r2, r4, 9);
555 SI0(r0, r3, r1, r2, r4); KL(r1, r4, r3, r0, r2, 8);
556 SI7(r1, r4, r3, r0, r2); KL(r4, r0, r1, r2, r3, 7);
557 SI6(r4, r0, r1, r2, r3); KL(r1, r3, r2, r4, r0, 6);
558 SI5(r1, r3, r2, r4, r0); KL(r3, r0, r1, r2, r4, 5);
559 SI4(r3, r0, r1, r2, r4); KL(r3, r1, r4, r2, r0, 4);
560 SI3(r3, r1, r4, r2, r0); KL(r4, r3, r0, r2, r1, 3);
561 SI2(r4, r3, r0, r2, r1); KL(r3, r1, r2, r0, r4, 2);
562 SI1(r3, r1, r2, r0, r4); KL(r4, r1, r2, r0, r3, 1);
563 SI0(r4, r1, r2, r0, r3); K(r2, r3, r1, r4, 0);
564
565 d[0] = cpu_to_le32(r2);
566 d[1] = cpu_to_le32(r3);
567 d[2] = cpu_to_le32(r1);
568 d[3] = cpu_to_le32(r4);
569}
570EXPORT_SYMBOL_GPL(__serpent_decrypt);
571
572static void serpent_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
573{
574 struct serpent_ctx *ctx = crypto_tfm_ctx(tfm);
575
576 __serpent_decrypt(ctx, dst, src);
577}
578
579static int tnepres_setkey(struct crypto_tfm *tfm, const u8 *key,
580 unsigned int keylen)
581{
582 u8 rev_key[SERPENT_MAX_KEY_SIZE];
583 int i;
584
585 for (i = 0; i < keylen; ++i)
586 rev_key[keylen - i - 1] = key[i];
587
588 return serpent_setkey(tfm, rev_key, keylen);
589}
590
591static void tnepres_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
592{
593 const u32 * const s = (const u32 * const)src;
594 u32 * const d = (u32 * const)dst;
595
596 u32 rs[4], rd[4];
597
598 rs[0] = swab32(s[3]);
599 rs[1] = swab32(s[2]);
600 rs[2] = swab32(s[1]);
601 rs[3] = swab32(s[0]);
602
603 serpent_encrypt(tfm, (u8 *)rd, (u8 *)rs);
604
605 d[0] = swab32(rd[3]);
606 d[1] = swab32(rd[2]);
607 d[2] = swab32(rd[1]);
608 d[3] = swab32(rd[0]);
609}
610
611static void tnepres_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
612{
613 const u32 * const s = (const u32 * const)src;
614 u32 * const d = (u32 * const)dst;
615
616 u32 rs[4], rd[4];
617
618 rs[0] = swab32(s[3]);
619 rs[1] = swab32(s[2]);
620 rs[2] = swab32(s[1]);
621 rs[3] = swab32(s[0]);
622
623 serpent_decrypt(tfm, (u8 *)rd, (u8 *)rs);
624
625 d[0] = swab32(rd[3]);
626 d[1] = swab32(rd[2]);
627 d[2] = swab32(rd[1]);
628 d[3] = swab32(rd[0]);
629}
630
631static struct crypto_alg srp_algs[2] = { {
632 .cra_name = "serpent",
633 .cra_driver_name = "serpent-generic",
634 .cra_priority = 100,
635 .cra_flags = CRYPTO_ALG_TYPE_CIPHER,
636 .cra_blocksize = SERPENT_BLOCK_SIZE,
637 .cra_ctxsize = sizeof(struct serpent_ctx),
638 .cra_alignmask = 3,
639 .cra_module = THIS_MODULE,
640 .cra_u = { .cipher = {
641 .cia_min_keysize = SERPENT_MIN_KEY_SIZE,
642 .cia_max_keysize = SERPENT_MAX_KEY_SIZE,
643 .cia_setkey = serpent_setkey,
644 .cia_encrypt = serpent_encrypt,
645 .cia_decrypt = serpent_decrypt } }
646}, {
647 .cra_name = "tnepres",
648 .cra_driver_name = "tnepres-generic",
649 .cra_flags = CRYPTO_ALG_TYPE_CIPHER,
650 .cra_blocksize = SERPENT_BLOCK_SIZE,
651 .cra_ctxsize = sizeof(struct serpent_ctx),
652 .cra_alignmask = 3,
653 .cra_module = THIS_MODULE,
654 .cra_u = { .cipher = {
655 .cia_min_keysize = SERPENT_MIN_KEY_SIZE,
656 .cia_max_keysize = SERPENT_MAX_KEY_SIZE,
657 .cia_setkey = tnepres_setkey,
658 .cia_encrypt = tnepres_encrypt,
659 .cia_decrypt = tnepres_decrypt } }
660} };
661
662static int __init serpent_mod_init(void)
663{
664 return crypto_register_algs(srp_algs, ARRAY_SIZE(srp_algs));
665}
666
667static void __exit serpent_mod_fini(void)
668{
669 crypto_unregister_algs(srp_algs, ARRAY_SIZE(srp_algs));
670}
671
672subsys_initcall(serpent_mod_init);
673module_exit(serpent_mod_fini);
674
675MODULE_LICENSE("GPL");
676MODULE_DESCRIPTION("Serpent and tnepres (kerneli compatible serpent reversed) Cipher Algorithm");
677MODULE_AUTHOR("Dag Arne Osvik <osvik@ii.uib.no>");
678MODULE_ALIAS_CRYPTO("tnepres");
679MODULE_ALIAS_CRYPTO("serpent");
680MODULE_ALIAS_CRYPTO("serpent-generic");
1/*
2 * Cryptographic API.
3 *
4 * Serpent Cipher Algorithm.
5 *
6 * Copyright (C) 2002 Dag Arne Osvik <osvik@ii.uib.no>
7 * 2003 Herbert Valerio Riedel <hvr@gnu.org>
8 *
9 * Added tnepres support:
10 * Ruben Jesus Garcia Hernandez <ruben@ugr.es>, 18.10.2004
11 * Based on code by hvr
12 *
13 * This program is free software; you can redistribute it and/or modify
14 * it under the terms of the GNU General Public License as published by
15 * the Free Software Foundation; either version 2 of the License, or
16 * (at your option) any later version.
17 */
18
19#include <linux/init.h>
20#include <linux/module.h>
21#include <linux/errno.h>
22#include <asm/byteorder.h>
23#include <linux/crypto.h>
24#include <linux/types.h>
25#include <crypto/serpent.h>
26
27/* Key is padded to the maximum of 256 bits before round key generation.
28 * Any key length <= 256 bits (32 bytes) is allowed by the algorithm.
29 */
30
31#define PHI 0x9e3779b9UL
32
33#define keyiter(a, b, c, d, i, j) \
34 ({ b ^= d; b ^= c; b ^= a; b ^= PHI ^ i; b = rol32(b, 11); k[j] = b; })
35
36#define loadkeys(x0, x1, x2, x3, i) \
37 ({ x0 = k[i]; x1 = k[i+1]; x2 = k[i+2]; x3 = k[i+3]; })
38
39#define storekeys(x0, x1, x2, x3, i) \
40 ({ k[i] = x0; k[i+1] = x1; k[i+2] = x2; k[i+3] = x3; })
41
42#define store_and_load_keys(x0, x1, x2, x3, s, l) \
43 ({ storekeys(x0, x1, x2, x3, s); loadkeys(x0, x1, x2, x3, l); })
44
45#define K(x0, x1, x2, x3, i) ({ \
46 x3 ^= k[4*(i)+3]; x2 ^= k[4*(i)+2]; \
47 x1 ^= k[4*(i)+1]; x0 ^= k[4*(i)+0]; \
48 })
49
50#define LK(x0, x1, x2, x3, x4, i) ({ \
51 x0 = rol32(x0, 13);\
52 x2 = rol32(x2, 3); x1 ^= x0; x4 = x0 << 3; \
53 x3 ^= x2; x1 ^= x2; \
54 x1 = rol32(x1, 1); x3 ^= x4; \
55 x3 = rol32(x3, 7); x4 = x1; \
56 x0 ^= x1; x4 <<= 7; x2 ^= x3; \
57 x0 ^= x3; x2 ^= x4; x3 ^= k[4*i+3]; \
58 x1 ^= k[4*i+1]; x0 = rol32(x0, 5); x2 = rol32(x2, 22);\
59 x0 ^= k[4*i+0]; x2 ^= k[4*i+2]; \
60 })
61
62#define KL(x0, x1, x2, x3, x4, i) ({ \
63 x0 ^= k[4*i+0]; x1 ^= k[4*i+1]; x2 ^= k[4*i+2]; \
64 x3 ^= k[4*i+3]; x0 = ror32(x0, 5); x2 = ror32(x2, 22);\
65 x4 = x1; x2 ^= x3; x0 ^= x3; \
66 x4 <<= 7; x0 ^= x1; x1 = ror32(x1, 1); \
67 x2 ^= x4; x3 = ror32(x3, 7); x4 = x0 << 3; \
68 x1 ^= x0; x3 ^= x4; x0 = ror32(x0, 13);\
69 x1 ^= x2; x3 ^= x2; x2 = ror32(x2, 3); \
70 })
71
72#define S0(x0, x1, x2, x3, x4) ({ \
73 x4 = x3; \
74 x3 |= x0; x0 ^= x4; x4 ^= x2; \
75 x4 = ~x4; x3 ^= x1; x1 &= x0; \
76 x1 ^= x4; x2 ^= x0; x0 ^= x3; \
77 x4 |= x0; x0 ^= x2; x2 &= x1; \
78 x3 ^= x2; x1 = ~x1; x2 ^= x4; \
79 x1 ^= x2; \
80 })
81
82#define S1(x0, x1, x2, x3, x4) ({ \
83 x4 = x1; \
84 x1 ^= x0; x0 ^= x3; x3 = ~x3; \
85 x4 &= x1; x0 |= x1; x3 ^= x2; \
86 x0 ^= x3; x1 ^= x3; x3 ^= x4; \
87 x1 |= x4; x4 ^= x2; x2 &= x0; \
88 x2 ^= x1; x1 |= x0; x0 = ~x0; \
89 x0 ^= x2; x4 ^= x1; \
90 })
91
92#define S2(x0, x1, x2, x3, x4) ({ \
93 x3 = ~x3; \
94 x1 ^= x0; x4 = x0; x0 &= x2; \
95 x0 ^= x3; x3 |= x4; x2 ^= x1; \
96 x3 ^= x1; x1 &= x0; x0 ^= x2; \
97 x2 &= x3; x3 |= x1; x0 = ~x0; \
98 x3 ^= x0; x4 ^= x0; x0 ^= x2; \
99 x1 |= x2; \
100 })
101
102#define S3(x0, x1, x2, x3, x4) ({ \
103 x4 = x1; \
104 x1 ^= x3; x3 |= x0; x4 &= x0; \
105 x0 ^= x2; x2 ^= x1; x1 &= x3; \
106 x2 ^= x3; x0 |= x4; x4 ^= x3; \
107 x1 ^= x0; x0 &= x3; x3 &= x4; \
108 x3 ^= x2; x4 |= x1; x2 &= x1; \
109 x4 ^= x3; x0 ^= x3; x3 ^= x2; \
110 })
111
112#define S4(x0, x1, x2, x3, x4) ({ \
113 x4 = x3; \
114 x3 &= x0; x0 ^= x4; \
115 x3 ^= x2; x2 |= x4; x0 ^= x1; \
116 x4 ^= x3; x2 |= x0; \
117 x2 ^= x1; x1 &= x0; \
118 x1 ^= x4; x4 &= x2; x2 ^= x3; \
119 x4 ^= x0; x3 |= x1; x1 = ~x1; \
120 x3 ^= x0; \
121 })
122
123#define S5(x0, x1, x2, x3, x4) ({ \
124 x4 = x1; x1 |= x0; \
125 x2 ^= x1; x3 = ~x3; x4 ^= x0; \
126 x0 ^= x2; x1 &= x4; x4 |= x3; \
127 x4 ^= x0; x0 &= x3; x1 ^= x3; \
128 x3 ^= x2; x0 ^= x1; x2 &= x4; \
129 x1 ^= x2; x2 &= x0; \
130 x3 ^= x2; \
131 })
132
133#define S6(x0, x1, x2, x3, x4) ({ \
134 x4 = x1; \
135 x3 ^= x0; x1 ^= x2; x2 ^= x0; \
136 x0 &= x3; x1 |= x3; x4 = ~x4; \
137 x0 ^= x1; x1 ^= x2; \
138 x3 ^= x4; x4 ^= x0; x2 &= x0; \
139 x4 ^= x1; x2 ^= x3; x3 &= x1; \
140 x3 ^= x0; x1 ^= x2; \
141 })
142
143#define S7(x0, x1, x2, x3, x4) ({ \
144 x1 = ~x1; \
145 x4 = x1; x0 = ~x0; x1 &= x2; \
146 x1 ^= x3; x3 |= x4; x4 ^= x2; \
147 x2 ^= x3; x3 ^= x0; x0 |= x1; \
148 x2 &= x0; x0 ^= x4; x4 ^= x3; \
149 x3 &= x0; x4 ^= x1; \
150 x2 ^= x4; x3 ^= x1; x4 |= x0; \
151 x4 ^= x1; \
152 })
153
154#define SI0(x0, x1, x2, x3, x4) ({ \
155 x4 = x3; x1 ^= x0; \
156 x3 |= x1; x4 ^= x1; x0 = ~x0; \
157 x2 ^= x3; x3 ^= x0; x0 &= x1; \
158 x0 ^= x2; x2 &= x3; x3 ^= x4; \
159 x2 ^= x3; x1 ^= x3; x3 &= x0; \
160 x1 ^= x0; x0 ^= x2; x4 ^= x3; \
161 })
162
163#define SI1(x0, x1, x2, x3, x4) ({ \
164 x1 ^= x3; x4 = x0; \
165 x0 ^= x2; x2 = ~x2; x4 |= x1; \
166 x4 ^= x3; x3 &= x1; x1 ^= x2; \
167 x2 &= x4; x4 ^= x1; x1 |= x3; \
168 x3 ^= x0; x2 ^= x0; x0 |= x4; \
169 x2 ^= x4; x1 ^= x0; \
170 x4 ^= x1; \
171 })
172
173#define SI2(x0, x1, x2, x3, x4) ({ \
174 x2 ^= x1; x4 = x3; x3 = ~x3; \
175 x3 |= x2; x2 ^= x4; x4 ^= x0; \
176 x3 ^= x1; x1 |= x2; x2 ^= x0; \
177 x1 ^= x4; x4 |= x3; x2 ^= x3; \
178 x4 ^= x2; x2 &= x1; \
179 x2 ^= x3; x3 ^= x4; x4 ^= x0; \
180 })
181
182#define SI3(x0, x1, x2, x3, x4) ({ \
183 x2 ^= x1; \
184 x4 = x1; x1 &= x2; \
185 x1 ^= x0; x0 |= x4; x4 ^= x3; \
186 x0 ^= x3; x3 |= x1; x1 ^= x2; \
187 x1 ^= x3; x0 ^= x2; x2 ^= x3; \
188 x3 &= x1; x1 ^= x0; x0 &= x2; \
189 x4 ^= x3; x3 ^= x0; x0 ^= x1; \
190 })
191
192#define SI4(x0, x1, x2, x3, x4) ({ \
193 x2 ^= x3; x4 = x0; x0 &= x1; \
194 x0 ^= x2; x2 |= x3; x4 = ~x4; \
195 x1 ^= x0; x0 ^= x2; x2 &= x4; \
196 x2 ^= x0; x0 |= x4; \
197 x0 ^= x3; x3 &= x2; \
198 x4 ^= x3; x3 ^= x1; x1 &= x0; \
199 x4 ^= x1; x0 ^= x3; \
200 })
201
202#define SI5(x0, x1, x2, x3, x4) ({ \
203 x4 = x1; x1 |= x2; \
204 x2 ^= x4; x1 ^= x3; x3 &= x4; \
205 x2 ^= x3; x3 |= x0; x0 = ~x0; \
206 x3 ^= x2; x2 |= x0; x4 ^= x1; \
207 x2 ^= x4; x4 &= x0; x0 ^= x1; \
208 x1 ^= x3; x0 &= x2; x2 ^= x3; \
209 x0 ^= x2; x2 ^= x4; x4 ^= x3; \
210 })
211
212#define SI6(x0, x1, x2, x3, x4) ({ \
213 x0 ^= x2; \
214 x4 = x0; x0 &= x3; x2 ^= x3; \
215 x0 ^= x2; x3 ^= x1; x2 |= x4; \
216 x2 ^= x3; x3 &= x0; x0 = ~x0; \
217 x3 ^= x1; x1 &= x2; x4 ^= x0; \
218 x3 ^= x4; x4 ^= x2; x0 ^= x1; \
219 x2 ^= x0; \
220 })
221
222#define SI7(x0, x1, x2, x3, x4) ({ \
223 x4 = x3; x3 &= x0; x0 ^= x2; \
224 x2 |= x4; x4 ^= x1; x0 = ~x0; \
225 x1 |= x3; x4 ^= x0; x0 &= x2; \
226 x0 ^= x1; x1 &= x2; x3 ^= x2; \
227 x4 ^= x3; x2 &= x3; x3 |= x0; \
228 x1 ^= x4; x3 ^= x4; x4 &= x0; \
229 x4 ^= x2; \
230 })
231
232int __serpent_setkey(struct serpent_ctx *ctx, const u8 *key,
233 unsigned int keylen)
234{
235 u32 *k = ctx->expkey;
236 u8 *k8 = (u8 *)k;
237 u32 r0, r1, r2, r3, r4;
238 int i;
239
240 /* Copy key, add padding */
241
242 for (i = 0; i < keylen; ++i)
243 k8[i] = key[i];
244 if (i < SERPENT_MAX_KEY_SIZE)
245 k8[i++] = 1;
246 while (i < SERPENT_MAX_KEY_SIZE)
247 k8[i++] = 0;
248
249 /* Expand key using polynomial */
250
251 r0 = le32_to_cpu(k[3]);
252 r1 = le32_to_cpu(k[4]);
253 r2 = le32_to_cpu(k[5]);
254 r3 = le32_to_cpu(k[6]);
255 r4 = le32_to_cpu(k[7]);
256
257 keyiter(le32_to_cpu(k[0]), r0, r4, r2, 0, 0);
258 keyiter(le32_to_cpu(k[1]), r1, r0, r3, 1, 1);
259 keyiter(le32_to_cpu(k[2]), r2, r1, r4, 2, 2);
260 keyiter(le32_to_cpu(k[3]), r3, r2, r0, 3, 3);
261 keyiter(le32_to_cpu(k[4]), r4, r3, r1, 4, 4);
262 keyiter(le32_to_cpu(k[5]), r0, r4, r2, 5, 5);
263 keyiter(le32_to_cpu(k[6]), r1, r0, r3, 6, 6);
264 keyiter(le32_to_cpu(k[7]), r2, r1, r4, 7, 7);
265
266 keyiter(k[0], r3, r2, r0, 8, 8);
267 keyiter(k[1], r4, r3, r1, 9, 9);
268 keyiter(k[2], r0, r4, r2, 10, 10);
269 keyiter(k[3], r1, r0, r3, 11, 11);
270 keyiter(k[4], r2, r1, r4, 12, 12);
271 keyiter(k[5], r3, r2, r0, 13, 13);
272 keyiter(k[6], r4, r3, r1, 14, 14);
273 keyiter(k[7], r0, r4, r2, 15, 15);
274 keyiter(k[8], r1, r0, r3, 16, 16);
275 keyiter(k[9], r2, r1, r4, 17, 17);
276 keyiter(k[10], r3, r2, r0, 18, 18);
277 keyiter(k[11], r4, r3, r1, 19, 19);
278 keyiter(k[12], r0, r4, r2, 20, 20);
279 keyiter(k[13], r1, r0, r3, 21, 21);
280 keyiter(k[14], r2, r1, r4, 22, 22);
281 keyiter(k[15], r3, r2, r0, 23, 23);
282 keyiter(k[16], r4, r3, r1, 24, 24);
283 keyiter(k[17], r0, r4, r2, 25, 25);
284 keyiter(k[18], r1, r0, r3, 26, 26);
285 keyiter(k[19], r2, r1, r4, 27, 27);
286 keyiter(k[20], r3, r2, r0, 28, 28);
287 keyiter(k[21], r4, r3, r1, 29, 29);
288 keyiter(k[22], r0, r4, r2, 30, 30);
289 keyiter(k[23], r1, r0, r3, 31, 31);
290
291 k += 50;
292
293 keyiter(k[-26], r2, r1, r4, 32, -18);
294 keyiter(k[-25], r3, r2, r0, 33, -17);
295 keyiter(k[-24], r4, r3, r1, 34, -16);
296 keyiter(k[-23], r0, r4, r2, 35, -15);
297 keyiter(k[-22], r1, r0, r3, 36, -14);
298 keyiter(k[-21], r2, r1, r4, 37, -13);
299 keyiter(k[-20], r3, r2, r0, 38, -12);
300 keyiter(k[-19], r4, r3, r1, 39, -11);
301 keyiter(k[-18], r0, r4, r2, 40, -10);
302 keyiter(k[-17], r1, r0, r3, 41, -9);
303 keyiter(k[-16], r2, r1, r4, 42, -8);
304 keyiter(k[-15], r3, r2, r0, 43, -7);
305 keyiter(k[-14], r4, r3, r1, 44, -6);
306 keyiter(k[-13], r0, r4, r2, 45, -5);
307 keyiter(k[-12], r1, r0, r3, 46, -4);
308 keyiter(k[-11], r2, r1, r4, 47, -3);
309 keyiter(k[-10], r3, r2, r0, 48, -2);
310 keyiter(k[-9], r4, r3, r1, 49, -1);
311 keyiter(k[-8], r0, r4, r2, 50, 0);
312 keyiter(k[-7], r1, r0, r3, 51, 1);
313 keyiter(k[-6], r2, r1, r4, 52, 2);
314 keyiter(k[-5], r3, r2, r0, 53, 3);
315 keyiter(k[-4], r4, r3, r1, 54, 4);
316 keyiter(k[-3], r0, r4, r2, 55, 5);
317 keyiter(k[-2], r1, r0, r3, 56, 6);
318 keyiter(k[-1], r2, r1, r4, 57, 7);
319 keyiter(k[0], r3, r2, r0, 58, 8);
320 keyiter(k[1], r4, r3, r1, 59, 9);
321 keyiter(k[2], r0, r4, r2, 60, 10);
322 keyiter(k[3], r1, r0, r3, 61, 11);
323 keyiter(k[4], r2, r1, r4, 62, 12);
324 keyiter(k[5], r3, r2, r0, 63, 13);
325 keyiter(k[6], r4, r3, r1, 64, 14);
326 keyiter(k[7], r0, r4, r2, 65, 15);
327 keyiter(k[8], r1, r0, r3, 66, 16);
328 keyiter(k[9], r2, r1, r4, 67, 17);
329 keyiter(k[10], r3, r2, r0, 68, 18);
330 keyiter(k[11], r4, r3, r1, 69, 19);
331 keyiter(k[12], r0, r4, r2, 70, 20);
332 keyiter(k[13], r1, r0, r3, 71, 21);
333 keyiter(k[14], r2, r1, r4, 72, 22);
334 keyiter(k[15], r3, r2, r0, 73, 23);
335 keyiter(k[16], r4, r3, r1, 74, 24);
336 keyiter(k[17], r0, r4, r2, 75, 25);
337 keyiter(k[18], r1, r0, r3, 76, 26);
338 keyiter(k[19], r2, r1, r4, 77, 27);
339 keyiter(k[20], r3, r2, r0, 78, 28);
340 keyiter(k[21], r4, r3, r1, 79, 29);
341 keyiter(k[22], r0, r4, r2, 80, 30);
342 keyiter(k[23], r1, r0, r3, 81, 31);
343
344 k += 50;
345
346 keyiter(k[-26], r2, r1, r4, 82, -18);
347 keyiter(k[-25], r3, r2, r0, 83, -17);
348 keyiter(k[-24], r4, r3, r1, 84, -16);
349 keyiter(k[-23], r0, r4, r2, 85, -15);
350 keyiter(k[-22], r1, r0, r3, 86, -14);
351 keyiter(k[-21], r2, r1, r4, 87, -13);
352 keyiter(k[-20], r3, r2, r0, 88, -12);
353 keyiter(k[-19], r4, r3, r1, 89, -11);
354 keyiter(k[-18], r0, r4, r2, 90, -10);
355 keyiter(k[-17], r1, r0, r3, 91, -9);
356 keyiter(k[-16], r2, r1, r4, 92, -8);
357 keyiter(k[-15], r3, r2, r0, 93, -7);
358 keyiter(k[-14], r4, r3, r1, 94, -6);
359 keyiter(k[-13], r0, r4, r2, 95, -5);
360 keyiter(k[-12], r1, r0, r3, 96, -4);
361 keyiter(k[-11], r2, r1, r4, 97, -3);
362 keyiter(k[-10], r3, r2, r0, 98, -2);
363 keyiter(k[-9], r4, r3, r1, 99, -1);
364 keyiter(k[-8], r0, r4, r2, 100, 0);
365 keyiter(k[-7], r1, r0, r3, 101, 1);
366 keyiter(k[-6], r2, r1, r4, 102, 2);
367 keyiter(k[-5], r3, r2, r0, 103, 3);
368 keyiter(k[-4], r4, r3, r1, 104, 4);
369 keyiter(k[-3], r0, r4, r2, 105, 5);
370 keyiter(k[-2], r1, r0, r3, 106, 6);
371 keyiter(k[-1], r2, r1, r4, 107, 7);
372 keyiter(k[0], r3, r2, r0, 108, 8);
373 keyiter(k[1], r4, r3, r1, 109, 9);
374 keyiter(k[2], r0, r4, r2, 110, 10);
375 keyiter(k[3], r1, r0, r3, 111, 11);
376 keyiter(k[4], r2, r1, r4, 112, 12);
377 keyiter(k[5], r3, r2, r0, 113, 13);
378 keyiter(k[6], r4, r3, r1, 114, 14);
379 keyiter(k[7], r0, r4, r2, 115, 15);
380 keyiter(k[8], r1, r0, r3, 116, 16);
381 keyiter(k[9], r2, r1, r4, 117, 17);
382 keyiter(k[10], r3, r2, r0, 118, 18);
383 keyiter(k[11], r4, r3, r1, 119, 19);
384 keyiter(k[12], r0, r4, r2, 120, 20);
385 keyiter(k[13], r1, r0, r3, 121, 21);
386 keyiter(k[14], r2, r1, r4, 122, 22);
387 keyiter(k[15], r3, r2, r0, 123, 23);
388 keyiter(k[16], r4, r3, r1, 124, 24);
389 keyiter(k[17], r0, r4, r2, 125, 25);
390 keyiter(k[18], r1, r0, r3, 126, 26);
391 keyiter(k[19], r2, r1, r4, 127, 27);
392 keyiter(k[20], r3, r2, r0, 128, 28);
393 keyiter(k[21], r4, r3, r1, 129, 29);
394 keyiter(k[22], r0, r4, r2, 130, 30);
395 keyiter(k[23], r1, r0, r3, 131, 31);
396
397 /* Apply S-boxes */
398
399 S3(r3, r4, r0, r1, r2); store_and_load_keys(r1, r2, r4, r3, 28, 24);
400 S4(r1, r2, r4, r3, r0); store_and_load_keys(r2, r4, r3, r0, 24, 20);
401 S5(r2, r4, r3, r0, r1); store_and_load_keys(r1, r2, r4, r0, 20, 16);
402 S6(r1, r2, r4, r0, r3); store_and_load_keys(r4, r3, r2, r0, 16, 12);
403 S7(r4, r3, r2, r0, r1); store_and_load_keys(r1, r2, r0, r4, 12, 8);
404 S0(r1, r2, r0, r4, r3); store_and_load_keys(r0, r2, r4, r1, 8, 4);
405 S1(r0, r2, r4, r1, r3); store_and_load_keys(r3, r4, r1, r0, 4, 0);
406 S2(r3, r4, r1, r0, r2); store_and_load_keys(r2, r4, r3, r0, 0, -4);
407 S3(r2, r4, r3, r0, r1); store_and_load_keys(r0, r1, r4, r2, -4, -8);
408 S4(r0, r1, r4, r2, r3); store_and_load_keys(r1, r4, r2, r3, -8, -12);
409 S5(r1, r4, r2, r3, r0); store_and_load_keys(r0, r1, r4, r3, -12, -16);
410 S6(r0, r1, r4, r3, r2); store_and_load_keys(r4, r2, r1, r3, -16, -20);
411 S7(r4, r2, r1, r3, r0); store_and_load_keys(r0, r1, r3, r4, -20, -24);
412 S0(r0, r1, r3, r4, r2); store_and_load_keys(r3, r1, r4, r0, -24, -28);
413 k -= 50;
414 S1(r3, r1, r4, r0, r2); store_and_load_keys(r2, r4, r0, r3, 22, 18);
415 S2(r2, r4, r0, r3, r1); store_and_load_keys(r1, r4, r2, r3, 18, 14);
416 S3(r1, r4, r2, r3, r0); store_and_load_keys(r3, r0, r4, r1, 14, 10);
417 S4(r3, r0, r4, r1, r2); store_and_load_keys(r0, r4, r1, r2, 10, 6);
418 S5(r0, r4, r1, r2, r3); store_and_load_keys(r3, r0, r4, r2, 6, 2);
419 S6(r3, r0, r4, r2, r1); store_and_load_keys(r4, r1, r0, r2, 2, -2);
420 S7(r4, r1, r0, r2, r3); store_and_load_keys(r3, r0, r2, r4, -2, -6);
421 S0(r3, r0, r2, r4, r1); store_and_load_keys(r2, r0, r4, r3, -6, -10);
422 S1(r2, r0, r4, r3, r1); store_and_load_keys(r1, r4, r3, r2, -10, -14);
423 S2(r1, r4, r3, r2, r0); store_and_load_keys(r0, r4, r1, r2, -14, -18);
424 S3(r0, r4, r1, r2, r3); store_and_load_keys(r2, r3, r4, r0, -18, -22);
425 k -= 50;
426 S4(r2, r3, r4, r0, r1); store_and_load_keys(r3, r4, r0, r1, 28, 24);
427 S5(r3, r4, r0, r1, r2); store_and_load_keys(r2, r3, r4, r1, 24, 20);
428 S6(r2, r3, r4, r1, r0); store_and_load_keys(r4, r0, r3, r1, 20, 16);
429 S7(r4, r0, r3, r1, r2); store_and_load_keys(r2, r3, r1, r4, 16, 12);
430 S0(r2, r3, r1, r4, r0); store_and_load_keys(r1, r3, r4, r2, 12, 8);
431 S1(r1, r3, r4, r2, r0); store_and_load_keys(r0, r4, r2, r1, 8, 4);
432 S2(r0, r4, r2, r1, r3); store_and_load_keys(r3, r4, r0, r1, 4, 0);
433 S3(r3, r4, r0, r1, r2); storekeys(r1, r2, r4, r3, 0);
434
435 return 0;
436}
437EXPORT_SYMBOL_GPL(__serpent_setkey);
438
439int serpent_setkey(struct crypto_tfm *tfm, const u8 *key, unsigned int keylen)
440{
441 return __serpent_setkey(crypto_tfm_ctx(tfm), key, keylen);
442}
443EXPORT_SYMBOL_GPL(serpent_setkey);
444
445void __serpent_encrypt(struct serpent_ctx *ctx, u8 *dst, const u8 *src)
446{
447 const u32 *k = ctx->expkey;
448 const __le32 *s = (const __le32 *)src;
449 __le32 *d = (__le32 *)dst;
450 u32 r0, r1, r2, r3, r4;
451
452/*
453 * Note: The conversions between u8* and u32* might cause trouble
454 * on architectures with stricter alignment rules than x86
455 */
456
457 r0 = le32_to_cpu(s[0]);
458 r1 = le32_to_cpu(s[1]);
459 r2 = le32_to_cpu(s[2]);
460 r3 = le32_to_cpu(s[3]);
461
462 K(r0, r1, r2, r3, 0);
463 S0(r0, r1, r2, r3, r4); LK(r2, r1, r3, r0, r4, 1);
464 S1(r2, r1, r3, r0, r4); LK(r4, r3, r0, r2, r1, 2);
465 S2(r4, r3, r0, r2, r1); LK(r1, r3, r4, r2, r0, 3);
466 S3(r1, r3, r4, r2, r0); LK(r2, r0, r3, r1, r4, 4);
467 S4(r2, r0, r3, r1, r4); LK(r0, r3, r1, r4, r2, 5);
468 S5(r0, r3, r1, r4, r2); LK(r2, r0, r3, r4, r1, 6);
469 S6(r2, r0, r3, r4, r1); LK(r3, r1, r0, r4, r2, 7);
470 S7(r3, r1, r0, r4, r2); LK(r2, r0, r4, r3, r1, 8);
471 S0(r2, r0, r4, r3, r1); LK(r4, r0, r3, r2, r1, 9);
472 S1(r4, r0, r3, r2, r1); LK(r1, r3, r2, r4, r0, 10);
473 S2(r1, r3, r2, r4, r0); LK(r0, r3, r1, r4, r2, 11);
474 S3(r0, r3, r1, r4, r2); LK(r4, r2, r3, r0, r1, 12);
475 S4(r4, r2, r3, r0, r1); LK(r2, r3, r0, r1, r4, 13);
476 S5(r2, r3, r0, r1, r4); LK(r4, r2, r3, r1, r0, 14);
477 S6(r4, r2, r3, r1, r0); LK(r3, r0, r2, r1, r4, 15);
478 S7(r3, r0, r2, r1, r4); LK(r4, r2, r1, r3, r0, 16);
479 S0(r4, r2, r1, r3, r0); LK(r1, r2, r3, r4, r0, 17);
480 S1(r1, r2, r3, r4, r0); LK(r0, r3, r4, r1, r2, 18);
481 S2(r0, r3, r4, r1, r2); LK(r2, r3, r0, r1, r4, 19);
482 S3(r2, r3, r0, r1, r4); LK(r1, r4, r3, r2, r0, 20);
483 S4(r1, r4, r3, r2, r0); LK(r4, r3, r2, r0, r1, 21);
484 S5(r4, r3, r2, r0, r1); LK(r1, r4, r3, r0, r2, 22);
485 S6(r1, r4, r3, r0, r2); LK(r3, r2, r4, r0, r1, 23);
486 S7(r3, r2, r4, r0, r1); LK(r1, r4, r0, r3, r2, 24);
487 S0(r1, r4, r0, r3, r2); LK(r0, r4, r3, r1, r2, 25);
488 S1(r0, r4, r3, r1, r2); LK(r2, r3, r1, r0, r4, 26);
489 S2(r2, r3, r1, r0, r4); LK(r4, r3, r2, r0, r1, 27);
490 S3(r4, r3, r2, r0, r1); LK(r0, r1, r3, r4, r2, 28);
491 S4(r0, r1, r3, r4, r2); LK(r1, r3, r4, r2, r0, 29);
492 S5(r1, r3, r4, r2, r0); LK(r0, r1, r3, r2, r4, 30);
493 S6(r0, r1, r3, r2, r4); LK(r3, r4, r1, r2, r0, 31);
494 S7(r3, r4, r1, r2, r0); K(r0, r1, r2, r3, 32);
495
496 d[0] = cpu_to_le32(r0);
497 d[1] = cpu_to_le32(r1);
498 d[2] = cpu_to_le32(r2);
499 d[3] = cpu_to_le32(r3);
500}
501EXPORT_SYMBOL_GPL(__serpent_encrypt);
502
503static void serpent_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
504{
505 struct serpent_ctx *ctx = crypto_tfm_ctx(tfm);
506
507 __serpent_encrypt(ctx, dst, src);
508}
509
510void __serpent_decrypt(struct serpent_ctx *ctx, u8 *dst, const u8 *src)
511{
512 const u32 *k = ctx->expkey;
513 const __le32 *s = (const __le32 *)src;
514 __le32 *d = (__le32 *)dst;
515 u32 r0, r1, r2, r3, r4;
516
517 r0 = le32_to_cpu(s[0]);
518 r1 = le32_to_cpu(s[1]);
519 r2 = le32_to_cpu(s[2]);
520 r3 = le32_to_cpu(s[3]);
521
522 K(r0, r1, r2, r3, 32);
523 SI7(r0, r1, r2, r3, r4); KL(r1, r3, r0, r4, r2, 31);
524 SI6(r1, r3, r0, r4, r2); KL(r0, r2, r4, r1, r3, 30);
525 SI5(r0, r2, r4, r1, r3); KL(r2, r3, r0, r4, r1, 29);
526 SI4(r2, r3, r0, r4, r1); KL(r2, r0, r1, r4, r3, 28);
527 SI3(r2, r0, r1, r4, r3); KL(r1, r2, r3, r4, r0, 27);
528 SI2(r1, r2, r3, r4, r0); KL(r2, r0, r4, r3, r1, 26);
529 SI1(r2, r0, r4, r3, r1); KL(r1, r0, r4, r3, r2, 25);
530 SI0(r1, r0, r4, r3, r2); KL(r4, r2, r0, r1, r3, 24);
531 SI7(r4, r2, r0, r1, r3); KL(r2, r1, r4, r3, r0, 23);
532 SI6(r2, r1, r4, r3, r0); KL(r4, r0, r3, r2, r1, 22);
533 SI5(r4, r0, r3, r2, r1); KL(r0, r1, r4, r3, r2, 21);
534 SI4(r0, r1, r4, r3, r2); KL(r0, r4, r2, r3, r1, 20);
535 SI3(r0, r4, r2, r3, r1); KL(r2, r0, r1, r3, r4, 19);
536 SI2(r2, r0, r1, r3, r4); KL(r0, r4, r3, r1, r2, 18);
537 SI1(r0, r4, r3, r1, r2); KL(r2, r4, r3, r1, r0, 17);
538 SI0(r2, r4, r3, r1, r0); KL(r3, r0, r4, r2, r1, 16);
539 SI7(r3, r0, r4, r2, r1); KL(r0, r2, r3, r1, r4, 15);
540 SI6(r0, r2, r3, r1, r4); KL(r3, r4, r1, r0, r2, 14);
541 SI5(r3, r4, r1, r0, r2); KL(r4, r2, r3, r1, r0, 13);
542 SI4(r4, r2, r3, r1, r0); KL(r4, r3, r0, r1, r2, 12);
543 SI3(r4, r3, r0, r1, r2); KL(r0, r4, r2, r1, r3, 11);
544 SI2(r0, r4, r2, r1, r3); KL(r4, r3, r1, r2, r0, 10);
545 SI1(r4, r3, r1, r2, r0); KL(r0, r3, r1, r2, r4, 9);
546 SI0(r0, r3, r1, r2, r4); KL(r1, r4, r3, r0, r2, 8);
547 SI7(r1, r4, r3, r0, r2); KL(r4, r0, r1, r2, r3, 7);
548 SI6(r4, r0, r1, r2, r3); KL(r1, r3, r2, r4, r0, 6);
549 SI5(r1, r3, r2, r4, r0); KL(r3, r0, r1, r2, r4, 5);
550 SI4(r3, r0, r1, r2, r4); KL(r3, r1, r4, r2, r0, 4);
551 SI3(r3, r1, r4, r2, r0); KL(r4, r3, r0, r2, r1, 3);
552 SI2(r4, r3, r0, r2, r1); KL(r3, r1, r2, r0, r4, 2);
553 SI1(r3, r1, r2, r0, r4); KL(r4, r1, r2, r0, r3, 1);
554 SI0(r4, r1, r2, r0, r3); K(r2, r3, r1, r4, 0);
555
556 d[0] = cpu_to_le32(r2);
557 d[1] = cpu_to_le32(r3);
558 d[2] = cpu_to_le32(r1);
559 d[3] = cpu_to_le32(r4);
560}
561EXPORT_SYMBOL_GPL(__serpent_decrypt);
562
563static void serpent_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
564{
565 struct serpent_ctx *ctx = crypto_tfm_ctx(tfm);
566
567 __serpent_decrypt(ctx, dst, src);
568}
569
570static int tnepres_setkey(struct crypto_tfm *tfm, const u8 *key,
571 unsigned int keylen)
572{
573 u8 rev_key[SERPENT_MAX_KEY_SIZE];
574 int i;
575
576 for (i = 0; i < keylen; ++i)
577 rev_key[keylen - i - 1] = key[i];
578
579 return serpent_setkey(tfm, rev_key, keylen);
580}
581
582static void tnepres_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
583{
584 const u32 * const s = (const u32 * const)src;
585 u32 * const d = (u32 * const)dst;
586
587 u32 rs[4], rd[4];
588
589 rs[0] = swab32(s[3]);
590 rs[1] = swab32(s[2]);
591 rs[2] = swab32(s[1]);
592 rs[3] = swab32(s[0]);
593
594 serpent_encrypt(tfm, (u8 *)rd, (u8 *)rs);
595
596 d[0] = swab32(rd[3]);
597 d[1] = swab32(rd[2]);
598 d[2] = swab32(rd[1]);
599 d[3] = swab32(rd[0]);
600}
601
602static void tnepres_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
603{
604 const u32 * const s = (const u32 * const)src;
605 u32 * const d = (u32 * const)dst;
606
607 u32 rs[4], rd[4];
608
609 rs[0] = swab32(s[3]);
610 rs[1] = swab32(s[2]);
611 rs[2] = swab32(s[1]);
612 rs[3] = swab32(s[0]);
613
614 serpent_decrypt(tfm, (u8 *)rd, (u8 *)rs);
615
616 d[0] = swab32(rd[3]);
617 d[1] = swab32(rd[2]);
618 d[2] = swab32(rd[1]);
619 d[3] = swab32(rd[0]);
620}
621
622static struct crypto_alg srp_algs[2] = { {
623 .cra_name = "serpent",
624 .cra_driver_name = "serpent-generic",
625 .cra_priority = 100,
626 .cra_flags = CRYPTO_ALG_TYPE_CIPHER,
627 .cra_blocksize = SERPENT_BLOCK_SIZE,
628 .cra_ctxsize = sizeof(struct serpent_ctx),
629 .cra_alignmask = 3,
630 .cra_module = THIS_MODULE,
631 .cra_u = { .cipher = {
632 .cia_min_keysize = SERPENT_MIN_KEY_SIZE,
633 .cia_max_keysize = SERPENT_MAX_KEY_SIZE,
634 .cia_setkey = serpent_setkey,
635 .cia_encrypt = serpent_encrypt,
636 .cia_decrypt = serpent_decrypt } }
637}, {
638 .cra_name = "tnepres",
639 .cra_flags = CRYPTO_ALG_TYPE_CIPHER,
640 .cra_blocksize = SERPENT_BLOCK_SIZE,
641 .cra_ctxsize = sizeof(struct serpent_ctx),
642 .cra_alignmask = 3,
643 .cra_module = THIS_MODULE,
644 .cra_u = { .cipher = {
645 .cia_min_keysize = SERPENT_MIN_KEY_SIZE,
646 .cia_max_keysize = SERPENT_MAX_KEY_SIZE,
647 .cia_setkey = tnepres_setkey,
648 .cia_encrypt = tnepres_encrypt,
649 .cia_decrypt = tnepres_decrypt } }
650} };
651
652static int __init serpent_mod_init(void)
653{
654 return crypto_register_algs(srp_algs, ARRAY_SIZE(srp_algs));
655}
656
657static void __exit serpent_mod_fini(void)
658{
659 crypto_unregister_algs(srp_algs, ARRAY_SIZE(srp_algs));
660}
661
662module_init(serpent_mod_init);
663module_exit(serpent_mod_fini);
664
665MODULE_LICENSE("GPL");
666MODULE_DESCRIPTION("Serpent and tnepres (kerneli compatible serpent reversed) Cipher Algorithm");
667MODULE_AUTHOR("Dag Arne Osvik <osvik@ii.uib.no>");
668MODULE_ALIAS("tnepres");
669MODULE_ALIAS("serpent");