Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/drivers/crypto/intel/keembay/ocs-hcu.c
29278 views
1
// SPDX-License-Identifier: GPL-2.0-only
2
/*
3
* Intel Keem Bay OCS HCU Crypto Driver.
4
*
5
* Copyright (C) 2018-2020 Intel Corporation
6
*/
7
8
#include <linux/delay.h>
9
#include <linux/device.h>
10
#include <linux/iopoll.h>
11
#include <linux/irq.h>
12
#include <linux/module.h>
13
14
#include <crypto/sha2.h>
15
16
#include "ocs-hcu.h"
17
18
/* Registers. */
19
#define OCS_HCU_MODE 0x00
20
#define OCS_HCU_CHAIN 0x04
21
#define OCS_HCU_OPERATION 0x08
22
#define OCS_HCU_KEY_0 0x0C
23
#define OCS_HCU_ISR 0x50
24
#define OCS_HCU_IER 0x54
25
#define OCS_HCU_STATUS 0x58
26
#define OCS_HCU_MSG_LEN_LO 0x60
27
#define OCS_HCU_MSG_LEN_HI 0x64
28
#define OCS_HCU_KEY_BYTE_ORDER_CFG 0x80
29
#define OCS_HCU_DMA_SRC_ADDR 0x400
30
#define OCS_HCU_DMA_SRC_SIZE 0x408
31
#define OCS_HCU_DMA_DST_SIZE 0x40C
32
#define OCS_HCU_DMA_DMA_MODE 0x410
33
#define OCS_HCU_DMA_NEXT_SRC_DESCR 0x418
34
#define OCS_HCU_DMA_MSI_ISR 0x480
35
#define OCS_HCU_DMA_MSI_IER 0x484
36
#define OCS_HCU_DMA_MSI_MASK 0x488
37
38
/* Register bit definitions. */
39
#define HCU_MODE_ALGO_SHIFT 16
40
#define HCU_MODE_HMAC_SHIFT 22
41
42
#define HCU_STATUS_BUSY BIT(0)
43
44
#define HCU_BYTE_ORDER_SWAP BIT(0)
45
46
#define HCU_IRQ_HASH_DONE BIT(2)
47
#define HCU_IRQ_HASH_ERR_MASK (BIT(3) | BIT(1) | BIT(0))
48
49
#define HCU_DMA_IRQ_SRC_DONE BIT(0)
50
#define HCU_DMA_IRQ_SAI_ERR BIT(2)
51
#define HCU_DMA_IRQ_BAD_COMP_ERR BIT(3)
52
#define HCU_DMA_IRQ_INBUF_RD_ERR BIT(4)
53
#define HCU_DMA_IRQ_INBUF_WD_ERR BIT(5)
54
#define HCU_DMA_IRQ_OUTBUF_WR_ERR BIT(6)
55
#define HCU_DMA_IRQ_OUTBUF_RD_ERR BIT(7)
56
#define HCU_DMA_IRQ_CRD_ERR BIT(8)
57
#define HCU_DMA_IRQ_ERR_MASK (HCU_DMA_IRQ_SAI_ERR | \
58
HCU_DMA_IRQ_BAD_COMP_ERR | \
59
HCU_DMA_IRQ_INBUF_RD_ERR | \
60
HCU_DMA_IRQ_INBUF_WD_ERR | \
61
HCU_DMA_IRQ_OUTBUF_WR_ERR | \
62
HCU_DMA_IRQ_OUTBUF_RD_ERR | \
63
HCU_DMA_IRQ_CRD_ERR)
64
65
#define HCU_DMA_SNOOP_MASK (0x7 << 28)
66
#define HCU_DMA_SRC_LL_EN BIT(25)
67
#define HCU_DMA_EN BIT(31)
68
69
#define OCS_HCU_ENDIANNESS_VALUE 0x2A
70
71
#define HCU_DMA_MSI_UNMASK BIT(0)
72
#define HCU_DMA_MSI_DISABLE 0
73
#define HCU_IRQ_DISABLE 0
74
75
#define OCS_HCU_START BIT(0)
76
#define OCS_HCU_TERMINATE BIT(1)
77
78
#define OCS_LL_DMA_FLAG_TERMINATE BIT(31)
79
80
#define OCS_HCU_HW_KEY_LEN_U32 (OCS_HCU_HW_KEY_LEN / sizeof(u32))
81
82
#define HCU_DATA_WRITE_ENDIANNESS_OFFSET 26
83
84
#define OCS_HCU_NUM_CHAINS_SHA256_224_SM3 (SHA256_DIGEST_SIZE / sizeof(u32))
85
#define OCS_HCU_NUM_CHAINS_SHA384_512 (SHA512_DIGEST_SIZE / sizeof(u32))
86
87
/*
88
* While polling on a busy HCU, wait maximum 200us between one check and the
89
* other.
90
*/
91
#define OCS_HCU_WAIT_BUSY_RETRY_DELAY_US 200
92
/* Wait on a busy HCU for maximum 1 second. */
93
#define OCS_HCU_WAIT_BUSY_TIMEOUT_US 1000000
94
95
/**
96
* struct ocs_hcu_dma_entry - An entry in an OCS DMA linked list.
97
* @src_addr: Source address of the data.
98
* @src_len: Length of data to be fetched.
99
* @nxt_desc: Next descriptor to fetch.
100
* @ll_flags: Flags (Freeze @ terminate) for the DMA engine.
101
*/
102
struct ocs_hcu_dma_entry {
103
u32 src_addr;
104
u32 src_len;
105
u32 nxt_desc;
106
u32 ll_flags;
107
};
108
109
/**
110
* struct ocs_hcu_dma_list - OCS-specific DMA linked list.
111
* @head: The head of the list (points to the array backing the list).
112
* @tail: The current tail of the list; NULL if the list is empty.
113
* @dma_addr: The DMA address of @head (i.e., the DMA address of the backing
114
* array).
115
* @max_nents: Maximum number of entries in the list (i.e., number of elements
116
* in the backing array).
117
*
118
* The OCS DMA list is an array-backed list of OCS DMA descriptors. The array
119
* backing the list is allocated with dma_alloc_coherent() and pointed by
120
* @head.
121
*/
122
struct ocs_hcu_dma_list {
123
struct ocs_hcu_dma_entry *head;
124
struct ocs_hcu_dma_entry *tail;
125
dma_addr_t dma_addr;
126
size_t max_nents;
127
};
128
129
static inline u32 ocs_hcu_num_chains(enum ocs_hcu_algo algo)
130
{
131
switch (algo) {
132
case OCS_HCU_ALGO_SHA224:
133
case OCS_HCU_ALGO_SHA256:
134
case OCS_HCU_ALGO_SM3:
135
return OCS_HCU_NUM_CHAINS_SHA256_224_SM3;
136
case OCS_HCU_ALGO_SHA384:
137
case OCS_HCU_ALGO_SHA512:
138
return OCS_HCU_NUM_CHAINS_SHA384_512;
139
default:
140
return 0;
141
};
142
}
143
144
static inline u32 ocs_hcu_digest_size(enum ocs_hcu_algo algo)
145
{
146
switch (algo) {
147
case OCS_HCU_ALGO_SHA224:
148
return SHA224_DIGEST_SIZE;
149
case OCS_HCU_ALGO_SHA256:
150
case OCS_HCU_ALGO_SM3:
151
/* SM3 shares the same block size. */
152
return SHA256_DIGEST_SIZE;
153
case OCS_HCU_ALGO_SHA384:
154
return SHA384_DIGEST_SIZE;
155
case OCS_HCU_ALGO_SHA512:
156
return SHA512_DIGEST_SIZE;
157
default:
158
return 0;
159
}
160
}
161
162
/**
163
* ocs_hcu_wait_busy() - Wait for HCU OCS hardware to became usable.
164
* @hcu_dev: OCS HCU device to wait for.
165
*
166
* Return: 0 if device free, -ETIMEOUT if device busy and internal timeout has
167
* expired.
168
*/
169
static int ocs_hcu_wait_busy(struct ocs_hcu_dev *hcu_dev)
170
{
171
long val;
172
173
return readl_poll_timeout(hcu_dev->io_base + OCS_HCU_STATUS, val,
174
!(val & HCU_STATUS_BUSY),
175
OCS_HCU_WAIT_BUSY_RETRY_DELAY_US,
176
OCS_HCU_WAIT_BUSY_TIMEOUT_US);
177
}
178
179
static void ocs_hcu_done_irq_en(struct ocs_hcu_dev *hcu_dev)
180
{
181
/* Clear any pending interrupts. */
182
writel(0xFFFFFFFF, hcu_dev->io_base + OCS_HCU_ISR);
183
hcu_dev->irq_err = false;
184
/* Enable error and HCU done interrupts. */
185
writel(HCU_IRQ_HASH_DONE | HCU_IRQ_HASH_ERR_MASK,
186
hcu_dev->io_base + OCS_HCU_IER);
187
}
188
189
static void ocs_hcu_dma_irq_en(struct ocs_hcu_dev *hcu_dev)
190
{
191
/* Clear any pending interrupts. */
192
writel(0xFFFFFFFF, hcu_dev->io_base + OCS_HCU_DMA_MSI_ISR);
193
hcu_dev->irq_err = false;
194
/* Only operating on DMA source completion and error interrupts. */
195
writel(HCU_DMA_IRQ_ERR_MASK | HCU_DMA_IRQ_SRC_DONE,
196
hcu_dev->io_base + OCS_HCU_DMA_MSI_IER);
197
/* Unmask */
198
writel(HCU_DMA_MSI_UNMASK, hcu_dev->io_base + OCS_HCU_DMA_MSI_MASK);
199
}
200
201
static void ocs_hcu_irq_dis(struct ocs_hcu_dev *hcu_dev)
202
{
203
writel(HCU_IRQ_DISABLE, hcu_dev->io_base + OCS_HCU_IER);
204
writel(HCU_DMA_MSI_DISABLE, hcu_dev->io_base + OCS_HCU_DMA_MSI_IER);
205
}
206
207
static int ocs_hcu_wait_and_disable_irq(struct ocs_hcu_dev *hcu_dev)
208
{
209
int rc;
210
211
rc = wait_for_completion_interruptible(&hcu_dev->irq_done);
212
if (rc)
213
goto exit;
214
215
if (hcu_dev->irq_err) {
216
/* Unset flag and return error. */
217
hcu_dev->irq_err = false;
218
rc = -EIO;
219
goto exit;
220
}
221
222
exit:
223
ocs_hcu_irq_dis(hcu_dev);
224
225
return rc;
226
}
227
228
/**
229
* ocs_hcu_get_intermediate_data() - Get intermediate data.
230
* @hcu_dev: The target HCU device.
231
* @data: Where to store the intermediate.
232
* @algo: The algorithm being used.
233
*
234
* This function is used to save the current hashing process state in order to
235
* continue it in the future.
236
*
237
* Note: once all data has been processed, the intermediate data actually
238
* contains the hashing result. So this function is also used to retrieve the
239
* final result of a hashing process.
240
*
241
* Return: 0 on success, negative error code otherwise.
242
*/
243
static int ocs_hcu_get_intermediate_data(struct ocs_hcu_dev *hcu_dev,
244
struct ocs_hcu_idata *data,
245
enum ocs_hcu_algo algo)
246
{
247
const int n = ocs_hcu_num_chains(algo);
248
u32 *chain;
249
int rc;
250
int i;
251
252
/* Data not requested. */
253
if (!data)
254
return -EINVAL;
255
256
chain = (u32 *)data->digest;
257
258
/* Ensure that the OCS is no longer busy before reading the chains. */
259
rc = ocs_hcu_wait_busy(hcu_dev);
260
if (rc)
261
return rc;
262
263
/*
264
* This loops is safe because data->digest is an array of
265
* SHA512_DIGEST_SIZE bytes and the maximum value returned by
266
* ocs_hcu_num_chains() is OCS_HCU_NUM_CHAINS_SHA384_512 which is equal
267
* to SHA512_DIGEST_SIZE / sizeof(u32).
268
*/
269
for (i = 0; i < n; i++)
270
chain[i] = readl(hcu_dev->io_base + OCS_HCU_CHAIN);
271
272
data->msg_len_lo = readl(hcu_dev->io_base + OCS_HCU_MSG_LEN_LO);
273
data->msg_len_hi = readl(hcu_dev->io_base + OCS_HCU_MSG_LEN_HI);
274
275
return 0;
276
}
277
278
/**
279
* ocs_hcu_set_intermediate_data() - Set intermediate data.
280
* @hcu_dev: The target HCU device.
281
* @data: The intermediate data to be set.
282
* @algo: The algorithm being used.
283
*
284
* This function is used to continue a previous hashing process.
285
*/
286
static void ocs_hcu_set_intermediate_data(struct ocs_hcu_dev *hcu_dev,
287
const struct ocs_hcu_idata *data,
288
enum ocs_hcu_algo algo)
289
{
290
const int n = ocs_hcu_num_chains(algo);
291
u32 *chain = (u32 *)data->digest;
292
int i;
293
294
/*
295
* This loops is safe because data->digest is an array of
296
* SHA512_DIGEST_SIZE bytes and the maximum value returned by
297
* ocs_hcu_num_chains() is OCS_HCU_NUM_CHAINS_SHA384_512 which is equal
298
* to SHA512_DIGEST_SIZE / sizeof(u32).
299
*/
300
for (i = 0; i < n; i++)
301
writel(chain[i], hcu_dev->io_base + OCS_HCU_CHAIN);
302
303
writel(data->msg_len_lo, hcu_dev->io_base + OCS_HCU_MSG_LEN_LO);
304
writel(data->msg_len_hi, hcu_dev->io_base + OCS_HCU_MSG_LEN_HI);
305
}
306
307
static int ocs_hcu_get_digest(struct ocs_hcu_dev *hcu_dev,
308
enum ocs_hcu_algo algo, u8 *dgst, size_t dgst_len)
309
{
310
u32 *chain;
311
int rc;
312
int i;
313
314
if (!dgst)
315
return -EINVAL;
316
317
/* Length of the output buffer must match the algo digest size. */
318
if (dgst_len != ocs_hcu_digest_size(algo))
319
return -EINVAL;
320
321
/* Ensure that the OCS is no longer busy before reading the chains. */
322
rc = ocs_hcu_wait_busy(hcu_dev);
323
if (rc)
324
return rc;
325
326
chain = (u32 *)dgst;
327
for (i = 0; i < dgst_len / sizeof(u32); i++)
328
chain[i] = readl(hcu_dev->io_base + OCS_HCU_CHAIN);
329
330
return 0;
331
}
332
333
/**
334
* ocs_hcu_hw_cfg() - Configure the HCU hardware.
335
* @hcu_dev: The HCU device to configure.
336
* @algo: The algorithm to be used by the HCU device.
337
* @use_hmac: Whether or not HW HMAC should be used.
338
*
339
* Return: 0 on success, negative error code otherwise.
340
*/
341
static int ocs_hcu_hw_cfg(struct ocs_hcu_dev *hcu_dev, enum ocs_hcu_algo algo,
342
bool use_hmac)
343
{
344
u32 cfg;
345
int rc;
346
347
if (algo != OCS_HCU_ALGO_SHA256 && algo != OCS_HCU_ALGO_SHA224 &&
348
algo != OCS_HCU_ALGO_SHA384 && algo != OCS_HCU_ALGO_SHA512 &&
349
algo != OCS_HCU_ALGO_SM3)
350
return -EINVAL;
351
352
rc = ocs_hcu_wait_busy(hcu_dev);
353
if (rc)
354
return rc;
355
356
/* Ensure interrupts are disabled. */
357
ocs_hcu_irq_dis(hcu_dev);
358
359
/* Configure endianness, hashing algorithm and HW HMAC (if needed) */
360
cfg = OCS_HCU_ENDIANNESS_VALUE << HCU_DATA_WRITE_ENDIANNESS_OFFSET;
361
cfg |= algo << HCU_MODE_ALGO_SHIFT;
362
if (use_hmac)
363
cfg |= BIT(HCU_MODE_HMAC_SHIFT);
364
365
writel(cfg, hcu_dev->io_base + OCS_HCU_MODE);
366
367
return 0;
368
}
369
370
/**
371
* ocs_hcu_clear_key() - Clear key stored in OCS HMAC KEY registers.
372
* @hcu_dev: The OCS HCU device whose key registers should be cleared.
373
*/
374
static void ocs_hcu_clear_key(struct ocs_hcu_dev *hcu_dev)
375
{
376
int reg_off;
377
378
/* Clear OCS_HCU_KEY_[0..15] */
379
for (reg_off = 0; reg_off < OCS_HCU_HW_KEY_LEN; reg_off += sizeof(u32))
380
writel(0, hcu_dev->io_base + OCS_HCU_KEY_0 + reg_off);
381
}
382
383
/**
384
* ocs_hcu_write_key() - Write key to OCS HMAC KEY registers.
385
* @hcu_dev: The OCS HCU device the key should be written to.
386
* @key: The key to be written.
387
* @len: The size of the key to write. It must be OCS_HCU_HW_KEY_LEN.
388
*
389
* Return: 0 on success, negative error code otherwise.
390
*/
391
static int ocs_hcu_write_key(struct ocs_hcu_dev *hcu_dev, const u8 *key, size_t len)
392
{
393
u32 key_u32[OCS_HCU_HW_KEY_LEN_U32];
394
int i;
395
396
if (len > OCS_HCU_HW_KEY_LEN)
397
return -EINVAL;
398
399
/* Copy key into temporary u32 array. */
400
memcpy(key_u32, key, len);
401
402
/*
403
* Hardware requires all the bytes of the HW Key vector to be
404
* written. So pad with zero until we reach OCS_HCU_HW_KEY_LEN.
405
*/
406
memzero_explicit((u8 *)key_u32 + len, OCS_HCU_HW_KEY_LEN - len);
407
408
/*
409
* OCS hardware expects the MSB of the key to be written at the highest
410
* address of the HCU Key vector; in other word, the key must be
411
* written in reverse order.
412
*
413
* Therefore, we first enable byte swapping for the HCU key vector;
414
* so that bytes of 32-bit word written to OCS_HCU_KEY_[0..15] will be
415
* swapped:
416
* 3 <---> 0, 2 <---> 1.
417
*/
418
writel(HCU_BYTE_ORDER_SWAP,
419
hcu_dev->io_base + OCS_HCU_KEY_BYTE_ORDER_CFG);
420
/*
421
* And then we write the 32-bit words composing the key starting from
422
* the end of the key.
423
*/
424
for (i = 0; i < OCS_HCU_HW_KEY_LEN_U32; i++)
425
writel(key_u32[OCS_HCU_HW_KEY_LEN_U32 - 1 - i],
426
hcu_dev->io_base + OCS_HCU_KEY_0 + (sizeof(u32) * i));
427
428
memzero_explicit(key_u32, OCS_HCU_HW_KEY_LEN);
429
430
return 0;
431
}
432
433
/**
434
* ocs_hcu_ll_dma_start() - Start OCS HCU hashing via DMA
435
* @hcu_dev: The OCS HCU device to use.
436
* @dma_list: The OCS DMA list mapping the data to hash.
437
* @finalize: Whether or not this is the last hashing operation and therefore
438
* the final hash should be compute even if data is not
439
* block-aligned.
440
*
441
* Return: 0 on success, negative error code otherwise.
442
*/
443
static int ocs_hcu_ll_dma_start(struct ocs_hcu_dev *hcu_dev,
444
const struct ocs_hcu_dma_list *dma_list,
445
bool finalize)
446
{
447
u32 cfg = HCU_DMA_SNOOP_MASK | HCU_DMA_SRC_LL_EN | HCU_DMA_EN;
448
int rc;
449
450
if (!dma_list)
451
return -EINVAL;
452
453
/*
454
* For final requests we use HCU_DONE IRQ to be notified when all input
455
* data has been processed by the HCU; however, we cannot do so for
456
* non-final requests, because we don't get a HCU_DONE IRQ when we
457
* don't terminate the operation.
458
*
459
* Therefore, for non-final requests, we use the DMA IRQ, which
460
* triggers when DMA has finishing feeding all the input data to the
461
* HCU, but the HCU may still be processing it. This is fine, since we
462
* will wait for the HCU processing to be completed when we try to read
463
* intermediate results, in ocs_hcu_get_intermediate_data().
464
*/
465
if (finalize)
466
ocs_hcu_done_irq_en(hcu_dev);
467
else
468
ocs_hcu_dma_irq_en(hcu_dev);
469
470
reinit_completion(&hcu_dev->irq_done);
471
writel(dma_list->dma_addr, hcu_dev->io_base + OCS_HCU_DMA_NEXT_SRC_DESCR);
472
writel(0, hcu_dev->io_base + OCS_HCU_DMA_SRC_SIZE);
473
writel(0, hcu_dev->io_base + OCS_HCU_DMA_DST_SIZE);
474
475
writel(OCS_HCU_START, hcu_dev->io_base + OCS_HCU_OPERATION);
476
477
writel(cfg, hcu_dev->io_base + OCS_HCU_DMA_DMA_MODE);
478
479
if (finalize)
480
writel(OCS_HCU_TERMINATE, hcu_dev->io_base + OCS_HCU_OPERATION);
481
482
rc = ocs_hcu_wait_and_disable_irq(hcu_dev);
483
if (rc)
484
return rc;
485
486
return 0;
487
}
488
489
struct ocs_hcu_dma_list *ocs_hcu_dma_list_alloc(struct ocs_hcu_dev *hcu_dev,
490
int max_nents)
491
{
492
struct ocs_hcu_dma_list *dma_list;
493
494
dma_list = kmalloc(sizeof(*dma_list), GFP_KERNEL);
495
if (!dma_list)
496
return NULL;
497
498
/* Total size of the DMA list to allocate. */
499
dma_list->head = dma_alloc_coherent(hcu_dev->dev,
500
sizeof(*dma_list->head) * max_nents,
501
&dma_list->dma_addr, GFP_KERNEL);
502
if (!dma_list->head) {
503
kfree(dma_list);
504
return NULL;
505
}
506
dma_list->max_nents = max_nents;
507
dma_list->tail = NULL;
508
509
return dma_list;
510
}
511
512
void ocs_hcu_dma_list_free(struct ocs_hcu_dev *hcu_dev,
513
struct ocs_hcu_dma_list *dma_list)
514
{
515
if (!dma_list)
516
return;
517
518
dma_free_coherent(hcu_dev->dev,
519
sizeof(*dma_list->head) * dma_list->max_nents,
520
dma_list->head, dma_list->dma_addr);
521
522
kfree(dma_list);
523
}
524
525
/* Add a new DMA entry at the end of the OCS DMA list. */
526
int ocs_hcu_dma_list_add_tail(struct ocs_hcu_dev *hcu_dev,
527
struct ocs_hcu_dma_list *dma_list,
528
dma_addr_t addr, u32 len)
529
{
530
struct device *dev = hcu_dev->dev;
531
struct ocs_hcu_dma_entry *old_tail;
532
struct ocs_hcu_dma_entry *new_tail;
533
534
if (!len)
535
return 0;
536
537
if (!dma_list)
538
return -EINVAL;
539
540
if (addr & ~OCS_HCU_DMA_BIT_MASK) {
541
dev_err(dev,
542
"Unexpected error: Invalid DMA address for OCS HCU\n");
543
return -EINVAL;
544
}
545
546
old_tail = dma_list->tail;
547
new_tail = old_tail ? old_tail + 1 : dma_list->head;
548
549
/* Check if list is full. */
550
if (new_tail - dma_list->head >= dma_list->max_nents)
551
return -ENOMEM;
552
553
/*
554
* If there was an old tail (i.e., this is not the first element we are
555
* adding), un-terminate the old tail and make it point to the new one.
556
*/
557
if (old_tail) {
558
old_tail->ll_flags &= ~OCS_LL_DMA_FLAG_TERMINATE;
559
/*
560
* The old tail 'nxt_desc' must point to the DMA address of the
561
* new tail.
562
*/
563
old_tail->nxt_desc = dma_list->dma_addr +
564
sizeof(*dma_list->tail) * (new_tail -
565
dma_list->head);
566
}
567
568
new_tail->src_addr = (u32)addr;
569
new_tail->src_len = (u32)len;
570
new_tail->ll_flags = OCS_LL_DMA_FLAG_TERMINATE;
571
new_tail->nxt_desc = 0;
572
573
/* Update list tail with new tail. */
574
dma_list->tail = new_tail;
575
576
return 0;
577
}
578
579
/**
580
* ocs_hcu_hash_init() - Initialize hash operation context.
581
* @ctx: The context to initialize.
582
* @algo: The hashing algorithm to use.
583
*
584
* Return: 0 on success, negative error code otherwise.
585
*/
586
int ocs_hcu_hash_init(struct ocs_hcu_hash_ctx *ctx, enum ocs_hcu_algo algo)
587
{
588
if (!ctx)
589
return -EINVAL;
590
591
ctx->algo = algo;
592
ctx->idata.msg_len_lo = 0;
593
ctx->idata.msg_len_hi = 0;
594
/* No need to set idata.digest to 0. */
595
596
return 0;
597
}
598
599
/**
600
* ocs_hcu_hash_update() - Perform a hashing iteration.
601
* @hcu_dev: The OCS HCU device to use.
602
* @ctx: The OCS HCU hashing context.
603
* @dma_list: The OCS DMA list mapping the input data to process.
604
*
605
* Return: 0 on success; negative error code otherwise.
606
*/
607
int ocs_hcu_hash_update(struct ocs_hcu_dev *hcu_dev,
608
struct ocs_hcu_hash_ctx *ctx,
609
const struct ocs_hcu_dma_list *dma_list)
610
{
611
int rc;
612
613
if (!hcu_dev || !ctx)
614
return -EINVAL;
615
616
/* Configure the hardware for the current request. */
617
rc = ocs_hcu_hw_cfg(hcu_dev, ctx->algo, false);
618
if (rc)
619
return rc;
620
621
/* If we already processed some data, idata needs to be set. */
622
if (ctx->idata.msg_len_lo || ctx->idata.msg_len_hi)
623
ocs_hcu_set_intermediate_data(hcu_dev, &ctx->idata, ctx->algo);
624
625
/* Start linked-list DMA hashing. */
626
rc = ocs_hcu_ll_dma_start(hcu_dev, dma_list, false);
627
if (rc)
628
return rc;
629
630
/* Update idata and return. */
631
return ocs_hcu_get_intermediate_data(hcu_dev, &ctx->idata, ctx->algo);
632
}
633
634
/**
635
* ocs_hcu_hash_finup() - Update and finalize hash computation.
636
* @hcu_dev: The OCS HCU device to use.
637
* @ctx: The OCS HCU hashing context.
638
* @dma_list: The OCS DMA list mapping the input data to process.
639
* @dgst: The buffer where to save the computed digest.
640
* @dgst_len: The length of @dgst.
641
*
642
* Return: 0 on success; negative error code otherwise.
643
*/
644
int ocs_hcu_hash_finup(struct ocs_hcu_dev *hcu_dev,
645
const struct ocs_hcu_hash_ctx *ctx,
646
const struct ocs_hcu_dma_list *dma_list,
647
u8 *dgst, size_t dgst_len)
648
{
649
int rc;
650
651
if (!hcu_dev || !ctx)
652
return -EINVAL;
653
654
/* Configure the hardware for the current request. */
655
rc = ocs_hcu_hw_cfg(hcu_dev, ctx->algo, false);
656
if (rc)
657
return rc;
658
659
/* If we already processed some data, idata needs to be set. */
660
if (ctx->idata.msg_len_lo || ctx->idata.msg_len_hi)
661
ocs_hcu_set_intermediate_data(hcu_dev, &ctx->idata, ctx->algo);
662
663
/* Start linked-list DMA hashing. */
664
rc = ocs_hcu_ll_dma_start(hcu_dev, dma_list, true);
665
if (rc)
666
return rc;
667
668
/* Get digest and return. */
669
return ocs_hcu_get_digest(hcu_dev, ctx->algo, dgst, dgst_len);
670
}
671
672
/**
673
* ocs_hcu_hash_final() - Finalize hash computation.
674
* @hcu_dev: The OCS HCU device to use.
675
* @ctx: The OCS HCU hashing context.
676
* @dgst: The buffer where to save the computed digest.
677
* @dgst_len: The length of @dgst.
678
*
679
* Return: 0 on success; negative error code otherwise.
680
*/
681
int ocs_hcu_hash_final(struct ocs_hcu_dev *hcu_dev,
682
const struct ocs_hcu_hash_ctx *ctx, u8 *dgst,
683
size_t dgst_len)
684
{
685
int rc;
686
687
if (!hcu_dev || !ctx)
688
return -EINVAL;
689
690
/* Configure the hardware for the current request. */
691
rc = ocs_hcu_hw_cfg(hcu_dev, ctx->algo, false);
692
if (rc)
693
return rc;
694
695
/* If we already processed some data, idata needs to be set. */
696
if (ctx->idata.msg_len_lo || ctx->idata.msg_len_hi)
697
ocs_hcu_set_intermediate_data(hcu_dev, &ctx->idata, ctx->algo);
698
699
/*
700
* Enable HCU interrupts, so that HCU_DONE will be triggered once the
701
* final hash is computed.
702
*/
703
ocs_hcu_done_irq_en(hcu_dev);
704
reinit_completion(&hcu_dev->irq_done);
705
writel(OCS_HCU_TERMINATE, hcu_dev->io_base + OCS_HCU_OPERATION);
706
707
rc = ocs_hcu_wait_and_disable_irq(hcu_dev);
708
if (rc)
709
return rc;
710
711
/* Get digest and return. */
712
return ocs_hcu_get_digest(hcu_dev, ctx->algo, dgst, dgst_len);
713
}
714
715
/**
716
* ocs_hcu_digest() - Compute hash digest.
717
* @hcu_dev: The OCS HCU device to use.
718
* @algo: The hash algorithm to use.
719
* @data: The input data to process.
720
* @data_len: The length of @data.
721
* @dgst: The buffer where to save the computed digest.
722
* @dgst_len: The length of @dgst.
723
*
724
* Return: 0 on success; negative error code otherwise.
725
*/
726
int ocs_hcu_digest(struct ocs_hcu_dev *hcu_dev, enum ocs_hcu_algo algo,
727
void *data, size_t data_len, u8 *dgst, size_t dgst_len)
728
{
729
struct device *dev = hcu_dev->dev;
730
dma_addr_t dma_handle;
731
u32 reg;
732
int rc;
733
734
/* Configure the hardware for the current request. */
735
rc = ocs_hcu_hw_cfg(hcu_dev, algo, false);
736
if (rc)
737
return rc;
738
739
dma_handle = dma_map_single(dev, data, data_len, DMA_TO_DEVICE);
740
if (dma_mapping_error(dev, dma_handle))
741
return -EIO;
742
743
reg = HCU_DMA_SNOOP_MASK | HCU_DMA_EN;
744
745
ocs_hcu_done_irq_en(hcu_dev);
746
747
reinit_completion(&hcu_dev->irq_done);
748
749
writel(dma_handle, hcu_dev->io_base + OCS_HCU_DMA_SRC_ADDR);
750
writel(data_len, hcu_dev->io_base + OCS_HCU_DMA_SRC_SIZE);
751
writel(OCS_HCU_START, hcu_dev->io_base + OCS_HCU_OPERATION);
752
writel(reg, hcu_dev->io_base + OCS_HCU_DMA_DMA_MODE);
753
754
writel(OCS_HCU_TERMINATE, hcu_dev->io_base + OCS_HCU_OPERATION);
755
756
rc = ocs_hcu_wait_and_disable_irq(hcu_dev);
757
if (rc)
758
return rc;
759
760
dma_unmap_single(dev, dma_handle, data_len, DMA_TO_DEVICE);
761
762
return ocs_hcu_get_digest(hcu_dev, algo, dgst, dgst_len);
763
}
764
765
/**
766
* ocs_hcu_hmac() - Compute HMAC.
767
* @hcu_dev: The OCS HCU device to use.
768
* @algo: The hash algorithm to use with HMAC.
769
* @key: The key to use.
770
* @dma_list: The OCS DMA list mapping the input data to process.
771
* @key_len: The length of @key.
772
* @dgst: The buffer where to save the computed HMAC.
773
* @dgst_len: The length of @dgst.
774
*
775
* Return: 0 on success; negative error code otherwise.
776
*/
777
int ocs_hcu_hmac(struct ocs_hcu_dev *hcu_dev, enum ocs_hcu_algo algo,
778
const u8 *key, size_t key_len,
779
const struct ocs_hcu_dma_list *dma_list,
780
u8 *dgst, size_t dgst_len)
781
{
782
int rc;
783
784
/* Ensure 'key' is not NULL. */
785
if (!key || key_len == 0)
786
return -EINVAL;
787
788
/* Configure the hardware for the current request. */
789
rc = ocs_hcu_hw_cfg(hcu_dev, algo, true);
790
if (rc)
791
return rc;
792
793
rc = ocs_hcu_write_key(hcu_dev, key, key_len);
794
if (rc)
795
return rc;
796
797
rc = ocs_hcu_ll_dma_start(hcu_dev, dma_list, true);
798
799
/* Clear HW key before processing return code. */
800
ocs_hcu_clear_key(hcu_dev);
801
802
if (rc)
803
return rc;
804
805
return ocs_hcu_get_digest(hcu_dev, algo, dgst, dgst_len);
806
}
807
808
irqreturn_t ocs_hcu_irq_handler(int irq, void *dev_id)
809
{
810
struct ocs_hcu_dev *hcu_dev = dev_id;
811
u32 hcu_irq;
812
u32 dma_irq;
813
814
/* Read and clear the HCU interrupt. */
815
hcu_irq = readl(hcu_dev->io_base + OCS_HCU_ISR);
816
writel(hcu_irq, hcu_dev->io_base + OCS_HCU_ISR);
817
818
/* Read and clear the HCU DMA interrupt. */
819
dma_irq = readl(hcu_dev->io_base + OCS_HCU_DMA_MSI_ISR);
820
writel(dma_irq, hcu_dev->io_base + OCS_HCU_DMA_MSI_ISR);
821
822
/* Check for errors. */
823
if (hcu_irq & HCU_IRQ_HASH_ERR_MASK || dma_irq & HCU_DMA_IRQ_ERR_MASK) {
824
hcu_dev->irq_err = true;
825
goto complete;
826
}
827
828
/* Check for DONE IRQs. */
829
if (hcu_irq & HCU_IRQ_HASH_DONE || dma_irq & HCU_DMA_IRQ_SRC_DONE)
830
goto complete;
831
832
return IRQ_NONE;
833
834
complete:
835
complete(&hcu_dev->irq_done);
836
837
return IRQ_HANDLED;
838
}
839
840
MODULE_DESCRIPTION("Intel Keem Bay OCS HCU Crypto Driver");
841
MODULE_LICENSE("GPL");
842
843