GNU Linux-libre 6.8.7-gnu
[releases.git] / drivers / crypto / caam / caamalg_qi2.c
1 // SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
2 /*
3  * Copyright 2015-2016 Freescale Semiconductor Inc.
4  * Copyright 2017-2019 NXP
5  */
6
7 #include "compat.h"
8 #include "regs.h"
9 #include "caamalg_qi2.h"
10 #include "dpseci_cmd.h"
11 #include "desc_constr.h"
12 #include "error.h"
13 #include "sg_sw_sec4.h"
14 #include "sg_sw_qm2.h"
15 #include "key_gen.h"
16 #include "caamalg_desc.h"
17 #include "caamhash_desc.h"
18 #include "dpseci-debugfs.h"
19 #include <linux/dma-mapping.h>
20 #include <linux/fsl/mc.h>
21 #include <linux/kernel.h>
22 #include <soc/fsl/dpaa2-io.h>
23 #include <soc/fsl/dpaa2-fd.h>
24 #include <crypto/xts.h>
25 #include <asm/unaligned.h>
26
27 #define CAAM_CRA_PRIORITY       2000
28
29 /* max key is sum of AES_MAX_KEY_SIZE, max split key size */
30 #define CAAM_MAX_KEY_SIZE       (AES_MAX_KEY_SIZE + CTR_RFC3686_NONCE_SIZE + \
31                                  SHA512_DIGEST_SIZE * 2)
32
33 /*
34  * This is a cache of buffers, from which the users of CAAM QI driver
35  * can allocate short buffers. It's speedier than doing kmalloc on the hotpath.
36  * NOTE: A more elegant solution would be to have some headroom in the frames
37  *       being processed. This can be added by the dpaa2-eth driver. This would
38  *       pose a problem for userspace application processing which cannot
39  *       know of this limitation. So for now, this will work.
40  * NOTE: The memcache is SMP-safe. No need to handle spinlocks in-here
41  */
42 static struct kmem_cache *qi_cache;
43
44 struct caam_alg_entry {
45         struct device *dev;
46         int class1_alg_type;
47         int class2_alg_type;
48         bool rfc3686;
49         bool geniv;
50         bool nodkp;
51 };
52
53 struct caam_aead_alg {
54         struct aead_alg aead;
55         struct caam_alg_entry caam;
56         bool registered;
57 };
58
59 struct caam_skcipher_alg {
60         struct skcipher_alg skcipher;
61         struct caam_alg_entry caam;
62         bool registered;
63 };
64
65 /**
66  * struct caam_ctx - per-session context
67  * @flc: Flow Contexts array
68  * @key:  [authentication key], encryption key
69  * @flc_dma: I/O virtual addresses of the Flow Contexts
70  * @key_dma: I/O virtual address of the key
71  * @dir: DMA direction for mapping key and Flow Contexts
72  * @dev: dpseci device
73  * @adata: authentication algorithm details
74  * @cdata: encryption algorithm details
75  * @authsize: authentication tag (a.k.a. ICV / MAC) size
76  * @xts_key_fallback: true if fallback tfm needs to be used due
77  *                    to unsupported xts key lengths
78  * @fallback: xts fallback tfm
79  */
80 struct caam_ctx {
81         struct caam_flc flc[NUM_OP];
82         u8 key[CAAM_MAX_KEY_SIZE];
83         dma_addr_t flc_dma[NUM_OP];
84         dma_addr_t key_dma;
85         enum dma_data_direction dir;
86         struct device *dev;
87         struct alginfo adata;
88         struct alginfo cdata;
89         unsigned int authsize;
90         bool xts_key_fallback;
91         struct crypto_skcipher *fallback;
92 };
93
94 static void *dpaa2_caam_iova_to_virt(struct dpaa2_caam_priv *priv,
95                                      dma_addr_t iova_addr)
96 {
97         phys_addr_t phys_addr;
98
99         phys_addr = priv->domain ? iommu_iova_to_phys(priv->domain, iova_addr) :
100                                    iova_addr;
101
102         return phys_to_virt(phys_addr);
103 }
104
105 /*
106  * qi_cache_zalloc - Allocate buffers from CAAM-QI cache
107  *
108  * Allocate data on the hotpath. Instead of using kzalloc, one can use the
109  * services of the CAAM QI memory cache (backed by kmem_cache). The buffers
110  * will have a size of CAAM_QI_MEMCACHE_SIZE, which should be sufficient for
111  * hosting 16 SG entries.
112  *
113  * @flags - flags that would be used for the equivalent kmalloc(..) call
114  *
115  * Returns a pointer to a retrieved buffer on success or NULL on failure.
116  */
117 static inline void *qi_cache_zalloc(gfp_t flags)
118 {
119         return kmem_cache_zalloc(qi_cache, flags);
120 }
121
122 /*
123  * qi_cache_free - Frees buffers allocated from CAAM-QI cache
124  *
125  * @obj - buffer previously allocated by qi_cache_zalloc
126  *
127  * No checking is being done, the call is a passthrough call to
128  * kmem_cache_free(...)
129  */
130 static inline void qi_cache_free(void *obj)
131 {
132         kmem_cache_free(qi_cache, obj);
133 }
134
135 static struct caam_request *to_caam_req(struct crypto_async_request *areq)
136 {
137         switch (crypto_tfm_alg_type(areq->tfm)) {
138         case CRYPTO_ALG_TYPE_SKCIPHER:
139                 return skcipher_request_ctx_dma(skcipher_request_cast(areq));
140         case CRYPTO_ALG_TYPE_AEAD:
141                 return aead_request_ctx_dma(
142                         container_of(areq, struct aead_request, base));
143         case CRYPTO_ALG_TYPE_AHASH:
144                 return ahash_request_ctx_dma(ahash_request_cast(areq));
145         default:
146                 return ERR_PTR(-EINVAL);
147         }
148 }
149
150 static void caam_unmap(struct device *dev, struct scatterlist *src,
151                        struct scatterlist *dst, int src_nents,
152                        int dst_nents, dma_addr_t iv_dma, int ivsize,
153                        enum dma_data_direction iv_dir, dma_addr_t qm_sg_dma,
154                        int qm_sg_bytes)
155 {
156         if (dst != src) {
157                 if (src_nents)
158                         dma_unmap_sg(dev, src, src_nents, DMA_TO_DEVICE);
159                 if (dst_nents)
160                         dma_unmap_sg(dev, dst, dst_nents, DMA_FROM_DEVICE);
161         } else {
162                 dma_unmap_sg(dev, src, src_nents, DMA_BIDIRECTIONAL);
163         }
164
165         if (iv_dma)
166                 dma_unmap_single(dev, iv_dma, ivsize, iv_dir);
167
168         if (qm_sg_bytes)
169                 dma_unmap_single(dev, qm_sg_dma, qm_sg_bytes, DMA_TO_DEVICE);
170 }
171
172 static int aead_set_sh_desc(struct crypto_aead *aead)
173 {
174         struct caam_aead_alg *alg = container_of(crypto_aead_alg(aead),
175                                                  typeof(*alg), aead);
176         struct caam_ctx *ctx = crypto_aead_ctx_dma(aead);
177         unsigned int ivsize = crypto_aead_ivsize(aead);
178         struct device *dev = ctx->dev;
179         struct dpaa2_caam_priv *priv = dev_get_drvdata(dev);
180         struct caam_flc *flc;
181         u32 *desc;
182         u32 ctx1_iv_off = 0;
183         u32 *nonce = NULL;
184         unsigned int data_len[2];
185         u32 inl_mask;
186         const bool ctr_mode = ((ctx->cdata.algtype & OP_ALG_AAI_MASK) ==
187                                OP_ALG_AAI_CTR_MOD128);
188         const bool is_rfc3686 = alg->caam.rfc3686;
189
190         if (!ctx->cdata.keylen || !ctx->authsize)
191                 return 0;
192
193         /*
194          * AES-CTR needs to load IV in CONTEXT1 reg
195          * at an offset of 128bits (16bytes)
196          * CONTEXT1[255:128] = IV
197          */
198         if (ctr_mode)
199                 ctx1_iv_off = 16;
200
201         /*
202          * RFC3686 specific:
203          *      CONTEXT1[255:128] = {NONCE, IV, COUNTER}
204          */
205         if (is_rfc3686) {
206                 ctx1_iv_off = 16 + CTR_RFC3686_NONCE_SIZE;
207                 nonce = (u32 *)((void *)ctx->key + ctx->adata.keylen_pad +
208                                 ctx->cdata.keylen - CTR_RFC3686_NONCE_SIZE);
209         }
210
211         /*
212          * In case |user key| > |derived key|, using DKP<imm,imm> would result
213          * in invalid opcodes (last bytes of user key) in the resulting
214          * descriptor. Use DKP<ptr,imm> instead => both virtual and dma key
215          * addresses are needed.
216          */
217         ctx->adata.key_virt = ctx->key;
218         ctx->adata.key_dma = ctx->key_dma;
219
220         ctx->cdata.key_virt = ctx->key + ctx->adata.keylen_pad;
221         ctx->cdata.key_dma = ctx->key_dma + ctx->adata.keylen_pad;
222
223         data_len[0] = ctx->adata.keylen_pad;
224         data_len[1] = ctx->cdata.keylen;
225
226         /* aead_encrypt shared descriptor */
227         if (desc_inline_query((alg->caam.geniv ? DESC_QI_AEAD_GIVENC_LEN :
228                                                  DESC_QI_AEAD_ENC_LEN) +
229                               (is_rfc3686 ? DESC_AEAD_CTR_RFC3686_LEN : 0),
230                               DESC_JOB_IO_LEN, data_len, &inl_mask,
231                               ARRAY_SIZE(data_len)) < 0)
232                 return -EINVAL;
233
234         ctx->adata.key_inline = !!(inl_mask & 1);
235         ctx->cdata.key_inline = !!(inl_mask & 2);
236
237         flc = &ctx->flc[ENCRYPT];
238         desc = flc->sh_desc;
239
240         if (alg->caam.geniv)
241                 cnstr_shdsc_aead_givencap(desc, &ctx->cdata, &ctx->adata,
242                                           ivsize, ctx->authsize, is_rfc3686,
243                                           nonce, ctx1_iv_off, true,
244                                           priv->sec_attr.era);
245         else
246                 cnstr_shdsc_aead_encap(desc, &ctx->cdata, &ctx->adata,
247                                        ivsize, ctx->authsize, is_rfc3686, nonce,
248                                        ctx1_iv_off, true, priv->sec_attr.era);
249
250         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
251         dma_sync_single_for_device(dev, ctx->flc_dma[ENCRYPT],
252                                    sizeof(flc->flc) + desc_bytes(desc),
253                                    ctx->dir);
254
255         /* aead_decrypt shared descriptor */
256         if (desc_inline_query(DESC_QI_AEAD_DEC_LEN +
257                               (is_rfc3686 ? DESC_AEAD_CTR_RFC3686_LEN : 0),
258                               DESC_JOB_IO_LEN, data_len, &inl_mask,
259                               ARRAY_SIZE(data_len)) < 0)
260                 return -EINVAL;
261
262         ctx->adata.key_inline = !!(inl_mask & 1);
263         ctx->cdata.key_inline = !!(inl_mask & 2);
264
265         flc = &ctx->flc[DECRYPT];
266         desc = flc->sh_desc;
267         cnstr_shdsc_aead_decap(desc, &ctx->cdata, &ctx->adata,
268                                ivsize, ctx->authsize, alg->caam.geniv,
269                                is_rfc3686, nonce, ctx1_iv_off, true,
270                                priv->sec_attr.era);
271         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
272         dma_sync_single_for_device(dev, ctx->flc_dma[DECRYPT],
273                                    sizeof(flc->flc) + desc_bytes(desc),
274                                    ctx->dir);
275
276         return 0;
277 }
278
279 static int aead_setauthsize(struct crypto_aead *authenc, unsigned int authsize)
280 {
281         struct caam_ctx *ctx = crypto_aead_ctx_dma(authenc);
282
283         ctx->authsize = authsize;
284         aead_set_sh_desc(authenc);
285
286         return 0;
287 }
288
289 static int aead_setkey(struct crypto_aead *aead, const u8 *key,
290                        unsigned int keylen)
291 {
292         struct caam_ctx *ctx = crypto_aead_ctx_dma(aead);
293         struct device *dev = ctx->dev;
294         struct crypto_authenc_keys keys;
295
296         if (crypto_authenc_extractkeys(&keys, key, keylen) != 0)
297                 goto badkey;
298
299         dev_dbg(dev, "keylen %d enckeylen %d authkeylen %d\n",
300                 keys.authkeylen + keys.enckeylen, keys.enckeylen,
301                 keys.authkeylen);
302         print_hex_dump_debug("key in @" __stringify(__LINE__)": ",
303                              DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
304
305         ctx->adata.keylen = keys.authkeylen;
306         ctx->adata.keylen_pad = split_key_len(ctx->adata.algtype &
307                                               OP_ALG_ALGSEL_MASK);
308
309         if (ctx->adata.keylen_pad + keys.enckeylen > CAAM_MAX_KEY_SIZE)
310                 goto badkey;
311
312         memcpy(ctx->key, keys.authkey, keys.authkeylen);
313         memcpy(ctx->key + ctx->adata.keylen_pad, keys.enckey, keys.enckeylen);
314         dma_sync_single_for_device(dev, ctx->key_dma, ctx->adata.keylen_pad +
315                                    keys.enckeylen, ctx->dir);
316         print_hex_dump_debug("ctx.key@" __stringify(__LINE__)": ",
317                              DUMP_PREFIX_ADDRESS, 16, 4, ctx->key,
318                              ctx->adata.keylen_pad + keys.enckeylen, 1);
319
320         ctx->cdata.keylen = keys.enckeylen;
321
322         memzero_explicit(&keys, sizeof(keys));
323         return aead_set_sh_desc(aead);
324 badkey:
325         memzero_explicit(&keys, sizeof(keys));
326         return -EINVAL;
327 }
328
329 static int des3_aead_setkey(struct crypto_aead *aead, const u8 *key,
330                             unsigned int keylen)
331 {
332         struct crypto_authenc_keys keys;
333         int err;
334
335         err = crypto_authenc_extractkeys(&keys, key, keylen);
336         if (unlikely(err))
337                 goto out;
338
339         err = -EINVAL;
340         if (keys.enckeylen != DES3_EDE_KEY_SIZE)
341                 goto out;
342
343         err = crypto_des3_ede_verify_key(crypto_aead_tfm(aead), keys.enckey) ?:
344               aead_setkey(aead, key, keylen);
345
346 out:
347         memzero_explicit(&keys, sizeof(keys));
348         return err;
349 }
350
351 static struct aead_edesc *aead_edesc_alloc(struct aead_request *req,
352                                            bool encrypt)
353 {
354         struct crypto_aead *aead = crypto_aead_reqtfm(req);
355         struct caam_request *req_ctx = aead_request_ctx_dma(req);
356         struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
357         struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
358         struct caam_ctx *ctx = crypto_aead_ctx_dma(aead);
359         struct caam_aead_alg *alg = container_of(crypto_aead_alg(aead),
360                                                  typeof(*alg), aead);
361         struct device *dev = ctx->dev;
362         gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
363                       GFP_KERNEL : GFP_ATOMIC;
364         int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0;
365         int src_len, dst_len = 0;
366         struct aead_edesc *edesc;
367         dma_addr_t qm_sg_dma, iv_dma = 0;
368         int ivsize = 0;
369         unsigned int authsize = ctx->authsize;
370         int qm_sg_index = 0, qm_sg_nents = 0, qm_sg_bytes;
371         int in_len, out_len;
372         struct dpaa2_sg_entry *sg_table;
373
374         /* allocate space for base edesc, link tables and IV */
375         edesc = qi_cache_zalloc(flags);
376         if (unlikely(!edesc)) {
377                 dev_err(dev, "could not allocate extended descriptor\n");
378                 return ERR_PTR(-ENOMEM);
379         }
380
381         if (unlikely(req->dst != req->src)) {
382                 src_len = req->assoclen + req->cryptlen;
383                 dst_len = src_len + (encrypt ? authsize : (-authsize));
384
385                 src_nents = sg_nents_for_len(req->src, src_len);
386                 if (unlikely(src_nents < 0)) {
387                         dev_err(dev, "Insufficient bytes (%d) in src S/G\n",
388                                 src_len);
389                         qi_cache_free(edesc);
390                         return ERR_PTR(src_nents);
391                 }
392
393                 dst_nents = sg_nents_for_len(req->dst, dst_len);
394                 if (unlikely(dst_nents < 0)) {
395                         dev_err(dev, "Insufficient bytes (%d) in dst S/G\n",
396                                 dst_len);
397                         qi_cache_free(edesc);
398                         return ERR_PTR(dst_nents);
399                 }
400
401                 if (src_nents) {
402                         mapped_src_nents = dma_map_sg(dev, req->src, src_nents,
403                                                       DMA_TO_DEVICE);
404                         if (unlikely(!mapped_src_nents)) {
405                                 dev_err(dev, "unable to map source\n");
406                                 qi_cache_free(edesc);
407                                 return ERR_PTR(-ENOMEM);
408                         }
409                 } else {
410                         mapped_src_nents = 0;
411                 }
412
413                 if (dst_nents) {
414                         mapped_dst_nents = dma_map_sg(dev, req->dst, dst_nents,
415                                                       DMA_FROM_DEVICE);
416                         if (unlikely(!mapped_dst_nents)) {
417                                 dev_err(dev, "unable to map destination\n");
418                                 dma_unmap_sg(dev, req->src, src_nents,
419                                              DMA_TO_DEVICE);
420                                 qi_cache_free(edesc);
421                                 return ERR_PTR(-ENOMEM);
422                         }
423                 } else {
424                         mapped_dst_nents = 0;
425                 }
426         } else {
427                 src_len = req->assoclen + req->cryptlen +
428                           (encrypt ? authsize : 0);
429
430                 src_nents = sg_nents_for_len(req->src, src_len);
431                 if (unlikely(src_nents < 0)) {
432                         dev_err(dev, "Insufficient bytes (%d) in src S/G\n",
433                                 src_len);
434                         qi_cache_free(edesc);
435                         return ERR_PTR(src_nents);
436                 }
437
438                 mapped_src_nents = dma_map_sg(dev, req->src, src_nents,
439                                               DMA_BIDIRECTIONAL);
440                 if (unlikely(!mapped_src_nents)) {
441                         dev_err(dev, "unable to map source\n");
442                         qi_cache_free(edesc);
443                         return ERR_PTR(-ENOMEM);
444                 }
445         }
446
447         if ((alg->caam.rfc3686 && encrypt) || !alg->caam.geniv)
448                 ivsize = crypto_aead_ivsize(aead);
449
450         /*
451          * Create S/G table: req->assoclen, [IV,] req->src [, req->dst].
452          * Input is not contiguous.
453          * HW reads 4 S/G entries at a time; make sure the reads don't go beyond
454          * the end of the table by allocating more S/G entries. Logic:
455          * if (src != dst && output S/G)
456          *      pad output S/G, if needed
457          * else if (src == dst && S/G)
458          *      overlapping S/Gs; pad one of them
459          * else if (input S/G) ...
460          *      pad input S/G, if needed
461          */
462         qm_sg_nents = 1 + !!ivsize + mapped_src_nents;
463         if (mapped_dst_nents > 1)
464                 qm_sg_nents += pad_sg_nents(mapped_dst_nents);
465         else if ((req->src == req->dst) && (mapped_src_nents > 1))
466                 qm_sg_nents = max(pad_sg_nents(qm_sg_nents),
467                                   1 + !!ivsize +
468                                   pad_sg_nents(mapped_src_nents));
469         else
470                 qm_sg_nents = pad_sg_nents(qm_sg_nents);
471
472         sg_table = &edesc->sgt[0];
473         qm_sg_bytes = qm_sg_nents * sizeof(*sg_table);
474         if (unlikely(offsetof(struct aead_edesc, sgt) + qm_sg_bytes + ivsize >
475                      CAAM_QI_MEMCACHE_SIZE)) {
476                 dev_err(dev, "No space for %d S/G entries and/or %dB IV\n",
477                         qm_sg_nents, ivsize);
478                 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0,
479                            0, DMA_NONE, 0, 0);
480                 qi_cache_free(edesc);
481                 return ERR_PTR(-ENOMEM);
482         }
483
484         if (ivsize) {
485                 u8 *iv = (u8 *)(sg_table + qm_sg_nents);
486
487                 /* Make sure IV is located in a DMAable area */
488                 memcpy(iv, req->iv, ivsize);
489
490                 iv_dma = dma_map_single(dev, iv, ivsize, DMA_TO_DEVICE);
491                 if (dma_mapping_error(dev, iv_dma)) {
492                         dev_err(dev, "unable to map IV\n");
493                         caam_unmap(dev, req->src, req->dst, src_nents,
494                                    dst_nents, 0, 0, DMA_NONE, 0, 0);
495                         qi_cache_free(edesc);
496                         return ERR_PTR(-ENOMEM);
497                 }
498         }
499
500         edesc->src_nents = src_nents;
501         edesc->dst_nents = dst_nents;
502         edesc->iv_dma = iv_dma;
503
504         if ((alg->caam.class1_alg_type & OP_ALG_ALGSEL_MASK) ==
505             OP_ALG_ALGSEL_CHACHA20 && ivsize != CHACHAPOLY_IV_SIZE)
506                 /*
507                  * The associated data comes already with the IV but we need
508                  * to skip it when we authenticate or encrypt...
509                  */
510                 edesc->assoclen = cpu_to_caam32(req->assoclen - ivsize);
511         else
512                 edesc->assoclen = cpu_to_caam32(req->assoclen);
513         edesc->assoclen_dma = dma_map_single(dev, &edesc->assoclen, 4,
514                                              DMA_TO_DEVICE);
515         if (dma_mapping_error(dev, edesc->assoclen_dma)) {
516                 dev_err(dev, "unable to map assoclen\n");
517                 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents,
518                            iv_dma, ivsize, DMA_TO_DEVICE, 0, 0);
519                 qi_cache_free(edesc);
520                 return ERR_PTR(-ENOMEM);
521         }
522
523         dma_to_qm_sg_one(sg_table, edesc->assoclen_dma, 4, 0);
524         qm_sg_index++;
525         if (ivsize) {
526                 dma_to_qm_sg_one(sg_table + qm_sg_index, iv_dma, ivsize, 0);
527                 qm_sg_index++;
528         }
529         sg_to_qm_sg_last(req->src, src_len, sg_table + qm_sg_index, 0);
530         qm_sg_index += mapped_src_nents;
531
532         if (mapped_dst_nents > 1)
533                 sg_to_qm_sg_last(req->dst, dst_len, sg_table + qm_sg_index, 0);
534
535         qm_sg_dma = dma_map_single(dev, sg_table, qm_sg_bytes, DMA_TO_DEVICE);
536         if (dma_mapping_error(dev, qm_sg_dma)) {
537                 dev_err(dev, "unable to map S/G table\n");
538                 dma_unmap_single(dev, edesc->assoclen_dma, 4, DMA_TO_DEVICE);
539                 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents,
540                            iv_dma, ivsize, DMA_TO_DEVICE, 0, 0);
541                 qi_cache_free(edesc);
542                 return ERR_PTR(-ENOMEM);
543         }
544
545         edesc->qm_sg_dma = qm_sg_dma;
546         edesc->qm_sg_bytes = qm_sg_bytes;
547
548         out_len = req->assoclen + req->cryptlen +
549                   (encrypt ? ctx->authsize : (-ctx->authsize));
550         in_len = 4 + ivsize + req->assoclen + req->cryptlen;
551
552         memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
553         dpaa2_fl_set_final(in_fle, true);
554         dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
555         dpaa2_fl_set_addr(in_fle, qm_sg_dma);
556         dpaa2_fl_set_len(in_fle, in_len);
557
558         if (req->dst == req->src) {
559                 if (mapped_src_nents == 1) {
560                         dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
561                         dpaa2_fl_set_addr(out_fle, sg_dma_address(req->src));
562                 } else {
563                         dpaa2_fl_set_format(out_fle, dpaa2_fl_sg);
564                         dpaa2_fl_set_addr(out_fle, qm_sg_dma +
565                                           (1 + !!ivsize) * sizeof(*sg_table));
566                 }
567         } else if (!mapped_dst_nents) {
568                 /*
569                  * crypto engine requires the output entry to be present when
570                  * "frame list" FD is used.
571                  * Since engine does not support FMT=2'b11 (unused entry type),
572                  * leaving out_fle zeroized is the best option.
573                  */
574                 goto skip_out_fle;
575         } else if (mapped_dst_nents == 1) {
576                 dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
577                 dpaa2_fl_set_addr(out_fle, sg_dma_address(req->dst));
578         } else {
579                 dpaa2_fl_set_format(out_fle, dpaa2_fl_sg);
580                 dpaa2_fl_set_addr(out_fle, qm_sg_dma + qm_sg_index *
581                                   sizeof(*sg_table));
582         }
583
584         dpaa2_fl_set_len(out_fle, out_len);
585
586 skip_out_fle:
587         return edesc;
588 }
589
590 static int chachapoly_set_sh_desc(struct crypto_aead *aead)
591 {
592         struct caam_ctx *ctx = crypto_aead_ctx_dma(aead);
593         unsigned int ivsize = crypto_aead_ivsize(aead);
594         struct device *dev = ctx->dev;
595         struct caam_flc *flc;
596         u32 *desc;
597
598         if (!ctx->cdata.keylen || !ctx->authsize)
599                 return 0;
600
601         flc = &ctx->flc[ENCRYPT];
602         desc = flc->sh_desc;
603         cnstr_shdsc_chachapoly(desc, &ctx->cdata, &ctx->adata, ivsize,
604                                ctx->authsize, true, true);
605         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
606         dma_sync_single_for_device(dev, ctx->flc_dma[ENCRYPT],
607                                    sizeof(flc->flc) + desc_bytes(desc),
608                                    ctx->dir);
609
610         flc = &ctx->flc[DECRYPT];
611         desc = flc->sh_desc;
612         cnstr_shdsc_chachapoly(desc, &ctx->cdata, &ctx->adata, ivsize,
613                                ctx->authsize, false, true);
614         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
615         dma_sync_single_for_device(dev, ctx->flc_dma[DECRYPT],
616                                    sizeof(flc->flc) + desc_bytes(desc),
617                                    ctx->dir);
618
619         return 0;
620 }
621
622 static int chachapoly_setauthsize(struct crypto_aead *aead,
623                                   unsigned int authsize)
624 {
625         struct caam_ctx *ctx = crypto_aead_ctx_dma(aead);
626
627         if (authsize != POLY1305_DIGEST_SIZE)
628                 return -EINVAL;
629
630         ctx->authsize = authsize;
631         return chachapoly_set_sh_desc(aead);
632 }
633
634 static int chachapoly_setkey(struct crypto_aead *aead, const u8 *key,
635                              unsigned int keylen)
636 {
637         struct caam_ctx *ctx = crypto_aead_ctx_dma(aead);
638         unsigned int ivsize = crypto_aead_ivsize(aead);
639         unsigned int saltlen = CHACHAPOLY_IV_SIZE - ivsize;
640
641         if (keylen != CHACHA_KEY_SIZE + saltlen)
642                 return -EINVAL;
643
644         memcpy(ctx->key, key, keylen);
645         ctx->cdata.key_virt = ctx->key;
646         ctx->cdata.keylen = keylen - saltlen;
647
648         return chachapoly_set_sh_desc(aead);
649 }
650
651 static int gcm_set_sh_desc(struct crypto_aead *aead)
652 {
653         struct caam_ctx *ctx = crypto_aead_ctx_dma(aead);
654         struct device *dev = ctx->dev;
655         unsigned int ivsize = crypto_aead_ivsize(aead);
656         struct caam_flc *flc;
657         u32 *desc;
658         int rem_bytes = CAAM_DESC_BYTES_MAX - DESC_JOB_IO_LEN -
659                         ctx->cdata.keylen;
660
661         if (!ctx->cdata.keylen || !ctx->authsize)
662                 return 0;
663
664         /*
665          * AES GCM encrypt shared descriptor
666          * Job Descriptor and Shared Descriptor
667          * must fit into the 64-word Descriptor h/w Buffer
668          */
669         if (rem_bytes >= DESC_QI_GCM_ENC_LEN) {
670                 ctx->cdata.key_inline = true;
671                 ctx->cdata.key_virt = ctx->key;
672         } else {
673                 ctx->cdata.key_inline = false;
674                 ctx->cdata.key_dma = ctx->key_dma;
675         }
676
677         flc = &ctx->flc[ENCRYPT];
678         desc = flc->sh_desc;
679         cnstr_shdsc_gcm_encap(desc, &ctx->cdata, ivsize, ctx->authsize, true);
680         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
681         dma_sync_single_for_device(dev, ctx->flc_dma[ENCRYPT],
682                                    sizeof(flc->flc) + desc_bytes(desc),
683                                    ctx->dir);
684
685         /*
686          * Job Descriptor and Shared Descriptors
687          * must all fit into the 64-word Descriptor h/w Buffer
688          */
689         if (rem_bytes >= DESC_QI_GCM_DEC_LEN) {
690                 ctx->cdata.key_inline = true;
691                 ctx->cdata.key_virt = ctx->key;
692         } else {
693                 ctx->cdata.key_inline = false;
694                 ctx->cdata.key_dma = ctx->key_dma;
695         }
696
697         flc = &ctx->flc[DECRYPT];
698         desc = flc->sh_desc;
699         cnstr_shdsc_gcm_decap(desc, &ctx->cdata, ivsize, ctx->authsize, true);
700         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
701         dma_sync_single_for_device(dev, ctx->flc_dma[DECRYPT],
702                                    sizeof(flc->flc) + desc_bytes(desc),
703                                    ctx->dir);
704
705         return 0;
706 }
707
708 static int gcm_setauthsize(struct crypto_aead *authenc, unsigned int authsize)
709 {
710         struct caam_ctx *ctx = crypto_aead_ctx_dma(authenc);
711         int err;
712
713         err = crypto_gcm_check_authsize(authsize);
714         if (err)
715                 return err;
716
717         ctx->authsize = authsize;
718         gcm_set_sh_desc(authenc);
719
720         return 0;
721 }
722
723 static int gcm_setkey(struct crypto_aead *aead,
724                       const u8 *key, unsigned int keylen)
725 {
726         struct caam_ctx *ctx = crypto_aead_ctx_dma(aead);
727         struct device *dev = ctx->dev;
728         int ret;
729
730         ret = aes_check_keylen(keylen);
731         if (ret)
732                 return ret;
733         print_hex_dump_debug("key in @" __stringify(__LINE__)": ",
734                              DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
735
736         memcpy(ctx->key, key, keylen);
737         dma_sync_single_for_device(dev, ctx->key_dma, keylen, ctx->dir);
738         ctx->cdata.keylen = keylen;
739
740         return gcm_set_sh_desc(aead);
741 }
742
743 static int rfc4106_set_sh_desc(struct crypto_aead *aead)
744 {
745         struct caam_ctx *ctx = crypto_aead_ctx_dma(aead);
746         struct device *dev = ctx->dev;
747         unsigned int ivsize = crypto_aead_ivsize(aead);
748         struct caam_flc *flc;
749         u32 *desc;
750         int rem_bytes = CAAM_DESC_BYTES_MAX - DESC_JOB_IO_LEN -
751                         ctx->cdata.keylen;
752
753         if (!ctx->cdata.keylen || !ctx->authsize)
754                 return 0;
755
756         ctx->cdata.key_virt = ctx->key;
757
758         /*
759          * RFC4106 encrypt shared descriptor
760          * Job Descriptor and Shared Descriptor
761          * must fit into the 64-word Descriptor h/w Buffer
762          */
763         if (rem_bytes >= DESC_QI_RFC4106_ENC_LEN) {
764                 ctx->cdata.key_inline = true;
765         } else {
766                 ctx->cdata.key_inline = false;
767                 ctx->cdata.key_dma = ctx->key_dma;
768         }
769
770         flc = &ctx->flc[ENCRYPT];
771         desc = flc->sh_desc;
772         cnstr_shdsc_rfc4106_encap(desc, &ctx->cdata, ivsize, ctx->authsize,
773                                   true);
774         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
775         dma_sync_single_for_device(dev, ctx->flc_dma[ENCRYPT],
776                                    sizeof(flc->flc) + desc_bytes(desc),
777                                    ctx->dir);
778
779         /*
780          * Job Descriptor and Shared Descriptors
781          * must all fit into the 64-word Descriptor h/w Buffer
782          */
783         if (rem_bytes >= DESC_QI_RFC4106_DEC_LEN) {
784                 ctx->cdata.key_inline = true;
785         } else {
786                 ctx->cdata.key_inline = false;
787                 ctx->cdata.key_dma = ctx->key_dma;
788         }
789
790         flc = &ctx->flc[DECRYPT];
791         desc = flc->sh_desc;
792         cnstr_shdsc_rfc4106_decap(desc, &ctx->cdata, ivsize, ctx->authsize,
793                                   true);
794         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
795         dma_sync_single_for_device(dev, ctx->flc_dma[DECRYPT],
796                                    sizeof(flc->flc) + desc_bytes(desc),
797                                    ctx->dir);
798
799         return 0;
800 }
801
802 static int rfc4106_setauthsize(struct crypto_aead *authenc,
803                                unsigned int authsize)
804 {
805         struct caam_ctx *ctx = crypto_aead_ctx_dma(authenc);
806         int err;
807
808         err = crypto_rfc4106_check_authsize(authsize);
809         if (err)
810                 return err;
811
812         ctx->authsize = authsize;
813         rfc4106_set_sh_desc(authenc);
814
815         return 0;
816 }
817
818 static int rfc4106_setkey(struct crypto_aead *aead,
819                           const u8 *key, unsigned int keylen)
820 {
821         struct caam_ctx *ctx = crypto_aead_ctx_dma(aead);
822         struct device *dev = ctx->dev;
823         int ret;
824
825         ret = aes_check_keylen(keylen - 4);
826         if (ret)
827                 return ret;
828
829         print_hex_dump_debug("key in @" __stringify(__LINE__)": ",
830                              DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
831
832         memcpy(ctx->key, key, keylen);
833         /*
834          * The last four bytes of the key material are used as the salt value
835          * in the nonce. Update the AES key length.
836          */
837         ctx->cdata.keylen = keylen - 4;
838         dma_sync_single_for_device(dev, ctx->key_dma, ctx->cdata.keylen,
839                                    ctx->dir);
840
841         return rfc4106_set_sh_desc(aead);
842 }
843
844 static int rfc4543_set_sh_desc(struct crypto_aead *aead)
845 {
846         struct caam_ctx *ctx = crypto_aead_ctx_dma(aead);
847         struct device *dev = ctx->dev;
848         unsigned int ivsize = crypto_aead_ivsize(aead);
849         struct caam_flc *flc;
850         u32 *desc;
851         int rem_bytes = CAAM_DESC_BYTES_MAX - DESC_JOB_IO_LEN -
852                         ctx->cdata.keylen;
853
854         if (!ctx->cdata.keylen || !ctx->authsize)
855                 return 0;
856
857         ctx->cdata.key_virt = ctx->key;
858
859         /*
860          * RFC4543 encrypt shared descriptor
861          * Job Descriptor and Shared Descriptor
862          * must fit into the 64-word Descriptor h/w Buffer
863          */
864         if (rem_bytes >= DESC_QI_RFC4543_ENC_LEN) {
865                 ctx->cdata.key_inline = true;
866         } else {
867                 ctx->cdata.key_inline = false;
868                 ctx->cdata.key_dma = ctx->key_dma;
869         }
870
871         flc = &ctx->flc[ENCRYPT];
872         desc = flc->sh_desc;
873         cnstr_shdsc_rfc4543_encap(desc, &ctx->cdata, ivsize, ctx->authsize,
874                                   true);
875         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
876         dma_sync_single_for_device(dev, ctx->flc_dma[ENCRYPT],
877                                    sizeof(flc->flc) + desc_bytes(desc),
878                                    ctx->dir);
879
880         /*
881          * Job Descriptor and Shared Descriptors
882          * must all fit into the 64-word Descriptor h/w Buffer
883          */
884         if (rem_bytes >= DESC_QI_RFC4543_DEC_LEN) {
885                 ctx->cdata.key_inline = true;
886         } else {
887                 ctx->cdata.key_inline = false;
888                 ctx->cdata.key_dma = ctx->key_dma;
889         }
890
891         flc = &ctx->flc[DECRYPT];
892         desc = flc->sh_desc;
893         cnstr_shdsc_rfc4543_decap(desc, &ctx->cdata, ivsize, ctx->authsize,
894                                   true);
895         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
896         dma_sync_single_for_device(dev, ctx->flc_dma[DECRYPT],
897                                    sizeof(flc->flc) + desc_bytes(desc),
898                                    ctx->dir);
899
900         return 0;
901 }
902
903 static int rfc4543_setauthsize(struct crypto_aead *authenc,
904                                unsigned int authsize)
905 {
906         struct caam_ctx *ctx = crypto_aead_ctx_dma(authenc);
907
908         if (authsize != 16)
909                 return -EINVAL;
910
911         ctx->authsize = authsize;
912         rfc4543_set_sh_desc(authenc);
913
914         return 0;
915 }
916
917 static int rfc4543_setkey(struct crypto_aead *aead,
918                           const u8 *key, unsigned int keylen)
919 {
920         struct caam_ctx *ctx = crypto_aead_ctx_dma(aead);
921         struct device *dev = ctx->dev;
922         int ret;
923
924         ret = aes_check_keylen(keylen - 4);
925         if (ret)
926                 return ret;
927
928         print_hex_dump_debug("key in @" __stringify(__LINE__)": ",
929                              DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
930
931         memcpy(ctx->key, key, keylen);
932         /*
933          * The last four bytes of the key material are used as the salt value
934          * in the nonce. Update the AES key length.
935          */
936         ctx->cdata.keylen = keylen - 4;
937         dma_sync_single_for_device(dev, ctx->key_dma, ctx->cdata.keylen,
938                                    ctx->dir);
939
940         return rfc4543_set_sh_desc(aead);
941 }
942
943 static int skcipher_setkey(struct crypto_skcipher *skcipher, const u8 *key,
944                            unsigned int keylen, const u32 ctx1_iv_off)
945 {
946         struct caam_ctx *ctx = crypto_skcipher_ctx_dma(skcipher);
947         struct caam_skcipher_alg *alg =
948                 container_of(crypto_skcipher_alg(skcipher),
949                              struct caam_skcipher_alg, skcipher);
950         struct device *dev = ctx->dev;
951         struct caam_flc *flc;
952         unsigned int ivsize = crypto_skcipher_ivsize(skcipher);
953         u32 *desc;
954         const bool is_rfc3686 = alg->caam.rfc3686;
955
956         print_hex_dump_debug("key in @" __stringify(__LINE__)": ",
957                              DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
958
959         ctx->cdata.keylen = keylen;
960         ctx->cdata.key_virt = key;
961         ctx->cdata.key_inline = true;
962
963         /* skcipher_encrypt shared descriptor */
964         flc = &ctx->flc[ENCRYPT];
965         desc = flc->sh_desc;
966         cnstr_shdsc_skcipher_encap(desc, &ctx->cdata, ivsize, is_rfc3686,
967                                    ctx1_iv_off);
968         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
969         dma_sync_single_for_device(dev, ctx->flc_dma[ENCRYPT],
970                                    sizeof(flc->flc) + desc_bytes(desc),
971                                    ctx->dir);
972
973         /* skcipher_decrypt shared descriptor */
974         flc = &ctx->flc[DECRYPT];
975         desc = flc->sh_desc;
976         cnstr_shdsc_skcipher_decap(desc, &ctx->cdata, ivsize, is_rfc3686,
977                                    ctx1_iv_off);
978         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
979         dma_sync_single_for_device(dev, ctx->flc_dma[DECRYPT],
980                                    sizeof(flc->flc) + desc_bytes(desc),
981                                    ctx->dir);
982
983         return 0;
984 }
985
986 static int aes_skcipher_setkey(struct crypto_skcipher *skcipher,
987                                const u8 *key, unsigned int keylen)
988 {
989         int err;
990
991         err = aes_check_keylen(keylen);
992         if (err)
993                 return err;
994
995         return skcipher_setkey(skcipher, key, keylen, 0);
996 }
997
998 static int rfc3686_skcipher_setkey(struct crypto_skcipher *skcipher,
999                                    const u8 *key, unsigned int keylen)
1000 {
1001         u32 ctx1_iv_off;
1002         int err;
1003
1004         /*
1005          * RFC3686 specific:
1006          *      | CONTEXT1[255:128] = {NONCE, IV, COUNTER}
1007          *      | *key = {KEY, NONCE}
1008          */
1009         ctx1_iv_off = 16 + CTR_RFC3686_NONCE_SIZE;
1010         keylen -= CTR_RFC3686_NONCE_SIZE;
1011
1012         err = aes_check_keylen(keylen);
1013         if (err)
1014                 return err;
1015
1016         return skcipher_setkey(skcipher, key, keylen, ctx1_iv_off);
1017 }
1018
1019 static int ctr_skcipher_setkey(struct crypto_skcipher *skcipher,
1020                                const u8 *key, unsigned int keylen)
1021 {
1022         u32 ctx1_iv_off;
1023         int err;
1024
1025         /*
1026          * AES-CTR needs to load IV in CONTEXT1 reg
1027          * at an offset of 128bits (16bytes)
1028          * CONTEXT1[255:128] = IV
1029          */
1030         ctx1_iv_off = 16;
1031
1032         err = aes_check_keylen(keylen);
1033         if (err)
1034                 return err;
1035
1036         return skcipher_setkey(skcipher, key, keylen, ctx1_iv_off);
1037 }
1038
1039 static int chacha20_skcipher_setkey(struct crypto_skcipher *skcipher,
1040                                     const u8 *key, unsigned int keylen)
1041 {
1042         if (keylen != CHACHA_KEY_SIZE)
1043                 return -EINVAL;
1044
1045         return skcipher_setkey(skcipher, key, keylen, 0);
1046 }
1047
1048 static int des_skcipher_setkey(struct crypto_skcipher *skcipher,
1049                                const u8 *key, unsigned int keylen)
1050 {
1051         return verify_skcipher_des_key(skcipher, key) ?:
1052                skcipher_setkey(skcipher, key, keylen, 0);
1053 }
1054
1055 static int des3_skcipher_setkey(struct crypto_skcipher *skcipher,
1056                                 const u8 *key, unsigned int keylen)
1057 {
1058         return verify_skcipher_des3_key(skcipher, key) ?:
1059                skcipher_setkey(skcipher, key, keylen, 0);
1060 }
1061
1062 static int xts_skcipher_setkey(struct crypto_skcipher *skcipher, const u8 *key,
1063                                unsigned int keylen)
1064 {
1065         struct caam_ctx *ctx = crypto_skcipher_ctx_dma(skcipher);
1066         struct device *dev = ctx->dev;
1067         struct dpaa2_caam_priv *priv = dev_get_drvdata(dev);
1068         struct caam_flc *flc;
1069         u32 *desc;
1070         int err;
1071
1072         err = xts_verify_key(skcipher, key, keylen);
1073         if (err) {
1074                 dev_dbg(dev, "key size mismatch\n");
1075                 return err;
1076         }
1077
1078         if (keylen != 2 * AES_KEYSIZE_128 && keylen != 2 * AES_KEYSIZE_256)
1079                 ctx->xts_key_fallback = true;
1080
1081         if (priv->sec_attr.era <= 8 || ctx->xts_key_fallback) {
1082                 err = crypto_skcipher_setkey(ctx->fallback, key, keylen);
1083                 if (err)
1084                         return err;
1085         }
1086
1087         ctx->cdata.keylen = keylen;
1088         ctx->cdata.key_virt = key;
1089         ctx->cdata.key_inline = true;
1090
1091         /* xts_skcipher_encrypt shared descriptor */
1092         flc = &ctx->flc[ENCRYPT];
1093         desc = flc->sh_desc;
1094         cnstr_shdsc_xts_skcipher_encap(desc, &ctx->cdata);
1095         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
1096         dma_sync_single_for_device(dev, ctx->flc_dma[ENCRYPT],
1097                                    sizeof(flc->flc) + desc_bytes(desc),
1098                                    ctx->dir);
1099
1100         /* xts_skcipher_decrypt shared descriptor */
1101         flc = &ctx->flc[DECRYPT];
1102         desc = flc->sh_desc;
1103         cnstr_shdsc_xts_skcipher_decap(desc, &ctx->cdata);
1104         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
1105         dma_sync_single_for_device(dev, ctx->flc_dma[DECRYPT],
1106                                    sizeof(flc->flc) + desc_bytes(desc),
1107                                    ctx->dir);
1108
1109         return 0;
1110 }
1111
1112 static struct skcipher_edesc *skcipher_edesc_alloc(struct skcipher_request *req)
1113 {
1114         struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req);
1115         struct caam_request *req_ctx = skcipher_request_ctx_dma(req);
1116         struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
1117         struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
1118         struct caam_ctx *ctx = crypto_skcipher_ctx_dma(skcipher);
1119         struct device *dev = ctx->dev;
1120         gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
1121                        GFP_KERNEL : GFP_ATOMIC;
1122         int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0;
1123         struct skcipher_edesc *edesc;
1124         dma_addr_t iv_dma;
1125         u8 *iv;
1126         int ivsize = crypto_skcipher_ivsize(skcipher);
1127         int dst_sg_idx, qm_sg_ents, qm_sg_bytes;
1128         struct dpaa2_sg_entry *sg_table;
1129
1130         src_nents = sg_nents_for_len(req->src, req->cryptlen);
1131         if (unlikely(src_nents < 0)) {
1132                 dev_err(dev, "Insufficient bytes (%d) in src S/G\n",
1133                         req->cryptlen);
1134                 return ERR_PTR(src_nents);
1135         }
1136
1137         if (unlikely(req->dst != req->src)) {
1138                 dst_nents = sg_nents_for_len(req->dst, req->cryptlen);
1139                 if (unlikely(dst_nents < 0)) {
1140                         dev_err(dev, "Insufficient bytes (%d) in dst S/G\n",
1141                                 req->cryptlen);
1142                         return ERR_PTR(dst_nents);
1143                 }
1144
1145                 mapped_src_nents = dma_map_sg(dev, req->src, src_nents,
1146                                               DMA_TO_DEVICE);
1147                 if (unlikely(!mapped_src_nents)) {
1148                         dev_err(dev, "unable to map source\n");
1149                         return ERR_PTR(-ENOMEM);
1150                 }
1151
1152                 mapped_dst_nents = dma_map_sg(dev, req->dst, dst_nents,
1153                                               DMA_FROM_DEVICE);
1154                 if (unlikely(!mapped_dst_nents)) {
1155                         dev_err(dev, "unable to map destination\n");
1156                         dma_unmap_sg(dev, req->src, src_nents, DMA_TO_DEVICE);
1157                         return ERR_PTR(-ENOMEM);
1158                 }
1159         } else {
1160                 mapped_src_nents = dma_map_sg(dev, req->src, src_nents,
1161                                               DMA_BIDIRECTIONAL);
1162                 if (unlikely(!mapped_src_nents)) {
1163                         dev_err(dev, "unable to map source\n");
1164                         return ERR_PTR(-ENOMEM);
1165                 }
1166         }
1167
1168         qm_sg_ents = 1 + mapped_src_nents;
1169         dst_sg_idx = qm_sg_ents;
1170
1171         /*
1172          * Input, output HW S/G tables: [IV, src][dst, IV]
1173          * IV entries point to the same buffer
1174          * If src == dst, S/G entries are reused (S/G tables overlap)
1175          *
1176          * HW reads 4 S/G entries at a time; make sure the reads don't go beyond
1177          * the end of the table by allocating more S/G entries.
1178          */
1179         if (req->src != req->dst)
1180                 qm_sg_ents += pad_sg_nents(mapped_dst_nents + 1);
1181         else
1182                 qm_sg_ents = 1 + pad_sg_nents(qm_sg_ents);
1183
1184         qm_sg_bytes = qm_sg_ents * sizeof(struct dpaa2_sg_entry);
1185         if (unlikely(offsetof(struct skcipher_edesc, sgt) + qm_sg_bytes +
1186                      ivsize > CAAM_QI_MEMCACHE_SIZE)) {
1187                 dev_err(dev, "No space for %d S/G entries and/or %dB IV\n",
1188                         qm_sg_ents, ivsize);
1189                 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0,
1190                            0, DMA_NONE, 0, 0);
1191                 return ERR_PTR(-ENOMEM);
1192         }
1193
1194         /* allocate space for base edesc, link tables and IV */
1195         edesc = qi_cache_zalloc(flags);
1196         if (unlikely(!edesc)) {
1197                 dev_err(dev, "could not allocate extended descriptor\n");
1198                 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0,
1199                            0, DMA_NONE, 0, 0);
1200                 return ERR_PTR(-ENOMEM);
1201         }
1202
1203         /* Make sure IV is located in a DMAable area */
1204         sg_table = &edesc->sgt[0];
1205         iv = (u8 *)(sg_table + qm_sg_ents);
1206         memcpy(iv, req->iv, ivsize);
1207
1208         iv_dma = dma_map_single(dev, iv, ivsize, DMA_BIDIRECTIONAL);
1209         if (dma_mapping_error(dev, iv_dma)) {
1210                 dev_err(dev, "unable to map IV\n");
1211                 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0,
1212                            0, DMA_NONE, 0, 0);
1213                 qi_cache_free(edesc);
1214                 return ERR_PTR(-ENOMEM);
1215         }
1216
1217         edesc->src_nents = src_nents;
1218         edesc->dst_nents = dst_nents;
1219         edesc->iv_dma = iv_dma;
1220         edesc->qm_sg_bytes = qm_sg_bytes;
1221
1222         dma_to_qm_sg_one(sg_table, iv_dma, ivsize, 0);
1223         sg_to_qm_sg(req->src, req->cryptlen, sg_table + 1, 0);
1224
1225         if (req->src != req->dst)
1226                 sg_to_qm_sg(req->dst, req->cryptlen, sg_table + dst_sg_idx, 0);
1227
1228         dma_to_qm_sg_one(sg_table + dst_sg_idx + mapped_dst_nents, iv_dma,
1229                          ivsize, 0);
1230
1231         edesc->qm_sg_dma = dma_map_single(dev, sg_table, edesc->qm_sg_bytes,
1232                                           DMA_TO_DEVICE);
1233         if (dma_mapping_error(dev, edesc->qm_sg_dma)) {
1234                 dev_err(dev, "unable to map S/G table\n");
1235                 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents,
1236                            iv_dma, ivsize, DMA_BIDIRECTIONAL, 0, 0);
1237                 qi_cache_free(edesc);
1238                 return ERR_PTR(-ENOMEM);
1239         }
1240
1241         memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
1242         dpaa2_fl_set_final(in_fle, true);
1243         dpaa2_fl_set_len(in_fle, req->cryptlen + ivsize);
1244         dpaa2_fl_set_len(out_fle, req->cryptlen + ivsize);
1245
1246         dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
1247         dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
1248
1249         dpaa2_fl_set_format(out_fle, dpaa2_fl_sg);
1250
1251         if (req->src == req->dst)
1252                 dpaa2_fl_set_addr(out_fle, edesc->qm_sg_dma +
1253                                   sizeof(*sg_table));
1254         else
1255                 dpaa2_fl_set_addr(out_fle, edesc->qm_sg_dma + dst_sg_idx *
1256                                   sizeof(*sg_table));
1257
1258         return edesc;
1259 }
1260
1261 static void aead_unmap(struct device *dev, struct aead_edesc *edesc,
1262                        struct aead_request *req)
1263 {
1264         struct crypto_aead *aead = crypto_aead_reqtfm(req);
1265         int ivsize = crypto_aead_ivsize(aead);
1266
1267         caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents,
1268                    edesc->iv_dma, ivsize, DMA_TO_DEVICE, edesc->qm_sg_dma,
1269                    edesc->qm_sg_bytes);
1270         dma_unmap_single(dev, edesc->assoclen_dma, 4, DMA_TO_DEVICE);
1271 }
1272
1273 static void skcipher_unmap(struct device *dev, struct skcipher_edesc *edesc,
1274                            struct skcipher_request *req)
1275 {
1276         struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req);
1277         int ivsize = crypto_skcipher_ivsize(skcipher);
1278
1279         caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents,
1280                    edesc->iv_dma, ivsize, DMA_BIDIRECTIONAL, edesc->qm_sg_dma,
1281                    edesc->qm_sg_bytes);
1282 }
1283
1284 static void aead_encrypt_done(void *cbk_ctx, u32 status)
1285 {
1286         struct crypto_async_request *areq = cbk_ctx;
1287         struct aead_request *req = container_of(areq, struct aead_request,
1288                                                 base);
1289         struct caam_request *req_ctx = to_caam_req(areq);
1290         struct aead_edesc *edesc = req_ctx->edesc;
1291         struct crypto_aead *aead = crypto_aead_reqtfm(req);
1292         struct caam_ctx *ctx = crypto_aead_ctx_dma(aead);
1293         int ecode = 0;
1294
1295         dev_dbg(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
1296
1297         if (unlikely(status))
1298                 ecode = caam_qi2_strstatus(ctx->dev, status);
1299
1300         aead_unmap(ctx->dev, edesc, req);
1301         qi_cache_free(edesc);
1302         aead_request_complete(req, ecode);
1303 }
1304
1305 static void aead_decrypt_done(void *cbk_ctx, u32 status)
1306 {
1307         struct crypto_async_request *areq = cbk_ctx;
1308         struct aead_request *req = container_of(areq, struct aead_request,
1309                                                 base);
1310         struct caam_request *req_ctx = to_caam_req(areq);
1311         struct aead_edesc *edesc = req_ctx->edesc;
1312         struct crypto_aead *aead = crypto_aead_reqtfm(req);
1313         struct caam_ctx *ctx = crypto_aead_ctx_dma(aead);
1314         int ecode = 0;
1315
1316         dev_dbg(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
1317
1318         if (unlikely(status))
1319                 ecode = caam_qi2_strstatus(ctx->dev, status);
1320
1321         aead_unmap(ctx->dev, edesc, req);
1322         qi_cache_free(edesc);
1323         aead_request_complete(req, ecode);
1324 }
1325
1326 static int aead_encrypt(struct aead_request *req)
1327 {
1328         struct aead_edesc *edesc;
1329         struct crypto_aead *aead = crypto_aead_reqtfm(req);
1330         struct caam_ctx *ctx = crypto_aead_ctx_dma(aead);
1331         struct caam_request *caam_req = aead_request_ctx_dma(req);
1332         int ret;
1333
1334         /* allocate extended descriptor */
1335         edesc = aead_edesc_alloc(req, true);
1336         if (IS_ERR(edesc))
1337                 return PTR_ERR(edesc);
1338
1339         caam_req->flc = &ctx->flc[ENCRYPT];
1340         caam_req->flc_dma = ctx->flc_dma[ENCRYPT];
1341         caam_req->cbk = aead_encrypt_done;
1342         caam_req->ctx = &req->base;
1343         caam_req->edesc = edesc;
1344         ret = dpaa2_caam_enqueue(ctx->dev, caam_req);
1345         if (ret != -EINPROGRESS &&
1346             !(ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG)) {
1347                 aead_unmap(ctx->dev, edesc, req);
1348                 qi_cache_free(edesc);
1349         }
1350
1351         return ret;
1352 }
1353
1354 static int aead_decrypt(struct aead_request *req)
1355 {
1356         struct aead_edesc *edesc;
1357         struct crypto_aead *aead = crypto_aead_reqtfm(req);
1358         struct caam_ctx *ctx = crypto_aead_ctx_dma(aead);
1359         struct caam_request *caam_req = aead_request_ctx_dma(req);
1360         int ret;
1361
1362         /* allocate extended descriptor */
1363         edesc = aead_edesc_alloc(req, false);
1364         if (IS_ERR(edesc))
1365                 return PTR_ERR(edesc);
1366
1367         caam_req->flc = &ctx->flc[DECRYPT];
1368         caam_req->flc_dma = ctx->flc_dma[DECRYPT];
1369         caam_req->cbk = aead_decrypt_done;
1370         caam_req->ctx = &req->base;
1371         caam_req->edesc = edesc;
1372         ret = dpaa2_caam_enqueue(ctx->dev, caam_req);
1373         if (ret != -EINPROGRESS &&
1374             !(ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG)) {
1375                 aead_unmap(ctx->dev, edesc, req);
1376                 qi_cache_free(edesc);
1377         }
1378
1379         return ret;
1380 }
1381
1382 static int ipsec_gcm_encrypt(struct aead_request *req)
1383 {
1384         return crypto_ipsec_check_assoclen(req->assoclen) ? : aead_encrypt(req);
1385 }
1386
1387 static int ipsec_gcm_decrypt(struct aead_request *req)
1388 {
1389         return crypto_ipsec_check_assoclen(req->assoclen) ? : aead_decrypt(req);
1390 }
1391
1392 static void skcipher_encrypt_done(void *cbk_ctx, u32 status)
1393 {
1394         struct crypto_async_request *areq = cbk_ctx;
1395         struct skcipher_request *req = skcipher_request_cast(areq);
1396         struct caam_request *req_ctx = to_caam_req(areq);
1397         struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req);
1398         struct caam_ctx *ctx = crypto_skcipher_ctx_dma(skcipher);
1399         struct skcipher_edesc *edesc = req_ctx->edesc;
1400         int ecode = 0;
1401         int ivsize = crypto_skcipher_ivsize(skcipher);
1402
1403         dev_dbg(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
1404
1405         if (unlikely(status))
1406                 ecode = caam_qi2_strstatus(ctx->dev, status);
1407
1408         print_hex_dump_debug("dstiv  @" __stringify(__LINE__)": ",
1409                              DUMP_PREFIX_ADDRESS, 16, 4, req->iv,
1410                              edesc->src_nents > 1 ? 100 : ivsize, 1);
1411         caam_dump_sg("dst    @" __stringify(__LINE__)": ",
1412                      DUMP_PREFIX_ADDRESS, 16, 4, req->dst,
1413                      edesc->dst_nents > 1 ? 100 : req->cryptlen, 1);
1414
1415         skcipher_unmap(ctx->dev, edesc, req);
1416
1417         /*
1418          * The crypto API expects us to set the IV (req->iv) to the last
1419          * ciphertext block (CBC mode) or last counter (CTR mode).
1420          * This is used e.g. by the CTS mode.
1421          */
1422         if (!ecode)
1423                 memcpy(req->iv, (u8 *)&edesc->sgt[0] + edesc->qm_sg_bytes,
1424                        ivsize);
1425
1426         qi_cache_free(edesc);
1427         skcipher_request_complete(req, ecode);
1428 }
1429
1430 static void skcipher_decrypt_done(void *cbk_ctx, u32 status)
1431 {
1432         struct crypto_async_request *areq = cbk_ctx;
1433         struct skcipher_request *req = skcipher_request_cast(areq);
1434         struct caam_request *req_ctx = to_caam_req(areq);
1435         struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req);
1436         struct caam_ctx *ctx = crypto_skcipher_ctx_dma(skcipher);
1437         struct skcipher_edesc *edesc = req_ctx->edesc;
1438         int ecode = 0;
1439         int ivsize = crypto_skcipher_ivsize(skcipher);
1440
1441         dev_dbg(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
1442
1443         if (unlikely(status))
1444                 ecode = caam_qi2_strstatus(ctx->dev, status);
1445
1446         print_hex_dump_debug("dstiv  @" __stringify(__LINE__)": ",
1447                              DUMP_PREFIX_ADDRESS, 16, 4, req->iv,
1448                              edesc->src_nents > 1 ? 100 : ivsize, 1);
1449         caam_dump_sg("dst    @" __stringify(__LINE__)": ",
1450                      DUMP_PREFIX_ADDRESS, 16, 4, req->dst,
1451                      edesc->dst_nents > 1 ? 100 : req->cryptlen, 1);
1452
1453         skcipher_unmap(ctx->dev, edesc, req);
1454
1455         /*
1456          * The crypto API expects us to set the IV (req->iv) to the last
1457          * ciphertext block (CBC mode) or last counter (CTR mode).
1458          * This is used e.g. by the CTS mode.
1459          */
1460         if (!ecode)
1461                 memcpy(req->iv, (u8 *)&edesc->sgt[0] + edesc->qm_sg_bytes,
1462                        ivsize);
1463
1464         qi_cache_free(edesc);
1465         skcipher_request_complete(req, ecode);
1466 }
1467
1468 static inline bool xts_skcipher_ivsize(struct skcipher_request *req)
1469 {
1470         struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req);
1471         unsigned int ivsize = crypto_skcipher_ivsize(skcipher);
1472
1473         return !!get_unaligned((u64 *)(req->iv + (ivsize / 2)));
1474 }
1475
1476 static int skcipher_encrypt(struct skcipher_request *req)
1477 {
1478         struct skcipher_edesc *edesc;
1479         struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req);
1480         struct caam_ctx *ctx = crypto_skcipher_ctx_dma(skcipher);
1481         struct caam_request *caam_req = skcipher_request_ctx_dma(req);
1482         struct dpaa2_caam_priv *priv = dev_get_drvdata(ctx->dev);
1483         int ret;
1484
1485         /*
1486          * XTS is expected to return an error even for input length = 0
1487          * Note that the case input length < block size will be caught during
1488          * HW offloading and return an error.
1489          */
1490         if (!req->cryptlen && !ctx->fallback)
1491                 return 0;
1492
1493         if (ctx->fallback && ((priv->sec_attr.era <= 8 && xts_skcipher_ivsize(req)) ||
1494                               ctx->xts_key_fallback)) {
1495                 skcipher_request_set_tfm(&caam_req->fallback_req, ctx->fallback);
1496                 skcipher_request_set_callback(&caam_req->fallback_req,
1497                                               req->base.flags,
1498                                               req->base.complete,
1499                                               req->base.data);
1500                 skcipher_request_set_crypt(&caam_req->fallback_req, req->src,
1501                                            req->dst, req->cryptlen, req->iv);
1502
1503                 return crypto_skcipher_encrypt(&caam_req->fallback_req);
1504         }
1505
1506         /* allocate extended descriptor */
1507         edesc = skcipher_edesc_alloc(req);
1508         if (IS_ERR(edesc))
1509                 return PTR_ERR(edesc);
1510
1511         caam_req->flc = &ctx->flc[ENCRYPT];
1512         caam_req->flc_dma = ctx->flc_dma[ENCRYPT];
1513         caam_req->cbk = skcipher_encrypt_done;
1514         caam_req->ctx = &req->base;
1515         caam_req->edesc = edesc;
1516         ret = dpaa2_caam_enqueue(ctx->dev, caam_req);
1517         if (ret != -EINPROGRESS &&
1518             !(ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG)) {
1519                 skcipher_unmap(ctx->dev, edesc, req);
1520                 qi_cache_free(edesc);
1521         }
1522
1523         return ret;
1524 }
1525
1526 static int skcipher_decrypt(struct skcipher_request *req)
1527 {
1528         struct skcipher_edesc *edesc;
1529         struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req);
1530         struct caam_ctx *ctx = crypto_skcipher_ctx_dma(skcipher);
1531         struct caam_request *caam_req = skcipher_request_ctx_dma(req);
1532         struct dpaa2_caam_priv *priv = dev_get_drvdata(ctx->dev);
1533         int ret;
1534
1535         /*
1536          * XTS is expected to return an error even for input length = 0
1537          * Note that the case input length < block size will be caught during
1538          * HW offloading and return an error.
1539          */
1540         if (!req->cryptlen && !ctx->fallback)
1541                 return 0;
1542
1543         if (ctx->fallback && ((priv->sec_attr.era <= 8 && xts_skcipher_ivsize(req)) ||
1544                               ctx->xts_key_fallback)) {
1545                 skcipher_request_set_tfm(&caam_req->fallback_req, ctx->fallback);
1546                 skcipher_request_set_callback(&caam_req->fallback_req,
1547                                               req->base.flags,
1548                                               req->base.complete,
1549                                               req->base.data);
1550                 skcipher_request_set_crypt(&caam_req->fallback_req, req->src,
1551                                            req->dst, req->cryptlen, req->iv);
1552
1553                 return crypto_skcipher_decrypt(&caam_req->fallback_req);
1554         }
1555
1556         /* allocate extended descriptor */
1557         edesc = skcipher_edesc_alloc(req);
1558         if (IS_ERR(edesc))
1559                 return PTR_ERR(edesc);
1560
1561         caam_req->flc = &ctx->flc[DECRYPT];
1562         caam_req->flc_dma = ctx->flc_dma[DECRYPT];
1563         caam_req->cbk = skcipher_decrypt_done;
1564         caam_req->ctx = &req->base;
1565         caam_req->edesc = edesc;
1566         ret = dpaa2_caam_enqueue(ctx->dev, caam_req);
1567         if (ret != -EINPROGRESS &&
1568             !(ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG)) {
1569                 skcipher_unmap(ctx->dev, edesc, req);
1570                 qi_cache_free(edesc);
1571         }
1572
1573         return ret;
1574 }
1575
1576 static int caam_cra_init(struct caam_ctx *ctx, struct caam_alg_entry *caam,
1577                          bool uses_dkp)
1578 {
1579         dma_addr_t dma_addr;
1580         int i;
1581
1582         /* copy descriptor header template value */
1583         ctx->cdata.algtype = OP_TYPE_CLASS1_ALG | caam->class1_alg_type;
1584         ctx->adata.algtype = OP_TYPE_CLASS2_ALG | caam->class2_alg_type;
1585
1586         ctx->dev = caam->dev;
1587         ctx->dir = uses_dkp ? DMA_BIDIRECTIONAL : DMA_TO_DEVICE;
1588
1589         dma_addr = dma_map_single_attrs(ctx->dev, ctx->flc,
1590                                         offsetof(struct caam_ctx, flc_dma),
1591                                         ctx->dir, DMA_ATTR_SKIP_CPU_SYNC);
1592         if (dma_mapping_error(ctx->dev, dma_addr)) {
1593                 dev_err(ctx->dev, "unable to map key, shared descriptors\n");
1594                 return -ENOMEM;
1595         }
1596
1597         for (i = 0; i < NUM_OP; i++)
1598                 ctx->flc_dma[i] = dma_addr + i * sizeof(ctx->flc[i]);
1599         ctx->key_dma = dma_addr + NUM_OP * sizeof(ctx->flc[0]);
1600
1601         return 0;
1602 }
1603
1604 static int caam_cra_init_skcipher(struct crypto_skcipher *tfm)
1605 {
1606         struct skcipher_alg *alg = crypto_skcipher_alg(tfm);
1607         struct caam_skcipher_alg *caam_alg =
1608                 container_of(alg, typeof(*caam_alg), skcipher);
1609         struct caam_ctx *ctx = crypto_skcipher_ctx_dma(tfm);
1610         u32 alg_aai = caam_alg->caam.class1_alg_type & OP_ALG_AAI_MASK;
1611         int ret = 0;
1612
1613         if (alg_aai == OP_ALG_AAI_XTS) {
1614                 const char *tfm_name = crypto_tfm_alg_name(&tfm->base);
1615                 struct crypto_skcipher *fallback;
1616
1617                 fallback = crypto_alloc_skcipher(tfm_name, 0,
1618                                                  CRYPTO_ALG_NEED_FALLBACK);
1619                 if (IS_ERR(fallback)) {
1620                         dev_err(caam_alg->caam.dev,
1621                                 "Failed to allocate %s fallback: %ld\n",
1622                                 tfm_name, PTR_ERR(fallback));
1623                         return PTR_ERR(fallback);
1624                 }
1625
1626                 ctx->fallback = fallback;
1627                 crypto_skcipher_set_reqsize_dma(
1628                         tfm, sizeof(struct caam_request) +
1629                              crypto_skcipher_reqsize(fallback));
1630         } else {
1631                 crypto_skcipher_set_reqsize_dma(tfm,
1632                                                 sizeof(struct caam_request));
1633         }
1634
1635         ret = caam_cra_init(ctx, &caam_alg->caam, false);
1636         if (ret && ctx->fallback)
1637                 crypto_free_skcipher(ctx->fallback);
1638
1639         return ret;
1640 }
1641
1642 static int caam_cra_init_aead(struct crypto_aead *tfm)
1643 {
1644         struct aead_alg *alg = crypto_aead_alg(tfm);
1645         struct caam_aead_alg *caam_alg = container_of(alg, typeof(*caam_alg),
1646                                                       aead);
1647
1648         crypto_aead_set_reqsize_dma(tfm, sizeof(struct caam_request));
1649         return caam_cra_init(crypto_aead_ctx_dma(tfm), &caam_alg->caam,
1650                              !caam_alg->caam.nodkp);
1651 }
1652
1653 static void caam_exit_common(struct caam_ctx *ctx)
1654 {
1655         dma_unmap_single_attrs(ctx->dev, ctx->flc_dma[0],
1656                                offsetof(struct caam_ctx, flc_dma), ctx->dir,
1657                                DMA_ATTR_SKIP_CPU_SYNC);
1658 }
1659
1660 static void caam_cra_exit(struct crypto_skcipher *tfm)
1661 {
1662         struct caam_ctx *ctx = crypto_skcipher_ctx_dma(tfm);
1663
1664         if (ctx->fallback)
1665                 crypto_free_skcipher(ctx->fallback);
1666         caam_exit_common(ctx);
1667 }
1668
1669 static void caam_cra_exit_aead(struct crypto_aead *tfm)
1670 {
1671         caam_exit_common(crypto_aead_ctx_dma(tfm));
1672 }
1673
1674 static struct caam_skcipher_alg driver_algs[] = {
1675         {
1676                 .skcipher = {
1677                         .base = {
1678                                 .cra_name = "cbc(aes)",
1679                                 .cra_driver_name = "cbc-aes-caam-qi2",
1680                                 .cra_blocksize = AES_BLOCK_SIZE,
1681                         },
1682                         .setkey = aes_skcipher_setkey,
1683                         .encrypt = skcipher_encrypt,
1684                         .decrypt = skcipher_decrypt,
1685                         .min_keysize = AES_MIN_KEY_SIZE,
1686                         .max_keysize = AES_MAX_KEY_SIZE,
1687                         .ivsize = AES_BLOCK_SIZE,
1688                 },
1689                 .caam.class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
1690         },
1691         {
1692                 .skcipher = {
1693                         .base = {
1694                                 .cra_name = "cbc(des3_ede)",
1695                                 .cra_driver_name = "cbc-3des-caam-qi2",
1696                                 .cra_blocksize = DES3_EDE_BLOCK_SIZE,
1697                         },
1698                         .setkey = des3_skcipher_setkey,
1699                         .encrypt = skcipher_encrypt,
1700                         .decrypt = skcipher_decrypt,
1701                         .min_keysize = DES3_EDE_KEY_SIZE,
1702                         .max_keysize = DES3_EDE_KEY_SIZE,
1703                         .ivsize = DES3_EDE_BLOCK_SIZE,
1704                 },
1705                 .caam.class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
1706         },
1707         {
1708                 .skcipher = {
1709                         .base = {
1710                                 .cra_name = "cbc(des)",
1711                                 .cra_driver_name = "cbc-des-caam-qi2",
1712                                 .cra_blocksize = DES_BLOCK_SIZE,
1713                         },
1714                         .setkey = des_skcipher_setkey,
1715                         .encrypt = skcipher_encrypt,
1716                         .decrypt = skcipher_decrypt,
1717                         .min_keysize = DES_KEY_SIZE,
1718                         .max_keysize = DES_KEY_SIZE,
1719                         .ivsize = DES_BLOCK_SIZE,
1720                 },
1721                 .caam.class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
1722         },
1723         {
1724                 .skcipher = {
1725                         .base = {
1726                                 .cra_name = "ctr(aes)",
1727                                 .cra_driver_name = "ctr-aes-caam-qi2",
1728                                 .cra_blocksize = 1,
1729                         },
1730                         .setkey = ctr_skcipher_setkey,
1731                         .encrypt = skcipher_encrypt,
1732                         .decrypt = skcipher_decrypt,
1733                         .min_keysize = AES_MIN_KEY_SIZE,
1734                         .max_keysize = AES_MAX_KEY_SIZE,
1735                         .ivsize = AES_BLOCK_SIZE,
1736                         .chunksize = AES_BLOCK_SIZE,
1737                 },
1738                 .caam.class1_alg_type = OP_ALG_ALGSEL_AES |
1739                                         OP_ALG_AAI_CTR_MOD128,
1740         },
1741         {
1742                 .skcipher = {
1743                         .base = {
1744                                 .cra_name = "rfc3686(ctr(aes))",
1745                                 .cra_driver_name = "rfc3686-ctr-aes-caam-qi2",
1746                                 .cra_blocksize = 1,
1747                         },
1748                         .setkey = rfc3686_skcipher_setkey,
1749                         .encrypt = skcipher_encrypt,
1750                         .decrypt = skcipher_decrypt,
1751                         .min_keysize = AES_MIN_KEY_SIZE +
1752                                        CTR_RFC3686_NONCE_SIZE,
1753                         .max_keysize = AES_MAX_KEY_SIZE +
1754                                        CTR_RFC3686_NONCE_SIZE,
1755                         .ivsize = CTR_RFC3686_IV_SIZE,
1756                         .chunksize = AES_BLOCK_SIZE,
1757                 },
1758                 .caam = {
1759                         .class1_alg_type = OP_ALG_ALGSEL_AES |
1760                                            OP_ALG_AAI_CTR_MOD128,
1761                         .rfc3686 = true,
1762                 },
1763         },
1764         {
1765                 .skcipher = {
1766                         .base = {
1767                                 .cra_name = "xts(aes)",
1768                                 .cra_driver_name = "xts-aes-caam-qi2",
1769                                 .cra_flags = CRYPTO_ALG_NEED_FALLBACK,
1770                                 .cra_blocksize = AES_BLOCK_SIZE,
1771                         },
1772                         .setkey = xts_skcipher_setkey,
1773                         .encrypt = skcipher_encrypt,
1774                         .decrypt = skcipher_decrypt,
1775                         .min_keysize = 2 * AES_MIN_KEY_SIZE,
1776                         .max_keysize = 2 * AES_MAX_KEY_SIZE,
1777                         .ivsize = AES_BLOCK_SIZE,
1778                 },
1779                 .caam.class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_XTS,
1780         },
1781         {
1782                 .skcipher = {
1783                         .base = {
1784                                 .cra_name = "chacha20",
1785                                 .cra_driver_name = "chacha20-caam-qi2",
1786                                 .cra_blocksize = 1,
1787                         },
1788                         .setkey = chacha20_skcipher_setkey,
1789                         .encrypt = skcipher_encrypt,
1790                         .decrypt = skcipher_decrypt,
1791                         .min_keysize = CHACHA_KEY_SIZE,
1792                         .max_keysize = CHACHA_KEY_SIZE,
1793                         .ivsize = CHACHA_IV_SIZE,
1794                 },
1795                 .caam.class1_alg_type = OP_ALG_ALGSEL_CHACHA20,
1796         },
1797 };
1798
1799 static struct caam_aead_alg driver_aeads[] = {
1800         {
1801                 .aead = {
1802                         .base = {
1803                                 .cra_name = "rfc4106(gcm(aes))",
1804                                 .cra_driver_name = "rfc4106-gcm-aes-caam-qi2",
1805                                 .cra_blocksize = 1,
1806                         },
1807                         .setkey = rfc4106_setkey,
1808                         .setauthsize = rfc4106_setauthsize,
1809                         .encrypt = ipsec_gcm_encrypt,
1810                         .decrypt = ipsec_gcm_decrypt,
1811                         .ivsize = 8,
1812                         .maxauthsize = AES_BLOCK_SIZE,
1813                 },
1814                 .caam = {
1815                         .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_GCM,
1816                         .nodkp = true,
1817                 },
1818         },
1819         {
1820                 .aead = {
1821                         .base = {
1822                                 .cra_name = "rfc4543(gcm(aes))",
1823                                 .cra_driver_name = "rfc4543-gcm-aes-caam-qi2",
1824                                 .cra_blocksize = 1,
1825                         },
1826                         .setkey = rfc4543_setkey,
1827                         .setauthsize = rfc4543_setauthsize,
1828                         .encrypt = ipsec_gcm_encrypt,
1829                         .decrypt = ipsec_gcm_decrypt,
1830                         .ivsize = 8,
1831                         .maxauthsize = AES_BLOCK_SIZE,
1832                 },
1833                 .caam = {
1834                         .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_GCM,
1835                         .nodkp = true,
1836                 },
1837         },
1838         /* Galois Counter Mode */
1839         {
1840                 .aead = {
1841                         .base = {
1842                                 .cra_name = "gcm(aes)",
1843                                 .cra_driver_name = "gcm-aes-caam-qi2",
1844                                 .cra_blocksize = 1,
1845                         },
1846                         .setkey = gcm_setkey,
1847                         .setauthsize = gcm_setauthsize,
1848                         .encrypt = aead_encrypt,
1849                         .decrypt = aead_decrypt,
1850                         .ivsize = 12,
1851                         .maxauthsize = AES_BLOCK_SIZE,
1852                 },
1853                 .caam = {
1854                         .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_GCM,
1855                         .nodkp = true,
1856                 }
1857         },
1858         /* single-pass ipsec_esp descriptor */
1859         {
1860                 .aead = {
1861                         .base = {
1862                                 .cra_name = "authenc(hmac(md5),cbc(aes))",
1863                                 .cra_driver_name = "authenc-hmac-md5-"
1864                                                    "cbc-aes-caam-qi2",
1865                                 .cra_blocksize = AES_BLOCK_SIZE,
1866                         },
1867                         .setkey = aead_setkey,
1868                         .setauthsize = aead_setauthsize,
1869                         .encrypt = aead_encrypt,
1870                         .decrypt = aead_decrypt,
1871                         .ivsize = AES_BLOCK_SIZE,
1872                         .maxauthsize = MD5_DIGEST_SIZE,
1873                 },
1874                 .caam = {
1875                         .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
1876                         .class2_alg_type = OP_ALG_ALGSEL_MD5 |
1877                                            OP_ALG_AAI_HMAC_PRECOMP,
1878                 }
1879         },
1880         {
1881                 .aead = {
1882                         .base = {
1883                                 .cra_name = "echainiv(authenc(hmac(md5),"
1884                                             "cbc(aes)))",
1885                                 .cra_driver_name = "echainiv-authenc-hmac-md5-"
1886                                                    "cbc-aes-caam-qi2",
1887                                 .cra_blocksize = AES_BLOCK_SIZE,
1888                         },
1889                         .setkey = aead_setkey,
1890                         .setauthsize = aead_setauthsize,
1891                         .encrypt = aead_encrypt,
1892                         .decrypt = aead_decrypt,
1893                         .ivsize = AES_BLOCK_SIZE,
1894                         .maxauthsize = MD5_DIGEST_SIZE,
1895                 },
1896                 .caam = {
1897                         .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
1898                         .class2_alg_type = OP_ALG_ALGSEL_MD5 |
1899                                            OP_ALG_AAI_HMAC_PRECOMP,
1900                         .geniv = true,
1901                 }
1902         },
1903         {
1904                 .aead = {
1905                         .base = {
1906                                 .cra_name = "authenc(hmac(sha1),cbc(aes))",
1907                                 .cra_driver_name = "authenc-hmac-sha1-"
1908                                                    "cbc-aes-caam-qi2",
1909                                 .cra_blocksize = AES_BLOCK_SIZE,
1910                         },
1911                         .setkey = aead_setkey,
1912                         .setauthsize = aead_setauthsize,
1913                         .encrypt = aead_encrypt,
1914                         .decrypt = aead_decrypt,
1915                         .ivsize = AES_BLOCK_SIZE,
1916                         .maxauthsize = SHA1_DIGEST_SIZE,
1917                 },
1918                 .caam = {
1919                         .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
1920                         .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
1921                                            OP_ALG_AAI_HMAC_PRECOMP,
1922                 }
1923         },
1924         {
1925                 .aead = {
1926                         .base = {
1927                                 .cra_name = "echainiv(authenc(hmac(sha1),"
1928                                             "cbc(aes)))",
1929                                 .cra_driver_name = "echainiv-authenc-"
1930                                                    "hmac-sha1-cbc-aes-caam-qi2",
1931                                 .cra_blocksize = AES_BLOCK_SIZE,
1932                         },
1933                         .setkey = aead_setkey,
1934                         .setauthsize = aead_setauthsize,
1935                         .encrypt = aead_encrypt,
1936                         .decrypt = aead_decrypt,
1937                         .ivsize = AES_BLOCK_SIZE,
1938                         .maxauthsize = SHA1_DIGEST_SIZE,
1939                 },
1940                 .caam = {
1941                         .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
1942                         .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
1943                                            OP_ALG_AAI_HMAC_PRECOMP,
1944                         .geniv = true,
1945                 },
1946         },
1947         {
1948                 .aead = {
1949                         .base = {
1950                                 .cra_name = "authenc(hmac(sha224),cbc(aes))",
1951                                 .cra_driver_name = "authenc-hmac-sha224-"
1952                                                    "cbc-aes-caam-qi2",
1953                                 .cra_blocksize = AES_BLOCK_SIZE,
1954                         },
1955                         .setkey = aead_setkey,
1956                         .setauthsize = aead_setauthsize,
1957                         .encrypt = aead_encrypt,
1958                         .decrypt = aead_decrypt,
1959                         .ivsize = AES_BLOCK_SIZE,
1960                         .maxauthsize = SHA224_DIGEST_SIZE,
1961                 },
1962                 .caam = {
1963                         .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
1964                         .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
1965                                            OP_ALG_AAI_HMAC_PRECOMP,
1966                 }
1967         },
1968         {
1969                 .aead = {
1970                         .base = {
1971                                 .cra_name = "echainiv(authenc(hmac(sha224),"
1972                                             "cbc(aes)))",
1973                                 .cra_driver_name = "echainiv-authenc-"
1974                                                    "hmac-sha224-cbc-aes-caam-qi2",
1975                                 .cra_blocksize = AES_BLOCK_SIZE,
1976                         },
1977                         .setkey = aead_setkey,
1978                         .setauthsize = aead_setauthsize,
1979                         .encrypt = aead_encrypt,
1980                         .decrypt = aead_decrypt,
1981                         .ivsize = AES_BLOCK_SIZE,
1982                         .maxauthsize = SHA224_DIGEST_SIZE,
1983                 },
1984                 .caam = {
1985                         .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
1986                         .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
1987                                            OP_ALG_AAI_HMAC_PRECOMP,
1988                         .geniv = true,
1989                 }
1990         },
1991         {
1992                 .aead = {
1993                         .base = {
1994                                 .cra_name = "authenc(hmac(sha256),cbc(aes))",
1995                                 .cra_driver_name = "authenc-hmac-sha256-"
1996                                                    "cbc-aes-caam-qi2",
1997                                 .cra_blocksize = AES_BLOCK_SIZE,
1998                         },
1999                         .setkey = aead_setkey,
2000                         .setauthsize = aead_setauthsize,
2001                         .encrypt = aead_encrypt,
2002                         .decrypt = aead_decrypt,
2003                         .ivsize = AES_BLOCK_SIZE,
2004                         .maxauthsize = SHA256_DIGEST_SIZE,
2005                 },
2006                 .caam = {
2007                         .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
2008                         .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
2009                                            OP_ALG_AAI_HMAC_PRECOMP,
2010                 }
2011         },
2012         {
2013                 .aead = {
2014                         .base = {
2015                                 .cra_name = "echainiv(authenc(hmac(sha256),"
2016                                             "cbc(aes)))",
2017                                 .cra_driver_name = "echainiv-authenc-"
2018                                                    "hmac-sha256-cbc-aes-"
2019                                                    "caam-qi2",
2020                                 .cra_blocksize = AES_BLOCK_SIZE,
2021                         },
2022                         .setkey = aead_setkey,
2023                         .setauthsize = aead_setauthsize,
2024                         .encrypt = aead_encrypt,
2025                         .decrypt = aead_decrypt,
2026                         .ivsize = AES_BLOCK_SIZE,
2027                         .maxauthsize = SHA256_DIGEST_SIZE,
2028                 },
2029                 .caam = {
2030                         .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
2031                         .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
2032                                            OP_ALG_AAI_HMAC_PRECOMP,
2033                         .geniv = true,
2034                 }
2035         },
2036         {
2037                 .aead = {
2038                         .base = {
2039                                 .cra_name = "authenc(hmac(sha384),cbc(aes))",
2040                                 .cra_driver_name = "authenc-hmac-sha384-"
2041                                                    "cbc-aes-caam-qi2",
2042                                 .cra_blocksize = AES_BLOCK_SIZE,
2043                         },
2044                         .setkey = aead_setkey,
2045                         .setauthsize = aead_setauthsize,
2046                         .encrypt = aead_encrypt,
2047                         .decrypt = aead_decrypt,
2048                         .ivsize = AES_BLOCK_SIZE,
2049                         .maxauthsize = SHA384_DIGEST_SIZE,
2050                 },
2051                 .caam = {
2052                         .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
2053                         .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
2054                                            OP_ALG_AAI_HMAC_PRECOMP,
2055                 }
2056         },
2057         {
2058                 .aead = {
2059                         .base = {
2060                                 .cra_name = "echainiv(authenc(hmac(sha384),"
2061                                             "cbc(aes)))",
2062                                 .cra_driver_name = "echainiv-authenc-"
2063                                                    "hmac-sha384-cbc-aes-"
2064                                                    "caam-qi2",
2065                                 .cra_blocksize = AES_BLOCK_SIZE,
2066                         },
2067                         .setkey = aead_setkey,
2068                         .setauthsize = aead_setauthsize,
2069                         .encrypt = aead_encrypt,
2070                         .decrypt = aead_decrypt,
2071                         .ivsize = AES_BLOCK_SIZE,
2072                         .maxauthsize = SHA384_DIGEST_SIZE,
2073                 },
2074                 .caam = {
2075                         .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
2076                         .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
2077                                            OP_ALG_AAI_HMAC_PRECOMP,
2078                         .geniv = true,
2079                 }
2080         },
2081         {
2082                 .aead = {
2083                         .base = {
2084                                 .cra_name = "authenc(hmac(sha512),cbc(aes))",
2085                                 .cra_driver_name = "authenc-hmac-sha512-"
2086                                                    "cbc-aes-caam-qi2",
2087                                 .cra_blocksize = AES_BLOCK_SIZE,
2088                         },
2089                         .setkey = aead_setkey,
2090                         .setauthsize = aead_setauthsize,
2091                         .encrypt = aead_encrypt,
2092                         .decrypt = aead_decrypt,
2093                         .ivsize = AES_BLOCK_SIZE,
2094                         .maxauthsize = SHA512_DIGEST_SIZE,
2095                 },
2096                 .caam = {
2097                         .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
2098                         .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
2099                                            OP_ALG_AAI_HMAC_PRECOMP,
2100                 }
2101         },
2102         {
2103                 .aead = {
2104                         .base = {
2105                                 .cra_name = "echainiv(authenc(hmac(sha512),"
2106                                             "cbc(aes)))",
2107                                 .cra_driver_name = "echainiv-authenc-"
2108                                                    "hmac-sha512-cbc-aes-"
2109                                                    "caam-qi2",
2110                                 .cra_blocksize = AES_BLOCK_SIZE,
2111                         },
2112                         .setkey = aead_setkey,
2113                         .setauthsize = aead_setauthsize,
2114                         .encrypt = aead_encrypt,
2115                         .decrypt = aead_decrypt,
2116                         .ivsize = AES_BLOCK_SIZE,
2117                         .maxauthsize = SHA512_DIGEST_SIZE,
2118                 },
2119                 .caam = {
2120                         .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
2121                         .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
2122                                            OP_ALG_AAI_HMAC_PRECOMP,
2123                         .geniv = true,
2124                 }
2125         },
2126         {
2127                 .aead = {
2128                         .base = {
2129                                 .cra_name = "authenc(hmac(md5),cbc(des3_ede))",
2130                                 .cra_driver_name = "authenc-hmac-md5-"
2131                                                    "cbc-des3_ede-caam-qi2",
2132                                 .cra_blocksize = DES3_EDE_BLOCK_SIZE,
2133                         },
2134                         .setkey = des3_aead_setkey,
2135                         .setauthsize = aead_setauthsize,
2136                         .encrypt = aead_encrypt,
2137                         .decrypt = aead_decrypt,
2138                         .ivsize = DES3_EDE_BLOCK_SIZE,
2139                         .maxauthsize = MD5_DIGEST_SIZE,
2140                 },
2141                 .caam = {
2142                         .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
2143                         .class2_alg_type = OP_ALG_ALGSEL_MD5 |
2144                                            OP_ALG_AAI_HMAC_PRECOMP,
2145                 }
2146         },
2147         {
2148                 .aead = {
2149                         .base = {
2150                                 .cra_name = "echainiv(authenc(hmac(md5),"
2151                                             "cbc(des3_ede)))",
2152                                 .cra_driver_name = "echainiv-authenc-hmac-md5-"
2153                                                    "cbc-des3_ede-caam-qi2",
2154                                 .cra_blocksize = DES3_EDE_BLOCK_SIZE,
2155                         },
2156                         .setkey = des3_aead_setkey,
2157                         .setauthsize = aead_setauthsize,
2158                         .encrypt = aead_encrypt,
2159                         .decrypt = aead_decrypt,
2160                         .ivsize = DES3_EDE_BLOCK_SIZE,
2161                         .maxauthsize = MD5_DIGEST_SIZE,
2162                 },
2163                 .caam = {
2164                         .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
2165                         .class2_alg_type = OP_ALG_ALGSEL_MD5 |
2166                                            OP_ALG_AAI_HMAC_PRECOMP,
2167                         .geniv = true,
2168                 }
2169         },
2170         {
2171                 .aead = {
2172                         .base = {
2173                                 .cra_name = "authenc(hmac(sha1),"
2174                                             "cbc(des3_ede))",
2175                                 .cra_driver_name = "authenc-hmac-sha1-"
2176                                                    "cbc-des3_ede-caam-qi2",
2177                                 .cra_blocksize = DES3_EDE_BLOCK_SIZE,
2178                         },
2179                         .setkey = des3_aead_setkey,
2180                         .setauthsize = aead_setauthsize,
2181                         .encrypt = aead_encrypt,
2182                         .decrypt = aead_decrypt,
2183                         .ivsize = DES3_EDE_BLOCK_SIZE,
2184                         .maxauthsize = SHA1_DIGEST_SIZE,
2185                 },
2186                 .caam = {
2187                         .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
2188                         .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
2189                                            OP_ALG_AAI_HMAC_PRECOMP,
2190                 },
2191         },
2192         {
2193                 .aead = {
2194                         .base = {
2195                                 .cra_name = "echainiv(authenc(hmac(sha1),"
2196                                             "cbc(des3_ede)))",
2197                                 .cra_driver_name = "echainiv-authenc-"
2198                                                    "hmac-sha1-"
2199                                                    "cbc-des3_ede-caam-qi2",
2200                                 .cra_blocksize = DES3_EDE_BLOCK_SIZE,
2201                         },
2202                         .setkey = des3_aead_setkey,
2203                         .setauthsize = aead_setauthsize,
2204                         .encrypt = aead_encrypt,
2205                         .decrypt = aead_decrypt,
2206                         .ivsize = DES3_EDE_BLOCK_SIZE,
2207                         .maxauthsize = SHA1_DIGEST_SIZE,
2208                 },
2209                 .caam = {
2210                         .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
2211                         .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
2212                                            OP_ALG_AAI_HMAC_PRECOMP,
2213                         .geniv = true,
2214                 }
2215         },
2216         {
2217                 .aead = {
2218                         .base = {
2219                                 .cra_name = "authenc(hmac(sha224),"
2220                                             "cbc(des3_ede))",
2221                                 .cra_driver_name = "authenc-hmac-sha224-"
2222                                                    "cbc-des3_ede-caam-qi2",
2223                                 .cra_blocksize = DES3_EDE_BLOCK_SIZE,
2224                         },
2225                         .setkey = des3_aead_setkey,
2226                         .setauthsize = aead_setauthsize,
2227                         .encrypt = aead_encrypt,
2228                         .decrypt = aead_decrypt,
2229                         .ivsize = DES3_EDE_BLOCK_SIZE,
2230                         .maxauthsize = SHA224_DIGEST_SIZE,
2231                 },
2232                 .caam = {
2233                         .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
2234                         .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
2235                                            OP_ALG_AAI_HMAC_PRECOMP,
2236                 },
2237         },
2238         {
2239                 .aead = {
2240                         .base = {
2241                                 .cra_name = "echainiv(authenc(hmac(sha224),"
2242                                             "cbc(des3_ede)))",
2243                                 .cra_driver_name = "echainiv-authenc-"
2244                                                    "hmac-sha224-"
2245                                                    "cbc-des3_ede-caam-qi2",
2246                                 .cra_blocksize = DES3_EDE_BLOCK_SIZE,
2247                         },
2248                         .setkey = des3_aead_setkey,
2249                         .setauthsize = aead_setauthsize,
2250                         .encrypt = aead_encrypt,
2251                         .decrypt = aead_decrypt,
2252                         .ivsize = DES3_EDE_BLOCK_SIZE,
2253                         .maxauthsize = SHA224_DIGEST_SIZE,
2254                 },
2255                 .caam = {
2256                         .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
2257                         .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
2258                                            OP_ALG_AAI_HMAC_PRECOMP,
2259                         .geniv = true,
2260                 }
2261         },
2262         {
2263                 .aead = {
2264                         .base = {
2265                                 .cra_name = "authenc(hmac(sha256),"
2266                                             "cbc(des3_ede))",
2267                                 .cra_driver_name = "authenc-hmac-sha256-"
2268                                                    "cbc-des3_ede-caam-qi2",
2269                                 .cra_blocksize = DES3_EDE_BLOCK_SIZE,
2270                         },
2271                         .setkey = des3_aead_setkey,
2272                         .setauthsize = aead_setauthsize,
2273                         .encrypt = aead_encrypt,
2274                         .decrypt = aead_decrypt,
2275                         .ivsize = DES3_EDE_BLOCK_SIZE,
2276                         .maxauthsize = SHA256_DIGEST_SIZE,
2277                 },
2278                 .caam = {
2279                         .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
2280                         .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
2281                                            OP_ALG_AAI_HMAC_PRECOMP,
2282                 },
2283         },
2284         {
2285                 .aead = {
2286                         .base = {
2287                                 .cra_name = "echainiv(authenc(hmac(sha256),"
2288                                             "cbc(des3_ede)))",
2289                                 .cra_driver_name = "echainiv-authenc-"
2290                                                    "hmac-sha256-"
2291                                                    "cbc-des3_ede-caam-qi2",
2292                                 .cra_blocksize = DES3_EDE_BLOCK_SIZE,
2293                         },
2294                         .setkey = des3_aead_setkey,
2295                         .setauthsize = aead_setauthsize,
2296                         .encrypt = aead_encrypt,
2297                         .decrypt = aead_decrypt,
2298                         .ivsize = DES3_EDE_BLOCK_SIZE,
2299                         .maxauthsize = SHA256_DIGEST_SIZE,
2300                 },
2301                 .caam = {
2302                         .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
2303                         .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
2304                                            OP_ALG_AAI_HMAC_PRECOMP,
2305                         .geniv = true,
2306                 }
2307         },
2308         {
2309                 .aead = {
2310                         .base = {
2311                                 .cra_name = "authenc(hmac(sha384),"
2312                                             "cbc(des3_ede))",
2313                                 .cra_driver_name = "authenc-hmac-sha384-"
2314                                                    "cbc-des3_ede-caam-qi2",
2315                                 .cra_blocksize = DES3_EDE_BLOCK_SIZE,
2316                         },
2317                         .setkey = des3_aead_setkey,
2318                         .setauthsize = aead_setauthsize,
2319                         .encrypt = aead_encrypt,
2320                         .decrypt = aead_decrypt,
2321                         .ivsize = DES3_EDE_BLOCK_SIZE,
2322                         .maxauthsize = SHA384_DIGEST_SIZE,
2323                 },
2324                 .caam = {
2325                         .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
2326                         .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
2327                                            OP_ALG_AAI_HMAC_PRECOMP,
2328                 },
2329         },
2330         {
2331                 .aead = {
2332                         .base = {
2333                                 .cra_name = "echainiv(authenc(hmac(sha384),"
2334                                             "cbc(des3_ede)))",
2335                                 .cra_driver_name = "echainiv-authenc-"
2336                                                    "hmac-sha384-"
2337                                                    "cbc-des3_ede-caam-qi2",
2338                                 .cra_blocksize = DES3_EDE_BLOCK_SIZE,
2339                         },
2340                         .setkey = des3_aead_setkey,
2341                         .setauthsize = aead_setauthsize,
2342                         .encrypt = aead_encrypt,
2343                         .decrypt = aead_decrypt,
2344                         .ivsize = DES3_EDE_BLOCK_SIZE,
2345                         .maxauthsize = SHA384_DIGEST_SIZE,
2346                 },
2347                 .caam = {
2348                         .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
2349                         .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
2350                                            OP_ALG_AAI_HMAC_PRECOMP,
2351                         .geniv = true,
2352                 }
2353         },
2354         {
2355                 .aead = {
2356                         .base = {
2357                                 .cra_name = "authenc(hmac(sha512),"
2358                                             "cbc(des3_ede))",
2359                                 .cra_driver_name = "authenc-hmac-sha512-"
2360                                                    "cbc-des3_ede-caam-qi2",
2361                                 .cra_blocksize = DES3_EDE_BLOCK_SIZE,
2362                         },
2363                         .setkey = des3_aead_setkey,
2364                         .setauthsize = aead_setauthsize,
2365                         .encrypt = aead_encrypt,
2366                         .decrypt = aead_decrypt,
2367                         .ivsize = DES3_EDE_BLOCK_SIZE,
2368                         .maxauthsize = SHA512_DIGEST_SIZE,
2369                 },
2370                 .caam = {
2371                         .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
2372                         .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
2373                                            OP_ALG_AAI_HMAC_PRECOMP,
2374                 },
2375         },
2376         {
2377                 .aead = {
2378                         .base = {
2379                                 .cra_name = "echainiv(authenc(hmac(sha512),"
2380                                             "cbc(des3_ede)))",
2381                                 .cra_driver_name = "echainiv-authenc-"
2382                                                    "hmac-sha512-"
2383                                                    "cbc-des3_ede-caam-qi2",
2384                                 .cra_blocksize = DES3_EDE_BLOCK_SIZE,
2385                         },
2386                         .setkey = des3_aead_setkey,
2387                         .setauthsize = aead_setauthsize,
2388                         .encrypt = aead_encrypt,
2389                         .decrypt = aead_decrypt,
2390                         .ivsize = DES3_EDE_BLOCK_SIZE,
2391                         .maxauthsize = SHA512_DIGEST_SIZE,
2392                 },
2393                 .caam = {
2394                         .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
2395                         .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
2396                                            OP_ALG_AAI_HMAC_PRECOMP,
2397                         .geniv = true,
2398                 }
2399         },
2400         {
2401                 .aead = {
2402                         .base = {
2403                                 .cra_name = "authenc(hmac(md5),cbc(des))",
2404                                 .cra_driver_name = "authenc-hmac-md5-"
2405                                                    "cbc-des-caam-qi2",
2406                                 .cra_blocksize = DES_BLOCK_SIZE,
2407                         },
2408                         .setkey = aead_setkey,
2409                         .setauthsize = aead_setauthsize,
2410                         .encrypt = aead_encrypt,
2411                         .decrypt = aead_decrypt,
2412                         .ivsize = DES_BLOCK_SIZE,
2413                         .maxauthsize = MD5_DIGEST_SIZE,
2414                 },
2415                 .caam = {
2416                         .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
2417                         .class2_alg_type = OP_ALG_ALGSEL_MD5 |
2418                                            OP_ALG_AAI_HMAC_PRECOMP,
2419                 },
2420         },
2421         {
2422                 .aead = {
2423                         .base = {
2424                                 .cra_name = "echainiv(authenc(hmac(md5),"
2425                                             "cbc(des)))",
2426                                 .cra_driver_name = "echainiv-authenc-hmac-md5-"
2427                                                    "cbc-des-caam-qi2",
2428                                 .cra_blocksize = DES_BLOCK_SIZE,
2429                         },
2430                         .setkey = aead_setkey,
2431                         .setauthsize = aead_setauthsize,
2432                         .encrypt = aead_encrypt,
2433                         .decrypt = aead_decrypt,
2434                         .ivsize = DES_BLOCK_SIZE,
2435                         .maxauthsize = MD5_DIGEST_SIZE,
2436                 },
2437                 .caam = {
2438                         .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
2439                         .class2_alg_type = OP_ALG_ALGSEL_MD5 |
2440                                            OP_ALG_AAI_HMAC_PRECOMP,
2441                         .geniv = true,
2442                 }
2443         },
2444         {
2445                 .aead = {
2446                         .base = {
2447                                 .cra_name = "authenc(hmac(sha1),cbc(des))",
2448                                 .cra_driver_name = "authenc-hmac-sha1-"
2449                                                    "cbc-des-caam-qi2",
2450                                 .cra_blocksize = DES_BLOCK_SIZE,
2451                         },
2452                         .setkey = aead_setkey,
2453                         .setauthsize = aead_setauthsize,
2454                         .encrypt = aead_encrypt,
2455                         .decrypt = aead_decrypt,
2456                         .ivsize = DES_BLOCK_SIZE,
2457                         .maxauthsize = SHA1_DIGEST_SIZE,
2458                 },
2459                 .caam = {
2460                         .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
2461                         .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
2462                                            OP_ALG_AAI_HMAC_PRECOMP,
2463                 },
2464         },
2465         {
2466                 .aead = {
2467                         .base = {
2468                                 .cra_name = "echainiv(authenc(hmac(sha1),"
2469                                             "cbc(des)))",
2470                                 .cra_driver_name = "echainiv-authenc-"
2471                                                    "hmac-sha1-cbc-des-caam-qi2",
2472                                 .cra_blocksize = DES_BLOCK_SIZE,
2473                         },
2474                         .setkey = aead_setkey,
2475                         .setauthsize = aead_setauthsize,
2476                         .encrypt = aead_encrypt,
2477                         .decrypt = aead_decrypt,
2478                         .ivsize = DES_BLOCK_SIZE,
2479                         .maxauthsize = SHA1_DIGEST_SIZE,
2480                 },
2481                 .caam = {
2482                         .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
2483                         .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
2484                                            OP_ALG_AAI_HMAC_PRECOMP,
2485                         .geniv = true,
2486                 }
2487         },
2488         {
2489                 .aead = {
2490                         .base = {
2491                                 .cra_name = "authenc(hmac(sha224),cbc(des))",
2492                                 .cra_driver_name = "authenc-hmac-sha224-"
2493                                                    "cbc-des-caam-qi2",
2494                                 .cra_blocksize = DES_BLOCK_SIZE,
2495                         },
2496                         .setkey = aead_setkey,
2497                         .setauthsize = aead_setauthsize,
2498                         .encrypt = aead_encrypt,
2499                         .decrypt = aead_decrypt,
2500                         .ivsize = DES_BLOCK_SIZE,
2501                         .maxauthsize = SHA224_DIGEST_SIZE,
2502                 },
2503                 .caam = {
2504                         .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
2505                         .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
2506                                            OP_ALG_AAI_HMAC_PRECOMP,
2507                 },
2508         },
2509         {
2510                 .aead = {
2511                         .base = {
2512                                 .cra_name = "echainiv(authenc(hmac(sha224),"
2513                                             "cbc(des)))",
2514                                 .cra_driver_name = "echainiv-authenc-"
2515                                                    "hmac-sha224-cbc-des-"
2516                                                    "caam-qi2",
2517                                 .cra_blocksize = DES_BLOCK_SIZE,
2518                         },
2519                         .setkey = aead_setkey,
2520                         .setauthsize = aead_setauthsize,
2521                         .encrypt = aead_encrypt,
2522                         .decrypt = aead_decrypt,
2523                         .ivsize = DES_BLOCK_SIZE,
2524                         .maxauthsize = SHA224_DIGEST_SIZE,
2525                 },
2526                 .caam = {
2527                         .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
2528                         .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
2529                                            OP_ALG_AAI_HMAC_PRECOMP,
2530                         .geniv = true,
2531                 }
2532         },
2533         {
2534                 .aead = {
2535                         .base = {
2536                                 .cra_name = "authenc(hmac(sha256),cbc(des))",
2537                                 .cra_driver_name = "authenc-hmac-sha256-"
2538                                                    "cbc-des-caam-qi2",
2539                                 .cra_blocksize = DES_BLOCK_SIZE,
2540                         },
2541                         .setkey = aead_setkey,
2542                         .setauthsize = aead_setauthsize,
2543                         .encrypt = aead_encrypt,
2544                         .decrypt = aead_decrypt,
2545                         .ivsize = DES_BLOCK_SIZE,
2546                         .maxauthsize = SHA256_DIGEST_SIZE,
2547                 },
2548                 .caam = {
2549                         .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
2550                         .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
2551                                            OP_ALG_AAI_HMAC_PRECOMP,
2552                 },
2553         },
2554         {
2555                 .aead = {
2556                         .base = {
2557                                 .cra_name = "echainiv(authenc(hmac(sha256),"
2558                                             "cbc(des)))",
2559                                 .cra_driver_name = "echainiv-authenc-"
2560                                                    "hmac-sha256-cbc-des-"
2561                                                    "caam-qi2",
2562                                 .cra_blocksize = DES_BLOCK_SIZE,
2563                         },
2564                         .setkey = aead_setkey,
2565                         .setauthsize = aead_setauthsize,
2566                         .encrypt = aead_encrypt,
2567                         .decrypt = aead_decrypt,
2568                         .ivsize = DES_BLOCK_SIZE,
2569                         .maxauthsize = SHA256_DIGEST_SIZE,
2570                 },
2571                 .caam = {
2572                         .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
2573                         .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
2574                                            OP_ALG_AAI_HMAC_PRECOMP,
2575                         .geniv = true,
2576                 },
2577         },
2578         {
2579                 .aead = {
2580                         .base = {
2581                                 .cra_name = "authenc(hmac(sha384),cbc(des))",
2582                                 .cra_driver_name = "authenc-hmac-sha384-"
2583                                                    "cbc-des-caam-qi2",
2584                                 .cra_blocksize = DES_BLOCK_SIZE,
2585                         },
2586                         .setkey = aead_setkey,
2587                         .setauthsize = aead_setauthsize,
2588                         .encrypt = aead_encrypt,
2589                         .decrypt = aead_decrypt,
2590                         .ivsize = DES_BLOCK_SIZE,
2591                         .maxauthsize = SHA384_DIGEST_SIZE,
2592                 },
2593                 .caam = {
2594                         .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
2595                         .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
2596                                            OP_ALG_AAI_HMAC_PRECOMP,
2597                 },
2598         },
2599         {
2600                 .aead = {
2601                         .base = {
2602                                 .cra_name = "echainiv(authenc(hmac(sha384),"
2603                                             "cbc(des)))",
2604                                 .cra_driver_name = "echainiv-authenc-"
2605                                                    "hmac-sha384-cbc-des-"
2606                                                    "caam-qi2",
2607                                 .cra_blocksize = DES_BLOCK_SIZE,
2608                         },
2609                         .setkey = aead_setkey,
2610                         .setauthsize = aead_setauthsize,
2611                         .encrypt = aead_encrypt,
2612                         .decrypt = aead_decrypt,
2613                         .ivsize = DES_BLOCK_SIZE,
2614                         .maxauthsize = SHA384_DIGEST_SIZE,
2615                 },
2616                 .caam = {
2617                         .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
2618                         .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
2619                                            OP_ALG_AAI_HMAC_PRECOMP,
2620                         .geniv = true,
2621                 }
2622         },
2623         {
2624                 .aead = {
2625                         .base = {
2626                                 .cra_name = "authenc(hmac(sha512),cbc(des))",
2627                                 .cra_driver_name = "authenc-hmac-sha512-"
2628                                                    "cbc-des-caam-qi2",
2629                                 .cra_blocksize = DES_BLOCK_SIZE,
2630                         },
2631                         .setkey = aead_setkey,
2632                         .setauthsize = aead_setauthsize,
2633                         .encrypt = aead_encrypt,
2634                         .decrypt = aead_decrypt,
2635                         .ivsize = DES_BLOCK_SIZE,
2636                         .maxauthsize = SHA512_DIGEST_SIZE,
2637                 },
2638                 .caam = {
2639                         .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
2640                         .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
2641                                            OP_ALG_AAI_HMAC_PRECOMP,
2642                 }
2643         },
2644         {
2645                 .aead = {
2646                         .base = {
2647                                 .cra_name = "echainiv(authenc(hmac(sha512),"
2648                                             "cbc(des)))",
2649                                 .cra_driver_name = "echainiv-authenc-"
2650                                                    "hmac-sha512-cbc-des-"
2651                                                    "caam-qi2",
2652                                 .cra_blocksize = DES_BLOCK_SIZE,
2653                         },
2654                         .setkey = aead_setkey,
2655                         .setauthsize = aead_setauthsize,
2656                         .encrypt = aead_encrypt,
2657                         .decrypt = aead_decrypt,
2658                         .ivsize = DES_BLOCK_SIZE,
2659                         .maxauthsize = SHA512_DIGEST_SIZE,
2660                 },
2661                 .caam = {
2662                         .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
2663                         .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
2664                                            OP_ALG_AAI_HMAC_PRECOMP,
2665                         .geniv = true,
2666                 }
2667         },
2668         {
2669                 .aead = {
2670                         .base = {
2671                                 .cra_name = "authenc(hmac(md5),"
2672                                             "rfc3686(ctr(aes)))",
2673                                 .cra_driver_name = "authenc-hmac-md5-"
2674                                                    "rfc3686-ctr-aes-caam-qi2",
2675                                 .cra_blocksize = 1,
2676                         },
2677                         .setkey = aead_setkey,
2678                         .setauthsize = aead_setauthsize,
2679                         .encrypt = aead_encrypt,
2680                         .decrypt = aead_decrypt,
2681                         .ivsize = CTR_RFC3686_IV_SIZE,
2682                         .maxauthsize = MD5_DIGEST_SIZE,
2683                 },
2684                 .caam = {
2685                         .class1_alg_type = OP_ALG_ALGSEL_AES |
2686                                            OP_ALG_AAI_CTR_MOD128,
2687                         .class2_alg_type = OP_ALG_ALGSEL_MD5 |
2688                                            OP_ALG_AAI_HMAC_PRECOMP,
2689                         .rfc3686 = true,
2690                 },
2691         },
2692         {
2693                 .aead = {
2694                         .base = {
2695                                 .cra_name = "seqiv(authenc("
2696                                             "hmac(md5),rfc3686(ctr(aes))))",
2697                                 .cra_driver_name = "seqiv-authenc-hmac-md5-"
2698                                                    "rfc3686-ctr-aes-caam-qi2",
2699                                 .cra_blocksize = 1,
2700                         },
2701                         .setkey = aead_setkey,
2702                         .setauthsize = aead_setauthsize,
2703                         .encrypt = aead_encrypt,
2704                         .decrypt = aead_decrypt,
2705                         .ivsize = CTR_RFC3686_IV_SIZE,
2706                         .maxauthsize = MD5_DIGEST_SIZE,
2707                 },
2708                 .caam = {
2709                         .class1_alg_type = OP_ALG_ALGSEL_AES |
2710                                            OP_ALG_AAI_CTR_MOD128,
2711                         .class2_alg_type = OP_ALG_ALGSEL_MD5 |
2712                                            OP_ALG_AAI_HMAC_PRECOMP,
2713                         .rfc3686 = true,
2714                         .geniv = true,
2715                 },
2716         },
2717         {
2718                 .aead = {
2719                         .base = {
2720                                 .cra_name = "authenc(hmac(sha1),"
2721                                             "rfc3686(ctr(aes)))",
2722                                 .cra_driver_name = "authenc-hmac-sha1-"
2723                                                    "rfc3686-ctr-aes-caam-qi2",
2724                                 .cra_blocksize = 1,
2725                         },
2726                         .setkey = aead_setkey,
2727                         .setauthsize = aead_setauthsize,
2728                         .encrypt = aead_encrypt,
2729                         .decrypt = aead_decrypt,
2730                         .ivsize = CTR_RFC3686_IV_SIZE,
2731                         .maxauthsize = SHA1_DIGEST_SIZE,
2732                 },
2733                 .caam = {
2734                         .class1_alg_type = OP_ALG_ALGSEL_AES |
2735                                            OP_ALG_AAI_CTR_MOD128,
2736                         .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
2737                                            OP_ALG_AAI_HMAC_PRECOMP,
2738                         .rfc3686 = true,
2739                 },
2740         },
2741         {
2742                 .aead = {
2743                         .base = {
2744                                 .cra_name = "seqiv(authenc("
2745                                             "hmac(sha1),rfc3686(ctr(aes))))",
2746                                 .cra_driver_name = "seqiv-authenc-hmac-sha1-"
2747                                                    "rfc3686-ctr-aes-caam-qi2",
2748                                 .cra_blocksize = 1,
2749                         },
2750                         .setkey = aead_setkey,
2751                         .setauthsize = aead_setauthsize,
2752                         .encrypt = aead_encrypt,
2753                         .decrypt = aead_decrypt,
2754                         .ivsize = CTR_RFC3686_IV_SIZE,
2755                         .maxauthsize = SHA1_DIGEST_SIZE,
2756                 },
2757                 .caam = {
2758                         .class1_alg_type = OP_ALG_ALGSEL_AES |
2759                                            OP_ALG_AAI_CTR_MOD128,
2760                         .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
2761                                            OP_ALG_AAI_HMAC_PRECOMP,
2762                         .rfc3686 = true,
2763                         .geniv = true,
2764                 },
2765         },
2766         {
2767                 .aead = {
2768                         .base = {
2769                                 .cra_name = "authenc(hmac(sha224),"
2770                                             "rfc3686(ctr(aes)))",
2771                                 .cra_driver_name = "authenc-hmac-sha224-"
2772                                                    "rfc3686-ctr-aes-caam-qi2",
2773                                 .cra_blocksize = 1,
2774                         },
2775                         .setkey = aead_setkey,
2776                         .setauthsize = aead_setauthsize,
2777                         .encrypt = aead_encrypt,
2778                         .decrypt = aead_decrypt,
2779                         .ivsize = CTR_RFC3686_IV_SIZE,
2780                         .maxauthsize = SHA224_DIGEST_SIZE,
2781                 },
2782                 .caam = {
2783                         .class1_alg_type = OP_ALG_ALGSEL_AES |
2784                                            OP_ALG_AAI_CTR_MOD128,
2785                         .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
2786                                            OP_ALG_AAI_HMAC_PRECOMP,
2787                         .rfc3686 = true,
2788                 },
2789         },
2790         {
2791                 .aead = {
2792                         .base = {
2793                                 .cra_name = "seqiv(authenc("
2794                                             "hmac(sha224),rfc3686(ctr(aes))))",
2795                                 .cra_driver_name = "seqiv-authenc-hmac-sha224-"
2796                                                    "rfc3686-ctr-aes-caam-qi2",
2797                                 .cra_blocksize = 1,
2798                         },
2799                         .setkey = aead_setkey,
2800                         .setauthsize = aead_setauthsize,
2801                         .encrypt = aead_encrypt,
2802                         .decrypt = aead_decrypt,
2803                         .ivsize = CTR_RFC3686_IV_SIZE,
2804                         .maxauthsize = SHA224_DIGEST_SIZE,
2805                 },
2806                 .caam = {
2807                         .class1_alg_type = OP_ALG_ALGSEL_AES |
2808                                            OP_ALG_AAI_CTR_MOD128,
2809                         .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
2810                                            OP_ALG_AAI_HMAC_PRECOMP,
2811                         .rfc3686 = true,
2812                         .geniv = true,
2813                 },
2814         },
2815         {
2816                 .aead = {
2817                         .base = {
2818                                 .cra_name = "authenc(hmac(sha256),"
2819                                             "rfc3686(ctr(aes)))",
2820                                 .cra_driver_name = "authenc-hmac-sha256-"
2821                                                    "rfc3686-ctr-aes-caam-qi2",
2822                                 .cra_blocksize = 1,
2823                         },
2824                         .setkey = aead_setkey,
2825                         .setauthsize = aead_setauthsize,
2826                         .encrypt = aead_encrypt,
2827                         .decrypt = aead_decrypt,
2828                         .ivsize = CTR_RFC3686_IV_SIZE,
2829                         .maxauthsize = SHA256_DIGEST_SIZE,
2830                 },
2831                 .caam = {
2832                         .class1_alg_type = OP_ALG_ALGSEL_AES |
2833                                            OP_ALG_AAI_CTR_MOD128,
2834                         .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
2835                                            OP_ALG_AAI_HMAC_PRECOMP,
2836                         .rfc3686 = true,
2837                 },
2838         },
2839         {
2840                 .aead = {
2841                         .base = {
2842                                 .cra_name = "seqiv(authenc(hmac(sha256),"
2843                                             "rfc3686(ctr(aes))))",
2844                                 .cra_driver_name = "seqiv-authenc-hmac-sha256-"
2845                                                    "rfc3686-ctr-aes-caam-qi2",
2846                                 .cra_blocksize = 1,
2847                         },
2848                         .setkey = aead_setkey,
2849                         .setauthsize = aead_setauthsize,
2850                         .encrypt = aead_encrypt,
2851                         .decrypt = aead_decrypt,
2852                         .ivsize = CTR_RFC3686_IV_SIZE,
2853                         .maxauthsize = SHA256_DIGEST_SIZE,
2854                 },
2855                 .caam = {
2856                         .class1_alg_type = OP_ALG_ALGSEL_AES |
2857                                            OP_ALG_AAI_CTR_MOD128,
2858                         .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
2859                                            OP_ALG_AAI_HMAC_PRECOMP,
2860                         .rfc3686 = true,
2861                         .geniv = true,
2862                 },
2863         },
2864         {
2865                 .aead = {
2866                         .base = {
2867                                 .cra_name = "authenc(hmac(sha384),"
2868                                             "rfc3686(ctr(aes)))",
2869                                 .cra_driver_name = "authenc-hmac-sha384-"
2870                                                    "rfc3686-ctr-aes-caam-qi2",
2871                                 .cra_blocksize = 1,
2872                         },
2873                         .setkey = aead_setkey,
2874                         .setauthsize = aead_setauthsize,
2875                         .encrypt = aead_encrypt,
2876                         .decrypt = aead_decrypt,
2877                         .ivsize = CTR_RFC3686_IV_SIZE,
2878                         .maxauthsize = SHA384_DIGEST_SIZE,
2879                 },
2880                 .caam = {
2881                         .class1_alg_type = OP_ALG_ALGSEL_AES |
2882                                            OP_ALG_AAI_CTR_MOD128,
2883                         .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
2884                                            OP_ALG_AAI_HMAC_PRECOMP,
2885                         .rfc3686 = true,
2886                 },
2887         },
2888         {
2889                 .aead = {
2890                         .base = {
2891                                 .cra_name = "seqiv(authenc(hmac(sha384),"
2892                                             "rfc3686(ctr(aes))))",
2893                                 .cra_driver_name = "seqiv-authenc-hmac-sha384-"
2894                                                    "rfc3686-ctr-aes-caam-qi2",
2895                                 .cra_blocksize = 1,
2896                         },
2897                         .setkey = aead_setkey,
2898                         .setauthsize = aead_setauthsize,
2899                         .encrypt = aead_encrypt,
2900                         .decrypt = aead_decrypt,
2901                         .ivsize = CTR_RFC3686_IV_SIZE,
2902                         .maxauthsize = SHA384_DIGEST_SIZE,
2903                 },
2904                 .caam = {
2905                         .class1_alg_type = OP_ALG_ALGSEL_AES |
2906                                            OP_ALG_AAI_CTR_MOD128,
2907                         .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
2908                                            OP_ALG_AAI_HMAC_PRECOMP,
2909                         .rfc3686 = true,
2910                         .geniv = true,
2911                 },
2912         },
2913         {
2914                 .aead = {
2915                         .base = {
2916                                 .cra_name = "rfc7539(chacha20,poly1305)",
2917                                 .cra_driver_name = "rfc7539-chacha20-poly1305-"
2918                                                    "caam-qi2",
2919                                 .cra_blocksize = 1,
2920                         },
2921                         .setkey = chachapoly_setkey,
2922                         .setauthsize = chachapoly_setauthsize,
2923                         .encrypt = aead_encrypt,
2924                         .decrypt = aead_decrypt,
2925                         .ivsize = CHACHAPOLY_IV_SIZE,
2926                         .maxauthsize = POLY1305_DIGEST_SIZE,
2927                 },
2928                 .caam = {
2929                         .class1_alg_type = OP_ALG_ALGSEL_CHACHA20 |
2930                                            OP_ALG_AAI_AEAD,
2931                         .class2_alg_type = OP_ALG_ALGSEL_POLY1305 |
2932                                            OP_ALG_AAI_AEAD,
2933                         .nodkp = true,
2934                 },
2935         },
2936         {
2937                 .aead = {
2938                         .base = {
2939                                 .cra_name = "rfc7539esp(chacha20,poly1305)",
2940                                 .cra_driver_name = "rfc7539esp-chacha20-"
2941                                                    "poly1305-caam-qi2",
2942                                 .cra_blocksize = 1,
2943                         },
2944                         .setkey = chachapoly_setkey,
2945                         .setauthsize = chachapoly_setauthsize,
2946                         .encrypt = aead_encrypt,
2947                         .decrypt = aead_decrypt,
2948                         .ivsize = 8,
2949                         .maxauthsize = POLY1305_DIGEST_SIZE,
2950                 },
2951                 .caam = {
2952                         .class1_alg_type = OP_ALG_ALGSEL_CHACHA20 |
2953                                            OP_ALG_AAI_AEAD,
2954                         .class2_alg_type = OP_ALG_ALGSEL_POLY1305 |
2955                                            OP_ALG_AAI_AEAD,
2956                         .nodkp = true,
2957                 },
2958         },
2959         {
2960                 .aead = {
2961                         .base = {
2962                                 .cra_name = "authenc(hmac(sha512),"
2963                                             "rfc3686(ctr(aes)))",
2964                                 .cra_driver_name = "authenc-hmac-sha512-"
2965                                                    "rfc3686-ctr-aes-caam-qi2",
2966                                 .cra_blocksize = 1,
2967                         },
2968                         .setkey = aead_setkey,
2969                         .setauthsize = aead_setauthsize,
2970                         .encrypt = aead_encrypt,
2971                         .decrypt = aead_decrypt,
2972                         .ivsize = CTR_RFC3686_IV_SIZE,
2973                         .maxauthsize = SHA512_DIGEST_SIZE,
2974                 },
2975                 .caam = {
2976                         .class1_alg_type = OP_ALG_ALGSEL_AES |
2977                                            OP_ALG_AAI_CTR_MOD128,
2978                         .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
2979                                            OP_ALG_AAI_HMAC_PRECOMP,
2980                         .rfc3686 = true,
2981                 },
2982         },
2983         {
2984                 .aead = {
2985                         .base = {
2986                                 .cra_name = "seqiv(authenc(hmac(sha512),"
2987                                             "rfc3686(ctr(aes))))",
2988                                 .cra_driver_name = "seqiv-authenc-hmac-sha512-"
2989                                                    "rfc3686-ctr-aes-caam-qi2",
2990                                 .cra_blocksize = 1,
2991                         },
2992                         .setkey = aead_setkey,
2993                         .setauthsize = aead_setauthsize,
2994                         .encrypt = aead_encrypt,
2995                         .decrypt = aead_decrypt,
2996                         .ivsize = CTR_RFC3686_IV_SIZE,
2997                         .maxauthsize = SHA512_DIGEST_SIZE,
2998                 },
2999                 .caam = {
3000                         .class1_alg_type = OP_ALG_ALGSEL_AES |
3001                                            OP_ALG_AAI_CTR_MOD128,
3002                         .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
3003                                            OP_ALG_AAI_HMAC_PRECOMP,
3004                         .rfc3686 = true,
3005                         .geniv = true,
3006                 },
3007         },
3008 };
3009
3010 static void caam_skcipher_alg_init(struct caam_skcipher_alg *t_alg)
3011 {
3012         struct skcipher_alg *alg = &t_alg->skcipher;
3013
3014         alg->base.cra_module = THIS_MODULE;
3015         alg->base.cra_priority = CAAM_CRA_PRIORITY;
3016         alg->base.cra_ctxsize = sizeof(struct caam_ctx) + crypto_dma_padding();
3017         alg->base.cra_flags |= (CRYPTO_ALG_ASYNC | CRYPTO_ALG_ALLOCATES_MEMORY |
3018                               CRYPTO_ALG_KERN_DRIVER_ONLY);
3019
3020         alg->init = caam_cra_init_skcipher;
3021         alg->exit = caam_cra_exit;
3022 }
3023
3024 static void caam_aead_alg_init(struct caam_aead_alg *t_alg)
3025 {
3026         struct aead_alg *alg = &t_alg->aead;
3027
3028         alg->base.cra_module = THIS_MODULE;
3029         alg->base.cra_priority = CAAM_CRA_PRIORITY;
3030         alg->base.cra_ctxsize = sizeof(struct caam_ctx) + crypto_dma_padding();
3031         alg->base.cra_flags = CRYPTO_ALG_ASYNC | CRYPTO_ALG_ALLOCATES_MEMORY |
3032                               CRYPTO_ALG_KERN_DRIVER_ONLY;
3033
3034         alg->init = caam_cra_init_aead;
3035         alg->exit = caam_cra_exit_aead;
3036 }
3037
3038 /* max hash key is max split key size */
3039 #define CAAM_MAX_HASH_KEY_SIZE          (SHA512_DIGEST_SIZE * 2)
3040
3041 #define CAAM_MAX_HASH_BLOCK_SIZE        SHA512_BLOCK_SIZE
3042
3043 /* caam context sizes for hashes: running digest + 8 */
3044 #define HASH_MSG_LEN                    8
3045 #define MAX_CTX_LEN                     (HASH_MSG_LEN + SHA512_DIGEST_SIZE)
3046
3047 enum hash_optype {
3048         UPDATE = 0,
3049         UPDATE_FIRST,
3050         FINALIZE,
3051         DIGEST,
3052         HASH_NUM_OP
3053 };
3054
3055 /**
3056  * struct caam_hash_ctx - ahash per-session context
3057  * @flc: Flow Contexts array
3058  * @key: authentication key
3059  * @flc_dma: I/O virtual addresses of the Flow Contexts
3060  * @dev: dpseci device
3061  * @ctx_len: size of Context Register
3062  * @adata: hashing algorithm details
3063  */
3064 struct caam_hash_ctx {
3065         struct caam_flc flc[HASH_NUM_OP];
3066         u8 key[CAAM_MAX_HASH_BLOCK_SIZE] ____cacheline_aligned;
3067         dma_addr_t flc_dma[HASH_NUM_OP];
3068         struct device *dev;
3069         int ctx_len;
3070         struct alginfo adata;
3071 };
3072
3073 /* ahash state */
3074 struct caam_hash_state {
3075         struct caam_request caam_req;
3076         dma_addr_t buf_dma;
3077         dma_addr_t ctx_dma;
3078         int ctx_dma_len;
3079         u8 buf[CAAM_MAX_HASH_BLOCK_SIZE] ____cacheline_aligned;
3080         int buflen;
3081         int next_buflen;
3082         u8 caam_ctx[MAX_CTX_LEN] ____cacheline_aligned;
3083         int (*update)(struct ahash_request *req);
3084         int (*final)(struct ahash_request *req);
3085         int (*finup)(struct ahash_request *req);
3086 };
3087
3088 struct caam_export_state {
3089         u8 buf[CAAM_MAX_HASH_BLOCK_SIZE];
3090         u8 caam_ctx[MAX_CTX_LEN];
3091         int buflen;
3092         int (*update)(struct ahash_request *req);
3093         int (*final)(struct ahash_request *req);
3094         int (*finup)(struct ahash_request *req);
3095 };
3096
3097 /* Map current buffer in state (if length > 0) and put it in link table */
3098 static inline int buf_map_to_qm_sg(struct device *dev,
3099                                    struct dpaa2_sg_entry *qm_sg,
3100                                    struct caam_hash_state *state)
3101 {
3102         int buflen = state->buflen;
3103
3104         if (!buflen)
3105                 return 0;
3106
3107         state->buf_dma = dma_map_single(dev, state->buf, buflen,
3108                                         DMA_TO_DEVICE);
3109         if (dma_mapping_error(dev, state->buf_dma)) {
3110                 dev_err(dev, "unable to map buf\n");
3111                 state->buf_dma = 0;
3112                 return -ENOMEM;
3113         }
3114
3115         dma_to_qm_sg_one(qm_sg, state->buf_dma, buflen, 0);
3116
3117         return 0;
3118 }
3119
3120 /* Map state->caam_ctx, and add it to link table */
3121 static inline int ctx_map_to_qm_sg(struct device *dev,
3122                                    struct caam_hash_state *state, int ctx_len,
3123                                    struct dpaa2_sg_entry *qm_sg, u32 flag)
3124 {
3125         state->ctx_dma_len = ctx_len;
3126         state->ctx_dma = dma_map_single(dev, state->caam_ctx, ctx_len, flag);
3127         if (dma_mapping_error(dev, state->ctx_dma)) {
3128                 dev_err(dev, "unable to map ctx\n");
3129                 state->ctx_dma = 0;
3130                 return -ENOMEM;
3131         }
3132
3133         dma_to_qm_sg_one(qm_sg, state->ctx_dma, ctx_len, 0);
3134
3135         return 0;
3136 }
3137
3138 static int ahash_set_sh_desc(struct crypto_ahash *ahash)
3139 {
3140         struct caam_hash_ctx *ctx = crypto_ahash_ctx_dma(ahash);
3141         int digestsize = crypto_ahash_digestsize(ahash);
3142         struct dpaa2_caam_priv *priv = dev_get_drvdata(ctx->dev);
3143         struct caam_flc *flc;
3144         u32 *desc;
3145
3146         /* ahash_update shared descriptor */
3147         flc = &ctx->flc[UPDATE];
3148         desc = flc->sh_desc;
3149         cnstr_shdsc_ahash(desc, &ctx->adata, OP_ALG_AS_UPDATE, ctx->ctx_len,
3150                           ctx->ctx_len, true, priv->sec_attr.era);
3151         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
3152         dma_sync_single_for_device(ctx->dev, ctx->flc_dma[UPDATE],
3153                                    desc_bytes(desc), DMA_BIDIRECTIONAL);
3154         print_hex_dump_debug("ahash update shdesc@" __stringify(__LINE__)": ",
3155                              DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc),
3156                              1);
3157
3158         /* ahash_update_first shared descriptor */
3159         flc = &ctx->flc[UPDATE_FIRST];
3160         desc = flc->sh_desc;
3161         cnstr_shdsc_ahash(desc, &ctx->adata, OP_ALG_AS_INIT, ctx->ctx_len,
3162                           ctx->ctx_len, false, priv->sec_attr.era);
3163         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
3164         dma_sync_single_for_device(ctx->dev, ctx->flc_dma[UPDATE_FIRST],
3165                                    desc_bytes(desc), DMA_BIDIRECTIONAL);
3166         print_hex_dump_debug("ahash update first shdesc@" __stringify(__LINE__)": ",
3167                              DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc),
3168                              1);
3169
3170         /* ahash_final shared descriptor */
3171         flc = &ctx->flc[FINALIZE];
3172         desc = flc->sh_desc;
3173         cnstr_shdsc_ahash(desc, &ctx->adata, OP_ALG_AS_FINALIZE, digestsize,
3174                           ctx->ctx_len, true, priv->sec_attr.era);
3175         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
3176         dma_sync_single_for_device(ctx->dev, ctx->flc_dma[FINALIZE],
3177                                    desc_bytes(desc), DMA_BIDIRECTIONAL);
3178         print_hex_dump_debug("ahash final shdesc@" __stringify(__LINE__)": ",
3179                              DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc),
3180                              1);
3181
3182         /* ahash_digest shared descriptor */
3183         flc = &ctx->flc[DIGEST];
3184         desc = flc->sh_desc;
3185         cnstr_shdsc_ahash(desc, &ctx->adata, OP_ALG_AS_INITFINAL, digestsize,
3186                           ctx->ctx_len, false, priv->sec_attr.era);
3187         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
3188         dma_sync_single_for_device(ctx->dev, ctx->flc_dma[DIGEST],
3189                                    desc_bytes(desc), DMA_BIDIRECTIONAL);
3190         print_hex_dump_debug("ahash digest shdesc@" __stringify(__LINE__)": ",
3191                              DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc),
3192                              1);
3193
3194         return 0;
3195 }
3196
3197 struct split_key_sh_result {
3198         struct completion completion;
3199         int err;
3200         struct device *dev;
3201 };
3202
3203 static void split_key_sh_done(void *cbk_ctx, u32 err)
3204 {
3205         struct split_key_sh_result *res = cbk_ctx;
3206
3207         dev_dbg(res->dev, "%s %d: err 0x%x\n", __func__, __LINE__, err);
3208
3209         res->err = err ? caam_qi2_strstatus(res->dev, err) : 0;
3210         complete(&res->completion);
3211 }
3212
3213 /* Digest hash size if it is too large */
3214 static int hash_digest_key(struct caam_hash_ctx *ctx, u32 *keylen, u8 *key,
3215                            u32 digestsize)
3216 {
3217         struct caam_request *req_ctx;
3218         u32 *desc;
3219         struct split_key_sh_result result;
3220         dma_addr_t key_dma;
3221         struct caam_flc *flc;
3222         dma_addr_t flc_dma;
3223         int ret = -ENOMEM;
3224         struct dpaa2_fl_entry *in_fle, *out_fle;
3225
3226         req_ctx = kzalloc(sizeof(*req_ctx), GFP_KERNEL);
3227         if (!req_ctx)
3228                 return -ENOMEM;
3229
3230         in_fle = &req_ctx->fd_flt[1];
3231         out_fle = &req_ctx->fd_flt[0];
3232
3233         flc = kzalloc(sizeof(*flc), GFP_KERNEL);
3234         if (!flc)
3235                 goto err_flc;
3236
3237         key_dma = dma_map_single(ctx->dev, key, *keylen, DMA_BIDIRECTIONAL);
3238         if (dma_mapping_error(ctx->dev, key_dma)) {
3239                 dev_err(ctx->dev, "unable to map key memory\n");
3240                 goto err_key_dma;
3241         }
3242
3243         desc = flc->sh_desc;
3244
3245         init_sh_desc(desc, 0);
3246
3247         /* descriptor to perform unkeyed hash on key_in */
3248         append_operation(desc, ctx->adata.algtype | OP_ALG_ENCRYPT |
3249                          OP_ALG_AS_INITFINAL);
3250         append_seq_fifo_load(desc, *keylen, FIFOLD_CLASS_CLASS2 |
3251                              FIFOLD_TYPE_LAST2 | FIFOLD_TYPE_MSG);
3252         append_seq_store(desc, digestsize, LDST_CLASS_2_CCB |
3253                          LDST_SRCDST_BYTE_CONTEXT);
3254
3255         flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
3256         flc_dma = dma_map_single(ctx->dev, flc, sizeof(flc->flc) +
3257                                  desc_bytes(desc), DMA_TO_DEVICE);
3258         if (dma_mapping_error(ctx->dev, flc_dma)) {
3259                 dev_err(ctx->dev, "unable to map shared descriptor\n");
3260                 goto err_flc_dma;
3261         }
3262
3263         dpaa2_fl_set_final(in_fle, true);
3264         dpaa2_fl_set_format(in_fle, dpaa2_fl_single);
3265         dpaa2_fl_set_addr(in_fle, key_dma);
3266         dpaa2_fl_set_len(in_fle, *keylen);
3267         dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
3268         dpaa2_fl_set_addr(out_fle, key_dma);
3269         dpaa2_fl_set_len(out_fle, digestsize);
3270
3271         print_hex_dump_debug("key_in@" __stringify(__LINE__)": ",
3272                              DUMP_PREFIX_ADDRESS, 16, 4, key, *keylen, 1);
3273         print_hex_dump_debug("shdesc@" __stringify(__LINE__)": ",
3274                              DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc),
3275                              1);
3276
3277         result.err = 0;
3278         init_completion(&result.completion);
3279         result.dev = ctx->dev;
3280
3281         req_ctx->flc = flc;
3282         req_ctx->flc_dma = flc_dma;
3283         req_ctx->cbk = split_key_sh_done;
3284         req_ctx->ctx = &result;
3285
3286         ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
3287         if (ret == -EINPROGRESS) {
3288                 /* in progress */
3289                 wait_for_completion(&result.completion);
3290                 ret = result.err;
3291                 print_hex_dump_debug("digested key@" __stringify(__LINE__)": ",
3292                                      DUMP_PREFIX_ADDRESS, 16, 4, key,
3293                                      digestsize, 1);
3294         }
3295
3296         dma_unmap_single(ctx->dev, flc_dma, sizeof(flc->flc) + desc_bytes(desc),
3297                          DMA_TO_DEVICE);
3298 err_flc_dma:
3299         dma_unmap_single(ctx->dev, key_dma, *keylen, DMA_BIDIRECTIONAL);
3300 err_key_dma:
3301         kfree(flc);
3302 err_flc:
3303         kfree(req_ctx);
3304
3305         *keylen = digestsize;
3306
3307         return ret;
3308 }
3309
3310 static int ahash_setkey(struct crypto_ahash *ahash, const u8 *key,
3311                         unsigned int keylen)
3312 {
3313         struct caam_hash_ctx *ctx = crypto_ahash_ctx_dma(ahash);
3314         unsigned int blocksize = crypto_tfm_alg_blocksize(&ahash->base);
3315         unsigned int digestsize = crypto_ahash_digestsize(ahash);
3316         int ret;
3317         u8 *hashed_key = NULL;
3318
3319         dev_dbg(ctx->dev, "keylen %d blocksize %d\n", keylen, blocksize);
3320
3321         if (keylen > blocksize) {
3322                 unsigned int aligned_len =
3323                         ALIGN(keylen, dma_get_cache_alignment());
3324
3325                 if (aligned_len < keylen)
3326                         return -EOVERFLOW;
3327
3328                 hashed_key = kmemdup(key, aligned_len, GFP_KERNEL);
3329                 if (!hashed_key)
3330                         return -ENOMEM;
3331                 ret = hash_digest_key(ctx, &keylen, hashed_key, digestsize);
3332                 if (ret)
3333                         goto bad_free_key;
3334                 key = hashed_key;
3335         }
3336
3337         ctx->adata.keylen = keylen;
3338         ctx->adata.keylen_pad = split_key_len(ctx->adata.algtype &
3339                                               OP_ALG_ALGSEL_MASK);
3340         if (ctx->adata.keylen_pad > CAAM_MAX_HASH_KEY_SIZE)
3341                 goto bad_free_key;
3342
3343         ctx->adata.key_virt = key;
3344         ctx->adata.key_inline = true;
3345
3346         /*
3347          * In case |user key| > |derived key|, using DKP<imm,imm> would result
3348          * in invalid opcodes (last bytes of user key) in the resulting
3349          * descriptor. Use DKP<ptr,imm> instead => both virtual and dma key
3350          * addresses are needed.
3351          */
3352         if (keylen > ctx->adata.keylen_pad) {
3353                 memcpy(ctx->key, key, keylen);
3354                 dma_sync_single_for_device(ctx->dev, ctx->adata.key_dma,
3355                                            ctx->adata.keylen_pad,
3356                                            DMA_TO_DEVICE);
3357         }
3358
3359         ret = ahash_set_sh_desc(ahash);
3360         kfree(hashed_key);
3361         return ret;
3362 bad_free_key:
3363         kfree(hashed_key);
3364         return -EINVAL;
3365 }
3366
3367 static inline void ahash_unmap(struct device *dev, struct ahash_edesc *edesc,
3368                                struct ahash_request *req)
3369 {
3370         struct caam_hash_state *state = ahash_request_ctx_dma(req);
3371
3372         if (edesc->src_nents)
3373                 dma_unmap_sg(dev, req->src, edesc->src_nents, DMA_TO_DEVICE);
3374
3375         if (edesc->qm_sg_bytes)
3376                 dma_unmap_single(dev, edesc->qm_sg_dma, edesc->qm_sg_bytes,
3377                                  DMA_TO_DEVICE);
3378
3379         if (state->buf_dma) {
3380                 dma_unmap_single(dev, state->buf_dma, state->buflen,
3381                                  DMA_TO_DEVICE);
3382                 state->buf_dma = 0;
3383         }
3384 }
3385
3386 static inline void ahash_unmap_ctx(struct device *dev,
3387                                    struct ahash_edesc *edesc,
3388                                    struct ahash_request *req, u32 flag)
3389 {
3390         struct caam_hash_state *state = ahash_request_ctx_dma(req);
3391
3392         if (state->ctx_dma) {
3393                 dma_unmap_single(dev, state->ctx_dma, state->ctx_dma_len, flag);
3394                 state->ctx_dma = 0;
3395         }
3396         ahash_unmap(dev, edesc, req);
3397 }
3398
3399 static void ahash_done(void *cbk_ctx, u32 status)
3400 {
3401         struct crypto_async_request *areq = cbk_ctx;
3402         struct ahash_request *req = ahash_request_cast(areq);
3403         struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
3404         struct caam_hash_state *state = ahash_request_ctx_dma(req);
3405         struct ahash_edesc *edesc = state->caam_req.edesc;
3406         struct caam_hash_ctx *ctx = crypto_ahash_ctx_dma(ahash);
3407         int digestsize = crypto_ahash_digestsize(ahash);
3408         int ecode = 0;
3409
3410         dev_dbg(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
3411
3412         if (unlikely(status))
3413                 ecode = caam_qi2_strstatus(ctx->dev, status);
3414
3415         ahash_unmap_ctx(ctx->dev, edesc, req, DMA_FROM_DEVICE);
3416         memcpy(req->result, state->caam_ctx, digestsize);
3417         qi_cache_free(edesc);
3418
3419         print_hex_dump_debug("ctx@" __stringify(__LINE__)": ",
3420                              DUMP_PREFIX_ADDRESS, 16, 4, state->caam_ctx,
3421                              ctx->ctx_len, 1);
3422
3423         ahash_request_complete(req, ecode);
3424 }
3425
3426 static void ahash_done_bi(void *cbk_ctx, u32 status)
3427 {
3428         struct crypto_async_request *areq = cbk_ctx;
3429         struct ahash_request *req = ahash_request_cast(areq);
3430         struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
3431         struct caam_hash_state *state = ahash_request_ctx_dma(req);
3432         struct ahash_edesc *edesc = state->caam_req.edesc;
3433         struct caam_hash_ctx *ctx = crypto_ahash_ctx_dma(ahash);
3434         int ecode = 0;
3435
3436         dev_dbg(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
3437
3438         if (unlikely(status))
3439                 ecode = caam_qi2_strstatus(ctx->dev, status);
3440
3441         ahash_unmap_ctx(ctx->dev, edesc, req, DMA_BIDIRECTIONAL);
3442         qi_cache_free(edesc);
3443
3444         scatterwalk_map_and_copy(state->buf, req->src,
3445                                  req->nbytes - state->next_buflen,
3446                                  state->next_buflen, 0);
3447         state->buflen = state->next_buflen;
3448
3449         print_hex_dump_debug("buf@" __stringify(__LINE__)": ",
3450                              DUMP_PREFIX_ADDRESS, 16, 4, state->buf,
3451                              state->buflen, 1);
3452
3453         print_hex_dump_debug("ctx@" __stringify(__LINE__)": ",
3454                              DUMP_PREFIX_ADDRESS, 16, 4, state->caam_ctx,
3455                              ctx->ctx_len, 1);
3456         if (req->result)
3457                 print_hex_dump_debug("result@" __stringify(__LINE__)": ",
3458                                      DUMP_PREFIX_ADDRESS, 16, 4, req->result,
3459                                      crypto_ahash_digestsize(ahash), 1);
3460
3461         ahash_request_complete(req, ecode);
3462 }
3463
3464 static void ahash_done_ctx_src(void *cbk_ctx, u32 status)
3465 {
3466         struct crypto_async_request *areq = cbk_ctx;
3467         struct ahash_request *req = ahash_request_cast(areq);
3468         struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
3469         struct caam_hash_state *state = ahash_request_ctx_dma(req);
3470         struct ahash_edesc *edesc = state->caam_req.edesc;
3471         struct caam_hash_ctx *ctx = crypto_ahash_ctx_dma(ahash);
3472         int digestsize = crypto_ahash_digestsize(ahash);
3473         int ecode = 0;
3474
3475         dev_dbg(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
3476
3477         if (unlikely(status))
3478                 ecode = caam_qi2_strstatus(ctx->dev, status);
3479
3480         ahash_unmap_ctx(ctx->dev, edesc, req, DMA_BIDIRECTIONAL);
3481         memcpy(req->result, state->caam_ctx, digestsize);
3482         qi_cache_free(edesc);
3483
3484         print_hex_dump_debug("ctx@" __stringify(__LINE__)": ",
3485                              DUMP_PREFIX_ADDRESS, 16, 4, state->caam_ctx,
3486                              ctx->ctx_len, 1);
3487
3488         ahash_request_complete(req, ecode);
3489 }
3490
3491 static void ahash_done_ctx_dst(void *cbk_ctx, u32 status)
3492 {
3493         struct crypto_async_request *areq = cbk_ctx;
3494         struct ahash_request *req = ahash_request_cast(areq);
3495         struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
3496         struct caam_hash_state *state = ahash_request_ctx_dma(req);
3497         struct ahash_edesc *edesc = state->caam_req.edesc;
3498         struct caam_hash_ctx *ctx = crypto_ahash_ctx_dma(ahash);
3499         int ecode = 0;
3500
3501         dev_dbg(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
3502
3503         if (unlikely(status))
3504                 ecode = caam_qi2_strstatus(ctx->dev, status);
3505
3506         ahash_unmap_ctx(ctx->dev, edesc, req, DMA_FROM_DEVICE);
3507         qi_cache_free(edesc);
3508
3509         scatterwalk_map_and_copy(state->buf, req->src,
3510                                  req->nbytes - state->next_buflen,
3511                                  state->next_buflen, 0);
3512         state->buflen = state->next_buflen;
3513
3514         print_hex_dump_debug("buf@" __stringify(__LINE__)": ",
3515                              DUMP_PREFIX_ADDRESS, 16, 4, state->buf,
3516                              state->buflen, 1);
3517
3518         print_hex_dump_debug("ctx@" __stringify(__LINE__)": ",
3519                              DUMP_PREFIX_ADDRESS, 16, 4, state->caam_ctx,
3520                              ctx->ctx_len, 1);
3521         if (req->result)
3522                 print_hex_dump_debug("result@" __stringify(__LINE__)": ",
3523                                      DUMP_PREFIX_ADDRESS, 16, 4, req->result,
3524                                      crypto_ahash_digestsize(ahash), 1);
3525
3526         ahash_request_complete(req, ecode);
3527 }
3528
3529 static int ahash_update_ctx(struct ahash_request *req)
3530 {
3531         struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
3532         struct caam_hash_ctx *ctx = crypto_ahash_ctx_dma(ahash);
3533         struct caam_hash_state *state = ahash_request_ctx_dma(req);
3534         struct caam_request *req_ctx = &state->caam_req;
3535         struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
3536         struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
3537         gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
3538                       GFP_KERNEL : GFP_ATOMIC;
3539         u8 *buf = state->buf;
3540         int *buflen = &state->buflen;
3541         int *next_buflen = &state->next_buflen;
3542         int in_len = *buflen + req->nbytes, to_hash;
3543         int src_nents, mapped_nents, qm_sg_bytes, qm_sg_src_index;
3544         struct ahash_edesc *edesc;
3545         int ret = 0;
3546
3547         *next_buflen = in_len & (crypto_tfm_alg_blocksize(&ahash->base) - 1);
3548         to_hash = in_len - *next_buflen;
3549
3550         if (to_hash) {
3551                 struct dpaa2_sg_entry *sg_table;
3552                 int src_len = req->nbytes - *next_buflen;
3553
3554                 src_nents = sg_nents_for_len(req->src, src_len);
3555                 if (src_nents < 0) {
3556                         dev_err(ctx->dev, "Invalid number of src SG.\n");
3557                         return src_nents;
3558                 }
3559
3560                 if (src_nents) {
3561                         mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
3562                                                   DMA_TO_DEVICE);
3563                         if (!mapped_nents) {
3564                                 dev_err(ctx->dev, "unable to DMA map source\n");
3565                                 return -ENOMEM;
3566                         }
3567                 } else {
3568                         mapped_nents = 0;
3569                 }
3570
3571                 /* allocate space for base edesc and link tables */
3572                 edesc = qi_cache_zalloc(flags);
3573                 if (!edesc) {
3574                         dma_unmap_sg(ctx->dev, req->src, src_nents,
3575                                      DMA_TO_DEVICE);
3576                         return -ENOMEM;
3577                 }
3578
3579                 edesc->src_nents = src_nents;
3580                 qm_sg_src_index = 1 + (*buflen ? 1 : 0);
3581                 qm_sg_bytes = pad_sg_nents(qm_sg_src_index + mapped_nents) *
3582                               sizeof(*sg_table);
3583                 sg_table = &edesc->sgt[0];
3584
3585                 ret = ctx_map_to_qm_sg(ctx->dev, state, ctx->ctx_len, sg_table,
3586                                        DMA_BIDIRECTIONAL);
3587                 if (ret)
3588                         goto unmap_ctx;
3589
3590                 ret = buf_map_to_qm_sg(ctx->dev, sg_table + 1, state);
3591                 if (ret)
3592                         goto unmap_ctx;
3593
3594                 if (mapped_nents) {
3595                         sg_to_qm_sg_last(req->src, src_len,
3596                                          sg_table + qm_sg_src_index, 0);
3597                 } else {
3598                         dpaa2_sg_set_final(sg_table + qm_sg_src_index - 1,
3599                                            true);
3600                 }
3601
3602                 edesc->qm_sg_dma = dma_map_single(ctx->dev, sg_table,
3603                                                   qm_sg_bytes, DMA_TO_DEVICE);
3604                 if (dma_mapping_error(ctx->dev, edesc->qm_sg_dma)) {
3605                         dev_err(ctx->dev, "unable to map S/G table\n");
3606                         ret = -ENOMEM;
3607                         goto unmap_ctx;
3608                 }
3609                 edesc->qm_sg_bytes = qm_sg_bytes;
3610
3611                 memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
3612                 dpaa2_fl_set_final(in_fle, true);
3613                 dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
3614                 dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
3615                 dpaa2_fl_set_len(in_fle, ctx->ctx_len + to_hash);
3616                 dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
3617                 dpaa2_fl_set_addr(out_fle, state->ctx_dma);
3618                 dpaa2_fl_set_len(out_fle, ctx->ctx_len);
3619
3620                 req_ctx->flc = &ctx->flc[UPDATE];
3621                 req_ctx->flc_dma = ctx->flc_dma[UPDATE];
3622                 req_ctx->cbk = ahash_done_bi;
3623                 req_ctx->ctx = &req->base;
3624                 req_ctx->edesc = edesc;
3625
3626                 ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
3627                 if (ret != -EINPROGRESS &&
3628                     !(ret == -EBUSY &&
3629                       req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG))
3630                         goto unmap_ctx;
3631         } else if (*next_buflen) {
3632                 scatterwalk_map_and_copy(buf + *buflen, req->src, 0,
3633                                          req->nbytes, 0);
3634                 *buflen = *next_buflen;
3635
3636                 print_hex_dump_debug("buf@" __stringify(__LINE__)": ",
3637                                      DUMP_PREFIX_ADDRESS, 16, 4, buf,
3638                                      *buflen, 1);
3639         }
3640
3641         return ret;
3642 unmap_ctx:
3643         ahash_unmap_ctx(ctx->dev, edesc, req, DMA_BIDIRECTIONAL);
3644         qi_cache_free(edesc);
3645         return ret;
3646 }
3647
3648 static int ahash_final_ctx(struct ahash_request *req)
3649 {
3650         struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
3651         struct caam_hash_ctx *ctx = crypto_ahash_ctx_dma(ahash);
3652         struct caam_hash_state *state = ahash_request_ctx_dma(req);
3653         struct caam_request *req_ctx = &state->caam_req;
3654         struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
3655         struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
3656         gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
3657                       GFP_KERNEL : GFP_ATOMIC;
3658         int buflen = state->buflen;
3659         int qm_sg_bytes;
3660         int digestsize = crypto_ahash_digestsize(ahash);
3661         struct ahash_edesc *edesc;
3662         struct dpaa2_sg_entry *sg_table;
3663         int ret;
3664
3665         /* allocate space for base edesc and link tables */
3666         edesc = qi_cache_zalloc(flags);
3667         if (!edesc)
3668                 return -ENOMEM;
3669
3670         qm_sg_bytes = pad_sg_nents(1 + (buflen ? 1 : 0)) * sizeof(*sg_table);
3671         sg_table = &edesc->sgt[0];
3672
3673         ret = ctx_map_to_qm_sg(ctx->dev, state, ctx->ctx_len, sg_table,
3674                                DMA_BIDIRECTIONAL);
3675         if (ret)
3676                 goto unmap_ctx;
3677
3678         ret = buf_map_to_qm_sg(ctx->dev, sg_table + 1, state);
3679         if (ret)
3680                 goto unmap_ctx;
3681
3682         dpaa2_sg_set_final(sg_table + (buflen ? 1 : 0), true);
3683
3684         edesc->qm_sg_dma = dma_map_single(ctx->dev, sg_table, qm_sg_bytes,
3685                                           DMA_TO_DEVICE);
3686         if (dma_mapping_error(ctx->dev, edesc->qm_sg_dma)) {
3687                 dev_err(ctx->dev, "unable to map S/G table\n");
3688                 ret = -ENOMEM;
3689                 goto unmap_ctx;
3690         }
3691         edesc->qm_sg_bytes = qm_sg_bytes;
3692
3693         memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
3694         dpaa2_fl_set_final(in_fle, true);
3695         dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
3696         dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
3697         dpaa2_fl_set_len(in_fle, ctx->ctx_len + buflen);
3698         dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
3699         dpaa2_fl_set_addr(out_fle, state->ctx_dma);
3700         dpaa2_fl_set_len(out_fle, digestsize);
3701
3702         req_ctx->flc = &ctx->flc[FINALIZE];
3703         req_ctx->flc_dma = ctx->flc_dma[FINALIZE];
3704         req_ctx->cbk = ahash_done_ctx_src;
3705         req_ctx->ctx = &req->base;
3706         req_ctx->edesc = edesc;
3707
3708         ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
3709         if (ret == -EINPROGRESS ||
3710             (ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG))
3711                 return ret;
3712
3713 unmap_ctx:
3714         ahash_unmap_ctx(ctx->dev, edesc, req, DMA_BIDIRECTIONAL);
3715         qi_cache_free(edesc);
3716         return ret;
3717 }
3718
3719 static int ahash_finup_ctx(struct ahash_request *req)
3720 {
3721         struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
3722         struct caam_hash_ctx *ctx = crypto_ahash_ctx_dma(ahash);
3723         struct caam_hash_state *state = ahash_request_ctx_dma(req);
3724         struct caam_request *req_ctx = &state->caam_req;
3725         struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
3726         struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
3727         gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
3728                       GFP_KERNEL : GFP_ATOMIC;
3729         int buflen = state->buflen;
3730         int qm_sg_bytes, qm_sg_src_index;
3731         int src_nents, mapped_nents;
3732         int digestsize = crypto_ahash_digestsize(ahash);
3733         struct ahash_edesc *edesc;
3734         struct dpaa2_sg_entry *sg_table;
3735         int ret;
3736
3737         src_nents = sg_nents_for_len(req->src, req->nbytes);
3738         if (src_nents < 0) {
3739                 dev_err(ctx->dev, "Invalid number of src SG.\n");
3740                 return src_nents;
3741         }
3742
3743         if (src_nents) {
3744                 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
3745                                           DMA_TO_DEVICE);
3746                 if (!mapped_nents) {
3747                         dev_err(ctx->dev, "unable to DMA map source\n");
3748                         return -ENOMEM;
3749                 }
3750         } else {
3751                 mapped_nents = 0;
3752         }
3753
3754         /* allocate space for base edesc and link tables */
3755         edesc = qi_cache_zalloc(flags);
3756         if (!edesc) {
3757                 dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE);
3758                 return -ENOMEM;
3759         }
3760
3761         edesc->src_nents = src_nents;
3762         qm_sg_src_index = 1 + (buflen ? 1 : 0);
3763         qm_sg_bytes = pad_sg_nents(qm_sg_src_index + mapped_nents) *
3764                       sizeof(*sg_table);
3765         sg_table = &edesc->sgt[0];
3766
3767         ret = ctx_map_to_qm_sg(ctx->dev, state, ctx->ctx_len, sg_table,
3768                                DMA_BIDIRECTIONAL);
3769         if (ret)
3770                 goto unmap_ctx;
3771
3772         ret = buf_map_to_qm_sg(ctx->dev, sg_table + 1, state);
3773         if (ret)
3774                 goto unmap_ctx;
3775
3776         sg_to_qm_sg_last(req->src, req->nbytes, sg_table + qm_sg_src_index, 0);
3777
3778         edesc->qm_sg_dma = dma_map_single(ctx->dev, sg_table, qm_sg_bytes,
3779                                           DMA_TO_DEVICE);
3780         if (dma_mapping_error(ctx->dev, edesc->qm_sg_dma)) {
3781                 dev_err(ctx->dev, "unable to map S/G table\n");
3782                 ret = -ENOMEM;
3783                 goto unmap_ctx;
3784         }
3785         edesc->qm_sg_bytes = qm_sg_bytes;
3786
3787         memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
3788         dpaa2_fl_set_final(in_fle, true);
3789         dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
3790         dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
3791         dpaa2_fl_set_len(in_fle, ctx->ctx_len + buflen + req->nbytes);
3792         dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
3793         dpaa2_fl_set_addr(out_fle, state->ctx_dma);
3794         dpaa2_fl_set_len(out_fle, digestsize);
3795
3796         req_ctx->flc = &ctx->flc[FINALIZE];
3797         req_ctx->flc_dma = ctx->flc_dma[FINALIZE];
3798         req_ctx->cbk = ahash_done_ctx_src;
3799         req_ctx->ctx = &req->base;
3800         req_ctx->edesc = edesc;
3801
3802         ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
3803         if (ret == -EINPROGRESS ||
3804             (ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG))
3805                 return ret;
3806
3807 unmap_ctx:
3808         ahash_unmap_ctx(ctx->dev, edesc, req, DMA_BIDIRECTIONAL);
3809         qi_cache_free(edesc);
3810         return ret;
3811 }
3812
3813 static int ahash_digest(struct ahash_request *req)
3814 {
3815         struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
3816         struct caam_hash_ctx *ctx = crypto_ahash_ctx_dma(ahash);
3817         struct caam_hash_state *state = ahash_request_ctx_dma(req);
3818         struct caam_request *req_ctx = &state->caam_req;
3819         struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
3820         struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
3821         gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
3822                       GFP_KERNEL : GFP_ATOMIC;
3823         int digestsize = crypto_ahash_digestsize(ahash);
3824         int src_nents, mapped_nents;
3825         struct ahash_edesc *edesc;
3826         int ret = -ENOMEM;
3827
3828         state->buf_dma = 0;
3829
3830         src_nents = sg_nents_for_len(req->src, req->nbytes);
3831         if (src_nents < 0) {
3832                 dev_err(ctx->dev, "Invalid number of src SG.\n");
3833                 return src_nents;
3834         }
3835
3836         if (src_nents) {
3837                 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
3838                                           DMA_TO_DEVICE);
3839                 if (!mapped_nents) {
3840                         dev_err(ctx->dev, "unable to map source for DMA\n");
3841                         return ret;
3842                 }
3843         } else {
3844                 mapped_nents = 0;
3845         }
3846
3847         /* allocate space for base edesc and link tables */
3848         edesc = qi_cache_zalloc(flags);
3849         if (!edesc) {
3850                 dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE);
3851                 return ret;
3852         }
3853
3854         edesc->src_nents = src_nents;
3855         memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
3856
3857         if (mapped_nents > 1) {
3858                 int qm_sg_bytes;
3859                 struct dpaa2_sg_entry *sg_table = &edesc->sgt[0];
3860
3861                 qm_sg_bytes = pad_sg_nents(mapped_nents) * sizeof(*sg_table);
3862                 sg_to_qm_sg_last(req->src, req->nbytes, sg_table, 0);
3863                 edesc->qm_sg_dma = dma_map_single(ctx->dev, sg_table,
3864                                                   qm_sg_bytes, DMA_TO_DEVICE);
3865                 if (dma_mapping_error(ctx->dev, edesc->qm_sg_dma)) {
3866                         dev_err(ctx->dev, "unable to map S/G table\n");
3867                         goto unmap;
3868                 }
3869                 edesc->qm_sg_bytes = qm_sg_bytes;
3870                 dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
3871                 dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
3872         } else {
3873                 dpaa2_fl_set_format(in_fle, dpaa2_fl_single);
3874                 dpaa2_fl_set_addr(in_fle, sg_dma_address(req->src));
3875         }
3876
3877         state->ctx_dma_len = digestsize;
3878         state->ctx_dma = dma_map_single(ctx->dev, state->caam_ctx, digestsize,
3879                                         DMA_FROM_DEVICE);
3880         if (dma_mapping_error(ctx->dev, state->ctx_dma)) {
3881                 dev_err(ctx->dev, "unable to map ctx\n");
3882                 state->ctx_dma = 0;
3883                 goto unmap;
3884         }
3885
3886         dpaa2_fl_set_final(in_fle, true);
3887         dpaa2_fl_set_len(in_fle, req->nbytes);
3888         dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
3889         dpaa2_fl_set_addr(out_fle, state->ctx_dma);
3890         dpaa2_fl_set_len(out_fle, digestsize);
3891
3892         req_ctx->flc = &ctx->flc[DIGEST];
3893         req_ctx->flc_dma = ctx->flc_dma[DIGEST];
3894         req_ctx->cbk = ahash_done;
3895         req_ctx->ctx = &req->base;
3896         req_ctx->edesc = edesc;
3897         ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
3898         if (ret == -EINPROGRESS ||
3899             (ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG))
3900                 return ret;
3901
3902 unmap:
3903         ahash_unmap_ctx(ctx->dev, edesc, req, DMA_FROM_DEVICE);
3904         qi_cache_free(edesc);
3905         return ret;
3906 }
3907
3908 static int ahash_final_no_ctx(struct ahash_request *req)
3909 {
3910         struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
3911         struct caam_hash_ctx *ctx = crypto_ahash_ctx_dma(ahash);
3912         struct caam_hash_state *state = ahash_request_ctx_dma(req);
3913         struct caam_request *req_ctx = &state->caam_req;
3914         struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
3915         struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
3916         gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
3917                       GFP_KERNEL : GFP_ATOMIC;
3918         u8 *buf = state->buf;
3919         int buflen = state->buflen;
3920         int digestsize = crypto_ahash_digestsize(ahash);
3921         struct ahash_edesc *edesc;
3922         int ret = -ENOMEM;
3923
3924         /* allocate space for base edesc and link tables */
3925         edesc = qi_cache_zalloc(flags);
3926         if (!edesc)
3927                 return ret;
3928
3929         if (buflen) {
3930                 state->buf_dma = dma_map_single(ctx->dev, buf, buflen,
3931                                                 DMA_TO_DEVICE);
3932                 if (dma_mapping_error(ctx->dev, state->buf_dma)) {
3933                         dev_err(ctx->dev, "unable to map src\n");
3934                         goto unmap;
3935                 }
3936         }
3937
3938         state->ctx_dma_len = digestsize;
3939         state->ctx_dma = dma_map_single(ctx->dev, state->caam_ctx, digestsize,
3940                                         DMA_FROM_DEVICE);
3941         if (dma_mapping_error(ctx->dev, state->ctx_dma)) {
3942                 dev_err(ctx->dev, "unable to map ctx\n");
3943                 state->ctx_dma = 0;
3944                 goto unmap;
3945         }
3946
3947         memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
3948         dpaa2_fl_set_final(in_fle, true);
3949         /*
3950          * crypto engine requires the input entry to be present when
3951          * "frame list" FD is used.
3952          * Since engine does not support FMT=2'b11 (unused entry type), leaving
3953          * in_fle zeroized (except for "Final" flag) is the best option.
3954          */
3955         if (buflen) {
3956                 dpaa2_fl_set_format(in_fle, dpaa2_fl_single);
3957                 dpaa2_fl_set_addr(in_fle, state->buf_dma);
3958                 dpaa2_fl_set_len(in_fle, buflen);
3959         }
3960         dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
3961         dpaa2_fl_set_addr(out_fle, state->ctx_dma);
3962         dpaa2_fl_set_len(out_fle, digestsize);
3963
3964         req_ctx->flc = &ctx->flc[DIGEST];
3965         req_ctx->flc_dma = ctx->flc_dma[DIGEST];
3966         req_ctx->cbk = ahash_done;
3967         req_ctx->ctx = &req->base;
3968         req_ctx->edesc = edesc;
3969
3970         ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
3971         if (ret == -EINPROGRESS ||
3972             (ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG))
3973                 return ret;
3974
3975 unmap:
3976         ahash_unmap_ctx(ctx->dev, edesc, req, DMA_FROM_DEVICE);
3977         qi_cache_free(edesc);
3978         return ret;
3979 }
3980
3981 static int ahash_update_no_ctx(struct ahash_request *req)
3982 {
3983         struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
3984         struct caam_hash_ctx *ctx = crypto_ahash_ctx_dma(ahash);
3985         struct caam_hash_state *state = ahash_request_ctx_dma(req);
3986         struct caam_request *req_ctx = &state->caam_req;
3987         struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
3988         struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
3989         gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
3990                       GFP_KERNEL : GFP_ATOMIC;
3991         u8 *buf = state->buf;
3992         int *buflen = &state->buflen;
3993         int *next_buflen = &state->next_buflen;
3994         int in_len = *buflen + req->nbytes, to_hash;
3995         int qm_sg_bytes, src_nents, mapped_nents;
3996         struct ahash_edesc *edesc;
3997         int ret = 0;
3998
3999         *next_buflen = in_len & (crypto_tfm_alg_blocksize(&ahash->base) - 1);
4000         to_hash = in_len - *next_buflen;
4001
4002         if (to_hash) {
4003                 struct dpaa2_sg_entry *sg_table;
4004                 int src_len = req->nbytes - *next_buflen;
4005
4006                 src_nents = sg_nents_for_len(req->src, src_len);
4007                 if (src_nents < 0) {
4008                         dev_err(ctx->dev, "Invalid number of src SG.\n");
4009                         return src_nents;
4010                 }
4011
4012                 if (src_nents) {
4013                         mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
4014                                                   DMA_TO_DEVICE);
4015                         if (!mapped_nents) {
4016                                 dev_err(ctx->dev, "unable to DMA map source\n");
4017                                 return -ENOMEM;
4018                         }
4019                 } else {
4020                         mapped_nents = 0;
4021                 }
4022
4023                 /* allocate space for base edesc and link tables */
4024                 edesc = qi_cache_zalloc(flags);
4025                 if (!edesc) {
4026                         dma_unmap_sg(ctx->dev, req->src, src_nents,
4027                                      DMA_TO_DEVICE);
4028                         return -ENOMEM;
4029                 }
4030
4031                 edesc->src_nents = src_nents;
4032                 qm_sg_bytes = pad_sg_nents(1 + mapped_nents) *
4033                               sizeof(*sg_table);
4034                 sg_table = &edesc->sgt[0];
4035
4036                 ret = buf_map_to_qm_sg(ctx->dev, sg_table, state);
4037                 if (ret)
4038                         goto unmap_ctx;
4039
4040                 sg_to_qm_sg_last(req->src, src_len, sg_table + 1, 0);
4041
4042                 edesc->qm_sg_dma = dma_map_single(ctx->dev, sg_table,
4043                                                   qm_sg_bytes, DMA_TO_DEVICE);
4044                 if (dma_mapping_error(ctx->dev, edesc->qm_sg_dma)) {
4045                         dev_err(ctx->dev, "unable to map S/G table\n");
4046                         ret = -ENOMEM;
4047                         goto unmap_ctx;
4048                 }
4049                 edesc->qm_sg_bytes = qm_sg_bytes;
4050
4051                 state->ctx_dma_len = ctx->ctx_len;
4052                 state->ctx_dma = dma_map_single(ctx->dev, state->caam_ctx,
4053                                                 ctx->ctx_len, DMA_FROM_DEVICE);
4054                 if (dma_mapping_error(ctx->dev, state->ctx_dma)) {
4055                         dev_err(ctx->dev, "unable to map ctx\n");
4056                         state->ctx_dma = 0;
4057                         ret = -ENOMEM;
4058                         goto unmap_ctx;
4059                 }
4060
4061                 memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
4062                 dpaa2_fl_set_final(in_fle, true);
4063                 dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
4064                 dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
4065                 dpaa2_fl_set_len(in_fle, to_hash);
4066                 dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
4067                 dpaa2_fl_set_addr(out_fle, state->ctx_dma);
4068                 dpaa2_fl_set_len(out_fle, ctx->ctx_len);
4069
4070                 req_ctx->flc = &ctx->flc[UPDATE_FIRST];
4071                 req_ctx->flc_dma = ctx->flc_dma[UPDATE_FIRST];
4072                 req_ctx->cbk = ahash_done_ctx_dst;
4073                 req_ctx->ctx = &req->base;
4074                 req_ctx->edesc = edesc;
4075
4076                 ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
4077                 if (ret != -EINPROGRESS &&
4078                     !(ret == -EBUSY &&
4079                       req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG))
4080                         goto unmap_ctx;
4081
4082                 state->update = ahash_update_ctx;
4083                 state->finup = ahash_finup_ctx;
4084                 state->final = ahash_final_ctx;
4085         } else if (*next_buflen) {
4086                 scatterwalk_map_and_copy(buf + *buflen, req->src, 0,
4087                                          req->nbytes, 0);
4088                 *buflen = *next_buflen;
4089
4090                 print_hex_dump_debug("buf@" __stringify(__LINE__)": ",
4091                                      DUMP_PREFIX_ADDRESS, 16, 4, buf,
4092                                      *buflen, 1);
4093         }
4094
4095         return ret;
4096 unmap_ctx:
4097         ahash_unmap_ctx(ctx->dev, edesc, req, DMA_TO_DEVICE);
4098         qi_cache_free(edesc);
4099         return ret;
4100 }
4101
4102 static int ahash_finup_no_ctx(struct ahash_request *req)
4103 {
4104         struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
4105         struct caam_hash_ctx *ctx = crypto_ahash_ctx_dma(ahash);
4106         struct caam_hash_state *state = ahash_request_ctx_dma(req);
4107         struct caam_request *req_ctx = &state->caam_req;
4108         struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
4109         struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
4110         gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
4111                       GFP_KERNEL : GFP_ATOMIC;
4112         int buflen = state->buflen;
4113         int qm_sg_bytes, src_nents, mapped_nents;
4114         int digestsize = crypto_ahash_digestsize(ahash);
4115         struct ahash_edesc *edesc;
4116         struct dpaa2_sg_entry *sg_table;
4117         int ret = -ENOMEM;
4118
4119         src_nents = sg_nents_for_len(req->src, req->nbytes);
4120         if (src_nents < 0) {
4121                 dev_err(ctx->dev, "Invalid number of src SG.\n");
4122                 return src_nents;
4123         }
4124
4125         if (src_nents) {
4126                 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
4127                                           DMA_TO_DEVICE);
4128                 if (!mapped_nents) {
4129                         dev_err(ctx->dev, "unable to DMA map source\n");
4130                         return ret;
4131                 }
4132         } else {
4133                 mapped_nents = 0;
4134         }
4135
4136         /* allocate space for base edesc and link tables */
4137         edesc = qi_cache_zalloc(flags);
4138         if (!edesc) {
4139                 dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE);
4140                 return ret;
4141         }
4142
4143         edesc->src_nents = src_nents;
4144         qm_sg_bytes = pad_sg_nents(2 + mapped_nents) * sizeof(*sg_table);
4145         sg_table = &edesc->sgt[0];
4146
4147         ret = buf_map_to_qm_sg(ctx->dev, sg_table, state);
4148         if (ret)
4149                 goto unmap;
4150
4151         sg_to_qm_sg_last(req->src, req->nbytes, sg_table + 1, 0);
4152
4153         edesc->qm_sg_dma = dma_map_single(ctx->dev, sg_table, qm_sg_bytes,
4154                                           DMA_TO_DEVICE);
4155         if (dma_mapping_error(ctx->dev, edesc->qm_sg_dma)) {
4156                 dev_err(ctx->dev, "unable to map S/G table\n");
4157                 ret = -ENOMEM;
4158                 goto unmap;
4159         }
4160         edesc->qm_sg_bytes = qm_sg_bytes;
4161
4162         state->ctx_dma_len = digestsize;
4163         state->ctx_dma = dma_map_single(ctx->dev, state->caam_ctx, digestsize,
4164                                         DMA_FROM_DEVICE);
4165         if (dma_mapping_error(ctx->dev, state->ctx_dma)) {
4166                 dev_err(ctx->dev, "unable to map ctx\n");
4167                 state->ctx_dma = 0;
4168                 ret = -ENOMEM;
4169                 goto unmap;
4170         }
4171
4172         memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
4173         dpaa2_fl_set_final(in_fle, true);
4174         dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
4175         dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
4176         dpaa2_fl_set_len(in_fle, buflen + req->nbytes);
4177         dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
4178         dpaa2_fl_set_addr(out_fle, state->ctx_dma);
4179         dpaa2_fl_set_len(out_fle, digestsize);
4180
4181         req_ctx->flc = &ctx->flc[DIGEST];
4182         req_ctx->flc_dma = ctx->flc_dma[DIGEST];
4183         req_ctx->cbk = ahash_done;
4184         req_ctx->ctx = &req->base;
4185         req_ctx->edesc = edesc;
4186         ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
4187         if (ret != -EINPROGRESS &&
4188             !(ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG))
4189                 goto unmap;
4190
4191         return ret;
4192 unmap:
4193         ahash_unmap_ctx(ctx->dev, edesc, req, DMA_FROM_DEVICE);
4194         qi_cache_free(edesc);
4195         return ret;
4196 }
4197
4198 static int ahash_update_first(struct ahash_request *req)
4199 {
4200         struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
4201         struct caam_hash_ctx *ctx = crypto_ahash_ctx_dma(ahash);
4202         struct caam_hash_state *state = ahash_request_ctx_dma(req);
4203         struct caam_request *req_ctx = &state->caam_req;
4204         struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
4205         struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
4206         gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
4207                       GFP_KERNEL : GFP_ATOMIC;
4208         u8 *buf = state->buf;
4209         int *buflen = &state->buflen;
4210         int *next_buflen = &state->next_buflen;
4211         int to_hash;
4212         int src_nents, mapped_nents;
4213         struct ahash_edesc *edesc;
4214         int ret = 0;
4215
4216         *next_buflen = req->nbytes & (crypto_tfm_alg_blocksize(&ahash->base) -
4217                                       1);
4218         to_hash = req->nbytes - *next_buflen;
4219
4220         if (to_hash) {
4221                 struct dpaa2_sg_entry *sg_table;
4222                 int src_len = req->nbytes - *next_buflen;
4223
4224                 src_nents = sg_nents_for_len(req->src, src_len);
4225                 if (src_nents < 0) {
4226                         dev_err(ctx->dev, "Invalid number of src SG.\n");
4227                         return src_nents;
4228                 }
4229
4230                 if (src_nents) {
4231                         mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
4232                                                   DMA_TO_DEVICE);
4233                         if (!mapped_nents) {
4234                                 dev_err(ctx->dev, "unable to map source for DMA\n");
4235                                 return -ENOMEM;
4236                         }
4237                 } else {
4238                         mapped_nents = 0;
4239                 }
4240
4241                 /* allocate space for base edesc and link tables */
4242                 edesc = qi_cache_zalloc(flags);
4243                 if (!edesc) {
4244                         dma_unmap_sg(ctx->dev, req->src, src_nents,
4245                                      DMA_TO_DEVICE);
4246                         return -ENOMEM;
4247                 }
4248
4249                 edesc->src_nents = src_nents;
4250                 sg_table = &edesc->sgt[0];
4251
4252                 memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
4253                 dpaa2_fl_set_final(in_fle, true);
4254                 dpaa2_fl_set_len(in_fle, to_hash);
4255
4256                 if (mapped_nents > 1) {
4257                         int qm_sg_bytes;
4258
4259                         sg_to_qm_sg_last(req->src, src_len, sg_table, 0);
4260                         qm_sg_bytes = pad_sg_nents(mapped_nents) *
4261                                       sizeof(*sg_table);
4262                         edesc->qm_sg_dma = dma_map_single(ctx->dev, sg_table,
4263                                                           qm_sg_bytes,
4264                                                           DMA_TO_DEVICE);
4265                         if (dma_mapping_error(ctx->dev, edesc->qm_sg_dma)) {
4266                                 dev_err(ctx->dev, "unable to map S/G table\n");
4267                                 ret = -ENOMEM;
4268                                 goto unmap_ctx;
4269                         }
4270                         edesc->qm_sg_bytes = qm_sg_bytes;
4271                         dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
4272                         dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
4273                 } else {
4274                         dpaa2_fl_set_format(in_fle, dpaa2_fl_single);
4275                         dpaa2_fl_set_addr(in_fle, sg_dma_address(req->src));
4276                 }
4277
4278                 state->ctx_dma_len = ctx->ctx_len;
4279                 state->ctx_dma = dma_map_single(ctx->dev, state->caam_ctx,
4280                                                 ctx->ctx_len, DMA_FROM_DEVICE);
4281                 if (dma_mapping_error(ctx->dev, state->ctx_dma)) {
4282                         dev_err(ctx->dev, "unable to map ctx\n");
4283                         state->ctx_dma = 0;
4284                         ret = -ENOMEM;
4285                         goto unmap_ctx;
4286                 }
4287
4288                 dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
4289                 dpaa2_fl_set_addr(out_fle, state->ctx_dma);
4290                 dpaa2_fl_set_len(out_fle, ctx->ctx_len);
4291
4292                 req_ctx->flc = &ctx->flc[UPDATE_FIRST];
4293                 req_ctx->flc_dma = ctx->flc_dma[UPDATE_FIRST];
4294                 req_ctx->cbk = ahash_done_ctx_dst;
4295                 req_ctx->ctx = &req->base;
4296                 req_ctx->edesc = edesc;
4297
4298                 ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
4299                 if (ret != -EINPROGRESS &&
4300                     !(ret == -EBUSY && req->base.flags &
4301                       CRYPTO_TFM_REQ_MAY_BACKLOG))
4302                         goto unmap_ctx;
4303
4304                 state->update = ahash_update_ctx;
4305                 state->finup = ahash_finup_ctx;
4306                 state->final = ahash_final_ctx;
4307         } else if (*next_buflen) {
4308                 state->update = ahash_update_no_ctx;
4309                 state->finup = ahash_finup_no_ctx;
4310                 state->final = ahash_final_no_ctx;
4311                 scatterwalk_map_and_copy(buf, req->src, 0,
4312                                          req->nbytes, 0);
4313                 *buflen = *next_buflen;
4314
4315                 print_hex_dump_debug("buf@" __stringify(__LINE__)": ",
4316                                      DUMP_PREFIX_ADDRESS, 16, 4, buf,
4317                                      *buflen, 1);
4318         }
4319
4320         return ret;
4321 unmap_ctx:
4322         ahash_unmap_ctx(ctx->dev, edesc, req, DMA_TO_DEVICE);
4323         qi_cache_free(edesc);
4324         return ret;
4325 }
4326
4327 static int ahash_finup_first(struct ahash_request *req)
4328 {
4329         return ahash_digest(req);
4330 }
4331
4332 static int ahash_init(struct ahash_request *req)
4333 {
4334         struct caam_hash_state *state = ahash_request_ctx_dma(req);
4335
4336         state->update = ahash_update_first;
4337         state->finup = ahash_finup_first;
4338         state->final = ahash_final_no_ctx;
4339
4340         state->ctx_dma = 0;
4341         state->ctx_dma_len = 0;
4342         state->buf_dma = 0;
4343         state->buflen = 0;
4344         state->next_buflen = 0;
4345
4346         return 0;
4347 }
4348
4349 static int ahash_update(struct ahash_request *req)
4350 {
4351         struct caam_hash_state *state = ahash_request_ctx_dma(req);
4352
4353         return state->update(req);
4354 }
4355
4356 static int ahash_finup(struct ahash_request *req)
4357 {
4358         struct caam_hash_state *state = ahash_request_ctx_dma(req);
4359
4360         return state->finup(req);
4361 }
4362
4363 static int ahash_final(struct ahash_request *req)
4364 {
4365         struct caam_hash_state *state = ahash_request_ctx_dma(req);
4366
4367         return state->final(req);
4368 }
4369
4370 static int ahash_export(struct ahash_request *req, void *out)
4371 {
4372         struct caam_hash_state *state = ahash_request_ctx_dma(req);
4373         struct caam_export_state *export = out;
4374         u8 *buf = state->buf;
4375         int len = state->buflen;
4376
4377         memcpy(export->buf, buf, len);
4378         memcpy(export->caam_ctx, state->caam_ctx, sizeof(export->caam_ctx));
4379         export->buflen = len;
4380         export->update = state->update;
4381         export->final = state->final;
4382         export->finup = state->finup;
4383
4384         return 0;
4385 }
4386
4387 static int ahash_import(struct ahash_request *req, const void *in)
4388 {
4389         struct caam_hash_state *state = ahash_request_ctx_dma(req);
4390         const struct caam_export_state *export = in;
4391
4392         memset(state, 0, sizeof(*state));
4393         memcpy(state->buf, export->buf, export->buflen);
4394         memcpy(state->caam_ctx, export->caam_ctx, sizeof(state->caam_ctx));
4395         state->buflen = export->buflen;
4396         state->update = export->update;
4397         state->final = export->final;
4398         state->finup = export->finup;
4399
4400         return 0;
4401 }
4402
4403 struct caam_hash_template {
4404         char name[CRYPTO_MAX_ALG_NAME];
4405         char driver_name[CRYPTO_MAX_ALG_NAME];
4406         char hmac_name[CRYPTO_MAX_ALG_NAME];
4407         char hmac_driver_name[CRYPTO_MAX_ALG_NAME];
4408         unsigned int blocksize;
4409         struct ahash_alg template_ahash;
4410         u32 alg_type;
4411 };
4412
4413 /* ahash descriptors */
4414 static struct caam_hash_template driver_hash[] = {
4415         {
4416                 .name = "sha1",
4417                 .driver_name = "sha1-caam-qi2",
4418                 .hmac_name = "hmac(sha1)",
4419                 .hmac_driver_name = "hmac-sha1-caam-qi2",
4420                 .blocksize = SHA1_BLOCK_SIZE,
4421                 .template_ahash = {
4422                         .init = ahash_init,
4423                         .update = ahash_update,
4424                         .final = ahash_final,
4425                         .finup = ahash_finup,
4426                         .digest = ahash_digest,
4427                         .export = ahash_export,
4428                         .import = ahash_import,
4429                         .setkey = ahash_setkey,
4430                         .halg = {
4431                                 .digestsize = SHA1_DIGEST_SIZE,
4432                                 .statesize = sizeof(struct caam_export_state),
4433                         },
4434                 },
4435                 .alg_type = OP_ALG_ALGSEL_SHA1,
4436         }, {
4437                 .name = "sha224",
4438                 .driver_name = "sha224-caam-qi2",
4439                 .hmac_name = "hmac(sha224)",
4440                 .hmac_driver_name = "hmac-sha224-caam-qi2",
4441                 .blocksize = SHA224_BLOCK_SIZE,
4442                 .template_ahash = {
4443                         .init = ahash_init,
4444                         .update = ahash_update,
4445                         .final = ahash_final,
4446                         .finup = ahash_finup,
4447                         .digest = ahash_digest,
4448                         .export = ahash_export,
4449                         .import = ahash_import,
4450                         .setkey = ahash_setkey,
4451                         .halg = {
4452                                 .digestsize = SHA224_DIGEST_SIZE,
4453                                 .statesize = sizeof(struct caam_export_state),
4454                         },
4455                 },
4456                 .alg_type = OP_ALG_ALGSEL_SHA224,
4457         }, {
4458                 .name = "sha256",
4459                 .driver_name = "sha256-caam-qi2",
4460                 .hmac_name = "hmac(sha256)",
4461                 .hmac_driver_name = "hmac-sha256-caam-qi2",
4462                 .blocksize = SHA256_BLOCK_SIZE,
4463                 .template_ahash = {
4464                         .init = ahash_init,
4465                         .update = ahash_update,
4466                         .final = ahash_final,
4467                         .finup = ahash_finup,
4468                         .digest = ahash_digest,
4469                         .export = ahash_export,
4470                         .import = ahash_import,
4471                         .setkey = ahash_setkey,
4472                         .halg = {
4473                                 .digestsize = SHA256_DIGEST_SIZE,
4474                                 .statesize = sizeof(struct caam_export_state),
4475                         },
4476                 },
4477                 .alg_type = OP_ALG_ALGSEL_SHA256,
4478         }, {
4479                 .name = "sha384",
4480                 .driver_name = "sha384-caam-qi2",
4481                 .hmac_name = "hmac(sha384)",
4482                 .hmac_driver_name = "hmac-sha384-caam-qi2",
4483                 .blocksize = SHA384_BLOCK_SIZE,
4484                 .template_ahash = {
4485                         .init = ahash_init,
4486                         .update = ahash_update,
4487                         .final = ahash_final,
4488                         .finup = ahash_finup,
4489                         .digest = ahash_digest,
4490                         .export = ahash_export,
4491                         .import = ahash_import,
4492                         .setkey = ahash_setkey,
4493                         .halg = {
4494                                 .digestsize = SHA384_DIGEST_SIZE,
4495                                 .statesize = sizeof(struct caam_export_state),
4496                         },
4497                 },
4498                 .alg_type = OP_ALG_ALGSEL_SHA384,
4499         }, {
4500                 .name = "sha512",
4501                 .driver_name = "sha512-caam-qi2",
4502                 .hmac_name = "hmac(sha512)",
4503                 .hmac_driver_name = "hmac-sha512-caam-qi2",
4504                 .blocksize = SHA512_BLOCK_SIZE,
4505                 .template_ahash = {
4506                         .init = ahash_init,
4507                         .update = ahash_update,
4508                         .final = ahash_final,
4509                         .finup = ahash_finup,
4510                         .digest = ahash_digest,
4511                         .export = ahash_export,
4512                         .import = ahash_import,
4513                         .setkey = ahash_setkey,
4514                         .halg = {
4515                                 .digestsize = SHA512_DIGEST_SIZE,
4516                                 .statesize = sizeof(struct caam_export_state),
4517                         },
4518                 },
4519                 .alg_type = OP_ALG_ALGSEL_SHA512,
4520         }, {
4521                 .name = "md5",
4522                 .driver_name = "md5-caam-qi2",
4523                 .hmac_name = "hmac(md5)",
4524                 .hmac_driver_name = "hmac-md5-caam-qi2",
4525                 .blocksize = MD5_BLOCK_WORDS * 4,
4526                 .template_ahash = {
4527                         .init = ahash_init,
4528                         .update = ahash_update,
4529                         .final = ahash_final,
4530                         .finup = ahash_finup,
4531                         .digest = ahash_digest,
4532                         .export = ahash_export,
4533                         .import = ahash_import,
4534                         .setkey = ahash_setkey,
4535                         .halg = {
4536                                 .digestsize = MD5_DIGEST_SIZE,
4537                                 .statesize = sizeof(struct caam_export_state),
4538                         },
4539                 },
4540                 .alg_type = OP_ALG_ALGSEL_MD5,
4541         }
4542 };
4543
4544 struct caam_hash_alg {
4545         struct list_head entry;
4546         struct device *dev;
4547         int alg_type;
4548         bool is_hmac;
4549         struct ahash_alg ahash_alg;
4550 };
4551
4552 static int caam_hash_cra_init(struct crypto_tfm *tfm)
4553 {
4554         struct crypto_ahash *ahash = __crypto_ahash_cast(tfm);
4555         struct crypto_alg *base = tfm->__crt_alg;
4556         struct hash_alg_common *halg =
4557                  container_of(base, struct hash_alg_common, base);
4558         struct ahash_alg *alg =
4559                  container_of(halg, struct ahash_alg, halg);
4560         struct caam_hash_alg *caam_hash =
4561                  container_of(alg, struct caam_hash_alg, ahash_alg);
4562         struct caam_hash_ctx *ctx = crypto_tfm_ctx_dma(tfm);
4563         /* Sizes for MDHA running digests: MD5, SHA1, 224, 256, 384, 512 */
4564         static const u8 runninglen[] = { HASH_MSG_LEN + MD5_DIGEST_SIZE,
4565                                          HASH_MSG_LEN + SHA1_DIGEST_SIZE,
4566                                          HASH_MSG_LEN + 32,
4567                                          HASH_MSG_LEN + SHA256_DIGEST_SIZE,
4568                                          HASH_MSG_LEN + 64,
4569                                          HASH_MSG_LEN + SHA512_DIGEST_SIZE };
4570         dma_addr_t dma_addr;
4571         int i;
4572
4573         ctx->dev = caam_hash->dev;
4574
4575         if (caam_hash->is_hmac) {
4576                 ctx->adata.key_dma = dma_map_single_attrs(ctx->dev, ctx->key,
4577                                                           ARRAY_SIZE(ctx->key),
4578                                                           DMA_TO_DEVICE,
4579                                                           DMA_ATTR_SKIP_CPU_SYNC);
4580                 if (dma_mapping_error(ctx->dev, ctx->adata.key_dma)) {
4581                         dev_err(ctx->dev, "unable to map key\n");
4582                         return -ENOMEM;
4583                 }
4584         }
4585
4586         dma_addr = dma_map_single_attrs(ctx->dev, ctx->flc, sizeof(ctx->flc),
4587                                         DMA_BIDIRECTIONAL,
4588                                         DMA_ATTR_SKIP_CPU_SYNC);
4589         if (dma_mapping_error(ctx->dev, dma_addr)) {
4590                 dev_err(ctx->dev, "unable to map shared descriptors\n");
4591                 if (ctx->adata.key_dma)
4592                         dma_unmap_single_attrs(ctx->dev, ctx->adata.key_dma,
4593                                                ARRAY_SIZE(ctx->key),
4594                                                DMA_TO_DEVICE,
4595                                                DMA_ATTR_SKIP_CPU_SYNC);
4596                 return -ENOMEM;
4597         }
4598
4599         for (i = 0; i < HASH_NUM_OP; i++)
4600                 ctx->flc_dma[i] = dma_addr + i * sizeof(ctx->flc[i]);
4601
4602         /* copy descriptor header template value */
4603         ctx->adata.algtype = OP_TYPE_CLASS2_ALG | caam_hash->alg_type;
4604
4605         ctx->ctx_len = runninglen[(ctx->adata.algtype &
4606                                    OP_ALG_ALGSEL_SUBMASK) >>
4607                                   OP_ALG_ALGSEL_SHIFT];
4608
4609         crypto_ahash_set_reqsize_dma(ahash, sizeof(struct caam_hash_state));
4610
4611         /*
4612          * For keyed hash algorithms shared descriptors
4613          * will be created later in setkey() callback
4614          */
4615         return caam_hash->is_hmac ? 0 : ahash_set_sh_desc(ahash);
4616 }
4617
4618 static void caam_hash_cra_exit(struct crypto_tfm *tfm)
4619 {
4620         struct caam_hash_ctx *ctx = crypto_tfm_ctx_dma(tfm);
4621
4622         dma_unmap_single_attrs(ctx->dev, ctx->flc_dma[0], sizeof(ctx->flc),
4623                                DMA_BIDIRECTIONAL, DMA_ATTR_SKIP_CPU_SYNC);
4624         if (ctx->adata.key_dma)
4625                 dma_unmap_single_attrs(ctx->dev, ctx->adata.key_dma,
4626                                        ARRAY_SIZE(ctx->key), DMA_TO_DEVICE,
4627                                        DMA_ATTR_SKIP_CPU_SYNC);
4628 }
4629
4630 static struct caam_hash_alg *caam_hash_alloc(struct device *dev,
4631         struct caam_hash_template *template, bool keyed)
4632 {
4633         struct caam_hash_alg *t_alg;
4634         struct ahash_alg *halg;
4635         struct crypto_alg *alg;
4636
4637         t_alg = kzalloc(sizeof(*t_alg), GFP_KERNEL);
4638         if (!t_alg)
4639                 return ERR_PTR(-ENOMEM);
4640
4641         t_alg->ahash_alg = template->template_ahash;
4642         halg = &t_alg->ahash_alg;
4643         alg = &halg->halg.base;
4644
4645         if (keyed) {
4646                 snprintf(alg->cra_name, CRYPTO_MAX_ALG_NAME, "%s",
4647                          template->hmac_name);
4648                 snprintf(alg->cra_driver_name, CRYPTO_MAX_ALG_NAME, "%s",
4649                          template->hmac_driver_name);
4650                 t_alg->is_hmac = true;
4651         } else {
4652                 snprintf(alg->cra_name, CRYPTO_MAX_ALG_NAME, "%s",
4653                          template->name);
4654                 snprintf(alg->cra_driver_name, CRYPTO_MAX_ALG_NAME, "%s",
4655                          template->driver_name);
4656                 t_alg->ahash_alg.setkey = NULL;
4657                 t_alg->is_hmac = false;
4658         }
4659         alg->cra_module = THIS_MODULE;
4660         alg->cra_init = caam_hash_cra_init;
4661         alg->cra_exit = caam_hash_cra_exit;
4662         alg->cra_ctxsize = sizeof(struct caam_hash_ctx) + crypto_dma_padding();
4663         alg->cra_priority = CAAM_CRA_PRIORITY;
4664         alg->cra_blocksize = template->blocksize;
4665         alg->cra_alignmask = 0;
4666         alg->cra_flags = CRYPTO_ALG_ASYNC | CRYPTO_ALG_ALLOCATES_MEMORY;
4667
4668         t_alg->alg_type = template->alg_type;
4669         t_alg->dev = dev;
4670
4671         return t_alg;
4672 }
4673
4674 static void dpaa2_caam_fqdan_cb(struct dpaa2_io_notification_ctx *nctx)
4675 {
4676         struct dpaa2_caam_priv_per_cpu *ppriv;
4677
4678         ppriv = container_of(nctx, struct dpaa2_caam_priv_per_cpu, nctx);
4679         napi_schedule_irqoff(&ppriv->napi);
4680 }
4681
4682 static int __cold dpaa2_dpseci_dpio_setup(struct dpaa2_caam_priv *priv)
4683 {
4684         struct device *dev = priv->dev;
4685         struct dpaa2_io_notification_ctx *nctx;
4686         struct dpaa2_caam_priv_per_cpu *ppriv;
4687         int err, i = 0, cpu;
4688
4689         for_each_online_cpu(cpu) {
4690                 ppriv = per_cpu_ptr(priv->ppriv, cpu);
4691                 ppriv->priv = priv;
4692                 nctx = &ppriv->nctx;
4693                 nctx->is_cdan = 0;
4694                 nctx->id = ppriv->rsp_fqid;
4695                 nctx->desired_cpu = cpu;
4696                 nctx->cb = dpaa2_caam_fqdan_cb;
4697
4698                 /* Register notification callbacks */
4699                 ppriv->dpio = dpaa2_io_service_select(cpu);
4700                 err = dpaa2_io_service_register(ppriv->dpio, nctx, dev);
4701                 if (unlikely(err)) {
4702                         dev_dbg(dev, "No affine DPIO for cpu %d\n", cpu);
4703                         nctx->cb = NULL;
4704                         /*
4705                          * If no affine DPIO for this core, there's probably
4706                          * none available for next cores either. Signal we want
4707                          * to retry later, in case the DPIO devices weren't
4708                          * probed yet.
4709                          */
4710                         err = -EPROBE_DEFER;
4711                         goto err;
4712                 }
4713
4714                 ppriv->store = dpaa2_io_store_create(DPAA2_CAAM_STORE_SIZE,
4715                                                      dev);
4716                 if (unlikely(!ppriv->store)) {
4717                         dev_err(dev, "dpaa2_io_store_create() failed\n");
4718                         err = -ENOMEM;
4719                         goto err;
4720                 }
4721
4722                 if (++i == priv->num_pairs)
4723                         break;
4724         }
4725
4726         return 0;
4727
4728 err:
4729         for_each_online_cpu(cpu) {
4730                 ppriv = per_cpu_ptr(priv->ppriv, cpu);
4731                 if (!ppriv->nctx.cb)
4732                         break;
4733                 dpaa2_io_service_deregister(ppriv->dpio, &ppriv->nctx, dev);
4734         }
4735
4736         for_each_online_cpu(cpu) {
4737                 ppriv = per_cpu_ptr(priv->ppriv, cpu);
4738                 if (!ppriv->store)
4739                         break;
4740                 dpaa2_io_store_destroy(ppriv->store);
4741         }
4742
4743         return err;
4744 }
4745
4746 static void __cold dpaa2_dpseci_dpio_free(struct dpaa2_caam_priv *priv)
4747 {
4748         struct dpaa2_caam_priv_per_cpu *ppriv;
4749         int i = 0, cpu;
4750
4751         for_each_online_cpu(cpu) {
4752                 ppriv = per_cpu_ptr(priv->ppriv, cpu);
4753                 dpaa2_io_service_deregister(ppriv->dpio, &ppriv->nctx,
4754                                             priv->dev);
4755                 dpaa2_io_store_destroy(ppriv->store);
4756
4757                 if (++i == priv->num_pairs)
4758                         return;
4759         }
4760 }
4761
4762 static int dpaa2_dpseci_bind(struct dpaa2_caam_priv *priv)
4763 {
4764         struct dpseci_rx_queue_cfg rx_queue_cfg;
4765         struct device *dev = priv->dev;
4766         struct fsl_mc_device *ls_dev = to_fsl_mc_device(dev);
4767         struct dpaa2_caam_priv_per_cpu *ppriv;
4768         int err = 0, i = 0, cpu;
4769
4770         /* Configure Rx queues */
4771         for_each_online_cpu(cpu) {
4772                 ppriv = per_cpu_ptr(priv->ppriv, cpu);
4773
4774                 rx_queue_cfg.options = DPSECI_QUEUE_OPT_DEST |
4775                                        DPSECI_QUEUE_OPT_USER_CTX;
4776                 rx_queue_cfg.order_preservation_en = 0;
4777                 rx_queue_cfg.dest_cfg.dest_type = DPSECI_DEST_DPIO;
4778                 rx_queue_cfg.dest_cfg.dest_id = ppriv->nctx.dpio_id;
4779                 /*
4780                  * Rx priority (WQ) doesn't really matter, since we use
4781                  * pull mode, i.e. volatile dequeues from specific FQs
4782                  */
4783                 rx_queue_cfg.dest_cfg.priority = 0;
4784                 rx_queue_cfg.user_ctx = ppriv->nctx.qman64;
4785
4786                 err = dpseci_set_rx_queue(priv->mc_io, 0, ls_dev->mc_handle, i,
4787                                           &rx_queue_cfg);
4788                 if (err) {
4789                         dev_err(dev, "dpseci_set_rx_queue() failed with err %d\n",
4790                                 err);
4791                         return err;
4792                 }
4793
4794                 if (++i == priv->num_pairs)
4795                         break;
4796         }
4797
4798         return err;
4799 }
4800
4801 static void dpaa2_dpseci_congestion_free(struct dpaa2_caam_priv *priv)
4802 {
4803         struct device *dev = priv->dev;
4804
4805         if (!priv->cscn_mem)
4806                 return;
4807
4808         dma_unmap_single(dev, priv->cscn_dma, DPAA2_CSCN_SIZE, DMA_FROM_DEVICE);
4809         kfree(priv->cscn_mem);
4810 }
4811
4812 static void dpaa2_dpseci_free(struct dpaa2_caam_priv *priv)
4813 {
4814         struct device *dev = priv->dev;
4815         struct fsl_mc_device *ls_dev = to_fsl_mc_device(dev);
4816         int err;
4817
4818         if (DPSECI_VER(priv->major_ver, priv->minor_ver) > DPSECI_VER(5, 3)) {
4819                 err = dpseci_reset(priv->mc_io, 0, ls_dev->mc_handle);
4820                 if (err)
4821                         dev_err(dev, "dpseci_reset() failed\n");
4822         }
4823
4824         dpaa2_dpseci_congestion_free(priv);
4825         dpseci_close(priv->mc_io, 0, ls_dev->mc_handle);
4826 }
4827
4828 static void dpaa2_caam_process_fd(struct dpaa2_caam_priv *priv,
4829                                   const struct dpaa2_fd *fd)
4830 {
4831         struct caam_request *req;
4832         u32 fd_err;
4833
4834         if (dpaa2_fd_get_format(fd) != dpaa2_fd_list) {
4835                 dev_err(priv->dev, "Only Frame List FD format is supported!\n");
4836                 return;
4837         }
4838
4839         fd_err = dpaa2_fd_get_ctrl(fd) & FD_CTRL_ERR_MASK;
4840         if (unlikely(fd_err))
4841                 dev_err_ratelimited(priv->dev, "FD error: %08x\n", fd_err);
4842
4843         /*
4844          * FD[ADDR] is guaranteed to be valid, irrespective of errors reported
4845          * in FD[ERR] or FD[FRC].
4846          */
4847         req = dpaa2_caam_iova_to_virt(priv, dpaa2_fd_get_addr(fd));
4848         dma_unmap_single(priv->dev, req->fd_flt_dma, sizeof(req->fd_flt),
4849                          DMA_BIDIRECTIONAL);
4850         req->cbk(req->ctx, dpaa2_fd_get_frc(fd));
4851 }
4852
4853 static int dpaa2_caam_pull_fq(struct dpaa2_caam_priv_per_cpu *ppriv)
4854 {
4855         int err;
4856
4857         /* Retry while portal is busy */
4858         do {
4859                 err = dpaa2_io_service_pull_fq(ppriv->dpio, ppriv->rsp_fqid,
4860                                                ppriv->store);
4861         } while (err == -EBUSY);
4862
4863         if (unlikely(err))
4864                 dev_err(ppriv->priv->dev, "dpaa2_io_service_pull err %d", err);
4865
4866         return err;
4867 }
4868
4869 static int dpaa2_caam_store_consume(struct dpaa2_caam_priv_per_cpu *ppriv)
4870 {
4871         struct dpaa2_dq *dq;
4872         int cleaned = 0, is_last;
4873
4874         do {
4875                 dq = dpaa2_io_store_next(ppriv->store, &is_last);
4876                 if (unlikely(!dq)) {
4877                         if (unlikely(!is_last)) {
4878                                 dev_dbg(ppriv->priv->dev,
4879                                         "FQ %d returned no valid frames\n",
4880                                         ppriv->rsp_fqid);
4881                                 /*
4882                                  * MUST retry until we get some sort of
4883                                  * valid response token (be it "empty dequeue"
4884                                  * or a valid frame).
4885                                  */
4886                                 continue;
4887                         }
4888                         break;
4889                 }
4890
4891                 /* Process FD */
4892                 dpaa2_caam_process_fd(ppriv->priv, dpaa2_dq_fd(dq));
4893                 cleaned++;
4894         } while (!is_last);
4895
4896         return cleaned;
4897 }
4898
4899 static int dpaa2_dpseci_poll(struct napi_struct *napi, int budget)
4900 {
4901         struct dpaa2_caam_priv_per_cpu *ppriv;
4902         struct dpaa2_caam_priv *priv;
4903         int err, cleaned = 0, store_cleaned;
4904
4905         ppriv = container_of(napi, struct dpaa2_caam_priv_per_cpu, napi);
4906         priv = ppriv->priv;
4907
4908         if (unlikely(dpaa2_caam_pull_fq(ppriv)))
4909                 return 0;
4910
4911         do {
4912                 store_cleaned = dpaa2_caam_store_consume(ppriv);
4913                 cleaned += store_cleaned;
4914
4915                 if (store_cleaned == 0 ||
4916                     cleaned > budget - DPAA2_CAAM_STORE_SIZE)
4917                         break;
4918
4919                 /* Try to dequeue some more */
4920                 err = dpaa2_caam_pull_fq(ppriv);
4921                 if (unlikely(err))
4922                         break;
4923         } while (1);
4924
4925         if (cleaned < budget) {
4926                 napi_complete_done(napi, cleaned);
4927                 err = dpaa2_io_service_rearm(ppriv->dpio, &ppriv->nctx);
4928                 if (unlikely(err))
4929                         dev_err(priv->dev, "Notification rearm failed: %d\n",
4930                                 err);
4931         }
4932
4933         return cleaned;
4934 }
4935
4936 static int dpaa2_dpseci_congestion_setup(struct dpaa2_caam_priv *priv,
4937                                          u16 token)
4938 {
4939         struct dpseci_congestion_notification_cfg cong_notif_cfg = { 0 };
4940         struct device *dev = priv->dev;
4941         unsigned int alignmask;
4942         int err;
4943
4944         /*
4945          * Congestion group feature supported starting with DPSECI API v5.1
4946          * and only when object has been created with this capability.
4947          */
4948         if ((DPSECI_VER(priv->major_ver, priv->minor_ver) < DPSECI_VER(5, 1)) ||
4949             !(priv->dpseci_attr.options & DPSECI_OPT_HAS_CG))
4950                 return 0;
4951
4952         alignmask = DPAA2_CSCN_ALIGN - 1;
4953         alignmask |= dma_get_cache_alignment() - 1;
4954         priv->cscn_mem = kzalloc(ALIGN(DPAA2_CSCN_SIZE, alignmask + 1),
4955                                  GFP_KERNEL);
4956         if (!priv->cscn_mem)
4957                 return -ENOMEM;
4958
4959         priv->cscn_dma = dma_map_single(dev, priv->cscn_mem,
4960                                         DPAA2_CSCN_SIZE, DMA_FROM_DEVICE);
4961         if (dma_mapping_error(dev, priv->cscn_dma)) {
4962                 dev_err(dev, "Error mapping CSCN memory area\n");
4963                 err = -ENOMEM;
4964                 goto err_dma_map;
4965         }
4966
4967         cong_notif_cfg.units = DPSECI_CONGESTION_UNIT_BYTES;
4968         cong_notif_cfg.threshold_entry = DPAA2_SEC_CONG_ENTRY_THRESH;
4969         cong_notif_cfg.threshold_exit = DPAA2_SEC_CONG_EXIT_THRESH;
4970         cong_notif_cfg.message_ctx = (uintptr_t)priv;
4971         cong_notif_cfg.message_iova = priv->cscn_dma;
4972         cong_notif_cfg.notification_mode = DPSECI_CGN_MODE_WRITE_MEM_ON_ENTER |
4973                                         DPSECI_CGN_MODE_WRITE_MEM_ON_EXIT |
4974                                         DPSECI_CGN_MODE_COHERENT_WRITE;
4975
4976         err = dpseci_set_congestion_notification(priv->mc_io, 0, token,
4977                                                  &cong_notif_cfg);
4978         if (err) {
4979                 dev_err(dev, "dpseci_set_congestion_notification failed\n");
4980                 goto err_set_cong;
4981         }
4982
4983         return 0;
4984
4985 err_set_cong:
4986         dma_unmap_single(dev, priv->cscn_dma, DPAA2_CSCN_SIZE, DMA_FROM_DEVICE);
4987 err_dma_map:
4988         kfree(priv->cscn_mem);
4989
4990         return err;
4991 }
4992
4993 static int __cold dpaa2_dpseci_setup(struct fsl_mc_device *ls_dev)
4994 {
4995         struct device *dev = &ls_dev->dev;
4996         struct dpaa2_caam_priv *priv;
4997         struct dpaa2_caam_priv_per_cpu *ppriv;
4998         int err, cpu;
4999         u8 i;
5000
5001         priv = dev_get_drvdata(dev);
5002
5003         priv->dev = dev;
5004         priv->dpsec_id = ls_dev->obj_desc.id;
5005
5006         /* Get a handle for the DPSECI this interface is associate with */
5007         err = dpseci_open(priv->mc_io, 0, priv->dpsec_id, &ls_dev->mc_handle);
5008         if (err) {
5009                 dev_err(dev, "dpseci_open() failed: %d\n", err);
5010                 goto err_open;
5011         }
5012
5013         err = dpseci_get_api_version(priv->mc_io, 0, &priv->major_ver,
5014                                      &priv->minor_ver);
5015         if (err) {
5016                 dev_err(dev, "dpseci_get_api_version() failed\n");
5017                 goto err_get_vers;
5018         }
5019
5020         dev_info(dev, "dpseci v%d.%d\n", priv->major_ver, priv->minor_ver);
5021
5022         if (DPSECI_VER(priv->major_ver, priv->minor_ver) > DPSECI_VER(5, 3)) {
5023                 err = dpseci_reset(priv->mc_io, 0, ls_dev->mc_handle);
5024                 if (err) {
5025                         dev_err(dev, "dpseci_reset() failed\n");
5026                         goto err_get_vers;
5027                 }
5028         }
5029
5030         err = dpseci_get_attributes(priv->mc_io, 0, ls_dev->mc_handle,
5031                                     &priv->dpseci_attr);
5032         if (err) {
5033                 dev_err(dev, "dpseci_get_attributes() failed\n");
5034                 goto err_get_vers;
5035         }
5036
5037         err = dpseci_get_sec_attr(priv->mc_io, 0, ls_dev->mc_handle,
5038                                   &priv->sec_attr);
5039         if (err) {
5040                 dev_err(dev, "dpseci_get_sec_attr() failed\n");
5041                 goto err_get_vers;
5042         }
5043
5044         err = dpaa2_dpseci_congestion_setup(priv, ls_dev->mc_handle);
5045         if (err) {
5046                 dev_err(dev, "setup_congestion() failed\n");
5047                 goto err_get_vers;
5048         }
5049
5050         priv->num_pairs = min(priv->dpseci_attr.num_rx_queues,
5051                               priv->dpseci_attr.num_tx_queues);
5052         if (priv->num_pairs > num_online_cpus()) {
5053                 dev_warn(dev, "%d queues won't be used\n",
5054                          priv->num_pairs - num_online_cpus());
5055                 priv->num_pairs = num_online_cpus();
5056         }
5057
5058         for (i = 0; i < priv->dpseci_attr.num_rx_queues; i++) {
5059                 err = dpseci_get_rx_queue(priv->mc_io, 0, ls_dev->mc_handle, i,
5060                                           &priv->rx_queue_attr[i]);
5061                 if (err) {
5062                         dev_err(dev, "dpseci_get_rx_queue() failed\n");
5063                         goto err_get_rx_queue;
5064                 }
5065         }
5066
5067         for (i = 0; i < priv->dpseci_attr.num_tx_queues; i++) {
5068                 err = dpseci_get_tx_queue(priv->mc_io, 0, ls_dev->mc_handle, i,
5069                                           &priv->tx_queue_attr[i]);
5070                 if (err) {
5071                         dev_err(dev, "dpseci_get_tx_queue() failed\n");
5072                         goto err_get_rx_queue;
5073                 }
5074         }
5075
5076         i = 0;
5077         for_each_online_cpu(cpu) {
5078                 u8 j;
5079
5080                 j = i % priv->num_pairs;
5081
5082                 ppriv = per_cpu_ptr(priv->ppriv, cpu);
5083                 ppriv->req_fqid = priv->tx_queue_attr[j].fqid;
5084
5085                 /*
5086                  * Allow all cores to enqueue, while only some of them
5087                  * will take part in dequeuing.
5088                  */
5089                 if (++i > priv->num_pairs)
5090                         continue;
5091
5092                 ppriv->rsp_fqid = priv->rx_queue_attr[j].fqid;
5093                 ppriv->prio = j;
5094
5095                 dev_dbg(dev, "pair %d: rx queue %d, tx queue %d\n", j,
5096                         priv->rx_queue_attr[j].fqid,
5097                         priv->tx_queue_attr[j].fqid);
5098
5099                 ppriv->net_dev.dev = *dev;
5100                 INIT_LIST_HEAD(&ppriv->net_dev.napi_list);
5101                 netif_napi_add_tx_weight(&ppriv->net_dev, &ppriv->napi,
5102                                          dpaa2_dpseci_poll,
5103                                          DPAA2_CAAM_NAPI_WEIGHT);
5104         }
5105
5106         return 0;
5107
5108 err_get_rx_queue:
5109         dpaa2_dpseci_congestion_free(priv);
5110 err_get_vers:
5111         dpseci_close(priv->mc_io, 0, ls_dev->mc_handle);
5112 err_open:
5113         return err;
5114 }
5115
5116 static int dpaa2_dpseci_enable(struct dpaa2_caam_priv *priv)
5117 {
5118         struct device *dev = priv->dev;
5119         struct fsl_mc_device *ls_dev = to_fsl_mc_device(dev);
5120         struct dpaa2_caam_priv_per_cpu *ppriv;
5121         int i;
5122
5123         for (i = 0; i < priv->num_pairs; i++) {
5124                 ppriv = per_cpu_ptr(priv->ppriv, i);
5125                 napi_enable(&ppriv->napi);
5126         }
5127
5128         return dpseci_enable(priv->mc_io, 0, ls_dev->mc_handle);
5129 }
5130
5131 static int __cold dpaa2_dpseci_disable(struct dpaa2_caam_priv *priv)
5132 {
5133         struct device *dev = priv->dev;
5134         struct dpaa2_caam_priv_per_cpu *ppriv;
5135         struct fsl_mc_device *ls_dev = to_fsl_mc_device(dev);
5136         int i, err = 0, enabled;
5137
5138         err = dpseci_disable(priv->mc_io, 0, ls_dev->mc_handle);
5139         if (err) {
5140                 dev_err(dev, "dpseci_disable() failed\n");
5141                 return err;
5142         }
5143
5144         err = dpseci_is_enabled(priv->mc_io, 0, ls_dev->mc_handle, &enabled);
5145         if (err) {
5146                 dev_err(dev, "dpseci_is_enabled() failed\n");
5147                 return err;
5148         }
5149
5150         dev_dbg(dev, "disable: %s\n", enabled ? "false" : "true");
5151
5152         for (i = 0; i < priv->num_pairs; i++) {
5153                 ppriv = per_cpu_ptr(priv->ppriv, i);
5154                 napi_disable(&ppriv->napi);
5155                 netif_napi_del(&ppriv->napi);
5156         }
5157
5158         return 0;
5159 }
5160
5161 static struct list_head hash_list;
5162
5163 static int dpaa2_caam_probe(struct fsl_mc_device *dpseci_dev)
5164 {
5165         struct device *dev;
5166         struct dpaa2_caam_priv *priv;
5167         int i, err = 0;
5168         bool registered = false;
5169
5170         /*
5171          * There is no way to get CAAM endianness - there is no direct register
5172          * space access and MC f/w does not provide this attribute.
5173          * All DPAA2-based SoCs have little endian CAAM, thus hard-code this
5174          * property.
5175          */
5176         caam_little_end = true;
5177
5178         caam_imx = false;
5179
5180         dev = &dpseci_dev->dev;
5181
5182         priv = devm_kzalloc(dev, sizeof(*priv), GFP_KERNEL);
5183         if (!priv)
5184                 return -ENOMEM;
5185
5186         dev_set_drvdata(dev, priv);
5187
5188         priv->domain = iommu_get_domain_for_dev(dev);
5189
5190         qi_cache = kmem_cache_create("dpaa2_caamqicache", CAAM_QI_MEMCACHE_SIZE,
5191                                      0, 0, NULL);
5192         if (!qi_cache) {
5193                 dev_err(dev, "Can't allocate SEC cache\n");
5194                 return -ENOMEM;
5195         }
5196
5197         err = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(49));
5198         if (err) {
5199                 dev_err(dev, "dma_set_mask_and_coherent() failed\n");
5200                 goto err_dma_mask;
5201         }
5202
5203         /* Obtain a MC portal */
5204         err = fsl_mc_portal_allocate(dpseci_dev, 0, &priv->mc_io);
5205         if (err) {
5206                 if (err == -ENXIO)
5207                         err = -EPROBE_DEFER;
5208                 else
5209                         dev_err(dev, "MC portal allocation failed\n");
5210
5211                 goto err_dma_mask;
5212         }
5213
5214         priv->ppriv = alloc_percpu(*priv->ppriv);
5215         if (!priv->ppriv) {
5216                 dev_err(dev, "alloc_percpu() failed\n");
5217                 err = -ENOMEM;
5218                 goto err_alloc_ppriv;
5219         }
5220
5221         /* DPSECI initialization */
5222         err = dpaa2_dpseci_setup(dpseci_dev);
5223         if (err) {
5224                 dev_err(dev, "dpaa2_dpseci_setup() failed\n");
5225                 goto err_dpseci_setup;
5226         }
5227
5228         /* DPIO */
5229         err = dpaa2_dpseci_dpio_setup(priv);
5230         if (err) {
5231                 dev_err_probe(dev, err, "dpaa2_dpseci_dpio_setup() failed\n");
5232                 goto err_dpio_setup;
5233         }
5234
5235         /* DPSECI binding to DPIO */
5236         err = dpaa2_dpseci_bind(priv);
5237         if (err) {
5238                 dev_err(dev, "dpaa2_dpseci_bind() failed\n");
5239                 goto err_bind;
5240         }
5241
5242         /* DPSECI enable */
5243         err = dpaa2_dpseci_enable(priv);
5244         if (err) {
5245                 dev_err(dev, "dpaa2_dpseci_enable() failed\n");
5246                 goto err_bind;
5247         }
5248
5249         dpaa2_dpseci_debugfs_init(priv);
5250
5251         /* register crypto algorithms the device supports */
5252         for (i = 0; i < ARRAY_SIZE(driver_algs); i++) {
5253                 struct caam_skcipher_alg *t_alg = driver_algs + i;
5254                 u32 alg_sel = t_alg->caam.class1_alg_type & OP_ALG_ALGSEL_MASK;
5255
5256                 /* Skip DES algorithms if not supported by device */
5257                 if (!priv->sec_attr.des_acc_num &&
5258                     (alg_sel == OP_ALG_ALGSEL_3DES ||
5259                      alg_sel == OP_ALG_ALGSEL_DES))
5260                         continue;
5261
5262                 /* Skip AES algorithms if not supported by device */
5263                 if (!priv->sec_attr.aes_acc_num &&
5264                     alg_sel == OP_ALG_ALGSEL_AES)
5265                         continue;
5266
5267                 /* Skip CHACHA20 algorithms if not supported by device */
5268                 if (alg_sel == OP_ALG_ALGSEL_CHACHA20 &&
5269                     !priv->sec_attr.ccha_acc_num)
5270                         continue;
5271
5272                 t_alg->caam.dev = dev;
5273                 caam_skcipher_alg_init(t_alg);
5274
5275                 err = crypto_register_skcipher(&t_alg->skcipher);
5276                 if (err) {
5277                         dev_warn(dev, "%s alg registration failed: %d\n",
5278                                  t_alg->skcipher.base.cra_driver_name, err);
5279                         continue;
5280                 }
5281
5282                 t_alg->registered = true;
5283                 registered = true;
5284         }
5285
5286         for (i = 0; i < ARRAY_SIZE(driver_aeads); i++) {
5287                 struct caam_aead_alg *t_alg = driver_aeads + i;
5288                 u32 c1_alg_sel = t_alg->caam.class1_alg_type &
5289                                  OP_ALG_ALGSEL_MASK;
5290                 u32 c2_alg_sel = t_alg->caam.class2_alg_type &
5291                                  OP_ALG_ALGSEL_MASK;
5292
5293                 /* Skip DES algorithms if not supported by device */
5294                 if (!priv->sec_attr.des_acc_num &&
5295                     (c1_alg_sel == OP_ALG_ALGSEL_3DES ||
5296                      c1_alg_sel == OP_ALG_ALGSEL_DES))
5297                         continue;
5298
5299                 /* Skip AES algorithms if not supported by device */
5300                 if (!priv->sec_attr.aes_acc_num &&
5301                     c1_alg_sel == OP_ALG_ALGSEL_AES)
5302                         continue;
5303
5304                 /* Skip CHACHA20 algorithms if not supported by device */
5305                 if (c1_alg_sel == OP_ALG_ALGSEL_CHACHA20 &&
5306                     !priv->sec_attr.ccha_acc_num)
5307                         continue;
5308
5309                 /* Skip POLY1305 algorithms if not supported by device */
5310                 if (c2_alg_sel == OP_ALG_ALGSEL_POLY1305 &&
5311                     !priv->sec_attr.ptha_acc_num)
5312                         continue;
5313
5314                 /*
5315                  * Skip algorithms requiring message digests
5316                  * if MD not supported by device.
5317                  */
5318                 if ((c2_alg_sel & ~OP_ALG_ALGSEL_SUBMASK) == 0x40 &&
5319                     !priv->sec_attr.md_acc_num)
5320                         continue;
5321
5322                 t_alg->caam.dev = dev;
5323                 caam_aead_alg_init(t_alg);
5324
5325                 err = crypto_register_aead(&t_alg->aead);
5326                 if (err) {
5327                         dev_warn(dev, "%s alg registration failed: %d\n",
5328                                  t_alg->aead.base.cra_driver_name, err);
5329                         continue;
5330                 }
5331
5332                 t_alg->registered = true;
5333                 registered = true;
5334         }
5335         if (registered)
5336                 dev_info(dev, "algorithms registered in /proc/crypto\n");
5337
5338         /* register hash algorithms the device supports */
5339         INIT_LIST_HEAD(&hash_list);
5340
5341         /*
5342          * Skip registration of any hashing algorithms if MD block
5343          * is not present.
5344          */
5345         if (!priv->sec_attr.md_acc_num)
5346                 return 0;
5347
5348         for (i = 0; i < ARRAY_SIZE(driver_hash); i++) {
5349                 struct caam_hash_alg *t_alg;
5350                 struct caam_hash_template *alg = driver_hash + i;
5351
5352                 /* register hmac version */
5353                 t_alg = caam_hash_alloc(dev, alg, true);
5354                 if (IS_ERR(t_alg)) {
5355                         err = PTR_ERR(t_alg);
5356                         dev_warn(dev, "%s hash alg allocation failed: %d\n",
5357                                  alg->hmac_driver_name, err);
5358                         continue;
5359                 }
5360
5361                 err = crypto_register_ahash(&t_alg->ahash_alg);
5362                 if (err) {
5363                         dev_warn(dev, "%s alg registration failed: %d\n",
5364                                  t_alg->ahash_alg.halg.base.cra_driver_name,
5365                                  err);
5366                         kfree(t_alg);
5367                 } else {
5368                         list_add_tail(&t_alg->entry, &hash_list);
5369                 }
5370
5371                 /* register unkeyed version */
5372                 t_alg = caam_hash_alloc(dev, alg, false);
5373                 if (IS_ERR(t_alg)) {
5374                         err = PTR_ERR(t_alg);
5375                         dev_warn(dev, "%s alg allocation failed: %d\n",
5376                                  alg->driver_name, err);
5377                         continue;
5378                 }
5379
5380                 err = crypto_register_ahash(&t_alg->ahash_alg);
5381                 if (err) {
5382                         dev_warn(dev, "%s alg registration failed: %d\n",
5383                                  t_alg->ahash_alg.halg.base.cra_driver_name,
5384                                  err);
5385                         kfree(t_alg);
5386                 } else {
5387                         list_add_tail(&t_alg->entry, &hash_list);
5388                 }
5389         }
5390         if (!list_empty(&hash_list))
5391                 dev_info(dev, "hash algorithms registered in /proc/crypto\n");
5392
5393         return err;
5394
5395 err_bind:
5396         dpaa2_dpseci_dpio_free(priv);
5397 err_dpio_setup:
5398         dpaa2_dpseci_free(priv);
5399 err_dpseci_setup:
5400         free_percpu(priv->ppriv);
5401 err_alloc_ppriv:
5402         fsl_mc_portal_free(priv->mc_io);
5403 err_dma_mask:
5404         kmem_cache_destroy(qi_cache);
5405
5406         return err;
5407 }
5408
5409 static void __cold dpaa2_caam_remove(struct fsl_mc_device *ls_dev)
5410 {
5411         struct device *dev;
5412         struct dpaa2_caam_priv *priv;
5413         int i;
5414
5415         dev = &ls_dev->dev;
5416         priv = dev_get_drvdata(dev);
5417
5418         dpaa2_dpseci_debugfs_exit(priv);
5419
5420         for (i = 0; i < ARRAY_SIZE(driver_aeads); i++) {
5421                 struct caam_aead_alg *t_alg = driver_aeads + i;
5422
5423                 if (t_alg->registered)
5424                         crypto_unregister_aead(&t_alg->aead);
5425         }
5426
5427         for (i = 0; i < ARRAY_SIZE(driver_algs); i++) {
5428                 struct caam_skcipher_alg *t_alg = driver_algs + i;
5429
5430                 if (t_alg->registered)
5431                         crypto_unregister_skcipher(&t_alg->skcipher);
5432         }
5433
5434         if (hash_list.next) {
5435                 struct caam_hash_alg *t_hash_alg, *p;
5436
5437                 list_for_each_entry_safe(t_hash_alg, p, &hash_list, entry) {
5438                         crypto_unregister_ahash(&t_hash_alg->ahash_alg);
5439                         list_del(&t_hash_alg->entry);
5440                         kfree(t_hash_alg);
5441                 }
5442         }
5443
5444         dpaa2_dpseci_disable(priv);
5445         dpaa2_dpseci_dpio_free(priv);
5446         dpaa2_dpseci_free(priv);
5447         free_percpu(priv->ppriv);
5448         fsl_mc_portal_free(priv->mc_io);
5449         kmem_cache_destroy(qi_cache);
5450 }
5451
5452 int dpaa2_caam_enqueue(struct device *dev, struct caam_request *req)
5453 {
5454         struct dpaa2_fd fd;
5455         struct dpaa2_caam_priv *priv = dev_get_drvdata(dev);
5456         struct dpaa2_caam_priv_per_cpu *ppriv;
5457         int err = 0, i;
5458
5459         if (IS_ERR(req))
5460                 return PTR_ERR(req);
5461
5462         if (priv->cscn_mem) {
5463                 dma_sync_single_for_cpu(priv->dev, priv->cscn_dma,
5464                                         DPAA2_CSCN_SIZE,
5465                                         DMA_FROM_DEVICE);
5466                 if (unlikely(dpaa2_cscn_state_congested(priv->cscn_mem))) {
5467                         dev_dbg_ratelimited(dev, "Dropping request\n");
5468                         return -EBUSY;
5469                 }
5470         }
5471
5472         dpaa2_fl_set_flc(&req->fd_flt[1], req->flc_dma);
5473
5474         req->fd_flt_dma = dma_map_single(dev, req->fd_flt, sizeof(req->fd_flt),
5475                                          DMA_BIDIRECTIONAL);
5476         if (dma_mapping_error(dev, req->fd_flt_dma)) {
5477                 dev_err(dev, "DMA mapping error for QI enqueue request\n");
5478                 goto err_out;
5479         }
5480
5481         memset(&fd, 0, sizeof(fd));
5482         dpaa2_fd_set_format(&fd, dpaa2_fd_list);
5483         dpaa2_fd_set_addr(&fd, req->fd_flt_dma);
5484         dpaa2_fd_set_len(&fd, dpaa2_fl_get_len(&req->fd_flt[1]));
5485         dpaa2_fd_set_flc(&fd, req->flc_dma);
5486
5487         ppriv = raw_cpu_ptr(priv->ppriv);
5488         for (i = 0; i < (priv->dpseci_attr.num_tx_queues << 1); i++) {
5489                 err = dpaa2_io_service_enqueue_fq(ppriv->dpio, ppriv->req_fqid,
5490                                                   &fd);
5491                 if (err != -EBUSY)
5492                         break;
5493
5494                 cpu_relax();
5495         }
5496
5497         if (unlikely(err)) {
5498                 dev_err_ratelimited(dev, "Error enqueuing frame: %d\n", err);
5499                 goto err_out;
5500         }
5501
5502         return -EINPROGRESS;
5503
5504 err_out:
5505         dma_unmap_single(dev, req->fd_flt_dma, sizeof(req->fd_flt),
5506                          DMA_BIDIRECTIONAL);
5507         return -EIO;
5508 }
5509 EXPORT_SYMBOL(dpaa2_caam_enqueue);
5510
5511 static const struct fsl_mc_device_id dpaa2_caam_match_id_table[] = {
5512         {
5513                 .vendor = FSL_MC_VENDOR_FREESCALE,
5514                 .obj_type = "dpseci",
5515         },
5516         { .vendor = 0x0 }
5517 };
5518 MODULE_DEVICE_TABLE(fslmc, dpaa2_caam_match_id_table);
5519
5520 static struct fsl_mc_driver dpaa2_caam_driver = {
5521         .driver = {
5522                 .name           = KBUILD_MODNAME,
5523                 .owner          = THIS_MODULE,
5524         },
5525         .probe          = dpaa2_caam_probe,
5526         .remove         = dpaa2_caam_remove,
5527         .match_id_table = dpaa2_caam_match_id_table
5528 };
5529
5530 MODULE_LICENSE("Dual BSD/GPL");
5531 MODULE_AUTHOR("Freescale Semiconductor, Inc");
5532 MODULE_DESCRIPTION("Freescale DPAA2 CAAM Driver");
5533
5534 module_fsl_mc_driver(dpaa2_caam_driver);