crypto: mxs-dcp - Use skcipher for fallback
This patch replaces use of the obsolete ablkcipher with skcipher. Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
This commit is contained in:
parent
241118de58
commit
29406bb923
|
@ -11,7 +11,6 @@
|
||||||
* http://www.gnu.org/copyleft/gpl.html
|
* http://www.gnu.org/copyleft/gpl.html
|
||||||
*/
|
*/
|
||||||
|
|
||||||
#include <linux/crypto.h>
|
|
||||||
#include <linux/dma-mapping.h>
|
#include <linux/dma-mapping.h>
|
||||||
#include <linux/interrupt.h>
|
#include <linux/interrupt.h>
|
||||||
#include <linux/io.h>
|
#include <linux/io.h>
|
||||||
|
@ -25,6 +24,7 @@
|
||||||
#include <crypto/aes.h>
|
#include <crypto/aes.h>
|
||||||
#include <crypto/sha.h>
|
#include <crypto/sha.h>
|
||||||
#include <crypto/internal/hash.h>
|
#include <crypto/internal/hash.h>
|
||||||
|
#include <crypto/internal/skcipher.h>
|
||||||
|
|
||||||
#define DCP_MAX_CHANS 4
|
#define DCP_MAX_CHANS 4
|
||||||
#define DCP_BUF_SZ PAGE_SIZE
|
#define DCP_BUF_SZ PAGE_SIZE
|
||||||
|
@ -84,7 +84,7 @@ struct dcp_async_ctx {
|
||||||
unsigned int hot:1;
|
unsigned int hot:1;
|
||||||
|
|
||||||
/* Crypto-specific context */
|
/* Crypto-specific context */
|
||||||
struct crypto_ablkcipher *fallback;
|
struct crypto_skcipher *fallback;
|
||||||
unsigned int key_len;
|
unsigned int key_len;
|
||||||
uint8_t key[AES_KEYSIZE_128];
|
uint8_t key[AES_KEYSIZE_128];
|
||||||
};
|
};
|
||||||
|
@ -374,20 +374,22 @@ static int dcp_chan_thread_aes(void *data)
|
||||||
|
|
||||||
static int mxs_dcp_block_fallback(struct ablkcipher_request *req, int enc)
|
static int mxs_dcp_block_fallback(struct ablkcipher_request *req, int enc)
|
||||||
{
|
{
|
||||||
struct crypto_tfm *tfm =
|
struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
|
||||||
crypto_ablkcipher_tfm(crypto_ablkcipher_reqtfm(req));
|
struct dcp_async_ctx *ctx = crypto_ablkcipher_ctx(tfm);
|
||||||
struct dcp_async_ctx *ctx = crypto_ablkcipher_ctx(
|
SKCIPHER_REQUEST_ON_STACK(subreq, ctx->fallback);
|
||||||
crypto_ablkcipher_reqtfm(req));
|
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
ablkcipher_request_set_tfm(req, ctx->fallback);
|
skcipher_request_set_tfm(subreq, ctx->fallback);
|
||||||
|
skcipher_request_set_callback(subreq, req->base.flags, NULL, NULL);
|
||||||
|
skcipher_request_set_crypt(subreq, req->src, req->dst,
|
||||||
|
req->nbytes, req->info);
|
||||||
|
|
||||||
if (enc)
|
if (enc)
|
||||||
ret = crypto_ablkcipher_encrypt(req);
|
ret = crypto_skcipher_encrypt(subreq);
|
||||||
else
|
else
|
||||||
ret = crypto_ablkcipher_decrypt(req);
|
ret = crypto_skcipher_decrypt(subreq);
|
||||||
|
|
||||||
ablkcipher_request_set_tfm(req, __crypto_ablkcipher_cast(tfm));
|
skcipher_request_zero(subreq);
|
||||||
|
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
@ -453,28 +455,22 @@ static int mxs_dcp_aes_setkey(struct crypto_ablkcipher *tfm, const u8 *key,
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Check if the key size is supported by kernel at all. */
|
|
||||||
if (len != AES_KEYSIZE_192 && len != AES_KEYSIZE_256) {
|
|
||||||
tfm->base.crt_flags |= CRYPTO_TFM_RES_BAD_KEY_LEN;
|
|
||||||
return -EINVAL;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* If the requested AES key size is not supported by the hardware,
|
* If the requested AES key size is not supported by the hardware,
|
||||||
* but is supported by in-kernel software implementation, we use
|
* but is supported by in-kernel software implementation, we use
|
||||||
* software fallback.
|
* software fallback.
|
||||||
*/
|
*/
|
||||||
actx->fallback->base.crt_flags &= ~CRYPTO_TFM_REQ_MASK;
|
crypto_skcipher_clear_flags(actx->fallback, CRYPTO_TFM_REQ_MASK);
|
||||||
actx->fallback->base.crt_flags |=
|
crypto_skcipher_set_flags(actx->fallback,
|
||||||
tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK;
|
tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK);
|
||||||
|
|
||||||
ret = crypto_ablkcipher_setkey(actx->fallback, key, len);
|
ret = crypto_skcipher_setkey(actx->fallback, key, len);
|
||||||
if (!ret)
|
if (!ret)
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
tfm->base.crt_flags &= ~CRYPTO_TFM_RES_MASK;
|
tfm->base.crt_flags &= ~CRYPTO_TFM_RES_MASK;
|
||||||
tfm->base.crt_flags |=
|
tfm->base.crt_flags |= crypto_skcipher_get_flags(actx->fallback) &
|
||||||
actx->fallback->base.crt_flags & CRYPTO_TFM_RES_MASK;
|
CRYPTO_TFM_RES_MASK;
|
||||||
|
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
@ -484,9 +480,9 @@ static int mxs_dcp_aes_fallback_init(struct crypto_tfm *tfm)
|
||||||
const char *name = crypto_tfm_alg_name(tfm);
|
const char *name = crypto_tfm_alg_name(tfm);
|
||||||
const uint32_t flags = CRYPTO_ALG_ASYNC | CRYPTO_ALG_NEED_FALLBACK;
|
const uint32_t flags = CRYPTO_ALG_ASYNC | CRYPTO_ALG_NEED_FALLBACK;
|
||||||
struct dcp_async_ctx *actx = crypto_tfm_ctx(tfm);
|
struct dcp_async_ctx *actx = crypto_tfm_ctx(tfm);
|
||||||
struct crypto_ablkcipher *blk;
|
struct crypto_skcipher *blk;
|
||||||
|
|
||||||
blk = crypto_alloc_ablkcipher(name, 0, flags);
|
blk = crypto_alloc_skcipher(name, 0, flags);
|
||||||
if (IS_ERR(blk))
|
if (IS_ERR(blk))
|
||||||
return PTR_ERR(blk);
|
return PTR_ERR(blk);
|
||||||
|
|
||||||
|
@ -499,8 +495,7 @@ static void mxs_dcp_aes_fallback_exit(struct crypto_tfm *tfm)
|
||||||
{
|
{
|
||||||
struct dcp_async_ctx *actx = crypto_tfm_ctx(tfm);
|
struct dcp_async_ctx *actx = crypto_tfm_ctx(tfm);
|
||||||
|
|
||||||
crypto_free_ablkcipher(actx->fallback);
|
crypto_free_skcipher(actx->fallback);
|
||||||
actx->fallback = NULL;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
Loading…
Reference in New Issue