123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315 |
- // SPDX-License-Identifier: GPL-2.0
- /*
- * Crypto-API module for CRC-32 algorithms implemented with the
- * z/Architecture Vector Extension Facility.
- *
- * Copyright IBM Corp. 2015
- * Author(s): Hendrik Brueckner <brueckner@linux.vnet.ibm.com>
- */
- #define KMSG_COMPONENT "crc32-vx"
- #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
- #include <linux/module.h>
- #include <linux/cpufeature.h>
- #include <linux/crc32.h>
- #include <crypto/internal/hash.h>
- #include <asm/fpu/api.h>
- #define CRC32_BLOCK_SIZE 1
- #define CRC32_DIGEST_SIZE 4
- #define VX_MIN_LEN 64
- #define VX_ALIGNMENT 16L
- #define VX_ALIGN_MASK (VX_ALIGNMENT - 1)
- struct crc_ctx {
- u32 key;
- };
- struct crc_desc_ctx {
- u32 crc;
- };
- /* Prototypes for functions in assembly files */
- u32 crc32_le_vgfm_16(u32 crc, unsigned char const *buf, size_t size);
- u32 crc32_be_vgfm_16(u32 crc, unsigned char const *buf, size_t size);
- u32 crc32c_le_vgfm_16(u32 crc, unsigned char const *buf, size_t size);
- /*
- * DEFINE_CRC32_VX() - Define a CRC-32 function using the vector extension
- *
- * Creates a function to perform a particular CRC-32 computation. Depending
- * on the message buffer, the hardware-accelerated or software implementation
- * is used. Note that the message buffer is aligned to improve fetch
- * operations of VECTOR LOAD MULTIPLE instructions.
- *
- */
- #define DEFINE_CRC32_VX(___fname, ___crc32_vx, ___crc32_sw) \
- static u32 __pure ___fname(u32 crc, \
- unsigned char const *data, size_t datalen) \
- { \
- struct kernel_fpu vxstate; \
- unsigned long prealign, aligned, remaining; \
- \
- if (datalen < VX_MIN_LEN + VX_ALIGN_MASK) \
- return ___crc32_sw(crc, data, datalen); \
- \
- if ((unsigned long)data & VX_ALIGN_MASK) { \
- prealign = VX_ALIGNMENT - \
- ((unsigned long)data & VX_ALIGN_MASK); \
- datalen -= prealign; \
- crc = ___crc32_sw(crc, data, prealign); \
- data = (void *)((unsigned long)data + prealign); \
- } \
- \
- aligned = datalen & ~VX_ALIGN_MASK; \
- remaining = datalen & VX_ALIGN_MASK; \
- \
- kernel_fpu_begin(&vxstate, KERNEL_VXR_LOW); \
- crc = ___crc32_vx(crc, data, aligned); \
- kernel_fpu_end(&vxstate, KERNEL_VXR_LOW); \
- \
- if (remaining) \
- crc = ___crc32_sw(crc, data + aligned, remaining); \
- \
- return crc; \
- }
- DEFINE_CRC32_VX(crc32_le_vx, crc32_le_vgfm_16, crc32_le)
- DEFINE_CRC32_VX(crc32_be_vx, crc32_be_vgfm_16, crc32_be)
- DEFINE_CRC32_VX(crc32c_le_vx, crc32c_le_vgfm_16, __crc32c_le)
- static int crc32_vx_cra_init_zero(struct crypto_tfm *tfm)
- {
- struct crc_ctx *mctx = crypto_tfm_ctx(tfm);
- mctx->key = 0;
- return 0;
- }
- static int crc32_vx_cra_init_invert(struct crypto_tfm *tfm)
- {
- struct crc_ctx *mctx = crypto_tfm_ctx(tfm);
- mctx->key = ~0;
- return 0;
- }
- static int crc32_vx_init(struct shash_desc *desc)
- {
- struct crc_ctx *mctx = crypto_shash_ctx(desc->tfm);
- struct crc_desc_ctx *ctx = shash_desc_ctx(desc);
- ctx->crc = mctx->key;
- return 0;
- }
- static int crc32_vx_setkey(struct crypto_shash *tfm, const u8 *newkey,
- unsigned int newkeylen)
- {
- struct crc_ctx *mctx = crypto_shash_ctx(tfm);
- if (newkeylen != sizeof(mctx->key)) {
- crypto_shash_set_flags(tfm, CRYPTO_TFM_RES_BAD_KEY_LEN);
- return -EINVAL;
- }
- mctx->key = le32_to_cpu(*(__le32 *)newkey);
- return 0;
- }
- static int crc32be_vx_setkey(struct crypto_shash *tfm, const u8 *newkey,
- unsigned int newkeylen)
- {
- struct crc_ctx *mctx = crypto_shash_ctx(tfm);
- if (newkeylen != sizeof(mctx->key)) {
- crypto_shash_set_flags(tfm, CRYPTO_TFM_RES_BAD_KEY_LEN);
- return -EINVAL;
- }
- mctx->key = be32_to_cpu(*(__be32 *)newkey);
- return 0;
- }
- static int crc32le_vx_final(struct shash_desc *desc, u8 *out)
- {
- struct crc_desc_ctx *ctx = shash_desc_ctx(desc);
- *(__le32 *)out = cpu_to_le32p(&ctx->crc);
- return 0;
- }
- static int crc32be_vx_final(struct shash_desc *desc, u8 *out)
- {
- struct crc_desc_ctx *ctx = shash_desc_ctx(desc);
- *(__be32 *)out = cpu_to_be32p(&ctx->crc);
- return 0;
- }
- static int crc32c_vx_final(struct shash_desc *desc, u8 *out)
- {
- struct crc_desc_ctx *ctx = shash_desc_ctx(desc);
- /*
- * Perform a final XOR with 0xFFFFFFFF to be in sync
- * with the generic crc32c shash implementation.
- */
- *(__le32 *)out = ~cpu_to_le32p(&ctx->crc);
- return 0;
- }
- static int __crc32le_vx_finup(u32 *crc, const u8 *data, unsigned int len,
- u8 *out)
- {
- *(__le32 *)out = cpu_to_le32(crc32_le_vx(*crc, data, len));
- return 0;
- }
- static int __crc32be_vx_finup(u32 *crc, const u8 *data, unsigned int len,
- u8 *out)
- {
- *(__be32 *)out = cpu_to_be32(crc32_be_vx(*crc, data, len));
- return 0;
- }
- static int __crc32c_vx_finup(u32 *crc, const u8 *data, unsigned int len,
- u8 *out)
- {
- /*
- * Perform a final XOR with 0xFFFFFFFF to be in sync
- * with the generic crc32c shash implementation.
- */
- *(__le32 *)out = ~cpu_to_le32(crc32c_le_vx(*crc, data, len));
- return 0;
- }
- #define CRC32_VX_FINUP(alg, func) \
- static int alg ## _vx_finup(struct shash_desc *desc, const u8 *data, \
- unsigned int datalen, u8 *out) \
- { \
- return __ ## alg ## _vx_finup(shash_desc_ctx(desc), \
- data, datalen, out); \
- }
- CRC32_VX_FINUP(crc32le, crc32_le_vx)
- CRC32_VX_FINUP(crc32be, crc32_be_vx)
- CRC32_VX_FINUP(crc32c, crc32c_le_vx)
- #define CRC32_VX_DIGEST(alg, func) \
- static int alg ## _vx_digest(struct shash_desc *desc, const u8 *data, \
- unsigned int len, u8 *out) \
- { \
- return __ ## alg ## _vx_finup(crypto_shash_ctx(desc->tfm), \
- data, len, out); \
- }
- CRC32_VX_DIGEST(crc32le, crc32_le_vx)
- CRC32_VX_DIGEST(crc32be, crc32_be_vx)
- CRC32_VX_DIGEST(crc32c, crc32c_le_vx)
- #define CRC32_VX_UPDATE(alg, func) \
- static int alg ## _vx_update(struct shash_desc *desc, const u8 *data, \
- unsigned int datalen) \
- { \
- struct crc_desc_ctx *ctx = shash_desc_ctx(desc); \
- ctx->crc = func(ctx->crc, data, datalen); \
- return 0; \
- }
- CRC32_VX_UPDATE(crc32le, crc32_le_vx)
- CRC32_VX_UPDATE(crc32be, crc32_be_vx)
- CRC32_VX_UPDATE(crc32c, crc32c_le_vx)
- static struct shash_alg crc32_vx_algs[] = {
- /* CRC-32 LE */
- {
- .init = crc32_vx_init,
- .setkey = crc32_vx_setkey,
- .update = crc32le_vx_update,
- .final = crc32le_vx_final,
- .finup = crc32le_vx_finup,
- .digest = crc32le_vx_digest,
- .descsize = sizeof(struct crc_desc_ctx),
- .digestsize = CRC32_DIGEST_SIZE,
- .base = {
- .cra_name = "crc32",
- .cra_driver_name = "crc32-vx",
- .cra_priority = 200,
- .cra_flags = CRYPTO_ALG_OPTIONAL_KEY,
- .cra_blocksize = CRC32_BLOCK_SIZE,
- .cra_ctxsize = sizeof(struct crc_ctx),
- .cra_module = THIS_MODULE,
- .cra_init = crc32_vx_cra_init_zero,
- },
- },
- /* CRC-32 BE */
- {
- .init = crc32_vx_init,
- .setkey = crc32be_vx_setkey,
- .update = crc32be_vx_update,
- .final = crc32be_vx_final,
- .finup = crc32be_vx_finup,
- .digest = crc32be_vx_digest,
- .descsize = sizeof(struct crc_desc_ctx),
- .digestsize = CRC32_DIGEST_SIZE,
- .base = {
- .cra_name = "crc32be",
- .cra_driver_name = "crc32be-vx",
- .cra_priority = 200,
- .cra_flags = CRYPTO_ALG_OPTIONAL_KEY,
- .cra_blocksize = CRC32_BLOCK_SIZE,
- .cra_ctxsize = sizeof(struct crc_ctx),
- .cra_module = THIS_MODULE,
- .cra_init = crc32_vx_cra_init_zero,
- },
- },
- /* CRC-32C LE */
- {
- .init = crc32_vx_init,
- .setkey = crc32_vx_setkey,
- .update = crc32c_vx_update,
- .final = crc32c_vx_final,
- .finup = crc32c_vx_finup,
- .digest = crc32c_vx_digest,
- .descsize = sizeof(struct crc_desc_ctx),
- .digestsize = CRC32_DIGEST_SIZE,
- .base = {
- .cra_name = "crc32c",
- .cra_driver_name = "crc32c-vx",
- .cra_priority = 200,
- .cra_flags = CRYPTO_ALG_OPTIONAL_KEY,
- .cra_blocksize = CRC32_BLOCK_SIZE,
- .cra_ctxsize = sizeof(struct crc_ctx),
- .cra_module = THIS_MODULE,
- .cra_init = crc32_vx_cra_init_invert,
- },
- },
- };
- static int __init crc_vx_mod_init(void)
- {
- return crypto_register_shashes(crc32_vx_algs,
- ARRAY_SIZE(crc32_vx_algs));
- }
- static void __exit crc_vx_mod_exit(void)
- {
- crypto_unregister_shashes(crc32_vx_algs, ARRAY_SIZE(crc32_vx_algs));
- }
- module_cpu_feature_match(VXRS, crc_vx_mod_init);
- module_exit(crc_vx_mod_exit);
- MODULE_AUTHOR("Hendrik Brueckner <brueckner@linux.vnet.ibm.com>");
- MODULE_LICENSE("GPL");
- MODULE_ALIAS_CRYPTO("crc32");
- MODULE_ALIAS_CRYPTO("crc32-vx");
- MODULE_ALIAS_CRYPTO("crc32c");
- MODULE_ALIAS_CRYPTO("crc32c-vx");
|