Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

at v5.6 121 lines 3.2 kB view raw
1// SPDX-License-Identifier: GPL-2.0-or-later 2/* 3 * Glue code for the SHA256 Secure Hash Algorithm assembly implementation 4 * using optimized ARM assembler and NEON instructions. 5 * 6 * Copyright © 2015 Google Inc. 7 * 8 * This file is based on sha256_ssse3_glue.c: 9 * Copyright (C) 2013 Intel Corporation 10 * Author: Tim Chen <tim.c.chen@linux.intel.com> 11 */ 12 13#include <crypto/internal/hash.h> 14#include <linux/crypto.h> 15#include <linux/init.h> 16#include <linux/module.h> 17#include <linux/mm.h> 18#include <linux/cryptohash.h> 19#include <linux/types.h> 20#include <linux/string.h> 21#include <crypto/sha.h> 22#include <crypto/sha256_base.h> 23#include <asm/simd.h> 24#include <asm/neon.h> 25 26#include "sha256_glue.h" 27 28asmlinkage void sha256_block_data_order(u32 *digest, const void *data, 29 unsigned int num_blks); 30 31int crypto_sha256_arm_update(struct shash_desc *desc, const u8 *data, 32 unsigned int len) 33{ 34 /* make sure casting to sha256_block_fn() is safe */ 35 BUILD_BUG_ON(offsetof(struct sha256_state, state) != 0); 36 37 return sha256_base_do_update(desc, data, len, 38 (sha256_block_fn *)sha256_block_data_order); 39} 40EXPORT_SYMBOL(crypto_sha256_arm_update); 41 42static int crypto_sha256_arm_final(struct shash_desc *desc, u8 *out) 43{ 44 sha256_base_do_finalize(desc, 45 (sha256_block_fn *)sha256_block_data_order); 46 return sha256_base_finish(desc, out); 47} 48 49int crypto_sha256_arm_finup(struct shash_desc *desc, const u8 *data, 50 unsigned int len, u8 *out) 51{ 52 sha256_base_do_update(desc, data, len, 53 (sha256_block_fn *)sha256_block_data_order); 54 return crypto_sha256_arm_final(desc, out); 55} 56EXPORT_SYMBOL(crypto_sha256_arm_finup); 57 58static struct shash_alg algs[] = { { 59 .digestsize = SHA256_DIGEST_SIZE, 60 .init = sha256_base_init, 61 .update = crypto_sha256_arm_update, 62 .final = crypto_sha256_arm_final, 63 .finup = crypto_sha256_arm_finup, 64 .descsize = sizeof(struct sha256_state), 65 .base = { 66 .cra_name = "sha256", 67 .cra_driver_name = "sha256-asm", 68 .cra_priority = 150, 69 .cra_blocksize = SHA256_BLOCK_SIZE, 70 .cra_module = THIS_MODULE, 71 } 72}, { 73 .digestsize = SHA224_DIGEST_SIZE, 74 .init = sha224_base_init, 75 .update = crypto_sha256_arm_update, 76 .final = crypto_sha256_arm_final, 77 .finup = crypto_sha256_arm_finup, 78 .descsize = sizeof(struct sha256_state), 79 .base = { 80 .cra_name = "sha224", 81 .cra_driver_name = "sha224-asm", 82 .cra_priority = 150, 83 .cra_blocksize = SHA224_BLOCK_SIZE, 84 .cra_module = THIS_MODULE, 85 } 86} }; 87 88static int __init sha256_mod_init(void) 89{ 90 int res = crypto_register_shashes(algs, ARRAY_SIZE(algs)); 91 92 if (res < 0) 93 return res; 94 95 if (IS_ENABLED(CONFIG_KERNEL_MODE_NEON) && cpu_has_neon()) { 96 res = crypto_register_shashes(sha256_neon_algs, 97 ARRAY_SIZE(sha256_neon_algs)); 98 99 if (res < 0) 100 crypto_unregister_shashes(algs, ARRAY_SIZE(algs)); 101 } 102 103 return res; 104} 105 106static void __exit sha256_mod_fini(void) 107{ 108 crypto_unregister_shashes(algs, ARRAY_SIZE(algs)); 109 110 if (IS_ENABLED(CONFIG_KERNEL_MODE_NEON) && cpu_has_neon()) 111 crypto_unregister_shashes(sha256_neon_algs, 112 ARRAY_SIZE(sha256_neon_algs)); 113} 114 115module_init(sha256_mod_init); 116module_exit(sha256_mod_fini); 117 118MODULE_LICENSE("GPL"); 119MODULE_DESCRIPTION("SHA256 Secure Hash Algorithm (ARM), including NEON"); 120 121MODULE_ALIAS_CRYPTO("sha256");