Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

crypto: x86/glue_helper - rename glue_skwalk_fpu_begin()

There are no users of the original glue_fpu_begin() anymore, so rename
glue_skwalk_fpu_begin() to glue_fpu_begin() so that it matches
glue_fpu_end() again.

Signed-off-by: Eric Biggers <ebiggers@google.com>
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>

authored by

Eric Biggers and committed by
Herbert Xu
75d8a553 0d87d0f4

+14 -18
+2 -2
arch/x86/crypto/cast5_avx_glue.c
··· 50 50 static inline bool cast5_fpu_begin(bool fpu_enabled, struct skcipher_walk *walk, 51 51 unsigned int nbytes) 52 52 { 53 - return glue_skwalk_fpu_begin(CAST5_BLOCK_SIZE, CAST5_PARALLEL_BLOCKS, 54 - walk, fpu_enabled, nbytes); 53 + return glue_fpu_begin(CAST5_BLOCK_SIZE, CAST5_PARALLEL_BLOCKS, 54 + walk, fpu_enabled, nbytes); 55 55 } 56 56 57 57 static inline void cast5_fpu_end(bool fpu_enabled)
+9 -12
arch/x86/crypto/glue_helper.c
··· 50 50 unsigned int func_bytes; 51 51 unsigned int i; 52 52 53 - fpu_enabled = glue_skwalk_fpu_begin(bsize, 54 - gctx->fpu_blocks_limit, 55 - &walk, fpu_enabled, nbytes); 53 + fpu_enabled = glue_fpu_begin(bsize, gctx->fpu_blocks_limit, 54 + &walk, fpu_enabled, nbytes); 56 55 for (i = 0; i < gctx->num_funcs; i++) { 57 56 func_bytes = bsize * gctx->funcs[i].num_blocks; 58 57 ··· 128 129 unsigned int i; 129 130 u128 last_iv; 130 131 131 - fpu_enabled = glue_skwalk_fpu_begin(bsize, 132 - gctx->fpu_blocks_limit, 133 - &walk, fpu_enabled, nbytes); 132 + fpu_enabled = glue_fpu_begin(bsize, gctx->fpu_blocks_limit, 133 + &walk, fpu_enabled, nbytes); 134 134 /* Start of the last block. */ 135 135 src += nbytes / bsize - 1; 136 136 dst += nbytes / bsize - 1; ··· 188 190 unsigned int i; 189 191 le128 ctrblk; 190 192 191 - fpu_enabled = glue_skwalk_fpu_begin(bsize, 192 - gctx->fpu_blocks_limit, 193 - &walk, fpu_enabled, nbytes); 193 + fpu_enabled = glue_fpu_begin(bsize, gctx->fpu_blocks_limit, 194 + &walk, fpu_enabled, nbytes); 194 195 195 196 be128_to_le128(&ctrblk, (be128 *)walk.iv); 196 197 ··· 288 291 return err; 289 292 290 293 /* set minimum length to bsize, for tweak_fn */ 291 - fpu_enabled = glue_skwalk_fpu_begin(bsize, gctx->fpu_blocks_limit, 292 - &walk, fpu_enabled, 293 - nbytes < bsize ? bsize : nbytes); 294 + fpu_enabled = glue_fpu_begin(bsize, gctx->fpu_blocks_limit, 295 + &walk, fpu_enabled, 296 + nbytes < bsize ? bsize : nbytes); 294 297 295 298 /* calculate first value of T */ 296 299 tweak_fn(tweak_ctx, walk.iv, walk.iv);
+3 -4
arch/x86/include/asm/crypto/glue_helper.h
··· 44 44 struct common_glue_func_entry funcs[]; 45 45 }; 46 46 47 - static inline bool glue_skwalk_fpu_begin(unsigned int bsize, 48 - int fpu_blocks_limit, 49 - struct skcipher_walk *walk, 50 - bool fpu_enabled, unsigned int nbytes) 47 + static inline bool glue_fpu_begin(unsigned int bsize, int fpu_blocks_limit, 48 + struct skcipher_walk *walk, 49 + bool fpu_enabled, unsigned int nbytes) 51 50 { 52 51 if (likely(fpu_blocks_limit < 0)) 53 52 return false;