Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

at v3.19-rc4 156 lines 3.8 kB view raw
1/* 2 * lz4defs.h -- architecture specific defines 3 * 4 * Copyright (C) 2013, LG Electronics, Kyungsik Lee <kyungsik.lee@lge.com> 5 * 6 * This program is free software; you can redistribute it and/or modify 7 * it under the terms of the GNU General Public License version 2 as 8 * published by the Free Software Foundation. 9 */ 10 11/* 12 * Detects 64 bits mode 13 */ 14#if (defined(__x86_64__) || defined(__x86_64) || defined(__amd64__) \ 15 || defined(__ppc64__) || defined(__LP64__)) 16#define LZ4_ARCH64 1 17#else 18#define LZ4_ARCH64 0 19#endif 20 21/* 22 * Architecture-specific macros 23 */ 24#define BYTE u8 25typedef struct _U16_S { u16 v; } U16_S; 26typedef struct _U32_S { u32 v; } U32_S; 27typedef struct _U64_S { u64 v; } U64_S; 28#if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) \ 29 || defined(CONFIG_ARM) && __LINUX_ARM_ARCH__ >= 6 \ 30 && defined(ARM_EFFICIENT_UNALIGNED_ACCESS) 31 32#define A16(x) (((U16_S *)(x))->v) 33#define A32(x) (((U32_S *)(x))->v) 34#define A64(x) (((U64_S *)(x))->v) 35 36#define PUT4(s, d) (A32(d) = A32(s)) 37#define PUT8(s, d) (A64(d) = A64(s)) 38#define LZ4_WRITE_LITTLEENDIAN_16(p, v) \ 39 do { \ 40 A16(p) = v; \ 41 p += 2; \ 42 } while (0) 43#else /* CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS */ 44 45#define A64(x) get_unaligned((u64 *)&(((U16_S *)(x))->v)) 46#define A32(x) get_unaligned((u32 *)&(((U16_S *)(x))->v)) 47#define A16(x) get_unaligned((u16 *)&(((U16_S *)(x))->v)) 48 49#define PUT4(s, d) \ 50 put_unaligned(get_unaligned((const u32 *) s), (u32 *) d) 51#define PUT8(s, d) \ 52 put_unaligned(get_unaligned((const u64 *) s), (u64 *) d) 53 54#define LZ4_WRITE_LITTLEENDIAN_16(p, v) \ 55 do { \ 56 put_unaligned(v, (u16 *)(p)); \ 57 p += 2; \ 58 } while (0) 59#endif 60 61#define COPYLENGTH 8 62#define ML_BITS 4 63#define ML_MASK ((1U << ML_BITS) - 1) 64#define RUN_BITS (8 - ML_BITS) 65#define RUN_MASK ((1U << RUN_BITS) - 1) 66#define MEMORY_USAGE 14 67#define MINMATCH 4 68#define SKIPSTRENGTH 6 69#define LASTLITERALS 5 70#define MFLIMIT (COPYLENGTH + MINMATCH) 71#define MINLENGTH (MFLIMIT + 1) 72#define MAXD_LOG 16 73#define MAXD (1 << MAXD_LOG) 74#define MAXD_MASK (u32)(MAXD - 1) 75#define MAX_DISTANCE (MAXD - 1) 76#define HASH_LOG (MAXD_LOG - 1) 77#define HASHTABLESIZE (1 << HASH_LOG) 78#define MAX_NB_ATTEMPTS 256 79#define OPTIMAL_ML (int)((ML_MASK-1)+MINMATCH) 80#define LZ4_64KLIMIT ((1<<16) + (MFLIMIT - 1)) 81#define HASHLOG64K ((MEMORY_USAGE - 2) + 1) 82#define HASH64KTABLESIZE (1U << HASHLOG64K) 83#define LZ4_HASH_VALUE(p) (((A32(p)) * 2654435761U) >> \ 84 ((MINMATCH * 8) - (MEMORY_USAGE-2))) 85#define LZ4_HASH64K_VALUE(p) (((A32(p)) * 2654435761U) >> \ 86 ((MINMATCH * 8) - HASHLOG64K)) 87#define HASH_VALUE(p) (((A32(p)) * 2654435761U) >> \ 88 ((MINMATCH * 8) - HASH_LOG)) 89 90#if LZ4_ARCH64/* 64-bit */ 91#define STEPSIZE 8 92 93#define LZ4_COPYSTEP(s, d) \ 94 do { \ 95 PUT8(s, d); \ 96 d += 8; \ 97 s += 8; \ 98 } while (0) 99 100#define LZ4_COPYPACKET(s, d) LZ4_COPYSTEP(s, d) 101 102#define LZ4_SECURECOPY(s, d, e) \ 103 do { \ 104 if (d < e) { \ 105 LZ4_WILDCOPY(s, d, e); \ 106 } \ 107 } while (0) 108#define HTYPE u32 109 110#ifdef __BIG_ENDIAN 111#define LZ4_NBCOMMONBYTES(val) (__builtin_clzll(val) >> 3) 112#else 113#define LZ4_NBCOMMONBYTES(val) (__builtin_ctzll(val) >> 3) 114#endif 115 116#else /* 32-bit */ 117#define STEPSIZE 4 118 119#define LZ4_COPYSTEP(s, d) \ 120 do { \ 121 PUT4(s, d); \ 122 d += 4; \ 123 s += 4; \ 124 } while (0) 125 126#define LZ4_COPYPACKET(s, d) \ 127 do { \ 128 LZ4_COPYSTEP(s, d); \ 129 LZ4_COPYSTEP(s, d); \ 130 } while (0) 131 132#define LZ4_SECURECOPY LZ4_WILDCOPY 133#define HTYPE const u8* 134 135#ifdef __BIG_ENDIAN 136#define LZ4_NBCOMMONBYTES(val) (__builtin_clz(val) >> 3) 137#else 138#define LZ4_NBCOMMONBYTES(val) (__builtin_ctz(val) >> 3) 139#endif 140 141#endif 142 143#define LZ4_READ_LITTLEENDIAN_16(d, s, p) \ 144 (d = s - get_unaligned_le16(p)) 145 146#define LZ4_WILDCOPY(s, d, e) \ 147 do { \ 148 LZ4_COPYPACKET(s, d); \ 149 } while (d < e) 150 151#define LZ4_BLINDCOPY(s, d, l) \ 152 do { \ 153 u8 *e = (d) + l; \ 154 LZ4_WILDCOPY(s, d, e); \ 155 d = e; \ 156 } while (0)