- Copy Brotli algorithm 3rd party source code for tool Cc: Liming Gao <liming.gao@intel.com> Contributed-under: TianoCore Contribution Agreement 1.0 Signed-off-by: Bell Song <binx.song@intel.com> Reviewed-by: Liming Gao <liming.gao@intel.com>
		
			
				
	
	
		
			169 lines
		
	
	
		
			5.2 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			169 lines
		
	
	
		
			5.2 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /* Copyright 2013 Google Inc. All Rights Reserved.
 | |
| 
 | |
|    Distributed under MIT license.
 | |
|    See file LICENSE for detail or copy at https://opensource.org/licenses/MIT
 | |
| */
 | |
| 
 | |
| /* Macros for endianness, branch prediction and unaligned loads and stores. */
 | |
| 
 | |
| #ifndef BROTLI_ENC_PORT_H_
 | |
| #define BROTLI_ENC_PORT_H_
 | |
| 
 | |
| #include <assert.h>
 | |
| #include <string.h>  /* memcpy */
 | |
| 
 | |
| #include "../common/port.h"
 | |
| #include "../common/types.h"
 | |
| 
 | |
| #if defined OS_LINUX || defined OS_CYGWIN
 | |
| #include <endian.h>
 | |
| #elif defined OS_FREEBSD
 | |
| #include <machine/endian.h>
 | |
| #elif defined OS_MACOSX
 | |
| #include <machine/endian.h>
 | |
| /* Let's try and follow the Linux convention */
 | |
| #define __BYTE_ORDER  BYTE_ORDER
 | |
| #define __LITTLE_ENDIAN LITTLE_ENDIAN
 | |
| #endif
 | |
| 
 | |
| /* define the macro IS_LITTLE_ENDIAN
 | |
|    using the above endian definitions from endian.h if
 | |
|    endian.h was included */
 | |
| #ifdef __BYTE_ORDER
 | |
| #if __BYTE_ORDER == __LITTLE_ENDIAN
 | |
| #define IS_LITTLE_ENDIAN
 | |
| #endif
 | |
| 
 | |
| #else
 | |
| 
 | |
| #if defined(__LITTLE_ENDIAN__)
 | |
| #define IS_LITTLE_ENDIAN
 | |
| #endif
 | |
| #endif  /* __BYTE_ORDER */
 | |
| 
 | |
| #if defined(__BYTE_ORDER__) && (__BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__)
 | |
| #define IS_LITTLE_ENDIAN
 | |
| #endif
 | |
| 
 | |
| /* Enable little-endian optimization for x64 architecture on Windows. */
 | |
| #if (defined(_WIN32) || defined(_WIN64)) && defined(_M_X64)
 | |
| #define IS_LITTLE_ENDIAN
 | |
| #endif
 | |
| 
 | |
| /* Portable handling of unaligned loads, stores, and copies.
 | |
|    On some platforms, like ARM, the copy functions can be more efficient
 | |
|    then a load and a store. */
 | |
| 
 | |
| #if defined(ARCH_PIII) || \
 | |
|   defined(ARCH_ATHLON) || defined(ARCH_K8) || defined(_ARCH_PPC)
 | |
| 
 | |
| /* x86 and x86-64 can perform unaligned loads/stores directly;
 | |
|    modern PowerPC hardware can also do unaligned integer loads and stores;
 | |
|    but note: the FPU still sends unaligned loads and stores to a trap handler!
 | |
| */
 | |
| 
 | |
| #define BROTLI_UNALIGNED_LOAD32(_p) (*(const uint32_t *)(_p))
 | |
| #define BROTLI_UNALIGNED_LOAD64(_p) (*(const uint64_t *)(_p))
 | |
| 
 | |
| #define BROTLI_UNALIGNED_STORE32(_p, _val) \
 | |
|   (*(uint32_t *)(_p) = (_val))
 | |
| #define BROTLI_UNALIGNED_STORE64(_p, _val) \
 | |
|   (*(uint64_t *)(_p) = (_val))
 | |
| 
 | |
| #elif defined(__arm__) && \
 | |
|   !defined(__ARM_ARCH_5__) && \
 | |
|   !defined(__ARM_ARCH_5T__) && \
 | |
|   !defined(__ARM_ARCH_5TE__) && \
 | |
|   !defined(__ARM_ARCH_5TEJ__) && \
 | |
|   !defined(__ARM_ARCH_6__) && \
 | |
|   !defined(__ARM_ARCH_6J__) && \
 | |
|   !defined(__ARM_ARCH_6K__) && \
 | |
|   !defined(__ARM_ARCH_6Z__) && \
 | |
|   !defined(__ARM_ARCH_6ZK__) && \
 | |
|   !defined(__ARM_ARCH_6T2__)
 | |
| 
 | |
| /* ARMv7 and newer support native unaligned accesses, but only of 16-bit
 | |
|    and 32-bit values (not 64-bit); older versions either raise a fatal signal,
 | |
|    do an unaligned read and rotate the words around a bit, or do the reads very
 | |
|    slowly (trip through kernel mode). */
 | |
| 
 | |
| #define BROTLI_UNALIGNED_LOAD32(_p) (*(const uint32_t *)(_p))
 | |
| #define BROTLI_UNALIGNED_STORE32(_p, _val) \
 | |
|   (*(uint32_t *)(_p) = (_val))
 | |
| 
 | |
| static BROTLI_INLINE uint64_t BROTLI_UNALIGNED_LOAD64(const void *p) {
 | |
|   uint64_t t;
 | |
|   memcpy(&t, p, sizeof t);
 | |
|   return t;
 | |
| }
 | |
| 
 | |
| static BROTLI_INLINE void BROTLI_UNALIGNED_STORE64(void *p, uint64_t v) {
 | |
|   memcpy(p, &v, sizeof v);
 | |
| }
 | |
| 
 | |
| #else
 | |
| 
 | |
| /* These functions are provided for architectures that don't support */
 | |
| /* unaligned loads and stores. */
 | |
| 
 | |
| static BROTLI_INLINE uint32_t BROTLI_UNALIGNED_LOAD32(const void *p) {
 | |
|   uint32_t t;
 | |
|   memcpy(&t, p, sizeof t);
 | |
|   return t;
 | |
| }
 | |
| 
 | |
| static BROTLI_INLINE uint64_t BROTLI_UNALIGNED_LOAD64(const void *p) {
 | |
|   uint64_t t;
 | |
|   memcpy(&t, p, sizeof t);
 | |
|   return t;
 | |
| }
 | |
| 
 | |
| static BROTLI_INLINE void BROTLI_UNALIGNED_STORE32(void *p, uint32_t v) {
 | |
|   memcpy(p, &v, sizeof v);
 | |
| }
 | |
| 
 | |
| static BROTLI_INLINE void BROTLI_UNALIGNED_STORE64(void *p, uint64_t v) {
 | |
|   memcpy(p, &v, sizeof v);
 | |
| }
 | |
| 
 | |
| #endif
 | |
| 
 | |
| #if !defined(__cplusplus) && !defined(c_plusplus) && __STDC_VERSION__ >= 199901L
 | |
| #define BROTLI_RESTRICT restrict
 | |
| #elif BROTLI_GCC_VERSION > 295 || defined(__llvm__)
 | |
| #define BROTLI_RESTRICT __restrict
 | |
| #else
 | |
| #define BROTLI_RESTRICT
 | |
| #endif
 | |
| 
 | |
| #define _TEMPLATE(T)                                                           \
 | |
|   static BROTLI_INLINE T brotli_min_ ## T (T a, T b) { return a < b ? a : b; } \
 | |
|   static BROTLI_INLINE T brotli_max_ ## T (T a, T b) { return a > b ? a : b; }
 | |
| _TEMPLATE(double) _TEMPLATE(float) _TEMPLATE(int)
 | |
| _TEMPLATE(size_t) _TEMPLATE(uint32_t) _TEMPLATE(uint8_t)
 | |
| #undef _TEMPLATE
 | |
| #define BROTLI_MIN(T, A, B) (brotli_min_ ## T((A), (B)))
 | |
| #define BROTLI_MAX(T, A, B) (brotli_max_ ## T((A), (B)))
 | |
| 
 | |
| #define BROTLI_SWAP(T, A, I, J) { \
 | |
|   T __brotli_swap_tmp = (A)[(I)]; \
 | |
|   (A)[(I)] = (A)[(J)];            \
 | |
|   (A)[(J)] = __brotli_swap_tmp;   \
 | |
| }
 | |
| 
 | |
| #define BROTLI_ENSURE_CAPACITY(M, T, A, C, R) {  \
 | |
|   if (C < (R)) {                                 \
 | |
|     size_t _new_size = (C == 0) ? (R) : C;       \
 | |
|     T* new_array;                                \
 | |
|     while (_new_size < (R)) _new_size *= 2;      \
 | |
|     new_array = BROTLI_ALLOC((M), T, _new_size); \
 | |
|     if (!BROTLI_IS_OOM(m) && C != 0)             \
 | |
|       memcpy(new_array, A, C * sizeof(T));       \
 | |
|     BROTLI_FREE((M), A);                         \
 | |
|     A = new_array;                               \
 | |
|     C = _new_size;                               \
 | |
|   }                                              \
 | |
| }
 | |
| 
 | |
| #endif  /* BROTLI_ENC_PORT_H_ */
 |