mirror of
https://github.com/google/brotli.git
synced 2024-11-22 19:50:06 +00:00
169 lines
5.0 KiB
C
169 lines
5.0 KiB
C
/* Copyright 2013 Google Inc. All Rights Reserved.
|
|
|
|
Distributed under MIT license.
|
|
See file LICENSE for detail or copy at https://opensource.org/licenses/MIT
|
|
*/
|
|
|
|
/* Macros for endianness, branch prediction and unaligned loads and stores. */
|
|
|
|
#ifndef BROTLI_ENC_PORT_H_
|
|
#define BROTLI_ENC_PORT_H_
|
|
|
|
#include <assert.h>
|
|
#include <string.h> /* memcpy */
|
|
|
|
#include "../common/port.h"
|
|
#include "../common/types.h"
|
|
|
|
#if defined OS_LINUX || defined OS_CYGWIN
|
|
#include <endian.h>
|
|
#elif defined OS_FREEBSD
|
|
#include <machine/endian.h>
|
|
#elif defined OS_MACOSX
|
|
#include <machine/endian.h>
|
|
/* Let's try and follow the Linux convention */
|
|
#define __BYTE_ORDER BYTE_ORDER
|
|
#define __LITTLE_ENDIAN LITTLE_ENDIAN
|
|
#endif
|
|
|
|
/* define the macro IS_LITTLE_ENDIAN
|
|
using the above endian definitions from endian.h if
|
|
endian.h was included */
|
|
#ifdef __BYTE_ORDER
|
|
#if __BYTE_ORDER == __LITTLE_ENDIAN
|
|
#define IS_LITTLE_ENDIAN
|
|
#endif
|
|
|
|
#else
|
|
|
|
#if defined(__LITTLE_ENDIAN__)
|
|
#define IS_LITTLE_ENDIAN
|
|
#endif
|
|
#endif /* __BYTE_ORDER */
|
|
|
|
#if defined(__BYTE_ORDER__) && (__BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__)
|
|
#define IS_LITTLE_ENDIAN
|
|
#endif
|
|
|
|
/* Enable little-endian optimization for x64 architecture on Windows. */
|
|
#if (defined(_WIN32) || defined(_WIN64)) && defined(_M_X64)
|
|
#define IS_LITTLE_ENDIAN
|
|
#endif
|
|
|
|
/* Portable handling of unaligned loads, stores, and copies.
|
|
On some platforms, like ARM, the copy functions can be more efficient
|
|
then a load and a store. */
|
|
|
|
#if defined(ARCH_PIII) || \
|
|
defined(ARCH_ATHLON) || defined(ARCH_K8) || defined(_ARCH_PPC)
|
|
|
|
/* x86 and x86-64 can perform unaligned loads/stores directly;
|
|
modern PowerPC hardware can also do unaligned integer loads and stores;
|
|
but note: the FPU still sends unaligned loads and stores to a trap handler!
|
|
*/
|
|
|
|
#define BROTLI_UNALIGNED_LOAD32(_p) (*(const uint32_t *)(_p))
|
|
#define BROTLI_UNALIGNED_LOAD64(_p) (*(const uint64_t *)(_p))
|
|
|
|
#define BROTLI_UNALIGNED_STORE32(_p, _val) \
|
|
(*(uint32_t *)(_p) = (_val))
|
|
#define BROTLI_UNALIGNED_STORE64(_p, _val) \
|
|
(*(uint64_t *)(_p) = (_val))
|
|
|
|
#elif defined(__arm__) && \
|
|
!defined(__ARM_ARCH_5__) && \
|
|
!defined(__ARM_ARCH_5T__) && \
|
|
!defined(__ARM_ARCH_5TE__) && \
|
|
!defined(__ARM_ARCH_5TEJ__) && \
|
|
!defined(__ARM_ARCH_6__) && \
|
|
!defined(__ARM_ARCH_6J__) && \
|
|
!defined(__ARM_ARCH_6K__) && \
|
|
!defined(__ARM_ARCH_6Z__) && \
|
|
!defined(__ARM_ARCH_6ZK__) && \
|
|
!defined(__ARM_ARCH_6T2__)
|
|
|
|
/* ARMv7 and newer support native unaligned accesses, but only of 16-bit
|
|
and 32-bit values (not 64-bit); older versions either raise a fatal signal,
|
|
do an unaligned read and rotate the words around a bit, or do the reads very
|
|
slowly (trip through kernel mode). */
|
|
|
|
#define BROTLI_UNALIGNED_LOAD32(_p) (*(const uint32_t *)(_p))
|
|
#define BROTLI_UNALIGNED_STORE32(_p, _val) \
|
|
(*(uint32_t *)(_p) = (_val))
|
|
|
|
static BROTLI_INLINE uint64_t BROTLI_UNALIGNED_LOAD64(const void *p) {
|
|
uint64_t t;
|
|
memcpy(&t, p, sizeof t);
|
|
return t;
|
|
}
|
|
|
|
static BROTLI_INLINE void BROTLI_UNALIGNED_STORE64(void *p, uint64_t v) {
|
|
memcpy(p, &v, sizeof v);
|
|
}
|
|
|
|
#else
|
|
|
|
/* These functions are provided for architectures that don't support */
|
|
/* unaligned loads and stores. */
|
|
|
|
static BROTLI_INLINE uint32_t BROTLI_UNALIGNED_LOAD32(const void *p) {
|
|
uint32_t t;
|
|
memcpy(&t, p, sizeof t);
|
|
return t;
|
|
}
|
|
|
|
static BROTLI_INLINE uint64_t BROTLI_UNALIGNED_LOAD64(const void *p) {
|
|
uint64_t t;
|
|
memcpy(&t, p, sizeof t);
|
|
return t;
|
|
}
|
|
|
|
static BROTLI_INLINE void BROTLI_UNALIGNED_STORE32(void *p, uint32_t v) {
|
|
memcpy(p, &v, sizeof v);
|
|
}
|
|
|
|
static BROTLI_INLINE void BROTLI_UNALIGNED_STORE64(void *p, uint64_t v) {
|
|
memcpy(p, &v, sizeof v);
|
|
}
|
|
|
|
#endif
|
|
|
|
#if !defined(__cplusplus) && !defined(c_plusplus) && __STDC_VERSION__ >= 199901L
|
|
#define BROTLI_RESTRICT restrict
|
|
#elif BROTLI_GCC_VERSION > 295 || defined(__llvm__)
|
|
#define BROTLI_RESTRICT __restrict
|
|
#else
|
|
#define BROTLI_RESTRICT
|
|
#endif
|
|
|
|
#define _TEMPLATE(T) \
|
|
static BROTLI_INLINE T brotli_min_ ## T (T a, T b) { return a < b ? a : b; } \
|
|
static BROTLI_INLINE T brotli_max_ ## T (T a, T b) { return a > b ? a : b; }
|
|
_TEMPLATE(double) _TEMPLATE(float) _TEMPLATE(int)
|
|
_TEMPLATE(size_t) _TEMPLATE(uint32_t) _TEMPLATE(uint8_t)
|
|
#undef _TEMPLATE
|
|
#define BROTLI_MIN(T, A, B) (brotli_min_ ## T((A), (B)))
|
|
#define BROTLI_MAX(T, A, B) (brotli_max_ ## T((A), (B)))
|
|
|
|
#define BROTLI_SWAP(T, A, I, J) { \
|
|
T __brotli_swap_tmp = (A)[(I)]; \
|
|
(A)[(I)] = (A)[(J)]; \
|
|
(A)[(J)] = __brotli_swap_tmp; \
|
|
}
|
|
|
|
#define BROTLI_ENSURE_CAPACITY(M, T, A, C, R) { \
|
|
if (C < (R)) { \
|
|
size_t _new_size = (C == 0) ? (R) : C; \
|
|
T* new_array; \
|
|
while (_new_size < (R)) _new_size *= 2; \
|
|
new_array = BROTLI_ALLOC((M), T, _new_size); \
|
|
if (!BROTLI_IS_OOM(m)) \
|
|
memcpy(new_array, A, C * sizeof(T)); \
|
|
BROTLI_FREE((M), A); \
|
|
A = new_array; \
|
|
C = _new_size; \
|
|
} \
|
|
}
|
|
|
|
#endif /* BROTLI_ENC_PORT_H_ */
|