| Index: third_party/lzma_sdk/CpuArch.h
|
| diff --git a/third_party/lzma_sdk/CpuArch.h b/third_party/lzma_sdk/CpuArch.h
|
| index 8c0c84d6fcb78621dac68d2f2facd15cad64904a..f6a28ba7e67119118c95eaba8f264e22399e24d0 100644
|
| --- a/third_party/lzma_sdk/CpuArch.h
|
| +++ b/third_party/lzma_sdk/CpuArch.h
|
| @@ -1,28 +1,34 @@
|
| /* CpuArch.h -- CPU specific code
|
| -2010-10-26: Igor Pavlov : Public domain */
|
| +2015-12-01: Igor Pavlov : Public domain */
|
|
|
| #ifndef __CPU_ARCH_H
|
| #define __CPU_ARCH_H
|
|
|
| -#include <stdlib.h>
|
| -#include "Types.h"
|
| +#include "7zTypes.h"
|
|
|
| EXTERN_C_BEGIN
|
|
|
| /*
|
| MY_CPU_LE means that CPU is LITTLE ENDIAN.
|
| -If MY_CPU_LE is not defined, we don't know about that property of platform (it can be LITTLE ENDIAN).
|
| +MY_CPU_BE means that CPU is BIG ENDIAN.
|
| +If MY_CPU_LE and MY_CPU_BE are not defined, we don't know about ENDIANNESS of platform.
|
|
|
| MY_CPU_LE_UNALIGN means that CPU is LITTLE ENDIAN and CPU supports unaligned memory accesses.
|
| -If MY_CPU_LE_UNALIGN is not defined, we don't know about these properties of platform.
|
| */
|
|
|
| -#if defined(_M_X64) || defined(_M_AMD64) || defined(__x86_64__)
|
| -#define MY_CPU_AMD64
|
| +#if defined(_M_X64) \
|
| + || defined(_M_AMD64) \
|
| + || defined(__x86_64__) \
|
| + || defined(__AMD64__) \
|
| + || defined(__amd64__)
|
| + #define MY_CPU_AMD64
|
| #endif
|
|
|
| -#if defined(MY_CPU_AMD64) || defined(_M_IA64)
|
| -#define MY_CPU_64BIT
|
| +#if defined(MY_CPU_AMD64) \
|
| + || defined(_M_IA64) \
|
| + || defined(__AARCH64EL__) \
|
| + || defined(__AARCH64EB__)
|
| + #define MY_CPU_64BIT
|
| #endif
|
|
|
| #if defined(_M_IX86) || defined(__i386__)
|
| @@ -33,8 +39,13 @@ If MY_CPU_LE_UNALIGN is not defined, we don't know about these properties of pla
|
| #define MY_CPU_X86_OR_AMD64
|
| #endif
|
|
|
| -#if defined(MY_CPU_X86) || defined(_M_ARM)
|
| -#define MY_CPU_32BIT
|
| +#if defined(MY_CPU_X86) \
|
| + || defined(_M_ARM) \
|
| + || defined(__ARMEL__) \
|
| + || defined(__THUMBEL__) \
|
| + || defined(__ARMEB__) \
|
| + || defined(__THUMBEB__)
|
| + #define MY_CPU_32BIT
|
| #endif
|
|
|
| #if defined(_WIN32) && defined(_M_ARM)
|
| @@ -45,34 +56,63 @@ If MY_CPU_LE_UNALIGN is not defined, we don't know about these properties of pla
|
| #define MY_CPU_IA64_LE
|
| #endif
|
|
|
| -#if defined(MY_CPU_X86_OR_AMD64)
|
| -#define MY_CPU_LE_UNALIGN
|
| -#endif
|
| -
|
| -#if defined(MY_CPU_X86_OR_AMD64) || defined(MY_CPU_ARM_LE) || defined(MY_CPU_IA64_LE) || defined(__ARMEL__) || defined(__MIPSEL__) || defined(__LITTLE_ENDIAN__)
|
| -#define MY_CPU_LE
|
| +#if defined(MY_CPU_X86_OR_AMD64) \
|
| + || defined(MY_CPU_ARM_LE) \
|
| + || defined(MY_CPU_IA64_LE) \
|
| + || defined(__LITTLE_ENDIAN__) \
|
| + || defined(__ARMEL__) \
|
| + || defined(__THUMBEL__) \
|
| + || defined(__AARCH64EL__) \
|
| + || defined(__MIPSEL__) \
|
| + || defined(__MIPSEL) \
|
| + || defined(_MIPSEL) \
|
| + || (defined(__BYTE_ORDER__) && (__BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__))
|
| + #define MY_CPU_LE
|
| #endif
|
|
|
| -#if defined(__BIG_ENDIAN__)
|
| -#define MY_CPU_BE
|
| +#if defined(__BIG_ENDIAN__) \
|
| + || defined(__ARMEB__) \
|
| + || defined(__THUMBEB__) \
|
| + || defined(__AARCH64EB__) \
|
| + || defined(__MIPSEB__) \
|
| + || defined(__MIPSEB) \
|
| + || defined(_MIPSEB) \
|
| + || defined(__m68k__) \
|
| + || defined(__s390__) \
|
| + || defined(__s390x__) \
|
| + || defined(__zarch__) \
|
| + || (defined(__BYTE_ORDER__) && (__BYTE_ORDER__ == __ORDER_BIG_ENDIAN__))
|
| + #define MY_CPU_BE
|
| #endif
|
|
|
| #if defined(MY_CPU_LE) && defined(MY_CPU_BE)
|
| Stop_Compiling_Bad_Endian
|
| #endif
|
|
|
| +
|
| +#ifdef MY_CPU_LE
|
| + #if defined(MY_CPU_X86_OR_AMD64) \
|
| + /* || defined(__AARCH64EL__) */
|
| + #define MY_CPU_LE_UNALIGN
|
| + #endif
|
| +#endif
|
| +
|
| +
|
| #ifdef MY_CPU_LE_UNALIGN
|
|
|
| -#define GetUi16(p) (*(const UInt16 *)(p))
|
| -#define GetUi32(p) (*(const UInt32 *)(p))
|
| -#define GetUi64(p) (*(const UInt64 *)(p))
|
| -#define SetUi16(p, d) *(UInt16 *)(p) = (d);
|
| -#define SetUi32(p, d) *(UInt32 *)(p) = (d);
|
| -#define SetUi64(p, d) *(UInt64 *)(p) = (d);
|
| +#define GetUi16(p) (*(const UInt16 *)(const void *)(p))
|
| +#define GetUi32(p) (*(const UInt32 *)(const void *)(p))
|
| +#define GetUi64(p) (*(const UInt64 *)(const void *)(p))
|
| +
|
| +#define SetUi16(p, v) { *(UInt16 *)(p) = (v); }
|
| +#define SetUi32(p, v) { *(UInt32 *)(p) = (v); }
|
| +#define SetUi64(p, v) { *(UInt64 *)(p) = (v); }
|
|
|
| #else
|
|
|
| -#define GetUi16(p) (((const Byte *)(p))[0] | ((UInt16)((const Byte *)(p))[1] << 8))
|
| +#define GetUi16(p) ( (UInt16) ( \
|
| + ((const Byte *)(p))[0] | \
|
| + ((UInt16)((const Byte *)(p))[1] << 8) ))
|
|
|
| #define GetUi32(p) ( \
|
| ((const Byte *)(p))[0] | \
|
| @@ -82,29 +122,43 @@ Stop_Compiling_Bad_Endian
|
|
|
| #define GetUi64(p) (GetUi32(p) | ((UInt64)GetUi32(((const Byte *)(p)) + 4) << 32))
|
|
|
| -#define SetUi16(p, d) { UInt32 _x_ = (d); \
|
| - ((Byte *)(p))[0] = (Byte)_x_; \
|
| - ((Byte *)(p))[1] = (Byte)(_x_ >> 8); }
|
| +#define SetUi16(p, v) { Byte *_ppp_ = (Byte *)(p); UInt32 _vvv_ = (v); \
|
| + _ppp_[0] = (Byte)_vvv_; \
|
| + _ppp_[1] = (Byte)(_vvv_ >> 8); }
|
|
|
| -#define SetUi32(p, d) { UInt32 _x_ = (d); \
|
| - ((Byte *)(p))[0] = (Byte)_x_; \
|
| - ((Byte *)(p))[1] = (Byte)(_x_ >> 8); \
|
| - ((Byte *)(p))[2] = (Byte)(_x_ >> 16); \
|
| - ((Byte *)(p))[3] = (Byte)(_x_ >> 24); }
|
| +#define SetUi32(p, v) { Byte *_ppp_ = (Byte *)(p); UInt32 _vvv_ = (v); \
|
| + _ppp_[0] = (Byte)_vvv_; \
|
| + _ppp_[1] = (Byte)(_vvv_ >> 8); \
|
| + _ppp_[2] = (Byte)(_vvv_ >> 16); \
|
| + _ppp_[3] = (Byte)(_vvv_ >> 24); }
|
|
|
| -#define SetUi64(p, d) { UInt64 _x64_ = (d); \
|
| - SetUi32(p, (UInt32)_x64_); \
|
| - SetUi32(((Byte *)(p)) + 4, (UInt32)(_x64_ >> 32)); }
|
| +#define SetUi64(p, v) { Byte *_ppp2_ = (Byte *)(p); UInt64 _vvv2_ = (v); \
|
| + SetUi32(_ppp2_ , (UInt32)_vvv2_); \
|
| + SetUi32(_ppp2_ + 4, (UInt32)(_vvv2_ >> 32)); }
|
|
|
| #endif
|
|
|
| -#if defined(MY_CPU_LE_UNALIGN) && defined(_WIN64) && (_MSC_VER >= 1300)
|
| +
|
| +#if defined(MY_CPU_LE_UNALIGN) && /* defined(_WIN64) && */ (_MSC_VER >= 1300)
|
| +
|
| +/* Note: we use bswap instruction, that is unsupported in 386 cpu */
|
| +
|
| +#include <stdlib.h>
|
|
|
| #pragma intrinsic(_byteswap_ulong)
|
| #pragma intrinsic(_byteswap_uint64)
|
| #define GetBe32(p) _byteswap_ulong(*(const UInt32 *)(const Byte *)(p))
|
| #define GetBe64(p) _byteswap_uint64(*(const UInt64 *)(const Byte *)(p))
|
|
|
| +#define SetBe32(p, v) (*(UInt32 *)(void *)(p)) = _byteswap_ulong(v)
|
| +
|
| +#elif defined(MY_CPU_LE_UNALIGN) && defined (__GNUC__) && (__GNUC__ > 4 || (__GNUC__ == 4 && __GNUC_MINOR__ >= 3))
|
| +
|
| +#define GetBe32(p) __builtin_bswap32(*(const UInt32 *)(const Byte *)(p))
|
| +#define GetBe64(p) __builtin_bswap64(*(const UInt64 *)(const Byte *)(p))
|
| +
|
| +#define SetBe32(p, v) (*(UInt32 *)(void *)(p)) = __builtin_bswap32(v)
|
| +
|
| #else
|
|
|
| #define GetBe32(p) ( \
|
| @@ -115,9 +169,19 @@ Stop_Compiling_Bad_Endian
|
|
|
| #define GetBe64(p) (((UInt64)GetBe32(p) << 32) | GetBe32(((const Byte *)(p)) + 4))
|
|
|
| +#define SetBe32(p, v) { Byte *_ppp_ = (Byte *)(p); UInt32 _vvv_ = (v); \
|
| + _ppp_[0] = (Byte)(_vvv_ >> 24); \
|
| + _ppp_[1] = (Byte)(_vvv_ >> 16); \
|
| + _ppp_[2] = (Byte)(_vvv_ >> 8); \
|
| + _ppp_[3] = (Byte)_vvv_; }
|
| +
|
| #endif
|
|
|
| -#define GetBe16(p) (((UInt16)((const Byte *)(p))[0] << 8) | ((const Byte *)(p))[1])
|
| +
|
| +#define GetBe16(p) ( (UInt16) ( \
|
| + ((UInt16)((const Byte *)(p))[0] << 8) | \
|
| + ((const Byte *)(p))[1] ))
|
| +
|
|
|
|
|
| #ifdef MY_CPU_X86_OR_AMD64
|
| @@ -139,12 +203,14 @@ enum
|
| CPU_FIRM_VIA
|
| };
|
|
|
| +void MyCPUID(UInt32 function, UInt32 *a, UInt32 *b, UInt32 *c, UInt32 *d);
|
| +
|
| Bool x86cpuid_CheckAndRead(Cx86cpuid *p);
|
| int x86cpuid_GetFirm(const Cx86cpuid *p);
|
|
|
| -#define x86cpuid_GetFamily(p) (((p)->ver >> 8) & 0xFF00F)
|
| -#define x86cpuid_GetModel(p) (((p)->ver >> 4) & 0xF00F)
|
| -#define x86cpuid_GetStepping(p) ((p)->ver & 0xF)
|
| +#define x86cpuid_GetFamily(ver) (((ver >> 16) & 0xFF0) | ((ver >> 8) & 0xF))
|
| +#define x86cpuid_GetModel(ver) (((ver >> 12) & 0xF0) | ((ver >> 4) & 0xF))
|
| +#define x86cpuid_GetStepping(ver) (ver & 0xF)
|
|
|
| Bool CPU_Is_InOrder();
|
| Bool CPU_Is_Aes_Supported();
|
|
|