Index: third_party/lzma_sdk/CpuArch.h |
diff --git a/third_party/lzma_sdk/CpuArch.h b/third_party/lzma_sdk/CpuArch.h |
index 8c0c84d6fcb78621dac68d2f2facd15cad64904a..f6a28ba7e67119118c95eaba8f264e22399e24d0 100644 |
--- a/third_party/lzma_sdk/CpuArch.h |
+++ b/third_party/lzma_sdk/CpuArch.h |
@@ -1,28 +1,34 @@ |
/* CpuArch.h -- CPU specific code |
-2010-10-26: Igor Pavlov : Public domain */ |
+2015-12-01: Igor Pavlov : Public domain */ |
#ifndef __CPU_ARCH_H |
#define __CPU_ARCH_H |
-#include <stdlib.h> |
-#include "Types.h" |
+#include "7zTypes.h" |
EXTERN_C_BEGIN |
/* |
MY_CPU_LE means that CPU is LITTLE ENDIAN. |
-If MY_CPU_LE is not defined, we don't know about that property of platform (it can be LITTLE ENDIAN). |
+MY_CPU_BE means that CPU is BIG ENDIAN. |
+If MY_CPU_LE and MY_CPU_BE are not defined, we don't know about ENDIANNESS of platform. |
MY_CPU_LE_UNALIGN means that CPU is LITTLE ENDIAN and CPU supports unaligned memory accesses. |
-If MY_CPU_LE_UNALIGN is not defined, we don't know about these properties of platform. |
*/ |
-#if defined(_M_X64) || defined(_M_AMD64) || defined(__x86_64__) |
-#define MY_CPU_AMD64 |
+#if defined(_M_X64) \ |
+ || defined(_M_AMD64) \ |
+ || defined(__x86_64__) \ |
+ || defined(__AMD64__) \ |
+ || defined(__amd64__) |
+ #define MY_CPU_AMD64 |
#endif |
-#if defined(MY_CPU_AMD64) || defined(_M_IA64) |
-#define MY_CPU_64BIT |
+#if defined(MY_CPU_AMD64) \ |
+ || defined(_M_IA64) \ |
+ || defined(__AARCH64EL__) \ |
+ || defined(__AARCH64EB__) |
+ #define MY_CPU_64BIT |
#endif |
#if defined(_M_IX86) || defined(__i386__) |
@@ -33,8 +39,13 @@ If MY_CPU_LE_UNALIGN is not defined, we don't know about these properties of pla |
#define MY_CPU_X86_OR_AMD64 |
#endif |
-#if defined(MY_CPU_X86) || defined(_M_ARM) |
-#define MY_CPU_32BIT |
+#if defined(MY_CPU_X86) \ |
+ || defined(_M_ARM) \ |
+ || defined(__ARMEL__) \ |
+ || defined(__THUMBEL__) \ |
+ || defined(__ARMEB__) \ |
+ || defined(__THUMBEB__) |
+ #define MY_CPU_32BIT |
#endif |
#if defined(_WIN32) && defined(_M_ARM) |
@@ -45,34 +56,63 @@ If MY_CPU_LE_UNALIGN is not defined, we don't know about these properties of pla |
#define MY_CPU_IA64_LE |
#endif |
-#if defined(MY_CPU_X86_OR_AMD64) |
-#define MY_CPU_LE_UNALIGN |
-#endif |
- |
-#if defined(MY_CPU_X86_OR_AMD64) || defined(MY_CPU_ARM_LE) || defined(MY_CPU_IA64_LE) || defined(__ARMEL__) || defined(__MIPSEL__) || defined(__LITTLE_ENDIAN__) |
-#define MY_CPU_LE |
+#if defined(MY_CPU_X86_OR_AMD64) \ |
+ || defined(MY_CPU_ARM_LE) \ |
+ || defined(MY_CPU_IA64_LE) \ |
+ || defined(__LITTLE_ENDIAN__) \ |
+ || defined(__ARMEL__) \ |
+ || defined(__THUMBEL__) \ |
+ || defined(__AARCH64EL__) \ |
+ || defined(__MIPSEL__) \ |
+ || defined(__MIPSEL) \ |
+ || defined(_MIPSEL) \ |
+ || (defined(__BYTE_ORDER__) && (__BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__)) |
+ #define MY_CPU_LE |
#endif |
-#if defined(__BIG_ENDIAN__) |
-#define MY_CPU_BE |
+#if defined(__BIG_ENDIAN__) \ |
+ || defined(__ARMEB__) \ |
+ || defined(__THUMBEB__) \ |
+ || defined(__AARCH64EB__) \ |
+ || defined(__MIPSEB__) \ |
+ || defined(__MIPSEB) \ |
+ || defined(_MIPSEB) \ |
+ || defined(__m68k__) \ |
+ || defined(__s390__) \ |
+ || defined(__s390x__) \ |
+ || defined(__zarch__) \ |
+ || (defined(__BYTE_ORDER__) && (__BYTE_ORDER__ == __ORDER_BIG_ENDIAN__)) |
+ #define MY_CPU_BE |
#endif |
#if defined(MY_CPU_LE) && defined(MY_CPU_BE) |
Stop_Compiling_Bad_Endian |
#endif |
+ |
+#ifdef MY_CPU_LE |
+ #if defined(MY_CPU_X86_OR_AMD64) \ |
+ /* || defined(__AARCH64EL__) */ |
+ #define MY_CPU_LE_UNALIGN |
+ #endif |
+#endif |
+ |
+ |
#ifdef MY_CPU_LE_UNALIGN |
-#define GetUi16(p) (*(const UInt16 *)(p)) |
-#define GetUi32(p) (*(const UInt32 *)(p)) |
-#define GetUi64(p) (*(const UInt64 *)(p)) |
-#define SetUi16(p, d) *(UInt16 *)(p) = (d); |
-#define SetUi32(p, d) *(UInt32 *)(p) = (d); |
-#define SetUi64(p, d) *(UInt64 *)(p) = (d); |
+#define GetUi16(p) (*(const UInt16 *)(const void *)(p)) |
+#define GetUi32(p) (*(const UInt32 *)(const void *)(p)) |
+#define GetUi64(p) (*(const UInt64 *)(const void *)(p)) |
+ |
+#define SetUi16(p, v) { *(UInt16 *)(p) = (v); } |
+#define SetUi32(p, v) { *(UInt32 *)(p) = (v); } |
+#define SetUi64(p, v) { *(UInt64 *)(p) = (v); } |
#else |
-#define GetUi16(p) (((const Byte *)(p))[0] | ((UInt16)((const Byte *)(p))[1] << 8)) |
+#define GetUi16(p) ( (UInt16) ( \ |
+ ((const Byte *)(p))[0] | \ |
+ ((UInt16)((const Byte *)(p))[1] << 8) )) |
#define GetUi32(p) ( \ |
((const Byte *)(p))[0] | \ |
@@ -82,29 +122,43 @@ Stop_Compiling_Bad_Endian |
#define GetUi64(p) (GetUi32(p) | ((UInt64)GetUi32(((const Byte *)(p)) + 4) << 32)) |
-#define SetUi16(p, d) { UInt32 _x_ = (d); \ |
- ((Byte *)(p))[0] = (Byte)_x_; \ |
- ((Byte *)(p))[1] = (Byte)(_x_ >> 8); } |
+#define SetUi16(p, v) { Byte *_ppp_ = (Byte *)(p); UInt32 _vvv_ = (v); \ |
+ _ppp_[0] = (Byte)_vvv_; \ |
+ _ppp_[1] = (Byte)(_vvv_ >> 8); } |
-#define SetUi32(p, d) { UInt32 _x_ = (d); \ |
- ((Byte *)(p))[0] = (Byte)_x_; \ |
- ((Byte *)(p))[1] = (Byte)(_x_ >> 8); \ |
- ((Byte *)(p))[2] = (Byte)(_x_ >> 16); \ |
- ((Byte *)(p))[3] = (Byte)(_x_ >> 24); } |
+#define SetUi32(p, v) { Byte *_ppp_ = (Byte *)(p); UInt32 _vvv_ = (v); \ |
+ _ppp_[0] = (Byte)_vvv_; \ |
+ _ppp_[1] = (Byte)(_vvv_ >> 8); \ |
+ _ppp_[2] = (Byte)(_vvv_ >> 16); \ |
+ _ppp_[3] = (Byte)(_vvv_ >> 24); } |
-#define SetUi64(p, d) { UInt64 _x64_ = (d); \ |
- SetUi32(p, (UInt32)_x64_); \ |
- SetUi32(((Byte *)(p)) + 4, (UInt32)(_x64_ >> 32)); } |
+#define SetUi64(p, v) { Byte *_ppp2_ = (Byte *)(p); UInt64 _vvv2_ = (v); \ |
+ SetUi32(_ppp2_ , (UInt32)_vvv2_); \ |
+ SetUi32(_ppp2_ + 4, (UInt32)(_vvv2_ >> 32)); } |
#endif |
-#if defined(MY_CPU_LE_UNALIGN) && defined(_WIN64) && (_MSC_VER >= 1300) |
+ |
+#if defined(MY_CPU_LE_UNALIGN) && /* defined(_WIN64) && */ (_MSC_VER >= 1300) |
+ |
+/* Note: we use bswap instruction, that is unsupported in 386 cpu */ |
+ |
+#include <stdlib.h> |
#pragma intrinsic(_byteswap_ulong) |
#pragma intrinsic(_byteswap_uint64) |
#define GetBe32(p) _byteswap_ulong(*(const UInt32 *)(const Byte *)(p)) |
#define GetBe64(p) _byteswap_uint64(*(const UInt64 *)(const Byte *)(p)) |
+#define SetBe32(p, v) (*(UInt32 *)(void *)(p)) = _byteswap_ulong(v) |
+ |
+#elif defined(MY_CPU_LE_UNALIGN) && defined (__GNUC__) && (__GNUC__ > 4 || (__GNUC__ == 4 && __GNUC_MINOR__ >= 3)) |
+ |
+#define GetBe32(p) __builtin_bswap32(*(const UInt32 *)(const Byte *)(p)) |
+#define GetBe64(p) __builtin_bswap64(*(const UInt64 *)(const Byte *)(p)) |
+ |
+#define SetBe32(p, v) (*(UInt32 *)(void *)(p)) = __builtin_bswap32(v) |
+ |
#else |
#define GetBe32(p) ( \ |
@@ -115,9 +169,19 @@ Stop_Compiling_Bad_Endian |
#define GetBe64(p) (((UInt64)GetBe32(p) << 32) | GetBe32(((const Byte *)(p)) + 4)) |
+#define SetBe32(p, v) { Byte *_ppp_ = (Byte *)(p); UInt32 _vvv_ = (v); \ |
+ _ppp_[0] = (Byte)(_vvv_ >> 24); \ |
+ _ppp_[1] = (Byte)(_vvv_ >> 16); \ |
+ _ppp_[2] = (Byte)(_vvv_ >> 8); \ |
+ _ppp_[3] = (Byte)_vvv_; } |
+ |
#endif |
-#define GetBe16(p) (((UInt16)((const Byte *)(p))[0] << 8) | ((const Byte *)(p))[1]) |
+ |
+#define GetBe16(p) ( (UInt16) ( \ |
+ ((UInt16)((const Byte *)(p))[0] << 8) | \ |
+ ((const Byte *)(p))[1] )) |
+ |
#ifdef MY_CPU_X86_OR_AMD64 |
@@ -139,12 +203,14 @@ enum |
CPU_FIRM_VIA |
}; |
+void MyCPUID(UInt32 function, UInt32 *a, UInt32 *b, UInt32 *c, UInt32 *d); |
+ |
Bool x86cpuid_CheckAndRead(Cx86cpuid *p); |
int x86cpuid_GetFirm(const Cx86cpuid *p); |
-#define x86cpuid_GetFamily(p) (((p)->ver >> 8) & 0xFF00F) |
-#define x86cpuid_GetModel(p) (((p)->ver >> 4) & 0xF00F) |
-#define x86cpuid_GetStepping(p) ((p)->ver & 0xF) |
+#define x86cpuid_GetFamily(ver) (((ver >> 16) & 0xFF0) | ((ver >> 8) & 0xF)) |
+#define x86cpuid_GetModel(ver) (((ver >> 12) & 0xF0) | ((ver >> 4) & 0xF)) |
+#define x86cpuid_GetStepping(ver) (ver & 0xF) |
Bool CPU_Is_InOrder(); |
Bool CPU_Is_Aes_Supported(); |