// AsmJit - Machine code generation for C++ // // * Official AsmJit Home Page: https://asmjit.com // * Official Github Repository: https://github.com/asmjit/asmjit // // Copyright (c) 2008-2020 The AsmJit Authors // // This software is provided 'as-is', without any express or implied // warranty. In no event will the authors be held liable for any damages // arising from the use of this software. // // Permission is granted to anyone to use this software for any purpose, // including commercial applications, and to alter it and redistribute it // freely, subject to the following restrictions: // // 1. The origin of this software must not be misrepresented; you must not // claim that you wrote the original software. If you use this software // in a product, an acknowledgment in the product documentation would be // appreciated but is not required. // 2. Altered source versions must be plainly marked as such, and must not be // misrepresented as being the original software. // 3. This notice may not be removed or altered from any source distribution. #ifndef ASMJIT_CORE_SUPPORT_H_INCLUDED #define ASMJIT_CORE_SUPPORT_H_INCLUDED #include "../core/globals.h" #if defined(_MSC_VER) #include #endif ASMJIT_BEGIN_NAMESPACE //! \addtogroup asmjit_utilities //! \{ //! Contains support classes and functions that may be used by AsmJit source //! and header files. Anything defined here is considered internal and should //! not be used outside of AsmJit and related projects like AsmTK. namespace Support { // ============================================================================ // [asmjit::Support - Architecture Features & Constraints] // ============================================================================ //! \cond INTERNAL static constexpr bool kUnalignedAccess16 = ASMJIT_ARCH_X86 != 0; static constexpr bool kUnalignedAccess32 = ASMJIT_ARCH_X86 != 0; static constexpr bool kUnalignedAccess64 = ASMJIT_ARCH_X86 != 0; //! \endcond // ============================================================================ // [asmjit::Support - Internal] // ============================================================================ //! \cond INTERNAL namespace Internal { template struct AlignedInt {}; template<> struct AlignedInt { typedef uint16_t ASMJIT_ALIGN_TYPE(T, 1); }; template<> struct AlignedInt { typedef uint16_t T; }; template<> struct AlignedInt { typedef uint32_t ASMJIT_ALIGN_TYPE(T, 1); }; template<> struct AlignedInt { typedef uint32_t ASMJIT_ALIGN_TYPE(T, 2); }; template<> struct AlignedInt { typedef uint32_t T; }; template<> struct AlignedInt { typedef uint64_t ASMJIT_ALIGN_TYPE(T, 1); }; template<> struct AlignedInt { typedef uint64_t ASMJIT_ALIGN_TYPE(T, 2); }; template<> struct AlignedInt { typedef uint64_t ASMJIT_ALIGN_TYPE(T, 4); }; template<> struct AlignedInt { typedef uint64_t T; }; // StdInt - Make an int-type by size (signed or unsigned) that is the // same as types defined by . // Int32Or64 - Make an int-type that has at least 32 bits: [u]int[32|64]_t. template struct StdInt {}; // Fail if not specialized. template<> struct StdInt<1, 0> { typedef int8_t Type; }; template<> struct StdInt<1, 1> { typedef uint8_t Type; }; template<> struct StdInt<2, 0> { typedef int16_t Type; }; template<> struct StdInt<2, 1> { typedef uint16_t Type; }; template<> struct StdInt<4, 0> { typedef int32_t Type; }; template<> struct StdInt<4, 1> { typedef uint32_t Type; }; template<> struct StdInt<8, 0> { typedef int64_t Type; }; template<> struct StdInt<8, 1> { typedef uint64_t Type; }; template::value> struct Int32Or64 : public StdInt {}; } //! \endcond // ============================================================================ // [asmjit::Support - Basic Traits] // ============================================================================ template static constexpr bool isUnsigned() noexcept { return std::is_unsigned::value; } // ============================================================================ // [asmjit::Support - FastUInt8] // ============================================================================ #if ASMJIT_ARCH_X86 typedef uint8_t FastUInt8; #else typedef unsigned int FastUInt8; #endif // ============================================================================ // [asmjit::Support - asInt / asUInt / asNormalized] // ============================================================================ //! Casts an integer `x` to either `int32_t` or `int64_t` depending on `T`. template static constexpr typename Internal::Int32Or64::Type asInt(const T& x) noexcept { return (typename Internal::Int32Or64::Type)x; } //! Casts an integer `x` to either `uint32_t` or `uint64_t` depending on `T`. template static constexpr typename Internal::Int32Or64::Type asUInt(const T& x) noexcept { return (typename Internal::Int32Or64::Type)x; } //! Casts an integer `x` to either `int32_t`, uint32_t`, `int64_t`, or `uint64_t` depending on `T`. template static constexpr typename Internal::Int32Or64::Type asNormalized(const T& x) noexcept { return (typename Internal::Int32Or64::Type)x; } //! Casts an integer `x` to the same type as defined by ``. template static constexpr typename Internal::StdInt()>::Type asStdInt(const T& x) noexcept { return (typename Internal::StdInt()>::Type)x; } // ============================================================================ // [asmjit::Support - BitCast] // ============================================================================ //! \cond namespace Internal { template union BitCastUnion { ASMJIT_INLINE BitCastUnion(SrcT src) noexcept : src(src) {} SrcT src; DstT dst; }; } //! \endcond //! Bit-casts from `Src` type to `Dst` type. //! //! Useful to bit-cast between integers and floating points. template static inline Dst bitCast(const Src& x) noexcept { return Internal::BitCastUnion(x).dst; } // ============================================================================ // [asmjit::Support - BitOps] // ============================================================================ //! Storage used to store a pack of bits (should by compatible with a machine word). typedef Internal::StdInt::Type BitWord; template static constexpr uint32_t bitSizeOf() noexcept { return uint32_t(sizeof(T) * 8u); } //! Number of bits stored in a single `BitWord`. static constexpr uint32_t kBitWordSizeInBits = bitSizeOf(); //! Returns `0 - x` in a safe way (no undefined behavior), works for unsigned numbers as well. template static constexpr T neg(const T& x) noexcept { typedef typename std::make_unsigned::type U; return T(U(0) - U(x)); } template static constexpr T allOnes() noexcept { return neg(T(1)); } //! Returns `x << y` (shift left logical) by explicitly casting `x` to an unsigned type and back. template static constexpr X shl(const X& x, const Y& y) noexcept { typedef typename std::make_unsigned::type U; return X(U(x) << y); } //! Returns `x >> y` (shift right logical) by explicitly casting `x` to an unsigned type and back. template static constexpr X shr(const X& x, const Y& y) noexcept { typedef typename std::make_unsigned::type U; return X(U(x) >> y); } //! Returns `x >> y` (shift right arithmetic) by explicitly casting `x` to a signed type and back. template static constexpr X sar(const X& x, const Y& y) noexcept { typedef typename std::make_signed::type S; return X(S(x) >> y); } //! Returns `x | (x >> y)` - helper used by some bit manipulation helpers. template static constexpr X or_shr(const X& x, const Y& y) noexcept { return X(x | shr(x, y)); } //! Returns `x & -x` - extracts lowest set isolated bit (like BLSI instruction). template static constexpr T blsi(T x) noexcept { typedef typename std::make_unsigned::type U; return T(U(x) & neg(U(x))); } //! Generate a trailing bit-mask that has `n` least significant (trailing) bits set. template static constexpr T lsbMask(const CountT& n) noexcept { typedef typename std::make_unsigned::type U; return (sizeof(U) < sizeof(uintptr_t)) // Prevent undefined behavior by using a larger type than T. ? T(U((uintptr_t(1) << n) - uintptr_t(1))) // Prevent undefined behavior by checking `n` before shift. : n ? T(shr(allOnes(), bitSizeOf() - size_t(n))) : T(0); } //! Tests whether the given value `x` has `n`th bit set. template static constexpr bool bitTest(T x, IndexT n) noexcept { typedef typename std::make_unsigned::type U; return (U(x) & (U(1) << n)) != 0; } //! Returns a bit-mask that has `x` bit set. template static constexpr uint32_t bitMask(T x) noexcept { return (1u << x); } //! Returns a bit-mask that has `x` bit set (multiple arguments). template static constexpr uint32_t bitMask(T x, Args... args) noexcept { return bitMask(x) | bitMask(args...); } //! Converts a boolean value `b` to zero or full mask (all bits set). template static constexpr DstT bitMaskFromBool(SrcT b) noexcept { typedef typename std::make_unsigned::type U; return DstT(U(0) - U(b)); } //! \cond namespace Internal { // Fills all trailing bits right from the first most significant bit set. static constexpr uint8_t fillTrailingBitsImpl(uint8_t x) noexcept { return or_shr(or_shr(or_shr(x, 1), 2), 4); } // Fills all trailing bits right from the first most significant bit set. static constexpr uint16_t fillTrailingBitsImpl(uint16_t x) noexcept { return or_shr(or_shr(or_shr(or_shr(x, 1), 2), 4), 8); } // Fills all trailing bits right from the first most significant bit set. static constexpr uint32_t fillTrailingBitsImpl(uint32_t x) noexcept { return or_shr(or_shr(or_shr(or_shr(or_shr(x, 1), 2), 4), 8), 16); } // Fills all trailing bits right from the first most significant bit set. static constexpr uint64_t fillTrailingBitsImpl(uint64_t x) noexcept { return or_shr(or_shr(or_shr(or_shr(or_shr(or_shr(x, 1), 2), 4), 8), 16), 32); } } //! \endcond // Fills all trailing bits right from the first most significant bit set. template static constexpr T fillTrailingBits(const T& x) noexcept { typedef typename std::make_unsigned::type U; return T(Internal::fillTrailingBitsImpl(U(x))); } // ============================================================================ // [asmjit::Support - CTZ] // ============================================================================ //! \cond namespace Internal { namespace { template struct BitScanData { T x; uint32_t n; }; template struct BitScanCalc { static constexpr BitScanData advanceLeft(const BitScanData& data, uint32_t n) noexcept { return BitScanData { data.x << n, data.n + n }; } static constexpr BitScanData advanceRight(const BitScanData& data, uint32_t n) noexcept { return BitScanData { data.x >> n, data.n + n }; } static constexpr BitScanData clz(const BitScanData& data) noexcept { return BitScanCalc::clz(advanceLeft(data, data.x & (allOnes() << (bitSizeOf() - N)) ? uint32_t(0) : N)); } static constexpr BitScanData ctz(const BitScanData& data) noexcept { return BitScanCalc::ctz(advanceRight(data, data.x & (allOnes() >> (bitSizeOf() - N)) ? uint32_t(0) : N)); } }; template struct BitScanCalc { static constexpr BitScanData clz(const BitScanData& ctx) noexcept { return BitScanData { 0, ctx.n - uint32_t(ctx.x >> (bitSizeOf() - 1)) }; } static constexpr BitScanData ctz(const BitScanData& ctx) noexcept { return BitScanData { 0, ctx.n - uint32_t(ctx.x & 0x1) }; } }; template constexpr uint32_t clzFallback(const T& x) noexcept { return BitScanCalc() / 2u>::clz(BitScanData{x, 1}).n; } template constexpr uint32_t ctzFallback(const T& x) noexcept { return BitScanCalc() / 2u>::ctz(BitScanData{x, 1}).n; } template constexpr uint32_t constClz(const T& x) noexcept { return clzFallback(asUInt(x)); } template constexpr uint32_t constCtz(const T& x) noexcept { return ctzFallback(asUInt(x)); } template inline uint32_t clzImpl(const T& x) noexcept { return constClz(x); } template inline uint32_t ctzImpl(const T& x) noexcept { return constCtz(x); } #if !defined(ASMJIT_NO_INTRINSICS) # if defined(__GNUC__) template<> inline uint32_t clzImpl(const uint32_t& x) noexcept { return uint32_t(__builtin_clz(x)); } template<> inline uint32_t clzImpl(const uint64_t& x) noexcept { return uint32_t(__builtin_clzll(x)); } template<> inline uint32_t ctzImpl(const uint32_t& x) noexcept { return uint32_t(__builtin_ctz(x)); } template<> inline uint32_t ctzImpl(const uint64_t& x) noexcept { return uint32_t(__builtin_ctzll(x)); } # elif defined(_MSC_VER) template<> inline uint32_t clzImpl(const uint32_t& x) noexcept { unsigned long i; _BitScanReverse(&i, x); return uint32_t(i ^ 31); } template<> inline uint32_t ctzImpl(const uint32_t& x) noexcept { unsigned long i; _BitScanForward(&i, x); return uint32_t(i); } # if ASMJIT_ARCH_X86 == 64 || ASMJIT_ARCH_ARM == 64 template<> inline uint32_t clzImpl(const uint64_t& x) noexcept { unsigned long i; _BitScanReverse64(&i, x); return uint32_t(i ^ 63); } template<> inline uint32_t ctzImpl(const uint64_t& x) noexcept { unsigned long i; _BitScanForward64(&i, x); return uint32_t(i); } # endif # endif #endif } // {anonymous} } // {Internal} //! \endcond //! Count leading zeros in `x` (returns a position of a first bit set in `x`). //! //! \note The input MUST NOT be zero, otherwise the result is undefined. template static inline uint32_t clz(T x) noexcept { return Internal::clzImpl(asUInt(x)); } //! Count leading zeros in `x` (constant expression). template static constexpr inline uint32_t constClz(T x) noexcept { return Internal::constClz(asUInt(x)); } //! Count trailing zeros in `x` (returns a position of a first bit set in `x`). //! //! \note The input MUST NOT be zero, otherwise the result is undefined. template static inline uint32_t ctz(T x) noexcept { return Internal::ctzImpl(asUInt(x)); } //! Count trailing zeros in `x` (constant expression). template static constexpr inline uint32_t constCtz(T x) noexcept { return Internal::constCtz(asUInt(x)); } // ============================================================================ // [asmjit::Support - PopCnt] // ============================================================================ // Based on the following resource: // http://graphics.stanford.edu/~seander/bithacks.html // // Alternatively, for a very small number of bits in `x`: // uint32_t n = 0; // while (x) { // x &= x - 1; // n++; // } // return n; //! \cond namespace Internal { static inline uint32_t constPopcntImpl(uint32_t x) noexcept { x = x - ((x >> 1) & 0x55555555u); x = (x & 0x33333333u) + ((x >> 2) & 0x33333333u); return (((x + (x >> 4)) & 0x0F0F0F0Fu) * 0x01010101u) >> 24; } static inline uint32_t constPopcntImpl(uint64_t x) noexcept { if (ASMJIT_ARCH_BITS >= 64) { x = x - ((x >> 1) & 0x5555555555555555u); x = (x & 0x3333333333333333u) + ((x >> 2) & 0x3333333333333333u); return uint32_t((((x + (x >> 4)) & 0x0F0F0F0F0F0F0F0Fu) * 0x0101010101010101u) >> 56); } else { return constPopcntImpl(uint32_t(x >> 32)) + constPopcntImpl(uint32_t(x & 0xFFFFFFFFu)); } } static inline uint32_t popcntImpl(uint32_t x) noexcept { #if defined(__GNUC__) return uint32_t(__builtin_popcount(x)); #else return constPopcntImpl(asUInt(x)); #endif } static inline uint32_t popcntImpl(uint64_t x) noexcept { #if defined(__GNUC__) return uint32_t(__builtin_popcountll(x)); #else return constPopcntImpl(asUInt(x)); #endif } } //! \endcond //! Calculates count of bits in `x`. template static inline uint32_t popcnt(T x) noexcept { return Internal::popcntImpl(asUInt(x)); } //! Calculates count of bits in `x` (useful in constant expressions). template static inline uint32_t constPopcnt(T x) noexcept { return Internal::constPopcntImpl(asUInt(x)); } // ============================================================================ // [asmjit::Support - Min/Max] // ============================================================================ // NOTE: These are constexpr `min()` and `max()` implementations that are not // exactly the same as `std::min()` and `std::max()`. The return value is not // a reference to `a` or `b` but it's a new value instead. template static constexpr T min(const T& a, const T& b) noexcept { return b < a ? b : a; } template static constexpr T min(const T& a, const T& b, Args&&... args) noexcept { return min(min(a, b), std::forward(args)...); } template static constexpr T max(const T& a, const T& b) noexcept { return a < b ? b : a; } template static constexpr T max(const T& a, const T& b, Args&&... args) noexcept { return max(max(a, b), std::forward(args)...); } // ============================================================================ // [asmjit::Support - Immediate Helpers] // ============================================================================ namespace Internal { template struct ImmConv { static inline int64_t fromT(const T& x) noexcept { return int64_t(x); } static inline T toT(int64_t x) noexcept { return T(uint64_t(x) & Support::allOnes::type>()); } }; template struct ImmConv { static inline int64_t fromT(const T& x) noexcept { return int64_t(bitCast(double(x))); } static inline T toT(int64_t x) noexcept { return T(bitCast(x)); } }; } template static inline int64_t immediateFromT(const T& x) noexcept { return Internal::ImmConv::value>::fromT(x); } template static inline T immediateToT(int64_t x) noexcept { return Internal::ImmConv::value>::toT(x); } // ============================================================================ // [asmjit::Support - Overflow Arithmetic] // ============================================================================ //! \cond namespace Internal { template ASMJIT_INLINE T addOverflowFallback(T x, T y, FastUInt8* of) noexcept { typedef typename std::make_unsigned::type U; U result = U(x) + U(y); *of = FastUInt8(*of | FastUInt8(isUnsigned() ? result < U(x) : T((U(x) ^ ~U(y)) & (U(x) ^ result)) < 0)); return T(result); } template ASMJIT_INLINE T subOverflowFallback(T x, T y, FastUInt8* of) noexcept { typedef typename std::make_unsigned::type U; U result = U(x) - U(y); *of = FastUInt8(*of | FastUInt8(isUnsigned() ? result > U(x) : T((U(x) ^ U(y)) & (U(x) ^ result)) < 0)); return T(result); } template ASMJIT_INLINE T mulOverflowFallback(T x, T y, FastUInt8* of) noexcept { typedef typename Internal::StdInt()>::Type I; typedef typename std::make_unsigned::type U; U mask = allOnes(); if (std::is_signed::value) { U prod = U(I(x)) * U(I(y)); *of = FastUInt8(*of | FastUInt8(I(prod) < I(std::numeric_limits::lowest()) || I(prod) > I(std::numeric_limits::max()))); return T(I(prod & mask)); } else { U prod = U(x) * U(y); *of = FastUInt8(*of | FastUInt8((prod & ~mask) != 0)); return T(prod & mask); } } template<> ASMJIT_INLINE int64_t mulOverflowFallback(int64_t x, int64_t y, FastUInt8* of) noexcept { int64_t result = int64_t(uint64_t(x) * uint64_t(y)); *of = FastUInt8(*of | FastUInt8(x && (result / x != y))); return result; } template<> ASMJIT_INLINE uint64_t mulOverflowFallback(uint64_t x, uint64_t y, FastUInt8* of) noexcept { uint64_t result = x * y; *of = FastUInt8(*of | FastUInt8(y != 0 && allOnes() / y < x)); return result; } // These can be specialized. template ASMJIT_INLINE T addOverflowImpl(const T& x, const T& y, FastUInt8* of) noexcept { return addOverflowFallback(x, y, of); } template ASMJIT_INLINE T subOverflowImpl(const T& x, const T& y, FastUInt8* of) noexcept { return subOverflowFallback(x, y, of); } template ASMJIT_INLINE T mulOverflowImpl(const T& x, const T& y, FastUInt8* of) noexcept { return mulOverflowFallback(x, y, of); } #if defined(__GNUC__) && !defined(ASMJIT_NO_INTRINSICS) #if defined(__clang__) || __GNUC__ >= 5 #define ASMJIT_ARITH_OVERFLOW_SPECIALIZE(FUNC, T, RESULT_T, BUILTIN) \ template<> \ ASMJIT_INLINE T FUNC(const T& x, const T& y, FastUInt8* of) noexcept { \ RESULT_T result; \ *of = FastUInt8(*of | (BUILTIN((RESULT_T)x, (RESULT_T)y, &result))); \ return T(result); \ } ASMJIT_ARITH_OVERFLOW_SPECIALIZE(addOverflowImpl, int32_t , int , __builtin_sadd_overflow ) ASMJIT_ARITH_OVERFLOW_SPECIALIZE(addOverflowImpl, uint32_t, unsigned int , __builtin_uadd_overflow ) ASMJIT_ARITH_OVERFLOW_SPECIALIZE(addOverflowImpl, int64_t , long long , __builtin_saddll_overflow) ASMJIT_ARITH_OVERFLOW_SPECIALIZE(addOverflowImpl, uint64_t, unsigned long long, __builtin_uaddll_overflow) ASMJIT_ARITH_OVERFLOW_SPECIALIZE(subOverflowImpl, int32_t , int , __builtin_ssub_overflow ) ASMJIT_ARITH_OVERFLOW_SPECIALIZE(subOverflowImpl, uint32_t, unsigned int , __builtin_usub_overflow ) ASMJIT_ARITH_OVERFLOW_SPECIALIZE(subOverflowImpl, int64_t , long long , __builtin_ssubll_overflow) ASMJIT_ARITH_OVERFLOW_SPECIALIZE(subOverflowImpl, uint64_t, unsigned long long, __builtin_usubll_overflow) ASMJIT_ARITH_OVERFLOW_SPECIALIZE(mulOverflowImpl, int32_t , int , __builtin_smul_overflow ) ASMJIT_ARITH_OVERFLOW_SPECIALIZE(mulOverflowImpl, uint32_t, unsigned int , __builtin_umul_overflow ) ASMJIT_ARITH_OVERFLOW_SPECIALIZE(mulOverflowImpl, int64_t , long long , __builtin_smulll_overflow) ASMJIT_ARITH_OVERFLOW_SPECIALIZE(mulOverflowImpl, uint64_t, unsigned long long, __builtin_umulll_overflow) #undef ASMJIT_ARITH_OVERFLOW_SPECIALIZE #endif #endif // There is a bug in MSVC that makes these specializations unusable, maybe in the future... #if defined(_MSC_VER) && 0 #define ASMJIT_ARITH_OVERFLOW_SPECIALIZE(FUNC, T, ALT_T, BUILTIN) \ template<> \ ASMJIT_INLINE T FUNC(T x, T y, FastUInt8* of) noexcept { \ ALT_T result; \ *of = FastUInt8(*of | BUILTIN(0, (ALT_T)x, (ALT_T)y, &result)); \ return T(result); \ } ASMJIT_ARITH_OVERFLOW_SPECIALIZE(addOverflowImpl, uint32_t, unsigned int , _addcarry_u32 ) ASMJIT_ARITH_OVERFLOW_SPECIALIZE(subOverflowImpl, uint32_t, unsigned int , _subborrow_u32) #if ARCH_BITS >= 64 ASMJIT_ARITH_OVERFLOW_SPECIALIZE(addOverflowImpl, uint64_t, unsigned __int64 , _addcarry_u64 ) ASMJIT_ARITH_OVERFLOW_SPECIALIZE(subOverflowImpl, uint64_t, unsigned __int64 , _subborrow_u64) #endif #undef ASMJIT_ARITH_OVERFLOW_SPECIALIZE #endif } // {Internal} //! \endcond template static ASMJIT_INLINE T addOverflow(const T& x, const T& y, FastUInt8* of) noexcept { return T(Internal::addOverflowImpl(asStdInt(x), asStdInt(y), of)); } template static ASMJIT_INLINE T subOverflow(const T& x, const T& y, FastUInt8* of) noexcept { return T(Internal::subOverflowImpl(asStdInt(x), asStdInt(y), of)); } template static ASMJIT_INLINE T mulOverflow(const T& x, const T& y, FastUInt8* of) noexcept { return T(Internal::mulOverflowImpl(asStdInt(x), asStdInt(y), of)); } // ============================================================================ // [asmjit::Support - Alignment] // ============================================================================ template static constexpr bool isAligned(X base, Y alignment) noexcept { typedef typename Internal::StdInt::Type U; return ((U)base % (U)alignment) == 0; } //! Tests whether the `x` is a power of two (only one bit is set). template static constexpr bool isPowerOf2(T x) noexcept { typedef typename std::make_unsigned::type U; return x && !(U(x) & (U(x) - U(1))); } template static constexpr X alignUp(X x, Y alignment) noexcept { typedef typename Internal::StdInt::Type U; return (X)( ((U)x + ((U)(alignment) - 1u)) & ~((U)(alignment) - 1u) ); } template static constexpr T alignUpPowerOf2(T x) noexcept { typedef typename Internal::StdInt::Type U; return (T)(fillTrailingBits(U(x) - 1u) + 1u); } //! Returns either zero or a positive difference between `base` and `base` when //! aligned to `alignment`. template static constexpr typename Internal::StdInt::Type alignUpDiff(X base, Y alignment) noexcept { typedef typename Internal::StdInt::Type U; return alignUp(U(base), alignment) - U(base); } template static constexpr X alignDown(X x, Y alignment) noexcept { typedef typename Internal::StdInt::Type U; return (X)( (U)x & ~((U)(alignment) - 1u) ); } // ============================================================================ // [asmjit::Support - NumGranularized] // ============================================================================ //! Calculates the number of elements that would be required if `base` is //! granularized by `granularity`. This function can be used to calculate //! the number of BitWords to represent N bits, for example. template static constexpr X numGranularized(X base, Y granularity) noexcept { typedef typename Internal::StdInt::Type U; return X((U(base) + U(granularity) - 1) / U(granularity)); } // ============================================================================ // [asmjit::Support - IsBetween] // ============================================================================ //! Checks whether `x` is greater than or equal to `a` and lesser than or equal to `b`. template static constexpr bool isBetween(const T& x, const T& a, const T& b) noexcept { return x >= a && x <= b; } // ============================================================================ // [asmjit::Support - IsInt / IsUInt] // ============================================================================ //! Checks whether the given integer `x` can be casted to a 4-bit signed integer. template static constexpr bool isInt4(T x) noexcept { typedef typename std::make_signed::type S; typedef typename std::make_unsigned::type U; return std::is_signed::value ? isBetween(S(x), -8, 7) : U(x) <= U(7u); } //! Checks whether the given integer `x` can be casted to a 7-bit signed integer. template static constexpr bool isInt7(T x) noexcept { typedef typename std::make_signed::type S; typedef typename std::make_unsigned::type U; return std::is_signed::value ? isBetween(S(x), -64, 63) : U(x) <= U(63u); } //! Checks whether the given integer `x` can be casted to an 8-bit signed integer. template static constexpr bool isInt8(T x) noexcept { typedef typename std::make_signed::type S; typedef typename std::make_unsigned::type U; return std::is_signed::value ? sizeof(T) <= 1 || isBetween(S(x), -128, 127) : U(x) <= U(127u); } //! Checks whether the given integer `x` can be casted to a 9-bit signed integer. template static constexpr bool isInt9(T x) noexcept { typedef typename std::make_signed::type S; typedef typename std::make_unsigned::type U; return std::is_signed::value ? sizeof(T) <= 1 || isBetween(S(x), -256, 255) : sizeof(T) <= 1 || U(x) <= U(255u); } //! Checks whether the given integer `x` can be casted to a 10-bit signed integer. template static constexpr bool isInt10(T x) noexcept { typedef typename std::make_signed::type S; typedef typename std::make_unsigned::type U; return std::is_signed::value ? sizeof(T) <= 1 || isBetween(S(x), -512, 511) : sizeof(T) <= 1 || U(x) <= U(511u); } //! Checks whether the given integer `x` can be casted to a 16-bit signed integer. template static constexpr bool isInt16(T x) noexcept { typedef typename std::make_signed::type S; typedef typename std::make_unsigned::type U; return std::is_signed::value ? sizeof(T) <= 2 || isBetween(S(x), -32768, 32767) : sizeof(T) <= 1 || U(x) <= U(32767u); } //! Checks whether the given integer `x` can be casted to a 32-bit signed integer. template static constexpr bool isInt32(T x) noexcept { typedef typename std::make_signed::type S; typedef typename std::make_unsigned::type U; return std::is_signed::value ? sizeof(T) <= 4 || isBetween(S(x), -2147483647 - 1, 2147483647) : sizeof(T) <= 2 || U(x) <= U(2147483647u); } //! Checks whether the given integer `x` can be casted to a 4-bit unsigned integer. template static constexpr bool isUInt4(T x) noexcept { typedef typename std::make_unsigned::type U; return std::is_signed::value ? x >= T(0) && x <= T(15) : U(x) <= U(15u); } //! Checks whether the given integer `x` can be casted to an 8-bit unsigned integer. template static constexpr bool isUInt8(T x) noexcept { typedef typename std::make_unsigned::type U; return std::is_signed::value ? (sizeof(T) <= 1 || T(x) <= T(255)) && x >= T(0) : (sizeof(T) <= 1 || U(x) <= U(255u)); } //! Checks whether the given integer `x` can be casted to a 12-bit unsigned integer (ARM specific). template static constexpr bool isUInt12(T x) noexcept { typedef typename std::make_unsigned::type U; return std::is_signed::value ? (sizeof(T) <= 1 || T(x) <= T(4095)) && x >= T(0) : (sizeof(T) <= 1 || U(x) <= U(4095u)); } //! Checks whether the given integer `x` can be casted to a 16-bit unsigned integer. template static constexpr bool isUInt16(T x) noexcept { typedef typename std::make_unsigned::type U; return std::is_signed::value ? (sizeof(T) <= 2 || T(x) <= T(65535)) && x >= T(0) : (sizeof(T) <= 2 || U(x) <= U(65535u)); } //! Checks whether the given integer `x` can be casted to a 32-bit unsigned integer. template static constexpr bool isUInt32(T x) noexcept { typedef typename std::make_unsigned::type U; return std::is_signed::value ? (sizeof(T) <= 4 || T(x) <= T(4294967295u)) && x >= T(0) : (sizeof(T) <= 4 || U(x) <= U(4294967295u)); } //! Checks whether the given integer `x` can be casted to a 32-bit unsigned integer. template static constexpr bool isIntOrUInt32(T x) noexcept { return sizeof(T) <= 4 ? true : (uint32_t(uint64_t(x) >> 32) + 1u) <= 1u; } static bool inline isEncodableOffset32(int32_t offset, uint32_t nBits) noexcept { uint32_t nRev = 32 - nBits; return Support::sar(Support::shl(offset, nRev), nRev) == offset; } static bool inline isEncodableOffset64(int64_t offset, uint32_t nBits) noexcept { uint32_t nRev = 64 - nBits; return Support::sar(Support::shl(offset, nRev), nRev) == offset; } // ============================================================================ // [asmjit::Support - ByteSwap] // ============================================================================ static constexpr uint32_t byteswap32(uint32_t x) noexcept { return (x << 24) | (x >> 24) | ((x << 8) & 0x00FF0000u) | ((x >> 8) & 0x0000FF00); } // ============================================================================ // [asmjit::Support - BytePack / Unpack] // ============================================================================ //! Pack four 8-bit integer into a 32-bit integer as it is an array of `{b0,b1,b2,b3}`. static constexpr uint32_t bytepack32_4x8(uint32_t a, uint32_t b, uint32_t c, uint32_t d) noexcept { return ASMJIT_ARCH_LE ? (a | (b << 8) | (c << 16) | (d << 24)) : (d | (c << 8) | (b << 16) | (a << 24)); } template static constexpr uint32_t unpackU32At0(T x) noexcept { return ASMJIT_ARCH_LE ? uint32_t(uint64_t(x) & 0xFFFFFFFFu) : uint32_t(uint64_t(x) >> 32); } template static constexpr uint32_t unpackU32At1(T x) noexcept { return ASMJIT_ARCH_BE ? uint32_t(uint64_t(x) & 0xFFFFFFFFu) : uint32_t(uint64_t(x) >> 32); } // ============================================================================ // [asmjit::Support - Position of byte (in bit-shift)] // ============================================================================ static inline uint32_t byteShiftOfDWordStruct(uint32_t index) noexcept { return ASMJIT_ARCH_LE ? index * 8 : (uint32_t(sizeof(uint32_t)) - 1u - index) * 8; } // ============================================================================ // [asmjit::Support - String Utilities] // ============================================================================ template static constexpr T asciiToLower(T c) noexcept { return T(c ^ T(T(c >= T('A') && c <= T('Z')) << 5)); } template static constexpr T asciiToUpper(T c) noexcept { return T(c ^ T(T(c >= T('a') && c <= T('z')) << 5)); } static ASMJIT_INLINE size_t strLen(const char* s, size_t maxSize) noexcept { size_t i = 0; while (i < maxSize && s[i] != '\0') i++; return i; } static constexpr uint32_t hashRound(uint32_t hash, uint32_t c) noexcept { return hash * 65599 + c; } // Gets a hash of the given string `data` of size `size`. Size must be valid // as this function doesn't check for a null terminator and allows it in the // middle of the string. static inline uint32_t hashString(const char* data, size_t size) noexcept { uint32_t hashCode = 0; for (uint32_t i = 0; i < size; i++) hashCode = hashRound(hashCode, uint8_t(data[i])); return hashCode; } static ASMJIT_INLINE const char* findPackedString(const char* p, uint32_t id) noexcept { uint32_t i = 0; while (i < id) { while (p[0]) p++; p++; i++; } return p; } //! Compares two instruction names. //! //! `a` is a null terminated instruction name from arch-specific `nameData[]` //! table. `b` is a possibly non-null terminated instruction name passed to //! `InstAPI::stringToInstId()`. static ASMJIT_INLINE int cmpInstName(const char* a, const char* b, size_t size) noexcept { for (size_t i = 0; i < size; i++) { int c = int(uint8_t(a[i])) - int(uint8_t(b[i])); if (c != 0) return c; } return int(uint8_t(a[size])); } // ============================================================================ // [asmjit::Support - Read / Write] // ============================================================================ static inline uint32_t readU8(const void* p) noexcept { return uint32_t(static_cast(p)[0]); } static inline int32_t readI8(const void* p) noexcept { return int32_t(static_cast(p)[0]); } template static inline uint32_t readU16x(const void* p) noexcept { if (BO == ByteOrder::kNative && (kUnalignedAccess16 || Alignment >= 2)) { typedef typename Internal::AlignedInt::T U16AlignedToN; return uint32_t(static_cast(p)[0]); } else { uint32_t hi = readU8(static_cast(p) + (BO == ByteOrder::kLE ? 1 : 0)); uint32_t lo = readU8(static_cast(p) + (BO == ByteOrder::kLE ? 0 : 1)); return shl(hi, 8) | lo; } } template static inline int32_t readI16x(const void* p) noexcept { if (BO == ByteOrder::kNative && (kUnalignedAccess16 || Alignment >= 2)) { typedef typename Internal::AlignedInt::T U16AlignedToN; return int32_t(int16_t(static_cast(p)[0])); } else { int32_t hi = readI8(static_cast(p) + (BO == ByteOrder::kLE ? 1 : 0)); uint32_t lo = readU8(static_cast(p) + (BO == ByteOrder::kLE ? 0 : 1)); return shl(hi, 8) | int32_t(lo); } } template static inline uint32_t readU24u(const void* p) noexcept { uint32_t b0 = readU8(static_cast(p) + (BO == ByteOrder::kLE ? 2 : 0)); uint32_t b1 = readU8(static_cast(p) + (BO == ByteOrder::kLE ? 1 : 1)); uint32_t b2 = readU8(static_cast(p) + (BO == ByteOrder::kLE ? 0 : 2)); return shl(b0, 16) | shl(b1, 8) | b2; } template static inline uint32_t readU32x(const void* p) noexcept { if (kUnalignedAccess32 || Alignment >= 4) { typedef typename Internal::AlignedInt::T U32AlignedToN; uint32_t x = static_cast(p)[0]; return BO == ByteOrder::kNative ? x : byteswap32(x); } else { uint32_t hi = readU16x= 2 ? size_t(2) : Alignment>(static_cast(p) + (BO == ByteOrder::kLE ? 2 : 0)); uint32_t lo = readU16x= 2 ? size_t(2) : Alignment>(static_cast(p) + (BO == ByteOrder::kLE ? 0 : 2)); return shl(hi, 16) | lo; } } template static inline uint64_t readU64x(const void* p) noexcept { if (BO == ByteOrder::kNative && (kUnalignedAccess64 || Alignment >= 8)) { typedef typename Internal::AlignedInt::T U64AlignedToN; return static_cast(p)[0]; } else { uint32_t hi = readU32x= 4 ? size_t(4) : Alignment>(static_cast(p) + (BO == ByteOrder::kLE ? 4 : 0)); uint32_t lo = readU32x= 4 ? size_t(4) : Alignment>(static_cast(p) + (BO == ByteOrder::kLE ? 0 : 4)); return shl(uint64_t(hi), 32) | lo; } } template static inline int32_t readI32x(const void* p) noexcept { return int32_t(readU32x(p)); } template static inline int64_t readI64x(const void* p) noexcept { return int64_t(readU64x(p)); } template static inline int32_t readI16xLE(const void* p) noexcept { return readI16x(p); } template static inline int32_t readI16xBE(const void* p) noexcept { return readI16x(p); } template static inline uint32_t readU16xLE(const void* p) noexcept { return readU16x(p); } template static inline uint32_t readU16xBE(const void* p) noexcept { return readU16x(p); } template static inline int32_t readI32xLE(const void* p) noexcept { return readI32x(p); } template static inline int32_t readI32xBE(const void* p) noexcept { return readI32x(p); } template static inline uint32_t readU32xLE(const void* p) noexcept { return readU32x(p); } template static inline uint32_t readU32xBE(const void* p) noexcept { return readU32x(p); } template static inline int64_t readI64xLE(const void* p) noexcept { return readI64x(p); } template static inline int64_t readI64xBE(const void* p) noexcept { return readI64x(p); } template static inline uint64_t readU64xLE(const void* p) noexcept { return readU64x(p); } template static inline uint64_t readU64xBE(const void* p) noexcept { return readU64x(p); } static inline int32_t readI16a(const void* p) noexcept { return readI16x(p); } static inline int32_t readI16u(const void* p) noexcept { return readI16x(p); } static inline uint32_t readU16a(const void* p) noexcept { return readU16x(p); } static inline uint32_t readU16u(const void* p) noexcept { return readU16x(p); } static inline int32_t readI16aLE(const void* p) noexcept { return readI16xLE<2>(p); } static inline int32_t readI16uLE(const void* p) noexcept { return readI16xLE<1>(p); } static inline uint32_t readU16aLE(const void* p) noexcept { return readU16xLE<2>(p); } static inline uint32_t readU16uLE(const void* p) noexcept { return readU16xLE<1>(p); } static inline int32_t readI16aBE(const void* p) noexcept { return readI16xBE<2>(p); } static inline int32_t readI16uBE(const void* p) noexcept { return readI16xBE<1>(p); } static inline uint32_t readU16aBE(const void* p) noexcept { return readU16xBE<2>(p); } static inline uint32_t readU16uBE(const void* p) noexcept { return readU16xBE<1>(p); } static inline uint32_t readU24uLE(const void* p) noexcept { return readU24u(p); } static inline uint32_t readU24uBE(const void* p) noexcept { return readU24u(p); } static inline int32_t readI32a(const void* p) noexcept { return readI32x(p); } static inline int32_t readI32u(const void* p) noexcept { return readI32x(p); } static inline uint32_t readU32a(const void* p) noexcept { return readU32x(p); } static inline uint32_t readU32u(const void* p) noexcept { return readU32x(p); } static inline int32_t readI32aLE(const void* p) noexcept { return readI32xLE<4>(p); } static inline int32_t readI32uLE(const void* p) noexcept { return readI32xLE<1>(p); } static inline uint32_t readU32aLE(const void* p) noexcept { return readU32xLE<4>(p); } static inline uint32_t readU32uLE(const void* p) noexcept { return readU32xLE<1>(p); } static inline int32_t readI32aBE(const void* p) noexcept { return readI32xBE<4>(p); } static inline int32_t readI32uBE(const void* p) noexcept { return readI32xBE<1>(p); } static inline uint32_t readU32aBE(const void* p) noexcept { return readU32xBE<4>(p); } static inline uint32_t readU32uBE(const void* p) noexcept { return readU32xBE<1>(p); } static inline int64_t readI64a(const void* p) noexcept { return readI64x(p); } static inline int64_t readI64u(const void* p) noexcept { return readI64x(p); } static inline uint64_t readU64a(const void* p) noexcept { return readU64x(p); } static inline uint64_t readU64u(const void* p) noexcept { return readU64x(p); } static inline int64_t readI64aLE(const void* p) noexcept { return readI64xLE<8>(p); } static inline int64_t readI64uLE(const void* p) noexcept { return readI64xLE<1>(p); } static inline uint64_t readU64aLE(const void* p) noexcept { return readU64xLE<8>(p); } static inline uint64_t readU64uLE(const void* p) noexcept { return readU64xLE<1>(p); } static inline int64_t readI64aBE(const void* p) noexcept { return readI64xBE<8>(p); } static inline int64_t readI64uBE(const void* p) noexcept { return readI64xBE<1>(p); } static inline uint64_t readU64aBE(const void* p) noexcept { return readU64xBE<8>(p); } static inline uint64_t readU64uBE(const void* p) noexcept { return readU64xBE<1>(p); } static inline void writeU8(void* p, uint32_t x) noexcept { static_cast(p)[0] = uint8_t(x & 0xFFu); } static inline void writeI8(void* p, int32_t x) noexcept { static_cast(p)[0] = uint8_t(x & 0xFF); } template static inline void writeU16x(void* p, uint32_t x) noexcept { if (BO == ByteOrder::kNative && (kUnalignedAccess16 || Alignment >= 2)) { typedef typename Internal::AlignedInt::T U16AlignedToN; static_cast(p)[0] = uint16_t(x & 0xFFFFu); } else { static_cast(p)[0] = uint8_t((x >> (BO == ByteOrder::kLE ? 0 : 8)) & 0xFFu); static_cast(p)[1] = uint8_t((x >> (BO == ByteOrder::kLE ? 8 : 0)) & 0xFFu); } } template static inline void writeU24u(void* p, uint32_t v) noexcept { static_cast(p)[0] = uint8_t((v >> (BO == ByteOrder::kLE ? 0 : 16)) & 0xFFu); static_cast(p)[1] = uint8_t((v >> (BO == ByteOrder::kLE ? 8 : 8)) & 0xFFu); static_cast(p)[2] = uint8_t((v >> (BO == ByteOrder::kLE ? 16 : 0)) & 0xFFu); } template static inline void writeU32x(void* p, uint32_t x) noexcept { if (kUnalignedAccess32 || Alignment >= 4) { typedef typename Internal::AlignedInt::T U32AlignedToN; static_cast(p)[0] = (BO == ByteOrder::kNative) ? x : Support::byteswap32(x); } else { writeU16x= 2 ? size_t(2) : Alignment>(static_cast(p) + 0, x >> (BO == ByteOrder::kLE ? 0 : 16)); writeU16x= 2 ? size_t(2) : Alignment>(static_cast(p) + 2, x >> (BO == ByteOrder::kLE ? 16 : 0)); } } template static inline void writeU64x(void* p, uint64_t x) noexcept { if (BO == ByteOrder::kNative && (kUnalignedAccess64 || Alignment >= 8)) { typedef typename Internal::AlignedInt::T U64AlignedToN; static_cast(p)[0] = x; } else { writeU32x= 4 ? size_t(4) : Alignment>(static_cast(p) + 0, uint32_t((x >> (BO == ByteOrder::kLE ? 0 : 32)) & 0xFFFFFFFFu)); writeU32x= 4 ? size_t(4) : Alignment>(static_cast(p) + 4, uint32_t((x >> (BO == ByteOrder::kLE ? 32 : 0)) & 0xFFFFFFFFu)); } } template static inline void writeI16x(void* p, int32_t x) noexcept { writeU16x(p, uint32_t(x)); } template static inline void writeI32x(void* p, int32_t x) noexcept { writeU32x(p, uint32_t(x)); } template static inline void writeI64x(void* p, int64_t x) noexcept { writeU64x(p, uint64_t(x)); } template static inline void writeI16xLE(void* p, int32_t x) noexcept { writeI16x(p, x); } template static inline void writeI16xBE(void* p, int32_t x) noexcept { writeI16x(p, x); } template static inline void writeU16xLE(void* p, uint32_t x) noexcept { writeU16x(p, x); } template static inline void writeU16xBE(void* p, uint32_t x) noexcept { writeU16x(p, x); } template static inline void writeI32xLE(void* p, int32_t x) noexcept { writeI32x(p, x); } template static inline void writeI32xBE(void* p, int32_t x) noexcept { writeI32x(p, x); } template static inline void writeU32xLE(void* p, uint32_t x) noexcept { writeU32x(p, x); } template static inline void writeU32xBE(void* p, uint32_t x) noexcept { writeU32x(p, x); } template static inline void writeI64xLE(void* p, int64_t x) noexcept { writeI64x(p, x); } template static inline void writeI64xBE(void* p, int64_t x) noexcept { writeI64x(p, x); } template static inline void writeU64xLE(void* p, uint64_t x) noexcept { writeU64x(p, x); } template static inline void writeU64xBE(void* p, uint64_t x) noexcept { writeU64x(p, x); } static inline void writeI16a(void* p, int32_t x) noexcept { writeI16x(p, x); } static inline void writeI16u(void* p, int32_t x) noexcept { writeI16x(p, x); } static inline void writeU16a(void* p, uint32_t x) noexcept { writeU16x(p, x); } static inline void writeU16u(void* p, uint32_t x) noexcept { writeU16x(p, x); } static inline void writeI16aLE(void* p, int32_t x) noexcept { writeI16xLE<2>(p, x); } static inline void writeI16uLE(void* p, int32_t x) noexcept { writeI16xLE<1>(p, x); } static inline void writeU16aLE(void* p, uint32_t x) noexcept { writeU16xLE<2>(p, x); } static inline void writeU16uLE(void* p, uint32_t x) noexcept { writeU16xLE<1>(p, x); } static inline void writeI16aBE(void* p, int32_t x) noexcept { writeI16xBE<2>(p, x); } static inline void writeI16uBE(void* p, int32_t x) noexcept { writeI16xBE<1>(p, x); } static inline void writeU16aBE(void* p, uint32_t x) noexcept { writeU16xBE<2>(p, x); } static inline void writeU16uBE(void* p, uint32_t x) noexcept { writeU16xBE<1>(p, x); } static inline void writeU24uLE(void* p, uint32_t v) noexcept { writeU24u(p, v); } static inline void writeU24uBE(void* p, uint32_t v) noexcept { writeU24u(p, v); } static inline void writeI32a(void* p, int32_t x) noexcept { writeI32x(p, x); } static inline void writeI32u(void* p, int32_t x) noexcept { writeI32x(p, x); } static inline void writeU32a(void* p, uint32_t x) noexcept { writeU32x(p, x); } static inline void writeU32u(void* p, uint32_t x) noexcept { writeU32x(p, x); } static inline void writeI32aLE(void* p, int32_t x) noexcept { writeI32xLE<4>(p, x); } static inline void writeI32uLE(void* p, int32_t x) noexcept { writeI32xLE<1>(p, x); } static inline void writeU32aLE(void* p, uint32_t x) noexcept { writeU32xLE<4>(p, x); } static inline void writeU32uLE(void* p, uint32_t x) noexcept { writeU32xLE<1>(p, x); } static inline void writeI32aBE(void* p, int32_t x) noexcept { writeI32xBE<4>(p, x); } static inline void writeI32uBE(void* p, int32_t x) noexcept { writeI32xBE<1>(p, x); } static inline void writeU32aBE(void* p, uint32_t x) noexcept { writeU32xBE<4>(p, x); } static inline void writeU32uBE(void* p, uint32_t x) noexcept { writeU32xBE<1>(p, x); } static inline void writeI64a(void* p, int64_t x) noexcept { writeI64x(p, x); } static inline void writeI64u(void* p, int64_t x) noexcept { writeI64x(p, x); } static inline void writeU64a(void* p, uint64_t x) noexcept { writeU64x(p, x); } static inline void writeU64u(void* p, uint64_t x) noexcept { writeU64x(p, x); } static inline void writeI64aLE(void* p, int64_t x) noexcept { writeI64xLE<8>(p, x); } static inline void writeI64uLE(void* p, int64_t x) noexcept { writeI64xLE<1>(p, x); } static inline void writeU64aLE(void* p, uint64_t x) noexcept { writeU64xLE<8>(p, x); } static inline void writeU64uLE(void* p, uint64_t x) noexcept { writeU64xLE<1>(p, x); } static inline void writeI64aBE(void* p, int64_t x) noexcept { writeI64xBE<8>(p, x); } static inline void writeI64uBE(void* p, int64_t x) noexcept { writeI64xBE<1>(p, x); } static inline void writeU64aBE(void* p, uint64_t x) noexcept { writeU64xBE<8>(p, x); } static inline void writeU64uBE(void* p, uint64_t x) noexcept { writeU64xBE<1>(p, x); } // ============================================================================ // [asmjit::Support - Operators] // ============================================================================ //! \cond INTERNAL struct Set { template static inline T op(T x, T y) noexcept { DebugUtils::unused(x); return y; } }; struct SetNot { template static inline T op(T x, T y) noexcept { DebugUtils::unused(x); return ~y; } }; struct And { template static inline T op(T x, T y) noexcept { return x & y; } }; struct AndNot { template static inline T op(T x, T y) noexcept { return x & ~y; } }; struct NotAnd { template static inline T op(T x, T y) noexcept { return ~x & y; } }; struct Or { template static inline T op(T x, T y) noexcept { return x | y; } }; struct Xor { template static inline T op(T x, T y) noexcept { return x ^ y; } }; struct Add { template static inline T op(T x, T y) noexcept { return x + y; } }; struct Sub { template static inline T op(T x, T y) noexcept { return x - y; } }; struct Min { template static inline T op(T x, T y) noexcept { return min(x, y); } }; struct Max { template static inline T op(T x, T y) noexcept { return max(x, y); } }; //! \endcond // ============================================================================ // [asmjit::Support - BitWordIterator] // ============================================================================ //! Iterates over each bit in a number which is set to 1. //! //! Example of use: //! //! ``` //! uint32_t bitsToIterate = 0x110F; //! Support::BitWordIterator it(bitsToIterate); //! //! while (it.hasNext()) { //! uint32_t bitIndex = it.next(); //! std::printf("Bit at %u is set\n", unsigned(bitIndex)); //! } //! ``` template class BitWordIterator { public: inline explicit BitWordIterator(T bitWord) noexcept : _bitWord(bitWord) {} inline void init(T bitWord) noexcept { _bitWord = bitWord; } inline bool hasNext() const noexcept { return _bitWord != 0; } inline uint32_t next() noexcept { ASMJIT_ASSERT(_bitWord != 0); uint32_t index = ctz(_bitWord); _bitWord ^= T(1u) << index; return index; } T _bitWord; }; // ============================================================================ // [asmjit::Support - BitWordFlipIterator] // ============================================================================ template class BitWordFlipIterator { public: inline explicit BitWordFlipIterator(T bitWord) noexcept : _bitWord(bitWord) {} inline void init(T bitWord) noexcept { _bitWord = bitWord; } inline bool hasNext() const noexcept { return _bitWord != 0; } inline uint32_t nextAndFlip() noexcept { ASMJIT_ASSERT(_bitWord != 0); uint32_t index = ctz(_bitWord); _bitWord ^= T(1u) << index; return index; } T _bitWord; T _xorMask; }; // ============================================================================ // [asmjit::Support - BitVectorOps] // ============================================================================ //! \cond namespace Internal { template static inline void bitVectorOp(T* buf, size_t index, size_t count) noexcept { if (count == 0) return; const size_t kTSizeInBits = bitSizeOf(); size_t vecIndex = index / kTSizeInBits; // T[] size_t bitIndex = index % kTSizeInBits; // T[][] buf += vecIndex; // The first BitWord requires special handling to preserve bits outside the fill region. const T kFillMask = allOnes(); size_t firstNBits = min(kTSizeInBits - bitIndex, count); buf[0] = OperatorT::op(buf[0], (kFillMask >> (kTSizeInBits - firstNBits)) << bitIndex); buf++; count -= firstNBits; // All bits between the first and last affected BitWords can be just filled. while (count >= kTSizeInBits) { buf[0] = FullWordOpT::op(buf[0], kFillMask); buf++; count -= kTSizeInBits; } // The last BitWord requires special handling as well if (count) buf[0] = OperatorT::op(buf[0], kFillMask >> (kTSizeInBits - count)); } } //! \endcond //! Sets bit in a bit-vector `buf` at `index`. template static inline bool bitVectorGetBit(T* buf, size_t index) noexcept { const size_t kTSizeInBits = bitSizeOf(); size_t vecIndex = index / kTSizeInBits; size_t bitIndex = index % kTSizeInBits; return bool((buf[vecIndex] >> bitIndex) & 0x1u); } //! Sets bit in a bit-vector `buf` at `index` to `value`. template static inline void bitVectorSetBit(T* buf, size_t index, bool value) noexcept { const size_t kTSizeInBits = bitSizeOf(); size_t vecIndex = index / kTSizeInBits; size_t bitIndex = index % kTSizeInBits; T bitMask = T(1u) << bitIndex; if (value) buf[vecIndex] |= bitMask; else buf[vecIndex] &= ~bitMask; } //! Sets bit in a bit-vector `buf` at `index` to `value`. template static inline void bitVectorFlipBit(T* buf, size_t index) noexcept { const size_t kTSizeInBits = bitSizeOf(); size_t vecIndex = index / kTSizeInBits; size_t bitIndex = index % kTSizeInBits; T bitMask = T(1u) << bitIndex; buf[vecIndex] ^= bitMask; } //! Fills `count` bits in bit-vector `buf` starting at bit-index `index`. template static inline void bitVectorFill(T* buf, size_t index, size_t count) noexcept { Internal::bitVectorOp(buf, index, count); } //! Clears `count` bits in bit-vector `buf` starting at bit-index `index`. template static inline void bitVectorClear(T* buf, size_t index, size_t count) noexcept { Internal::bitVectorOp(buf, index, count); } template static inline size_t bitVectorIndexOf(T* buf, size_t start, bool value) noexcept { const size_t kTSizeInBits = bitSizeOf(); size_t vecIndex = start / kTSizeInBits; // T[] size_t bitIndex = start % kTSizeInBits; // T[][] T* p = buf + vecIndex; // We always look for zeros, if value is `true` we have to flip all bits before the search. const T kFillMask = allOnes(); const T kFlipMask = value ? T(0) : kFillMask; // The first BitWord requires special handling as there are some bits we want to ignore. T bits = (*p ^ kFlipMask) & (kFillMask << bitIndex); for (;;) { if (bits) return (size_t)(p - buf) * kTSizeInBits + ctz(bits); bits = *++p ^ kFlipMask; } } // ============================================================================ // [asmjit::Support - BitVectorIterator] // ============================================================================ template class BitVectorIterator { public: const T* _ptr; size_t _idx; size_t _end; T _current; ASMJIT_INLINE BitVectorIterator(const BitVectorIterator& other) noexcept = default; ASMJIT_INLINE BitVectorIterator(const T* data, size_t numBitWords, size_t start = 0) noexcept { init(data, numBitWords, start); } ASMJIT_INLINE void init(const T* data, size_t numBitWords, size_t start = 0) noexcept { const T* ptr = data + (start / bitSizeOf()); size_t idx = alignDown(start, bitSizeOf()); size_t end = numBitWords * bitSizeOf(); T bitWord = T(0); if (idx < end) { bitWord = *ptr++ & (allOnes() << (start % bitSizeOf())); while (!bitWord && (idx += bitSizeOf()) < end) bitWord = *ptr++; } _ptr = ptr; _idx = idx; _end = end; _current = bitWord; } ASMJIT_INLINE bool hasNext() const noexcept { return _current != T(0); } ASMJIT_INLINE size_t next() noexcept { T bitWord = _current; ASMJIT_ASSERT(bitWord != T(0)); uint32_t bit = ctz(bitWord); bitWord ^= T(1u) << bit; size_t n = _idx + bit; while (!bitWord && (_idx += bitSizeOf()) < _end) bitWord = *_ptr++; _current = bitWord; return n; } ASMJIT_INLINE size_t peekNext() const noexcept { ASMJIT_ASSERT(_current != T(0)); return _idx + ctz(_current); } }; // ============================================================================ // [asmjit::Support - BitVectorOpIterator] // ============================================================================ template class BitVectorOpIterator { public: static constexpr uint32_t kTSizeInBits = bitSizeOf(); const T* _aPtr; const T* _bPtr; size_t _idx; size_t _end; T _current; ASMJIT_INLINE BitVectorOpIterator(const T* aData, const T* bData, size_t numBitWords, size_t start = 0) noexcept { init(aData, bData, numBitWords, start); } ASMJIT_INLINE void init(const T* aData, const T* bData, size_t numBitWords, size_t start = 0) noexcept { const T* aPtr = aData + (start / bitSizeOf()); const T* bPtr = bData + (start / bitSizeOf()); size_t idx = alignDown(start, bitSizeOf()); size_t end = numBitWords * bitSizeOf(); T bitWord = T(0); if (idx < end) { bitWord = OperatorT::op(*aPtr++, *bPtr++) & (allOnes() << (start % bitSizeOf())); while (!bitWord && (idx += kTSizeInBits) < end) bitWord = OperatorT::op(*aPtr++, *bPtr++); } _aPtr = aPtr; _bPtr = bPtr; _idx = idx; _end = end; _current = bitWord; } ASMJIT_INLINE bool hasNext() noexcept { return _current != T(0); } ASMJIT_INLINE size_t next() noexcept { T bitWord = _current; ASMJIT_ASSERT(bitWord != T(0)); uint32_t bit = ctz(bitWord); bitWord ^= T(1u) << bit; size_t n = _idx + bit; while (!bitWord && (_idx += kTSizeInBits) < _end) bitWord = OperatorT::op(*_aPtr++, *_bPtr++); _current = bitWord; return n; } }; // ============================================================================ // [asmjit::Support - Sorting] // ============================================================================ //! Sort order. enum SortOrder : uint32_t { kSortAscending = 0, //!< Ascending. kSortDescending = 1 //!< Descending. }; //! A helper class that provides comparison of any user-defined type that //! implements `<` and `>` operators (primitive types are supported as well). template struct Compare { template inline int operator()(const A& a, const B& b) const noexcept { return Order == kSortAscending ? int(a > b) - int(a < b) : int(a < b) - int(a > b); } }; //! Insertion sort. template> static inline void iSort(T* base, size_t size, const CompareT& cmp = CompareT()) noexcept { for (T* pm = base + 1; pm < base + size; pm++) for (T* pl = pm; pl > base && cmp(pl[-1], pl[0]) > 0; pl--) std::swap(pl[-1], pl[0]); } //! \cond namespace Internal { //! Quick-sort implementation. template struct QSortImpl { static constexpr size_t kStackSize = 64 * 2; static constexpr size_t kISortThreshold = 7; // Based on "PDCLib - Public Domain C Library" and rewritten to C++. static void sort(T* base, size_t size, const CompareT& cmp) noexcept { T* end = base + size; T* stack[kStackSize]; T** stackptr = stack; for (;;) { if ((size_t)(end - base) > kISortThreshold) { // We work from second to last - first will be pivot element. T* pi = base + 1; T* pj = end - 1; std::swap(base[(size_t)(end - base) / 2], base[0]); if (cmp(*pi , *pj ) > 0) std::swap(*pi , *pj ); if (cmp(*base, *pj ) > 0) std::swap(*base, *pj ); if (cmp(*pi , *base) > 0) std::swap(*pi , *base); // Now we have the median for pivot element, entering main loop. for (;;) { while (pi < pj && cmp(*++pi, *base) < 0) continue; // Move `i` right until `*i >= pivot`. while (pj > base && cmp(*--pj, *base) > 0) continue; // Move `j` left until `*j <= pivot`. if (pi > pj) break; std::swap(*pi, *pj); } // Move pivot into correct place. std::swap(*base, *pj); // Larger subfile base / end to stack, sort smaller. if (pj - base > end - pi) { // Left is larger. *stackptr++ = base; *stackptr++ = pj; base = pi; } else { // Right is larger. *stackptr++ = pi; *stackptr++ = end; end = pj; } ASMJIT_ASSERT(stackptr <= stack + kStackSize); } else { // UB sanitizer doesn't like applying offset to a nullptr base. if (base != end) iSort(base, (size_t)(end - base), cmp); if (stackptr == stack) break; end = *--stackptr; base = *--stackptr; } } } }; } //! \endcond //! Quick sort implementation. //! //! The main reason to provide a custom qsort implementation is that we needed //! something that will never throw `bad_alloc` exception. This implementation //! doesn't use dynamic memory allocation. template> static inline void qSort(T* base, size_t size, const CompareT& cmp = CompareT()) noexcept { Internal::QSortImpl::sort(base, size, cmp); } // ============================================================================ // [asmjit::Support::Temporary] // ============================================================================ //! Used to pass a temporary buffer to: //! //! - Containers that use user-passed buffer as an initial storage (still can grow). //! - Zone allocator that would use the temporary buffer as a first block. struct Temporary { void* _data; size_t _size; //! \name Construction & Destruction //! \{ constexpr Temporary(const Temporary& other) noexcept = default; constexpr Temporary(void* data, size_t size) noexcept : _data(data), _size(size) {} //! \} //! \name Overloaded Operators //! \{ inline Temporary& operator=(const Temporary& other) noexcept = default; //! \} //! \name Accessors //! \{ //! Returns the data storage. template constexpr T* data() const noexcept { return static_cast(_data); } //! Returns the data storage size in bytes. constexpr size_t size() const noexcept { return _size; } //! \} }; } // {Support} //! \} ASMJIT_END_NAMESPACE #endif // ASMJIT_CORE_SUPPORT_H_INCLUDED