From 14f5305c92008015de4e91cd59e41e9128067101 Mon Sep 17 00:00:00 2001 From: Jeffrey Walton Date: Sun, 22 Jan 2017 20:53:19 -0500 Subject: [PATCH] VEXT_8 -> VEXT_U8 --- cpu.h | 2 +- gcm.cpp | 16 ++++++++-------- 2 files changed, 9 insertions(+), 9 deletions(-) diff --git a/cpu.h b/cpu.h index 3285230e..78c0f3f2 100644 --- a/cpu.h +++ b/cpu.h @@ -17,7 +17,7 @@ # pragma GCC diagnostic ignored "-Wsign-conversion" #endif -// ARM32/ARM64 Headers +// ARM32 and ARM64 Headers #if (CRYPTOPP_BOOL_ARM32 || CRYPTOPP_BOOL_ARM64) # if defined(__GNUC__) # include diff --git a/gcm.cpp b/gcm.cpp index 3dbf1011..a9552e6e 100644 --- a/gcm.cpp +++ b/gcm.cpp @@ -87,7 +87,7 @@ inline uint64x2_t PMULL_11(const uint64x2_t a, const uint64x2_t b) return r; } -inline uint64x2_t VEXT_8(uint64x2_t a, uint64x2_t b, unsigned int c) +inline uint64x2_t VEXT_U8(uint64x2_t a, uint64x2_t b, unsigned int c) { uint64x2_t r; __asm __volatile("ext %0.16b, %1.16b, %2.16b, %3 \n\t" @@ -97,7 +97,7 @@ inline uint64x2_t VEXT_8(uint64x2_t a, uint64x2_t b, unsigned int c) // https://github.com/weidai11/cryptopp/issues/366 template -inline uint64x2_t VEXT_8(uint64x2_t a, uint64x2_t b) +inline uint64x2_t VEXT_U8(uint64x2_t a, uint64x2_t b) { uint64x2_t r; __asm __volatile("ext %0.16b, %1.16b, %2.16b, %3 \n\t" @@ -131,14 +131,14 @@ inline uint64x2_t PMULL_11(const uint64x2_t a, const uint64x2_t b) vgetq_lane_u64(vreinterpretq_u64_u8(b),1))); } -inline uint64x2_t VEXT_8(uint64x2_t a, uint64x2_t b, unsigned int c) +inline uint64x2_t VEXT_U8(uint64x2_t a, uint64x2_t b, unsigned int c) { return (uint64x2_t)vextq_u8(vreinterpretq_u8_u64(a), vreinterpretq_u8_u64(b), c); } // https://github.com/weidai11/cryptopp/issues/366 template -inline uint64x2_t VEXT_8(uint64x2_t a, uint64x2_t b) +inline uint64x2_t VEXT_U8(uint64x2_t a, uint64x2_t b) { return (uint64x2_t)vextq_u8(vreinterpretq_u8_u64(a), vreinterpretq_u8_u64(b), C); } @@ -297,13 +297,13 @@ static const unsigned int s_clmulTableSizeInBlocks = 8; inline uint64x2_t PMULL_Reduce(uint64x2_t c0, uint64x2_t c1, uint64x2_t c2, const uint64x2_t &r) { // See comments fo CLMUL_Reduce - c1 = veorq_u64(c1, VEXT_8<8>(vdupq_n_u64(0), c0)); + c1 = veorq_u64(c1, VEXT_U8<8>(vdupq_n_u64(0), c0)); c1 = veorq_u64(c1, PMULL_01(c0, r)); - c0 = VEXT_8<8>(c0, vdupq_n_u64(0)); + c0 = VEXT_U8<8>(c0, vdupq_n_u64(0)); c0 = vshlq_n_u64(veorq_u64(c0, c1), 1); c0 = PMULL_00(c0, r); c2 = veorq_u64(c2, c0); - c2 = veorq_u64(c2, VEXT_8<8>(c1, vdupq_n_u64(0))); + c2 = veorq_u64(c2, VEXT_U8<8>(c1, vdupq_n_u64(0))); c1 = vshrq_n_u64(vcombine_u64(vget_low_u64(c1), vget_low_u64(c2)), 63); c2 = vshlq_n_u64(c2, 1); @@ -750,7 +750,7 @@ size_t GCM_Base::AuthenticateBlocks(const byte *data, size_t len) #if CRYPTOPP_BOOL_SSE2_ASM_AVAILABLE || defined(CRYPTOPP_X64_MASM_AVAILABLE) + HasSSE2() //#elif CRYPTOPP_BOOL_NEON_INTRINSICS_AVAILABLE -// + HasNEON() +// + HasNEON() #endif ) {