From dd65a73af480b727de6f7ceb925c5f7ea75bdc13 Mon Sep 17 00:00:00 2001 From: Duncan Sands Date: Wed, 28 Nov 2007 10:36:19 +0000 Subject: [PATCH] My compiler complains that "x always evaluates to true" in this call: Result.IntVal = APInt(80, 2, x); What is x? uint16_t x[8]; I deduce that the APInt constructor being used is this one: APInt(uint32_t numBits, uint64_t val, bool isSigned = false); rather than this one: APInt(uint32_t numBits, uint32_t numWords, const uint64_t bigVal[]); That doesn't seem right! This fix compiles but is otherwise completely untested. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@44400 91177308-0d34-0410-b5e6-96231b3b80d8 --- lib/ExecutionEngine/ExecutionEngine.cpp | 8 ++++++-- 1 file changed, 6 insertions(+), 2 deletions(-) diff --git a/lib/ExecutionEngine/ExecutionEngine.cpp b/lib/ExecutionEngine/ExecutionEngine.cpp index 72db4e43603..cc3cc38305b 100644 --- a/lib/ExecutionEngine/ExecutionEngine.cpp +++ b/lib/ExecutionEngine/ExecutionEngine.cpp @@ -712,13 +712,17 @@ void ExecutionEngine::LoadValueFromMemory(GenericValue &Result, break; case Type::X86_FP80TyID: { // This is endian dependent, but it will only work on x86 anyway. - uint16_t x[8], *p = (uint16_t*)Ptr; + uint16_t *p = (uint16_t*)Ptr; + union { + uint16_t x[8]; + uint64_t y[2]; + }; x[0] = p[1]; x[1] = p[2]; x[2] = p[3]; x[3] = p[4]; x[4] = p[0]; - Result.IntVal = APInt(80, 2, x); + Result.IntVal = APInt(80, 2, y); break; } default: