diff options
author | tstellar <tstellar@91177308-0d34-0410-b5e6-96231b3b80d8> | 2012-09-24 15:52:33 +0000 |
---|---|---|
committer | tstellar <tstellar@91177308-0d34-0410-b5e6-96231b3b80d8> | 2012-09-24 15:52:33 +0000 |
commit | 163fa0e8c815415396cf422130e755c753bde651 (patch) | |
tree | 8edf7959952fe0a68c3559c3b99b63ed7186f7c4 /lib/Transforms | |
parent | 69a60197fd6e6974846bf2c6525f6e1521343e28 (diff) |
Switch to a signed representation for the dynamic offsets while walking
across the uses of the alloca. It's entirely possible for negative
numbers to come up here, and in some rare cases simply doing the 2's
complement arithmetic isn't the correct decision. Notably, we can't zext
the index of the GEP. The definition of GEP is that these offsets are
sign extended or truncated to the size of the pointer, and then wrapping
2's complement arithmetic used.
This patch fixes an issue that comes up with *no* input from the
buildbots or bootstrap afaict. The only place where it manifested,
disturbingly, is Clang's own regression test suite. A reduced and
targeted collection of tests are added to cope with this. Note that I've
tried to pin down the potential cases of overflow, but may have missed
some cases. I've tried to add a few cases to test this, but its hard
because LLVM has quite limited support for >64bit constructs.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/branches/R600/@164528 91177308-0d34-0410-b5e6-96231b3b80d8
Diffstat (limited to 'lib/Transforms')
-rw-r--r-- | lib/Transforms/Scalar/SROA.cpp | 101 |
1 files changed, 75 insertions, 26 deletions
diff --git a/lib/Transforms/Scalar/SROA.cpp b/lib/Transforms/Scalar/SROA.cpp index e1f02868568..5a9247ea1b5 100644 --- a/lib/Transforms/Scalar/SROA.cpp +++ b/lib/Transforms/Scalar/SROA.cpp @@ -403,15 +403,15 @@ protected: struct OffsetUse { Use *U; - uint64_t Offset; + int64_t Offset; }; SmallVector<OffsetUse, 8> Queue; // The active offset and use while visiting. Use *U; - uint64_t Offset; + int64_t Offset; - void enqueueUsers(Instruction &I, uint64_t UserOffset) { + void enqueueUsers(Instruction &I, int64_t UserOffset) { SmallPtrSet<User *, 8> UserSet; for (Value::use_iterator UI = I.use_begin(), UE = I.use_end(); UI != UE; ++UI) { @@ -423,7 +423,7 @@ protected: } } - bool computeConstantGEPOffset(GetElementPtrInst &GEPI, uint64_t &GEPOffset) { + bool computeConstantGEPOffset(GetElementPtrInst &GEPI, int64_t &GEPOffset) { GEPOffset = Offset; for (gep_type_iterator GTI = gep_type_begin(GEPI), GTE = gep_type_end(GEPI); GTI != GTE; ++GTI) { @@ -437,12 +437,37 @@ protected: if (StructType *STy = dyn_cast<StructType>(*GTI)) { unsigned ElementIdx = OpC->getZExtValue(); const StructLayout *SL = TD.getStructLayout(STy); - GEPOffset += SL->getElementOffset(ElementIdx); + uint64_t ElementOffset = SL->getElementOffset(ElementIdx); + // Check that we can continue to model this GEP in a signed 64-bit offset. + if (ElementOffset > INT64_MAX || + (GEPOffset >= 0 && + ((uint64_t)GEPOffset + ElementOffset) > INT64_MAX)) { + DEBUG(dbgs() << "WARNING: Encountered a cumulative offset exceeding " + << "what can be represented in an int64_t!\n" + << " alloca: " << P.AI << "\n"); + return false; + } + if (GEPOffset < 0) + GEPOffset = ElementOffset + (uint64_t)-GEPOffset; + else + GEPOffset += ElementOffset; continue; } - GEPOffset - += OpC->getZExtValue() * TD.getTypeAllocSize(GTI.getIndexedType()); + APInt Index = OpC->getValue().sextOrTrunc(TD.getPointerSizeInBits()); + Index *= APInt(Index.getBitWidth(), + TD.getTypeAllocSize(GTI.getIndexedType())); + Index += APInt(Index.getBitWidth(), (uint64_t)GEPOffset, + /*isSigned*/true); + // Check if the result can be stored in our int64_t offset. + if (!Index.isSignedIntN(sizeof(GEPOffset) * 8)) { + DEBUG(dbgs() << "WARNING: Encountered a cumulative offset exceeding " + << "what can be represented in an int64_t!\n" + << " alloca: " << P.AI << "\n"); + return false; + } + + GEPOffset = Index.getSExtValue(); } return true; } @@ -495,12 +520,11 @@ private: return false; } - void insertUse(Instruction &I, uint64_t Offset, uint64_t Size, + void insertUse(Instruction &I, int64_t Offset, uint64_t Size, bool IsSplittable = false) { - uint64_t BeginOffset = Offset, EndOffset = Offset + Size; - - // Completely skip uses which start outside of the allocation. - if (BeginOffset >= AllocSize) { + // Completely skip uses which don't overlap the allocation. + if ((Offset >= 0 && (uint64_t)Offset >= AllocSize) || + (Offset < 0 && (uint64_t)-Offset >= Size)) { DEBUG(dbgs() << "WARNING: Ignoring " << Size << " byte use @" << Offset << " which starts past the end of the " << AllocSize << " byte alloca:\n" @@ -509,8 +533,22 @@ private: return; } - // Clamp the size to the allocation. - if (EndOffset > AllocSize) { + // Clamp the start to the beginning of the allocation. + if (Offset < 0) { + DEBUG(dbgs() << "WARNING: Clamping a " << Size << " byte use @" << Offset + << " to start at the beginning of the alloca:\n" + << " alloca: " << P.AI << "\n" + << " use: " << I << "\n"); + Size -= (uint64_t)-Offset; + Offset = 0; + } + + uint64_t BeginOffset = Offset, EndOffset = BeginOffset + Size; + + // Clamp the end offset to the end of the allocation. Note that this is + // formulated to handle even the case where "BeginOffset + Size" overflows. + assert(AllocSize >= BeginOffset); // Established above. + if (Size > AllocSize - BeginOffset) { DEBUG(dbgs() << "WARNING: Clamping a " << Size << " byte use @" << Offset << " to remain within the " << AllocSize << " byte alloca:\n" << " alloca: " << P.AI << "\n" @@ -530,7 +568,7 @@ private: P.Partitions.push_back(New); } - bool handleLoadOrStore(Type *Ty, Instruction &I, uint64_t Offset) { + bool handleLoadOrStore(Type *Ty, Instruction &I, int64_t Offset) { uint64_t Size = TD.getTypeStoreSize(Ty); // If this memory access can be shown to *statically* extend outside the @@ -540,7 +578,8 @@ private: // risk of overflow. // FIXME: We should instead consider the pointer to have escaped if this // function is being instrumented for addressing bugs or race conditions. - if (Offset >= AllocSize || Size > AllocSize || Offset + Size > AllocSize) { + if (Offset < 0 || (uint64_t)Offset >= AllocSize || + Size > (AllocSize - (uint64_t)Offset)) { DEBUG(dbgs() << "WARNING: Ignoring " << Size << " byte " << (isa<LoadInst>(I) ? "load" : "store") << " @" << Offset << " which extends past the end of the " << AllocSize @@ -560,7 +599,7 @@ private: } bool visitGetElementPtrInst(GetElementPtrInst &GEPI) { - uint64_t GEPOffset; + int64_t GEPOffset; if (!computeConstantGEPOffset(GEPI, GEPOffset)) return markAsEscaping(GEPI); @@ -784,16 +823,25 @@ private: P.DeadUsers.push_back(&I); } - void insertUse(Instruction &User, uint64_t Offset, uint64_t Size) { - uint64_t BeginOffset = Offset, EndOffset = Offset + Size; - + void insertUse(Instruction &User, int64_t Offset, uint64_t Size) { // If the use extends outside of the allocation, record it as a dead use // for elimination later. - if (BeginOffset >= AllocSize || Size == 0) + if ((uint64_t)Offset >= AllocSize || + (Offset < 0 && (uint64_t)-Offset >= Size)) return markAsDead(User); - // Bound the use by the size of the allocation. - if (EndOffset > AllocSize) + // Clamp the start to the beginning of the allocation. + if (Offset < 0) { + Size -= (uint64_t)-Offset; + Offset = 0; + } + + uint64_t BeginOffset = Offset, EndOffset = BeginOffset + Size; + + // Clamp the end offset to the end of the allocation. Note that this is + // formulated to handle even the case where "BeginOffset + Size" overflows. + assert(AllocSize >= BeginOffset); // Established above. + if (Size > AllocSize - BeginOffset) EndOffset = AllocSize; // NB: This only works if we have zero overlapping partitions. @@ -812,14 +860,15 @@ private: } } - void handleLoadOrStore(Type *Ty, Instruction &I, uint64_t Offset) { + void handleLoadOrStore(Type *Ty, Instruction &I, int64_t Offset) { uint64_t Size = TD.getTypeStoreSize(Ty); // If this memory access can be shown to *statically* extend outside the // bounds of of the allocation, it's behavior is undefined, so simply // ignore it. Note that this is more strict than the generic clamping // behavior of insertUse. - if (Offset >= AllocSize || Size > AllocSize || Offset + Size > AllocSize) + if (Offset < 0 || (uint64_t)Offset >= AllocSize || + Size > (AllocSize - (uint64_t)Offset)) return markAsDead(I); insertUse(I, Offset, Size); @@ -836,7 +885,7 @@ private: if (GEPI.use_empty()) return markAsDead(GEPI); - uint64_t GEPOffset; + int64_t GEPOffset; if (!computeConstantGEPOffset(GEPI, GEPOffset)) llvm_unreachable("Unable to compute constant offset for use"); |