diff options
author | Eli Friedman <eli.friedman@gmail.com> | 2011-06-14 01:37:52 +0000 |
---|---|---|
committer | Eli Friedman <eli.friedman@gmail.com> | 2011-06-14 01:37:52 +0000 |
commit | ea5e4da116aae0eb932d32b73e7cec2f04932248 (patch) | |
tree | 8bc1c234b5ee38e8dffdb8fe46ce55196d9666de /lib/CodeGen/CGCall.cpp | |
parent | bd32beee8a0f22e1d5245112f5e34ad4669994ae (diff) |
The LLVM IR representation of byval arguments has a rather strange property: if the alignment of an argument to a call is less than the specified byval alignment for that argument, there is no way to specify the alignment of the implied copy. Therefore, we must ensure that the alignment of the argument is at least the byval alignment. To do this, we have to mess with the alignment of relevant alloca's in some cases, and insert a copy that conceptually shouldn't be necessary in some cases.
This patch tries relatively hard to avoid creating an extra copy if it can be avoided (see test3 in the included testcase), but it is not possible to avoid in some cases (like test2 in the included testcase).
rdar://9483886
git-svn-id: https://llvm.org/svn/llvm-project/cfe/trunk@132957 91177308-0d34-0410-b5e6-96231b3b80d8
Diffstat (limited to 'lib/CodeGen/CGCall.cpp')
-rw-r--r-- | lib/CodeGen/CGCall.cpp | 49 |
1 files changed, 44 insertions, 5 deletions
diff --git a/lib/CodeGen/CGCall.cpp b/lib/CodeGen/CGCall.cpp index 712ae89a48..5e9ecd574b 100644 --- a/lib/CodeGen/CGCall.cpp +++ b/lib/CodeGen/CGCall.cpp @@ -1263,12 +1263,51 @@ RValue CodeGenFunction::EmitCall(const CGFunctionInfo &CallInfo, Alignment, I->Ty); else StoreComplexToAddr(RV.getComplexVal(), Args.back(), false); - } else if (I->NeedsCopy && !ArgInfo.getIndirectByVal()) { - Args.push_back(CreateMemTemp(I->Ty)); - EmitAggregateCopy(Args.back(), RV.getAggregateAddr(), I->Ty, - RV.isVolatileQualified()); } else { - Args.push_back(RV.getAggregateAddr()); + // We want to avoid creating an unnecessary temporary+copy here; + // however, we need one in two cases: + // 1. If the argument is not byval, and we are required to copy the + // source. (This case doesn't occur on any common architecture.) + // 2. If the argument is byval, RV is not sufficiently aligned, and + // we cannot force it to be sufficiently aligned. + // FIXME: This code is ugly because we don't know the required + // alignment when RV is generated. + llvm::AllocaInst *AI = + dyn_cast<llvm::AllocaInst>(RV.getAggregateAddr()); + bool NeedsAggCopy = false; + if (I->NeedsCopy && !ArgInfo.getIndirectByVal()) + NeedsAggCopy = true; + if (ArgInfo.getIndirectByVal()) { + if (AI) { + // The source is an alloca; we can force appropriate alignment. + if (ArgInfo.getIndirectAlign() > AI->getAlignment()) + AI->setAlignment(ArgInfo.getIndirectAlign()); + } else if (llvm::Argument *A = + dyn_cast<llvm::Argument>(RV.getAggregateAddr())) { + // Check if the source is an appropriately aligned byval argument. + if (!A->hasByValAttr() || + A->getParamAlignment() < ArgInfo.getIndirectAlign()) + NeedsAggCopy = true; + } else { + // We don't know what the input is; force a temporary+copy if + // the type alignment is not sufficient. + assert(I->NeedsCopy && "Temporary must be AllocaInst"); + if (ArgInfo.getIndirectAlign() > Alignment) + NeedsAggCopy = true; + } + } + if (NeedsAggCopy) { + // Create an aligned temporary, and copy to it. + AI = CreateMemTemp(I->Ty); + if (ArgInfo.getIndirectAlign() > AI->getAlignment()) + AI->setAlignment(ArgInfo.getIndirectAlign()); + Args.push_back(AI); + EmitAggregateCopy(AI, RV.getAggregateAddr(), I->Ty, + RV.isVolatileQualified()); + } else { + // Skip the extra memcpy call. + Args.push_back(RV.getAggregateAddr()); + } } break; } |