r175389 - Re-apply r174919 - smarter copy/move assignment/construction, with fixes for

Chandler Carruth chandlerc at google.com
Sun Feb 17 01:45:13 PST 2013


On Sun, Feb 17, 2013 at 1:25 AM, Lang Hames <lhames at gmail.com> wrote:

> Hrm. I can't reproduce those locally. I suspect the check lines are just
> overly specific.
>
> Do you have access to that bot? Are you able to attach the bitcode that
> that clang command produces?
>

Have to tried with a release build? I suspect this is because preserving
basic block names only happens in +Asserts builds, and you're CHECK-ing the
name of a basic block in the test.

(You're also testing the output of -O1, which means that random LLVM
changes can break this test... sub-optimal, but not a critical issue as
there are a *lot* of violations of this principle in the Clang test
suite...)


>
> Can you point me to the asan failures?
>
> - Lang.
>
>
> On Sun, Feb 17, 2013 at 12:13 AM, Chandler Carruth <chandlerc at google.com>wrote:
>
>> On Sat, Feb 16, 2013 at 11:22 PM, Lang Hames <lhames at gmail.com> wrote:
>>
>>> Author: lhames
>>> Date: Sun Feb 17 01:22:09 2013
>>> New Revision: 175389
>>>
>>> URL: http://llvm.org/viewvc/llvm-project?rev=175389&view=rev
>>> Log:
>>> Re-apply r174919 - smarter copy/move assignment/construction, with fixes
>>> for
>>> bitfield related issues.
>>>
>>> The original commit broke Takumi's builder. The bug was caused by
>>> bitfield sizes
>>> being determined by their underlying type, rather than the field info. A
>>> similar
>>> issue with bitfield alignments showed up on closer testing. Both have
>>> been fixed
>>> in this patch.
>>>
>>
>> This seems to have caused (less exciting looking) test failures:
>> http://bb.pgr.jp/builders/cmake-clang-x86_64-linux/builds/4670
>>
>> I'm also seeing a bunch of asan failures, but not 100% certain they're
>> related to this patch.
>>
>>
>>>
>>>
>>> Added:
>>>     cfe/trunk/test/CodeGenCXX/pod-member-memcpys.cpp
>>> Modified:
>>>     cfe/trunk/lib/CodeGen/CGClass.cpp
>>>     cfe/trunk/lib/CodeGen/CodeGenFunction.cpp
>>>     cfe/trunk/lib/CodeGen/CodeGenFunction.h
>>>     cfe/trunk/test/CodeGenCXX/copy-assign-synthesis-1.cpp
>>>     cfe/trunk/test/CodeGenCXX/implicit-copy-assign-operator.cpp
>>>     cfe/trunk/test/CodeGenCXX/implicit-copy-constructor.cpp
>>>
>>> Modified: cfe/trunk/lib/CodeGen/CGClass.cpp
>>> URL:
>>> http://llvm.org/viewvc/llvm-project/cfe/trunk/lib/CodeGen/CGClass.cpp?rev=175389&r1=175388&r2=175389&view=diff
>>>
>>> ==============================================================================
>>> --- cfe/trunk/lib/CodeGen/CGClass.cpp (original)
>>> +++ cfe/trunk/lib/CodeGen/CGClass.cpp Sun Feb 17 01:22:09 2013
>>> @@ -13,11 +13,13 @@
>>>
>>>  #include "CGBlocks.h"
>>>  #include "CGDebugInfo.h"
>>> +#include "CGRecordLayout.h"
>>>  #include "CodeGenFunction.h"
>>>  #include "clang/AST/CXXInheritance.h"
>>>  #include "clang/AST/EvaluatedExprVisitor.h"
>>>  #include "clang/AST/RecordLayout.h"
>>>  #include "clang/AST/StmtCXX.h"
>>> +#include "clang/Basic/TargetBuiltins.h"
>>>  #include "clang/Frontend/CodeGenOptions.h"
>>>
>>>  using namespace clang;
>>> @@ -742,6 +744,352 @@ void CodeGenFunction::EmitConstructorBod
>>>      ExitCXXTryStmt(*cast<CXXTryStmt>(Body), true);
>>>  }
>>>
>>> +namespace {
>>> +  class FieldMemcpyizer {
>>> +  public:
>>> +    FieldMemcpyizer(CodeGenFunction &CGF, const CXXRecordDecl
>>> *ClassDecl,
>>> +                    const VarDecl *SrcRec)
>>> +      : CGF(CGF), ClassDecl(ClassDecl), SrcRec(SrcRec),
>>> +        RecLayout(CGF.getContext().getASTRecordLayout(ClassDecl)),
>>> +        FirstField(0), LastField(0), FirstFieldOffset(0),
>>> LastFieldOffset(0),
>>> +        LastAddedFieldIndex(0) { }
>>> +
>>> +    static bool isMemcpyableField(FieldDecl *F) {
>>> +      Qualifiers Qual = F->getType().getQualifiers();
>>> +      if (Qual.hasVolatile() || Qual.hasObjCLifetime())
>>> +        return false;
>>> +      return true;
>>> +    }
>>> +
>>> +    void addMemcpyableField(FieldDecl *F) {
>>> +      if (FirstField == 0)
>>> +        addInitialField(F);
>>> +      else
>>> +        addNextField(F);
>>> +    }
>>> +
>>> +    CharUnits getMemcpySize() const {
>>> +      unsigned LastFieldSize =
>>> +        LastField->isBitField() ?
>>> +          LastField->getBitWidthValue(CGF.getContext()) :
>>> +          CGF.getContext().getTypeSize(LastField->getType());
>>> +      uint64_t MemcpySizeBits =
>>> +        LastFieldOffset + LastFieldSize - FirstFieldOffset +
>>> +        CGF.getContext().getCharWidth() - 1;
>>> +      CharUnits MemcpySize =
>>> +        CGF.getContext().toCharUnitsFromBits(MemcpySizeBits);
>>> +      return MemcpySize;
>>> +    }
>>> +
>>> +    void emitMemcpy() {
>>> +      // Give the subclass a chance to bail out if it feels the memcpy
>>> isn't
>>> +      // worth it (e.g. Hasn't aggregated enough data).
>>> +      if (FirstField == 0) {
>>> +        return;
>>> +      }
>>> +
>>> +      unsigned FirstFieldAlign = ~0U; // Set to invalid.
>>> +
>>> +      if (FirstField->isBitField()) {
>>> +        const CGRecordLayout &RL =
>>> +          CGF.getTypes().getCGRecordLayout(FirstField->getParent());
>>> +        const CGBitFieldInfo &BFInfo = RL.getBitFieldInfo(FirstField);
>>> +        FirstFieldAlign = BFInfo.StorageAlignment;
>>> +      } else
>>> +        FirstFieldAlign =
>>> CGF.getContext().getTypeAlign(FirstField->getType());
>>> +
>>> +      assert(FirstFieldOffset % FirstFieldAlign == 0 && "Bad field
>>> alignment.");
>>> +      CharUnits Alignment =
>>> +        CGF.getContext().toCharUnitsFromBits(FirstFieldAlign);
>>> +      CharUnits MemcpySize = getMemcpySize();
>>> +      QualType RecordTy = CGF.getContext().getTypeDeclType(ClassDecl);
>>> +      llvm::Value *ThisPtr = CGF.LoadCXXThis();
>>> +      LValue DestLV = CGF.MakeNaturalAlignAddrLValue(ThisPtr, RecordTy);
>>> +      LValue Dest = CGF.EmitLValueForFieldInitialization(DestLV,
>>> FirstField);
>>> +      llvm::Value *SrcPtr =
>>> CGF.Builder.CreateLoad(CGF.GetAddrOfLocalVar(SrcRec));
>>> +      LValue SrcLV = CGF.MakeNaturalAlignAddrLValue(SrcPtr, RecordTy);
>>> +      LValue Src = CGF.EmitLValueForFieldInitialization(SrcLV,
>>> FirstField);
>>> +
>>> +      emitMemcpyIR(Dest.isBitField() ? Dest.getBitFieldAddr() :
>>> Dest.getAddress(),
>>> +                   Src.isBitField() ? Src.getBitFieldAddr() :
>>> Src.getAddress(),
>>> +                   MemcpySize, Alignment);
>>> +      reset();
>>> +    }
>>> +
>>> +    void reset() {
>>> +      FirstField = 0;
>>> +    }
>>> +
>>> +  protected:
>>> +    CodeGenFunction &CGF;
>>> +    const CXXRecordDecl *ClassDecl;
>>> +
>>> +  private:
>>> +
>>> +    void emitMemcpyIR(llvm::Value *DestPtr, llvm::Value *SrcPtr,
>>> +                      CharUnits Size, CharUnits Alignment) {
>>> +      llvm::PointerType *DPT =
>>> cast<llvm::PointerType>(DestPtr->getType());
>>> +      llvm::Type *DBP =
>>> +        llvm::Type::getInt8PtrTy(CGF.getLLVMContext(),
>>> DPT->getAddressSpace());
>>> +      DestPtr = CGF.Builder.CreateBitCast(DestPtr, DBP);
>>> +
>>> +      llvm::PointerType *SPT =
>>> cast<llvm::PointerType>(SrcPtr->getType());
>>> +      llvm::Type *SBP =
>>> +        llvm::Type::getInt8PtrTy(CGF.getLLVMContext(),
>>> SPT->getAddressSpace());
>>> +      SrcPtr = CGF.Builder.CreateBitCast(SrcPtr, SBP);
>>> +
>>> +      CGF.Builder.CreateMemCpy(DestPtr, SrcPtr, Size.getQuantity(),
>>> +                               Alignment.getQuantity());
>>> +    }
>>> +
>>> +    void addInitialField(FieldDecl *F) {
>>> +        FirstField = F;
>>> +        LastField = F;
>>> +        FirstFieldOffset = RecLayout.getFieldOffset(F->getFieldIndex());
>>> +        LastFieldOffset = FirstFieldOffset;
>>> +        LastAddedFieldIndex = F->getFieldIndex();
>>> +        return;
>>> +      }
>>> +
>>> +    void addNextField(FieldDecl *F) {
>>> +      assert(F->getFieldIndex() == LastAddedFieldIndex + 1 &&
>>> +             "Cannot aggregate non-contiguous fields.");
>>> +      LastAddedFieldIndex = F->getFieldIndex();
>>> +
>>> +      // The 'first' and 'last' fields are chosen by offset, rather
>>> than field
>>> +      // index. This allows the code to support bitfields, as well as
>>> regular
>>> +      // fields.
>>> +      uint64_t FOffset = RecLayout.getFieldOffset(F->getFieldIndex());
>>> +      if (FOffset < FirstFieldOffset) {
>>> +        FirstField = F;
>>> +        FirstFieldOffset = FOffset;
>>> +      } else if (FOffset > LastFieldOffset) {
>>> +        LastField = F;
>>> +        LastFieldOffset = FOffset;
>>> +      }
>>> +    }
>>> +
>>> +    const VarDecl *SrcRec;
>>> +    const ASTRecordLayout &RecLayout;
>>> +    FieldDecl *FirstField;
>>> +    FieldDecl *LastField;
>>> +    uint64_t FirstFieldOffset, LastFieldOffset;
>>> +    unsigned LastAddedFieldIndex;
>>> +  };
>>> +
>>> +  class ConstructorMemcpyizer : public FieldMemcpyizer {
>>> +  private:
>>> +
>>> +    /// Get source argument for copy constructor. Returns null if not a
>>> copy
>>> +    /// constructor.
>>> +    static const VarDecl* getTrivialCopySource(const CXXConstructorDecl
>>> *CD,
>>> +                                               FunctionArgList &Args) {
>>> +      if (CD->isCopyOrMoveConstructor() && CD->isImplicitlyDefined())
>>> +        return Args[Args.size() - 1];
>>> +      return 0;
>>> +    }
>>> +
>>> +    // Returns true if a CXXCtorInitializer represents a member
>>> initialization
>>> +    // that can be rolled into a memcpy.
>>> +    bool isMemberInitMemcpyable(CXXCtorInitializer *MemberInit) const {
>>> +      if (!MemcpyableCtor)
>>> +        return false;
>>> +      FieldDecl *Field = MemberInit->getMember();
>>> +      assert(Field != 0 && "No field for member init.");
>>> +      QualType FieldType = Field->getType();
>>> +      CXXConstructExpr *CE =
>>> dyn_cast<CXXConstructExpr>(MemberInit->getInit());
>>> +
>>> +      // Bail out on non-POD, not-trivially-constructable members.
>>> +      if (!(CE && CE->getConstructor()->isTrivial()) &&
>>> +          !(FieldType.isTriviallyCopyableType(CGF.getContext()) ||
>>> +            FieldType->isReferenceType()))
>>> +        return false;
>>> +
>>> +      // Bail out on volatile fields.
>>> +      if (!isMemcpyableField(Field))
>>> +        return false;
>>> +
>>> +      // Otherwise we're good.
>>> +      return true;
>>> +    }
>>> +
>>> +  public:
>>> +    ConstructorMemcpyizer(CodeGenFunction &CGF, const
>>> CXXConstructorDecl *CD,
>>> +                          FunctionArgList &Args)
>>> +      : FieldMemcpyizer(CGF, CD->getParent(), getTrivialCopySource(CD,
>>> Args)),
>>> +        ConstructorDecl(CD),
>>> +        MemcpyableCtor(CD->isImplicitlyDefined() &&
>>> +                       CD->isCopyOrMoveConstructor() &&
>>> +                       CGF.getLangOpts().getGC() == LangOptions::NonGC),
>>> +        Args(Args) { }
>>> +
>>> +    void addMemberInitializer(CXXCtorInitializer *MemberInit) {
>>> +      if (isMemberInitMemcpyable(MemberInit)) {
>>> +        AggregatedInits.push_back(MemberInit);
>>> +        addMemcpyableField(MemberInit->getMember());
>>> +      } else {
>>> +        emitAggregatedInits();
>>> +        EmitMemberInitializer(CGF, ConstructorDecl->getParent(),
>>> MemberInit,
>>> +                              ConstructorDecl, Args);
>>> +      }
>>> +    }
>>> +
>>> +    void emitAggregatedInits() {
>>> +      if (AggregatedInits.size() <= 1) {
>>> +        // This memcpy is too small to be worthwhile. Fall back on
>>> default
>>> +        // codegen.
>>> +        for (unsigned i = 0; i < AggregatedInits.size(); ++i) {
>>> +          EmitMemberInitializer(CGF, ConstructorDecl->getParent(),
>>> +                                AggregatedInits[i], ConstructorDecl,
>>> Args);
>>> +        }
>>> +        reset();
>>> +        return;
>>> +      }
>>> +
>>> +      pushEHDestructors();
>>> +      emitMemcpy();
>>> +      AggregatedInits.clear();
>>> +    }
>>> +
>>> +    void pushEHDestructors() {
>>> +      llvm::Value *ThisPtr = CGF.LoadCXXThis();
>>> +      QualType RecordTy = CGF.getContext().getTypeDeclType(ClassDecl);
>>> +      LValue LHS = CGF.MakeNaturalAlignAddrLValue(ThisPtr, RecordTy);
>>> +
>>> +      for (unsigned i = 0; i < AggregatedInits.size(); ++i) {
>>> +        QualType FieldType = AggregatedInits[i]->getMember()->getType();
>>> +        QualType::DestructionKind dtorKind =
>>> FieldType.isDestructedType();
>>> +        if (CGF.needsEHCleanup(dtorKind))
>>> +          CGF.pushEHDestroy(dtorKind, LHS.getAddress(), FieldType);
>>> +      }
>>> +    }
>>> +
>>> +    void finish() {
>>> +      emitAggregatedInits();
>>> +    }
>>> +
>>> +  private:
>>> +    const CXXConstructorDecl *ConstructorDecl;
>>> +    bool MemcpyableCtor;
>>> +    FunctionArgList &Args;
>>> +    SmallVector<CXXCtorInitializer*, 16> AggregatedInits;
>>> +  };
>>> +
>>> +  class AssignmentMemcpyizer : public FieldMemcpyizer {
>>> +  private:
>>> +
>>> +    // Returns the memcpyable field copied by the given statement, if
>>> one
>>> +    // exists. Otherwise r
>>> +    FieldDecl* getMemcpyableField(Stmt *S) {
>>> +      if (!AssignmentsMemcpyable)
>>> +        return 0;
>>> +      if (BinaryOperator *BO = dyn_cast<BinaryOperator>(S)) {
>>> +        // Recognise trivial assignments.
>>> +        if (BO->getOpcode() != BO_Assign)
>>> +          return 0;
>>> +        MemberExpr *ME = dyn_cast<MemberExpr>(BO->getLHS());
>>> +        if (!ME)
>>> +          return 0;
>>> +        FieldDecl *Field = dyn_cast<FieldDecl>(ME->getMemberDecl());
>>> +        if (!Field || !isMemcpyableField(Field))
>>> +          return 0;
>>> +        Stmt *RHS = BO->getRHS();
>>> +        if (ImplicitCastExpr *EC = dyn_cast<ImplicitCastExpr>(RHS))
>>> +          RHS = EC->getSubExpr();
>>> +        if (!RHS)
>>> +          return 0;
>>> +        MemberExpr *ME2 = dyn_cast<MemberExpr>(RHS);
>>> +        if (dyn_cast<FieldDecl>(ME2->getMemberDecl()) != Field)
>>> +          return 0;
>>> +        return Field;
>>> +      } else if (CXXMemberCallExpr *MCE =
>>> dyn_cast<CXXMemberCallExpr>(S)) {
>>> +        CXXMethodDecl *MD =
>>> dyn_cast<CXXMethodDecl>(MCE->getCalleeDecl());
>>> +        if (!(MD && (MD->isCopyAssignmentOperator() ||
>>> +                       MD->isMoveAssignmentOperator()) &&
>>> +              MD->isTrivial()))
>>> +          return 0;
>>> +        MemberExpr *IOA =
>>> dyn_cast<MemberExpr>(MCE->getImplicitObjectArgument());
>>> +        if (!IOA)
>>> +          return 0;
>>> +        FieldDecl *Field = dyn_cast<FieldDecl>(IOA->getMemberDecl());
>>> +        if (!Field || !isMemcpyableField(Field))
>>> +          return 0;
>>> +        MemberExpr *Arg0 = dyn_cast<MemberExpr>(MCE->getArg(0));
>>> +        if (!Arg0 || Field !=
>>> dyn_cast<FieldDecl>(Arg0->getMemberDecl()))
>>> +          return 0;
>>> +        return Field;
>>> +      } else if (CallExpr *CE = dyn_cast<CallExpr>(S)) {
>>> +        FunctionDecl *FD = dyn_cast<FunctionDecl>(CE->getCalleeDecl());
>>> +        if (!FD || FD->getBuiltinID() != Builtin::BI__builtin_memcpy)
>>> +          return 0;
>>> +        Expr *DstPtr = CE->getArg(0);
>>> +        if (ImplicitCastExpr *DC = dyn_cast<ImplicitCastExpr>(DstPtr))
>>> +          DstPtr = DC->getSubExpr();
>>> +        UnaryOperator *DUO = dyn_cast<UnaryOperator>(DstPtr);
>>> +        if (!DUO || DUO->getOpcode() != UO_AddrOf)
>>> +          return 0;
>>> +        MemberExpr *ME = dyn_cast<MemberExpr>(DUO->getSubExpr());
>>> +        if (!ME)
>>> +          return 0;
>>> +        FieldDecl *Field = dyn_cast<FieldDecl>(ME->getMemberDecl());
>>> +        if (!Field || !isMemcpyableField(Field))
>>> +          return 0;
>>> +        Expr *SrcPtr = CE->getArg(1);
>>> +        if (ImplicitCastExpr *SC = dyn_cast<ImplicitCastExpr>(SrcPtr))
>>> +          SrcPtr = SC->getSubExpr();
>>> +        UnaryOperator *SUO = dyn_cast<UnaryOperator>(SrcPtr);
>>> +        if (!SUO || SUO->getOpcode() != UO_AddrOf)
>>> +          return 0;
>>> +        MemberExpr *ME2 = dyn_cast<MemberExpr>(SUO->getSubExpr());
>>> +        if (!ME2 || Field != dyn_cast<FieldDecl>(ME2->getMemberDecl()))
>>> +          return 0;
>>> +        return Field;
>>> +      }
>>> +
>>> +      return 0;
>>> +    }
>>> +
>>> +    bool AssignmentsMemcpyable;
>>> +    SmallVector<Stmt*, 16> AggregatedStmts;
>>> +
>>> +  public:
>>> +
>>> +    AssignmentMemcpyizer(CodeGenFunction &CGF, const CXXMethodDecl *AD,
>>> +                         FunctionArgList &Args)
>>> +      : FieldMemcpyizer(CGF, AD->getParent(), Args[Args.size() - 1]),
>>> +        AssignmentsMemcpyable(CGF.getLangOpts().getGC() ==
>>> LangOptions::NonGC) {
>>> +      assert(Args.size() == 2);
>>> +    }
>>> +
>>> +    void emitAssignment(Stmt *S) {
>>> +      FieldDecl *F = getMemcpyableField(S);
>>> +      if (F) {
>>> +        addMemcpyableField(F);
>>> +        AggregatedStmts.push_back(S);
>>> +      } else {
>>> +        emitAggregatedStmts();
>>> +        CGF.EmitStmt(S);
>>> +      }
>>> +    }
>>> +
>>> +    void emitAggregatedStmts() {
>>> +      if (AggregatedStmts.size() <= 1) {
>>> +        for (unsigned i = 0; i < AggregatedStmts.size(); ++i)
>>> +          CGF.EmitStmt(AggregatedStmts[i]);
>>> +        reset();
>>> +      }
>>> +
>>> +      emitMemcpy();
>>> +      AggregatedStmts.clear();
>>> +    }
>>> +
>>> +    void finish() {
>>> +      emitAggregatedStmts();
>>> +    }
>>> +  };
>>> +
>>> +}
>>> +
>>>  /// EmitCtorPrologue - This routine generates necessary code to
>>> initialize
>>>  /// base classes and non-static data members belonging to this
>>> constructor.
>>>  void CodeGenFunction::EmitCtorPrologue(const CXXConstructorDecl *CD,
>>> @@ -770,8 +1118,10 @@ void CodeGenFunction::EmitCtorPrologue(c
>>>
>>>    InitializeVTablePointers(ClassDecl);
>>>
>>> +  ConstructorMemcpyizer CM(*this, CD, Args);
>>>    for (unsigned I = 0, E = MemberInitializers.size(); I != E; ++I)
>>> -    EmitMemberInitializer(*this, ClassDecl, MemberInitializers[I], CD,
>>> Args);
>>> +    CM.addMemberInitializer(MemberInitializers[I]);
>>> +  CM.finish();
>>>  }
>>>
>>>  static bool
>>> @@ -940,6 +1290,24 @@ void CodeGenFunction::EmitDestructorBody
>>>      ExitCXXTryStmt(*cast<CXXTryStmt>(Body), true);
>>>  }
>>>
>>> +void
>>> CodeGenFunction::emitImplicitAssignmentOperatorBody(FunctionArgList &Args) {
>>> +  const CXXMethodDecl *AssignOp = cast<CXXMethodDecl>(CurGD.getDecl());
>>> +  const Stmt *RootS = AssignOp->getBody();
>>> +  assert(isa<CompoundStmt>(RootS) &&
>>> +         "Body of an implicit assignment operator should be compound
>>> stmt.");
>>> +  const CompoundStmt *RootCS = cast<CompoundStmt>(RootS);
>>> +
>>> +  LexicalScope Scope(*this, RootCS->getSourceRange());
>>> +
>>> +  AssignmentMemcpyizer AM(*this, AssignOp, Args);
>>> +  for (CompoundStmt::const_body_iterator I = RootCS->body_begin(),
>>> +                                         E = RootCS->body_end();
>>> +       I != E; ++I) {
>>> +    AM.emitAssignment(*I);
>>> +  }
>>> +  AM.finish();
>>> +}
>>> +
>>>  namespace {
>>>    /// Call the operator delete associated with the current destructor.
>>>    struct CallDtorDelete : EHScopeStack::Cleanup {
>>>
>>> Modified: cfe/trunk/lib/CodeGen/CodeGenFunction.cpp
>>> URL:
>>> http://llvm.org/viewvc/llvm-project/cfe/trunk/lib/CodeGen/CodeGenFunction.cpp?rev=175389&r1=175388&r2=175389&view=diff
>>>
>>> ==============================================================================
>>> --- cfe/trunk/lib/CodeGen/CodeGenFunction.cpp (original)
>>> +++ cfe/trunk/lib/CodeGen/CodeGenFunction.cpp Sun Feb 17 01:22:09 2013
>>> @@ -560,6 +560,11 @@ void CodeGenFunction::GenerateCode(Globa
>>>      // The lambda "__invoke" function is special, because it forwards or
>>>      // clones the body of the function call operator (but is actually
>>> static).
>>>      EmitLambdaStaticInvokeFunction(cast<CXXMethodDecl>(FD));
>>> +  } else if (FD->isDefaulted() && isa<CXXMethodDecl>(FD) &&
>>> +             cast<CXXMethodDecl>(FD)->isCopyAssignmentOperator()) {
>>> +    // Implicit copy-assignment gets the same special treatment as
>>> implicit
>>> +    // copy-constructors.
>>> +    emitImplicitAssignmentOperatorBody(Args);
>>>    }
>>>    else
>>>      EmitFunctionBody(Args);
>>>
>>> Modified: cfe/trunk/lib/CodeGen/CodeGenFunction.h
>>> URL:
>>> http://llvm.org/viewvc/llvm-project/cfe/trunk/lib/CodeGen/CodeGenFunction.h?rev=175389&r1=175388&r2=175389&view=diff
>>>
>>> ==============================================================================
>>> --- cfe/trunk/lib/CodeGen/CodeGenFunction.h (original)
>>> +++ cfe/trunk/lib/CodeGen/CodeGenFunction.h Sun Feb 17 01:22:09 2013
>>> @@ -1401,6 +1401,7 @@ public:
>>>
>>>    void EmitConstructorBody(FunctionArgList &Args);
>>>    void EmitDestructorBody(FunctionArgList &Args);
>>> +  void emitImplicitAssignmentOperatorBody(FunctionArgList &Args);
>>>    void EmitFunctionBody(FunctionArgList &Args);
>>>
>>>    void EmitForwardingCallToLambda(const CXXRecordDecl *Lambda,
>>>
>>> Modified: cfe/trunk/test/CodeGenCXX/copy-assign-synthesis-1.cpp
>>> URL:
>>> http://llvm.org/viewvc/llvm-project/cfe/trunk/test/CodeGenCXX/copy-assign-synthesis-1.cpp?rev=175389&r1=175388&r2=175389&view=diff
>>>
>>> ==============================================================================
>>> --- cfe/trunk/test/CodeGenCXX/copy-assign-synthesis-1.cpp (original)
>>> +++ cfe/trunk/test/CodeGenCXX/copy-assign-synthesis-1.cpp Sun Feb 17
>>> 01:22:09 2013
>>> @@ -96,14 +96,8 @@ int main() {
>>>  // CHECK-LP64: .globl   __ZN1XaSERKS_
>>>  // CHECK-LP64: .weak_definition  __ZN1XaSERKS_
>>>  // CHECK-LP64: __ZN1XaSERKS_:
>>> -// CHECK-LP64: .globl   __ZN1QaSERKS_
>>> -// CHECK-LP64: .weak_definition  __ZN1QaSERKS_
>>> -// CHECK-LP64: __ZN1QaSERKS_:
>>>
>>>  // CHECK-LP32: .globl   __ZN1XaSERKS_
>>>  // CHECK-LP32: .weak_definition  __ZN1XaSERKS_
>>>  // CHECK-LP32: __ZN1XaSERKS_:
>>> -// CHECK-LP32: .globl   __ZN1QaSERKS_
>>> -// CHECK-LP32: .weak_definition  __ZN1QaSERKS_
>>> -// CHECK-LP32: __ZN1QaSERKS_:
>>>
>>>
>>> Modified: cfe/trunk/test/CodeGenCXX/implicit-copy-assign-operator.cpp
>>> URL:
>>> http://llvm.org/viewvc/llvm-project/cfe/trunk/test/CodeGenCXX/implicit-copy-assign-operator.cpp?rev=175389&r1=175388&r2=175389&view=diff
>>>
>>> ==============================================================================
>>> --- cfe/trunk/test/CodeGenCXX/implicit-copy-assign-operator.cpp
>>> (original)
>>> +++ cfe/trunk/test/CodeGenCXX/implicit-copy-assign-operator.cpp Sun Feb
>>> 17 01:22:09 2013
>>> @@ -44,7 +44,7 @@ void test_D(D d1, D d2) {
>>>  // CHECK: {{call.*_ZN1AaSERS_}}
>>>  // CHECK: {{call.*_ZN1BaSERS_}}
>>>  // CHECK: {{call.*_ZN1CaSERKS_}}
>>> -// CHECK: {{call void @llvm.memcpy.p0i8.p0i8.i64.*i64 24}}
>>> +// CHECK: {{call void @llvm.memcpy.p0i8.p0i8.i64.*i64 28}}
>>>  // CHECK: {{call.*_ZN1BaSERS_}}
>>>  // CHECK: br
>>>  // CHECK: {{call.*_ZN1CaSERKS_}}
>>>
>>> Modified: cfe/trunk/test/CodeGenCXX/implicit-copy-constructor.cpp
>>> URL:
>>> http://llvm.org/viewvc/llvm-project/cfe/trunk/test/CodeGenCXX/implicit-copy-constructor.cpp?rev=175389&r1=175388&r2=175389&view=diff
>>>
>>> ==============================================================================
>>> --- cfe/trunk/test/CodeGenCXX/implicit-copy-constructor.cpp (original)
>>> +++ cfe/trunk/test/CodeGenCXX/implicit-copy-constructor.cpp Sun Feb 17
>>> 01:22:09 2013
>>> @@ -46,7 +46,7 @@ void f(D d) {
>>>  // CHECK: call void @_ZN1AD1Ev
>>>  // CHECK: call void @_ZN1AC2ERS_
>>>  // CHECK: call void @_ZN1BC2ERS_
>>> -// CHECK: {{call void @llvm.memcpy.p0i8.p0i8.i64.*i64 24}}
>>> +// CHECK: {{call void @llvm.memcpy.p0i8.p0i8.i64.*i64 28}}
>>>  // CHECK: call void @_ZN1BC1ERS_
>>>  // CHECK: br
>>>  // CHECK: {{icmp ult.*, 2}}
>>> @@ -54,8 +54,7 @@ void f(D d) {
>>>  // CHECK: call void @_ZN1AC1Ev
>>>  // CHECK: call void @_ZN1CC1ERS_1A
>>>  // CHECK: call void @_ZN1AD1Ev
>>> -// CHECK: {{call void @llvm.memcpy.p0i8.p0i8.i64.*i64 288}}
>>> -// CHECK: {{call void @llvm.memcpy.p0i8.p0i8.i64.*i64 12}}
>>> +// CHECK: {{call void @llvm.memcpy.p0i8.p0i8.i64.*i64 300}}
>>>  // CHECK: ret void
>>>
>>>
>>>
>>> Added: cfe/trunk/test/CodeGenCXX/pod-member-memcpys.cpp
>>> URL:
>>> http://llvm.org/viewvc/llvm-project/cfe/trunk/test/CodeGenCXX/pod-member-memcpys.cpp?rev=175389&view=auto
>>>
>>> ==============================================================================
>>> --- cfe/trunk/test/CodeGenCXX/pod-member-memcpys.cpp (added)
>>> +++ cfe/trunk/test/CodeGenCXX/pod-member-memcpys.cpp Sun Feb 17 01:22:09
>>> 2013
>>> @@ -0,0 +1,224 @@
>>> +// RUN: %clang_cc1 -triple x86_64-apple-darwin10 -emit-llvm -std=c++03
>>> -fexceptions -fcxx-exceptions -O1 -o - %s | FileCheck %s
>>> +
>>> +struct POD {
>>> +  int w, x, y, z;
>>> +};
>>> +
>>> +struct PODLike {
>>> +  int w, x, y, z;
>>> +  PODLike();
>>> +  ~PODLike();
>>> +};
>>> +
>>> +struct NonPOD {
>>> +  NonPOD();
>>> +  NonPOD(const NonPOD&);
>>> +  NonPOD& operator=(const NonPOD&);
>>> +};
>>> +
>>> +struct Basic {
>>> +  int a, b, c, d;
>>> +  NonPOD np;
>>> +  int w, x, y, z;
>>> +};
>>> +
>>> +struct PODMember {
>>> +  int a, b, c, d;
>>> +  POD p;
>>> +  NonPOD np;
>>> +  int w, x, y, z;
>>> +};
>>> +
>>> +struct PODLikeMember {
>>> +  int a, b, c, d;
>>> +  PODLike pl;
>>> +  NonPOD np;
>>> +  int w, x, y, z;
>>> +};
>>> +
>>> +struct ArrayMember {
>>> +  int a, b, c, d;
>>> +  int e[12];
>>> +  NonPOD np;
>>> +  int f[12];
>>> +  int w, x, y, z;
>>> +};
>>> +
>>> +struct VolatileMember {
>>> +  int a, b, c, d;
>>> +  volatile int v;
>>> +  NonPOD np;
>>> +  int w, x, y, z;
>>> +};
>>> +
>>> +struct BitfieldMember {
>>> +  int a, b, c, d;
>>> +  NonPOD np;
>>> +  int w : 6;
>>> +  int x : 6;
>>> +  int y : 6;
>>> +  int z : 6;
>>> +};
>>> +
>>> +struct InnerClassMember {
>>> +  struct {
>>> +    int a, b, c, d;
>>> +  } a;
>>> +  int b, c, d, e;
>>> +  NonPOD np;
>>> +  int w, x, y, z;
>>> +};
>>> +
>>> +struct ReferenceMember {
>>> +  ReferenceMember(int &a, int &b, int &c, int &d)
>>> +    : a(a), b(b), c(c), d(d) {}
>>> +  int &a;
>>> +  int &b;
>>> +  NonPOD np;
>>> +  int &c;
>>> +  int &d;
>>> +};
>>> +
>>> +// COPY-ASSIGNMENT OPERATORS:
>>> +
>>> +// Assignment operators are output in the order they're encountered.
>>> +
>>> +#define CALL_AO(T) void callAO##T(T& a, const T& b) { a = b; }
>>> +
>>> +CALL_AO(Basic)
>>> +CALL_AO(PODMember)
>>> +CALL_AO(PODLikeMember)
>>> +CALL_AO(ArrayMember)
>>> +CALL_AO(VolatileMember)
>>> +CALL_AO(BitfieldMember)
>>> +CALL_AO(InnerClassMember)
>>> +
>>> +// Basic copy-assignment:
>>> +// CHECK: define linkonce_odr %struct.Basic*
>>> @_ZN5BasicaSERKS_(%struct.Basic* %this, %struct.Basic*)
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 4{{.*}})
>>> +// CHECK: tail call %struct.NonPOD* @_ZN6NonPODaSERKS_
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 4{{.*}})
>>> +// CHECK: ret %struct.Basic* %this
>>> +
>>> +// PODMember copy-assignment:
>>> +// CHECK: define linkonce_odr %struct.PODMember*
>>> @_ZN9PODMemberaSERKS_(%struct.PODMember* %this, %struct.PODMember*)
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 32, i32
>>> 4{{.*}})
>>> +// CHECK: tail call %struct.NonPOD* @_ZN6NonPODaSERKS_
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 4{{.*}})
>>> +// CHECK: ret %struct.PODMember* %this
>>> +
>>> +// PODLikeMember copy-assignment:
>>> +// CHECK: define linkonce_odr %struct.PODLikeMember*
>>> @_ZN13PODLikeMemberaSERKS_(%struct.PODLikeMember* %this,
>>> %struct.PODLikeMember*)
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 32, i32
>>> 4{{.*}})
>>> +// CHECK: tail call %struct.NonPOD* @_ZN6NonPODaSERKS_
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 4{{.*}})
>>> +// CHECK: ret %struct.PODLikeMember* %this
>>> +
>>> +// ArrayMember copy-assignment:
>>> +// CHECK: define linkonce_odr %struct.ArrayMember*
>>> @_ZN11ArrayMemberaSERKS_(%struct.ArrayMember* %this, %struct.ArrayMember*)
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 64, i32
>>> 4{{.*}})
>>> +// CHECK: tail call %struct.NonPOD* @_ZN6NonPODaSERKS_
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 64, i32
>>> 4{{.*}})
>>> +// CHECK: ret %struct.ArrayMember* %this
>>> +
>>> +// VolatileMember copy-assignment:
>>> +// CHECK: define linkonce_odr %struct.VolatileMember*
>>> @_ZN14VolatileMemberaSERKS_(%struct.VolatileMember* %this,
>>> %struct.VolatileMember*)
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 4{{.*}})
>>> +// CHECK: load volatile i32* {{.*}}, align 4
>>> +// CHECK: store volatile i32 {{.*}}, align 4
>>> +// CHECK: tail call %struct.NonPOD* @_ZN6NonPODaSERKS_
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 4{{.*}})
>>> +// CHECK: ret %struct.VolatileMember* %this
>>> +
>>> +// BitfieldMember copy-assignment:
>>> +// CHECK: define linkonce_odr %struct.BitfieldMember*
>>> @_ZN14BitfieldMemberaSERKS_(%struct.BitfieldMember* %this,
>>> %struct.BitfieldMember*)
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 4{{.*}})
>>> +// CHECK: tail call %struct.NonPOD* @_ZN6NonPODaSERKS_
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 3, i32
>>> 1{{.*}})
>>> +// CHECK: ret %struct.BitfieldMember* %this
>>> +
>>> +// InnerClass copy-assignment:
>>> +// CHECK: define linkonce_odr %struct.InnerClassMember*
>>> @_ZN16InnerClassMemberaSERKS_(%struct.InnerClassMember* %this,
>>> %struct.InnerClassMember*)
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 32, i32
>>> 4{{.*}})
>>> +// CHECK: tail call %struct.NonPOD* @_ZN6NonPODaSERKS_
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 4{{.*}})
>>> +// CHECK: ret %struct.InnerClassMember* %this
>>> +
>>> +// COPY-CONSTRUCTORS:
>>> +
>>> +// Clang outputs copy-constructors in the reverse of the order that
>>> +// copy-constructor calls are encountered. Add functions that call the
>>> copy
>>> +// constructors of the classes above in reverse order here.
>>> +
>>> +#define CALL_CC(T) T callCC##T(const T& b) { return b; }
>>> +
>>> +CALL_CC(ReferenceMember)
>>> +CALL_CC(InnerClassMember)
>>> +CALL_CC(BitfieldMember)
>>> +CALL_CC(VolatileMember)
>>> +CALL_CC(ArrayMember)
>>> +CALL_CC(PODLikeMember)
>>> +CALL_CC(PODMember)
>>> +CALL_CC(Basic)
>>> +
>>> +// Basic copy-constructor:
>>> +// CHECK: define linkonce_odr void @_ZN5BasicC2ERKS_(%struct.Basic*
>>> %this, %struct.Basic*)
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 4{{.*}})
>>> +// CHECK: tail call void @_ZN6NonPODC1ERKS_
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 4{{.*}})
>>> +// CHECK: ret void
>>> +
>>> +// PODMember copy-constructor:
>>> +// CHECK: define linkonce_odr void
>>> @_ZN9PODMemberC2ERKS_(%struct.PODMember* %this, %struct.PODMember*)
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 32, i32
>>> 4{{.*}})
>>> +// CHECK: tail call void @_ZN6NonPODC1ERKS_
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 4{{.*}})
>>> +// CHECK: ret void
>>> +
>>> +// PODLikeMember copy-constructor:
>>> +// CHECK: define linkonce_odr void
>>> @_ZN13PODLikeMemberC2ERKS_(%struct.PODLikeMember* %this,
>>> %struct.PODLikeMember*)
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 32, i32
>>> 4{{.*}})
>>> +// CHECK: invoke void @_ZN6NonPODC1ERKS_
>>> +// CHECK: invoke.cont:
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 4{{.*}})
>>> +// CHECK: ret void
>>> +// CHECK: lpad:
>>> +// CHECK: landingpad
>>> +// CHECK: invoke void @_ZN7PODLikeD1Ev
>>> +
>>> +// ArrayMember copy-constructor:
>>> +// CHECK: define linkonce_odr void
>>> @_ZN11ArrayMemberC2ERKS_(%struct.ArrayMember* %this, %struct.ArrayMember*)
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 64, i32
>>> 4{{.*}})
>>> +// CHECK: tail call void @_ZN6NonPODC1ERKS_
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 64, i32
>>> 4{{.*}})
>>> +// CHECK: ret void
>>> +
>>> +// VolatileMember copy-constructor:
>>> +// CHECK: define linkonce_odr void
>>> @_ZN14VolatileMemberC2ERKS_(%struct.VolatileMember* %this,
>>> %struct.VolatileMember*)
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 4{{.*}})
>>> +// CHECK: load volatile i32* {{.*}}, align 4
>>> +// CHECK: store volatile i32 {{.*}}, align 4
>>> +// CHECK: tail call void @_ZN6NonPODC1ERKS_
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 4{{.*}})
>>> +// CHECK: ret void
>>> +
>>> +// BitfieldMember copy-constructor:
>>> +// CHECK: define linkonce_odr void
>>> @_ZN14BitfieldMemberC2ERKS_(%struct.BitfieldMember* %this,
>>> %struct.BitfieldMember*)
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 4{{.*}})
>>> +// CHECK: tail call void @_ZN6NonPODC1ERKS_
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 3, i32
>>> 1{{.*}})
>>> +// CHECK: ret void
>>> +
>>> +// InnerClass copy-constructor:
>>> +// CHECK: define linkonce_odr void
>>> @_ZN16InnerClassMemberC2ERKS_(%struct.InnerClassMember* %this,
>>> %struct.InnerClassMember*)
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 32, i32
>>> 4{{.*}})
>>> +// CHECK: tail call void @_ZN6NonPODC1ERKS_
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 4{{.*}})
>>> +// CHECK: ret void
>>> +
>>> +// ReferenceMember copy-constructor:
>>> +// CHECK: define linkonce_odr void
>>> @_ZN15ReferenceMemberC2ERKS_(%struct.ReferenceMember* %this,
>>> %struct.ReferenceMember*)
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 8{{.*}})
>>> +// CHECK: tail call void @_ZN6NonPODC1ERKS_
>>> +// CHECK: tail call void @llvm.memcpy.p0i8.p0i8.i64({{.*}}i64 16, i32
>>> 8{{.*}})
>>> +// CHECK: ret void
>>>
>>>
>>> _______________________________________________
>>> cfe-commits mailing list
>>> cfe-commits at cs.uiuc.edu
>>> http://lists.cs.uiuc.edu/mailman/listinfo/cfe-commits
>>>
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.llvm.org/pipermail/cfe-commits/attachments/20130217/9d739555/attachment.html>


More information about the cfe-commits mailing list