[llvm-commits] [llvm] r154705 - in /llvm/trunk: lib/Target/X86/X86ISelLowering.cpp lib/Target/X86/X86ISelLowering.h test/CodeGen/X86/atomic_op.ll

Douglas Gregor dgregor at apple.com
Mon Apr 16 11:07:04 PDT 2012


On Apr 13, 2012, at 3:47 PM, Richard Smith <richard-llvm at metafoo.co.uk> wrote:

> Author: rsmith
> Date: Fri Apr 13 17:47:00 2012
> New Revision: 154705
> 
> URL: http://llvm.org/viewvc/llvm-project?rev=154705&view=rev
> Log:
> Fix X86 codegen for 'atomicrmw nand' to generate *x = ~(*x & y), not *x = ~*x & y.

The libc++ atomics tests are deadlocking now, which seems to have started with this changeā€¦

I'm seeing deadlocks with at least:

test/atomics/atomics.types.generic/integral.pass.cpp failed at run time
test/atomics/atomics.types.operations/atomics.types.operations.req/atomic_fetch_and.pass.cpp
test/atomics/atomics.types.operations/atomics.types.operations.req/atomic_fetch_and_explicit.pass.cpp
test/atomics/atomics.types.operations/atomics.types.operations.req/atomic_fetch_or.pass.cpp
test/atomics/atomics.types.operations/atomics.types.operations.req/atomic_fetch_or_explicit.pass.cpp
test/atomics/atomics.types.operations/atomics.types.operations.req/atomic_fetch_xor.pass.cpp
test/atomics/atomics.types.operations/atomics.types.operations.req/atomic_fetch_xor_explicit.pass.cpp

	- Doug

> Modified:
>    llvm/trunk/lib/Target/X86/X86ISelLowering.cpp
>    llvm/trunk/lib/Target/X86/X86ISelLowering.h
>    llvm/trunk/test/CodeGen/X86/atomic_op.ll
> 
> Modified: llvm/trunk/lib/Target/X86/X86ISelLowering.cpp
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/X86/X86ISelLowering.cpp?rev=154705&r1=154704&r2=154705&view=diff
> ==============================================================================
> --- llvm/trunk/lib/Target/X86/X86ISelLowering.cpp (original)
> +++ llvm/trunk/lib/Target/X86/X86ISelLowering.cpp Fri Apr 13 17:47:00 2012
> @@ -11298,14 +11298,15 @@
>                                                        unsigned notOpc,
>                                                        unsigned EAXreg,
>                                                  const TargetRegisterClass *RC,
> -                                                       bool invSrc) const {
> +                                                       bool Invert) const {
>   // For the atomic bitwise operator, we generate
>   //   thisMBB:
>   //   newMBB:
>   //     ld  t1 = [bitinstr.addr]
>   //     op  t2 = t1, [bitinstr.val]
> +  //     not t3 = t2  (if Invert)
>   //     mov EAX = t1
> -  //     lcs dest = [bitinstr.addr], t2  [EAX is implicit]
> +  //     lcs dest = [bitinstr.addr], t3  [EAX is implicit]
>   //     bz  newMBB
>   //     fallthrough -->nextMBB
>   const TargetInstrInfo *TII = getTargetMachine().getInstrInfo();
> @@ -11353,13 +11354,6 @@
>   for (int i=0; i <= lastAddrIndx; ++i)
>     (*MIB).addOperand(*argOpers[i]);
> 
> -  unsigned tt = F->getRegInfo().createVirtualRegister(RC);
> -  if (invSrc) {
> -    MIB = BuildMI(newMBB, dl, TII->get(notOpc), tt).addReg(t1);
> -  }
> -  else
> -    tt = t1;
> -
>   unsigned t2 = F->getRegInfo().createVirtualRegister(RC);
>   assert((argOpers[valArgIndx]->isReg() ||
>           argOpers[valArgIndx]->isImm()) &&
> @@ -11368,16 +11362,23 @@
>     MIB = BuildMI(newMBB, dl, TII->get(regOpc), t2);
>   else
>     MIB = BuildMI(newMBB, dl, TII->get(immOpc), t2);
> -  MIB.addReg(tt);
> +  MIB.addReg(t1);
>   (*MIB).addOperand(*argOpers[valArgIndx]);
> 
> +  unsigned t3 = F->getRegInfo().createVirtualRegister(RC);
> +  if (Invert) {
> +    MIB = BuildMI(newMBB, dl, TII->get(notOpc), t3).addReg(t2);
> +  }
> +  else
> +    t3 = t2;
> +
>   MIB = BuildMI(newMBB, dl, TII->get(TargetOpcode::COPY), EAXreg);
> -  MIB.addReg(t1);
> +  MIB.addReg(t3);
> 
>   MIB = BuildMI(newMBB, dl, TII->get(CXchgOpc));
>   for (int i=0; i <= lastAddrIndx; ++i)
>     (*MIB).addOperand(*argOpers[i]);
> -  MIB.addReg(t2);
> +  MIB.addReg(t3);
>   assert(bInstr->hasOneMemOperand() && "Unexpected number of memoperand");
>   (*MIB).setMemRefs(bInstr->memoperands_begin(),
>                     bInstr->memoperands_end());
> @@ -11400,7 +11401,7 @@
>                                                        unsigned regOpcH,
>                                                        unsigned immOpcL,
>                                                        unsigned immOpcH,
> -                                                       bool invSrc) const {
> +                                                       bool Invert) const {
>   // For the atomic bitwise operator, we generate
>   //   thisMBB (instructions are in pairs, except cmpxchg8b)
>   //     ld t1,t2 = [bitinstr.addr]
> @@ -11408,6 +11409,7 @@
>   //     out1, out2 = phi (thisMBB, t1/t2) (newMBB, t3/t4)
>   //     op  t5, t6 <- out1, out2, [bitinstr.val]
>   //      (for SWAP, substitute:  mov t5, t6 <- [bitinstr.val])
> +  //     neg t7, t8 < t5, t6  (if Invert)
>   //     mov ECX, EBX <- t5, t6
>   //     mov EAX, EDX <- t1, t2
>   //     cmpxchg8b [bitinstr.addr]  [EAX, EDX, EBX, ECX implicit]
> @@ -11491,16 +11493,9 @@
>     .addReg(t2).addMBB(thisMBB).addReg(t4).addMBB(newMBB);
> 
>   // The subsequent operations should be using the destination registers of
> -  //the PHI instructions.
> -  if (invSrc) {
> -    t1 = F->getRegInfo().createVirtualRegister(RC);
> -    t2 = F->getRegInfo().createVirtualRegister(RC);
> -    MIB = BuildMI(newMBB, dl, TII->get(NotOpc), t1).addReg(dest1Oper.getReg());
> -    MIB = BuildMI(newMBB, dl, TII->get(NotOpc), t2).addReg(dest2Oper.getReg());
> -  } else {
> -    t1 = dest1Oper.getReg();
> -    t2 = dest2Oper.getReg();
> -  }
> +  // the PHI instructions.
> +  t1 = dest1Oper.getReg();
> +  t2 = dest2Oper.getReg();
> 
>   int valArgIndx = lastAddrIndx + 1;
>   assert((argOpers[valArgIndx]->isReg() ||
> @@ -11527,15 +11522,26 @@
>     MIB.addReg(t2);
>   (*MIB).addOperand(*argOpers[valArgIndx + 1]);
> 
> +  unsigned t7, t8;
> +  if (Invert) {
> +    t7 = F->getRegInfo().createVirtualRegister(RC);
> +    t8 = F->getRegInfo().createVirtualRegister(RC);
> +    MIB = BuildMI(newMBB, dl, TII->get(NotOpc), t7).addReg(t5);
> +    MIB = BuildMI(newMBB, dl, TII->get(NotOpc), t8).addReg(t6);
> +  } else {
> +    t7 = t5;
> +    t8 = t6;
> +  }
> +
>   MIB = BuildMI(newMBB, dl, TII->get(TargetOpcode::COPY), X86::EAX);
>   MIB.addReg(t1);
>   MIB = BuildMI(newMBB, dl, TII->get(TargetOpcode::COPY), X86::EDX);
>   MIB.addReg(t2);
> 
>   MIB = BuildMI(newMBB, dl, TII->get(TargetOpcode::COPY), X86::EBX);
> -  MIB.addReg(t5);
> +  MIB.addReg(t7);
>   MIB = BuildMI(newMBB, dl, TII->get(TargetOpcode::COPY), X86::ECX);
> -  MIB.addReg(t6);
> +  MIB.addReg(t8);
> 
>   MIB = BuildMI(newMBB, dl, TII->get(X86::LCMPXCHG8B));
>   for (int i=0; i <= lastAddrIndx; ++i)
> 
> Modified: llvm/trunk/lib/Target/X86/X86ISelLowering.h
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/X86/X86ISelLowering.h?rev=154705&r1=154704&r2=154705&view=diff
> ==============================================================================
> --- llvm/trunk/lib/Target/X86/X86ISelLowering.h (original)
> +++ llvm/trunk/lib/Target/X86/X86ISelLowering.h Fri Apr 13 17:47:00 2012
> @@ -855,7 +855,7 @@
>                                                     unsigned notOpc,
>                                                     unsigned EAXreg,
>                                               const TargetRegisterClass *RC,
> -                                                    bool invSrc = false) const;
> +                                                    bool Invert = false) const;
> 
>     MachineBasicBlock *EmitAtomicBit6432WithCustomInserter(
>                                                     MachineInstr *BInstr,
> @@ -864,7 +864,7 @@
>                                                     unsigned regOpcH,
>                                                     unsigned immOpcL,
>                                                     unsigned immOpcH,
> -                                                    bool invSrc = false) const;
> +                                                    bool Invert = false) const;
> 
>     /// Utility function to emit atomic min and max.  It takes the min/max
>     /// instruction to expand, the associated basic block, and the associated
> 
> Modified: llvm/trunk/test/CodeGen/X86/atomic_op.ll
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/X86/atomic_op.ll?rev=154705&r1=154704&r2=154705&view=diff
> ==============================================================================
> --- llvm/trunk/test/CodeGen/X86/atomic_op.ll (original)
> +++ llvm/trunk/test/CodeGen/X86/atomic_op.ll Fri Apr 13 17:47:00 2012
> @@ -13,6 +13,7 @@
> 	%xort = alloca i32		; <i32*> [#uses=2]
> 	%old = alloca i32		; <i32*> [#uses=18]
> 	%temp = alloca i32		; <i32*> [#uses=2]
> +	%temp64 = alloca i64
> 	store i32 %argc, i32* %argc.addr
> 	store i8** %argv, i8*** %argv.addr
> 	store i32 0, i32* %val1
> @@ -106,6 +107,20 @@
>         ; CHECK: cmpxchgl
>   %17 = cmpxchg i32* %val2, i32 1976, i32 1 monotonic
> 	store i32 %17, i32* %old
> +        ; CHECK: andl
> +        ; CHECK: notl
> +        ; CHECK: lock
> +        ; CHECK: cmpxchgl
> +  %18 = atomicrmw nand i32* %val2, i32 1401 monotonic
> +  store i32 %18, i32* %old
> +        ; CHECK: andl
> +        ; CHECK: andl
> +        ; CHECK: notl
> +        ; CHECK: notl
> +        ; CHECK: lock
> +        ; CHECK: cmpxchg8b
> +  %19 = atomicrmw nand i64* %temp64, i64 17361641481138401520 monotonic
> +  store i64 %19, i64* %temp64
> 	ret void
> }
> 
> 
> 
> _______________________________________________
> llvm-commits mailing list
> llvm-commits at cs.uiuc.edu
> http://lists.cs.uiuc.edu/mailman/listinfo/llvm-commits





More information about the llvm-commits mailing list