[llvm-commits] [llvm] r167615 - in /llvm/trunk: lib/Transforms/Instrumentation/ThreadSanitizer.cpp test/Instrumentation/ThreadSanitizer/atomic.ll

Dmitry Vyukov dvyukov at google.com
Fri Nov 9 06:12:16 PST 2012


Author: dvyukov
Date: Fri Nov  9 08:12:16 2012
New Revision: 167615

URL: http://llvm.org/viewvc/llvm-project?rev=167615&view=rev
Log:
tsan: switch to new memory_order constants (ABI compatible)

Modified:
    llvm/trunk/lib/Transforms/Instrumentation/ThreadSanitizer.cpp
    llvm/trunk/test/Instrumentation/ThreadSanitizer/atomic.ll

Modified: llvm/trunk/lib/Transforms/Instrumentation/ThreadSanitizer.cpp
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Transforms/Instrumentation/ThreadSanitizer.cpp?rev=167615&r1=167614&r2=167615&view=diff
==============================================================================
--- llvm/trunk/lib/Transforms/Instrumentation/ThreadSanitizer.cpp (original)
+++ llvm/trunk/lib/Transforms/Instrumentation/ThreadSanitizer.cpp Fri Nov  9 08:12:16 2012
@@ -390,15 +390,14 @@
   switch (ord) {
     case NotAtomic:              assert(false);
     case Unordered:              // Fall-through.
-    case Monotonic:              v = 1 << 0; break;
-    // case Consume:                v = 1 << 1; break;  // Not specified yet.
-    case Acquire:                v = 1 << 2; break;
-    case Release:                v = 1 << 3; break;
-    case AcquireRelease:         v = 1 << 4; break;
-    case SequentiallyConsistent: v = 1 << 5; break;
+    case Monotonic:              v = 0; break;
+ // case Consume:                v = 1; break;  // Not specified yet.
+    case Acquire:                v = 2; break;
+    case Release:                v = 3; break;
+    case AcquireRelease:         v = 4; break;
+    case SequentiallyConsistent: v = 5; break;
   }
-  // +100500 is temporal to migrate to new enum values.
-  return IRB->getInt32(v + 100500);
+  return IRB->getInt32(v);
 }
 
 bool ThreadSanitizer::instrumentAtomic(Instruction *I) {

Modified: llvm/trunk/test/Instrumentation/ThreadSanitizer/atomic.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/Instrumentation/ThreadSanitizer/atomic.ll?rev=167615&r1=167614&r2=167615&view=diff
==============================================================================
--- llvm/trunk/test/Instrumentation/ThreadSanitizer/atomic.ll (original)
+++ llvm/trunk/test/Instrumentation/ThreadSanitizer/atomic.ll Fri Nov  9 08:12:16 2012
@@ -8,7 +8,7 @@
   ret i8 %0
 }
 ; CHECK: atomic8_load_unordered
-; CHECK: call i8 @__tsan_atomic8_load(i8* %a, i32 100501)
+; CHECK: call i8 @__tsan_atomic8_load(i8* %a, i32 0)
 
 define i8 @atomic8_load_monotonic(i8* %a) nounwind uwtable {
 entry:
@@ -16,7 +16,7 @@
   ret i8 %0
 }
 ; CHECK: atomic8_load_monotonic
-; CHECK: call i8 @__tsan_atomic8_load(i8* %a, i32 100501)
+; CHECK: call i8 @__tsan_atomic8_load(i8* %a, i32 0)
 
 define i8 @atomic8_load_acquire(i8* %a) nounwind uwtable {
 entry:
@@ -24,7 +24,7 @@
   ret i8 %0
 }
 ; CHECK: atomic8_load_acquire
-; CHECK: call i8 @__tsan_atomic8_load(i8* %a, i32 100504)
+; CHECK: call i8 @__tsan_atomic8_load(i8* %a, i32 2)
 
 define i8 @atomic8_load_seq_cst(i8* %a) nounwind uwtable {
 entry:
@@ -32,7 +32,7 @@
   ret i8 %0
 }
 ; CHECK: atomic8_load_seq_cst
-; CHECK: call i8 @__tsan_atomic8_load(i8* %a, i32 100532)
+; CHECK: call i8 @__tsan_atomic8_load(i8* %a, i32 5)
 
 define void @atomic8_store_unordered(i8* %a) nounwind uwtable {
 entry:
@@ -40,7 +40,7 @@
   ret void
 }
 ; CHECK: atomic8_store_unordered
-; CHECK: call void @__tsan_atomic8_store(i8* %a, i8 0, i32 100501)
+; CHECK: call void @__tsan_atomic8_store(i8* %a, i8 0, i32 0)
 
 define void @atomic8_store_monotonic(i8* %a) nounwind uwtable {
 entry:
@@ -48,7 +48,7 @@
   ret void
 }
 ; CHECK: atomic8_store_monotonic
-; CHECK: call void @__tsan_atomic8_store(i8* %a, i8 0, i32 100501)
+; CHECK: call void @__tsan_atomic8_store(i8* %a, i8 0, i32 0)
 
 define void @atomic8_store_release(i8* %a) nounwind uwtable {
 entry:
@@ -56,7 +56,7 @@
   ret void
 }
 ; CHECK: atomic8_store_release
-; CHECK: call void @__tsan_atomic8_store(i8* %a, i8 0, i32 100508)
+; CHECK: call void @__tsan_atomic8_store(i8* %a, i8 0, i32 3)
 
 define void @atomic8_store_seq_cst(i8* %a) nounwind uwtable {
 entry:
@@ -64,7 +64,7 @@
   ret void
 }
 ; CHECK: atomic8_store_seq_cst
-; CHECK: call void @__tsan_atomic8_store(i8* %a, i8 0, i32 100532)
+; CHECK: call void @__tsan_atomic8_store(i8* %a, i8 0, i32 5)
 
 define void @atomic8_xchg_monotonic(i8* %a) nounwind uwtable {
 entry:
@@ -72,7 +72,7 @@
   ret void
 }
 ; CHECK: atomic8_xchg_monotonic
-; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 100501)
+; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 0)
 
 define void @atomic8_add_monotonic(i8* %a) nounwind uwtable {
 entry:
@@ -80,7 +80,7 @@
   ret void
 }
 ; CHECK: atomic8_add_monotonic
-; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 100501)
+; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 0)
 
 define void @atomic8_sub_monotonic(i8* %a) nounwind uwtable {
 entry:
@@ -88,7 +88,7 @@
   ret void
 }
 ; CHECK: atomic8_sub_monotonic
-; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 100501)
+; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 0)
 
 define void @atomic8_and_monotonic(i8* %a) nounwind uwtable {
 entry:
@@ -96,7 +96,7 @@
   ret void
 }
 ; CHECK: atomic8_and_monotonic
-; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 100501)
+; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 0)
 
 define void @atomic8_or_monotonic(i8* %a) nounwind uwtable {
 entry:
@@ -104,7 +104,7 @@
   ret void
 }
 ; CHECK: atomic8_or_monotonic
-; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 100501)
+; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 0)
 
 define void @atomic8_xor_monotonic(i8* %a) nounwind uwtable {
 entry:
@@ -112,7 +112,7 @@
   ret void
 }
 ; CHECK: atomic8_xor_monotonic
-; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 100501)
+; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 0)
 
 define void @atomic8_xchg_acquire(i8* %a) nounwind uwtable {
 entry:
@@ -120,7 +120,7 @@
   ret void
 }
 ; CHECK: atomic8_xchg_acquire
-; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 100504)
+; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 2)
 
 define void @atomic8_add_acquire(i8* %a) nounwind uwtable {
 entry:
@@ -128,7 +128,7 @@
   ret void
 }
 ; CHECK: atomic8_add_acquire
-; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 100504)
+; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 2)
 
 define void @atomic8_sub_acquire(i8* %a) nounwind uwtable {
 entry:
@@ -136,7 +136,7 @@
   ret void
 }
 ; CHECK: atomic8_sub_acquire
-; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 100504)
+; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 2)
 
 define void @atomic8_and_acquire(i8* %a) nounwind uwtable {
 entry:
@@ -144,7 +144,7 @@
   ret void
 }
 ; CHECK: atomic8_and_acquire
-; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 100504)
+; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 2)
 
 define void @atomic8_or_acquire(i8* %a) nounwind uwtable {
 entry:
@@ -152,7 +152,7 @@
   ret void
 }
 ; CHECK: atomic8_or_acquire
-; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 100504)
+; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 2)
 
 define void @atomic8_xor_acquire(i8* %a) nounwind uwtable {
 entry:
@@ -160,7 +160,7 @@
   ret void
 }
 ; CHECK: atomic8_xor_acquire
-; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 100504)
+; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 2)
 
 define void @atomic8_xchg_release(i8* %a) nounwind uwtable {
 entry:
@@ -168,7 +168,7 @@
   ret void
 }
 ; CHECK: atomic8_xchg_release
-; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 100508)
+; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 3)
 
 define void @atomic8_add_release(i8* %a) nounwind uwtable {
 entry:
@@ -176,7 +176,7 @@
   ret void
 }
 ; CHECK: atomic8_add_release
-; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 100508)
+; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 3)
 
 define void @atomic8_sub_release(i8* %a) nounwind uwtable {
 entry:
@@ -184,7 +184,7 @@
   ret void
 }
 ; CHECK: atomic8_sub_release
-; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 100508)
+; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 3)
 
 define void @atomic8_and_release(i8* %a) nounwind uwtable {
 entry:
@@ -192,7 +192,7 @@
   ret void
 }
 ; CHECK: atomic8_and_release
-; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 100508)
+; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 3)
 
 define void @atomic8_or_release(i8* %a) nounwind uwtable {
 entry:
@@ -200,7 +200,7 @@
   ret void
 }
 ; CHECK: atomic8_or_release
-; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 100508)
+; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 3)
 
 define void @atomic8_xor_release(i8* %a) nounwind uwtable {
 entry:
@@ -208,7 +208,7 @@
   ret void
 }
 ; CHECK: atomic8_xor_release
-; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 100508)
+; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 3)
 
 define void @atomic8_xchg_acq_rel(i8* %a) nounwind uwtable {
 entry:
@@ -216,7 +216,7 @@
   ret void
 }
 ; CHECK: atomic8_xchg_acq_rel
-; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 100516)
+; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 4)
 
 define void @atomic8_add_acq_rel(i8* %a) nounwind uwtable {
 entry:
@@ -224,7 +224,7 @@
   ret void
 }
 ; CHECK: atomic8_add_acq_rel
-; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 100516)
+; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 4)
 
 define void @atomic8_sub_acq_rel(i8* %a) nounwind uwtable {
 entry:
@@ -232,7 +232,7 @@
   ret void
 }
 ; CHECK: atomic8_sub_acq_rel
-; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 100516)
+; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 4)
 
 define void @atomic8_and_acq_rel(i8* %a) nounwind uwtable {
 entry:
@@ -240,7 +240,7 @@
   ret void
 }
 ; CHECK: atomic8_and_acq_rel
-; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 100516)
+; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 4)
 
 define void @atomic8_or_acq_rel(i8* %a) nounwind uwtable {
 entry:
@@ -248,7 +248,7 @@
   ret void
 }
 ; CHECK: atomic8_or_acq_rel
-; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 100516)
+; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 4)
 
 define void @atomic8_xor_acq_rel(i8* %a) nounwind uwtable {
 entry:
@@ -256,7 +256,7 @@
   ret void
 }
 ; CHECK: atomic8_xor_acq_rel
-; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 100516)
+; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 4)
 
 define void @atomic8_xchg_seq_cst(i8* %a) nounwind uwtable {
 entry:
@@ -264,7 +264,7 @@
   ret void
 }
 ; CHECK: atomic8_xchg_seq_cst
-; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 100532)
+; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 5)
 
 define void @atomic8_add_seq_cst(i8* %a) nounwind uwtable {
 entry:
@@ -272,7 +272,7 @@
   ret void
 }
 ; CHECK: atomic8_add_seq_cst
-; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 100532)
+; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 5)
 
 define void @atomic8_sub_seq_cst(i8* %a) nounwind uwtable {
 entry:
@@ -280,7 +280,7 @@
   ret void
 }
 ; CHECK: atomic8_sub_seq_cst
-; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 100532)
+; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 5)
 
 define void @atomic8_and_seq_cst(i8* %a) nounwind uwtable {
 entry:
@@ -288,7 +288,7 @@
   ret void
 }
 ; CHECK: atomic8_and_seq_cst
-; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 100532)
+; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 5)
 
 define void @atomic8_or_seq_cst(i8* %a) nounwind uwtable {
 entry:
@@ -296,7 +296,7 @@
   ret void
 }
 ; CHECK: atomic8_or_seq_cst
-; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 100532)
+; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 5)
 
 define void @atomic8_xor_seq_cst(i8* %a) nounwind uwtable {
 entry:
@@ -304,7 +304,7 @@
   ret void
 }
 ; CHECK: atomic8_xor_seq_cst
-; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 100532)
+; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 5)
 
 define void @atomic8_cas_monotonic(i8* %a) nounwind uwtable {
 entry:
@@ -312,7 +312,7 @@
   ret void
 }
 ; CHECK: atomic8_cas_monotonic
-; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 100501)
+; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 0)
 
 define void @atomic8_cas_acquire(i8* %a) nounwind uwtable {
 entry:
@@ -320,7 +320,7 @@
   ret void
 }
 ; CHECK: atomic8_cas_acquire
-; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 100504)
+; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 2)
 
 define void @atomic8_cas_release(i8* %a) nounwind uwtable {
 entry:
@@ -328,7 +328,7 @@
   ret void
 }
 ; CHECK: atomic8_cas_release
-; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 100508)
+; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 3)
 
 define void @atomic8_cas_acq_rel(i8* %a) nounwind uwtable {
 entry:
@@ -336,7 +336,7 @@
   ret void
 }
 ; CHECK: atomic8_cas_acq_rel
-; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 100516)
+; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 4)
 
 define void @atomic8_cas_seq_cst(i8* %a) nounwind uwtable {
 entry:
@@ -344,7 +344,7 @@
   ret void
 }
 ; CHECK: atomic8_cas_seq_cst
-; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 100532)
+; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 5)
 
 define i16 @atomic16_load_unordered(i16* %a) nounwind uwtable {
 entry:
@@ -352,7 +352,7 @@
   ret i16 %0
 }
 ; CHECK: atomic16_load_unordered
-; CHECK: call i16 @__tsan_atomic16_load(i16* %a, i32 100501)
+; CHECK: call i16 @__tsan_atomic16_load(i16* %a, i32 0)
 
 define i16 @atomic16_load_monotonic(i16* %a) nounwind uwtable {
 entry:
@@ -360,7 +360,7 @@
   ret i16 %0
 }
 ; CHECK: atomic16_load_monotonic
-; CHECK: call i16 @__tsan_atomic16_load(i16* %a, i32 100501)
+; CHECK: call i16 @__tsan_atomic16_load(i16* %a, i32 0)
 
 define i16 @atomic16_load_acquire(i16* %a) nounwind uwtable {
 entry:
@@ -368,7 +368,7 @@
   ret i16 %0
 }
 ; CHECK: atomic16_load_acquire
-; CHECK: call i16 @__tsan_atomic16_load(i16* %a, i32 100504)
+; CHECK: call i16 @__tsan_atomic16_load(i16* %a, i32 2)
 
 define i16 @atomic16_load_seq_cst(i16* %a) nounwind uwtable {
 entry:
@@ -376,7 +376,7 @@
   ret i16 %0
 }
 ; CHECK: atomic16_load_seq_cst
-; CHECK: call i16 @__tsan_atomic16_load(i16* %a, i32 100532)
+; CHECK: call i16 @__tsan_atomic16_load(i16* %a, i32 5)
 
 define void @atomic16_store_unordered(i16* %a) nounwind uwtable {
 entry:
@@ -384,7 +384,7 @@
   ret void
 }
 ; CHECK: atomic16_store_unordered
-; CHECK: call void @__tsan_atomic16_store(i16* %a, i16 0, i32 100501)
+; CHECK: call void @__tsan_atomic16_store(i16* %a, i16 0, i32 0)
 
 define void @atomic16_store_monotonic(i16* %a) nounwind uwtable {
 entry:
@@ -392,7 +392,7 @@
   ret void
 }
 ; CHECK: atomic16_store_monotonic
-; CHECK: call void @__tsan_atomic16_store(i16* %a, i16 0, i32 100501)
+; CHECK: call void @__tsan_atomic16_store(i16* %a, i16 0, i32 0)
 
 define void @atomic16_store_release(i16* %a) nounwind uwtable {
 entry:
@@ -400,7 +400,7 @@
   ret void
 }
 ; CHECK: atomic16_store_release
-; CHECK: call void @__tsan_atomic16_store(i16* %a, i16 0, i32 100508)
+; CHECK: call void @__tsan_atomic16_store(i16* %a, i16 0, i32 3)
 
 define void @atomic16_store_seq_cst(i16* %a) nounwind uwtable {
 entry:
@@ -408,7 +408,7 @@
   ret void
 }
 ; CHECK: atomic16_store_seq_cst
-; CHECK: call void @__tsan_atomic16_store(i16* %a, i16 0, i32 100532)
+; CHECK: call void @__tsan_atomic16_store(i16* %a, i16 0, i32 5)
 
 define void @atomic16_xchg_monotonic(i16* %a) nounwind uwtable {
 entry:
@@ -416,7 +416,7 @@
   ret void
 }
 ; CHECK: atomic16_xchg_monotonic
-; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 100501)
+; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 0)
 
 define void @atomic16_add_monotonic(i16* %a) nounwind uwtable {
 entry:
@@ -424,7 +424,7 @@
   ret void
 }
 ; CHECK: atomic16_add_monotonic
-; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 100501)
+; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 0)
 
 define void @atomic16_sub_monotonic(i16* %a) nounwind uwtable {
 entry:
@@ -432,7 +432,7 @@
   ret void
 }
 ; CHECK: atomic16_sub_monotonic
-; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 100501)
+; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 0)
 
 define void @atomic16_and_monotonic(i16* %a) nounwind uwtable {
 entry:
@@ -440,7 +440,7 @@
   ret void
 }
 ; CHECK: atomic16_and_monotonic
-; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 100501)
+; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 0)
 
 define void @atomic16_or_monotonic(i16* %a) nounwind uwtable {
 entry:
@@ -448,7 +448,7 @@
   ret void
 }
 ; CHECK: atomic16_or_monotonic
-; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 100501)
+; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 0)
 
 define void @atomic16_xor_monotonic(i16* %a) nounwind uwtable {
 entry:
@@ -456,7 +456,7 @@
   ret void
 }
 ; CHECK: atomic16_xor_monotonic
-; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 100501)
+; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 0)
 
 define void @atomic16_xchg_acquire(i16* %a) nounwind uwtable {
 entry:
@@ -464,7 +464,7 @@
   ret void
 }
 ; CHECK: atomic16_xchg_acquire
-; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 100504)
+; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 2)
 
 define void @atomic16_add_acquire(i16* %a) nounwind uwtable {
 entry:
@@ -472,7 +472,7 @@
   ret void
 }
 ; CHECK: atomic16_add_acquire
-; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 100504)
+; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 2)
 
 define void @atomic16_sub_acquire(i16* %a) nounwind uwtable {
 entry:
@@ -480,7 +480,7 @@
   ret void
 }
 ; CHECK: atomic16_sub_acquire
-; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 100504)
+; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 2)
 
 define void @atomic16_and_acquire(i16* %a) nounwind uwtable {
 entry:
@@ -488,7 +488,7 @@
   ret void
 }
 ; CHECK: atomic16_and_acquire
-; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 100504)
+; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 2)
 
 define void @atomic16_or_acquire(i16* %a) nounwind uwtable {
 entry:
@@ -496,7 +496,7 @@
   ret void
 }
 ; CHECK: atomic16_or_acquire
-; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 100504)
+; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 2)
 
 define void @atomic16_xor_acquire(i16* %a) nounwind uwtable {
 entry:
@@ -504,7 +504,7 @@
   ret void
 }
 ; CHECK: atomic16_xor_acquire
-; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 100504)
+; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 2)
 
 define void @atomic16_xchg_release(i16* %a) nounwind uwtable {
 entry:
@@ -512,7 +512,7 @@
   ret void
 }
 ; CHECK: atomic16_xchg_release
-; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 100508)
+; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 3)
 
 define void @atomic16_add_release(i16* %a) nounwind uwtable {
 entry:
@@ -520,7 +520,7 @@
   ret void
 }
 ; CHECK: atomic16_add_release
-; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 100508)
+; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 3)
 
 define void @atomic16_sub_release(i16* %a) nounwind uwtable {
 entry:
@@ -528,7 +528,7 @@
   ret void
 }
 ; CHECK: atomic16_sub_release
-; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 100508)
+; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 3)
 
 define void @atomic16_and_release(i16* %a) nounwind uwtable {
 entry:
@@ -536,7 +536,7 @@
   ret void
 }
 ; CHECK: atomic16_and_release
-; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 100508)
+; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 3)
 
 define void @atomic16_or_release(i16* %a) nounwind uwtable {
 entry:
@@ -544,7 +544,7 @@
   ret void
 }
 ; CHECK: atomic16_or_release
-; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 100508)
+; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 3)
 
 define void @atomic16_xor_release(i16* %a) nounwind uwtable {
 entry:
@@ -552,7 +552,7 @@
   ret void
 }
 ; CHECK: atomic16_xor_release
-; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 100508)
+; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 3)
 
 define void @atomic16_xchg_acq_rel(i16* %a) nounwind uwtable {
 entry:
@@ -560,7 +560,7 @@
   ret void
 }
 ; CHECK: atomic16_xchg_acq_rel
-; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 100516)
+; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 4)
 
 define void @atomic16_add_acq_rel(i16* %a) nounwind uwtable {
 entry:
@@ -568,7 +568,7 @@
   ret void
 }
 ; CHECK: atomic16_add_acq_rel
-; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 100516)
+; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 4)
 
 define void @atomic16_sub_acq_rel(i16* %a) nounwind uwtable {
 entry:
@@ -576,7 +576,7 @@
   ret void
 }
 ; CHECK: atomic16_sub_acq_rel
-; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 100516)
+; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 4)
 
 define void @atomic16_and_acq_rel(i16* %a) nounwind uwtable {
 entry:
@@ -584,7 +584,7 @@
   ret void
 }
 ; CHECK: atomic16_and_acq_rel
-; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 100516)
+; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 4)
 
 define void @atomic16_or_acq_rel(i16* %a) nounwind uwtable {
 entry:
@@ -592,7 +592,7 @@
   ret void
 }
 ; CHECK: atomic16_or_acq_rel
-; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 100516)
+; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 4)
 
 define void @atomic16_xor_acq_rel(i16* %a) nounwind uwtable {
 entry:
@@ -600,7 +600,7 @@
   ret void
 }
 ; CHECK: atomic16_xor_acq_rel
-; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 100516)
+; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 4)
 
 define void @atomic16_xchg_seq_cst(i16* %a) nounwind uwtable {
 entry:
@@ -608,7 +608,7 @@
   ret void
 }
 ; CHECK: atomic16_xchg_seq_cst
-; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 100532)
+; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 5)
 
 define void @atomic16_add_seq_cst(i16* %a) nounwind uwtable {
 entry:
@@ -616,7 +616,7 @@
   ret void
 }
 ; CHECK: atomic16_add_seq_cst
-; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 100532)
+; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 5)
 
 define void @atomic16_sub_seq_cst(i16* %a) nounwind uwtable {
 entry:
@@ -624,7 +624,7 @@
   ret void
 }
 ; CHECK: atomic16_sub_seq_cst
-; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 100532)
+; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 5)
 
 define void @atomic16_and_seq_cst(i16* %a) nounwind uwtable {
 entry:
@@ -632,7 +632,7 @@
   ret void
 }
 ; CHECK: atomic16_and_seq_cst
-; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 100532)
+; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 5)
 
 define void @atomic16_or_seq_cst(i16* %a) nounwind uwtable {
 entry:
@@ -640,7 +640,7 @@
   ret void
 }
 ; CHECK: atomic16_or_seq_cst
-; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 100532)
+; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 5)
 
 define void @atomic16_xor_seq_cst(i16* %a) nounwind uwtable {
 entry:
@@ -648,7 +648,7 @@
   ret void
 }
 ; CHECK: atomic16_xor_seq_cst
-; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 100532)
+; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 5)
 
 define void @atomic16_cas_monotonic(i16* %a) nounwind uwtable {
 entry:
@@ -656,7 +656,7 @@
   ret void
 }
 ; CHECK: atomic16_cas_monotonic
-; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 100501)
+; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 0)
 
 define void @atomic16_cas_acquire(i16* %a) nounwind uwtable {
 entry:
@@ -664,7 +664,7 @@
   ret void
 }
 ; CHECK: atomic16_cas_acquire
-; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 100504)
+; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 2)
 
 define void @atomic16_cas_release(i16* %a) nounwind uwtable {
 entry:
@@ -672,7 +672,7 @@
   ret void
 }
 ; CHECK: atomic16_cas_release
-; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 100508)
+; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 3)
 
 define void @atomic16_cas_acq_rel(i16* %a) nounwind uwtable {
 entry:
@@ -680,7 +680,7 @@
   ret void
 }
 ; CHECK: atomic16_cas_acq_rel
-; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 100516)
+; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 4)
 
 define void @atomic16_cas_seq_cst(i16* %a) nounwind uwtable {
 entry:
@@ -688,7 +688,7 @@
   ret void
 }
 ; CHECK: atomic16_cas_seq_cst
-; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 100532)
+; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 5)
 
 define i32 @atomic32_load_unordered(i32* %a) nounwind uwtable {
 entry:
@@ -696,7 +696,7 @@
   ret i32 %0
 }
 ; CHECK: atomic32_load_unordered
-; CHECK: call i32 @__tsan_atomic32_load(i32* %a, i32 100501)
+; CHECK: call i32 @__tsan_atomic32_load(i32* %a, i32 0)
 
 define i32 @atomic32_load_monotonic(i32* %a) nounwind uwtable {
 entry:
@@ -704,7 +704,7 @@
   ret i32 %0
 }
 ; CHECK: atomic32_load_monotonic
-; CHECK: call i32 @__tsan_atomic32_load(i32* %a, i32 100501)
+; CHECK: call i32 @__tsan_atomic32_load(i32* %a, i32 0)
 
 define i32 @atomic32_load_acquire(i32* %a) nounwind uwtable {
 entry:
@@ -712,7 +712,7 @@
   ret i32 %0
 }
 ; CHECK: atomic32_load_acquire
-; CHECK: call i32 @__tsan_atomic32_load(i32* %a, i32 100504)
+; CHECK: call i32 @__tsan_atomic32_load(i32* %a, i32 2)
 
 define i32 @atomic32_load_seq_cst(i32* %a) nounwind uwtable {
 entry:
@@ -720,7 +720,7 @@
   ret i32 %0
 }
 ; CHECK: atomic32_load_seq_cst
-; CHECK: call i32 @__tsan_atomic32_load(i32* %a, i32 100532)
+; CHECK: call i32 @__tsan_atomic32_load(i32* %a, i32 5)
 
 define void @atomic32_store_unordered(i32* %a) nounwind uwtable {
 entry:
@@ -728,7 +728,7 @@
   ret void
 }
 ; CHECK: atomic32_store_unordered
-; CHECK: call void @__tsan_atomic32_store(i32* %a, i32 0, i32 100501)
+; CHECK: call void @__tsan_atomic32_store(i32* %a, i32 0, i32 0)
 
 define void @atomic32_store_monotonic(i32* %a) nounwind uwtable {
 entry:
@@ -736,7 +736,7 @@
   ret void
 }
 ; CHECK: atomic32_store_monotonic
-; CHECK: call void @__tsan_atomic32_store(i32* %a, i32 0, i32 100501)
+; CHECK: call void @__tsan_atomic32_store(i32* %a, i32 0, i32 0)
 
 define void @atomic32_store_release(i32* %a) nounwind uwtable {
 entry:
@@ -744,7 +744,7 @@
   ret void
 }
 ; CHECK: atomic32_store_release
-; CHECK: call void @__tsan_atomic32_store(i32* %a, i32 0, i32 100508)
+; CHECK: call void @__tsan_atomic32_store(i32* %a, i32 0, i32 3)
 
 define void @atomic32_store_seq_cst(i32* %a) nounwind uwtable {
 entry:
@@ -752,7 +752,7 @@
   ret void
 }
 ; CHECK: atomic32_store_seq_cst
-; CHECK: call void @__tsan_atomic32_store(i32* %a, i32 0, i32 100532)
+; CHECK: call void @__tsan_atomic32_store(i32* %a, i32 0, i32 5)
 
 define void @atomic32_xchg_monotonic(i32* %a) nounwind uwtable {
 entry:
@@ -760,7 +760,7 @@
   ret void
 }
 ; CHECK: atomic32_xchg_monotonic
-; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 100501)
+; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 0)
 
 define void @atomic32_add_monotonic(i32* %a) nounwind uwtable {
 entry:
@@ -768,7 +768,7 @@
   ret void
 }
 ; CHECK: atomic32_add_monotonic
-; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 100501)
+; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 0)
 
 define void @atomic32_sub_monotonic(i32* %a) nounwind uwtable {
 entry:
@@ -776,7 +776,7 @@
   ret void
 }
 ; CHECK: atomic32_sub_monotonic
-; CHECK: call i32 @__tsan_atomic32_fetch_sub(i32* %a, i32 0, i32 100501)
+; CHECK: call i32 @__tsan_atomic32_fetch_sub(i32* %a, i32 0, i32 0)
 
 define void @atomic32_and_monotonic(i32* %a) nounwind uwtable {
 entry:
@@ -784,7 +784,7 @@
   ret void
 }
 ; CHECK: atomic32_and_monotonic
-; CHECK: call i32 @__tsan_atomic32_fetch_and(i32* %a, i32 0, i32 100501)
+; CHECK: call i32 @__tsan_atomic32_fetch_and(i32* %a, i32 0, i32 0)
 
 define void @atomic32_or_monotonic(i32* %a) nounwind uwtable {
 entry:
@@ -792,7 +792,7 @@
   ret void
 }
 ; CHECK: atomic32_or_monotonic
-; CHECK: call i32 @__tsan_atomic32_fetch_or(i32* %a, i32 0, i32 100501)
+; CHECK: call i32 @__tsan_atomic32_fetch_or(i32* %a, i32 0, i32 0)
 
 define void @atomic32_xor_monotonic(i32* %a) nounwind uwtable {
 entry:
@@ -800,7 +800,7 @@
   ret void
 }
 ; CHECK: atomic32_xor_monotonic
-; CHECK: call i32 @__tsan_atomic32_fetch_xor(i32* %a, i32 0, i32 100501)
+; CHECK: call i32 @__tsan_atomic32_fetch_xor(i32* %a, i32 0, i32 0)
 
 define void @atomic32_xchg_acquire(i32* %a) nounwind uwtable {
 entry:
@@ -808,7 +808,7 @@
   ret void
 }
 ; CHECK: atomic32_xchg_acquire
-; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 100504)
+; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 2)
 
 define void @atomic32_add_acquire(i32* %a) nounwind uwtable {
 entry:
@@ -816,7 +816,7 @@
   ret void
 }
 ; CHECK: atomic32_add_acquire
-; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 100504)
+; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 2)
 
 define void @atomic32_sub_acquire(i32* %a) nounwind uwtable {
 entry:
@@ -824,7 +824,7 @@
   ret void
 }
 ; CHECK: atomic32_sub_acquire
-; CHECK: call i32 @__tsan_atomic32_fetch_sub(i32* %a, i32 0, i32 100504)
+; CHECK: call i32 @__tsan_atomic32_fetch_sub(i32* %a, i32 0, i32 2)
 
 define void @atomic32_and_acquire(i32* %a) nounwind uwtable {
 entry:
@@ -832,7 +832,7 @@
   ret void
 }
 ; CHECK: atomic32_and_acquire
-; CHECK: call i32 @__tsan_atomic32_fetch_and(i32* %a, i32 0, i32 100504)
+; CHECK: call i32 @__tsan_atomic32_fetch_and(i32* %a, i32 0, i32 2)
 
 define void @atomic32_or_acquire(i32* %a) nounwind uwtable {
 entry:
@@ -840,7 +840,7 @@
   ret void
 }
 ; CHECK: atomic32_or_acquire
-; CHECK: call i32 @__tsan_atomic32_fetch_or(i32* %a, i32 0, i32 100504)
+; CHECK: call i32 @__tsan_atomic32_fetch_or(i32* %a, i32 0, i32 2)
 
 define void @atomic32_xor_acquire(i32* %a) nounwind uwtable {
 entry:
@@ -848,7 +848,7 @@
   ret void
 }
 ; CHECK: atomic32_xor_acquire
-; CHECK: call i32 @__tsan_atomic32_fetch_xor(i32* %a, i32 0, i32 100504)
+; CHECK: call i32 @__tsan_atomic32_fetch_xor(i32* %a, i32 0, i32 2)
 
 define void @atomic32_xchg_release(i32* %a) nounwind uwtable {
 entry:
@@ -856,7 +856,7 @@
   ret void
 }
 ; CHECK: atomic32_xchg_release
-; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 100508)
+; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 3)
 
 define void @atomic32_add_release(i32* %a) nounwind uwtable {
 entry:
@@ -864,7 +864,7 @@
   ret void
 }
 ; CHECK: atomic32_add_release
-; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 100508)
+; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 3)
 
 define void @atomic32_sub_release(i32* %a) nounwind uwtable {
 entry:
@@ -872,7 +872,7 @@
   ret void
 }
 ; CHECK: atomic32_sub_release
-; CHECK: call i32 @__tsan_atomic32_fetch_sub(i32* %a, i32 0, i32 100508)
+; CHECK: call i32 @__tsan_atomic32_fetch_sub(i32* %a, i32 0, i32 3)
 
 define void @atomic32_and_release(i32* %a) nounwind uwtable {
 entry:
@@ -880,7 +880,7 @@
   ret void
 }
 ; CHECK: atomic32_and_release
-; CHECK: call i32 @__tsan_atomic32_fetch_and(i32* %a, i32 0, i32 100508)
+; CHECK: call i32 @__tsan_atomic32_fetch_and(i32* %a, i32 0, i32 3)
 
 define void @atomic32_or_release(i32* %a) nounwind uwtable {
 entry:
@@ -888,7 +888,7 @@
   ret void
 }
 ; CHECK: atomic32_or_release
-; CHECK: call i32 @__tsan_atomic32_fetch_or(i32* %a, i32 0, i32 100508)
+; CHECK: call i32 @__tsan_atomic32_fetch_or(i32* %a, i32 0, i32 3)
 
 define void @atomic32_xor_release(i32* %a) nounwind uwtable {
 entry:
@@ -896,7 +896,7 @@
   ret void
 }
 ; CHECK: atomic32_xor_release
-; CHECK: call i32 @__tsan_atomic32_fetch_xor(i32* %a, i32 0, i32 100508)
+; CHECK: call i32 @__tsan_atomic32_fetch_xor(i32* %a, i32 0, i32 3)
 
 define void @atomic32_xchg_acq_rel(i32* %a) nounwind uwtable {
 entry:
@@ -904,7 +904,7 @@
   ret void
 }
 ; CHECK: atomic32_xchg_acq_rel
-; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 100516)
+; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 4)
 
 define void @atomic32_add_acq_rel(i32* %a) nounwind uwtable {
 entry:
@@ -912,7 +912,7 @@
   ret void
 }
 ; CHECK: atomic32_add_acq_rel
-; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 100516)
+; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 4)
 
 define void @atomic32_sub_acq_rel(i32* %a) nounwind uwtable {
 entry:
@@ -920,7 +920,7 @@
   ret void
 }
 ; CHECK: atomic32_sub_acq_rel
-; CHECK: call i32 @__tsan_atomic32_fetch_sub(i32* %a, i32 0, i32 100516)
+; CHECK: call i32 @__tsan_atomic32_fetch_sub(i32* %a, i32 0, i32 4)
 
 define void @atomic32_and_acq_rel(i32* %a) nounwind uwtable {
 entry:
@@ -928,7 +928,7 @@
   ret void
 }
 ; CHECK: atomic32_and_acq_rel
-; CHECK: call i32 @__tsan_atomic32_fetch_and(i32* %a, i32 0, i32 100516)
+; CHECK: call i32 @__tsan_atomic32_fetch_and(i32* %a, i32 0, i32 4)
 
 define void @atomic32_or_acq_rel(i32* %a) nounwind uwtable {
 entry:
@@ -936,7 +936,7 @@
   ret void
 }
 ; CHECK: atomic32_or_acq_rel
-; CHECK: call i32 @__tsan_atomic32_fetch_or(i32* %a, i32 0, i32 100516)
+; CHECK: call i32 @__tsan_atomic32_fetch_or(i32* %a, i32 0, i32 4)
 
 define void @atomic32_xor_acq_rel(i32* %a) nounwind uwtable {
 entry:
@@ -944,7 +944,7 @@
   ret void
 }
 ; CHECK: atomic32_xor_acq_rel
-; CHECK: call i32 @__tsan_atomic32_fetch_xor(i32* %a, i32 0, i32 100516)
+; CHECK: call i32 @__tsan_atomic32_fetch_xor(i32* %a, i32 0, i32 4)
 
 define void @atomic32_xchg_seq_cst(i32* %a) nounwind uwtable {
 entry:
@@ -952,7 +952,7 @@
   ret void
 }
 ; CHECK: atomic32_xchg_seq_cst
-; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 100532)
+; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 5)
 
 define void @atomic32_add_seq_cst(i32* %a) nounwind uwtable {
 entry:
@@ -960,7 +960,7 @@
   ret void
 }
 ; CHECK: atomic32_add_seq_cst
-; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 100532)
+; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 5)
 
 define void @atomic32_sub_seq_cst(i32* %a) nounwind uwtable {
 entry:
@@ -968,7 +968,7 @@
   ret void
 }
 ; CHECK: atomic32_sub_seq_cst
-; CHECK: call i32 @__tsan_atomic32_fetch_sub(i32* %a, i32 0, i32 100532)
+; CHECK: call i32 @__tsan_atomic32_fetch_sub(i32* %a, i32 0, i32 5)
 
 define void @atomic32_and_seq_cst(i32* %a) nounwind uwtable {
 entry:
@@ -976,7 +976,7 @@
   ret void
 }
 ; CHECK: atomic32_and_seq_cst
-; CHECK: call i32 @__tsan_atomic32_fetch_and(i32* %a, i32 0, i32 100532)
+; CHECK: call i32 @__tsan_atomic32_fetch_and(i32* %a, i32 0, i32 5)
 
 define void @atomic32_or_seq_cst(i32* %a) nounwind uwtable {
 entry:
@@ -984,7 +984,7 @@
   ret void
 }
 ; CHECK: atomic32_or_seq_cst
-; CHECK: call i32 @__tsan_atomic32_fetch_or(i32* %a, i32 0, i32 100532)
+; CHECK: call i32 @__tsan_atomic32_fetch_or(i32* %a, i32 0, i32 5)
 
 define void @atomic32_xor_seq_cst(i32* %a) nounwind uwtable {
 entry:
@@ -992,7 +992,7 @@
   ret void
 }
 ; CHECK: atomic32_xor_seq_cst
-; CHECK: call i32 @__tsan_atomic32_fetch_xor(i32* %a, i32 0, i32 100532)
+; CHECK: call i32 @__tsan_atomic32_fetch_xor(i32* %a, i32 0, i32 5)
 
 define void @atomic32_cas_monotonic(i32* %a) nounwind uwtable {
 entry:
@@ -1000,7 +1000,7 @@
   ret void
 }
 ; CHECK: atomic32_cas_monotonic
-; CHECK: call i32 @__tsan_atomic32_compare_exchange_val(i32* %a, i32 0, i32 1, i32 100501)
+; CHECK: call i32 @__tsan_atomic32_compare_exchange_val(i32* %a, i32 0, i32 1, i32 0)
 
 define void @atomic32_cas_acquire(i32* %a) nounwind uwtable {
 entry:
@@ -1008,7 +1008,7 @@
   ret void
 }
 ; CHECK: atomic32_cas_acquire
-; CHECK: call i32 @__tsan_atomic32_compare_exchange_val(i32* %a, i32 0, i32 1, i32 100504)
+; CHECK: call i32 @__tsan_atomic32_compare_exchange_val(i32* %a, i32 0, i32 1, i32 2)
 
 define void @atomic32_cas_release(i32* %a) nounwind uwtable {
 entry:
@@ -1016,7 +1016,7 @@
   ret void
 }
 ; CHECK: atomic32_cas_release
-; CHECK: call i32 @__tsan_atomic32_compare_exchange_val(i32* %a, i32 0, i32 1, i32 100508)
+; CHECK: call i32 @__tsan_atomic32_compare_exchange_val(i32* %a, i32 0, i32 1, i32 3)
 
 define void @atomic32_cas_acq_rel(i32* %a) nounwind uwtable {
 entry:
@@ -1024,7 +1024,7 @@
   ret void
 }
 ; CHECK: atomic32_cas_acq_rel
-; CHECK: call i32 @__tsan_atomic32_compare_exchange_val(i32* %a, i32 0, i32 1, i32 100516)
+; CHECK: call i32 @__tsan_atomic32_compare_exchange_val(i32* %a, i32 0, i32 1, i32 4)
 
 define void @atomic32_cas_seq_cst(i32* %a) nounwind uwtable {
 entry:
@@ -1032,7 +1032,7 @@
   ret void
 }
 ; CHECK: atomic32_cas_seq_cst
-; CHECK: call i32 @__tsan_atomic32_compare_exchange_val(i32* %a, i32 0, i32 1, i32 100532)
+; CHECK: call i32 @__tsan_atomic32_compare_exchange_val(i32* %a, i32 0, i32 1, i32 5)
 
 define i64 @atomic64_load_unordered(i64* %a) nounwind uwtable {
 entry:
@@ -1040,7 +1040,7 @@
   ret i64 %0
 }
 ; CHECK: atomic64_load_unordered
-; CHECK: call i64 @__tsan_atomic64_load(i64* %a, i32 100501)
+; CHECK: call i64 @__tsan_atomic64_load(i64* %a, i32 0)
 
 define i64 @atomic64_load_monotonic(i64* %a) nounwind uwtable {
 entry:
@@ -1048,7 +1048,7 @@
   ret i64 %0
 }
 ; CHECK: atomic64_load_monotonic
-; CHECK: call i64 @__tsan_atomic64_load(i64* %a, i32 100501)
+; CHECK: call i64 @__tsan_atomic64_load(i64* %a, i32 0)
 
 define i64 @atomic64_load_acquire(i64* %a) nounwind uwtable {
 entry:
@@ -1056,7 +1056,7 @@
   ret i64 %0
 }
 ; CHECK: atomic64_load_acquire
-; CHECK: call i64 @__tsan_atomic64_load(i64* %a, i32 100504)
+; CHECK: call i64 @__tsan_atomic64_load(i64* %a, i32 2)
 
 define i64 @atomic64_load_seq_cst(i64* %a) nounwind uwtable {
 entry:
@@ -1064,7 +1064,7 @@
   ret i64 %0
 }
 ; CHECK: atomic64_load_seq_cst
-; CHECK: call i64 @__tsan_atomic64_load(i64* %a, i32 100532)
+; CHECK: call i64 @__tsan_atomic64_load(i64* %a, i32 5)
 
 define void @atomic64_store_unordered(i64* %a) nounwind uwtable {
 entry:
@@ -1072,7 +1072,7 @@
   ret void
 }
 ; CHECK: atomic64_store_unordered
-; CHECK: call void @__tsan_atomic64_store(i64* %a, i64 0, i32 100501)
+; CHECK: call void @__tsan_atomic64_store(i64* %a, i64 0, i32 0)
 
 define void @atomic64_store_monotonic(i64* %a) nounwind uwtable {
 entry:
@@ -1080,7 +1080,7 @@
   ret void
 }
 ; CHECK: atomic64_store_monotonic
-; CHECK: call void @__tsan_atomic64_store(i64* %a, i64 0, i32 100501)
+; CHECK: call void @__tsan_atomic64_store(i64* %a, i64 0, i32 0)
 
 define void @atomic64_store_release(i64* %a) nounwind uwtable {
 entry:
@@ -1088,7 +1088,7 @@
   ret void
 }
 ; CHECK: atomic64_store_release
-; CHECK: call void @__tsan_atomic64_store(i64* %a, i64 0, i32 100508)
+; CHECK: call void @__tsan_atomic64_store(i64* %a, i64 0, i32 3)
 
 define void @atomic64_store_seq_cst(i64* %a) nounwind uwtable {
 entry:
@@ -1096,7 +1096,7 @@
   ret void
 }
 ; CHECK: atomic64_store_seq_cst
-; CHECK: call void @__tsan_atomic64_store(i64* %a, i64 0, i32 100532)
+; CHECK: call void @__tsan_atomic64_store(i64* %a, i64 0, i32 5)
 
 define void @atomic64_xchg_monotonic(i64* %a) nounwind uwtable {
 entry:
@@ -1104,7 +1104,7 @@
   ret void
 }
 ; CHECK: atomic64_xchg_monotonic
-; CHECK: call i64 @__tsan_atomic64_exchange(i64* %a, i64 0, i32 100501)
+; CHECK: call i64 @__tsan_atomic64_exchange(i64* %a, i64 0, i32 0)
 
 define void @atomic64_add_monotonic(i64* %a) nounwind uwtable {
 entry:
@@ -1112,7 +1112,7 @@
   ret void
 }
 ; CHECK: atomic64_add_monotonic
-; CHECK: call i64 @__tsan_atomic64_fetch_add(i64* %a, i64 0, i32 100501)
+; CHECK: call i64 @__tsan_atomic64_fetch_add(i64* %a, i64 0, i32 0)
 
 define void @atomic64_sub_monotonic(i64* %a) nounwind uwtable {
 entry:
@@ -1120,7 +1120,7 @@
   ret void
 }
 ; CHECK: atomic64_sub_monotonic
-; CHECK: call i64 @__tsan_atomic64_fetch_sub(i64* %a, i64 0, i32 100501)
+; CHECK: call i64 @__tsan_atomic64_fetch_sub(i64* %a, i64 0, i32 0)
 
 define void @atomic64_and_monotonic(i64* %a) nounwind uwtable {
 entry:
@@ -1128,7 +1128,7 @@
   ret void
 }
 ; CHECK: atomic64_and_monotonic
-; CHECK: call i64 @__tsan_atomic64_fetch_and(i64* %a, i64 0, i32 100501)
+; CHECK: call i64 @__tsan_atomic64_fetch_and(i64* %a, i64 0, i32 0)
 
 define void @atomic64_or_monotonic(i64* %a) nounwind uwtable {
 entry:
@@ -1136,7 +1136,7 @@
   ret void
 }
 ; CHECK: atomic64_or_monotonic
-; CHECK: call i64 @__tsan_atomic64_fetch_or(i64* %a, i64 0, i32 100501)
+; CHECK: call i64 @__tsan_atomic64_fetch_or(i64* %a, i64 0, i32 0)
 
 define void @atomic64_xor_monotonic(i64* %a) nounwind uwtable {
 entry:
@@ -1144,7 +1144,7 @@
   ret void
 }
 ; CHECK: atomic64_xor_monotonic
-; CHECK: call i64 @__tsan_atomic64_fetch_xor(i64* %a, i64 0, i32 100501)
+; CHECK: call i64 @__tsan_atomic64_fetch_xor(i64* %a, i64 0, i32 0)
 
 define void @atomic64_xchg_acquire(i64* %a) nounwind uwtable {
 entry:
@@ -1152,7 +1152,7 @@
   ret void
 }
 ; CHECK: atomic64_xchg_acquire
-; CHECK: call i64 @__tsan_atomic64_exchange(i64* %a, i64 0, i32 100504)
+; CHECK: call i64 @__tsan_atomic64_exchange(i64* %a, i64 0, i32 2)
 
 define void @atomic64_add_acquire(i64* %a) nounwind uwtable {
 entry:
@@ -1160,7 +1160,7 @@
   ret void
 }
 ; CHECK: atomic64_add_acquire
-; CHECK: call i64 @__tsan_atomic64_fetch_add(i64* %a, i64 0, i32 100504)
+; CHECK: call i64 @__tsan_atomic64_fetch_add(i64* %a, i64 0, i32 2)
 
 define void @atomic64_sub_acquire(i64* %a) nounwind uwtable {
 entry:
@@ -1168,7 +1168,7 @@
   ret void
 }
 ; CHECK: atomic64_sub_acquire
-; CHECK: call i64 @__tsan_atomic64_fetch_sub(i64* %a, i64 0, i32 100504)
+; CHECK: call i64 @__tsan_atomic64_fetch_sub(i64* %a, i64 0, i32 2)
 
 define void @atomic64_and_acquire(i64* %a) nounwind uwtable {
 entry:
@@ -1176,7 +1176,7 @@
   ret void
 }
 ; CHECK: atomic64_and_acquire
-; CHECK: call i64 @__tsan_atomic64_fetch_and(i64* %a, i64 0, i32 100504)
+; CHECK: call i64 @__tsan_atomic64_fetch_and(i64* %a, i64 0, i32 2)
 
 define void @atomic64_or_acquire(i64* %a) nounwind uwtable {
 entry:
@@ -1184,7 +1184,7 @@
   ret void
 }
 ; CHECK: atomic64_or_acquire
-; CHECK: call i64 @__tsan_atomic64_fetch_or(i64* %a, i64 0, i32 100504)
+; CHECK: call i64 @__tsan_atomic64_fetch_or(i64* %a, i64 0, i32 2)
 
 define void @atomic64_xor_acquire(i64* %a) nounwind uwtable {
 entry:
@@ -1192,7 +1192,7 @@
   ret void
 }
 ; CHECK: atomic64_xor_acquire
-; CHECK: call i64 @__tsan_atomic64_fetch_xor(i64* %a, i64 0, i32 100504)
+; CHECK: call i64 @__tsan_atomic64_fetch_xor(i64* %a, i64 0, i32 2)
 
 define void @atomic64_xchg_release(i64* %a) nounwind uwtable {
 entry:
@@ -1200,7 +1200,7 @@
   ret void
 }
 ; CHECK: atomic64_xchg_release
-; CHECK: call i64 @__tsan_atomic64_exchange(i64* %a, i64 0, i32 100508)
+; CHECK: call i64 @__tsan_atomic64_exchange(i64* %a, i64 0, i32 3)
 
 define void @atomic64_add_release(i64* %a) nounwind uwtable {
 entry:
@@ -1208,7 +1208,7 @@
   ret void
 }
 ; CHECK: atomic64_add_release
-; CHECK: call i64 @__tsan_atomic64_fetch_add(i64* %a, i64 0, i32 100508)
+; CHECK: call i64 @__tsan_atomic64_fetch_add(i64* %a, i64 0, i32 3)
 
 define void @atomic64_sub_release(i64* %a) nounwind uwtable {
 entry:
@@ -1216,7 +1216,7 @@
   ret void
 }
 ; CHECK: atomic64_sub_release
-; CHECK: call i64 @__tsan_atomic64_fetch_sub(i64* %a, i64 0, i32 100508)
+; CHECK: call i64 @__tsan_atomic64_fetch_sub(i64* %a, i64 0, i32 3)
 
 define void @atomic64_and_release(i64* %a) nounwind uwtable {
 entry:
@@ -1224,7 +1224,7 @@
   ret void
 }
 ; CHECK: atomic64_and_release
-; CHECK: call i64 @__tsan_atomic64_fetch_and(i64* %a, i64 0, i32 100508)
+; CHECK: call i64 @__tsan_atomic64_fetch_and(i64* %a, i64 0, i32 3)
 
 define void @atomic64_or_release(i64* %a) nounwind uwtable {
 entry:
@@ -1232,7 +1232,7 @@
   ret void
 }
 ; CHECK: atomic64_or_release
-; CHECK: call i64 @__tsan_atomic64_fetch_or(i64* %a, i64 0, i32 100508)
+; CHECK: call i64 @__tsan_atomic64_fetch_or(i64* %a, i64 0, i32 3)
 
 define void @atomic64_xor_release(i64* %a) nounwind uwtable {
 entry:
@@ -1240,7 +1240,7 @@
   ret void
 }
 ; CHECK: atomic64_xor_release
-; CHECK: call i64 @__tsan_atomic64_fetch_xor(i64* %a, i64 0, i32 100508)
+; CHECK: call i64 @__tsan_atomic64_fetch_xor(i64* %a, i64 0, i32 3)
 
 define void @atomic64_xchg_acq_rel(i64* %a) nounwind uwtable {
 entry:
@@ -1248,7 +1248,7 @@
   ret void
 }
 ; CHECK: atomic64_xchg_acq_rel
-; CHECK: call i64 @__tsan_atomic64_exchange(i64* %a, i64 0, i32 100516)
+; CHECK: call i64 @__tsan_atomic64_exchange(i64* %a, i64 0, i32 4)
 
 define void @atomic64_add_acq_rel(i64* %a) nounwind uwtable {
 entry:
@@ -1256,7 +1256,7 @@
   ret void
 }
 ; CHECK: atomic64_add_acq_rel
-; CHECK: call i64 @__tsan_atomic64_fetch_add(i64* %a, i64 0, i32 100516)
+; CHECK: call i64 @__tsan_atomic64_fetch_add(i64* %a, i64 0, i32 4)
 
 define void @atomic64_sub_acq_rel(i64* %a) nounwind uwtable {
 entry:
@@ -1264,7 +1264,7 @@
   ret void
 }
 ; CHECK: atomic64_sub_acq_rel
-; CHECK: call i64 @__tsan_atomic64_fetch_sub(i64* %a, i64 0, i32 100516)
+; CHECK: call i64 @__tsan_atomic64_fetch_sub(i64* %a, i64 0, i32 4)
 
 define void @atomic64_and_acq_rel(i64* %a) nounwind uwtable {
 entry:
@@ -1272,7 +1272,7 @@
   ret void
 }
 ; CHECK: atomic64_and_acq_rel
-; CHECK: call i64 @__tsan_atomic64_fetch_and(i64* %a, i64 0, i32 100516)
+; CHECK: call i64 @__tsan_atomic64_fetch_and(i64* %a, i64 0, i32 4)
 
 define void @atomic64_or_acq_rel(i64* %a) nounwind uwtable {
 entry:
@@ -1280,7 +1280,7 @@
   ret void
 }
 ; CHECK: atomic64_or_acq_rel
-; CHECK: call i64 @__tsan_atomic64_fetch_or(i64* %a, i64 0, i32 100516)
+; CHECK: call i64 @__tsan_atomic64_fetch_or(i64* %a, i64 0, i32 4)
 
 define void @atomic64_xor_acq_rel(i64* %a) nounwind uwtable {
 entry:
@@ -1288,7 +1288,7 @@
   ret void
 }
 ; CHECK: atomic64_xor_acq_rel
-; CHECK: call i64 @__tsan_atomic64_fetch_xor(i64* %a, i64 0, i32 100516)
+; CHECK: call i64 @__tsan_atomic64_fetch_xor(i64* %a, i64 0, i32 4)
 
 define void @atomic64_xchg_seq_cst(i64* %a) nounwind uwtable {
 entry:
@@ -1296,7 +1296,7 @@
   ret void
 }
 ; CHECK: atomic64_xchg_seq_cst
-; CHECK: call i64 @__tsan_atomic64_exchange(i64* %a, i64 0, i32 100532)
+; CHECK: call i64 @__tsan_atomic64_exchange(i64* %a, i64 0, i32 5)
 
 define void @atomic64_add_seq_cst(i64* %a) nounwind uwtable {
 entry:
@@ -1304,7 +1304,7 @@
   ret void
 }
 ; CHECK: atomic64_add_seq_cst
-; CHECK: call i64 @__tsan_atomic64_fetch_add(i64* %a, i64 0, i32 100532)
+; CHECK: call i64 @__tsan_atomic64_fetch_add(i64* %a, i64 0, i32 5)
 
 define void @atomic64_sub_seq_cst(i64* %a) nounwind uwtable {
 entry:
@@ -1312,7 +1312,7 @@
   ret void
 }
 ; CHECK: atomic64_sub_seq_cst
-; CHECK: call i64 @__tsan_atomic64_fetch_sub(i64* %a, i64 0, i32 100532)
+; CHECK: call i64 @__tsan_atomic64_fetch_sub(i64* %a, i64 0, i32 5)
 
 define void @atomic64_and_seq_cst(i64* %a) nounwind uwtable {
 entry:
@@ -1320,7 +1320,7 @@
   ret void
 }
 ; CHECK: atomic64_and_seq_cst
-; CHECK: call i64 @__tsan_atomic64_fetch_and(i64* %a, i64 0, i32 100532)
+; CHECK: call i64 @__tsan_atomic64_fetch_and(i64* %a, i64 0, i32 5)
 
 define void @atomic64_or_seq_cst(i64* %a) nounwind uwtable {
 entry:
@@ -1328,7 +1328,7 @@
   ret void
 }
 ; CHECK: atomic64_or_seq_cst
-; CHECK: call i64 @__tsan_atomic64_fetch_or(i64* %a, i64 0, i32 100532)
+; CHECK: call i64 @__tsan_atomic64_fetch_or(i64* %a, i64 0, i32 5)
 
 define void @atomic64_xor_seq_cst(i64* %a) nounwind uwtable {
 entry:
@@ -1336,7 +1336,7 @@
   ret void
 }
 ; CHECK: atomic64_xor_seq_cst
-; CHECK: call i64 @__tsan_atomic64_fetch_xor(i64* %a, i64 0, i32 100532)
+; CHECK: call i64 @__tsan_atomic64_fetch_xor(i64* %a, i64 0, i32 5)
 
 define void @atomic64_cas_monotonic(i64* %a) nounwind uwtable {
 entry:
@@ -1344,7 +1344,7 @@
   ret void
 }
 ; CHECK: atomic64_cas_monotonic
-; CHECK: call i64 @__tsan_atomic64_compare_exchange_val(i64* %a, i64 0, i64 1, i32 100501)
+; CHECK: call i64 @__tsan_atomic64_compare_exchange_val(i64* %a, i64 0, i64 1, i32 0)
 
 define void @atomic64_cas_acquire(i64* %a) nounwind uwtable {
 entry:
@@ -1352,7 +1352,7 @@
   ret void
 }
 ; CHECK: atomic64_cas_acquire
-; CHECK: call i64 @__tsan_atomic64_compare_exchange_val(i64* %a, i64 0, i64 1, i32 100504)
+; CHECK: call i64 @__tsan_atomic64_compare_exchange_val(i64* %a, i64 0, i64 1, i32 2)
 
 define void @atomic64_cas_release(i64* %a) nounwind uwtable {
 entry:
@@ -1360,7 +1360,7 @@
   ret void
 }
 ; CHECK: atomic64_cas_release
-; CHECK: call i64 @__tsan_atomic64_compare_exchange_val(i64* %a, i64 0, i64 1, i32 100508)
+; CHECK: call i64 @__tsan_atomic64_compare_exchange_val(i64* %a, i64 0, i64 1, i32 3)
 
 define void @atomic64_cas_acq_rel(i64* %a) nounwind uwtable {
 entry:
@@ -1368,7 +1368,7 @@
   ret void
 }
 ; CHECK: atomic64_cas_acq_rel
-; CHECK: call i64 @__tsan_atomic64_compare_exchange_val(i64* %a, i64 0, i64 1, i32 100516)
+; CHECK: call i64 @__tsan_atomic64_compare_exchange_val(i64* %a, i64 0, i64 1, i32 4)
 
 define void @atomic64_cas_seq_cst(i64* %a) nounwind uwtable {
 entry:
@@ -1376,7 +1376,7 @@
   ret void
 }
 ; CHECK: atomic64_cas_seq_cst
-; CHECK: call i64 @__tsan_atomic64_compare_exchange_val(i64* %a, i64 0, i64 1, i32 100532)
+; CHECK: call i64 @__tsan_atomic64_compare_exchange_val(i64* %a, i64 0, i64 1, i32 5)
 
 define i128 @atomic128_load_unordered(i128* %a) nounwind uwtable {
 entry:
@@ -1384,7 +1384,7 @@
   ret i128 %0
 }
 ; CHECK: atomic128_load_unordered
-; CHECK: call i128 @__tsan_atomic128_load(i128* %a, i32 100501)
+; CHECK: call i128 @__tsan_atomic128_load(i128* %a, i32 0)
 
 define i128 @atomic128_load_monotonic(i128* %a) nounwind uwtable {
 entry:
@@ -1392,7 +1392,7 @@
   ret i128 %0
 }
 ; CHECK: atomic128_load_monotonic
-; CHECK: call i128 @__tsan_atomic128_load(i128* %a, i32 100501)
+; CHECK: call i128 @__tsan_atomic128_load(i128* %a, i32 0)
 
 define i128 @atomic128_load_acquire(i128* %a) nounwind uwtable {
 entry:
@@ -1400,7 +1400,7 @@
   ret i128 %0
 }
 ; CHECK: atomic128_load_acquire
-; CHECK: call i128 @__tsan_atomic128_load(i128* %a, i32 100504)
+; CHECK: call i128 @__tsan_atomic128_load(i128* %a, i32 2)
 
 define i128 @atomic128_load_seq_cst(i128* %a) nounwind uwtable {
 entry:
@@ -1408,7 +1408,7 @@
   ret i128 %0
 }
 ; CHECK: atomic128_load_seq_cst
-; CHECK: call i128 @__tsan_atomic128_load(i128* %a, i32 100532)
+; CHECK: call i128 @__tsan_atomic128_load(i128* %a, i32 5)
 
 define void @atomic128_store_unordered(i128* %a) nounwind uwtable {
 entry:
@@ -1416,7 +1416,7 @@
   ret void
 }
 ; CHECK: atomic128_store_unordered
-; CHECK: call void @__tsan_atomic128_store(i128* %a, i128 0, i32 100501)
+; CHECK: call void @__tsan_atomic128_store(i128* %a, i128 0, i32 0)
 
 define void @atomic128_store_monotonic(i128* %a) nounwind uwtable {
 entry:
@@ -1424,7 +1424,7 @@
   ret void
 }
 ; CHECK: atomic128_store_monotonic
-; CHECK: call void @__tsan_atomic128_store(i128* %a, i128 0, i32 100501)
+; CHECK: call void @__tsan_atomic128_store(i128* %a, i128 0, i32 0)
 
 define void @atomic128_store_release(i128* %a) nounwind uwtable {
 entry:
@@ -1432,7 +1432,7 @@
   ret void
 }
 ; CHECK: atomic128_store_release
-; CHECK: call void @__tsan_atomic128_store(i128* %a, i128 0, i32 100508)
+; CHECK: call void @__tsan_atomic128_store(i128* %a, i128 0, i32 3)
 
 define void @atomic128_store_seq_cst(i128* %a) nounwind uwtable {
 entry:
@@ -1440,7 +1440,7 @@
   ret void
 }
 ; CHECK: atomic128_store_seq_cst
-; CHECK: call void @__tsan_atomic128_store(i128* %a, i128 0, i32 100532)
+; CHECK: call void @__tsan_atomic128_store(i128* %a, i128 0, i32 5)
 
 define void @atomic128_xchg_monotonic(i128* %a) nounwind uwtable {
 entry:
@@ -1448,7 +1448,7 @@
   ret void
 }
 ; CHECK: atomic128_xchg_monotonic
-; CHECK: call i128 @__tsan_atomic128_exchange(i128* %a, i128 0, i32 100501)
+; CHECK: call i128 @__tsan_atomic128_exchange(i128* %a, i128 0, i32 0)
 
 define void @atomic128_add_monotonic(i128* %a) nounwind uwtable {
 entry:
@@ -1456,7 +1456,7 @@
   ret void
 }
 ; CHECK: atomic128_add_monotonic
-; CHECK: call i128 @__tsan_atomic128_fetch_add(i128* %a, i128 0, i32 100501)
+; CHECK: call i128 @__tsan_atomic128_fetch_add(i128* %a, i128 0, i32 0)
 
 define void @atomic128_sub_monotonic(i128* %a) nounwind uwtable {
 entry:
@@ -1464,7 +1464,7 @@
   ret void
 }
 ; CHECK: atomic128_sub_monotonic
-; CHECK: call i128 @__tsan_atomic128_fetch_sub(i128* %a, i128 0, i32 100501)
+; CHECK: call i128 @__tsan_atomic128_fetch_sub(i128* %a, i128 0, i32 0)
 
 define void @atomic128_and_monotonic(i128* %a) nounwind uwtable {
 entry:
@@ -1472,7 +1472,7 @@
   ret void
 }
 ; CHECK: atomic128_and_monotonic
-; CHECK: call i128 @__tsan_atomic128_fetch_and(i128* %a, i128 0, i32 100501)
+; CHECK: call i128 @__tsan_atomic128_fetch_and(i128* %a, i128 0, i32 0)
 
 define void @atomic128_or_monotonic(i128* %a) nounwind uwtable {
 entry:
@@ -1480,7 +1480,7 @@
   ret void
 }
 ; CHECK: atomic128_or_monotonic
-; CHECK: call i128 @__tsan_atomic128_fetch_or(i128* %a, i128 0, i32 100501)
+; CHECK: call i128 @__tsan_atomic128_fetch_or(i128* %a, i128 0, i32 0)
 
 define void @atomic128_xor_monotonic(i128* %a) nounwind uwtable {
 entry:
@@ -1488,7 +1488,7 @@
   ret void
 }
 ; CHECK: atomic128_xor_monotonic
-; CHECK: call i128 @__tsan_atomic128_fetch_xor(i128* %a, i128 0, i32 100501)
+; CHECK: call i128 @__tsan_atomic128_fetch_xor(i128* %a, i128 0, i32 0)
 
 define void @atomic128_xchg_acquire(i128* %a) nounwind uwtable {
 entry:
@@ -1496,7 +1496,7 @@
   ret void
 }
 ; CHECK: atomic128_xchg_acquire
-; CHECK: call i128 @__tsan_atomic128_exchange(i128* %a, i128 0, i32 100504)
+; CHECK: call i128 @__tsan_atomic128_exchange(i128* %a, i128 0, i32 2)
 
 define void @atomic128_add_acquire(i128* %a) nounwind uwtable {
 entry:
@@ -1504,7 +1504,7 @@
   ret void
 }
 ; CHECK: atomic128_add_acquire
-; CHECK: call i128 @__tsan_atomic128_fetch_add(i128* %a, i128 0, i32 100504)
+; CHECK: call i128 @__tsan_atomic128_fetch_add(i128* %a, i128 0, i32 2)
 
 define void @atomic128_sub_acquire(i128* %a) nounwind uwtable {
 entry:
@@ -1512,7 +1512,7 @@
   ret void
 }
 ; CHECK: atomic128_sub_acquire
-; CHECK: call i128 @__tsan_atomic128_fetch_sub(i128* %a, i128 0, i32 100504)
+; CHECK: call i128 @__tsan_atomic128_fetch_sub(i128* %a, i128 0, i32 2)
 
 define void @atomic128_and_acquire(i128* %a) nounwind uwtable {
 entry:
@@ -1520,7 +1520,7 @@
   ret void
 }
 ; CHECK: atomic128_and_acquire
-; CHECK: call i128 @__tsan_atomic128_fetch_and(i128* %a, i128 0, i32 100504)
+; CHECK: call i128 @__tsan_atomic128_fetch_and(i128* %a, i128 0, i32 2)
 
 define void @atomic128_or_acquire(i128* %a) nounwind uwtable {
 entry:
@@ -1528,7 +1528,7 @@
   ret void
 }
 ; CHECK: atomic128_or_acquire
-; CHECK: call i128 @__tsan_atomic128_fetch_or(i128* %a, i128 0, i32 100504)
+; CHECK: call i128 @__tsan_atomic128_fetch_or(i128* %a, i128 0, i32 2)
 
 define void @atomic128_xor_acquire(i128* %a) nounwind uwtable {
 entry:
@@ -1536,7 +1536,7 @@
   ret void
 }
 ; CHECK: atomic128_xor_acquire
-; CHECK: call i128 @__tsan_atomic128_fetch_xor(i128* %a, i128 0, i32 100504)
+; CHECK: call i128 @__tsan_atomic128_fetch_xor(i128* %a, i128 0, i32 2)
 
 define void @atomic128_xchg_release(i128* %a) nounwind uwtable {
 entry:
@@ -1544,7 +1544,7 @@
   ret void
 }
 ; CHECK: atomic128_xchg_release
-; CHECK: call i128 @__tsan_atomic128_exchange(i128* %a, i128 0, i32 100508)
+; CHECK: call i128 @__tsan_atomic128_exchange(i128* %a, i128 0, i32 3)
 
 define void @atomic128_add_release(i128* %a) nounwind uwtable {
 entry:
@@ -1552,7 +1552,7 @@
   ret void
 }
 ; CHECK: atomic128_add_release
-; CHECK: call i128 @__tsan_atomic128_fetch_add(i128* %a, i128 0, i32 100508)
+; CHECK: call i128 @__tsan_atomic128_fetch_add(i128* %a, i128 0, i32 3)
 
 define void @atomic128_sub_release(i128* %a) nounwind uwtable {
 entry:
@@ -1560,7 +1560,7 @@
   ret void
 }
 ; CHECK: atomic128_sub_release
-; CHECK: call i128 @__tsan_atomic128_fetch_sub(i128* %a, i128 0, i32 100508)
+; CHECK: call i128 @__tsan_atomic128_fetch_sub(i128* %a, i128 0, i32 3)
 
 define void @atomic128_and_release(i128* %a) nounwind uwtable {
 entry:
@@ -1568,7 +1568,7 @@
   ret void
 }
 ; CHECK: atomic128_and_release
-; CHECK: call i128 @__tsan_atomic128_fetch_and(i128* %a, i128 0, i32 100508)
+; CHECK: call i128 @__tsan_atomic128_fetch_and(i128* %a, i128 0, i32 3)
 
 define void @atomic128_or_release(i128* %a) nounwind uwtable {
 entry:
@@ -1576,7 +1576,7 @@
   ret void
 }
 ; CHECK: atomic128_or_release
-; CHECK: call i128 @__tsan_atomic128_fetch_or(i128* %a, i128 0, i32 100508)
+; CHECK: call i128 @__tsan_atomic128_fetch_or(i128* %a, i128 0, i32 3)
 
 define void @atomic128_xor_release(i128* %a) nounwind uwtable {
 entry:
@@ -1584,7 +1584,7 @@
   ret void
 }
 ; CHECK: atomic128_xor_release
-; CHECK: call i128 @__tsan_atomic128_fetch_xor(i128* %a, i128 0, i32 100508)
+; CHECK: call i128 @__tsan_atomic128_fetch_xor(i128* %a, i128 0, i32 3)
 
 define void @atomic128_xchg_acq_rel(i128* %a) nounwind uwtable {
 entry:
@@ -1592,7 +1592,7 @@
   ret void
 }
 ; CHECK: atomic128_xchg_acq_rel
-; CHECK: call i128 @__tsan_atomic128_exchange(i128* %a, i128 0, i32 100516)
+; CHECK: call i128 @__tsan_atomic128_exchange(i128* %a, i128 0, i32 4)
 
 define void @atomic128_add_acq_rel(i128* %a) nounwind uwtable {
 entry:
@@ -1600,7 +1600,7 @@
   ret void
 }
 ; CHECK: atomic128_add_acq_rel
-; CHECK: call i128 @__tsan_atomic128_fetch_add(i128* %a, i128 0, i32 100516)
+; CHECK: call i128 @__tsan_atomic128_fetch_add(i128* %a, i128 0, i32 4)
 
 define void @atomic128_sub_acq_rel(i128* %a) nounwind uwtable {
 entry:
@@ -1608,7 +1608,7 @@
   ret void
 }
 ; CHECK: atomic128_sub_acq_rel
-; CHECK: call i128 @__tsan_atomic128_fetch_sub(i128* %a, i128 0, i32 100516)
+; CHECK: call i128 @__tsan_atomic128_fetch_sub(i128* %a, i128 0, i32 4)
 
 define void @atomic128_and_acq_rel(i128* %a) nounwind uwtable {
 entry:
@@ -1616,7 +1616,7 @@
   ret void
 }
 ; CHECK: atomic128_and_acq_rel
-; CHECK: call i128 @__tsan_atomic128_fetch_and(i128* %a, i128 0, i32 100516)
+; CHECK: call i128 @__tsan_atomic128_fetch_and(i128* %a, i128 0, i32 4)
 
 define void @atomic128_or_acq_rel(i128* %a) nounwind uwtable {
 entry:
@@ -1624,7 +1624,7 @@
   ret void
 }
 ; CHECK: atomic128_or_acq_rel
-; CHECK: call i128 @__tsan_atomic128_fetch_or(i128* %a, i128 0, i32 100516)
+; CHECK: call i128 @__tsan_atomic128_fetch_or(i128* %a, i128 0, i32 4)
 
 define void @atomic128_xor_acq_rel(i128* %a) nounwind uwtable {
 entry:
@@ -1632,7 +1632,7 @@
   ret void
 }
 ; CHECK: atomic128_xor_acq_rel
-; CHECK: call i128 @__tsan_atomic128_fetch_xor(i128* %a, i128 0, i32 100516)
+; CHECK: call i128 @__tsan_atomic128_fetch_xor(i128* %a, i128 0, i32 4)
 
 define void @atomic128_xchg_seq_cst(i128* %a) nounwind uwtable {
 entry:
@@ -1640,7 +1640,7 @@
   ret void
 }
 ; CHECK: atomic128_xchg_seq_cst
-; CHECK: call i128 @__tsan_atomic128_exchange(i128* %a, i128 0, i32 100532)
+; CHECK: call i128 @__tsan_atomic128_exchange(i128* %a, i128 0, i32 5)
 
 define void @atomic128_add_seq_cst(i128* %a) nounwind uwtable {
 entry:
@@ -1648,7 +1648,7 @@
   ret void
 }
 ; CHECK: atomic128_add_seq_cst
-; CHECK: call i128 @__tsan_atomic128_fetch_add(i128* %a, i128 0, i32 100532)
+; CHECK: call i128 @__tsan_atomic128_fetch_add(i128* %a, i128 0, i32 5)
 
 define void @atomic128_sub_seq_cst(i128* %a) nounwind uwtable {
 entry:
@@ -1656,7 +1656,7 @@
   ret void
 }
 ; CHECK: atomic128_sub_seq_cst
-; CHECK: call i128 @__tsan_atomic128_fetch_sub(i128* %a, i128 0, i32 100532)
+; CHECK: call i128 @__tsan_atomic128_fetch_sub(i128* %a, i128 0, i32 5)
 
 define void @atomic128_and_seq_cst(i128* %a) nounwind uwtable {
 entry:
@@ -1664,7 +1664,7 @@
   ret void
 }
 ; CHECK: atomic128_and_seq_cst
-; CHECK: call i128 @__tsan_atomic128_fetch_and(i128* %a, i128 0, i32 100532)
+; CHECK: call i128 @__tsan_atomic128_fetch_and(i128* %a, i128 0, i32 5)
 
 define void @atomic128_or_seq_cst(i128* %a) nounwind uwtable {
 entry:
@@ -1672,7 +1672,7 @@
   ret void
 }
 ; CHECK: atomic128_or_seq_cst
-; CHECK: call i128 @__tsan_atomic128_fetch_or(i128* %a, i128 0, i32 100532)
+; CHECK: call i128 @__tsan_atomic128_fetch_or(i128* %a, i128 0, i32 5)
 
 define void @atomic128_xor_seq_cst(i128* %a) nounwind uwtable {
 entry:
@@ -1680,7 +1680,7 @@
   ret void
 }
 ; CHECK: atomic128_xor_seq_cst
-; CHECK: call i128 @__tsan_atomic128_fetch_xor(i128* %a, i128 0, i32 100532)
+; CHECK: call i128 @__tsan_atomic128_fetch_xor(i128* %a, i128 0, i32 5)
 
 define void @atomic128_cas_monotonic(i128* %a) nounwind uwtable {
 entry:
@@ -1688,7 +1688,7 @@
   ret void
 }
 ; CHECK: atomic128_cas_monotonic
-; CHECK: call i128 @__tsan_atomic128_compare_exchange_val(i128* %a, i128 0, i128 1, i32 100501)
+; CHECK: call i128 @__tsan_atomic128_compare_exchange_val(i128* %a, i128 0, i128 1, i32 0)
 
 define void @atomic128_cas_acquire(i128* %a) nounwind uwtable {
 entry:
@@ -1696,7 +1696,7 @@
   ret void
 }
 ; CHECK: atomic128_cas_acquire
-; CHECK: call i128 @__tsan_atomic128_compare_exchange_val(i128* %a, i128 0, i128 1, i32 100504)
+; CHECK: call i128 @__tsan_atomic128_compare_exchange_val(i128* %a, i128 0, i128 1, i32 2)
 
 define void @atomic128_cas_release(i128* %a) nounwind uwtable {
 entry:
@@ -1704,7 +1704,7 @@
   ret void
 }
 ; CHECK: atomic128_cas_release
-; CHECK: call i128 @__tsan_atomic128_compare_exchange_val(i128* %a, i128 0, i128 1, i32 100508)
+; CHECK: call i128 @__tsan_atomic128_compare_exchange_val(i128* %a, i128 0, i128 1, i32 3)
 
 define void @atomic128_cas_acq_rel(i128* %a) nounwind uwtable {
 entry:
@@ -1712,7 +1712,7 @@
   ret void
 }
 ; CHECK: atomic128_cas_acq_rel
-; CHECK: call i128 @__tsan_atomic128_compare_exchange_val(i128* %a, i128 0, i128 1, i32 100516)
+; CHECK: call i128 @__tsan_atomic128_compare_exchange_val(i128* %a, i128 0, i128 1, i32 4)
 
 define void @atomic128_cas_seq_cst(i128* %a) nounwind uwtable {
 entry:
@@ -1720,7 +1720,7 @@
   ret void
 }
 ; CHECK: atomic128_cas_seq_cst
-; CHECK: call i128 @__tsan_atomic128_compare_exchange_val(i128* %a, i128 0, i128 1, i32 100532)
+; CHECK: call i128 @__tsan_atomic128_compare_exchange_val(i128* %a, i128 0, i128 1, i32 5)
 
 define void @atomic_signal_fence_acquire() nounwind uwtable {
 entry:
@@ -1728,7 +1728,7 @@
   ret void
 }
 ; CHECK: atomic_signal_fence_acquire
-; CHECK: call void @__tsan_atomic_signal_fence(i32 100504)
+; CHECK: call void @__tsan_atomic_signal_fence(i32 2)
 
 define void @atomic_thread_fence_acquire() nounwind uwtable {
 entry:
@@ -1736,7 +1736,7 @@
   ret void
 }
 ; CHECK: atomic_thread_fence_acquire
-; CHECK: call void @__tsan_atomic_thread_fence(i32 100504)
+; CHECK: call void @__tsan_atomic_thread_fence(i32 2)
 
 define void @atomic_signal_fence_release() nounwind uwtable {
 entry:
@@ -1744,7 +1744,7 @@
   ret void
 }
 ; CHECK: atomic_signal_fence_release
-; CHECK: call void @__tsan_atomic_signal_fence(i32 100508)
+; CHECK: call void @__tsan_atomic_signal_fence(i32 3)
 
 define void @atomic_thread_fence_release() nounwind uwtable {
 entry:
@@ -1752,7 +1752,7 @@
   ret void
 }
 ; CHECK: atomic_thread_fence_release
-; CHECK: call void @__tsan_atomic_thread_fence(i32 100508)
+; CHECK: call void @__tsan_atomic_thread_fence(i32 3)
 
 define void @atomic_signal_fence_acq_rel() nounwind uwtable {
 entry:
@@ -1760,7 +1760,7 @@
   ret void
 }
 ; CHECK: atomic_signal_fence_acq_rel
-; CHECK: call void @__tsan_atomic_signal_fence(i32 100516)
+; CHECK: call void @__tsan_atomic_signal_fence(i32 4)
 
 define void @atomic_thread_fence_acq_rel() nounwind uwtable {
 entry:
@@ -1768,7 +1768,7 @@
   ret void
 }
 ; CHECK: atomic_thread_fence_acq_rel
-; CHECK: call void @__tsan_atomic_thread_fence(i32 100516)
+; CHECK: call void @__tsan_atomic_thread_fence(i32 4)
 
 define void @atomic_signal_fence_seq_cst() nounwind uwtable {
 entry:
@@ -1776,7 +1776,7 @@
   ret void
 }
 ; CHECK: atomic_signal_fence_seq_cst
-; CHECK: call void @__tsan_atomic_signal_fence(i32 100532)
+; CHECK: call void @__tsan_atomic_signal_fence(i32 5)
 
 define void @atomic_thread_fence_seq_cst() nounwind uwtable {
 entry:
@@ -1784,4 +1784,4 @@
   ret void
 }
 ; CHECK: atomic_thread_fence_seq_cst
-; CHECK: call void @__tsan_atomic_thread_fence(i32 100532)
+; CHECK: call void @__tsan_atomic_thread_fence(i32 5)





More information about the llvm-commits mailing list