diff options
author | Dmitry Vyukov <dvyukov@google.com> | 2012-11-09 14:12:16 +0000 |
---|---|---|
committer | Dmitry Vyukov <dvyukov@google.com> | 2012-11-09 14:12:16 +0000 |
commit | c2e9ca15fbb7ce2bc0bc79e44862d74cdd2802b9 (patch) | |
tree | d2db1d35c7e1ac1384544073592dfc2b05eb6d95 | |
parent | 9f8a90b3ce0e248e3b68b056d4c840295facbc02 (diff) |
tsan: switch to new memory_order constants (ABI compatible)
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@167615 91177308-0d34-0410-b5e6-96231b3b80d8
-rw-r--r-- | lib/Transforms/Instrumentation/ThreadSanitizer.cpp | 15 | ||||
-rw-r--r-- | test/Instrumentation/ThreadSanitizer/atomic.ll | 446 |
2 files changed, 230 insertions, 231 deletions
diff --git a/lib/Transforms/Instrumentation/ThreadSanitizer.cpp b/lib/Transforms/Instrumentation/ThreadSanitizer.cpp index 27514a5b72..9e10fc4416 100644 --- a/lib/Transforms/Instrumentation/ThreadSanitizer.cpp +++ b/lib/Transforms/Instrumentation/ThreadSanitizer.cpp @@ -390,15 +390,14 @@ static ConstantInt *createOrdering(IRBuilder<> *IRB, AtomicOrdering ord) { switch (ord) { case NotAtomic: assert(false); case Unordered: // Fall-through. - case Monotonic: v = 1 << 0; break; - // case Consume: v = 1 << 1; break; // Not specified yet. - case Acquire: v = 1 << 2; break; - case Release: v = 1 << 3; break; - case AcquireRelease: v = 1 << 4; break; - case SequentiallyConsistent: v = 1 << 5; break; + case Monotonic: v = 0; break; + // case Consume: v = 1; break; // Not specified yet. + case Acquire: v = 2; break; + case Release: v = 3; break; + case AcquireRelease: v = 4; break; + case SequentiallyConsistent: v = 5; break; } - // +100500 is temporal to migrate to new enum values. - return IRB->getInt32(v + 100500); + return IRB->getInt32(v); } bool ThreadSanitizer::instrumentAtomic(Instruction *I) { diff --git a/test/Instrumentation/ThreadSanitizer/atomic.ll b/test/Instrumentation/ThreadSanitizer/atomic.ll index 672e47e1f2..107dbdc0f2 100644 --- a/test/Instrumentation/ThreadSanitizer/atomic.ll +++ b/test/Instrumentation/ThreadSanitizer/atomic.ll @@ -8,7 +8,7 @@ entry: ret i8 %0 } ; CHECK: atomic8_load_unordered -; CHECK: call i8 @__tsan_atomic8_load(i8* %a, i32 100501) +; CHECK: call i8 @__tsan_atomic8_load(i8* %a, i32 0) define i8 @atomic8_load_monotonic(i8* %a) nounwind uwtable { entry: @@ -16,7 +16,7 @@ entry: ret i8 %0 } ; CHECK: atomic8_load_monotonic -; CHECK: call i8 @__tsan_atomic8_load(i8* %a, i32 100501) +; CHECK: call i8 @__tsan_atomic8_load(i8* %a, i32 0) define i8 @atomic8_load_acquire(i8* %a) nounwind uwtable { entry: @@ -24,7 +24,7 @@ entry: ret i8 %0 } ; CHECK: atomic8_load_acquire -; CHECK: call i8 @__tsan_atomic8_load(i8* %a, i32 100504) +; CHECK: call i8 @__tsan_atomic8_load(i8* %a, i32 2) define i8 @atomic8_load_seq_cst(i8* %a) nounwind uwtable { entry: @@ -32,7 +32,7 @@ entry: ret i8 %0 } ; CHECK: atomic8_load_seq_cst -; CHECK: call i8 @__tsan_atomic8_load(i8* %a, i32 100532) +; CHECK: call i8 @__tsan_atomic8_load(i8* %a, i32 5) define void @atomic8_store_unordered(i8* %a) nounwind uwtable { entry: @@ -40,7 +40,7 @@ entry: ret void } ; CHECK: atomic8_store_unordered -; CHECK: call void @__tsan_atomic8_store(i8* %a, i8 0, i32 100501) +; CHECK: call void @__tsan_atomic8_store(i8* %a, i8 0, i32 0) define void @atomic8_store_monotonic(i8* %a) nounwind uwtable { entry: @@ -48,7 +48,7 @@ entry: ret void } ; CHECK: atomic8_store_monotonic -; CHECK: call void @__tsan_atomic8_store(i8* %a, i8 0, i32 100501) +; CHECK: call void @__tsan_atomic8_store(i8* %a, i8 0, i32 0) define void @atomic8_store_release(i8* %a) nounwind uwtable { entry: @@ -56,7 +56,7 @@ entry: ret void } ; CHECK: atomic8_store_release -; CHECK: call void @__tsan_atomic8_store(i8* %a, i8 0, i32 100508) +; CHECK: call void @__tsan_atomic8_store(i8* %a, i8 0, i32 3) define void @atomic8_store_seq_cst(i8* %a) nounwind uwtable { entry: @@ -64,7 +64,7 @@ entry: ret void } ; CHECK: atomic8_store_seq_cst -; CHECK: call void @__tsan_atomic8_store(i8* %a, i8 0, i32 100532) +; CHECK: call void @__tsan_atomic8_store(i8* %a, i8 0, i32 5) define void @atomic8_xchg_monotonic(i8* %a) nounwind uwtable { entry: @@ -72,7 +72,7 @@ entry: ret void } ; CHECK: atomic8_xchg_monotonic -; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 100501) +; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 0) define void @atomic8_add_monotonic(i8* %a) nounwind uwtable { entry: @@ -80,7 +80,7 @@ entry: ret void } ; CHECK: atomic8_add_monotonic -; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 100501) +; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 0) define void @atomic8_sub_monotonic(i8* %a) nounwind uwtable { entry: @@ -88,7 +88,7 @@ entry: ret void } ; CHECK: atomic8_sub_monotonic -; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 100501) +; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 0) define void @atomic8_and_monotonic(i8* %a) nounwind uwtable { entry: @@ -96,7 +96,7 @@ entry: ret void } ; CHECK: atomic8_and_monotonic -; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 100501) +; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 0) define void @atomic8_or_monotonic(i8* %a) nounwind uwtable { entry: @@ -104,7 +104,7 @@ entry: ret void } ; CHECK: atomic8_or_monotonic -; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 100501) +; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 0) define void @atomic8_xor_monotonic(i8* %a) nounwind uwtable { entry: @@ -112,7 +112,7 @@ entry: ret void } ; CHECK: atomic8_xor_monotonic -; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 100501) +; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 0) define void @atomic8_xchg_acquire(i8* %a) nounwind uwtable { entry: @@ -120,7 +120,7 @@ entry: ret void } ; CHECK: atomic8_xchg_acquire -; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 100504) +; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 2) define void @atomic8_add_acquire(i8* %a) nounwind uwtable { entry: @@ -128,7 +128,7 @@ entry: ret void } ; CHECK: atomic8_add_acquire -; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 100504) +; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 2) define void @atomic8_sub_acquire(i8* %a) nounwind uwtable { entry: @@ -136,7 +136,7 @@ entry: ret void } ; CHECK: atomic8_sub_acquire -; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 100504) +; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 2) define void @atomic8_and_acquire(i8* %a) nounwind uwtable { entry: @@ -144,7 +144,7 @@ entry: ret void } ; CHECK: atomic8_and_acquire -; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 100504) +; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 2) define void @atomic8_or_acquire(i8* %a) nounwind uwtable { entry: @@ -152,7 +152,7 @@ entry: ret void } ; CHECK: atomic8_or_acquire -; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 100504) +; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 2) define void @atomic8_xor_acquire(i8* %a) nounwind uwtable { entry: @@ -160,7 +160,7 @@ entry: ret void } ; CHECK: atomic8_xor_acquire -; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 100504) +; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 2) define void @atomic8_xchg_release(i8* %a) nounwind uwtable { entry: @@ -168,7 +168,7 @@ entry: ret void } ; CHECK: atomic8_xchg_release -; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 100508) +; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 3) define void @atomic8_add_release(i8* %a) nounwind uwtable { entry: @@ -176,7 +176,7 @@ entry: ret void } ; CHECK: atomic8_add_release -; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 100508) +; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 3) define void @atomic8_sub_release(i8* %a) nounwind uwtable { entry: @@ -184,7 +184,7 @@ entry: ret void } ; CHECK: atomic8_sub_release -; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 100508) +; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 3) define void @atomic8_and_release(i8* %a) nounwind uwtable { entry: @@ -192,7 +192,7 @@ entry: ret void } ; CHECK: atomic8_and_release -; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 100508) +; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 3) define void @atomic8_or_release(i8* %a) nounwind uwtable { entry: @@ -200,7 +200,7 @@ entry: ret void } ; CHECK: atomic8_or_release -; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 100508) +; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 3) define void @atomic8_xor_release(i8* %a) nounwind uwtable { entry: @@ -208,7 +208,7 @@ entry: ret void } ; CHECK: atomic8_xor_release -; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 100508) +; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 3) define void @atomic8_xchg_acq_rel(i8* %a) nounwind uwtable { entry: @@ -216,7 +216,7 @@ entry: ret void } ; CHECK: atomic8_xchg_acq_rel -; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 100516) +; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 4) define void @atomic8_add_acq_rel(i8* %a) nounwind uwtable { entry: @@ -224,7 +224,7 @@ entry: ret void } ; CHECK: atomic8_add_acq_rel -; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 100516) +; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 4) define void @atomic8_sub_acq_rel(i8* %a) nounwind uwtable { entry: @@ -232,7 +232,7 @@ entry: ret void } ; CHECK: atomic8_sub_acq_rel -; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 100516) +; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 4) define void @atomic8_and_acq_rel(i8* %a) nounwind uwtable { entry: @@ -240,7 +240,7 @@ entry: ret void } ; CHECK: atomic8_and_acq_rel -; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 100516) +; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 4) define void @atomic8_or_acq_rel(i8* %a) nounwind uwtable { entry: @@ -248,7 +248,7 @@ entry: ret void } ; CHECK: atomic8_or_acq_rel -; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 100516) +; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 4) define void @atomic8_xor_acq_rel(i8* %a) nounwind uwtable { entry: @@ -256,7 +256,7 @@ entry: ret void } ; CHECK: atomic8_xor_acq_rel -; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 100516) +; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 4) define void @atomic8_xchg_seq_cst(i8* %a) nounwind uwtable { entry: @@ -264,7 +264,7 @@ entry: ret void } ; CHECK: atomic8_xchg_seq_cst -; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 100532) +; CHECK: call i8 @__tsan_atomic8_exchange(i8* %a, i8 0, i32 5) define void @atomic8_add_seq_cst(i8* %a) nounwind uwtable { entry: @@ -272,7 +272,7 @@ entry: ret void } ; CHECK: atomic8_add_seq_cst -; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 100532) +; CHECK: call i8 @__tsan_atomic8_fetch_add(i8* %a, i8 0, i32 5) define void @atomic8_sub_seq_cst(i8* %a) nounwind uwtable { entry: @@ -280,7 +280,7 @@ entry: ret void } ; CHECK: atomic8_sub_seq_cst -; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 100532) +; CHECK: call i8 @__tsan_atomic8_fetch_sub(i8* %a, i8 0, i32 5) define void @atomic8_and_seq_cst(i8* %a) nounwind uwtable { entry: @@ -288,7 +288,7 @@ entry: ret void } ; CHECK: atomic8_and_seq_cst -; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 100532) +; CHECK: call i8 @__tsan_atomic8_fetch_and(i8* %a, i8 0, i32 5) define void @atomic8_or_seq_cst(i8* %a) nounwind uwtable { entry: @@ -296,7 +296,7 @@ entry: ret void } ; CHECK: atomic8_or_seq_cst -; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 100532) +; CHECK: call i8 @__tsan_atomic8_fetch_or(i8* %a, i8 0, i32 5) define void @atomic8_xor_seq_cst(i8* %a) nounwind uwtable { entry: @@ -304,7 +304,7 @@ entry: ret void } ; CHECK: atomic8_xor_seq_cst -; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 100532) +; CHECK: call i8 @__tsan_atomic8_fetch_xor(i8* %a, i8 0, i32 5) define void @atomic8_cas_monotonic(i8* %a) nounwind uwtable { entry: @@ -312,7 +312,7 @@ entry: ret void } ; CHECK: atomic8_cas_monotonic -; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 100501) +; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 0) define void @atomic8_cas_acquire(i8* %a) nounwind uwtable { entry: @@ -320,7 +320,7 @@ entry: ret void } ; CHECK: atomic8_cas_acquire -; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 100504) +; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 2) define void @atomic8_cas_release(i8* %a) nounwind uwtable { entry: @@ -328,7 +328,7 @@ entry: ret void } ; CHECK: atomic8_cas_release -; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 100508) +; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 3) define void @atomic8_cas_acq_rel(i8* %a) nounwind uwtable { entry: @@ -336,7 +336,7 @@ entry: ret void } ; CHECK: atomic8_cas_acq_rel -; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 100516) +; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 4) define void @atomic8_cas_seq_cst(i8* %a) nounwind uwtable { entry: @@ -344,7 +344,7 @@ entry: ret void } ; CHECK: atomic8_cas_seq_cst -; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 100532) +; CHECK: call i8 @__tsan_atomic8_compare_exchange_val(i8* %a, i8 0, i8 1, i32 5) define i16 @atomic16_load_unordered(i16* %a) nounwind uwtable { entry: @@ -352,7 +352,7 @@ entry: ret i16 %0 } ; CHECK: atomic16_load_unordered -; CHECK: call i16 @__tsan_atomic16_load(i16* %a, i32 100501) +; CHECK: call i16 @__tsan_atomic16_load(i16* %a, i32 0) define i16 @atomic16_load_monotonic(i16* %a) nounwind uwtable { entry: @@ -360,7 +360,7 @@ entry: ret i16 %0 } ; CHECK: atomic16_load_monotonic -; CHECK: call i16 @__tsan_atomic16_load(i16* %a, i32 100501) +; CHECK: call i16 @__tsan_atomic16_load(i16* %a, i32 0) define i16 @atomic16_load_acquire(i16* %a) nounwind uwtable { entry: @@ -368,7 +368,7 @@ entry: ret i16 %0 } ; CHECK: atomic16_load_acquire -; CHECK: call i16 @__tsan_atomic16_load(i16* %a, i32 100504) +; CHECK: call i16 @__tsan_atomic16_load(i16* %a, i32 2) define i16 @atomic16_load_seq_cst(i16* %a) nounwind uwtable { entry: @@ -376,7 +376,7 @@ entry: ret i16 %0 } ; CHECK: atomic16_load_seq_cst -; CHECK: call i16 @__tsan_atomic16_load(i16* %a, i32 100532) +; CHECK: call i16 @__tsan_atomic16_load(i16* %a, i32 5) define void @atomic16_store_unordered(i16* %a) nounwind uwtable { entry: @@ -384,7 +384,7 @@ entry: ret void } ; CHECK: atomic16_store_unordered -; CHECK: call void @__tsan_atomic16_store(i16* %a, i16 0, i32 100501) +; CHECK: call void @__tsan_atomic16_store(i16* %a, i16 0, i32 0) define void @atomic16_store_monotonic(i16* %a) nounwind uwtable { entry: @@ -392,7 +392,7 @@ entry: ret void } ; CHECK: atomic16_store_monotonic -; CHECK: call void @__tsan_atomic16_store(i16* %a, i16 0, i32 100501) +; CHECK: call void @__tsan_atomic16_store(i16* %a, i16 0, i32 0) define void @atomic16_store_release(i16* %a) nounwind uwtable { entry: @@ -400,7 +400,7 @@ entry: ret void } ; CHECK: atomic16_store_release -; CHECK: call void @__tsan_atomic16_store(i16* %a, i16 0, i32 100508) +; CHECK: call void @__tsan_atomic16_store(i16* %a, i16 0, i32 3) define void @atomic16_store_seq_cst(i16* %a) nounwind uwtable { entry: @@ -408,7 +408,7 @@ entry: ret void } ; CHECK: atomic16_store_seq_cst -; CHECK: call void @__tsan_atomic16_store(i16* %a, i16 0, i32 100532) +; CHECK: call void @__tsan_atomic16_store(i16* %a, i16 0, i32 5) define void @atomic16_xchg_monotonic(i16* %a) nounwind uwtable { entry: @@ -416,7 +416,7 @@ entry: ret void } ; CHECK: atomic16_xchg_monotonic -; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 100501) +; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 0) define void @atomic16_add_monotonic(i16* %a) nounwind uwtable { entry: @@ -424,7 +424,7 @@ entry: ret void } ; CHECK: atomic16_add_monotonic -; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 100501) +; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 0) define void @atomic16_sub_monotonic(i16* %a) nounwind uwtable { entry: @@ -432,7 +432,7 @@ entry: ret void } ; CHECK: atomic16_sub_monotonic -; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 100501) +; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 0) define void @atomic16_and_monotonic(i16* %a) nounwind uwtable { entry: @@ -440,7 +440,7 @@ entry: ret void } ; CHECK: atomic16_and_monotonic -; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 100501) +; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 0) define void @atomic16_or_monotonic(i16* %a) nounwind uwtable { entry: @@ -448,7 +448,7 @@ entry: ret void } ; CHECK: atomic16_or_monotonic -; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 100501) +; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 0) define void @atomic16_xor_monotonic(i16* %a) nounwind uwtable { entry: @@ -456,7 +456,7 @@ entry: ret void } ; CHECK: atomic16_xor_monotonic -; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 100501) +; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 0) define void @atomic16_xchg_acquire(i16* %a) nounwind uwtable { entry: @@ -464,7 +464,7 @@ entry: ret void } ; CHECK: atomic16_xchg_acquire -; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 100504) +; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 2) define void @atomic16_add_acquire(i16* %a) nounwind uwtable { entry: @@ -472,7 +472,7 @@ entry: ret void } ; CHECK: atomic16_add_acquire -; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 100504) +; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 2) define void @atomic16_sub_acquire(i16* %a) nounwind uwtable { entry: @@ -480,7 +480,7 @@ entry: ret void } ; CHECK: atomic16_sub_acquire -; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 100504) +; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 2) define void @atomic16_and_acquire(i16* %a) nounwind uwtable { entry: @@ -488,7 +488,7 @@ entry: ret void } ; CHECK: atomic16_and_acquire -; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 100504) +; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 2) define void @atomic16_or_acquire(i16* %a) nounwind uwtable { entry: @@ -496,7 +496,7 @@ entry: ret void } ; CHECK: atomic16_or_acquire -; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 100504) +; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 2) define void @atomic16_xor_acquire(i16* %a) nounwind uwtable { entry: @@ -504,7 +504,7 @@ entry: ret void } ; CHECK: atomic16_xor_acquire -; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 100504) +; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 2) define void @atomic16_xchg_release(i16* %a) nounwind uwtable { entry: @@ -512,7 +512,7 @@ entry: ret void } ; CHECK: atomic16_xchg_release -; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 100508) +; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 3) define void @atomic16_add_release(i16* %a) nounwind uwtable { entry: @@ -520,7 +520,7 @@ entry: ret void } ; CHECK: atomic16_add_release -; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 100508) +; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 3) define void @atomic16_sub_release(i16* %a) nounwind uwtable { entry: @@ -528,7 +528,7 @@ entry: ret void } ; CHECK: atomic16_sub_release -; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 100508) +; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 3) define void @atomic16_and_release(i16* %a) nounwind uwtable { entry: @@ -536,7 +536,7 @@ entry: ret void } ; CHECK: atomic16_and_release -; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 100508) +; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 3) define void @atomic16_or_release(i16* %a) nounwind uwtable { entry: @@ -544,7 +544,7 @@ entry: ret void } ; CHECK: atomic16_or_release -; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 100508) +; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 3) define void @atomic16_xor_release(i16* %a) nounwind uwtable { entry: @@ -552,7 +552,7 @@ entry: ret void } ; CHECK: atomic16_xor_release -; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 100508) +; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 3) define void @atomic16_xchg_acq_rel(i16* %a) nounwind uwtable { entry: @@ -560,7 +560,7 @@ entry: ret void } ; CHECK: atomic16_xchg_acq_rel -; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 100516) +; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 4) define void @atomic16_add_acq_rel(i16* %a) nounwind uwtable { entry: @@ -568,7 +568,7 @@ entry: ret void } ; CHECK: atomic16_add_acq_rel -; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 100516) +; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 4) define void @atomic16_sub_acq_rel(i16* %a) nounwind uwtable { entry: @@ -576,7 +576,7 @@ entry: ret void } ; CHECK: atomic16_sub_acq_rel -; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 100516) +; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 4) define void @atomic16_and_acq_rel(i16* %a) nounwind uwtable { entry: @@ -584,7 +584,7 @@ entry: ret void } ; CHECK: atomic16_and_acq_rel -; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 100516) +; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 4) define void @atomic16_or_acq_rel(i16* %a) nounwind uwtable { entry: @@ -592,7 +592,7 @@ entry: ret void } ; CHECK: atomic16_or_acq_rel -; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 100516) +; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 4) define void @atomic16_xor_acq_rel(i16* %a) nounwind uwtable { entry: @@ -600,7 +600,7 @@ entry: ret void } ; CHECK: atomic16_xor_acq_rel -; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 100516) +; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 4) define void @atomic16_xchg_seq_cst(i16* %a) nounwind uwtable { entry: @@ -608,7 +608,7 @@ entry: ret void } ; CHECK: atomic16_xchg_seq_cst -; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 100532) +; CHECK: call i16 @__tsan_atomic16_exchange(i16* %a, i16 0, i32 5) define void @atomic16_add_seq_cst(i16* %a) nounwind uwtable { entry: @@ -616,7 +616,7 @@ entry: ret void } ; CHECK: atomic16_add_seq_cst -; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 100532) +; CHECK: call i16 @__tsan_atomic16_fetch_add(i16* %a, i16 0, i32 5) define void @atomic16_sub_seq_cst(i16* %a) nounwind uwtable { entry: @@ -624,7 +624,7 @@ entry: ret void } ; CHECK: atomic16_sub_seq_cst -; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 100532) +; CHECK: call i16 @__tsan_atomic16_fetch_sub(i16* %a, i16 0, i32 5) define void @atomic16_and_seq_cst(i16* %a) nounwind uwtable { entry: @@ -632,7 +632,7 @@ entry: ret void } ; CHECK: atomic16_and_seq_cst -; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 100532) +; CHECK: call i16 @__tsan_atomic16_fetch_and(i16* %a, i16 0, i32 5) define void @atomic16_or_seq_cst(i16* %a) nounwind uwtable { entry: @@ -640,7 +640,7 @@ entry: ret void } ; CHECK: atomic16_or_seq_cst -; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 100532) +; CHECK: call i16 @__tsan_atomic16_fetch_or(i16* %a, i16 0, i32 5) define void @atomic16_xor_seq_cst(i16* %a) nounwind uwtable { entry: @@ -648,7 +648,7 @@ entry: ret void } ; CHECK: atomic16_xor_seq_cst -; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 100532) +; CHECK: call i16 @__tsan_atomic16_fetch_xor(i16* %a, i16 0, i32 5) define void @atomic16_cas_monotonic(i16* %a) nounwind uwtable { entry: @@ -656,7 +656,7 @@ entry: ret void } ; CHECK: atomic16_cas_monotonic -; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 100501) +; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 0) define void @atomic16_cas_acquire(i16* %a) nounwind uwtable { entry: @@ -664,7 +664,7 @@ entry: ret void } ; CHECK: atomic16_cas_acquire -; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 100504) +; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 2) define void @atomic16_cas_release(i16* %a) nounwind uwtable { entry: @@ -672,7 +672,7 @@ entry: ret void } ; CHECK: atomic16_cas_release -; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 100508) +; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 3) define void @atomic16_cas_acq_rel(i16* %a) nounwind uwtable { entry: @@ -680,7 +680,7 @@ entry: ret void } ; CHECK: atomic16_cas_acq_rel -; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 100516) +; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 4) define void @atomic16_cas_seq_cst(i16* %a) nounwind uwtable { entry: @@ -688,7 +688,7 @@ entry: ret void } ; CHECK: atomic16_cas_seq_cst -; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 100532) +; CHECK: call i16 @__tsan_atomic16_compare_exchange_val(i16* %a, i16 0, i16 1, i32 5) define i32 @atomic32_load_unordered(i32* %a) nounwind uwtable { entry: @@ -696,7 +696,7 @@ entry: ret i32 %0 } ; CHECK: atomic32_load_unordered -; CHECK: call i32 @__tsan_atomic32_load(i32* %a, i32 100501) +; CHECK: call i32 @__tsan_atomic32_load(i32* %a, i32 0) define i32 @atomic32_load_monotonic(i32* %a) nounwind uwtable { entry: @@ -704,7 +704,7 @@ entry: ret i32 %0 } ; CHECK: atomic32_load_monotonic -; CHECK: call i32 @__tsan_atomic32_load(i32* %a, i32 100501) +; CHECK: call i32 @__tsan_atomic32_load(i32* %a, i32 0) define i32 @atomic32_load_acquire(i32* %a) nounwind uwtable { entry: @@ -712,7 +712,7 @@ entry: ret i32 %0 } ; CHECK: atomic32_load_acquire -; CHECK: call i32 @__tsan_atomic32_load(i32* %a, i32 100504) +; CHECK: call i32 @__tsan_atomic32_load(i32* %a, i32 2) define i32 @atomic32_load_seq_cst(i32* %a) nounwind uwtable { entry: @@ -720,7 +720,7 @@ entry: ret i32 %0 } ; CHECK: atomic32_load_seq_cst -; CHECK: call i32 @__tsan_atomic32_load(i32* %a, i32 100532) +; CHECK: call i32 @__tsan_atomic32_load(i32* %a, i32 5) define void @atomic32_store_unordered(i32* %a) nounwind uwtable { entry: @@ -728,7 +728,7 @@ entry: ret void } ; CHECK: atomic32_store_unordered -; CHECK: call void @__tsan_atomic32_store(i32* %a, i32 0, i32 100501) +; CHECK: call void @__tsan_atomic32_store(i32* %a, i32 0, i32 0) define void @atomic32_store_monotonic(i32* %a) nounwind uwtable { entry: @@ -736,7 +736,7 @@ entry: ret void } ; CHECK: atomic32_store_monotonic -; CHECK: call void @__tsan_atomic32_store(i32* %a, i32 0, i32 100501) +; CHECK: call void @__tsan_atomic32_store(i32* %a, i32 0, i32 0) define void @atomic32_store_release(i32* %a) nounwind uwtable { entry: @@ -744,7 +744,7 @@ entry: ret void } ; CHECK: atomic32_store_release -; CHECK: call void @__tsan_atomic32_store(i32* %a, i32 0, i32 100508) +; CHECK: call void @__tsan_atomic32_store(i32* %a, i32 0, i32 3) define void @atomic32_store_seq_cst(i32* %a) nounwind uwtable { entry: @@ -752,7 +752,7 @@ entry: ret void } ; CHECK: atomic32_store_seq_cst -; CHECK: call void @__tsan_atomic32_store(i32* %a, i32 0, i32 100532) +; CHECK: call void @__tsan_atomic32_store(i32* %a, i32 0, i32 5) define void @atomic32_xchg_monotonic(i32* %a) nounwind uwtable { entry: @@ -760,7 +760,7 @@ entry: ret void } ; CHECK: atomic32_xchg_monotonic -; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 100501) +; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 0) define void @atomic32_add_monotonic(i32* %a) nounwind uwtable { entry: @@ -768,7 +768,7 @@ entry: ret void } ; CHECK: atomic32_add_monotonic -; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 100501) +; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 0) define void @atomic32_sub_monotonic(i32* %a) nounwind uwtable { entry: @@ -776,7 +776,7 @@ entry: ret void } ; CHECK: atomic32_sub_monotonic -; CHECK: call i32 @__tsan_atomic32_fetch_sub(i32* %a, i32 0, i32 100501) +; CHECK: call i32 @__tsan_atomic32_fetch_sub(i32* %a, i32 0, i32 0) define void @atomic32_and_monotonic(i32* %a) nounwind uwtable { entry: @@ -784,7 +784,7 @@ entry: ret void } ; CHECK: atomic32_and_monotonic -; CHECK: call i32 @__tsan_atomic32_fetch_and(i32* %a, i32 0, i32 100501) +; CHECK: call i32 @__tsan_atomic32_fetch_and(i32* %a, i32 0, i32 0) define void @atomic32_or_monotonic(i32* %a) nounwind uwtable { entry: @@ -792,7 +792,7 @@ entry: ret void } ; CHECK: atomic32_or_monotonic -; CHECK: call i32 @__tsan_atomic32_fetch_or(i32* %a, i32 0, i32 100501) +; CHECK: call i32 @__tsan_atomic32_fetch_or(i32* %a, i32 0, i32 0) define void @atomic32_xor_monotonic(i32* %a) nounwind uwtable { entry: @@ -800,7 +800,7 @@ entry: ret void } ; CHECK: atomic32_xor_monotonic -; CHECK: call i32 @__tsan_atomic32_fetch_xor(i32* %a, i32 0, i32 100501) +; CHECK: call i32 @__tsan_atomic32_fetch_xor(i32* %a, i32 0, i32 0) define void @atomic32_xchg_acquire(i32* %a) nounwind uwtable { entry: @@ -808,7 +808,7 @@ entry: ret void } ; CHECK: atomic32_xchg_acquire -; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 100504) +; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 2) define void @atomic32_add_acquire(i32* %a) nounwind uwtable { entry: @@ -816,7 +816,7 @@ entry: ret void } ; CHECK: atomic32_add_acquire -; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 100504) +; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 2) define void @atomic32_sub_acquire(i32* %a) nounwind uwtable { entry: @@ -824,7 +824,7 @@ entry: ret void } ; CHECK: atomic32_sub_acquire -; CHECK: call i32 @__tsan_atomic32_fetch_sub(i32* %a, i32 0, i32 100504) +; CHECK: call i32 @__tsan_atomic32_fetch_sub(i32* %a, i32 0, i32 2) define void @atomic32_and_acquire(i32* %a) nounwind uwtable { entry: @@ -832,7 +832,7 @@ entry: ret void } ; CHECK: atomic32_and_acquire -; CHECK: call i32 @__tsan_atomic32_fetch_and(i32* %a, i32 0, i32 100504) +; CHECK: call i32 @__tsan_atomic32_fetch_and(i32* %a, i32 0, i32 2) define void @atomic32_or_acquire(i32* %a) nounwind uwtable { entry: @@ -840,7 +840,7 @@ entry: ret void } ; CHECK: atomic32_or_acquire -; CHECK: call i32 @__tsan_atomic32_fetch_or(i32* %a, i32 0, i32 100504) +; CHECK: call i32 @__tsan_atomic32_fetch_or(i32* %a, i32 0, i32 2) define void @atomic32_xor_acquire(i32* %a) nounwind uwtable { entry: @@ -848,7 +848,7 @@ entry: ret void } ; CHECK: atomic32_xor_acquire -; CHECK: call i32 @__tsan_atomic32_fetch_xor(i32* %a, i32 0, i32 100504) +; CHECK: call i32 @__tsan_atomic32_fetch_xor(i32* %a, i32 0, i32 2) define void @atomic32_xchg_release(i32* %a) nounwind uwtable { entry: @@ -856,7 +856,7 @@ entry: ret void } ; CHECK: atomic32_xchg_release -; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 100508) +; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 3) define void @atomic32_add_release(i32* %a) nounwind uwtable { entry: @@ -864,7 +864,7 @@ entry: ret void } ; CHECK: atomic32_add_release -; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 100508) +; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 3) define void @atomic32_sub_release(i32* %a) nounwind uwtable { entry: @@ -872,7 +872,7 @@ entry: ret void } ; CHECK: atomic32_sub_release -; CHECK: call i32 @__tsan_atomic32_fetch_sub(i32* %a, i32 0, i32 100508) +; CHECK: call i32 @__tsan_atomic32_fetch_sub(i32* %a, i32 0, i32 3) define void @atomic32_and_release(i32* %a) nounwind uwtable { entry: @@ -880,7 +880,7 @@ entry: ret void } ; CHECK: atomic32_and_release -; CHECK: call i32 @__tsan_atomic32_fetch_and(i32* %a, i32 0, i32 100508) +; CHECK: call i32 @__tsan_atomic32_fetch_and(i32* %a, i32 0, i32 3) define void @atomic32_or_release(i32* %a) nounwind uwtable { entry: @@ -888,7 +888,7 @@ entry: ret void } ; CHECK: atomic32_or_release -; CHECK: call i32 @__tsan_atomic32_fetch_or(i32* %a, i32 0, i32 100508) +; CHECK: call i32 @__tsan_atomic32_fetch_or(i32* %a, i32 0, i32 3) define void @atomic32_xor_release(i32* %a) nounwind uwtable { entry: @@ -896,7 +896,7 @@ entry: ret void } ; CHECK: atomic32_xor_release -; CHECK: call i32 @__tsan_atomic32_fetch_xor(i32* %a, i32 0, i32 100508) +; CHECK: call i32 @__tsan_atomic32_fetch_xor(i32* %a, i32 0, i32 3) define void @atomic32_xchg_acq_rel(i32* %a) nounwind uwtable { entry: @@ -904,7 +904,7 @@ entry: ret void } ; CHECK: atomic32_xchg_acq_rel -; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 100516) +; CHECK: call i32 @__tsan_atomic32_exchange(i32* %a, i32 0, i32 4) define void @atomic32_add_acq_rel(i32* %a) nounwind uwtable { entry: @@ -912,7 +912,7 @@ entry: ret void } ; CHECK: atomic32_add_acq_rel -; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 100516) +; CHECK: call i32 @__tsan_atomic32_fetch_add(i32* %a, i32 0, i32 4) de |