330 lines
10 KiB
LLVM
330 lines
10 KiB
LLVM
|
; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
|
||
|
; RUN: opt -basic-aa -dse -S < %s | FileCheck %s
|
||
|
|
||
|
target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128-n8:16:32:64"
|
||
|
target triple = "x86_64-apple-macosx10.7.0"
|
||
|
|
||
|
; Sanity tests for atomic stores.
|
||
|
; Note that it turns out essentially every transformation DSE does is legal on
|
||
|
; atomic ops, just some transformations are not allowed across release-acquire pairs.
|
||
|
|
||
|
@x = common global i32 0, align 4
|
||
|
@y = common global i32 0, align 4
|
||
|
|
||
|
declare void @randomop(i32*)
|
||
|
|
||
|
; DSE across unordered store (allowed)
|
||
|
define void @test1() {
|
||
|
; CHECK-LABEL: @test1(
|
||
|
; CHECK-NEXT: store atomic i32 0, i32* @y unordered, align 4
|
||
|
; CHECK-NEXT: store i32 1, i32* @x, align 4
|
||
|
; CHECK-NEXT: ret void
|
||
|
;
|
||
|
store i32 0, i32* @x
|
||
|
store atomic i32 0, i32* @y unordered, align 4
|
||
|
store i32 1, i32* @x
|
||
|
ret void
|
||
|
}
|
||
|
|
||
|
; DSE remove unordered store (allowed)
|
||
|
define void @test4() {
|
||
|
; CHECK-LABEL: @test4(
|
||
|
; CHECK-NEXT: store i32 1, i32* @x, align 4
|
||
|
; CHECK-NEXT: ret void
|
||
|
;
|
||
|
store atomic i32 0, i32* @x unordered, align 4
|
||
|
store i32 1, i32* @x
|
||
|
ret void
|
||
|
}
|
||
|
|
||
|
; DSE unordered store overwriting non-atomic store (allowed)
|
||
|
define void @test5() {
|
||
|
; CHECK-LABEL: @test5(
|
||
|
; CHECK-NEXT: store atomic i32 1, i32* @x unordered, align 4
|
||
|
; CHECK-NEXT: ret void
|
||
|
;
|
||
|
store i32 0, i32* @x
|
||
|
store atomic i32 1, i32* @x unordered, align 4
|
||
|
ret void
|
||
|
}
|
||
|
|
||
|
; DSE no-op unordered atomic store (allowed)
|
||
|
define void @test6() {
|
||
|
; CHECK-LABEL: @test6(
|
||
|
; CHECK-NEXT: ret void
|
||
|
;
|
||
|
%x = load atomic i32, i32* @x unordered, align 4
|
||
|
store atomic i32 %x, i32* @x unordered, align 4
|
||
|
ret void
|
||
|
}
|
||
|
|
||
|
; DSE seq_cst store (be conservative; DSE doesn't have infrastructure
|
||
|
; to reason about atomic operations).
|
||
|
define void @test7() {
|
||
|
; CHECK-LABEL: @test7(
|
||
|
; CHECK-NEXT: [[A:%.*]] = alloca i32, align 4
|
||
|
; CHECK-NEXT: store atomic i32 0, i32* [[A]] seq_cst, align 4
|
||
|
; CHECK-NEXT: ret void
|
||
|
;
|
||
|
%a = alloca i32
|
||
|
store atomic i32 0, i32* %a seq_cst, align 4
|
||
|
ret void
|
||
|
}
|
||
|
|
||
|
; DSE and seq_cst load (be conservative; DSE doesn't have infrastructure
|
||
|
; to reason about atomic operations).
|
||
|
define i32 @test8() {
|
||
|
; CHECK-LABEL: @test8(
|
||
|
; CHECK-NEXT: [[A:%.*]] = alloca i32, align 4
|
||
|
; CHECK-NEXT: call void @randomop(i32* [[A]])
|
||
|
; CHECK-NEXT: store i32 0, i32* [[A]], align 4
|
||
|
; CHECK-NEXT: [[X:%.*]] = load atomic i32, i32* @x seq_cst, align 4
|
||
|
; CHECK-NEXT: ret i32 [[X]]
|
||
|
;
|
||
|
%a = alloca i32
|
||
|
call void @randomop(i32* %a)
|
||
|
store i32 0, i32* %a, align 4
|
||
|
%x = load atomic i32, i32* @x seq_cst, align 4
|
||
|
ret i32 %x
|
||
|
}
|
||
|
|
||
|
; DSE across monotonic store (allowed as long as the eliminated store isUnordered)
|
||
|
define void @test10() {
|
||
|
; CHECK-LABEL: test10
|
||
|
; CHECK-NOT: store i32 0
|
||
|
; CHECK: store i32 1
|
||
|
store i32 0, i32* @x
|
||
|
store atomic i32 42, i32* @y monotonic, align 4
|
||
|
store i32 1, i32* @x
|
||
|
ret void
|
||
|
}
|
||
|
|
||
|
; DSE across monotonic load (forbidden since the eliminated store is atomic)
|
||
|
define i32 @test11() {
|
||
|
; CHECK-LABEL: @test11(
|
||
|
; CHECK-NEXT: store atomic i32 0, i32* @x monotonic, align 4
|
||
|
; CHECK-NEXT: [[X:%.*]] = load atomic i32, i32* @y monotonic, align 4
|
||
|
; CHECK-NEXT: store atomic i32 1, i32* @x monotonic, align 4
|
||
|
; CHECK-NEXT: ret i32 [[X]]
|
||
|
;
|
||
|
store atomic i32 0, i32* @x monotonic, align 4
|
||
|
%x = load atomic i32, i32* @y monotonic, align 4
|
||
|
store atomic i32 1, i32* @x monotonic, align 4
|
||
|
ret i32 %x
|
||
|
}
|
||
|
|
||
|
; DSE across monotonic store (forbidden since the eliminated store is atomic)
|
||
|
define void @test12() {
|
||
|
; CHECK-LABEL: @test12(
|
||
|
; CHECK-NEXT: store atomic i32 0, i32* @x monotonic, align 4
|
||
|
; CHECK-NEXT: store atomic i32 42, i32* @y monotonic, align 4
|
||
|
; CHECK-NEXT: store atomic i32 1, i32* @x monotonic, align 4
|
||
|
; CHECK-NEXT: ret void
|
||
|
;
|
||
|
store atomic i32 0, i32* @x monotonic, align 4
|
||
|
store atomic i32 42, i32* @y monotonic, align 4
|
||
|
store atomic i32 1, i32* @x monotonic, align 4
|
||
|
ret void
|
||
|
}
|
||
|
|
||
|
; But DSE is not allowed across a release-acquire pair.
|
||
|
define i32 @test15() {
|
||
|
; CHECK-LABEL: @test15(
|
||
|
; CHECK-NEXT: store i32 0, i32* @x, align 4
|
||
|
; CHECK-NEXT: store atomic i32 0, i32* @y release, align 4
|
||
|
; CHECK-NEXT: [[X:%.*]] = load atomic i32, i32* @y acquire, align 4
|
||
|
; CHECK-NEXT: store i32 1, i32* @x, align 4
|
||
|
; CHECK-NEXT: ret i32 [[X]]
|
||
|
;
|
||
|
store i32 0, i32* @x
|
||
|
store atomic i32 0, i32* @y release, align 4
|
||
|
%x = load atomic i32, i32* @y acquire, align 4
|
||
|
store i32 1, i32* @x
|
||
|
ret i32 %x
|
||
|
}
|
||
|
|
||
|
@z = common global i64 0, align 4
|
||
|
@a = common global i64 0, align 4
|
||
|
|
||
|
; Be conservative, do not kill regular store.
|
||
|
define i64 @test_atomicrmw_0() {
|
||
|
; CHECK-LABEL: @test_atomicrmw_0(
|
||
|
; CHECK-NEXT: store i64 1, i64* @z, align 8
|
||
|
; CHECK-NEXT: [[RES:%.*]] = atomicrmw add i64* @z, i64 -1 monotonic
|
||
|
; CHECK-NEXT: ret i64 [[RES]]
|
||
|
;
|
||
|
store i64 1, i64* @z
|
||
|
%res = atomicrmw add i64* @z, i64 -1 monotonic
|
||
|
ret i64 %res
|
||
|
}
|
||
|
|
||
|
; Be conservative, do not kill regular store.
|
||
|
define i64 @test_atomicrmw_1() {
|
||
|
; CHECK-LABEL: @test_atomicrmw_1(
|
||
|
; CHECK-NEXT: store i64 1, i64* @z, align 8
|
||
|
; CHECK-NEXT: [[RES:%.*]] = atomicrmw add i64* @z, i64 -1 acq_rel
|
||
|
; CHECK-NEXT: ret i64 [[RES]]
|
||
|
;
|
||
|
store i64 1, i64* @z
|
||
|
%res = atomicrmw add i64* @z, i64 -1 acq_rel
|
||
|
ret i64 %res
|
||
|
}
|
||
|
|
||
|
; Monotonic atomicrmw should not block eliminating no-aliasing stores.
|
||
|
define i64 @test_atomicrmw_2() {
|
||
|
; CHECK-LABEL: @test_atomicrmw_2(
|
||
|
; CHECK-NEXT: [[RES:%.*]] = atomicrmw add i64* @a, i64 -1 monotonic
|
||
|
; CHECK-NEXT: store i64 2, i64* @z, align 8
|
||
|
; CHECK-NEXT: ret i64 [[RES]]
|
||
|
;
|
||
|
store i64 1, i64* @z
|
||
|
%res = atomicrmw add i64* @a, i64 -1 monotonic
|
||
|
store i64 2, i64* @z
|
||
|
ret i64 %res
|
||
|
}
|
||
|
|
||
|
; Be conservative, do not eliminate stores across atomic operations > monotonic.
|
||
|
define i64 @test_atomicrmw_3() {
|
||
|
; CHECK-LABEL: @test_atomicrmw_3(
|
||
|
; CHECK-NEXT: store i64 1, i64* @z, align 8
|
||
|
; CHECK-NEXT: [[RES:%.*]] = atomicrmw add i64* @a, i64 -1 release
|
||
|
; CHECK-NEXT: store i64 2, i64* @z, align 8
|
||
|
; CHECK-NEXT: ret i64 [[RES]]
|
||
|
;
|
||
|
store i64 1, i64* @z
|
||
|
%res = atomicrmw add i64* @a, i64 -1 release
|
||
|
store i64 2, i64* @z
|
||
|
ret i64 %res
|
||
|
}
|
||
|
|
||
|
; Be conservative, do not eliminate may-alias stores.
|
||
|
define i64 @test_atomicrmw_4(i64* %ptr) {
|
||
|
; CHECK-LABEL: @test_atomicrmw_4(
|
||
|
; CHECK-NEXT: store i64 1, i64* @z, align 8
|
||
|
; CHECK-NEXT: [[RES:%.*]] = atomicrmw add i64* [[PTR:%.*]], i64 -1 monotonic
|
||
|
; CHECK-NEXT: store i64 2, i64* @z, align 8
|
||
|
; CHECK-NEXT: ret i64 [[RES]]
|
||
|
;
|
||
|
store i64 1, i64* @z
|
||
|
%res = atomicrmw add i64* %ptr, i64 -1 monotonic
|
||
|
store i64 2, i64* @z
|
||
|
ret i64 %res
|
||
|
}
|
||
|
|
||
|
; Be conservative, do not eliminate aliasing stores.
|
||
|
define i64 @test_atomicrmw_5() {
|
||
|
; CHECK-LABEL: @test_atomicrmw_5(
|
||
|
; CHECK-NEXT: store i64 1, i64* @z, align 8
|
||
|
; CHECK-NEXT: [[RES:%.*]] = atomicrmw add i64* @z, i64 -1 monotonic
|
||
|
; CHECK-NEXT: store i64 2, i64* @z, align 8
|
||
|
; CHECK-NEXT: ret i64 [[RES]]
|
||
|
;
|
||
|
store i64 1, i64* @z
|
||
|
%res = atomicrmw add i64* @z, i64 -1 monotonic
|
||
|
store i64 2, i64* @z
|
||
|
ret i64 %res
|
||
|
}
|
||
|
|
||
|
; Be conservative, do not eliminate non-monotonic cmpxchg.
|
||
|
define { i32, i1} @test_cmpxchg_1() {
|
||
|
; CHECK-LABEL: @test_cmpxchg_1(
|
||
|
; CHECK-NEXT: store i32 1, i32* @x, align 4
|
||
|
; CHECK-NEXT: [[RET:%.*]] = cmpxchg volatile i32* @x, i32 10, i32 20 seq_cst monotonic
|
||
|
; CHECK-NEXT: store i32 2, i32* @x, align 4
|
||
|
; CHECK-NEXT: ret { i32, i1 } [[RET]]
|
||
|
;
|
||
|
store i32 1, i32* @x
|
||
|
%ret = cmpxchg volatile i32* @x, i32 10, i32 20 seq_cst monotonic
|
||
|
store i32 2, i32* @x
|
||
|
ret { i32, i1 } %ret
|
||
|
}
|
||
|
|
||
|
; Monotonic cmpxchg should not block DSE for non-aliasing stores.
|
||
|
define { i32, i1} @test_cmpxchg_2() {
|
||
|
; CHECK-LABEL: @test_cmpxchg_2(
|
||
|
; CHECK-NEXT: [[RET:%.*]] = cmpxchg volatile i32* @y, i32 10, i32 20 monotonic monotonic
|
||
|
; CHECK-NEXT: store i32 2, i32* @x, align 4
|
||
|
; CHECK-NEXT: ret { i32, i1 } [[RET]]
|
||
|
;
|
||
|
store i32 1, i32* @x
|
||
|
%ret = cmpxchg volatile i32* @y, i32 10, i32 20 monotonic monotonic
|
||
|
store i32 2, i32* @x
|
||
|
ret { i32, i1 } %ret
|
||
|
}
|
||
|
|
||
|
; Be conservative, do not eliminate non-monotonic cmpxchg.
|
||
|
define { i32, i1} @test_cmpxchg_3() {
|
||
|
; CHECK-LABEL: @test_cmpxchg_3(
|
||
|
; CHECK-NEXT: store i32 1, i32* @x, align 4
|
||
|
; CHECK-NEXT: [[RET:%.*]] = cmpxchg volatile i32* @y, i32 10, i32 20 seq_cst seq_cst
|
||
|
; CHECK-NEXT: store i32 2, i32* @x, align 4
|
||
|
; CHECK-NEXT: ret { i32, i1 } [[RET]]
|
||
|
;
|
||
|
store i32 1, i32* @x
|
||
|
%ret = cmpxchg volatile i32* @y, i32 10, i32 20 seq_cst seq_cst
|
||
|
store i32 2, i32* @x
|
||
|
ret { i32, i1 } %ret
|
||
|
}
|
||
|
|
||
|
; Be conservative, do not eliminate may-alias stores.
|
||
|
define { i32, i1} @test_cmpxchg_4(i32* %ptr) {
|
||
|
; CHECK-LABEL: @test_cmpxchg_4(
|
||
|
; CHECK-NEXT: store i32 1, i32* @x, align 4
|
||
|
; CHECK-NEXT: [[RET:%.*]] = cmpxchg volatile i32* [[PTR:%.*]], i32 10, i32 20 monotonic monotonic
|
||
|
; CHECK-NEXT: store i32 2, i32* @x, align 4
|
||
|
; CHECK-NEXT: ret { i32, i1 } [[RET]]
|
||
|
;
|
||
|
store i32 1, i32* @x
|
||
|
%ret = cmpxchg volatile i32* %ptr, i32 10, i32 20 monotonic monotonic
|
||
|
store i32 2, i32* @x
|
||
|
ret { i32, i1 } %ret
|
||
|
}
|
||
|
|
||
|
; Be conservative, do not eliminate alias stores.
|
||
|
define { i32, i1} @test_cmpxchg_5(i32* %ptr) {
|
||
|
; CHECK-LABEL: @test_cmpxchg_5(
|
||
|
; CHECK-NEXT: store i32 1, i32* @x, align 4
|
||
|
; CHECK-NEXT: [[RET:%.*]] = cmpxchg volatile i32* @x, i32 10, i32 20 monotonic monotonic
|
||
|
; CHECK-NEXT: store i32 2, i32* @x, align 4
|
||
|
; CHECK-NEXT: ret { i32, i1 } [[RET]]
|
||
|
;
|
||
|
store i32 1, i32* @x
|
||
|
%ret = cmpxchg volatile i32* @x, i32 10, i32 20 monotonic monotonic
|
||
|
store i32 2, i32* @x
|
||
|
ret { i32, i1 } %ret
|
||
|
}
|
||
|
|
||
|
; **** Noop load->store tests **************************************************
|
||
|
|
||
|
; We can optimize unordered atomic loads or stores.
|
||
|
define void @test_load_atomic(i32* %Q) {
|
||
|
; CHECK-LABEL: @test_load_atomic(
|
||
|
; CHECK-NEXT: ret void
|
||
|
;
|
||
|
%a = load atomic i32, i32* %Q unordered, align 4
|
||
|
store atomic i32 %a, i32* %Q unordered, align 4
|
||
|
ret void
|
||
|
}
|
||
|
|
||
|
; We can optimize unordered atomic loads or stores.
|
||
|
define void @test_store_atomic(i32* %Q) {
|
||
|
; CHECK-LABEL: @test_store_atomic(
|
||
|
; CHECK-NEXT: ret void
|
||
|
;
|
||
|
%a = load i32, i32* %Q
|
||
|
store atomic i32 %a, i32* %Q unordered, align 4
|
||
|
ret void
|
||
|
}
|
||
|
|
||
|
; We can NOT optimize release atomic loads or stores.
|
||
|
define void @test_store_atomic_release(i32* %Q) {
|
||
|
; CHECK-LABEL: @test_store_atomic_release(
|
||
|
; CHECK-NEXT: [[A:%.*]] = load i32, i32* [[Q:%.*]], align 4
|
||
|
; CHECK-NEXT: store atomic i32 [[A]], i32* [[Q]] release, align 4
|
||
|
; CHECK-NEXT: ret void
|
||
|
;
|
||
|
%a = load i32, i32* %Q
|
||
|
store atomic i32 %a, i32* %Q release, align 4
|
||
|
ret void
|
||
|
}
|