62 lines
1.6 KiB
LLVM
62 lines
1.6 KiB
LLVM
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -verify-machineinstrs | FileCheck %s
|
|
; RUN: llc < %s -mtriple=i686-unknown-unknown -verify-machineinstrs | FileCheck %s
|
|
|
|
; Make sure that flags are properly preserved despite atomic optimizations.
|
|
|
|
define i32 @atomic_and_flags_1(i8* %p, i32 %a, i32 %b) {
|
|
; CHECK-LABEL: atomic_and_flags_1:
|
|
|
|
; Generate flags value, and use it.
|
|
; CHECK: cmpl
|
|
; CHECK-NEXT: jne
|
|
%cmp = icmp eq i32 %a, %b
|
|
br i1 %cmp, label %L1, label %L2
|
|
|
|
L1:
|
|
; The following pattern will get folded.
|
|
; CHECK: incb
|
|
%1 = load atomic i8, i8* %p seq_cst, align 1
|
|
%2 = add i8 %1, 1 ; This forces the INC instruction to be generated.
|
|
store atomic i8 %2, i8* %p release, align 1
|
|
|
|
; Use the comparison result again. We need to rematerialize the comparison
|
|
; somehow. This test checks that cmpl gets emitted again, but any
|
|
; rematerialization would work (the optimizer used to clobber the flags with
|
|
; the add).
|
|
; CHECK-NEXT: cmpl
|
|
; CHECK-NEXT: jne
|
|
br i1 %cmp, label %L3, label %L4
|
|
|
|
L2:
|
|
ret i32 2
|
|
|
|
L3:
|
|
ret i32 3
|
|
|
|
L4:
|
|
ret i32 4
|
|
}
|
|
|
|
; Same as above, but using 2 as immediate to avoid the INC instruction.
|
|
define i32 @atomic_and_flags_2(i8* %p, i32 %a, i32 %b) {
|
|
; CHECK-LABEL: atomic_and_flags_2:
|
|
; CHECK: cmpl
|
|
; CHECK-NEXT: jne
|
|
%cmp = icmp eq i32 %a, %b
|
|
br i1 %cmp, label %L1, label %L2
|
|
L1:
|
|
; CHECK: addb
|
|
%1 = load atomic i8, i8* %p seq_cst, align 1
|
|
%2 = add i8 %1, 2
|
|
store atomic i8 %2, i8* %p release, align 1
|
|
; CHECK-NEXT: cmpl
|
|
; CHECK-NEXT: jne
|
|
br i1 %cmp, label %L3, label %L4
|
|
L2:
|
|
ret i32 2
|
|
L3:
|
|
ret i32 3
|
|
L4:
|
|
ret i32 4
|
|
}
|