153 lines
5.2 KiB
LLVM
153 lines
5.2 KiB
LLVM
|
; RUN: llc -verify-machineinstrs -mtriple=x86_64-unknown < %s | FileCheck %s --check-prefix=X64 --check-prefix=X64-ALL
|
||
|
; RUN: llc -verify-machineinstrs -mtriple=x86_64-unknown --x86-lvi-load-no-cbranch < %s | FileCheck %s --check-prefix=X64
|
||
|
; RUN: llc -O0 -verify-machineinstrs -mtriple=x86_64-unknown < %s | FileCheck %s --check-prefix=X64-NOOPT
|
||
|
|
||
|
; Function Attrs: noinline nounwind optnone uwtable
|
||
|
define dso_local i32 @test(i32** %secret, i32 %secret_size) #0 {
|
||
|
; X64-LABEL: test:
|
||
|
entry:
|
||
|
%secret.addr = alloca i32**, align 8
|
||
|
%secret_size.addr = alloca i32, align 4
|
||
|
%ret_val = alloca i32, align 4
|
||
|
%i = alloca i32, align 4
|
||
|
store i32** %secret, i32*** %secret.addr, align 8
|
||
|
store i32 %secret_size, i32* %secret_size.addr, align 4
|
||
|
store i32 0, i32* %ret_val, align 4
|
||
|
call void @llvm.x86.sse2.lfence()
|
||
|
store i32 0, i32* %i, align 4
|
||
|
br label %for.cond
|
||
|
|
||
|
; X64: # %bb.0: # %entry
|
||
|
; X64-NEXT: movq %rdi, -{{[0-9]+}}(%rsp)
|
||
|
; X64-NEXT: movl %esi, -{{[0-9]+}}(%rsp)
|
||
|
; X64-NEXT: movl $0, -{{[0-9]+}}(%rsp)
|
||
|
; X64-NEXT: lfence
|
||
|
; X64-NEXT: movl $0, -{{[0-9]+}}(%rsp)
|
||
|
; X64-NEXT: jmp .LBB0_1
|
||
|
|
||
|
; X64-NOOPT: # %bb.0: # %entry
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: movq %rdi, -{{[0-9]+}}(%rsp)
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: movl %esi, -{{[0-9]+}}(%rsp)
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: movl $0, -{{[0-9]+}}(%rsp)
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: movl $0, -{{[0-9]+}}(%rsp)
|
||
|
|
||
|
for.cond: ; preds = %for.inc, %entry
|
||
|
%0 = load i32, i32* %i, align 4
|
||
|
%1 = load i32, i32* %secret_size.addr, align 4
|
||
|
%cmp = icmp slt i32 %0, %1
|
||
|
br i1 %cmp, label %for.body, label %for.end
|
||
|
|
||
|
; X64: .LBB0_1: # %for.cond
|
||
|
; X64-NEXT: # =>This Inner Loop Header: Depth=1
|
||
|
; X64-NEXT: movl -{{[0-9]+}}(%rsp), %eax
|
||
|
; X64-ALL-NEXT: lfence
|
||
|
; X64-NEXT: cmpl -{{[0-9]+}}(%rsp), %eax
|
||
|
; X64-ALL-NEXT: lfence
|
||
|
; X64-NEXT: jge .LBB0_5
|
||
|
|
||
|
; X64-NOOPT: .LBB0_1: # %for.cond
|
||
|
; X64-NOOPT-NEXT: # =>This Inner Loop Header: Depth=1
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: movl -{{[0-9]+}}(%rsp), %eax
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: cmpl -{{[0-9]+}}(%rsp), %eax
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: jge .LBB0_6
|
||
|
|
||
|
for.body: ; preds = %for.cond
|
||
|
%2 = load i32, i32* %i, align 4
|
||
|
%rem = srem i32 %2, 2
|
||
|
%cmp1 = icmp eq i32 %rem, 0
|
||
|
br i1 %cmp1, label %if.then, label %if.end
|
||
|
|
||
|
; X64: # %bb.2: # %for.body
|
||
|
; X64-NEXT: # in Loop: Header=BB0_1 Depth=1
|
||
|
; X64-NEXT: movl -{{[0-9]+}}(%rsp), %eax
|
||
|
; X64-ALL-NEXT: lfence
|
||
|
; X64-NEXT: movl %eax, %ecx
|
||
|
; X64-NEXT: shrl $31, %ecx
|
||
|
; X64-NEXT: addl %eax, %ecx
|
||
|
; X64-NEXT: andl $-2, %ecx
|
||
|
; X64-NEXT: cmpl %ecx, %eax
|
||
|
; X64-NEXT: jne .LBB0_4
|
||
|
|
||
|
; X64-NOOPT: # %bb.2: # %for.body
|
||
|
; X64-NOOPT-NEXT: # in Loop: Header=BB0_1 Depth=1
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: movl -{{[0-9]+}}(%rsp), %eax
|
||
|
; X64-NOOPT-NEXT: movl $2, %ecx
|
||
|
; X64-NOOPT-NEXT: cltd
|
||
|
; X64-NOOPT-NEXT: idivl %ecx
|
||
|
; X64-NOOPT-NEXT: cmpl $0, %edx
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: jne .LBB0_4
|
||
|
|
||
|
if.then: ; preds = %for.body
|
||
|
%3 = load i32**, i32*** %secret.addr, align 8
|
||
|
%4 = load i32, i32* %ret_val, align 4
|
||
|
%idxprom = sext i32 %4 to i64
|
||
|
%arrayidx = getelementptr inbounds i32*, i32** %3, i64 %idxprom
|
||
|
%5 = load i32*, i32** %arrayidx, align 8
|
||
|
%6 = load i32, i32* %5, align 4
|
||
|
store i32 %6, i32* %ret_val, align 4
|
||
|
br label %if.end
|
||
|
|
||
|
; X64: # %bb.3: # %if.then
|
||
|
; X64-NEXT: # in Loop: Header=BB0_1 Depth=1
|
||
|
; X64-NEXT: movq -{{[0-9]+}}(%rsp), %rax
|
||
|
; X64-NEXT: lfence
|
||
|
; X64-NEXT: movslq -{{[0-9]+}}(%rsp), %rcx
|
||
|
; X64-NEXT: lfence
|
||
|
; X64-NEXT: movq (%rax,%rcx,8), %rax
|
||
|
; X64-NEXT: lfence
|
||
|
; X64-NEXT: movl (%rax), %eax
|
||
|
; X64-NEXT: movl %eax, -{{[0-9]+}}(%rsp)
|
||
|
; X64-NEXT: jmp .LBB0_4
|
||
|
|
||
|
; X64-NOOPT: # %bb.3: # %if.then
|
||
|
; X64-NOOPT-NEXT: # in Loop: Header=BB0_1 Depth=1
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: movq -{{[0-9]+}}(%rsp), %rax
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: movslq -{{[0-9]+}}(%rsp), %rcx
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: movq (%rax,%rcx,8), %rax
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: movl (%rax), %eax
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: movl %eax, -{{[0-9]+}}(%rsp)
|
||
|
|
||
|
if.end: ; preds = %if.then, %for.body
|
||
|
br label %for.inc
|
||
|
|
||
|
for.inc: ; preds = %if.end
|
||
|
%7 = load i32, i32* %i, align 4
|
||
|
%inc = add nsw i32 %7, 1
|
||
|
store i32 %inc, i32* %i, align 4
|
||
|
br label %for.cond
|
||
|
|
||
|
; X64-NOOPT: .LBB0_5: # %for.inc
|
||
|
; X64-NOOPT-NEXT: # in Loop: Header=BB0_1 Depth=1
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: movl -{{[0-9]+}}(%rsp), %eax
|
||
|
; X64-NOOPT-NEXT: addl $1, %eax
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: movl %eax, -{{[0-9]+}}(%rsp)
|
||
|
; X64-NOOPT-NEXT: lfence
|
||
|
; X64-NOOPT-NEXT: jmp .LBB0_1
|
||
|
|
||
|
for.end: ; preds = %for.cond
|
||
|
%8 = load i32, i32* %ret_val, align 4
|
||
|
ret i32 %8
|
||
|
}
|
||
|
|
||
|
; Function Attrs: nounwind
|
||
|
declare void @llvm.x86.sse2.lfence() #1
|
||
|
|
||
|
attributes #0 = { "target-features"="+lvi-load-hardening" }
|
||
|
attributes #1 = { nounwind }
|