; RUN: opt < %s -loop-vectorize -mtriple=ve-linux -S | FileCheck %s -check-prefix=VE ; RUN: opt < %s -loop-vectorize -mtriple=x86_64-pc_linux -mcpu=core-avx2 -S | FileCheck %s -check-prefix=AVX ; Make sure LV does not trigger for VE on an appealing loop that vectorizes for x86 AVX. ; TODO: Remove this test once VE vector isel is deemed stable. ; VE-NOT: llvm.loop.isvectorized ; AVX: llvm.loop.isvectorized define dso_local void @foo(i32* noalias nocapture %A, i32* noalias nocapture readonly %B, i32 signext %n) local_unnamed_addr { entry: %cmp = icmp sgt i32 %n, 0 br i1 %cmp, label %omp.inner.for.body.preheader, label %simd.if.end omp.inner.for.body.preheader: ; preds = %entry %wide.trip.count = zext i32 %n to i64 br label %omp.inner.for.body omp.inner.for.body: ; preds = %omp.inner.for.body.preheader, %omp.inner.for.body %indvars.iv = phi i64 [ 0, %omp.inner.for.body.preheader ], [ %indvars.iv.next, %omp.inner.for.body ] %arrayidx = getelementptr inbounds i32, i32* %B, i64 %indvars.iv %0 = load i32, i32* %arrayidx, align 4, !llvm.access.group !6 %mul6 = mul nsw i32 %0, 3 %arrayidx8 = getelementptr inbounds i32, i32* %A, i64 %indvars.iv store i32 %mul6, i32* %arrayidx8, align 4, !llvm.access.group !6 %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1 %exitcond.not = icmp eq i64 %indvars.iv.next, %wide.trip.count br i1 %exitcond.not, label %simd.if.end, label %omp.inner.for.body, !llvm.loop !7 simd.if.end: ; preds = %omp.inner.for.body, %entry ret void } !6 = distinct !{} !7 = distinct !{!7, !8, !9} !8 = !{!"llvm.loop.parallel_accesses", !6} !9 = !{!"llvm.loop.vectorize.enable", i1 true}