; RUN: opt -mtriple=aarch64-linux-gnu -mattr=+sve -scalarize-masked-mem-intrin -S < %s 2>%t | FileCheck %s ; RUN: FileCheck --check-prefix=WARN --allow-empty %s <%t ; If this check fails please read test/CodeGen/AArch64/README for instructions on how to resolve it. ; WARN-NOT: warning ; Testing that masked gathers operating on scalable vectors that are ; packed in SVE registers are not scalarized. ; CHECK-LABEL: @masked_gather_nxv4i32( ; CHECK: call @llvm.masked.gather.nxv4i32 define @masked_gather_nxv4i32( %ld, %masks, %passthru) { %res = call @llvm.masked.gather.nxv4i32( %ld, i32 0, %masks, %passthru) ret %res } ; Testing that masked gathers operating on scalable vectors of FP data ; that is packed in SVE registers are not scalarized. ; CHECK-LABEL: @masked_gather_nxv2f64( ; CHECK: call @llvm.masked.gather.nxv2f64 define @masked_gather_nxv2f64( %ld, %masks, %passthru) { %res = call @llvm.masked.gather.nxv2f64( %ld, i32 0, %masks, %passthru) ret %res } ; Testing that masked gathers operating on scalable vectors of FP data ; that is unpacked in SVE registers are not scalarized. ; CHECK-LABEL: @masked_gather_nxv2f16( ; CHECK: call @llvm.masked.gather.nxv2f16 define @masked_gather_nxv2f16( %ld, %masks, %passthru) { %res = call @llvm.masked.gather.nxv2f16( %ld, i32 0, %masks, %passthru) ret %res } ; Testing that masked gathers operating on 64-bit fixed vectors are ; scalarized because NEON doesn't have support for masked gather ; instructions. ; CHECK-LABEL: @masked_gather_v2f32( ; CHECK-NOT: @llvm.masked.gather.v2f32( define <2 x float> @masked_gather_v2f32(<2 x float*> %ld, <2 x i1> %masks, <2 x float> %passthru) { %res = call <2 x float> @llvm.masked.gather.v2f32(<2 x float*> %ld, i32 0, <2 x i1> %masks, <2 x float> %passthru) ret <2 x float> %res } ; Testing that masked gathers operating on 128-bit fixed vectors are ; scalarized because NEON doesn't have support for masked gather ; instructions and because we are not targeting fixed width SVE. ; CHECK-LABEL: @masked_gather_v4i32( ; CHECK-NOT: @llvm.masked.gather.v4i32( define <4 x i32> @masked_gather_v4i32(<4 x i32*> %ld, <4 x i1> %masks, <4 x i32> %passthru) { %res = call <4 x i32> @llvm.masked.gather.v4i32(<4 x i32*> %ld, i32 0, <4 x i1> %masks, <4 x i32> %passthru) ret <4 x i32> %res } declare @llvm.masked.gather.nxv4i32( %ptrs, i32 %align, %masks, %passthru) declare @llvm.masked.gather.nxv2f64( %ptrs, i32 %align, %masks, %passthru) declare @llvm.masked.gather.nxv2f16( %ptrs, i32 %align, %masks, %passthru) declare <2 x float> @llvm.masked.gather.v2f32(<2 x float*> %ptrs, i32 %align, <2 x i1> %masks, <2 x float> %passthru) declare <4 x i32> @llvm.masked.gather.v4i32(<4 x i32*> %ptrs, i32 %align, <4 x i1> %masks, <4 x i32> %passthru)