; RUN: llc -mtriple=aarch64-linux-gnu -mattr=+sve2 < %s 2>%t | FileCheck %s ; RUN: FileCheck --check-prefix=WARN --allow-empty %s <%t ; If this check fails please read test/CodeGen/AArch64/README for instructions on how to resolve it. ; WARN-NOT: warning ; ; EOR3 (vector, bitwise, unpredicated) ; define @eor3_i8( %a, %b, %c) { ; CHECK-LABEL: eor3_i8 ; CHECK: eor3 z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.eor3.nxv16i8( %a, %b, %c) ret %res } define @eor3_i16( %a, %b, %c) { ; CHECK-LABEL: eor3_i16 ; CHECK: eor3 z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.eor3.nxv8i16( %a, %b, %c) ret %res } define @eor3_i32( %a, %b, %c) { ; CHECK-LABEL: eor3_i32 ; CHECK: eor3 z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.eor3.nxv4i32( %a, %b, %c) ret %res } define @eor3_i64( %a, %b, %c) { ; CHECK-LABEL: eor3_i64 ; CHECK: eor3 z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.eor3.nxv2i64( %a, %b, %c) ret %res } ; ; BCAX (vector, bitwise, unpredicated) ; define @bcax_i8( %a, %b, %c) { ; CHECK-LABEL: bcax_i8 ; CHECK: bcax z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.bcax.nxv16i8( %a, %b, %c) ret %res } define @bcax_i16( %a, %b, %c) { ; CHECK-LABEL: bcax_i16 ; CHECK: bcax z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.bcax.nxv8i16( %a, %b, %c) ret %res } define @bcax_i32( %a, %b, %c) { ; CHECK-LABEL: bcax_i32 ; CHECK: bcax z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.bcax.nxv4i32( %a, %b, %c) ret %res } define @bcax_i64( %a, %b, %c) { ; CHECK-LABEL: bcax_i64 ; CHECK: bcax z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.bcax.nxv2i64( %a, %b, %c) ret %res } ; ; BSL (vector, bitwise, unpredicated) ; define @bsl_i8( %a, %b, %c) { ; CHECK-LABEL: bsl_i8 ; CHECK: bsl z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.bsl.nxv16i8( %a, %b, %c) ret %res } define @bsl_i16( %a, %b, %c) { ; CHECK-LABEL: bsl_i16 ; CHECK: bsl z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.bsl.nxv8i16( %a, %b, %c) ret %res } define @bsl_i32( %a, %b, %c) { ; CHECK-LABEL: bsl_i32 ; CHECK: bsl z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.bsl.nxv4i32( %a, %b, %c) ret %res } define @bsl_i64( %a, %b, %c) { ; CHECK-LABEL: bsl_i64 ; CHECK: bsl z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.bsl.nxv2i64( %a, %b, %c) ret %res } ; ; BSL1N (vector, bitwise, unpredicated) ; define @bsl1n_i8( %a, %b, %c) { ; CHECK-LABEL: bsl1n_i8 ; CHECK: bsl1n z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.bsl1n.nxv16i8( %a, %b, %c) ret %res } define @bsl1n_i16( %a, %b, %c) { ; CHECK-LABEL: bsl1n_i16 ; CHECK: bsl1n z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.bsl1n.nxv8i16( %a, %b, %c) ret %res } define @bsl1n_i32( %a, %b, %c) { ; CHECK-LABEL: bsl1n_i32 ; CHECK: bsl1n z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.bsl1n.nxv4i32( %a, %b, %c) ret %res } define @bsl1n_i64( %a, %b, %c) { ; CHECK-LABEL: bsl1n_i64 ; CHECK: bsl1n z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.bsl1n.nxv2i64( %a, %b, %c) ret %res } ; ; BSL2N (vector, bitwise, unpredicated) ; define @bsl2n_i8( %a, %b, %c) { ; CHECK-LABEL: bsl2n_i8 ; CHECK: bsl2n z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.bsl2n.nxv16i8( %a, %b, %c) ret %res } define @bsl2n_i16( %a, %b, %c) { ; CHECK-LABEL: bsl2n_i16 ; CHECK: bsl2n z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.bsl2n.nxv8i16( %a, %b, %c) ret %res } define @bsl2n_i32( %a, %b, %c) { ; CHECK-LABEL: bsl2n_i32 ; CHECK: bsl2n z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.bsl2n.nxv4i32( %a, %b, %c) ret %res } define @bsl2n_i64( %a, %b, %c) { ; CHECK-LABEL: bsl2n_i64 ; CHECK: bsl2n z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.bsl2n.nxv2i64( %a, %b, %c) ret %res } ; ; NBSL (vector, bitwise, unpredicated) ; define @nbsl_i8( %a, %b, %c) { ; CHECK-LABEL: nbsl_i8 ; CHECK: nbsl z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.nbsl.nxv16i8( %a, %b, %c) ret %res } define @nbsl_i16( %a, %b, %c) { ; CHECK-LABEL: nbsl_i16 ; CHECK: nbsl z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.nbsl.nxv8i16( %a, %b, %c) ret %res } define @nbsl_i32( %a, %b, %c) { ; CHECK-LABEL: nbsl_i32 ; CHECK: nbsl z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.nbsl.nxv4i32( %a, %b, %c) ret %res } define @nbsl_i64( %a, %b, %c) { ; CHECK-LABEL: nbsl_i64 ; CHECK: nbsl z0.d, z0.d, z1.d, z2.d ; CHECK-NEXT: ret %res = call @llvm.aarch64.sve.nbsl.nxv2i64( %a, %b, %c) ret %res } ; ; XAR (vector, bitwise, unpredicated) ; define @xar_b( %a, %b) { ; CHECK-LABEL: xar_b: ; CHECK: xar z0.b, z0.b, z1.b, #1 ; CHECK-NEXT: ret %out = call @llvm.aarch64.sve.xar.nxv16i8( %a, %b, i32 1) ret %out } define @xar_h( %a, %b) { ; CHECK-LABEL: xar_h: ; CHECK: xar z0.h, z0.h, z1.h, #2 ; CHECK-NEXT: ret %out = call @llvm.aarch64.sve.xar.nxv8i16( %a, %b, i32 2) ret %out } define @xar_s( %a, %b) { ; CHECK-LABEL: xar_s: ; CHECK: xar z0.s, z0.s, z1.s, #3 ; CHECK-NEXT: ret %out = call @llvm.aarch64.sve.xar.nxv4i32( %a, %b, i32 3) ret %out } define @xar_d( %a, %b) { ; CHECK-LABEL: xar_d: ; CHECK: xar z0.d, z0.d, z1.d, #4 ; CHECK-NEXT: ret %out = call @llvm.aarch64.sve.xar.nxv2i64( %a, %b, i32 4) ret %out } declare @llvm.aarch64.sve.eor3.nxv16i8(,,) declare @llvm.aarch64.sve.eor3.nxv8i16(,,) declare @llvm.aarch64.sve.eor3.nxv4i32(,,) declare @llvm.aarch64.sve.eor3.nxv2i64(,,) declare @llvm.aarch64.sve.bcax.nxv16i8(,,) declare @llvm.aarch64.sve.bcax.nxv8i16(,,) declare @llvm.aarch64.sve.bcax.nxv4i32(,,) declare @llvm.aarch64.sve.bcax.nxv2i64(,,) declare @llvm.aarch64.sve.bsl.nxv16i8(,,) declare @llvm.aarch64.sve.bsl.nxv8i16(,,) declare @llvm.aarch64.sve.bsl.nxv4i32(,,) declare @llvm.aarch64.sve.bsl.nxv2i64(,,) declare @llvm.aarch64.sve.bsl1n.nxv16i8(,,) declare @llvm.aarch64.sve.bsl1n.nxv8i16(,,) declare @llvm.aarch64.sve.bsl1n.nxv4i32(,,) declare @llvm.aarch64.sve.bsl1n.nxv2i64(,,) declare @llvm.aarch64.sve.bsl2n.nxv16i8(,,) declare @llvm.aarch64.sve.bsl2n.nxv8i16(,,) declare @llvm.aarch64.sve.bsl2n.nxv4i32(,,) declare @llvm.aarch64.sve.bsl2n.nxv2i64(,,) declare @llvm.aarch64.sve.nbsl.nxv16i8(,,) declare @llvm.aarch64.sve.nbsl.nxv8i16(,,) declare @llvm.aarch64.sve.nbsl.nxv4i32(,,) declare @llvm.aarch64.sve.nbsl.nxv2i64(,,) declare @llvm.aarch64.sve.xar.nxv16i8(, , i32) declare @llvm.aarch64.sve.xar.nxv8i16(, , i32) declare @llvm.aarch64.sve.xar.nxv4i32(, , i32) declare @llvm.aarch64.sve.xar.nxv2i64(, , i32)