| ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py |
| ; RUN: llc -mtriple=riscv32 -target-abi=ilp32d -mattr=+v,+zvfh,+f,+d -verify-machineinstrs < %s | FileCheck %s -check-prefixes=CHECK,RV32 |
| ; RUN: llc -mtriple=riscv64 -target-abi=lp64d -mattr=+v,+zvfh,+f,+d -verify-machineinstrs < %s | FileCheck %s -check-prefixes=CHECK,RV64 |
| |
| define void @vselect_vv_v6i32(ptr %a, ptr %b, ptr %cc, ptr %z) { |
| ; RV32-LABEL: vselect_vv_v6i32: |
| ; RV32: # %bb.0: |
| ; RV32-NEXT: lbu a2, 0(a2) |
| ; RV32-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV32-NEXT: vle32.v v8, (a1) |
| ; RV32-NEXT: slli a1, a2, 30 |
| ; RV32-NEXT: andi a4, a2, 1 |
| ; RV32-NEXT: vsetivli zero, 8, e8, mf2, ta, ma |
| ; RV32-NEXT: vmv.v.x v10, a4 |
| ; RV32-NEXT: slli a4, a2, 29 |
| ; RV32-NEXT: srli a1, a1, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV32-NEXT: slli a1, a2, 28 |
| ; RV32-NEXT: srli a4, a4, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a4 |
| ; RV32-NEXT: slli a4, a2, 27 |
| ; RV32-NEXT: srli a2, a2, 5 |
| ; RV32-NEXT: srli a1, a1, 31 |
| ; RV32-NEXT: srli a4, a4, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a4 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a2 |
| ; RV32-NEXT: vslidedown.vi v10, v10, 2 |
| ; RV32-NEXT: vand.vi v10, v10, 1 |
| ; RV32-NEXT: vmsne.vi v0, v10, 0 |
| ; RV32-NEXT: vsetivli zero, 6, e32, m2, tu, mu |
| ; RV32-NEXT: vle32.v v8, (a0), v0.t |
| ; RV32-NEXT: vse32.v v8, (a3) |
| ; RV32-NEXT: ret |
| ; |
| ; RV64-LABEL: vselect_vv_v6i32: |
| ; RV64: # %bb.0: |
| ; RV64-NEXT: lbu a2, 0(a2) |
| ; RV64-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV64-NEXT: vle32.v v8, (a1) |
| ; RV64-NEXT: slli a1, a2, 62 |
| ; RV64-NEXT: andi a4, a2, 1 |
| ; RV64-NEXT: vsetivli zero, 8, e8, mf2, ta, ma |
| ; RV64-NEXT: vmv.v.x v10, a4 |
| ; RV64-NEXT: slli a4, a2, 61 |
| ; RV64-NEXT: srli a1, a1, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV64-NEXT: slli a1, a2, 60 |
| ; RV64-NEXT: srli a4, a4, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a4 |
| ; RV64-NEXT: slli a4, a2, 59 |
| ; RV64-NEXT: srli a2, a2, 5 |
| ; RV64-NEXT: srli a1, a1, 63 |
| ; RV64-NEXT: srli a4, a4, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a4 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a2 |
| ; RV64-NEXT: vslidedown.vi v10, v10, 2 |
| ; RV64-NEXT: vand.vi v10, v10, 1 |
| ; RV64-NEXT: vmsne.vi v0, v10, 0 |
| ; RV64-NEXT: vsetivli zero, 6, e32, m2, tu, mu |
| ; RV64-NEXT: vle32.v v8, (a0), v0.t |
| ; RV64-NEXT: vse32.v v8, (a3) |
| ; RV64-NEXT: ret |
| %va = load <6 x i32>, ptr %a |
| %vb = load <6 x i32>, ptr %b |
| %vcc = load <6 x i1>, ptr %cc |
| %vsel = select <6 x i1> %vcc, <6 x i32> %va, <6 x i32> %vb |
| store <6 x i32> %vsel, ptr %z |
| ret void |
| } |
| |
| define void @vselect_vx_v6i32(i32 %a, ptr %b, ptr %cc, ptr %z) { |
| ; RV32-LABEL: vselect_vx_v6i32: |
| ; RV32: # %bb.0: |
| ; RV32-NEXT: lbu a2, 0(a2) |
| ; RV32-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV32-NEXT: vle32.v v8, (a1) |
| ; RV32-NEXT: slli a1, a2, 30 |
| ; RV32-NEXT: andi a4, a2, 1 |
| ; RV32-NEXT: vsetivli zero, 8, e8, mf2, ta, ma |
| ; RV32-NEXT: vmv.v.x v10, a4 |
| ; RV32-NEXT: slli a4, a2, 29 |
| ; RV32-NEXT: srli a1, a1, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV32-NEXT: slli a1, a2, 28 |
| ; RV32-NEXT: srli a4, a4, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a4 |
| ; RV32-NEXT: slli a4, a2, 27 |
| ; RV32-NEXT: srli a2, a2, 5 |
| ; RV32-NEXT: srli a1, a1, 31 |
| ; RV32-NEXT: srli a4, a4, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a4 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a2 |
| ; RV32-NEXT: vslidedown.vi v10, v10, 2 |
| ; RV32-NEXT: vand.vi v10, v10, 1 |
| ; RV32-NEXT: vmsne.vi v0, v10, 0 |
| ; RV32-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV32-NEXT: vmerge.vxm v8, v8, a0, v0 |
| ; RV32-NEXT: vse32.v v8, (a3) |
| ; RV32-NEXT: ret |
| ; |
| ; RV64-LABEL: vselect_vx_v6i32: |
| ; RV64: # %bb.0: |
| ; RV64-NEXT: lbu a2, 0(a2) |
| ; RV64-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV64-NEXT: vle32.v v8, (a1) |
| ; RV64-NEXT: slli a1, a2, 62 |
| ; RV64-NEXT: andi a4, a2, 1 |
| ; RV64-NEXT: vsetivli zero, 8, e8, mf2, ta, ma |
| ; RV64-NEXT: vmv.v.x v10, a4 |
| ; RV64-NEXT: slli a4, a2, 61 |
| ; RV64-NEXT: srli a1, a1, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV64-NEXT: slli a1, a2, 60 |
| ; RV64-NEXT: srli a4, a4, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a4 |
| ; RV64-NEXT: slli a4, a2, 59 |
| ; RV64-NEXT: srli a2, a2, 5 |
| ; RV64-NEXT: srli a1, a1, 63 |
| ; RV64-NEXT: srli a4, a4, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a4 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a2 |
| ; RV64-NEXT: vslidedown.vi v10, v10, 2 |
| ; RV64-NEXT: vand.vi v10, v10, 1 |
| ; RV64-NEXT: vmsne.vi v0, v10, 0 |
| ; RV64-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV64-NEXT: vmerge.vxm v8, v8, a0, v0 |
| ; RV64-NEXT: vse32.v v8, (a3) |
| ; RV64-NEXT: ret |
| %vb = load <6 x i32>, ptr %b |
| %ahead = insertelement <6 x i32> poison, i32 %a, i32 0 |
| %va = shufflevector <6 x i32> %ahead, <6 x i32> poison, <6 x i32> zeroinitializer |
| %vcc = load <6 x i1>, ptr %cc |
| %vsel = select <6 x i1> %vcc, <6 x i32> %va, <6 x i32> %vb |
| store <6 x i32> %vsel, ptr %z |
| ret void |
| } |
| |
| define void @vselect_vi_v6i32(ptr %b, ptr %cc, ptr %z) { |
| ; RV32-LABEL: vselect_vi_v6i32: |
| ; RV32: # %bb.0: |
| ; RV32-NEXT: lbu a1, 0(a1) |
| ; RV32-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV32-NEXT: vle32.v v8, (a0) |
| ; RV32-NEXT: slli a0, a1, 30 |
| ; RV32-NEXT: andi a3, a1, 1 |
| ; RV32-NEXT: vsetivli zero, 8, e8, mf2, ta, ma |
| ; RV32-NEXT: vmv.v.x v10, a3 |
| ; RV32-NEXT: slli a3, a1, 29 |
| ; RV32-NEXT: srli a0, a0, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a0 |
| ; RV32-NEXT: slli a0, a1, 28 |
| ; RV32-NEXT: srli a3, a3, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a3 |
| ; RV32-NEXT: slli a3, a1, 27 |
| ; RV32-NEXT: srli a1, a1, 5 |
| ; RV32-NEXT: srli a0, a0, 31 |
| ; RV32-NEXT: srli a3, a3, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a0 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a3 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV32-NEXT: vslidedown.vi v10, v10, 2 |
| ; RV32-NEXT: vand.vi v10, v10, 1 |
| ; RV32-NEXT: vmsne.vi v0, v10, 0 |
| ; RV32-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV32-NEXT: vmerge.vim v8, v8, -1, v0 |
| ; RV32-NEXT: vse32.v v8, (a2) |
| ; RV32-NEXT: ret |
| ; |
| ; RV64-LABEL: vselect_vi_v6i32: |
| ; RV64: # %bb.0: |
| ; RV64-NEXT: lbu a1, 0(a1) |
| ; RV64-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV64-NEXT: vle32.v v8, (a0) |
| ; RV64-NEXT: slli a0, a1, 62 |
| ; RV64-NEXT: andi a3, a1, 1 |
| ; RV64-NEXT: vsetivli zero, 8, e8, mf2, ta, ma |
| ; RV64-NEXT: vmv.v.x v10, a3 |
| ; RV64-NEXT: slli a3, a1, 61 |
| ; RV64-NEXT: srli a0, a0, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a0 |
| ; RV64-NEXT: slli a0, a1, 60 |
| ; RV64-NEXT: srli a3, a3, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a3 |
| ; RV64-NEXT: slli a3, a1, 59 |
| ; RV64-NEXT: srli a1, a1, 5 |
| ; RV64-NEXT: srli a0, a0, 63 |
| ; RV64-NEXT: srli a3, a3, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a0 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a3 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV64-NEXT: vslidedown.vi v10, v10, 2 |
| ; RV64-NEXT: vand.vi v10, v10, 1 |
| ; RV64-NEXT: vmsne.vi v0, v10, 0 |
| ; RV64-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV64-NEXT: vmerge.vim v8, v8, -1, v0 |
| ; RV64-NEXT: vse32.v v8, (a2) |
| ; RV64-NEXT: ret |
| %vb = load <6 x i32>, ptr %b |
| %vcc = load <6 x i1>, ptr %cc |
| %vsel = select <6 x i1> %vcc, <6 x i32> splat (i32 -1), <6 x i32> %vb |
| store <6 x i32> %vsel, ptr %z |
| ret void |
| } |
| |
| |
| define void @vselect_vv_v6f32(ptr %a, ptr %b, ptr %cc, ptr %z) { |
| ; RV32-LABEL: vselect_vv_v6f32: |
| ; RV32: # %bb.0: |
| ; RV32-NEXT: lbu a2, 0(a2) |
| ; RV32-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV32-NEXT: vle32.v v8, (a1) |
| ; RV32-NEXT: slli a1, a2, 30 |
| ; RV32-NEXT: andi a4, a2, 1 |
| ; RV32-NEXT: vsetivli zero, 8, e8, mf2, ta, ma |
| ; RV32-NEXT: vmv.v.x v10, a4 |
| ; RV32-NEXT: slli a4, a2, 29 |
| ; RV32-NEXT: srli a1, a1, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV32-NEXT: slli a1, a2, 28 |
| ; RV32-NEXT: srli a4, a4, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a4 |
| ; RV32-NEXT: slli a4, a2, 27 |
| ; RV32-NEXT: srli a2, a2, 5 |
| ; RV32-NEXT: srli a1, a1, 31 |
| ; RV32-NEXT: srli a4, a4, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a4 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a2 |
| ; RV32-NEXT: vslidedown.vi v10, v10, 2 |
| ; RV32-NEXT: vand.vi v10, v10, 1 |
| ; RV32-NEXT: vmsne.vi v0, v10, 0 |
| ; RV32-NEXT: vsetivli zero, 6, e32, m2, tu, mu |
| ; RV32-NEXT: vle32.v v8, (a0), v0.t |
| ; RV32-NEXT: vse32.v v8, (a3) |
| ; RV32-NEXT: ret |
| ; |
| ; RV64-LABEL: vselect_vv_v6f32: |
| ; RV64: # %bb.0: |
| ; RV64-NEXT: lbu a2, 0(a2) |
| ; RV64-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV64-NEXT: vle32.v v8, (a1) |
| ; RV64-NEXT: slli a1, a2, 62 |
| ; RV64-NEXT: andi a4, a2, 1 |
| ; RV64-NEXT: vsetivli zero, 8, e8, mf2, ta, ma |
| ; RV64-NEXT: vmv.v.x v10, a4 |
| ; RV64-NEXT: slli a4, a2, 61 |
| ; RV64-NEXT: srli a1, a1, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV64-NEXT: slli a1, a2, 60 |
| ; RV64-NEXT: srli a4, a4, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a4 |
| ; RV64-NEXT: slli a4, a2, 59 |
| ; RV64-NEXT: srli a2, a2, 5 |
| ; RV64-NEXT: srli a1, a1, 63 |
| ; RV64-NEXT: srli a4, a4, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a4 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a2 |
| ; RV64-NEXT: vslidedown.vi v10, v10, 2 |
| ; RV64-NEXT: vand.vi v10, v10, 1 |
| ; RV64-NEXT: vmsne.vi v0, v10, 0 |
| ; RV64-NEXT: vsetivli zero, 6, e32, m2, tu, mu |
| ; RV64-NEXT: vle32.v v8, (a0), v0.t |
| ; RV64-NEXT: vse32.v v8, (a3) |
| ; RV64-NEXT: ret |
| %va = load <6 x float>, ptr %a |
| %vb = load <6 x float>, ptr %b |
| %vcc = load <6 x i1>, ptr %cc |
| %vsel = select <6 x i1> %vcc, <6 x float> %va, <6 x float> %vb |
| store <6 x float> %vsel, ptr %z |
| ret void |
| } |
| |
| define void @vselect_vx_v6f32(float %a, ptr %b, ptr %cc, ptr %z) { |
| ; RV32-LABEL: vselect_vx_v6f32: |
| ; RV32: # %bb.0: |
| ; RV32-NEXT: lbu a1, 0(a1) |
| ; RV32-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV32-NEXT: vle32.v v8, (a0) |
| ; RV32-NEXT: slli a0, a1, 30 |
| ; RV32-NEXT: andi a3, a1, 1 |
| ; RV32-NEXT: vsetivli zero, 8, e8, mf2, ta, ma |
| ; RV32-NEXT: vmv.v.x v10, a3 |
| ; RV32-NEXT: slli a3, a1, 29 |
| ; RV32-NEXT: srli a0, a0, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a0 |
| ; RV32-NEXT: slli a0, a1, 28 |
| ; RV32-NEXT: srli a3, a3, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a3 |
| ; RV32-NEXT: slli a3, a1, 27 |
| ; RV32-NEXT: srli a1, a1, 5 |
| ; RV32-NEXT: srli a0, a0, 31 |
| ; RV32-NEXT: srli a3, a3, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a0 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a3 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV32-NEXT: vslidedown.vi v10, v10, 2 |
| ; RV32-NEXT: vand.vi v10, v10, 1 |
| ; RV32-NEXT: vmsne.vi v0, v10, 0 |
| ; RV32-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV32-NEXT: vfmerge.vfm v8, v8, fa0, v0 |
| ; RV32-NEXT: vse32.v v8, (a2) |
| ; RV32-NEXT: ret |
| ; |
| ; RV64-LABEL: vselect_vx_v6f32: |
| ; RV64: # %bb.0: |
| ; RV64-NEXT: lbu a1, 0(a1) |
| ; RV64-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV64-NEXT: vle32.v v8, (a0) |
| ; RV64-NEXT: slli a0, a1, 62 |
| ; RV64-NEXT: andi a3, a1, 1 |
| ; RV64-NEXT: vsetivli zero, 8, e8, mf2, ta, ma |
| ; RV64-NEXT: vmv.v.x v10, a3 |
| ; RV64-NEXT: slli a3, a1, 61 |
| ; RV64-NEXT: srli a0, a0, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a0 |
| ; RV64-NEXT: slli a0, a1, 60 |
| ; RV64-NEXT: srli a3, a3, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a3 |
| ; RV64-NEXT: slli a3, a1, 59 |
| ; RV64-NEXT: srli a1, a1, 5 |
| ; RV64-NEXT: srli a0, a0, 63 |
| ; RV64-NEXT: srli a3, a3, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a0 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a3 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV64-NEXT: vslidedown.vi v10, v10, 2 |
| ; RV64-NEXT: vand.vi v10, v10, 1 |
| ; RV64-NEXT: vmsne.vi v0, v10, 0 |
| ; RV64-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV64-NEXT: vfmerge.vfm v8, v8, fa0, v0 |
| ; RV64-NEXT: vse32.v v8, (a2) |
| ; RV64-NEXT: ret |
| %vb = load <6 x float>, ptr %b |
| %ahead = insertelement <6 x float> poison, float %a, i32 0 |
| %va = shufflevector <6 x float> %ahead, <6 x float> poison, <6 x i32> zeroinitializer |
| %vcc = load <6 x i1>, ptr %cc |
| %vsel = select <6 x i1> %vcc, <6 x float> %va, <6 x float> %vb |
| store <6 x float> %vsel, ptr %z |
| ret void |
| } |
| |
| define void @vselect_vfpzero_v6f32(ptr %b, ptr %cc, ptr %z) { |
| ; RV32-LABEL: vselect_vfpzero_v6f32: |
| ; RV32: # %bb.0: |
| ; RV32-NEXT: lbu a1, 0(a1) |
| ; RV32-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV32-NEXT: vle32.v v8, (a0) |
| ; RV32-NEXT: slli a0, a1, 30 |
| ; RV32-NEXT: andi a3, a1, 1 |
| ; RV32-NEXT: vsetivli zero, 8, e8, mf2, ta, ma |
| ; RV32-NEXT: vmv.v.x v10, a3 |
| ; RV32-NEXT: slli a3, a1, 29 |
| ; RV32-NEXT: srli a0, a0, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a0 |
| ; RV32-NEXT: slli a0, a1, 28 |
| ; RV32-NEXT: srli a3, a3, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a3 |
| ; RV32-NEXT: slli a3, a1, 27 |
| ; RV32-NEXT: srli a1, a1, 5 |
| ; RV32-NEXT: srli a0, a0, 31 |
| ; RV32-NEXT: srli a3, a3, 31 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a0 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a3 |
| ; RV32-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV32-NEXT: vslidedown.vi v10, v10, 2 |
| ; RV32-NEXT: vand.vi v10, v10, 1 |
| ; RV32-NEXT: vmsne.vi v0, v10, 0 |
| ; RV32-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV32-NEXT: vmerge.vim v8, v8, 0, v0 |
| ; RV32-NEXT: vse32.v v8, (a2) |
| ; RV32-NEXT: ret |
| ; |
| ; RV64-LABEL: vselect_vfpzero_v6f32: |
| ; RV64: # %bb.0: |
| ; RV64-NEXT: lbu a1, 0(a1) |
| ; RV64-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV64-NEXT: vle32.v v8, (a0) |
| ; RV64-NEXT: slli a0, a1, 62 |
| ; RV64-NEXT: andi a3, a1, 1 |
| ; RV64-NEXT: vsetivli zero, 8, e8, mf2, ta, ma |
| ; RV64-NEXT: vmv.v.x v10, a3 |
| ; RV64-NEXT: slli a3, a1, 61 |
| ; RV64-NEXT: srli a0, a0, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a0 |
| ; RV64-NEXT: slli a0, a1, 60 |
| ; RV64-NEXT: srli a3, a3, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a3 |
| ; RV64-NEXT: slli a3, a1, 59 |
| ; RV64-NEXT: srli a1, a1, 5 |
| ; RV64-NEXT: srli a0, a0, 63 |
| ; RV64-NEXT: srli a3, a3, 63 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a0 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a3 |
| ; RV64-NEXT: vslide1down.vx v10, v10, a1 |
| ; RV64-NEXT: vslidedown.vi v10, v10, 2 |
| ; RV64-NEXT: vand.vi v10, v10, 1 |
| ; RV64-NEXT: vmsne.vi v0, v10, 0 |
| ; RV64-NEXT: vsetivli zero, 6, e32, m2, ta, ma |
| ; RV64-NEXT: vmerge.vim v8, v8, 0, v0 |
| ; RV64-NEXT: vse32.v v8, (a2) |
| ; RV64-NEXT: ret |
| %vb = load <6 x float>, ptr %b |
| %vcc = load <6 x i1>, ptr %cc |
| %vsel = select <6 x i1> %vcc, <6 x float> splat (float 0.0), <6 x float> %vb |
| store <6 x float> %vsel, ptr %z |
| ret void |
| } |
| |
| define void @vselect_vv_v8i32(ptr %a, ptr %b, ptr %cc, ptr %z) { |
| ; CHECK-LABEL: vselect_vv_v8i32: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vsetivli zero, 8, e32, m2, ta, mu |
| ; CHECK-NEXT: vlm.v v0, (a2) |
| ; CHECK-NEXT: vle32.v v8, (a1) |
| ; CHECK-NEXT: vle32.v v8, (a0), v0.t |
| ; CHECK-NEXT: vse32.v v8, (a3) |
| ; CHECK-NEXT: ret |
| %va = load <8 x i32>, ptr %a |
| %vb = load <8 x i32>, ptr %b |
| %vcc = load <8 x i1>, ptr %cc |
| %vsel = select <8 x i1> %vcc, <8 x i32> %va, <8 x i32> %vb |
| store <8 x i32> %vsel, ptr %z |
| ret void |
| } |
| |
| define void @vselect_vx_v8i32(i32 %a, ptr %b, ptr %cc, ptr %z) { |
| ; CHECK-LABEL: vselect_vx_v8i32: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vsetivli zero, 8, e32, m2, ta, ma |
| ; CHECK-NEXT: vlm.v v0, (a2) |
| ; CHECK-NEXT: vle32.v v8, (a1) |
| ; CHECK-NEXT: vmerge.vxm v8, v8, a0, v0 |
| ; CHECK-NEXT: vse32.v v8, (a3) |
| ; CHECK-NEXT: ret |
| %vb = load <8 x i32>, ptr %b |
| %ahead = insertelement <8 x i32> poison, i32 %a, i32 0 |
| %va = shufflevector <8 x i32> %ahead, <8 x i32> poison, <8 x i32> zeroinitializer |
| %vcc = load <8 x i1>, ptr %cc |
| %vsel = select <8 x i1> %vcc, <8 x i32> %va, <8 x i32> %vb |
| store <8 x i32> %vsel, ptr %z |
| ret void |
| } |
| |
| define void @vselect_vi_v8i32(ptr %b, ptr %cc, ptr %z) { |
| ; CHECK-LABEL: vselect_vi_v8i32: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vsetivli zero, 8, e32, m2, ta, ma |
| ; CHECK-NEXT: vlm.v v0, (a1) |
| ; CHECK-NEXT: vle32.v v8, (a0) |
| ; CHECK-NEXT: vmerge.vim v8, v8, -1, v0 |
| ; CHECK-NEXT: vse32.v v8, (a2) |
| ; CHECK-NEXT: ret |
| %vb = load <8 x i32>, ptr %b |
| %vcc = load <8 x i1>, ptr %cc |
| %vsel = select <8 x i1> %vcc, <8 x i32> splat (i32 -1), <8 x i32> %vb |
| store <8 x i32> %vsel, ptr %z |
| ret void |
| } |
| |
| define void @vselect_vv_v8f32(ptr %a, ptr %b, ptr %cc, ptr %z) { |
| ; CHECK-LABEL: vselect_vv_v8f32: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vsetivli zero, 8, e32, m2, ta, mu |
| ; CHECK-NEXT: vlm.v v0, (a2) |
| ; CHECK-NEXT: vle32.v v8, (a1) |
| ; CHECK-NEXT: vle32.v v8, (a0), v0.t |
| ; CHECK-NEXT: vse32.v v8, (a3) |
| ; CHECK-NEXT: ret |
| %va = load <8 x float>, ptr %a |
| %vb = load <8 x float>, ptr %b |
| %vcc = load <8 x i1>, ptr %cc |
| %vsel = select <8 x i1> %vcc, <8 x float> %va, <8 x float> %vb |
| store <8 x float> %vsel, ptr %z |
| ret void |
| } |
| |
| define void @vselect_vx_v8f32(float %a, ptr %b, ptr %cc, ptr %z) { |
| ; CHECK-LABEL: vselect_vx_v8f32: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vsetivli zero, 8, e32, m2, ta, ma |
| ; CHECK-NEXT: vlm.v v0, (a1) |
| ; CHECK-NEXT: vle32.v v8, (a0) |
| ; CHECK-NEXT: vfmerge.vfm v8, v8, fa0, v0 |
| ; CHECK-NEXT: vse32.v v8, (a2) |
| ; CHECK-NEXT: ret |
| %vb = load <8 x float>, ptr %b |
| %ahead = insertelement <8 x float> poison, float %a, i32 0 |
| %va = shufflevector <8 x float> %ahead, <8 x float> poison, <8 x i32> zeroinitializer |
| %vcc = load <8 x i1>, ptr %cc |
| %vsel = select <8 x i1> %vcc, <8 x float> %va, <8 x float> %vb |
| store <8 x float> %vsel, ptr %z |
| ret void |
| } |
| |
| define void @vselect_vfpzero_v8f32(ptr %b, ptr %cc, ptr %z) { |
| ; CHECK-LABEL: vselect_vfpzero_v8f32: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vsetivli zero, 8, e32, m2, ta, ma |
| ; CHECK-NEXT: vlm.v v0, (a1) |
| ; CHECK-NEXT: vle32.v v8, (a0) |
| ; CHECK-NEXT: vmerge.vim v8, v8, 0, v0 |
| ; CHECK-NEXT: vse32.v v8, (a2) |
| ; CHECK-NEXT: ret |
| %vb = load <8 x float>, ptr %b |
| %vcc = load <8 x i1>, ptr %cc |
| %vsel = select <8 x i1> %vcc, <8 x float> splat (float 0.0), <8 x float> %vb |
| store <8 x float> %vsel, ptr %z |
| ret void |
| } |
| |
| define void @vselect_vv_v16i16(ptr %a, ptr %b, ptr %cc, ptr %z) { |
| ; CHECK-LABEL: vselect_vv_v16i16: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vsetivli zero, 16, e16, m2, ta, mu |
| ; CHECK-NEXT: vlm.v v0, (a2) |
| ; CHECK-NEXT: vle16.v v8, (a1) |
| ; CHECK-NEXT: vle16.v v8, (a0), v0.t |
| ; CHECK-NEXT: vse16.v v8, (a3) |
| ; CHECK-NEXT: ret |
| %va = load <16 x i16>, ptr %a |
| %vb = load <16 x i16>, ptr %b |
| %vcc = load <16 x i1>, ptr %cc |
| %vsel = select <16 x i1> %vcc, <16 x i16> %va, <16 x i16> %vb |
| store <16 x i16> %vsel, ptr %z |
| ret void |
| } |
| |
| define void @vselect_vx_v16i16(i16 signext %a, ptr %b, ptr %cc, ptr %z) { |
| ; CHECK-LABEL: vselect_vx_v16i16: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vsetivli zero, 16, e16, m2, ta, ma |
| ; CHECK-NEXT: vlm.v v0, (a2) |
| ; CHECK-NEXT: vle16.v v8, (a1) |
| ; CHECK-NEXT: vmerge.vxm v8, v8, a0, v0 |
| ; CHECK-NEXT: vse16.v v8, (a3) |
| ; CHECK-NEXT: ret |
| %vb = load <16 x i16>, ptr %b |
| %ahead = insertelement <16 x i16> poison, i16 %a, i32 0 |
| %va = shufflevector <16 x i16> %ahead, <16 x i16> poison, <16 x i32> zeroinitializer |
| %vcc = load <16 x i1>, ptr %cc |
| %vsel = select <16 x i1> %vcc, <16 x i16> %va, <16 x i16> %vb |
| store <16 x i16> %vsel, ptr %z |
| ret void |
| } |
| |
| define void @vselect_vi_v16i16(ptr %b, ptr %cc, ptr %z) { |
| ; CHECK-LABEL: vselect_vi_v16i16: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vsetivli zero, 16, e16, m2, ta, ma |
| ; CHECK-NEXT: vlm.v v0, (a1) |
| ; CHECK-NEXT: vle16.v v8, (a0) |
| ; CHECK-NEXT: vmerge.vim v8, v8, 4, v0 |
| ; CHECK-NEXT: vse16.v v8, (a2) |
| ; CHECK-NEXT: ret |
| %vb = load <16 x i16>, ptr %b |
| %vcc = load <16 x i1>, ptr %cc |
| %vsel = select <16 x i1> %vcc, <16 x i16> splat (i16 4), <16 x i16> %vb |
| store <16 x i16> %vsel, ptr %z |
| ret void |
| } |
| |
| define void @vselect_vv_v32f16(ptr %a, ptr %b, ptr %cc, ptr %z) { |
| ; CHECK-LABEL: vselect_vv_v32f16: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: li a4, 32 |
| ; CHECK-NEXT: vsetvli zero, a4, e16, m4, ta, mu |
| ; CHECK-NEXT: vlm.v v0, (a2) |
| ; CHECK-NEXT: vle16.v v8, (a1) |
| ; CHECK-NEXT: vle16.v v8, (a0), v0.t |
| ; CHECK-NEXT: vse16.v v8, (a3) |
| ; CHECK-NEXT: ret |
| %va = load <32 x half>, ptr %a |
| %vb = load <32 x half>, ptr %b |
| %vcc = load <32 x i1>, ptr %cc |
| %vsel = select <32 x i1> %vcc, <32 x half> %va, <32 x half> %vb |
| store <32 x half> %vsel, ptr %z |
| ret void |
| } |
| |
| define void @vselect_vx_v32f16(half %a, ptr %b, ptr %cc, ptr %z) { |
| ; CHECK-LABEL: vselect_vx_v32f16: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: li a3, 32 |
| ; CHECK-NEXT: vsetvli zero, a3, e16, m4, ta, ma |
| ; CHECK-NEXT: vlm.v v0, (a1) |
| ; CHECK-NEXT: vle16.v v8, (a0) |
| ; CHECK-NEXT: vfmerge.vfm v8, v8, fa0, v0 |
| ; CHECK-NEXT: vse16.v v8, (a2) |
| ; CHECK-NEXT: ret |
| %vb = load <32 x half>, ptr %b |
| %ahead = insertelement <32 x half> poison, half %a, i32 0 |
| %va = shufflevector <32 x half> %ahead, <32 x half> poison, <32 x i32> zeroinitializer |
| %vcc = load <32 x i1>, ptr %cc |
| %vsel = select <32 x i1> %vcc, <32 x half> %va, <32 x half> %vb |
| store <32 x half> %vsel, ptr %z |
| ret void |
| } |
| |
| define void @vselect_vfpzero_v32f16(ptr %b, ptr %cc, ptr %z) { |
| ; CHECK-LABEL: vselect_vfpzero_v32f16: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: li a3, 32 |
| ; CHECK-NEXT: vsetvli zero, a3, e16, m4, ta, ma |
| ; CHECK-NEXT: vlm.v v0, (a1) |
| ; CHECK-NEXT: vle16.v v8, (a0) |
| ; CHECK-NEXT: vmerge.vim v8, v8, 0, v0 |
| ; CHECK-NEXT: vse16.v v8, (a2) |
| ; CHECK-NEXT: ret |
| %vb = load <32 x half>, ptr %b |
| %vcc = load <32 x i1>, ptr %cc |
| %vsel = select <32 x i1> %vcc, <32 x half> splat (half 0.0), <32 x half> %vb |
| store <32 x half> %vsel, ptr %z |
| ret void |
| } |
| |
| define <2 x i1> @vselect_v2i1(<2 x i1> %a, <2 x i1> %b, <2 x i1> %cc) { |
| ; CHECK-LABEL: vselect_v2i1: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vsetivli zero, 2, e8, mf8, ta, ma |
| ; CHECK-NEXT: vmandn.mm v8, v8, v9 |
| ; CHECK-NEXT: vmand.mm v9, v0, v9 |
| ; CHECK-NEXT: vmor.mm v0, v9, v8 |
| ; CHECK-NEXT: ret |
| %v = select <2 x i1> %cc, <2 x i1> %a, <2 x i1> %b |
| ret <2 x i1> %v |
| } |
| |
| define <4 x i1> @vselect_v4i1(<4 x i1> %a, <4 x i1> %b, <4 x i1> %cc) { |
| ; CHECK-LABEL: vselect_v4i1: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vsetivli zero, 4, e8, mf4, ta, ma |
| ; CHECK-NEXT: vmandn.mm v8, v8, v9 |
| ; CHECK-NEXT: vmand.mm v9, v0, v9 |
| ; CHECK-NEXT: vmor.mm v0, v9, v8 |
| ; CHECK-NEXT: ret |
| %v = select <4 x i1> %cc, <4 x i1> %a, <4 x i1> %b |
| ret <4 x i1> %v |
| } |
| |
| define <8 x i1> @vselect_v8i1(<8 x i1> %a, <8 x i1> %b, <8 x i1> %cc) { |
| ; CHECK-LABEL: vselect_v8i1: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vsetivli zero, 8, e8, mf2, ta, ma |
| ; CHECK-NEXT: vmandn.mm v8, v8, v9 |
| ; CHECK-NEXT: vmand.mm v9, v0, v9 |
| ; CHECK-NEXT: vmor.mm v0, v9, v8 |
| ; CHECK-NEXT: ret |
| %v = select <8 x i1> %cc, <8 x i1> %a, <8 x i1> %b |
| ret <8 x i1> %v |
| } |
| |
| define <16 x i1> @vselect_v16i1(<16 x i1> %a, <16 x i1> %b, <16 x i1> %cc) { |
| ; CHECK-LABEL: vselect_v16i1: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: vsetivli zero, 16, e8, m1, ta, ma |
| ; CHECK-NEXT: vmandn.mm v8, v8, v9 |
| ; CHECK-NEXT: vmand.mm v9, v0, v9 |
| ; CHECK-NEXT: vmor.mm v0, v9, v8 |
| ; CHECK-NEXT: ret |
| %v = select <16 x i1> %cc, <16 x i1> %a, <16 x i1> %b |
| ret <16 x i1> %v |
| } |
| |
| define <32 x i1> @vselect_v32i1(<32 x i1> %a, <32 x i1> %b, <32 x i1> %cc) { |
| ; CHECK-LABEL: vselect_v32i1: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: li a0, 32 |
| ; CHECK-NEXT: vsetvli zero, a0, e8, m2, ta, ma |
| ; CHECK-NEXT: vmandn.mm v8, v8, v9 |
| ; CHECK-NEXT: vmand.mm v9, v0, v9 |
| ; CHECK-NEXT: vmor.mm v0, v9, v8 |
| ; CHECK-NEXT: ret |
| %v = select <32 x i1> %cc, <32 x i1> %a, <32 x i1> %b |
| ret <32 x i1> %v |
| } |
| |
| define <64 x i1> @vselect_v64i1(<64 x i1> %a, <64 x i1> %b, <64 x i1> %cc) { |
| ; CHECK-LABEL: vselect_v64i1: |
| ; CHECK: # %bb.0: |
| ; CHECK-NEXT: li a0, 64 |
| ; CHECK-NEXT: vsetvli zero, a0, e8, m4, ta, ma |
| ; CHECK-NEXT: vmandn.mm v8, v8, v9 |
| ; CHECK-NEXT: vmand.mm v9, v0, v9 |
| ; CHECK-NEXT: vmor.mm v0, v9, v8 |
| ; CHECK-NEXT: ret |
| %v = select <64 x i1> %cc, <64 x i1> %a, <64 x i1> %b |
| ret <64 x i1> %v |
| } |