| ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 5 |
| ; RUN: llc %s -mtriple=aarch64 -o - | FileCheck %s |
| |
| define <4 x i16> @NarrowAShrI32By5(<4 x i32> %x) { |
| ; CHECK-LABEL: NarrowAShrI32By5: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: sqshrn v0.4h, v0.4s, #5 |
| ; CHECK-NEXT: ret |
| %s = ashr <4 x i32> %x, <i32 5, i32 5, i32 5, i32 5> |
| %r = tail call <4 x i16> @llvm.aarch64.neon.sqxtn.v4i16(<4 x i32> %s) |
| ret <4 x i16> %r |
| } |
| |
| define <4 x i16> @NarrowAShrU32By5(<4 x i32> %x) { |
| ; CHECK-LABEL: NarrowAShrU32By5: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: sshr v0.4s, v0.4s, #5 |
| ; CHECK-NEXT: uqxtn v0.4h, v0.4s |
| ; CHECK-NEXT: ret |
| %s = ashr <4 x i32> %x, <i32 5, i32 5, i32 5, i32 5> |
| %r = tail call <4 x i16> @llvm.aarch64.neon.uqxtn.v4i16(<4 x i32> %s) |
| ret <4 x i16> %r |
| } |
| |
| define <4 x i16> @NarrowAShrI32By5ToU16(<4 x i32> %x) { |
| ; CHECK-LABEL: NarrowAShrI32By5ToU16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: sqshrun v0.4h, v0.4s, #5 |
| ; CHECK-NEXT: ret |
| %s = ashr <4 x i32> %x, <i32 5, i32 5, i32 5, i32 5> |
| %r = tail call <4 x i16> @llvm.aarch64.neon.sqxtun.v4i16(<4 x i32> %s) |
| ret <4 x i16> %r |
| } |
| |
| define <4 x i16> @NarrowLShrI32By5(<4 x i32> %x) { |
| ; CHECK-LABEL: NarrowLShrI32By5: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ushr v0.4s, v0.4s, #5 |
| ; CHECK-NEXT: sqxtn v0.4h, v0.4s |
| ; CHECK-NEXT: ret |
| %s = lshr <4 x i32> %x, <i32 5, i32 5, i32 5, i32 5> |
| %r = tail call <4 x i16> @llvm.aarch64.neon.sqxtn.v4i16(<4 x i32> %s) |
| ret <4 x i16> %r |
| } |
| |
| define <4 x i16> @NarrowLShrU32By5(<4 x i32> %x) { |
| ; CHECK-LABEL: NarrowLShrU32By5: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uqshrn v0.4h, v0.4s, #5 |
| ; CHECK-NEXT: ret |
| %s = lshr <4 x i32> %x, <i32 5, i32 5, i32 5, i32 5> |
| %r = tail call <4 x i16> @llvm.aarch64.neon.uqxtn.v4i16(<4 x i32> %s) |
| ret <4 x i16> %r |
| } |
| |
| define <4 x i16> @NarrowLShrI32By5ToU16(<4 x i32> %x) { |
| ; CHECK-LABEL: NarrowLShrI32By5ToU16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ushr v0.4s, v0.4s, #5 |
| ; CHECK-NEXT: sqxtun v0.4h, v0.4s |
| ; CHECK-NEXT: ret |
| %s = lshr <4 x i32> %x, <i32 5, i32 5, i32 5, i32 5> |
| %r = tail call <4 x i16> @llvm.aarch64.neon.sqxtun.v4i16(<4 x i32> %s) |
| ret <4 x i16> %r |
| } |
| |
| |
| define <2 x i32> @NarrowAShri64By5(<2 x i64> %x) { |
| ; CHECK-LABEL: NarrowAShri64By5: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: sqshrn v0.2s, v0.2d, #5 |
| ; CHECK-NEXT: ret |
| %s = ashr <2 x i64> %x, <i64 5, i64 5> |
| %r = tail call <2 x i32> @llvm.aarch64.neon.sqxtn.v2i32(<2 x i64> %s) |
| ret <2 x i32> %r |
| } |
| |
| define <2 x i32> @NarrowAShrU64By5(<2 x i64> %x) { |
| ; CHECK-LABEL: NarrowAShrU64By5: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: sshr v0.2d, v0.2d, #5 |
| ; CHECK-NEXT: uqxtn v0.2s, v0.2d |
| ; CHECK-NEXT: ret |
| %s = ashr <2 x i64> %x, <i64 5, i64 5> |
| %r = tail call <2 x i32> @llvm.aarch64.neon.uqxtn.v2i32(<2 x i64> %s) |
| ret <2 x i32> %r |
| } |
| |
| define <2 x i32> @NarrowAShri64By5ToU32(<2 x i64> %x) { |
| ; CHECK-LABEL: NarrowAShri64By5ToU32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: sqshrun v0.2s, v0.2d, #5 |
| ; CHECK-NEXT: ret |
| %s = ashr <2 x i64> %x, <i64 5, i64 5> |
| %r = tail call <2 x i32> @llvm.aarch64.neon.sqxtun.v2i32(<2 x i64> %s) |
| ret <2 x i32> %r |
| } |
| |
| define <2 x i32> @NarrowLShri64By5(<2 x i64> %x) { |
| ; CHECK-LABEL: NarrowLShri64By5: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ushr v0.2d, v0.2d, #5 |
| ; CHECK-NEXT: sqxtn v0.2s, v0.2d |
| ; CHECK-NEXT: ret |
| %s = lshr <2 x i64> %x, <i64 5, i64 5> |
| %r = tail call <2 x i32> @llvm.aarch64.neon.sqxtn.v2i32(<2 x i64> %s) |
| ret <2 x i32> %r |
| } |
| |
| define <2 x i32> @NarrowLShrU64By5(<2 x i64> %x) { |
| ; CHECK-LABEL: NarrowLShrU64By5: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uqshrn v0.2s, v0.2d, #5 |
| ; CHECK-NEXT: ret |
| %s = lshr <2 x i64> %x, <i64 5, i64 5> |
| %r = tail call <2 x i32> @llvm.aarch64.neon.uqxtn.v2i32(<2 x i64> %s) |
| ret <2 x i32> %r |
| } |
| |
| define <2 x i32> @NarrowLShri64By5ToU32(<2 x i64> %x) { |
| ; CHECK-LABEL: NarrowLShri64By5ToU32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ushr v0.2d, v0.2d, #5 |
| ; CHECK-NEXT: sqxtun v0.2s, v0.2d |
| ; CHECK-NEXT: ret |
| %s = lshr <2 x i64> %x, <i64 5, i64 5> |
| %r = tail call <2 x i32> @llvm.aarch64.neon.sqxtun.v2i32(<2 x i64> %s) |
| ret <2 x i32> %r |
| } |
| |
| |
| define <8 x i8> @NarrowAShri16By5(<8 x i16> %x) { |
| ; CHECK-LABEL: NarrowAShri16By5: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: sqshrn v0.8b, v0.8h, #5 |
| ; CHECK-NEXT: ret |
| %s = ashr <8 x i16> %x, <i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5> |
| %r = tail call <8 x i8> @llvm.aarch64.neon.sqxtn.v8i8(<8 x i16> %s) |
| ret <8 x i8> %r |
| } |
| |
| define <8 x i8> @NarrowAShrU16By5(<8 x i16> %x) { |
| ; CHECK-LABEL: NarrowAShrU16By5: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: sshr v0.8h, v0.8h, #5 |
| ; CHECK-NEXT: uqxtn v0.8b, v0.8h |
| ; CHECK-NEXT: ret |
| %s = ashr <8 x i16> %x, <i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5> |
| %r = tail call <8 x i8> @llvm.aarch64.neon.uqxtn.v8i8(<8 x i16> %s) |
| ret <8 x i8> %r |
| } |
| |
| define <8 x i8> @NarrowAShri16By5ToU8(<8 x i16> %x) { |
| ; CHECK-LABEL: NarrowAShri16By5ToU8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: sqshrun v0.8b, v0.8h, #5 |
| ; CHECK-NEXT: ret |
| %s = ashr <8 x i16> %x, <i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5> |
| %r = tail call <8 x i8> @llvm.aarch64.neon.sqxtun.v8i8(<8 x i16> %s) |
| ret <8 x i8> %r |
| } |
| |
| define <8 x i8> @NarrowLShri16By5(<8 x i16> %x) { |
| ; CHECK-LABEL: NarrowLShri16By5: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ushr v0.8h, v0.8h, #5 |
| ; CHECK-NEXT: sqxtn v0.8b, v0.8h |
| ; CHECK-NEXT: ret |
| %s = lshr <8 x i16> %x, <i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5> |
| %r = tail call <8 x i8> @llvm.aarch64.neon.sqxtn.v8i8(<8 x i16> %s) |
| ret <8 x i8> %r |
| } |
| |
| define <8 x i8> @NarrowLShrU16By5(<8 x i16> %x) { |
| ; CHECK-LABEL: NarrowLShrU16By5: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uqshrn v0.8b, v0.8h, #5 |
| ; CHECK-NEXT: ret |
| %s = lshr <8 x i16> %x, <i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5> |
| %r = tail call <8 x i8> @llvm.aarch64.neon.uqxtn.v8i8(<8 x i16> %s) |
| ret <8 x i8> %r |
| } |
| |
| define <8 x i8> @NarrowLShri16By5ToU8(<8 x i16> %x) { |
| ; CHECK-LABEL: NarrowLShri16By5ToU8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ushr v0.8h, v0.8h, #5 |
| ; CHECK-NEXT: sqxtun v0.8b, v0.8h |
| ; CHECK-NEXT: ret |
| %s = lshr <8 x i16> %x, <i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5> |
| %r = tail call <8 x i8> @llvm.aarch64.neon.sqxtun.v8i8(<8 x i16> %s) |
| ret <8 x i8> %r |
| } |
| |
| |
| |
| |
| |
| define <4 x i16> @NarrowAShrI32By31(<4 x i32> %x) { |
| ; CHECK-LABEL: NarrowAShrI32By31: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: sqshrn v0.4h, v0.4s, #16 |
| ; CHECK-NEXT: ret |
| %s = ashr <4 x i32> %x, <i32 16, i32 16, i32 16, i32 16> |
| %r = tail call <4 x i16> @llvm.aarch64.neon.sqxtn.v4i16(<4 x i32> %s) |
| ret <4 x i16> %r |
| } |
| |
| define <4 x i16> @NarrowAShrI32By31ToU16(<4 x i32> %x) { |
| ; CHECK-LABEL: NarrowAShrI32By31ToU16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: sqshrun v0.4h, v0.4s, #16 |
| ; CHECK-NEXT: ret |
| %s = ashr <4 x i32> %x, <i32 16, i32 16, i32 16, i32 16> |
| %r = tail call <4 x i16> @llvm.aarch64.neon.sqxtun.v4i16(<4 x i32> %s) |
| ret <4 x i16> %r |
| } |
| |
| define <4 x i16> @NarrowLShrU32By31(<4 x i32> %x) { |
| ; CHECK-LABEL: NarrowLShrU32By31: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uqshrn v0.4h, v0.4s, #16 |
| ; CHECK-NEXT: ret |
| %s = lshr <4 x i32> %x, <i32 16, i32 16, i32 16, i32 16> |
| %r = tail call <4 x i16> @llvm.aarch64.neon.uqxtn.v4i16(<4 x i32> %s) |
| ret <4 x i16> %r |
| } |
| |
| |
| define <16 x i8> @signed_minmax_v8i16_to_v16i8(<16 x i16> %x) { |
| ; CHECK-LABEL: signed_minmax_v8i16_to_v16i8: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: sqshrn v0.8b, v0.8h, #5 |
| ; CHECK-NEXT: sqshrn2 v0.16b, v1.8h, #5 |
| ; CHECK-NEXT: ret |
| entry: |
| %s = ashr <16 x i16> %x, <i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5> |
| %min = call <16 x i16> @llvm.smin.v8i16(<16 x i16> %s, <16 x i16> <i16 127, i16 127, i16 127, i16 127, i16 127, i16 127, i16 127, i16 127, i16 127, i16 127, i16 127, i16 127, i16 127, i16 127, i16 127, i16 127>) |
| %max = call <16 x i16> @llvm.smax.v8i16(<16 x i16> %min, <16 x i16> <i16 -128, i16 -128, i16 -128, i16 -128, i16 -128, i16 -128, i16 -128, i16 -128, i16 -128, i16 -128, i16 -128, i16 -128, i16 -128, i16 -128, i16 -128, i16 -128>) |
| %trunc = trunc <16 x i16> %max to <16 x i8> |
| ret <16 x i8> %trunc |
| } |
| |
| define <16 x i8> @unsigned_minmax_v8i16_to_v16i8(<16 x i16> %x) { |
| ; CHECK-LABEL: unsigned_minmax_v8i16_to_v16i8: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: uqshrn v0.8b, v0.8h, #5 |
| ; CHECK-NEXT: uqshrn2 v0.16b, v1.8h, #5 |
| ; CHECK-NEXT: ret |
| entry: |
| %s = lshr <16 x i16> %x, <i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5> |
| %min = call <16 x i16> @llvm.umin.v8i16(<16 x i16> %s, <16 x i16> <i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255>) |
| %trunc = trunc <16 x i16> %min to <16 x i8> |
| ret <16 x i8> %trunc |
| } |
| |
| define <16 x i8> @unsigned_signed_minmax_v8i16_to_v16i8(<16 x i16> %x) { |
| ; CHECK-LABEL: unsigned_signed_minmax_v8i16_to_v16i8: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: sqshrun v0.8b, v0.8h, #5 |
| ; CHECK-NEXT: sqshrun2 v0.16b, v1.8h, #5 |
| ; CHECK-NEXT: ret |
| entry: |
| %s = ashr <16 x i16> %x, <i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5> |
| %max = call <16 x i16> @llvm.smax.v8i16(<16 x i16> %s, <16 x i16> <i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0, i16 0>) |
| %min = call <16 x i16> @llvm.umin.v8i16(<16 x i16> %max, <16 x i16> <i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255, i16 255>) |
| %trunc = trunc <16 x i16> %min to <16 x i8> |
| ret <16 x i8> %trunc |
| } |
| |
| |
| define <8 x i16> @signed_minmax_v4i32_to_v8i16(<8 x i32> %x) { |
| ; CHECK-LABEL: signed_minmax_v4i32_to_v8i16: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: sqshrn v0.4h, v0.4s, #5 |
| ; CHECK-NEXT: sqshrn2 v0.8h, v1.4s, #5 |
| ; CHECK-NEXT: ret |
| entry: |
| %s = ashr <8 x i32> %x, <i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5> |
| %min = call <8 x i32> @llvm.smin.v8i32(<8 x i32> %s, <8 x i32> <i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767, i32 32767>) |
| %max = call <8 x i32> @llvm.smax.v8i32(<8 x i32> %min, <8 x i32> <i32 -32768, i32 -32768, i32 -32768, i32 -32768, i32 -32768, i32 -32768, i32 -32768, i32 -32768>) |
| %trunc = trunc <8 x i32> %max to <8 x i16> |
| ret <8 x i16> %trunc |
| } |
| |
| define <8 x i16> @unsigned_minmax_v4i32_to_v8i16(<8 x i32> %x) { |
| ; CHECK-LABEL: unsigned_minmax_v4i32_to_v8i16: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: uqshrn v0.4h, v0.4s, #5 |
| ; CHECK-NEXT: uqshrn2 v0.8h, v1.4s, #5 |
| ; CHECK-NEXT: ret |
| entry: |
| %s = lshr <8 x i32> %x, <i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5> |
| %min = call <8 x i32> @llvm.umin.v8i32(<8 x i32> %s, <8 x i32> <i32 65535, i32 65535, i32 65535, i32 65535, i32 65535, i32 65535, i32 65535, i32 65535>) |
| %trunc = trunc <8 x i32> %min to <8 x i16> |
| ret <8 x i16> %trunc |
| } |
| |
| define <8 x i16> @unsigned_signed_minmax_v4i32_to_v8i16(<8 x i32> %x) { |
| ; CHECK-LABEL: unsigned_signed_minmax_v4i32_to_v8i16: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: sqshrun v0.4h, v0.4s, #5 |
| ; CHECK-NEXT: sqshrun2 v0.8h, v1.4s, #5 |
| ; CHECK-NEXT: ret |
| entry: |
| %s = ashr <8 x i32> %x, <i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5, i32 5> |
| %max = call <8 x i32> @llvm.smax.v8i32(<8 x i32> %s, <8 x i32> <i32 0, i32 0, i32 0, i32 0, i32 0, i32 0, i32 0, i32 0>) |
| %min = call <8 x i32> @llvm.umin.v8i32(<8 x i32> %max, <8 x i32> <i32 65535, i32 65535, i32 65535, i32 65535, i32 65535, i32 65535, i32 65535, i32 65535>) |
| %trunc = trunc <8 x i32> %min to <8 x i16> |
| ret <8 x i16> %trunc |
| } |
| |
| |
| define <4 x i32> @signed_minmax_v4i64_to_v8i32(<4 x i64> %x) { |
| ; CHECK-LABEL: signed_minmax_v4i64_to_v8i32: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: sqshrn v0.2s, v0.2d, #5 |
| ; CHECK-NEXT: sqshrn2 v0.4s, v1.2d, #5 |
| ; CHECK-NEXT: ret |
| entry: |
| %s = ashr <4 x i64> %x, <i64 5, i64 5, i64 5, i64 5> |
| %min = call <4 x i64> @llvm.smin.v8i64(<4 x i64> %s, <4 x i64> <i64 2147483647, i64 2147483647, i64 2147483647, i64 2147483647>) |
| %max = call <4 x i64> @llvm.smax.v8i64(<4 x i64> %min, <4 x i64> <i64 -2147483648, i64 -2147483648, i64 -2147483648, i64 -2147483648>) |
| %trunc = trunc <4 x i64> %max to <4 x i32> |
| ret <4 x i32> %trunc |
| } |
| |
| define <4 x i32> @unsigned_minmax_v4i64_to_v8i32(<4 x i64> %x) { |
| ; CHECK-LABEL: unsigned_minmax_v4i64_to_v8i32: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: uqshrn v0.2s, v0.2d, #5 |
| ; CHECK-NEXT: uqshrn2 v0.4s, v1.2d, #5 |
| ; CHECK-NEXT: ret |
| entry: |
| %s = lshr <4 x i64> %x, <i64 5, i64 5, i64 5, i64 5> |
| %min = call <4 x i64> @llvm.umin.v8i64(<4 x i64> %s, <4 x i64> <i64 4294967295, i64 4294967295, i64 4294967295, i64 4294967295>) |
| %trunc = trunc <4 x i64> %min to <4 x i32> |
| ret <4 x i32> %trunc |
| } |
| |
| define <4 x i32> @unsigned_signed_minmax_v4i64_to_v8i32(<4 x i64> %x) { |
| ; CHECK-LABEL: unsigned_signed_minmax_v4i64_to_v8i32: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: sqshrun v0.2s, v0.2d, #5 |
| ; CHECK-NEXT: sqshrun2 v0.4s, v1.2d, #5 |
| ; CHECK-NEXT: ret |
| entry: |
| %s = ashr <4 x i64> %x, <i64 5, i64 5, i64 5, i64 5> |
| %max = call <4 x i64> @llvm.smax.v8i64(<4 x i64> %s, <4 x i64> <i64 0, i64 0, i64 0, i64 0>) |
| %min = call <4 x i64> @llvm.umin.v8i64(<4 x i64> %max, <4 x i64> <i64 4294967295, i64 4294967295, i64 4294967295, i64 4294967295>) |
| %trunc = trunc <4 x i64> %min to <4 x i32> |
| ret <4 x i32> %trunc |
| } |