| // NOTE: Assertions have been autogenerated by utils/update_cc_test_checks.py | 
 | // RUN: %clang_cc1 -triple thumbv8.1m.main-none-none-eabi -target-feature +mve.fp -mfloat-abi hard -O0 -disable-O0-optnone -emit-llvm -o - %s | opt -S -passes=mem2reg | FileCheck %s | 
 | // RUN: %clang_cc1 -triple thumbv8.1m.main-none-none-eabi -target-feature +mve.fp -mfloat-abi hard -O0 -disable-O0-optnone -DPOLYMORPHIC -emit-llvm -o - %s | opt -S -passes=mem2reg | FileCheck %s | 
 |  | 
 | // REQUIRES: aarch64-registered-target || arm-registered-target | 
 |  | 
 | #include <arm_mve.h> | 
 |  | 
 | // CHECK-LABEL: @test_vld1q_f16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <8 x half>, ptr [[BASE:%.*]], align 2 | 
 | // CHECK-NEXT:    ret <8 x half> [[TMP1]] | 
 | // | 
 | float16x8_t test_vld1q_f16(const float16_t *base) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     return vld1q(base); | 
 | #else /* POLYMORPHIC */ | 
 |     return vld1q_f16(base); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vld1q_f32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <4 x float>, ptr [[BASE:%.*]], align 4 | 
 | // CHECK-NEXT:    ret <4 x float> [[TMP1]] | 
 | // | 
 | float32x4_t test_vld1q_f32(const float32_t *base) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     return vld1q(base); | 
 | #else /* POLYMORPHIC */ | 
 |     return vld1q_f32(base); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vld1q_s8( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <16 x i8>, ptr [[BASE:%.*]], align 1 | 
 | // CHECK-NEXT:    ret <16 x i8> [[TMP1]] | 
 | // | 
 | int8x16_t test_vld1q_s8(const int8_t *base) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     return vld1q(base); | 
 | #else /* POLYMORPHIC */ | 
 |     return vld1q_s8(base); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vld1q_s16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <8 x i16>, ptr [[BASE:%.*]], align 2 | 
 | // CHECK-NEXT:    ret <8 x i16> [[TMP1]] | 
 | // | 
 | int16x8_t test_vld1q_s16(const int16_t *base) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     return vld1q(base); | 
 | #else /* POLYMORPHIC */ | 
 |     return vld1q_s16(base); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vld1q_s32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <4 x i32>, ptr [[BASE:%.*]], align 4 | 
 | // CHECK-NEXT:    ret <4 x i32> [[TMP1]] | 
 | // | 
 | int32x4_t test_vld1q_s32(const int32_t *base) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     return vld1q(base); | 
 | #else /* POLYMORPHIC */ | 
 |     return vld1q_s32(base); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vld1q_u8( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <16 x i8>, ptr [[BASE:%.*]], align 1 | 
 | // CHECK-NEXT:    ret <16 x i8> [[TMP1]] | 
 | // | 
 | uint8x16_t test_vld1q_u8(const uint8_t *base) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     return vld1q(base); | 
 | #else /* POLYMORPHIC */ | 
 |     return vld1q_u8(base); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vld1q_u16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <8 x i16>, ptr [[BASE:%.*]], align 2 | 
 | // CHECK-NEXT:    ret <8 x i16> [[TMP1]] | 
 | // | 
 | uint16x8_t test_vld1q_u16(const uint16_t *base) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     return vld1q(base); | 
 | #else /* POLYMORPHIC */ | 
 |     return vld1q_u16(base); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vld1q_u32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <4 x i32>, ptr [[BASE:%.*]], align 4 | 
 | // CHECK-NEXT:    ret <4 x i32> [[TMP1]] | 
 | // | 
 | uint32x4_t test_vld1q_u32(const uint32_t *base) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     return vld1q(base); | 
 | #else /* POLYMORPHIC */ | 
 |     return vld1q_u32(base); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vld1q_z_f16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <8 x half> @llvm.masked.load.v8f16.p0(ptr [[BASE:%.*]], i32 2, <8 x i1> [[TMP2]], <8 x half> zeroinitializer) | 
 | // CHECK-NEXT:    ret <8 x half> [[TMP3]] | 
 | // | 
 | float16x8_t test_vld1q_z_f16(const float16_t *base, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     return vld1q_z(base, p); | 
 | #else /* POLYMORPHIC */ | 
 |     return vld1q_z_f16(base, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vld1q_z_f32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <4 x float> @llvm.masked.load.v4f32.p0(ptr [[BASE:%.*]], i32 4, <4 x i1> [[TMP2]], <4 x float> zeroinitializer) | 
 | // CHECK-NEXT:    ret <4 x float> [[TMP3]] | 
 | // | 
 | float32x4_t test_vld1q_z_f32(const float32_t *base, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     return vld1q_z(base, p); | 
 | #else /* POLYMORPHIC */ | 
 |     return vld1q_z_f32(base, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vld1q_z_s8( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <16 x i1> @llvm.arm.mve.pred.i2v.v16i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <16 x i8> @llvm.masked.load.v16i8.p0(ptr [[BASE:%.*]], i32 1, <16 x i1> [[TMP2]], <16 x i8> zeroinitializer) | 
 | // CHECK-NEXT:    ret <16 x i8> [[TMP3]] | 
 | // | 
 | int8x16_t test_vld1q_z_s8(const int8_t *base, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     return vld1q_z(base, p); | 
 | #else /* POLYMORPHIC */ | 
 |     return vld1q_z_s8(base, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vld1q_z_s16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <8 x i16> @llvm.masked.load.v8i16.p0(ptr [[BASE:%.*]], i32 2, <8 x i1> [[TMP2]], <8 x i16> zeroinitializer) | 
 | // CHECK-NEXT:    ret <8 x i16> [[TMP3]] | 
 | // | 
 | int16x8_t test_vld1q_z_s16(const int16_t *base, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     return vld1q_z(base, p); | 
 | #else /* POLYMORPHIC */ | 
 |     return vld1q_z_s16(base, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vld1q_z_s32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <4 x i32> @llvm.masked.load.v4i32.p0(ptr [[BASE:%.*]], i32 4, <4 x i1> [[TMP2]], <4 x i32> zeroinitializer) | 
 | // CHECK-NEXT:    ret <4 x i32> [[TMP3]] | 
 | // | 
 | int32x4_t test_vld1q_z_s32(const int32_t *base, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     return vld1q_z(base, p); | 
 | #else /* POLYMORPHIC */ | 
 |     return vld1q_z_s32(base, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vld1q_z_u8( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <16 x i1> @llvm.arm.mve.pred.i2v.v16i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <16 x i8> @llvm.masked.load.v16i8.p0(ptr [[BASE:%.*]], i32 1, <16 x i1> [[TMP2]], <16 x i8> zeroinitializer) | 
 | // CHECK-NEXT:    ret <16 x i8> [[TMP3]] | 
 | // | 
 | uint8x16_t test_vld1q_z_u8(const uint8_t *base, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     return vld1q_z(base, p); | 
 | #else /* POLYMORPHIC */ | 
 |     return vld1q_z_u8(base, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vld1q_z_u16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <8 x i16> @llvm.masked.load.v8i16.p0(ptr [[BASE:%.*]], i32 2, <8 x i1> [[TMP2]], <8 x i16> zeroinitializer) | 
 | // CHECK-NEXT:    ret <8 x i16> [[TMP3]] | 
 | // | 
 | uint16x8_t test_vld1q_z_u16(const uint16_t *base, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     return vld1q_z(base, p); | 
 | #else /* POLYMORPHIC */ | 
 |     return vld1q_z_u16(base, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vld1q_z_u32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <4 x i32> @llvm.masked.load.v4i32.p0(ptr [[BASE:%.*]], i32 4, <4 x i1> [[TMP2]], <4 x i32> zeroinitializer) | 
 | // CHECK-NEXT:    ret <4 x i32> [[TMP3]] | 
 | // | 
 | uint32x4_t test_vld1q_z_u32(const uint32_t *base, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     return vld1q_z(base, p); | 
 | #else /* POLYMORPHIC */ | 
 |     return vld1q_z_u32(base, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrbq_s8( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <16 x i8>, ptr [[BASE:%.*]], align 1 | 
 | // CHECK-NEXT:    ret <16 x i8> [[TMP1]] | 
 | // | 
 | int8x16_t test_vldrbq_s8(const int8_t *base) | 
 | { | 
 |     return vldrbq_s8(base); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrbq_s16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <8 x i8>, ptr [[BASE:%.*]], align 1 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = sext <8 x i8> [[TMP1]] to <8 x i16> | 
 | // CHECK-NEXT:    ret <8 x i16> [[TMP2]] | 
 | // | 
 | int16x8_t test_vldrbq_s16(const int8_t *base) | 
 | { | 
 |     return vldrbq_s16(base); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrbq_s32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <4 x i8>, ptr [[BASE:%.*]], align 1 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = sext <4 x i8> [[TMP1]] to <4 x i32> | 
 | // CHECK-NEXT:    ret <4 x i32> [[TMP2]] | 
 | // | 
 | int32x4_t test_vldrbq_s32(const int8_t *base) | 
 | { | 
 |     return vldrbq_s32(base); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrbq_u8( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <16 x i8>, ptr [[BASE:%.*]], align 1 | 
 | // CHECK-NEXT:    ret <16 x i8> [[TMP1]] | 
 | // | 
 | uint8x16_t test_vldrbq_u8(const uint8_t *base) | 
 | { | 
 |     return vldrbq_u8(base); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrbq_u16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <8 x i8>, ptr [[BASE:%.*]], align 1 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = zext <8 x i8> [[TMP1]] to <8 x i16> | 
 | // CHECK-NEXT:    ret <8 x i16> [[TMP2]] | 
 | // | 
 | uint16x8_t test_vldrbq_u16(const uint8_t *base) | 
 | { | 
 |     return vldrbq_u16(base); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrbq_u32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <4 x i8>, ptr [[BASE:%.*]], align 1 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = zext <4 x i8> [[TMP1]] to <4 x i32> | 
 | // CHECK-NEXT:    ret <4 x i32> [[TMP2]] | 
 | // | 
 | uint32x4_t test_vldrbq_u32(const uint8_t *base) | 
 | { | 
 |     return vldrbq_u32(base); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrbq_z_s8( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <16 x i1> @llvm.arm.mve.pred.i2v.v16i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <16 x i8> @llvm.masked.load.v16i8.p0(ptr [[BASE:%.*]], i32 1, <16 x i1> [[TMP2]], <16 x i8> zeroinitializer) | 
 | // CHECK-NEXT:    ret <16 x i8> [[TMP3]] | 
 | // | 
 | int8x16_t test_vldrbq_z_s8(const int8_t *base, mve_pred16_t p) | 
 | { | 
 |     return vldrbq_z_s8(base, p); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrbq_z_s16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <8 x i8> @llvm.masked.load.v8i8.p0(ptr [[BASE:%.*]], i32 1, <8 x i1> [[TMP2]], <8 x i8> zeroinitializer) | 
 | // CHECK-NEXT:    [[TMP4:%.*]] = sext <8 x i8> [[TMP3]] to <8 x i16> | 
 | // CHECK-NEXT:    ret <8 x i16> [[TMP4]] | 
 | // | 
 | int16x8_t test_vldrbq_z_s16(const int8_t *base, mve_pred16_t p) | 
 | { | 
 |     return vldrbq_z_s16(base, p); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrbq_z_s32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <4 x i8> @llvm.masked.load.v4i8.p0(ptr [[BASE:%.*]], i32 1, <4 x i1> [[TMP2]], <4 x i8> zeroinitializer) | 
 | // CHECK-NEXT:    [[TMP4:%.*]] = sext <4 x i8> [[TMP3]] to <4 x i32> | 
 | // CHECK-NEXT:    ret <4 x i32> [[TMP4]] | 
 | // | 
 | int32x4_t test_vldrbq_z_s32(const int8_t *base, mve_pred16_t p) | 
 | { | 
 |     return vldrbq_z_s32(base, p); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrbq_z_u8( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <16 x i1> @llvm.arm.mve.pred.i2v.v16i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <16 x i8> @llvm.masked.load.v16i8.p0(ptr [[BASE:%.*]], i32 1, <16 x i1> [[TMP2]], <16 x i8> zeroinitializer) | 
 | // CHECK-NEXT:    ret <16 x i8> [[TMP3]] | 
 | // | 
 | uint8x16_t test_vldrbq_z_u8(const uint8_t *base, mve_pred16_t p) | 
 | { | 
 |     return vldrbq_z_u8(base, p); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrbq_z_u16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <8 x i8> @llvm.masked.load.v8i8.p0(ptr [[BASE:%.*]], i32 1, <8 x i1> [[TMP2]], <8 x i8> zeroinitializer) | 
 | // CHECK-NEXT:    [[TMP4:%.*]] = zext <8 x i8> [[TMP3]] to <8 x i16> | 
 | // CHECK-NEXT:    ret <8 x i16> [[TMP4]] | 
 | // | 
 | uint16x8_t test_vldrbq_z_u16(const uint8_t *base, mve_pred16_t p) | 
 | { | 
 |     return vldrbq_z_u16(base, p); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrbq_z_u32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <4 x i8> @llvm.masked.load.v4i8.p0(ptr [[BASE:%.*]], i32 1, <4 x i1> [[TMP2]], <4 x i8> zeroinitializer) | 
 | // CHECK-NEXT:    [[TMP4:%.*]] = zext <4 x i8> [[TMP3]] to <4 x i32> | 
 | // CHECK-NEXT:    ret <4 x i32> [[TMP4]] | 
 | // | 
 | uint32x4_t test_vldrbq_z_u32(const uint8_t *base, mve_pred16_t p) | 
 | { | 
 |     return vldrbq_z_u32(base, p); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrhq_f16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <8 x half>, ptr [[BASE:%.*]], align 2 | 
 | // CHECK-NEXT:    ret <8 x half> [[TMP1]] | 
 | // | 
 | float16x8_t test_vldrhq_f16(const float16_t *base) | 
 | { | 
 |     return vldrhq_f16(base); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrhq_s16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <8 x i16>, ptr [[BASE:%.*]], align 2 | 
 | // CHECK-NEXT:    ret <8 x i16> [[TMP1]] | 
 | // | 
 | int16x8_t test_vldrhq_s16(const int16_t *base) | 
 | { | 
 |     return vldrhq_s16(base); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrhq_s32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <4 x i16>, ptr [[BASE:%.*]], align 2 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = sext <4 x i16> [[TMP1]] to <4 x i32> | 
 | // CHECK-NEXT:    ret <4 x i32> [[TMP2]] | 
 | // | 
 | int32x4_t test_vldrhq_s32(const int16_t *base) | 
 | { | 
 |     return vldrhq_s32(base); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrhq_u16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <8 x i16>, ptr [[BASE:%.*]], align 2 | 
 | // CHECK-NEXT:    ret <8 x i16> [[TMP1]] | 
 | // | 
 | uint16x8_t test_vldrhq_u16(const uint16_t *base) | 
 | { | 
 |     return vldrhq_u16(base); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrhq_u32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <4 x i16>, ptr [[BASE:%.*]], align 2 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = zext <4 x i16> [[TMP1]] to <4 x i32> | 
 | // CHECK-NEXT:    ret <4 x i32> [[TMP2]] | 
 | // | 
 | uint32x4_t test_vldrhq_u32(const uint16_t *base) | 
 | { | 
 |     return vldrhq_u32(base); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrhq_z_f16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <8 x half> @llvm.masked.load.v8f16.p0(ptr [[BASE:%.*]], i32 2, <8 x i1> [[TMP2]], <8 x half> zeroinitializer) | 
 | // CHECK-NEXT:    ret <8 x half> [[TMP3]] | 
 | // | 
 | float16x8_t test_vldrhq_z_f16(const float16_t *base, mve_pred16_t p) | 
 | { | 
 |     return vldrhq_z_f16(base, p); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrhq_z_s16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <8 x i16> @llvm.masked.load.v8i16.p0(ptr [[BASE:%.*]], i32 2, <8 x i1> [[TMP2]], <8 x i16> zeroinitializer) | 
 | // CHECK-NEXT:    ret <8 x i16> [[TMP3]] | 
 | // | 
 | int16x8_t test_vldrhq_z_s16(const int16_t *base, mve_pred16_t p) | 
 | { | 
 |     return vldrhq_z_s16(base, p); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrhq_z_s32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <4 x i16> @llvm.masked.load.v4i16.p0(ptr [[BASE:%.*]], i32 2, <4 x i1> [[TMP2]], <4 x i16> zeroinitializer) | 
 | // CHECK-NEXT:    [[TMP4:%.*]] = sext <4 x i16> [[TMP3]] to <4 x i32> | 
 | // CHECK-NEXT:    ret <4 x i32> [[TMP4]] | 
 | // | 
 | int32x4_t test_vldrhq_z_s32(const int16_t *base, mve_pred16_t p) | 
 | { | 
 |     return vldrhq_z_s32(base, p); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrhq_z_u16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <8 x i16> @llvm.masked.load.v8i16.p0(ptr [[BASE:%.*]], i32 2, <8 x i1> [[TMP2]], <8 x i16> zeroinitializer) | 
 | // CHECK-NEXT:    ret <8 x i16> [[TMP3]] | 
 | // | 
 | uint16x8_t test_vldrhq_z_u16(const uint16_t *base, mve_pred16_t p) | 
 | { | 
 |     return vldrhq_z_u16(base, p); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrhq_z_u32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <4 x i16> @llvm.masked.load.v4i16.p0(ptr [[BASE:%.*]], i32 2, <4 x i1> [[TMP2]], <4 x i16> zeroinitializer) | 
 | // CHECK-NEXT:    [[TMP4:%.*]] = zext <4 x i16> [[TMP3]] to <4 x i32> | 
 | // CHECK-NEXT:    ret <4 x i32> [[TMP4]] | 
 | // | 
 | uint32x4_t test_vldrhq_z_u32(const uint16_t *base, mve_pred16_t p) | 
 | { | 
 |     return vldrhq_z_u32(base, p); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrwq_f32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <4 x float>, ptr [[BASE:%.*]], align 4 | 
 | // CHECK-NEXT:    ret <4 x float> [[TMP1]] | 
 | // | 
 | float32x4_t test_vldrwq_f32(const float32_t *base) | 
 | { | 
 |     return vldrwq_f32(base); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrwq_s32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <4 x i32>, ptr [[BASE:%.*]], align 4 | 
 | // CHECK-NEXT:    ret <4 x i32> [[TMP1]] | 
 | // | 
 | int32x4_t test_vldrwq_s32(const int32_t *base) | 
 | { | 
 |     return vldrwq_s32(base); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrwq_u32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = load <4 x i32>, ptr [[BASE:%.*]], align 4 | 
 | // CHECK-NEXT:    ret <4 x i32> [[TMP1]] | 
 | // | 
 | uint32x4_t test_vldrwq_u32(const uint32_t *base) | 
 | { | 
 |     return vldrwq_u32(base); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrwq_z_f32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <4 x float> @llvm.masked.load.v4f32.p0(ptr [[BASE:%.*]], i32 4, <4 x i1> [[TMP2]], <4 x float> zeroinitializer) | 
 | // CHECK-NEXT:    ret <4 x float> [[TMP3]] | 
 | // | 
 | float32x4_t test_vldrwq_z_f32(const float32_t *base, mve_pred16_t p) | 
 | { | 
 |     return vldrwq_z_f32(base, p); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrwq_z_s32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <4 x i32> @llvm.masked.load.v4i32.p0(ptr [[BASE:%.*]], i32 4, <4 x i1> [[TMP2]], <4 x i32> zeroinitializer) | 
 | // CHECK-NEXT:    ret <4 x i32> [[TMP3]] | 
 | // | 
 | int32x4_t test_vldrwq_z_s32(const int32_t *base, mve_pred16_t p) | 
 | { | 
 |     return vldrwq_z_s32(base, p); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vldrwq_z_u32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <4 x i32> @llvm.masked.load.v4i32.p0(ptr [[BASE:%.*]], i32 4, <4 x i1> [[TMP2]], <4 x i32> zeroinitializer) | 
 | // CHECK-NEXT:    ret <4 x i32> [[TMP3]] | 
 | // | 
 | uint32x4_t test_vldrwq_z_u32(const uint32_t *base, mve_pred16_t p) | 
 | { | 
 |     return vldrwq_z_u32(base, p); | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vst1q_f16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    store <8 x half> [[VALUE:%.*]], ptr [[BASE:%.*]], align 2 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vst1q_f16(float16_t *base, float16x8_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vst1q(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vst1q_f16(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vst1q_f32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    store <4 x float> [[VALUE:%.*]], ptr [[BASE:%.*]], align 4 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vst1q_f32(float32_t *base, float32x4_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vst1q(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vst1q_f32(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vst1q_s8( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    store <16 x i8> [[VALUE:%.*]], ptr [[BASE:%.*]], align 1 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vst1q_s8(int8_t *base, int8x16_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vst1q(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vst1q_s8(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vst1q_s16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    store <8 x i16> [[VALUE:%.*]], ptr [[BASE:%.*]], align 2 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vst1q_s16(int16_t *base, int16x8_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vst1q(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vst1q_s16(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vst1q_s32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    store <4 x i32> [[VALUE:%.*]], ptr [[BASE:%.*]], align 4 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vst1q_s32(int32_t *base, int32x4_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vst1q(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vst1q_s32(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vst1q_u8( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    store <16 x i8> [[VALUE:%.*]], ptr [[BASE:%.*]], align 1 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vst1q_u8(uint8_t *base, uint8x16_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vst1q(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vst1q_u8(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vst1q_u16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    store <8 x i16> [[VALUE:%.*]], ptr [[BASE:%.*]], align 2 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vst1q_u16(uint16_t *base, uint16x8_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vst1q(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vst1q_u16(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vst1q_u32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    store <4 x i32> [[VALUE:%.*]], ptr [[BASE:%.*]], align 4 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vst1q_u32(uint32_t *base, uint32x4_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vst1q(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vst1q_u32(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vst1q_p_f16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v8f16.p0(<8 x half> [[VALUE:%.*]], ptr [[BASE:%.*]], i32 2, <8 x i1> [[TMP2]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vst1q_p_f16(float16_t *base, float16x8_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vst1q_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vst1q_p_f16(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vst1q_p_f32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v4f32.p0(<4 x float> [[VALUE:%.*]], ptr [[BASE:%.*]], i32 4, <4 x i1> [[TMP2]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vst1q_p_f32(float32_t *base, float32x4_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vst1q_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vst1q_p_f32(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vst1q_p_s8( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <16 x i1> @llvm.arm.mve.pred.i2v.v16i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v16i8.p0(<16 x i8> [[VALUE:%.*]], ptr [[BASE:%.*]], i32 1, <16 x i1> [[TMP2]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vst1q_p_s8(int8_t *base, int8x16_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vst1q_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vst1q_p_s8(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vst1q_p_s16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v8i16.p0(<8 x i16> [[VALUE:%.*]], ptr [[BASE:%.*]], i32 2, <8 x i1> [[TMP2]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vst1q_p_s16(int16_t *base, int16x8_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vst1q_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vst1q_p_s16(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vst1q_p_s32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v4i32.p0(<4 x i32> [[VALUE:%.*]], ptr [[BASE:%.*]], i32 4, <4 x i1> [[TMP2]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vst1q_p_s32(int32_t *base, int32x4_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vst1q_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vst1q_p_s32(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vst1q_p_u8( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <16 x i1> @llvm.arm.mve.pred.i2v.v16i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v16i8.p0(<16 x i8> [[VALUE:%.*]], ptr [[BASE:%.*]], i32 1, <16 x i1> [[TMP2]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vst1q_p_u8(uint8_t *base, uint8x16_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vst1q_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vst1q_p_u8(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vst1q_p_u16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v8i16.p0(<8 x i16> [[VALUE:%.*]], ptr [[BASE:%.*]], i32 2, <8 x i1> [[TMP2]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vst1q_p_u16(uint16_t *base, uint16x8_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vst1q_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vst1q_p_u16(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vst1q_p_u32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v4i32.p0(<4 x i32> [[VALUE:%.*]], ptr [[BASE:%.*]], i32 4, <4 x i1> [[TMP2]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vst1q_p_u32(uint32_t *base, uint32x4_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vst1q_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vst1q_p_u32(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrbq_s8( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    store <16 x i8> [[VALUE:%.*]], ptr [[BASE:%.*]], align 1 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrbq_s8(int8_t *base, int8x16_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrbq(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrbq_s8(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrbq_s16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP0:%.*]] = trunc <8 x i16> [[VALUE:%.*]] to <8 x i8> | 
 | // CHECK-NEXT:    store <8 x i8> [[TMP0]], ptr [[BASE:%.*]], align 1 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrbq_s16(int8_t *base, int16x8_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrbq(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrbq_s16(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrbq_s32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP0:%.*]] = trunc <4 x i32> [[VALUE:%.*]] to <4 x i8> | 
 | // CHECK-NEXT:    store <4 x i8> [[TMP0]], ptr [[BASE:%.*]], align 1 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrbq_s32(int8_t *base, int32x4_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrbq(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrbq_s32(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrbq_u8( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    store <16 x i8> [[VALUE:%.*]], ptr [[BASE:%.*]], align 1 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrbq_u8(uint8_t *base, uint8x16_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrbq(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrbq_u8(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrbq_u16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP0:%.*]] = trunc <8 x i16> [[VALUE:%.*]] to <8 x i8> | 
 | // CHECK-NEXT:    store <8 x i8> [[TMP0]], ptr [[BASE:%.*]], align 1 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrbq_u16(uint8_t *base, uint16x8_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrbq(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrbq_u16(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrbq_u32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP0:%.*]] = trunc <4 x i32> [[VALUE:%.*]] to <4 x i8> | 
 | // CHECK-NEXT:    store <4 x i8> [[TMP0]], ptr [[BASE:%.*]], align 1 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrbq_u32(uint8_t *base, uint32x4_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrbq(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrbq_u32(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrbq_p_s8( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <16 x i1> @llvm.arm.mve.pred.i2v.v16i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v16i8.p0(<16 x i8> [[VALUE:%.*]], ptr [[BASE:%.*]], i32 1, <16 x i1> [[TMP2]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrbq_p_s8(int8_t *base, int8x16_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrbq_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrbq_p_s8(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrbq_p_s16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP0:%.*]] = trunc <8 x i16> [[VALUE:%.*]] to <8 x i8> | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 [[TMP2]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v8i8.p0(<8 x i8> [[TMP0]], ptr [[BASE:%.*]], i32 1, <8 x i1> [[TMP3]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrbq_p_s16(int8_t *base, int16x8_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrbq_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrbq_p_s16(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrbq_p_s32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP0:%.*]] = trunc <4 x i32> [[VALUE:%.*]] to <4 x i8> | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP2]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v4i8.p0(<4 x i8> [[TMP0]], ptr [[BASE:%.*]], i32 1, <4 x i1> [[TMP3]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrbq_p_s32(int8_t *base, int32x4_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrbq_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrbq_p_s32(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrbq_p_u8( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <16 x i1> @llvm.arm.mve.pred.i2v.v16i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v16i8.p0(<16 x i8> [[VALUE:%.*]], ptr [[BASE:%.*]], i32 1, <16 x i1> [[TMP2]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrbq_p_u8(uint8_t *base, uint8x16_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrbq_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrbq_p_u8(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrbq_p_u16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP0:%.*]] = trunc <8 x i16> [[VALUE:%.*]] to <8 x i8> | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 [[TMP2]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v8i8.p0(<8 x i8> [[TMP0]], ptr [[BASE:%.*]], i32 1, <8 x i1> [[TMP3]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrbq_p_u16(uint8_t *base, uint16x8_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrbq_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrbq_p_u16(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrbq_p_u32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP0:%.*]] = trunc <4 x i32> [[VALUE:%.*]] to <4 x i8> | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP2]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v4i8.p0(<4 x i8> [[TMP0]], ptr [[BASE:%.*]], i32 1, <4 x i1> [[TMP3]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrbq_p_u32(uint8_t *base, uint32x4_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrbq_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrbq_p_u32(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrhq_f16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    store <8 x half> [[VALUE:%.*]], ptr [[BASE:%.*]], align 2 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrhq_f16(float16_t *base, float16x8_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrhq(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrhq_f16(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrhq_s16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    store <8 x i16> [[VALUE:%.*]], ptr [[BASE:%.*]], align 2 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrhq_s16(int16_t *base, int16x8_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrhq(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrhq_s16(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrhq_s32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP0:%.*]] = trunc <4 x i32> [[VALUE:%.*]] to <4 x i16> | 
 | // CHECK-NEXT:    store <4 x i16> [[TMP0]], ptr [[BASE:%.*]], align 2 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrhq_s32(int16_t *base, int32x4_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrhq(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrhq_s32(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrhq_u16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    store <8 x i16> [[VALUE:%.*]], ptr [[BASE:%.*]], align 2 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrhq_u16(uint16_t *base, uint16x8_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrhq(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrhq_u16(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrhq_u32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP0:%.*]] = trunc <4 x i32> [[VALUE:%.*]] to <4 x i16> | 
 | // CHECK-NEXT:    store <4 x i16> [[TMP0]], ptr [[BASE:%.*]], align 2 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrhq_u32(uint16_t *base, uint32x4_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrhq(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrhq_u32(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrhq_p_f16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v8f16.p0(<8 x half> [[VALUE:%.*]], ptr [[BASE:%.*]], i32 2, <8 x i1> [[TMP2]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrhq_p_f16(float16_t *base, float16x8_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrhq_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrhq_p_f16(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrhq_p_s16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v8i16.p0(<8 x i16> [[VALUE:%.*]], ptr [[BASE:%.*]], i32 2, <8 x i1> [[TMP2]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrhq_p_s16(int16_t *base, int16x8_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrhq_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrhq_p_s16(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrhq_p_s32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP0:%.*]] = trunc <4 x i32> [[VALUE:%.*]] to <4 x i16> | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP2]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v4i16.p0(<4 x i16> [[TMP0]], ptr [[BASE:%.*]], i32 2, <4 x i1> [[TMP3]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrhq_p_s32(int16_t *base, int32x4_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrhq_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrhq_p_s32(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrhq_p_u16( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v8i16.p0(<8 x i16> [[VALUE:%.*]], ptr [[BASE:%.*]], i32 2, <8 x i1> [[TMP2]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrhq_p_u16(uint16_t *base, uint16x8_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrhq_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrhq_p_u16(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrhq_p_u32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP0:%.*]] = trunc <4 x i32> [[VALUE:%.*]] to <4 x i16> | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP3:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP2]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v4i16.p0(<4 x i16> [[TMP0]], ptr [[BASE:%.*]], i32 2, <4 x i1> [[TMP3]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrhq_p_u32(uint16_t *base, uint32x4_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrhq_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrhq_p_u32(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrwq_f32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    store <4 x float> [[VALUE:%.*]], ptr [[BASE:%.*]], align 4 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrwq_f32(float32_t *base, float32x4_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrwq(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrwq_f32(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrwq_s32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    store <4 x i32> [[VALUE:%.*]], ptr [[BASE:%.*]], align 4 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrwq_s32(int32_t *base, int32x4_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrwq(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrwq_s32(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrwq_u32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    store <4 x i32> [[VALUE:%.*]], ptr [[BASE:%.*]], align 4 | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrwq_u32(uint32_t *base, uint32x4_t value) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrwq(base, value); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrwq_u32(base, value); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrwq_p_f32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v4f32.p0(<4 x float> [[VALUE:%.*]], ptr [[BASE:%.*]], i32 4, <4 x i1> [[TMP2]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrwq_p_f32(float32_t *base, float32x4_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrwq_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrwq_p_f32(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrwq_p_s32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v4i32.p0(<4 x i32> [[VALUE:%.*]], ptr [[BASE:%.*]], i32 4, <4 x i1> [[TMP2]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrwq_p_s32(int32_t *base, int32x4_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrwq_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrwq_p_s32(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } | 
 |  | 
 | // CHECK-LABEL: @test_vstrwq_p_u32( | 
 | // CHECK-NEXT:  entry: | 
 | // CHECK-NEXT:    [[TMP1:%.*]] = zext i16 [[P:%.*]] to i32 | 
 | // CHECK-NEXT:    [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.pred.i2v.v4i1(i32 [[TMP1]]) | 
 | // CHECK-NEXT:    call void @llvm.masked.store.v4i32.p0(<4 x i32> [[VALUE:%.*]], ptr [[BASE:%.*]], i32 4, <4 x i1> [[TMP2]]) | 
 | // CHECK-NEXT:    ret void | 
 | // | 
 | void test_vstrwq_p_u32(uint32_t *base, uint32x4_t value, mve_pred16_t p) | 
 | { | 
 | #ifdef POLYMORPHIC | 
 |     vstrwq_p(base, value, p); | 
 | #else /* POLYMORPHIC */ | 
 |     vstrwq_p_u32(base, value, p); | 
 | #endif /* POLYMORPHIC */ | 
 | } |