f4a2713ac8
Change-Id: Ia40e9ffdf29b5dab2f122f673ff6802a58bc690f
283 lines
8 KiB
C
283 lines
8 KiB
C
// REQUIRES: aarch64-registered-target
|
|
// RUN: %clang_cc1 -triple aarch64-none-linux-gnu -target-feature +neon \
|
|
// RUN: -ffp-contract=fast -S -O3 -o - %s | FileCheck %s
|
|
|
|
// Test new aarch64 intrinsics with poly64
|
|
|
|
#include <arm_neon.h>
|
|
|
|
uint64x1_t test_vceq_p64(poly64x1_t a, poly64x1_t b) {
|
|
// CHECK: test_vceq_p64
|
|
return vceq_p64(a, b);
|
|
// CHECK: cmeq {{d[0-9]+}}, {{d[0-9]+}}, {{d[0-9]+}}
|
|
}
|
|
|
|
uint64x2_t test_vceqq_p64(poly64x2_t a, poly64x2_t b) {
|
|
// CHECK: test_vceqq_p64
|
|
return vceqq_p64(a, b);
|
|
// CHECK: cmeq {{v[0-9]+}}.2d, {{v[0-9]+}}.2d, {{v[0-9]+}}.2d
|
|
}
|
|
|
|
uint64x1_t test_vtst_p64(poly64x1_t a, poly64x1_t b) {
|
|
// CHECK: test_vtst_p64
|
|
return vtst_p64(a, b);
|
|
// CHECK: cmtst {{d[0-9]+}}, {{d[0-9]+}}, {{d[0-9]+}}
|
|
}
|
|
|
|
uint64x2_t test_vtstq_p64(poly64x2_t a, poly64x2_t b) {
|
|
// CHECK: test_vtstq_p64
|
|
return vtstq_p64(a, b);
|
|
// CHECK: cmtst {{v[0-9]+}}.2d, {{v[0-9]+}}.2d, {{v[0-9]+}}.2d
|
|
}
|
|
|
|
poly64x1_t test_vbsl_p64(poly64x1_t a, poly64x1_t b, poly64x1_t c) {
|
|
// CHECK: test_vbsl_p64
|
|
return vbsl_p64(a, b, c);
|
|
// CHECK: bsl {{v[0-9]+}}.8b, {{v[0-9]+}}.8b, {{v[0-9]+}}.8b
|
|
}
|
|
|
|
poly64x2_t test_vbslq_p64(poly64x2_t a, poly64x2_t b, poly64x2_t c) {
|
|
// CHECK: test_vbslq_p64
|
|
return vbslq_p64(a, b, c);
|
|
// CHECK: bsl {{v[0-9]+}}.16b, {{v[0-9]+}}.16b, {{v[0-9]+}}.16b
|
|
}
|
|
|
|
poly64_t test_vget_lane_p64(poly64x1_t v) {
|
|
// CHECK: test_vget_lane_p64
|
|
return vget_lane_p64(v, 0);
|
|
// CHECK: fmov {{x[0-9]+}}, {{d[0-9]+}}
|
|
}
|
|
|
|
poly64_t test_vgetq_lane_p64(poly64x2_t v) {
|
|
// CHECK: test_vgetq_lane_p64
|
|
return vgetq_lane_p64(v, 1);
|
|
// CHECK: umov {{x[0-9]+}}, {{v[0-9]+}}.d[1]
|
|
}
|
|
|
|
poly64x1_t test_vset_lane_p64(poly64_t a, poly64x1_t v) {
|
|
// CHECK: test_vset_lane_p64
|
|
return vset_lane_p64(a, v, 0);
|
|
// CHECK: fmov {{d[0-9]+}}, {{x[0-9]+}}
|
|
}
|
|
|
|
poly64x2_t test_vsetq_lane_p64(poly64_t a, poly64x2_t v) {
|
|
// CHECK: test_vsetq_lane_p64
|
|
return vsetq_lane_p64(a, v, 1);
|
|
// CHECK: ins {{v[0-9]+}}.d[1], {{x[0-9]+}}
|
|
}
|
|
|
|
poly64x1_t test_vcopy_lane_p64(poly64x1_t a, poly64x1_t b) {
|
|
// CHECK: test_vcopy_lane_p64
|
|
return vcopy_lane_p64(a, 0, b, 0);
|
|
// CHECK: fmov {{d[0-9]+}}, {{d[0-9]+}}
|
|
}
|
|
|
|
poly64x2_t test_vcopyq_lane_p64(poly64x2_t a, poly64x1_t b) {
|
|
// CHECK: test_vcopyq_lane_p64
|
|
return vcopyq_lane_p64(a, 1, b, 0);
|
|
// CHECK: ins {{v[0-9]+}}.d[1], {{v[0-9]+}}.d[0]
|
|
}
|
|
|
|
poly64x2_t test_vcopyq_laneq_p64(poly64x2_t a, poly64x2_t b) {
|
|
// CHECK: test_vcopyq_laneq_p64
|
|
return vcopyq_laneq_p64(a, 1, b, 1);
|
|
// CHECK: ins {{v[0-9]+}}.d[1], {{v[0-9]+}}.d[1]
|
|
}
|
|
|
|
poly64x1_t test_vcreate_p64(uint64_t a) {
|
|
// CHECK: test_vcreate_p64
|
|
return vcreate_p64(a);
|
|
// CHECK: fmov {{d[0-9]+}}, {{x[0-9]+}}
|
|
}
|
|
|
|
poly64x1_t test_vdup_n_p64(poly64_t a) {
|
|
// CHECK: test_vdup_n_p64
|
|
return vdup_n_p64(a);
|
|
// CHECK: fmov {{d[0-9]+}}, {{x[0-9]+}}
|
|
}
|
|
poly64x2_t test_vdupq_n_p64(poly64_t a) {
|
|
// CHECK: test_vdup_n_p64
|
|
return vdupq_n_p64(a);
|
|
// CHECK: dup {{v[0-9]+}}.2d, {{x[0-9]+}}
|
|
}
|
|
|
|
poly64x1_t test_vdup_lane_p64(poly64x1_t vec) {
|
|
// CHECK: test_vdup_lane_p64
|
|
return vdup_lane_p64(vec, 0);
|
|
// CHECK: ret
|
|
}
|
|
|
|
poly64x2_t test_vdupq_lane_p64(poly64x1_t vec) {
|
|
// CHECK: test_vdupq_lane_p64
|
|
return vdupq_lane_p64(vec, 0);
|
|
// CHECK: dup {{v[0-9]+}}.2d, {{v[0-9]+}}.d[0]
|
|
}
|
|
|
|
poly64x2_t test_vdupq_laneq_p64(poly64x2_t vec) {
|
|
// CHECK: test_vdupq_laneq_p64
|
|
return vdupq_laneq_p64(vec, 1);
|
|
// CHECK: dup {{v[0-9]+}}.2d, {{v[0-9]+}}.d[1]
|
|
}
|
|
|
|
poly64x2_t test_vcombine_p64(poly64x1_t low, poly64x1_t high) {
|
|
// CHECK: test_vcombine_p64
|
|
return vcombine_p64(low, high);
|
|
// CHECK: ins {{v[0-9]+}}.d[1], {{v[0-9]+}}.d[0]
|
|
}
|
|
|
|
poly64x1_t test_vld1_p64(poly64_t const * ptr) {
|
|
// CHECK: test_vld1_p64
|
|
return vld1_p64(ptr);
|
|
// CHECK: ld1 {{{v[0-9]+}}.1d}, [{{x[0-9]+|sp}}]
|
|
}
|
|
|
|
poly64x2_t test_vld1q_p64(poly64_t const * ptr) {
|
|
// CHECK: test_vld1q_p64
|
|
return vld1q_p64(ptr);
|
|
// CHECK: ld1 {{{v[0-9]+}}.2d}, [{{x[0-9]+|sp}}]
|
|
}
|
|
|
|
void test_vst1_p64(poly64_t * ptr, poly64x1_t val) {
|
|
// CHECK: test_vst1_p64
|
|
return vst1_p64(ptr, val);
|
|
// CHECK: st1 {{{v[0-9]+}}.1d}, [{{x[0-9]+|sp}}]
|
|
}
|
|
|
|
void test_vst1q_p64(poly64_t * ptr, poly64x2_t val) {
|
|
// CHECK: test_vst1q_p64
|
|
return vst1q_p64(ptr, val);
|
|
// CHECK: st1 {{{v[0-9]+}}.2d}, [{{x[0-9]+|sp}}]
|
|
}
|
|
|
|
poly64x1x2_t test_vld2_p64(poly64_t const * ptr) {
|
|
// CHECK: test_vld2_p64
|
|
return vld2_p64(ptr);
|
|
// CHECK: ld1 {{{v[0-9]+}}.1d, {{v[0-9]+}}.1d}, [{{x[0-9]+|sp}}]
|
|
}
|
|
|
|
poly64x2x2_t test_vld2q_p64(poly64_t const * ptr) {
|
|
// CHECK: test_vld2q_p64
|
|
return vld2q_p64(ptr);
|
|
// CHECK: ld2 {{{v[0-9]+}}.2d, {{v[0-9]+}}.2d}, [{{x[0-9]+|sp}}]
|
|
}
|
|
|
|
poly64x1x3_t test_vld3_p64(poly64_t const * ptr) {
|
|
// CHECK: test_vld3_p64
|
|
return vld3_p64(ptr);
|
|
// CHECK: ld1 {{{v[0-9]+}}.1d, {{v[0-9]+}}.1d, {{v[0-9]+}}.1d}, [{{x[0-9]+|sp}}]
|
|
}
|
|
|
|
poly64x2x3_t test_vld3q_p64(poly64_t const * ptr) {
|
|
// CHECK: test_vld3q_p64
|
|
return vld3q_p64(ptr);
|
|
// CHECK: ld3 {{{v[0-9]+}}.2d, {{v[0-9]+}}.2d, {{v[0-9]+}}.2d}, [{{x[0-9]+|sp}}]
|
|
}
|
|
|
|
poly64x1x4_t test_vld4_p64(poly64_t const * ptr) {
|
|
// CHECK: test_vld4_p64
|
|
return vld4_p64(ptr);
|
|
// CHECK: ld1 {{{v[0-9]+}}.1d, {{v[0-9]+}}.1d, {{v[0-9]+}}.1d, {{v[0-9]+}}.1d}, [{{x[0-9]+|sp}}]
|
|
}
|
|
|
|
poly64x2x4_t test_vld4q_p64(poly64_t const * ptr) {
|
|
// CHECK: test_vld4q_p64
|
|
return vld4q_p64(ptr);
|
|
// CHECK: ld4 {{{v[0-9]+}}.2d, {{v[0-9]+}}.2d, {{v[0-9]+}}.2d, {{v[0-9]+}}.2d}, [{{x[0-9]+|sp}}]
|
|
}
|
|
|
|
void test_vst2_p64(poly64_t * ptr, poly64x1x2_t val) {
|
|
// CHECK: test_vst2_p64
|
|
return vst2_p64(ptr, val);
|
|
// CHECK: st1 {{{v[0-9]+}}.1d, {{v[0-9]+}}.1d}, [{{x[0-9]+|sp}}]
|
|
}
|
|
|
|
void test_vst2q_p64(poly64_t * ptr, poly64x2x2_t val) {
|
|
// CHECK: test_vst2q_p64
|
|
return vst2q_p64(ptr, val);
|
|
// CHECK: st2 {{{v[0-9]+}}.2d, {{v[0-9]+}}.2d}, [{{x[0-9]+|sp}}]
|
|
}
|
|
|
|
void test_vst3_p64(poly64_t * ptr, poly64x1x3_t val) {
|
|
// CHECK: test_vst3_p64
|
|
return vst3_p64(ptr, val);
|
|
// CHECK: st1 {{{v[0-9]+}}.1d, {{v[0-9]+}}.1d, {{v[0-9]+}}.1d}, [{{x[0-9]+|sp}}]
|
|
}
|
|
|
|
void test_vst3q_p64(poly64_t * ptr, poly64x2x3_t val) {
|
|
// CHECK: test_vst3q_p64
|
|
return vst3q_p64(ptr, val);
|
|
// CHECK: st3 {{{v[0-9]+}}.2d, {{v[0-9]+}}.2d, {{v[0-9]+}}.2d}, [{{x[0-9]+|sp}}]
|
|
}
|
|
|
|
void test_vst4_p64(poly64_t * ptr, poly64x1x4_t val) {
|
|
// CHECK: test_vst4_p64
|
|
return vst4_p64(ptr, val);
|
|
// CHECK: st1 {{{v[0-9]+}}.1d, {{v[0-9]+}}.1d, {{v[0-9]+}}.1d, {{v[0-9]+}}.1d}, [{{x[0-9]+|sp}}]
|
|
}
|
|
|
|
void test_vst4q_p64(poly64_t * ptr, poly64x2x4_t val) {
|
|
// CHECK: test_vst4q_p64
|
|
return vst4q_p64(ptr, val);
|
|
// CHECK: st4 {{{v[0-9]+}}.2d, {{v[0-9]+}}.2d, {{v[0-9]+}}.2d, {{v[0-9]+}}.2d}, [{{x[0-9]+|sp}}]
|
|
}
|
|
|
|
poly64x1_t test_vext_p64(poly64x1_t a, poly64x1_t b) {
|
|
// CHECK: test_vext_p64
|
|
return vext_u64(a, b, 0);
|
|
|
|
}
|
|
|
|
poly64x2_t test_vextq_p64(poly64x2_t a, poly64x2_t b) {
|
|
// CHECK: test_vextq_p64
|
|
return vextq_p64(a, b, 1);
|
|
// CHECK: ext {{v[0-9]+}}.16b, {{v[0-9]+}}.16b, {{v[0-9]+}}.16b, #0x8
|
|
}
|
|
|
|
poly64x2_t test_vzip1q_p64(poly64x2_t a, poly64x2_t b) {
|
|
// CHECK: test_vzip1q_p64
|
|
return vzip1q_p64(a, b);
|
|
// CHECK: ins {{v[0-9]+}}.d[1], {{v[0-9]+}}.d[0]
|
|
}
|
|
|
|
poly64x2_t test_vzip2q_p64(poly64x2_t a, poly64x2_t b) {
|
|
// CHECK: test_vzip2q_p64
|
|
return vzip2q_u64(a, b);
|
|
// CHECK: ins {{v[0-9]+}}.d[0], {{v[0-9]+}}.d[1]
|
|
}
|
|
|
|
poly64x2_t test_vuzp1q_p64(poly64x2_t a, poly64x2_t b) {
|
|
// CHECK: test_vuzp1q_p64
|
|
return vuzp1q_p64(a, b);
|
|
// CHECK: ins {{v[0-9]+}}.d[1], {{v[0-9]+}}.d[0]
|
|
}
|
|
|
|
poly64x2_t test_vuzp2q_p64(poly64x2_t a, poly64x2_t b) {
|
|
// CHECK: test_vuzp2q_p64
|
|
return vuzp2q_u64(a, b);
|
|
// CHECK: ins {{v[0-9]+}}.d[0], {{v[0-9]+}}.d[1]
|
|
}
|
|
|
|
poly64x2_t test_vtrn1q_p64(poly64x2_t a, poly64x2_t b) {
|
|
// CHECK: test_vtrn1q_p64
|
|
return vtrn1q_p64(a, b);
|
|
// CHECK: ins {{v[0-9]+}}.d[1], {{v[0-9]+}}.d[0]
|
|
}
|
|
|
|
poly64x2_t test_vtrn2q_p64(poly64x2_t a, poly64x2_t b) {
|
|
// CHECK: test_vtrn2q_p64
|
|
return vtrn2q_u64(a, b);
|
|
// CHECK: ins {{v[0-9]+}}.d[0], {{v[0-9]+}}.d[1]
|
|
}
|
|
|
|
poly64x1_t test_vsri_n_p64(poly64x1_t a, poly64x1_t b) {
|
|
// CHECK: test_vsri_n_p64
|
|
return vsri_n_p64(a, b, 33);
|
|
// CHECK: sri {{d[0-9]+}}, {{d[0-9]+}}, #33
|
|
}
|
|
|
|
poly64x2_t test_vsriq_n_p64(poly64x2_t a, poly64x2_t b) {
|
|
// CHECK: test_vsriq_n_p64
|
|
return vsriq_n_p64(a, b, 64);
|
|
// CHECK: sri {{v[0-9]+}}.2d, {{v[0-9]+}}.2d, #64
|
|
}
|
|
|