From 5366d0e0bc49c6cedf105ecb9b03eea75df46f4e Mon Sep 17 00:00:00 2001 From: Simon Pilgrim Date: Tue, 19 Jul 2016 17:04:28 +0000 Subject: [PATCH] [X86][AVX512] Added AVX512 subvector broadcast tests llvm-svn: 275994 --- llvm/test/CodeGen/X86/avx512-vbroadcasti128.ll | 326 +++++++++++++++++++++++++ llvm/test/CodeGen/X86/avx512-vbroadcasti256.ll | 138 +++++++++++ 2 files changed, 464 insertions(+) create mode 100644 llvm/test/CodeGen/X86/avx512-vbroadcasti128.ll create mode 100644 llvm/test/CodeGen/X86/avx512-vbroadcasti256.ll diff --git a/llvm/test/CodeGen/X86/avx512-vbroadcasti128.ll b/llvm/test/CodeGen/X86/avx512-vbroadcasti128.ll new file mode 100644 index 0000000..e2460d2 --- /dev/null +++ b/llvm/test/CodeGen/X86/avx512-vbroadcasti128.ll @@ -0,0 +1,326 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc < %s -mtriple=x86_64-apple-darwin -mattr=+avx512vl | FileCheck %s --check-prefix=X64-AVX512 --check-prefix=X64-AVX512VL +; RUN: llc < %s -mtriple=x86_64-apple-darwin -mattr=+avx512bw,+avx512vl | FileCheck %s --check-prefix=X64-AVX512 --check-prefix=X64-AVX512BWVL +; RUN: llc < %s -mtriple=x86_64-apple-darwin -mattr=+avx512dq,+avx512vl | FileCheck %s --check-prefix=X64-AVX512 --check-prefix=X64-AVX512DQVL + +; +; 128-bit Subvector Broadcast to 256-bit +; + +define <4 x double> @test_broadcast_2f64_4f64(<2 x double> *%p) nounwind { +; X64-AVX512VL-LABEL: test_broadcast_2f64_4f64: +; X64-AVX512VL: ## BB#0: +; X64-AVX512VL-NEXT: vmovapd (%rdi), %xmm0 +; X64-AVX512VL-NEXT: vinsertf32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512VL-NEXT: vaddpd {{.*}}(%rip), %ymm0, %ymm0 +; X64-AVX512VL-NEXT: retq +; +; X64-AVX512BWVL-LABEL: test_broadcast_2f64_4f64: +; X64-AVX512BWVL: ## BB#0: +; X64-AVX512BWVL-NEXT: vmovapd (%rdi), %xmm0 +; X64-AVX512BWVL-NEXT: vinsertf32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512BWVL-NEXT: vaddpd {{.*}}(%rip), %ymm0, %ymm0 +; X64-AVX512BWVL-NEXT: retq +; +; X64-AVX512DQVL-LABEL: test_broadcast_2f64_4f64: +; X64-AVX512DQVL: ## BB#0: +; X64-AVX512DQVL-NEXT: vmovapd (%rdi), %xmm0 +; X64-AVX512DQVL-NEXT: vinsertf64x2 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512DQVL-NEXT: vaddpd {{.*}}(%rip), %ymm0, %ymm0 +; X64-AVX512DQVL-NEXT: retq + %1 = load <2 x double>, <2 x double> *%p + %2 = shufflevector <2 x double> %1, <2 x double> undef, <4 x i32> + %3 = fadd <4 x double> %2, + ret <4 x double> %3 +} + +define <4 x i64> @test_broadcast_2i64_4i64(<2 x i64> *%p) nounwind { +; X64-AVX512VL-LABEL: test_broadcast_2i64_4i64: +; X64-AVX512VL: ## BB#0: +; X64-AVX512VL-NEXT: vmovdqa64 (%rdi), %xmm0 +; X64-AVX512VL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512VL-NEXT: vpaddq {{.*}}(%rip), %ymm0, %ymm0 +; X64-AVX512VL-NEXT: retq +; +; X64-AVX512BWVL-LABEL: test_broadcast_2i64_4i64: +; X64-AVX512BWVL: ## BB#0: +; X64-AVX512BWVL-NEXT: vmovdqa64 (%rdi), %xmm0 +; X64-AVX512BWVL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512BWVL-NEXT: vpaddq {{.*}}(%rip), %ymm0, %ymm0 +; X64-AVX512BWVL-NEXT: retq +; +; X64-AVX512DQVL-LABEL: test_broadcast_2i64_4i64: +; X64-AVX512DQVL: ## BB#0: +; X64-AVX512DQVL-NEXT: vmovdqa64 (%rdi), %xmm0 +; X64-AVX512DQVL-NEXT: vinserti64x2 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512DQVL-NEXT: vpaddq {{.*}}(%rip), %ymm0, %ymm0 +; X64-AVX512DQVL-NEXT: retq + %1 = load <2 x i64>, <2 x i64> *%p + %2 = shufflevector <2 x i64> %1, <2 x i64> undef, <4 x i32> + %3 = add <4 x i64> %2, + ret <4 x i64> %3 +} + +define <8 x float> @test_broadcast_4f32_8f32(<4 x float> *%p) nounwind { +; X64-AVX512-LABEL: test_broadcast_4f32_8f32: +; X64-AVX512: ## BB#0: +; X64-AVX512-NEXT: vmovaps (%rdi), %xmm0 +; X64-AVX512-NEXT: vinsertf32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512-NEXT: vaddps {{.*}}(%rip), %ymm0, %ymm0 +; X64-AVX512-NEXT: retq + %1 = load <4 x float>, <4 x float> *%p + %2 = shufflevector <4 x float> %1, <4 x float> undef, <8 x i32> + %3 = fadd <8 x float> %2, + ret <8 x float> %3 +} + +define <8 x i32> @test_broadcast_4i32_8i32(<4 x i32> *%p) nounwind { +; X64-AVX512-LABEL: test_broadcast_4i32_8i32: +; X64-AVX512: ## BB#0: +; X64-AVX512-NEXT: vmovdqa32 (%rdi), %xmm0 +; X64-AVX512-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512-NEXT: vpaddd {{.*}}(%rip), %ymm0, %ymm0 +; X64-AVX512-NEXT: retq + %1 = load <4 x i32>, <4 x i32> *%p + %2 = shufflevector <4 x i32> %1, <4 x i32> undef, <8 x i32> + %3 = add <8 x i32> %2, + ret <8 x i32> %3 +} + +define <16 x i16> @test_broadcast_8i16_16i16(<8 x i16> *%p) nounwind { +; X64-AVX512VL-LABEL: test_broadcast_8i16_16i16: +; X64-AVX512VL: ## BB#0: +; X64-AVX512VL-NEXT: vmovdqa64 (%rdi), %xmm0 +; X64-AVX512VL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512VL-NEXT: vpaddw {{.*}}(%rip), %ymm0, %ymm0 +; X64-AVX512VL-NEXT: retq +; +; X64-AVX512BWVL-LABEL: test_broadcast_8i16_16i16: +; X64-AVX512BWVL: ## BB#0: +; X64-AVX512BWVL-NEXT: vmovdqu16 (%rdi), %xmm0 +; X64-AVX512BWVL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512BWVL-NEXT: vpaddw {{.*}}(%rip), %ymm0, %ymm0 +; X64-AVX512BWVL-NEXT: retq +; +; X64-AVX512DQVL-LABEL: test_broadcast_8i16_16i16: +; X64-AVX512DQVL: ## BB#0: +; X64-AVX512DQVL-NEXT: vmovdqa64 (%rdi), %xmm0 +; X64-AVX512DQVL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512DQVL-NEXT: vpaddw {{.*}}(%rip), %ymm0, %ymm0 +; X64-AVX512DQVL-NEXT: retq + %1 = load <8 x i16>, <8 x i16> *%p + %2 = shufflevector <8 x i16> %1, <8 x i16> undef, <16 x i32> + %3 = add <16 x i16> %2, + ret <16 x i16> %3 +} + +define <32 x i8> @test_broadcast_16i8_32i8(<16 x i8> *%p) nounwind { +; X64-AVX512VL-LABEL: test_broadcast_16i8_32i8: +; X64-AVX512VL: ## BB#0: +; X64-AVX512VL-NEXT: vmovdqa64 (%rdi), %xmm0 +; X64-AVX512VL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512VL-NEXT: vpaddb {{.*}}(%rip), %ymm0, %ymm0 +; X64-AVX512VL-NEXT: retq +; +; X64-AVX512BWVL-LABEL: test_broadcast_16i8_32i8: +; X64-AVX512BWVL: ## BB#0: +; X64-AVX512BWVL-NEXT: vmovdqu8 (%rdi), %xmm0 +; X64-AVX512BWVL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512BWVL-NEXT: vpaddb {{.*}}(%rip), %ymm0, %ymm0 +; X64-AVX512BWVL-NEXT: retq +; +; X64-AVX512DQVL-LABEL: test_broadcast_16i8_32i8: +; X64-AVX512DQVL: ## BB#0: +; X64-AVX512DQVL-NEXT: vmovdqa64 (%rdi), %xmm0 +; X64-AVX512DQVL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512DQVL-NEXT: vpaddb {{.*}}(%rip), %ymm0, %ymm0 +; X64-AVX512DQVL-NEXT: retq + %1 = load <16 x i8>, <16 x i8> *%p + %2 = shufflevector <16 x i8> %1, <16 x i8> undef, <32 x i32> + %3 = add <32 x i8> %2, + ret <32 x i8> %3 +} + +; +; 128-bit Subvector Broadcast to 512-bit +; + +define <8 x double> @test_broadcast_2f64_8f64(<2 x double> *%p) nounwind { +; X64-AVX512VL-LABEL: test_broadcast_2f64_8f64: +; X64-AVX512VL: ## BB#0: +; X64-AVX512VL-NEXT: vmovapd (%rdi), %xmm0 +; X64-AVX512VL-NEXT: vinsertf32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512VL-NEXT: vinsertf64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512VL-NEXT: vaddpd {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512VL-NEXT: retq +; +; X64-AVX512BWVL-LABEL: test_broadcast_2f64_8f64: +; X64-AVX512BWVL: ## BB#0: +; X64-AVX512BWVL-NEXT: vmovapd (%rdi), %xmm0 +; X64-AVX512BWVL-NEXT: vinsertf32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512BWVL-NEXT: vinsertf64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: vaddpd {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: retq +; +; X64-AVX512DQVL-LABEL: test_broadcast_2f64_8f64: +; X64-AVX512DQVL: ## BB#0: +; X64-AVX512DQVL-NEXT: vmovapd (%rdi), %xmm0 +; X64-AVX512DQVL-NEXT: vinsertf64x2 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512DQVL-NEXT: vinsertf64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512DQVL-NEXT: vaddpd {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512DQVL-NEXT: retq + %1 = load <2 x double>, <2 x double> *%p + %2 = shufflevector <2 x double> %1, <2 x double> undef, <8 x i32> + %3 = fadd <8 x double> %2, + ret <8 x double> %3 +} + +define <8 x i64> @test_broadcast_2i64_8i64(<2 x i64> *%p) nounwind { +; X64-AVX512VL-LABEL: test_broadcast_2i64_8i64: +; X64-AVX512VL: ## BB#0: +; X64-AVX512VL-NEXT: vmovdqa64 (%rdi), %xmm0 +; X64-AVX512VL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512VL-NEXT: vinserti64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512VL-NEXT: vpaddq {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512VL-NEXT: retq +; +; X64-AVX512BWVL-LABEL: test_broadcast_2i64_8i64: +; X64-AVX512BWVL: ## BB#0: +; X64-AVX512BWVL-NEXT: vmovdqa64 (%rdi), %xmm0 +; X64-AVX512BWVL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512BWVL-NEXT: vinserti64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: vpaddq {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: retq +; +; X64-AVX512DQVL-LABEL: test_broadcast_2i64_8i64: +; X64-AVX512DQVL: ## BB#0: +; X64-AVX512DQVL-NEXT: vmovdqa64 (%rdi), %xmm0 +; X64-AVX512DQVL-NEXT: vinserti64x2 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512DQVL-NEXT: vinserti64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512DQVL-NEXT: vpaddq {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512DQVL-NEXT: retq + %1 = load <2 x i64>, <2 x i64> *%p + %2 = shufflevector <2 x i64> %1, <2 x i64> undef, <8 x i32> + %3 = add <8 x i64> %2, + ret <8 x i64> %3 +} + +define <16 x float> @test_broadcast_4f32_16f32(<4 x float> *%p) nounwind { +; X64-AVX512VL-LABEL: test_broadcast_4f32_16f32: +; X64-AVX512VL: ## BB#0: +; X64-AVX512VL-NEXT: vmovaps (%rdi), %xmm0 +; X64-AVX512VL-NEXT: vinsertf32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512VL-NEXT: vinsertf64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512VL-NEXT: vaddps {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512VL-NEXT: retq +; +; X64-AVX512BWVL-LABEL: test_broadcast_4f32_16f32: +; X64-AVX512BWVL: ## BB#0: +; X64-AVX512BWVL-NEXT: vmovaps (%rdi), %xmm0 +; X64-AVX512BWVL-NEXT: vinsertf32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512BWVL-NEXT: vinsertf64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: vaddps {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: retq +; +; X64-AVX512DQVL-LABEL: test_broadcast_4f32_16f32: +; X64-AVX512DQVL: ## BB#0: +; X64-AVX512DQVL-NEXT: vmovaps (%rdi), %xmm0 +; X64-AVX512DQVL-NEXT: vinsertf32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512DQVL-NEXT: vinsertf32x8 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512DQVL-NEXT: vaddps {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512DQVL-NEXT: retq + %1 = load <4 x float>, <4 x float> *%p + %2 = shufflevector <4 x float> %1, <4 x float> undef, <16 x i32> + %3 = fadd <16 x float> %2, + ret <16 x float> %3 +} + +define <16 x i32> @test_broadcast_4i32_16i32(<4 x i32> *%p) nounwind { +; X64-AVX512VL-LABEL: test_broadcast_4i32_16i32: +; X64-AVX512VL: ## BB#0: +; X64-AVX512VL-NEXT: vmovdqa32 (%rdi), %xmm0 +; X64-AVX512VL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512VL-NEXT: vinserti64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512VL-NEXT: vpaddd {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512VL-NEXT: retq +; +; X64-AVX512BWVL-LABEL: test_broadcast_4i32_16i32: +; X64-AVX512BWVL: ## BB#0: +; X64-AVX512BWVL-NEXT: vmovdqa32 (%rdi), %xmm0 +; X64-AVX512BWVL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512BWVL-NEXT: vinserti64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: vpaddd {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: retq +; +; X64-AVX512DQVL-LABEL: test_broadcast_4i32_16i32: +; X64-AVX512DQVL: ## BB#0: +; X64-AVX512DQVL-NEXT: vmovdqa32 (%rdi), %xmm0 +; X64-AVX512DQVL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512DQVL-NEXT: vinserti32x8 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512DQVL-NEXT: vpaddd {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512DQVL-NEXT: retq + %1 = load <4 x i32>, <4 x i32> *%p + %2 = shufflevector <4 x i32> %1, <4 x i32> undef, <16 x i32> + %3 = add <16 x i32> %2, + ret <16 x i32> %3 +} + +define <32 x i16> @test_broadcast_8i16_32i16(<8 x i16> *%p) nounwind { +; X64-AVX512VL-LABEL: test_broadcast_8i16_32i16: +; X64-AVX512VL: ## BB#0: +; X64-AVX512VL-NEXT: vmovdqa64 (%rdi), %xmm0 +; X64-AVX512VL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm1 +; X64-AVX512VL-NEXT: vpaddw {{.*}}(%rip), %ymm1, %ymm0 +; X64-AVX512VL-NEXT: vpaddw {{.*}}(%rip), %ymm1, %ymm1 +; X64-AVX512VL-NEXT: retq +; +; X64-AVX512BWVL-LABEL: test_broadcast_8i16_32i16: +; X64-AVX512BWVL: ## BB#0: +; X64-AVX512BWVL-NEXT: vmovdqu16 (%rdi), %xmm0 +; X64-AVX512BWVL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512BWVL-NEXT: vinserti64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: vpaddw {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: retq +; +; X64-AVX512DQVL-LABEL: test_broadcast_8i16_32i16: +; X64-AVX512DQVL: ## BB#0: +; X64-AVX512DQVL-NEXT: vmovdqa64 (%rdi), %xmm0 +; X64-AVX512DQVL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm1 +; X64-AVX512DQVL-NEXT: vpaddw {{.*}}(%rip), %ymm1, %ymm0 +; X64-AVX512DQVL-NEXT: vpaddw {{.*}}(%rip), %ymm1, %ymm1 +; X64-AVX512DQVL-NEXT: retq + %1 = load <8 x i16>, <8 x i16> *%p + %2 = shufflevector <8 x i16> %1, <8 x i16> undef, <32 x i32> + %3 = add <32 x i16> %2, + ret <32 x i16> %3 +} + +define <64 x i8> @test_broadcast_16i8_64i8(<16 x i8> *%p) nounwind { +; X64-AVX512VL-LABEL: test_broadcast_16i8_64i8: +; X64-AVX512VL: ## BB#0: +; X64-AVX512VL-NEXT: vmovdqa64 (%rdi), %xmm0 +; X64-AVX512VL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm1 +; X64-AVX512VL-NEXT: vpaddb {{.*}}(%rip), %ymm1, %ymm0 +; X64-AVX512VL-NEXT: vpaddb {{.*}}(%rip), %ymm1, %ymm1 +; X64-AVX512VL-NEXT: retq +; +; X64-AVX512BWVL-LABEL: test_broadcast_16i8_64i8: +; X64-AVX512BWVL: ## BB#0: +; X64-AVX512BWVL-NEXT: vmovdqu8 (%rdi), %xmm0 +; X64-AVX512BWVL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0 +; X64-AVX512BWVL-NEXT: vinserti64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: vpaddb {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: retq +; +; X64-AVX512DQVL-LABEL: test_broadcast_16i8_64i8: +; X64-AVX512DQVL: ## BB#0: +; X64-AVX512DQVL-NEXT: vmovdqa64 (%rdi), %xmm0 +; X64-AVX512DQVL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm1 +; X64-AVX512DQVL-NEXT: vpaddb {{.*}}(%rip), %ymm1, %ymm0 +; X64-AVX512DQVL-NEXT: vpaddb {{.*}}(%rip), %ymm1, %ymm1 +; X64-AVX512DQVL-NEXT: retq + %1 = load <16 x i8>, <16 x i8> *%p + %2 = shufflevector <16 x i8> %1, <16 x i8> undef, <64 x i32> + %3 = add <64 x i8> %2, + ret <64 x i8> %3 +} diff --git a/llvm/test/CodeGen/X86/avx512-vbroadcasti256.ll b/llvm/test/CodeGen/X86/avx512-vbroadcasti256.ll new file mode 100644 index 0000000..4a9e359 --- /dev/null +++ b/llvm/test/CodeGen/X86/avx512-vbroadcasti256.ll @@ -0,0 +1,138 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc < %s -mtriple=x86_64-apple-darwin -mattr=+avx512vl | FileCheck %s --check-prefix=X64-AVX512 --check-prefix=X64-AVX512VL +; RUN: llc < %s -mtriple=x86_64-apple-darwin -mattr=+avx512bw,+avx512vl | FileCheck %s --check-prefix=X64-AVX512 --check-prefix=X64-AVX512BWVL +; RUN: llc < %s -mtriple=x86_64-apple-darwin -mattr=+avx512dq,+avx512vl | FileCheck %s --check-prefix=X64-AVX512 --check-prefix=X64-AVX512DQVL + +define <8 x double> @test_broadcast_4f64_8f64(<4 x double> *%p) nounwind { +; X64-AVX512-LABEL: test_broadcast_4f64_8f64: +; X64-AVX512: ## BB#0: +; X64-AVX512-NEXT: vmovapd (%rdi), %ymm0 +; X64-AVX512-NEXT: vinsertf64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512-NEXT: vaddpd {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512-NEXT: retq + %1 = load <4 x double>, <4 x double> *%p + %2 = shufflevector <4 x double> %1, <4 x double> undef, <8 x i32> + %3 = fadd <8 x double> %2, + ret <8 x double> %3 +} + +define <8 x i64> @test_broadcast_4i64_8i64(<4 x i64> *%p) nounwind { +; X64-AVX512-LABEL: test_broadcast_4i64_8i64: +; X64-AVX512: ## BB#0: +; X64-AVX512-NEXT: vmovdqa64 (%rdi), %ymm0 +; X64-AVX512-NEXT: vinserti64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512-NEXT: vpaddq {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512-NEXT: retq + %1 = load <4 x i64>, <4 x i64> *%p + %2 = shufflevector <4 x i64> %1, <4 x i64> undef, <8 x i32> + %3 = add <8 x i64> %2, + ret <8 x i64> %3 +} + +define <16 x float> @test_broadcast_8f32_16f32(<8 x float> *%p) nounwind { +; X64-AVX512VL-LABEL: test_broadcast_8f32_16f32: +; X64-AVX512VL: ## BB#0: +; X64-AVX512VL-NEXT: vmovaps (%rdi), %ymm0 +; X64-AVX512VL-NEXT: vinsertf64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512VL-NEXT: vaddps {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512VL-NEXT: retq +; +; X64-AVX512BWVL-LABEL: test_broadcast_8f32_16f32: +; X64-AVX512BWVL: ## BB#0: +; X64-AVX512BWVL-NEXT: vmovaps (%rdi), %ymm0 +; X64-AVX512BWVL-NEXT: vinsertf64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: vaddps {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: retq +; +; X64-AVX512DQVL-LABEL: test_broadcast_8f32_16f32: +; X64-AVX512DQVL: ## BB#0: +; X64-AVX512DQVL-NEXT: vmovaps (%rdi), %ymm0 +; X64-AVX512DQVL-NEXT: vinsertf32x8 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512DQVL-NEXT: vaddps {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512DQVL-NEXT: retq + %1 = load <8 x float>, <8 x float> *%p + %2 = shufflevector <8 x float> %1, <8 x float> undef, <16 x i32> + %3 = fadd <16 x float> %2, + ret <16 x float> %3 +} + +define <16 x i32> @test_broadcast_8i32_16i32(<8 x i32> *%p) nounwind { +; X64-AVX512VL-LABEL: test_broadcast_8i32_16i32: +; X64-AVX512VL: ## BB#0: +; X64-AVX512VL-NEXT: vmovdqa32 (%rdi), %ymm0 +; X64-AVX512VL-NEXT: vinserti64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512VL-NEXT: vpaddd {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512VL-NEXT: retq +; +; X64-AVX512BWVL-LABEL: test_broadcast_8i32_16i32: +; X64-AVX512BWVL: ## BB#0: +; X64-AVX512BWVL-NEXT: vmovdqa32 (%rdi), %ymm0 +; X64-AVX512BWVL-NEXT: vinserti64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: vpaddd {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: retq +; +; X64-AVX512DQVL-LABEL: test_broadcast_8i32_16i32: +; X64-AVX512DQVL: ## BB#0: +; X64-AVX512DQVL-NEXT: vmovdqa32 (%rdi), %ymm0 +; X64-AVX512DQVL-NEXT: vinserti32x8 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512DQVL-NEXT: vpaddd {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512DQVL-NEXT: retq + %1 = load <8 x i32>, <8 x i32> *%p + %2 = shufflevector <8 x i32> %1, <8 x i32> undef, <16 x i32> + %3 = add <16 x i32> %2, + ret <16 x i32> %3 +} + +define <32 x i16> @test_broadcast_16i16_32i16(<16 x i16> *%p) nounwind { +; X64-AVX512VL-LABEL: test_broadcast_16i16_32i16: +; X64-AVX512VL: ## BB#0: +; X64-AVX512VL-NEXT: vmovdqa64 (%rdi), %ymm1 +; X64-AVX512VL-NEXT: vpaddw {{.*}}(%rip), %ymm1, %ymm0 +; X64-AVX512VL-NEXT: vpaddw {{.*}}(%rip), %ymm1, %ymm1 +; X64-AVX512VL-NEXT: retq +; +; X64-AVX512BWVL-LABEL: test_broadcast_16i16_32i16: +; X64-AVX512BWVL: ## BB#0: +; X64-AVX512BWVL-NEXT: vmovdqu16 (%rdi), %ymm0 +; X64-AVX512BWVL-NEXT: vinserti64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: vpaddw {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: retq +; +; X64-AVX512DQVL-LABEL: test_broadcast_16i16_32i16: +; X64-AVX512DQVL: ## BB#0: +; X64-AVX512DQVL-NEXT: vmovdqa64 (%rdi), %ymm1 +; X64-AVX512DQVL-NEXT: vpaddw {{.*}}(%rip), %ymm1, %ymm0 +; X64-AVX512DQVL-NEXT: vpaddw {{.*}}(%rip), %ymm1, %ymm1 +; X64-AVX512DQVL-NEXT: retq + %1 = load <16 x i16>, <16 x i16> *%p + %2 = shufflevector <16 x i16> %1, <16 x i16> undef, <32 x i32> + %3 = add <32 x i16> %2, + ret <32 x i16> %3 +} + +define <64 x i8> @test_broadcast_32i8_64i8(<32 x i8> *%p) nounwind { +; X64-AVX512VL-LABEL: test_broadcast_32i8_64i8: +; X64-AVX512VL: ## BB#0: +; X64-AVX512VL-NEXT: vmovdqa64 (%rdi), %ymm1 +; X64-AVX512VL-NEXT: vpaddb {{.*}}(%rip), %ymm1, %ymm0 +; X64-AVX512VL-NEXT: vpaddb {{.*}}(%rip), %ymm1, %ymm1 +; X64-AVX512VL-NEXT: retq +; +; X64-AVX512BWVL-LABEL: test_broadcast_32i8_64i8: +; X64-AVX512BWVL: ## BB#0: +; X64-AVX512BWVL-NEXT: vmovdqu8 (%rdi), %ymm0 +; X64-AVX512BWVL-NEXT: vinserti64x4 $1, %ymm0, %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: vpaddb {{.*}}(%rip), %zmm0, %zmm0 +; X64-AVX512BWVL-NEXT: retq +; +; X64-AVX512DQVL-LABEL: test_broadcast_32i8_64i8: +; X64-AVX512DQVL: ## BB#0: +; X64-AVX512DQVL-NEXT: vmovdqa64 (%rdi), %ymm1 +; X64-AVX512DQVL-NEXT: vpaddb {{.*}}(%rip), %ymm1, %ymm0 +; X64-AVX512DQVL-NEXT: vpaddb {{.*}}(%rip), %ymm1, %ymm1 +; X64-AVX512DQVL-NEXT: retq + %1 = load <32 x i8>, <32 x i8> *%p + %2 = shufflevector <32 x i8> %1, <32 x i8> undef, <64 x i32> + %3 = add <64 x i8> %2, + ret <64 x i8> %3 +} -- 2.7.4