[llvm] r334551 - [X86] add avx512 tests for potentially miscompiling cvttp2si/cvttp2ui (PR37751).

Craig Topper via llvm-commits llvm-commits at lists.llvm.org
Tue Jun 12 14:42:42 PDT 2018


Author: ctopper
Date: Tue Jun 12 14:42:42 2018
New Revision: 334551

URL: http://llvm.org/viewvc/llvm-project?rev=334551&view=rev
Log:
[X86] add avx512 tests for potentially miscompiling cvttp2si/cvttp2ui (PR37751).

Added:
    llvm/trunk/test/CodeGen/X86/avx512-cvttp2i.ll

Added: llvm/trunk/test/CodeGen/X86/avx512-cvttp2i.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/X86/avx512-cvttp2i.ll?rev=334551&view=auto
==============================================================================
--- llvm/trunk/test/CodeGen/X86/avx512-cvttp2i.ll (added)
+++ llvm/trunk/test/CodeGen/X86/avx512-cvttp2i.ll Tue Jun 12 14:42:42 2018
@@ -0,0 +1,358 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc < %s -mtriple=x86_64-- -mattr=avx512f,avx512vl,avx512dq | FileCheck %s --check-prefixes=CHECK
+
+; PR37751 - https://bugs.llvm.org/show_bug.cgi?id=37751
+; We can't combine into 'round' instructions because the behavior is different for out-of-range values.
+
+declare <16 x i32> @llvm.x86.avx512.mask.cvttps2dq.512(<16 x float>, <16 x i32>, i16, i32)
+declare <4 x i32> @llvm.x86.avx512.mask.cvttps2udq.128(<4 x float>, <4 x i32>, i8)
+declare <8 x i32> @llvm.x86.avx512.mask.cvttps2udq.256(<8 x float>, <8 x i32>, i8)
+declare <16 x i32> @llvm.x86.avx512.mask.cvttps2udq.512(<16 x float>, <16 x i32>, i16, i32)
+declare <4 x i32> @llvm.x86.avx512.mask.cvttpd2udq.256(<4 x double>, <4 x i32>, i8)
+declare <8 x i32> @llvm.x86.avx512.mask.cvttpd2udq.512(<8 x double>, <8 x i32>, i8, i32)
+declare <4 x i64> @llvm.x86.avx512.mask.cvttps2qq.256(<4 x float>, <4 x i64>, i8)
+declare <8 x i64> @llvm.x86.avx512.mask.cvttps2qq.512(<8 x float>, <8 x i64>, i8, i32)
+declare <4 x i64> @llvm.x86.avx512.mask.cvttps2uqq.256(<4 x float>, <4 x i64>, i8)
+declare <8 x i64> @llvm.x86.avx512.mask.cvttps2uqq.512(<8 x float>, <8 x i64>, i8, i32)
+declare <2 x i64> @llvm.x86.avx512.mask.cvttpd2qq.128(<2 x double>, <2 x i64>, i8)
+declare <4 x i64> @llvm.x86.avx512.mask.cvttpd2qq.256(<4 x double>, <4 x i64>, i8)
+declare <8 x i64> @llvm.x86.avx512.mask.cvttpd2qq.512(<8 x double>, <8 x i64>, i8, i32)
+declare <2 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.128(<2 x double>, <2 x i64>, i8)
+declare <4 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.256(<4 x double>, <4 x i64>, i8)
+declare <8 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.512(<8 x double>, <8 x i64>, i8, i32)
+
+define <16 x float> @float_to_sint_to_float_mem_v16f32(<16 x float>* %p) {
+; CHECK-LABEL: float_to_sint_to_float_mem_v16f32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vrndscaleps $11, (%rdi), %zmm0
+; CHECK-NEXT:    retq
+  %x = load <16 x float>, <16 x float>* %p
+  %fptosi = tail call <16 x i32> @llvm.x86.avx512.mask.cvttps2dq.512(<16 x float> %x, <16 x i32> undef, i16 -1, i32 4)
+  %sitofp = sitofp <16 x i32> %fptosi to <16 x float>
+  ret <16 x float> %sitofp
+}
+
+define <16 x float> @float_to_sint_to_float_reg_v16f32(<16 x float> %x) {
+; CHECK-LABEL: float_to_sint_to_float_reg_v16f32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vrndscaleps $11, %zmm0, %zmm0
+; CHECK-NEXT:    retq
+  %fptosi = tail call <16 x i32> @llvm.x86.avx512.mask.cvttps2dq.512(<16 x float> %x, <16 x i32> undef, i16 -1, i32 4)
+  %sitofp = sitofp <16 x i32> %fptosi to <16 x float>
+  ret <16 x float> %sitofp
+}
+
+define <16 x float> @float_to_uint_to_float_mem_v16f32(<16 x float>* %p) {
+; CHECK-LABEL: float_to_uint_to_float_mem_v16f32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vrndscaleps $11, (%rdi), %zmm0
+; CHECK-NEXT:    retq
+  %x = load <16 x float>, <16 x float>* %p
+  %fptoui = tail call <16 x i32> @llvm.x86.avx512.mask.cvttps2udq.512(<16 x float> %x, <16 x i32> undef, i16 -1, i32 4)
+  %uitofp = uitofp <16 x i32> %fptoui to <16 x float>
+  ret <16 x float> %uitofp
+}
+
+define <16 x float> @float_to_uint_to_float_reg_v16f32(<16 x float> %x) {
+; CHECK-LABEL: float_to_uint_to_float_reg_v16f32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vrndscaleps $11, %zmm0, %zmm0
+; CHECK-NEXT:    retq
+  %fptoui = tail call <16 x i32> @llvm.x86.avx512.mask.cvttps2udq.512(<16 x float> %x, <16 x i32> undef, i16 -1, i32 4)
+  %uitofp = uitofp <16 x i32> %fptoui to <16 x float>
+  ret <16 x float> %uitofp
+}
+
+define <4 x float> @float_to_uint_to_float_mem_v4f32(<4 x float>* %p) {
+; CHECK-LABEL: float_to_uint_to_float_mem_v4f32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundps $11, (%rdi), %xmm0
+; CHECK-NEXT:    retq
+  %x = load <4 x float>, <4 x float>* %p
+  %fptoui = tail call <4 x i32> @llvm.x86.avx512.mask.cvttps2udq.128(<4 x float> %x, <4 x i32> undef, i8 -1)
+  %uitofp = uitofp <4 x i32> %fptoui to <4 x float>
+  ret <4 x float> %uitofp
+}
+
+define <4 x float> @float_to_uint_to_float_reg_v4f32(<4 x float> %x) {
+; CHECK-LABEL: float_to_uint_to_float_reg_v4f32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundps $11, %xmm0, %xmm0
+; CHECK-NEXT:    retq
+  %fptoui = tail call <4 x i32> @llvm.x86.avx512.mask.cvttps2udq.128(<4 x float> %x, <4 x i32> undef, i8 -1)
+  %uitofp = uitofp <4 x i32> %fptoui to <4 x float>
+  ret <4 x float> %uitofp
+}
+
+define <8 x float> @float_to_uint_to_float_mem_v8f32(<8 x float>* %p) {
+; CHECK-LABEL: float_to_uint_to_float_mem_v8f32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundps $11, (%rdi), %ymm0
+; CHECK-NEXT:    retq
+  %x = load <8 x float>, <8 x float>* %p
+  %fptoui = tail call <8 x i32> @llvm.x86.avx512.mask.cvttps2udq.256(<8 x float> %x, <8 x i32> undef, i8 -1)
+  %uitofp = uitofp <8 x i32> %fptoui to <8 x float>
+  ret <8 x float> %uitofp
+}
+
+define <8 x float> @float_to_uint_to_float_reg_v8f32(<8 x float> %x) {
+; CHECK-LABEL: float_to_uint_to_float_reg_v8f32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundps $11, %ymm0, %ymm0
+; CHECK-NEXT:    retq
+  %fptoui = tail call <8 x i32> @llvm.x86.avx512.mask.cvttps2udq.256(<8 x float> %x, <8 x i32> undef, i8 -1)
+  %uitofp = uitofp <8 x i32> %fptoui to <8 x float>
+  ret <8 x float> %uitofp
+}
+
+define <4 x double> @double_to_uint_to_double_mem_v4f64(<4 x double>* %p) {
+; CHECK-LABEL: double_to_uint_to_double_mem_v4f64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundpd $11, (%rdi), %ymm0
+; CHECK-NEXT:    retq
+  %x = load <4 x double>, <4 x double>* %p
+  %fptoui = tail call <4 x i32> @llvm.x86.avx512.mask.cvttpd2udq.256(<4 x double> %x, <4 x i32> undef, i8 -1)
+  %uitofp = uitofp <4 x i32> %fptoui to <4 x double>
+  ret <4 x double> %uitofp
+}
+
+define <4 x double> @double_to_uint_to_double_reg_v4f64(<4 x double> %x) {
+; CHECK-LABEL: double_to_uint_to_double_reg_v4f64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundpd $11, %ymm0, %ymm0
+; CHECK-NEXT:    retq
+  %fptoui = tail call <4 x i32> @llvm.x86.avx512.mask.cvttpd2udq.256(<4 x double> %x, <4 x i32> undef, i8 -1)
+  %uitofp = uitofp <4 x i32> %fptoui to <4 x double>
+  ret <4 x double> %uitofp
+}
+
+define <8 x double> @double_to_uint_to_double_mem_v8f64(<8 x double>* %p) {
+; CHECK-LABEL: double_to_uint_to_double_mem_v8f64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vrndscalepd $11, (%rdi), %zmm0
+; CHECK-NEXT:    retq
+  %x = load <8 x double>, <8 x double>* %p
+  %fptoui = tail call <8 x i32> @llvm.x86.avx512.mask.cvttpd2udq.512(<8 x double> %x, <8 x i32> undef, i8 -1, i32 4)
+  %uitofp = uitofp <8 x i32> %fptoui to <8 x double>
+  ret <8 x double> %uitofp
+}
+
+define <8 x double> @double_to_uint_to_double_reg_v8f64(<8 x double> %x) {
+; CHECK-LABEL: double_to_uint_to_double_reg_v8f64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vrndscalepd $11, %zmm0, %zmm0
+; CHECK-NEXT:    retq
+  %fptoui = tail call <8 x i32> @llvm.x86.avx512.mask.cvttpd2udq.512(<8 x double> %x, <8 x i32> undef, i8 -1, i32 4)
+  %uitofp = uitofp <8 x i32> %fptoui to <8 x double>
+  ret <8 x double> %uitofp
+}
+
+define <4 x float> @float_to_sint64_to_float_mem_v4f32(<4 x float>* %p) {
+; CHECK-LABEL: float_to_sint64_to_float_mem_v4f32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundps $11, (%rdi), %xmm0
+; CHECK-NEXT:    retq
+  %x = load <4 x float>, <4 x float>* %p
+  %fptosi = tail call <4 x i64> @llvm.x86.avx512.mask.cvttps2qq.256(<4 x float> %x, <4 x i64> undef, i8 -1)
+  %sitofp = sitofp <4 x i64> %fptosi to <4 x float>
+  ret <4 x float> %sitofp
+}
+
+define <4 x float> @float_to_sint64_to_float_reg_v4f32(<4 x float> %x) {
+; CHECK-LABEL: float_to_sint64_to_float_reg_v4f32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundps $11, %xmm0, %xmm0
+; CHECK-NEXT:    retq
+  %fptosi = tail call <4 x i64> @llvm.x86.avx512.mask.cvttps2qq.256(<4 x float> %x, <4 x i64> undef, i8 -1)
+  %sitofp = sitofp <4 x i64> %fptosi to <4 x float>
+  ret <4 x float> %sitofp
+}
+
+define <4 x float> @float_to_uint64_to_float_mem_v4f32(<4 x float>* %p) {
+; CHECK-LABEL: float_to_uint64_to_float_mem_v4f32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundps $11, (%rdi), %xmm0
+; CHECK-NEXT:    retq
+  %x = load <4 x float>, <4 x float>* %p
+  %fptoui = tail call <4 x i64> @llvm.x86.avx512.mask.cvttps2uqq.256(<4 x float> %x, <4 x i64> undef, i8 -1)
+  %uitofp = uitofp <4 x i64> %fptoui to <4 x float>
+  ret <4 x float> %uitofp
+}
+
+define <4 x float> @float_to_uint64_to_float_reg_v4f32(<4 x float> %x) {
+; CHECK-LABEL: float_to_uint64_to_float_reg_v4f32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundps $11, %xmm0, %xmm0
+; CHECK-NEXT:    retq
+  %fptoui = tail call <4 x i64> @llvm.x86.avx512.mask.cvttps2uqq.256(<4 x float> %x, <4 x i64> undef, i8 -1)
+  %uitofp = uitofp <4 x i64> %fptoui to <4 x float>
+  ret <4 x float> %uitofp
+}
+
+define <8 x float> @float_to_sint64_to_float_mem_v8f32(<8 x float>* %p) {
+; CHECK-LABEL: float_to_sint64_to_float_mem_v8f32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundps $11, (%rdi), %ymm0
+; CHECK-NEXT:    retq
+  %x = load <8 x float>, <8 x float>* %p
+  %fptosi = tail call <8 x i64> @llvm.x86.avx512.mask.cvttps2qq.512(<8 x float> %x, <8 x i64> undef, i8 -1, i32 4)
+  %sitofp = sitofp <8 x i64> %fptosi to <8 x float>
+  ret <8 x float> %sitofp
+}
+
+define <8 x float> @float_to_sint64_to_float_reg_v8f32(<8 x float> %x) {
+; CHECK-LABEL: float_to_sint64_to_float_reg_v8f32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundps $11, %ymm0, %ymm0
+; CHECK-NEXT:    retq
+  %fptosi = tail call <8 x i64> @llvm.x86.avx512.mask.cvttps2qq.512(<8 x float> %x, <8 x i64> undef, i8 -1, i32 4)
+  %sitofp = sitofp <8 x i64> %fptosi to <8 x float>
+  ret <8 x float> %sitofp
+}
+
+define <8 x float> @float_to_uint64_to_float_mem_v8f32(<8 x float>* %p) {
+; CHECK-LABEL: float_to_uint64_to_float_mem_v8f32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundps $11, (%rdi), %ymm0
+; CHECK-NEXT:    retq
+  %x = load <8 x float>, <8 x float>* %p
+  %fptoui = tail call <8 x i64> @llvm.x86.avx512.mask.cvttps2uqq.512(<8 x float> %x, <8 x i64> undef, i8 -1, i32 4)
+  %uitofp = uitofp <8 x i64> %fptoui to <8 x float>
+  ret <8 x float> %uitofp
+}
+
+define <8 x float> @float_to_uint64_to_float_reg_v8f32(<8 x float> %x) {
+; CHECK-LABEL: float_to_uint64_to_float_reg_v8f32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundps $11, %ymm0, %ymm0
+; CHECK-NEXT:    retq
+  %fptoui = tail call <8 x i64> @llvm.x86.avx512.mask.cvttps2uqq.512(<8 x float> %x, <8 x i64> undef, i8 -1, i32 4)
+  %uitofp = uitofp <8 x i64> %fptoui to <8 x float>
+  ret <8 x float> %uitofp
+}
+
+define <2 x double> @double_to_sint64_to_double_mem_v2f64(<2 x double>* %p) {
+; CHECK-LABEL: double_to_sint64_to_double_mem_v2f64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundpd $11, (%rdi), %xmm0
+; CHECK-NEXT:    retq
+  %x = load <2 x double>, <2 x double>* %p
+  %fptosi = tail call <2 x i64> @llvm.x86.avx512.mask.cvttpd2qq.128(<2 x double> %x, <2 x i64> undef, i8 -1)
+  %sitofp = sitofp <2 x i64> %fptosi to <2 x double>
+  ret <2 x double> %sitofp
+}
+
+define <2 x double> @double_to_sint64_to_double_reg_v2f64(<2 x double> %x) {
+; CHECK-LABEL: double_to_sint64_to_double_reg_v2f64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundpd $11, %xmm0, %xmm0
+; CHECK-NEXT:    retq
+  %fptosi = tail call <2 x i64> @llvm.x86.avx512.mask.cvttpd2qq.128(<2 x double> %x, <2 x i64> undef, i8 -1)
+  %sitofp = sitofp <2 x i64> %fptosi to <2 x double>
+  ret <2 x double> %sitofp
+}
+
+define <2 x double> @double_to_uint64_to_double_mem_v2f64(<2 x double>* %p) {
+; CHECK-LABEL: double_to_uint64_to_double_mem_v2f64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundpd $11, (%rdi), %xmm0
+; CHECK-NEXT:    retq
+  %x = load <2 x double>, <2 x double>* %p
+  %fptoui = tail call <2 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.128(<2 x double> %x, <2 x i64> undef, i8 -1)
+  %uitofp = uitofp <2 x i64> %fptoui to <2 x double>
+  ret <2 x double> %uitofp
+}
+
+define <2 x double> @double_to_uint64_to_double_reg_v2f64(<2 x double> %x) {
+; CHECK-LABEL: double_to_uint64_to_double_reg_v2f64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundpd $11, %xmm0, %xmm0
+; CHECK-NEXT:    retq
+  %fptoui = tail call <2 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.128(<2 x double> %x, <2 x i64> undef, i8 -1)
+  %uitofp = uitofp <2 x i64> %fptoui to <2 x double>
+  ret <2 x double> %uitofp
+}
+
+define <4 x double> @double_to_sint64_to_double_mem_v4f64(<4 x double>* %p) {
+; CHECK-LABEL: double_to_sint64_to_double_mem_v4f64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundpd $11, (%rdi), %ymm0
+; CHECK-NEXT:    retq
+  %x = load <4 x double>, <4 x double>* %p
+  %fptosi = tail call <4 x i64> @llvm.x86.avx512.mask.cvttpd2qq.256(<4 x double> %x, <4 x i64> undef, i8 -1)
+  %sitofp = sitofp <4 x i64> %fptosi to <4 x double>
+  ret <4 x double> %sitofp
+}
+
+define <4 x double> @double_to_sint64_to_double_reg_v4f64(<4 x double> %x) {
+; CHECK-LABEL: double_to_sint64_to_double_reg_v4f64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundpd $11, %ymm0, %ymm0
+; CHECK-NEXT:    retq
+  %fptosi = tail call <4 x i64> @llvm.x86.avx512.mask.cvttpd2qq.256(<4 x double> %x, <4 x i64> undef, i8 -1)
+  %sitofp = sitofp <4 x i64> %fptosi to <4 x double>
+  ret <4 x double> %sitofp
+}
+
+define <4 x double> @double_to_uint64_to_double_mem_v4f64(<4 x double>* %p) {
+; CHECK-LABEL: double_to_uint64_to_double_mem_v4f64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundpd $11, (%rdi), %ymm0
+; CHECK-NEXT:    retq
+  %x = load <4 x double>, <4 x double>* %p
+  %fptoui = tail call <4 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.256(<4 x double> %x, <4 x i64> undef, i8 -1)
+  %uitofp = uitofp <4 x i64> %fptoui to <4 x double>
+  ret <4 x double> %uitofp
+}
+
+define <4 x double> @double_to_uint64_to_double_reg_v4f64(<4 x double> %x) {
+; CHECK-LABEL: double_to_uint64_to_double_reg_v4f64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vroundpd $11, %ymm0, %ymm0
+; CHECK-NEXT:    retq
+  %fptoui = tail call <4 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.256(<4 x double> %x, <4 x i64> undef, i8 -1)
+  %uitofp = uitofp <4 x i64> %fptoui to <4 x double>
+  ret <4 x double> %uitofp
+}
+
+define <8 x double> @double_to_sint64_to_double_mem_v8f64(<8 x double>* %p) {
+; CHECK-LABEL: double_to_sint64_to_double_mem_v8f64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vrndscalepd $11, (%rdi), %zmm0
+; CHECK-NEXT:    retq
+  %x = load <8 x double>, <8 x double>* %p
+  %fptosi = tail call <8 x i64> @llvm.x86.avx512.mask.cvttpd2qq.512(<8 x double> %x, <8 x i64> undef, i8 -1, i32 4)
+  %sitofp = sitofp <8 x i64> %fptosi to <8 x double>
+  ret <8 x double> %sitofp
+}
+
+define <8 x double> @double_to_sint64_to_double_reg_v8f64(<8 x double> %x) {
+; CHECK-LABEL: double_to_sint64_to_double_reg_v8f64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vrndscalepd $11, %zmm0, %zmm0
+; CHECK-NEXT:    retq
+  %fptosi = tail call <8 x i64> @llvm.x86.avx512.mask.cvttpd2qq.512(<8 x double> %x, <8 x i64> undef, i8 -1, i32 4)
+  %sitofp = sitofp <8 x i64> %fptosi to <8 x double>
+  ret <8 x double> %sitofp
+}
+
+define <8 x double> @double_to_uint64_to_double_mem_v8f64(<8 x double>* %p) {
+; CHECK-LABEL: double_to_uint64_to_double_mem_v8f64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vrndscalepd $11, (%rdi), %zmm0
+; CHECK-NEXT:    retq
+  %x = load <8 x double>, <8 x double>* %p
+  %fptoui = tail call <8 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.512(<8 x double> %x, <8 x i64> undef, i8 -1, i32 4)
+  %uitofp = uitofp <8 x i64> %fptoui to <8 x double>
+  ret <8 x double> %uitofp
+}
+
+define <8 x double> @double_to_uint64_to_double_reg_v8f64(<8 x double> %x) {
+; CHECK-LABEL: double_to_uint64_to_double_reg_v8f64:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    vrndscalepd $11, %zmm0, %zmm0
+; CHECK-NEXT:    retq
+  %fptoui = tail call <8 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.512(<8 x double> %x, <8 x i64> undef, i8 -1, i32 4)
+  %uitofp = uitofp <8 x i64> %fptoui to <8 x double>
+  ret <8 x double> %uitofp
+}




More information about the llvm-commits mailing list