[llvm] r281864 - [CostModel][X86] Added scalar float op costs

Simon Pilgrim via llvm-commits llvm-commits at lists.llvm.org
Sun Sep 18 14:01:20 PDT 2016


Author: rksimon
Date: Sun Sep 18 16:01:20 2016
New Revision: 281864

URL: http://llvm.org/viewvc/llvm-project?rev=281864&view=rev
Log:
[CostModel][X86] Added scalar float op costs

Modified:
    llvm/trunk/test/Analysis/CostModel/X86/arith-fp.ll

Modified: llvm/trunk/test/Analysis/CostModel/X86/arith-fp.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/Analysis/CostModel/X86/arith-fp.ll?rev=281864&r1=281863&r2=281864&view=diff
==============================================================================
--- llvm/trunk/test/Analysis/CostModel/X86/arith-fp.ll (original)
+++ llvm/trunk/test/Analysis/CostModel/X86/arith-fp.ll Sun Sep 18 16:01:20 2016
@@ -10,317 +10,405 @@ target triple = "x86_64-apple-macosx10.8
 
 ; CHECK-LABEL: 'fadd'
 define i32 @fadd(i32 %arg) {
-  ; SSE2: cost of 2 {{.*}} %A = fadd
-  ; SSE42: cost of 2 {{.*}} %A = fadd
-  ; AVX: cost of 2 {{.*}} %A = fadd
-  ; AVX2: cost of 2 {{.*}} %A = fadd
-  ; AVX512: cost of 2 {{.*}} %A = fadd
-  %A = fadd <4 x float> undef, undef
-  ; SSE2: cost of 4 {{.*}} %B = fadd
-  ; SSE42: cost of 4 {{.*}} %B = fadd
-  ; AVX: cost of 2 {{.*}} %B = fadd
-  ; AVX2: cost of 2 {{.*}} %B = fadd
-  ; AVX512: cost of 2 {{.*}} %B = fadd
-  %B = fadd <8 x float> undef, undef
-  ; SSE2: cost of 8 {{.*}} %C = fadd
-  ; SSE42: cost of 8 {{.*}} %C = fadd
-  ; AVX: cost of 4 {{.*}} %C = fadd
-  ; AVX2: cost of 4 {{.*}} %C = fadd
-  ; AVX512: cost of 2 {{.*}} %C = fadd
-  %C = fadd <16 x float> undef, undef
-
-  ; SSE2: cost of 2 {{.*}} %D = fadd
-  ; SSE42: cost of 2 {{.*}} %D = fadd
-  ; AVX: cost of 2 {{.*}} %D = fadd
-  ; AVX2: cost of 2 {{.*}} %D = fadd
-  ; AVX512: cost of 2 {{.*}} %D = fadd
-  %D = fadd <2 x double> undef, undef
-  ; SSE2: cost of 4 {{.*}} %E = fadd
-  ; SSE42: cost of 4 {{.*}} %E = fadd
-  ; AVX: cost of 2 {{.*}} %E = fadd
-  ; AVX2: cost of 2 {{.*}} %E = fadd
-  ; AVX512: cost of 2 {{.*}} %E = fadd
-  %E = fadd <4 x double> undef, undef
-  ; SSE2: cost of 8 {{.*}} %F = fadd
-  ; SSE42: cost of 8 {{.*}} %F = fadd
-  ; AVX: cost of 4 {{.*}} %F = fadd
-  ; AVX2: cost of 4 {{.*}} %F = fadd
-  ; AVX512: cost of 2 {{.*}} %F = fadd
-  %F = fadd <8 x double> undef, undef
+  ; SSE2: cost of 2 {{.*}} %F32 = fadd
+  ; SSE42: cost of 2 {{.*}} %F32 = fadd
+  ; AVX: cost of 2 {{.*}} %F32 = fadd
+  ; AVX2: cost of 2 {{.*}} %F32 = fadd
+  ; AVX512: cost of 2 {{.*}} %F32 = fadd
+  %F32 = fadd float undef, undef
+  ; SSE2: cost of 2 {{.*}} %V4F32 = fadd
+  ; SSE42: cost of 2 {{.*}} %V4F32 = fadd
+  ; AVX: cost of 2 {{.*}} %V4F32 = fadd
+  ; AVX2: cost of 2 {{.*}} %V4F32 = fadd
+  ; AVX512: cost of 2 {{.*}} %V4F32 = fadd
+  %V4F32 = fadd <4 x float> undef, undef
+  ; SSE2: cost of 4 {{.*}} %V8F32 = fadd
+  ; SSE42: cost of 4 {{.*}} %V8F32 = fadd
+  ; AVX: cost of 2 {{.*}} %V8F32 = fadd
+  ; AVX2: cost of 2 {{.*}} %V8F32 = fadd
+  ; AVX512: cost of 2 {{.*}} %V8F32 = fadd
+  %V8F32 = fadd <8 x float> undef, undef
+  ; SSE2: cost of 8 {{.*}} %V16F32 = fadd
+  ; SSE42: cost of 8 {{.*}} %V16F32 = fadd
+  ; AVX: cost of 4 {{.*}} %V16F32 = fadd
+  ; AVX2: cost of 4 {{.*}} %V16F32 = fadd
+  ; AVX512: cost of 2 {{.*}} %V16F32 = fadd
+  %V16F32 = fadd <16 x float> undef, undef
+
+  ; SSE2: cost of 2 {{.*}} %F64 = fadd
+  ; SSE42: cost of 2 {{.*}} %F64 = fadd
+  ; AVX: cost of 2 {{.*}} %F64 = fadd
+  ; AVX2: cost of 2 {{.*}} %F64 = fadd
+  ; AVX512: cost of 2 {{.*}} %F64 = fadd
+  %F64 = fadd double undef, undef
+  ; SSE2: cost of 2 {{.*}} %V2F64 = fadd
+  ; SSE42: cost of 2 {{.*}} %V2F64 = fadd
+  ; AVX: cost of 2 {{.*}} %V2F64 = fadd
+  ; AVX2: cost of 2 {{.*}} %V2F64 = fadd
+  ; AVX512: cost of 2 {{.*}} %V2F64 = fadd
+  %V2F64 = fadd <2 x double> undef, undef
+  ; SSE2: cost of 4 {{.*}} %V4F64 = fadd
+  ; SSE42: cost of 4 {{.*}} %V4F64 = fadd
+  ; AVX: cost of 2 {{.*}} %V4F64 = fadd
+  ; AVX2: cost of 2 {{.*}} %V4F64 = fadd
+  ; AVX512: cost of 2 {{.*}} %V4F64 = fadd
+  %V4F64 = fadd <4 x double> undef, undef
+  ; SSE2: cost of 8 {{.*}} %V8F64 = fadd
+  ; SSE42: cost of 8 {{.*}} %V8F64 = fadd
+  ; AVX: cost of 4 {{.*}} %V8F64 = fadd
+  ; AVX2: cost of 4 {{.*}} %V8F64 = fadd
+  ; AVX512: cost of 2 {{.*}} %V8F64 = fadd
+  %V8F64 = fadd <8 x double> undef, undef
 
   ret i32 undef
 }
 
 ; CHECK-LABEL: 'fsub'
 define i32 @fsub(i32 %arg) {
-  ; SSE2: cost of 2 {{.*}} %A = fsub
-  ; SSE42: cost of 2 {{.*}} %A = fsub
-  ; AVX: cost of 2 {{.*}} %A = fsub
-  ; AVX2: cost of 2 {{.*}} %A = fsub
-  ; AVX512: cost of 2 {{.*}} %A = fsub
-  %A = fsub <4 x float> undef, undef
-  ; SSE2: cost of 4 {{.*}} %B = fsub
-  ; SSE42: cost of 4 {{.*}} %B = fsub
-  ; AVX: cost of 2 {{.*}} %B = fsub
-  ; AVX2: cost of 2 {{.*}} %B = fsub
-  ; AVX512: cost of 2 {{.*}} %B = fsub
-  %B = fsub <8 x float> undef, undef
-  ; SSE2: cost of 8 {{.*}} %C = fsub
-  ; SSE42: cost of 8 {{.*}} %C = fsub
-  ; AVX: cost of 4 {{.*}} %C = fsub
-  ; AVX2: cost of 4 {{.*}} %C = fsub
-  ; AVX512: cost of 2 {{.*}} %C = fsub
-  %C = fsub <16 x float> undef, undef
-
-  ; SSE2: cost of 2 {{.*}} %D = fsub
-  ; SSE42: cost of 2 {{.*}} %D = fsub
-  ; AVX: cost of 2 {{.*}} %D = fsub
-  ; AVX2: cost of 2 {{.*}} %D = fsub
-  ; AVX512: cost of 2 {{.*}} %D = fsub
-  %D = fsub <2 x double> undef, undef
-  ; SSE2: cost of 4 {{.*}} %E = fsub
-  ; SSE42: cost of 4 {{.*}} %E = fsub
-  ; AVX: cost of 2 {{.*}} %E = fsub
-  ; AVX2: cost of 2 {{.*}} %E = fsub
-  ; AVX512: cost of 2 {{.*}} %E = fsub
-  %E = fsub <4 x double> undef, undef
-  ; SSE2: cost of 8 {{.*}} %F = fsub
-  ; SSE42: cost of 8 {{.*}} %F = fsub
-  ; AVX: cost of 4 {{.*}} %F = fsub
-  ; AVX2: cost of 4 {{.*}} %F = fsub
-  ; AVX512: cost of 2 {{.*}} %F = fsub
-  %F = fsub <8 x double> undef, undef
+  ; SSE2: cost of 2 {{.*}} %F32 = fsub
+  ; SSE42: cost of 2 {{.*}} %F32 = fsub
+  ; AVX: cost of 2 {{.*}} %F32 = fsub
+  ; AVX2: cost of 2 {{.*}} %F32 = fsub
+  ; AVX512: cost of 2 {{.*}} %F32 = fsub
+  %F32 = fsub float undef, undef
+  ; SSE2: cost of 2 {{.*}} %V4F32 = fsub
+  ; SSE42: cost of 2 {{.*}} %V4F32 = fsub
+  ; AVX: cost of 2 {{.*}} %V4F32 = fsub
+  ; AVX2: cost of 2 {{.*}} %V4F32 = fsub
+  ; AVX512: cost of 2 {{.*}} %V4F32 = fsub
+  %V4F32 = fsub <4 x float> undef, undef
+  ; SSE2: cost of 4 {{.*}} %V8F32 = fsub
+  ; SSE42: cost of 4 {{.*}} %V8F32 = fsub
+  ; AVX: cost of 2 {{.*}} %V8F32 = fsub
+  ; AVX2: cost of 2 {{.*}} %V8F32 = fsub
+  ; AVX512: cost of 2 {{.*}} %V8F32 = fsub
+  %V8F32 = fsub <8 x float> undef, undef
+  ; SSE2: cost of 8 {{.*}} %V16F32 = fsub
+  ; SSE42: cost of 8 {{.*}} %V16F32 = fsub
+  ; AVX: cost of 4 {{.*}} %V16F32 = fsub
+  ; AVX2: cost of 4 {{.*}} %V16F32 = fsub
+  ; AVX512: cost of 2 {{.*}} %V16F32 = fsub
+  %V16F32 = fsub <16 x float> undef, undef
+
+  ; SSE2: cost of 2 {{.*}} %F64 = fsub
+  ; SSE42: cost of 2 {{.*}} %F64 = fsub
+  ; AVX: cost of 2 {{.*}} %F64 = fsub
+  ; AVX2: cost of 2 {{.*}} %F64 = fsub
+  ; AVX512: cost of 2 {{.*}} %F64 = fsub
+  %F64 = fsub double undef, undef
+  ; SSE2: cost of 2 {{.*}} %V2F64 = fsub
+  ; SSE42: cost of 2 {{.*}} %V2F64 = fsub
+  ; AVX: cost of 2 {{.*}} %V2F64 = fsub
+  ; AVX2: cost of 2 {{.*}} %V2F64 = fsub
+  ; AVX512: cost of 2 {{.*}} %V2F64 = fsub
+  %V2F64 = fsub <2 x double> undef, undef
+  ; SSE2: cost of 4 {{.*}} %V4F64 = fsub
+  ; SSE42: cost of 4 {{.*}} %V4F64 = fsub
+  ; AVX: cost of 2 {{.*}} %V4F64 = fsub
+  ; AVX2: cost of 2 {{.*}} %V4F64 = fsub
+  ; AVX512: cost of 2 {{.*}} %V4F64 = fsub
+  %V4F64 = fsub <4 x double> undef, undef
+  ; SSE2: cost of 8 {{.*}} %V8F64 = fsub
+  ; SSE42: cost of 8 {{.*}} %V8F64 = fsub
+  ; AVX: cost of 4 {{.*}} %V8F64 = fsub
+  ; AVX2: cost of 4 {{.*}} %V8F64 = fsub
+  ; AVX512: cost of 2 {{.*}} %V8F64 = fsub
+  %V8F64 = fsub <8 x double> undef, undef
 
   ret i32 undef
 }
 
 ; CHECK-LABEL: 'fmul'
 define i32 @fmul(i32 %arg) {
-  ; SSE2: cost of 2 {{.*}} %A = fmul
-  ; SSE42: cost of 2 {{.*}} %A = fmul
-  ; AVX: cost of 2 {{.*}} %A = fmul
-  ; AVX2: cost of 2 {{.*}} %A = fmul
-  ; AVX512: cost of 2 {{.*}} %A = fmul
-  %A = fmul <4 x float> undef, undef
-  ; SSE2: cost of 4 {{.*}} %B = fmul
-  ; SSE42: cost of 4 {{.*}} %B = fmul
-  ; AVX: cost of 2 {{.*}} %B = fmul
-  ; AVX2: cost of 2 {{.*}} %B = fmul
-  ; AVX512: cost of 2 {{.*}} %B = fmul
-  %B = fmul <8 x float> undef, undef
-  ; SSE2: cost of 8 {{.*}} %C = fmul
-  ; SSE42: cost of 8 {{.*}} %C = fmul
-  ; AVX: cost of 4 {{.*}} %C = fmul
-  ; AVX2: cost of 4 {{.*}} %C = fmul
-  ; AVX512: cost of 2 {{.*}} %C = fmul
-  %C = fmul <16 x float> undef, undef
-
-  ; SSE2: cost of 2 {{.*}} %D = fmul
-  ; SSE42: cost of 2 {{.*}} %D = fmul
-  ; AVX: cost of 2 {{.*}} %D = fmul
-  ; AVX2: cost of 2 {{.*}} %D = fmul
-  ; AVX512: cost of 2 {{.*}} %D = fmul
-  %D = fmul <2 x double> undef, undef
-  ; SSE2: cost of 4 {{.*}} %E = fmul
-  ; SSE42: cost of 4 {{.*}} %E = fmul
-  ; AVX: cost of 2 {{.*}} %E = fmul
-  ; AVX2: cost of 2 {{.*}} %E = fmul
-  ; AVX512: cost of 2 {{.*}} %E = fmul
-  %E = fmul <4 x double> undef, undef
-  ; SSE2: cost of 8 {{.*}} %F = fmul
-  ; SSE42: cost of 8 {{.*}} %F = fmul
-  ; AVX: cost of 4 {{.*}} %F = fmul
-  ; AVX2: cost of 4 {{.*}} %F = fmul
-  ; AVX512: cost of 2 {{.*}} %F = fmul
-  %F = fmul <8 x double> undef, undef
+  ; SSE2: cost of 2 {{.*}} %F32 = fmul
+  ; SSE42: cost of 2 {{.*}} %F32 = fmul
+  ; AVX: cost of 2 {{.*}} %F32 = fmul
+  ; AVX2: cost of 2 {{.*}} %F32 = fmul
+  ; AVX512: cost of 2 {{.*}} %F32 = fmul
+  %F32 = fmul float undef, undef
+  ; SSE2: cost of 2 {{.*}} %V4F32 = fmul
+  ; SSE42: cost of 2 {{.*}} %V4F32 = fmul
+  ; AVX: cost of 2 {{.*}} %V4F32 = fmul
+  ; AVX2: cost of 2 {{.*}} %V4F32 = fmul
+  ; AVX512: cost of 2 {{.*}} %V4F32 = fmul
+  %V4F32 = fmul <4 x float> undef, undef
+  ; SSE2: cost of 4 {{.*}} %V8F32 = fmul
+  ; SSE42: cost of 4 {{.*}} %V8F32 = fmul
+  ; AVX: cost of 2 {{.*}} %V8F32 = fmul
+  ; AVX2: cost of 2 {{.*}} %V8F32 = fmul
+  ; AVX512: cost of 2 {{.*}} %V8F32 = fmul
+  %V8F32 = fmul <8 x float> undef, undef
+  ; SSE2: cost of 8 {{.*}} %V16F32 = fmul
+  ; SSE42: cost of 8 {{.*}} %V16F32 = fmul
+  ; AVX: cost of 4 {{.*}} %V16F32 = fmul
+  ; AVX2: cost of 4 {{.*}} %V16F32 = fmul
+  ; AVX512: cost of 2 {{.*}} %V16F32 = fmul
+  %V16F32 = fmul <16 x float> undef, undef
+
+  ; SSE2: cost of 2 {{.*}} %F64 = fmul
+  ; SSE42: cost of 2 {{.*}} %F64 = fmul
+  ; AVX: cost of 2 {{.*}} %F64 = fmul
+  ; AVX2: cost of 2 {{.*}} %F64 = fmul
+  ; AVX512: cost of 2 {{.*}} %F64 = fmul
+  %F64 = fmul double undef, undef
+  ; SSE2: cost of 2 {{.*}} %V2F64 = fmul
+  ; SSE42: cost of 2 {{.*}} %V2F64 = fmul
+  ; AVX: cost of 2 {{.*}} %V2F64 = fmul
+  ; AVX2: cost of 2 {{.*}} %V2F64 = fmul
+  ; AVX512: cost of 2 {{.*}} %V2F64 = fmul
+  %V2F64 = fmul <2 x double> undef, undef
+  ; SSE2: cost of 4 {{.*}} %V4F64 = fmul
+  ; SSE42: cost of 4 {{.*}} %V4F64 = fmul
+  ; AVX: cost of 2 {{.*}} %V4F64 = fmul
+  ; AVX2: cost of 2 {{.*}} %V4F64 = fmul
+  ; AVX512: cost of 2 {{.*}} %V4F64 = fmul
+  %V4F64 = fmul <4 x double> undef, undef
+  ; SSE2: cost of 8 {{.*}} %V8F64 = fmul
+  ; SSE42: cost of 8 {{.*}} %V8F64 = fmul
+  ; AVX: cost of 4 {{.*}} %V8F64 = fmul
+  ; AVX2: cost of 4 {{.*}} %V8F64 = fmul
+  ; AVX512: cost of 2 {{.*}} %V8F64 = fmul
+  %V8F64 = fmul <8 x double> undef, undef
 
   ret i32 undef
 }
 
 ; CHECK-LABEL: 'fdiv'
 define i32 @fdiv(i32 %arg) {
-  ; SSE2: cost of 2 {{.*}} %A = fdiv
-  ; SSE42: cost of 2 {{.*}} %A = fdiv
-  ; AVX: cost of 2 {{.*}} %A = fdiv
-  ; AVX2: cost of 2 {{.*}} %A = fdiv
-  ; AVX512: cost of 2 {{.*}} %A = fdiv
-  %A = fdiv <4 x float> undef, undef
-  ; SSE2: cost of 4 {{.*}} %B = fdiv
-  ; SSE42: cost of 4 {{.*}} %B = fdiv
-  ; AVX: cost of 2 {{.*}} %B = fdiv
-  ; AVX2: cost of 2 {{.*}} %B = fdiv
-  ; AVX512: cost of 2 {{.*}} %B = fdiv
-  %B = fdiv <8 x float> undef, undef
-  ; SSE2: cost of 8 {{.*}} %C = fdiv
-  ; SSE42: cost of 8 {{.*}} %C = fdiv
-  ; AVX: cost of 4 {{.*}} %C = fdiv
-  ; AVX2: cost of 4 {{.*}} %C = fdiv
-  ; AVX512: cost of 2 {{.*}} %C = fdiv
-  %C = fdiv <16 x float> undef, undef
-
-  ; SSE2: cost of 2 {{.*}} %D = fdiv
-  ; SSE42: cost of 2 {{.*}} %D = fdiv
-  ; AVX: cost of 2 {{.*}} %D = fdiv
-  ; AVX2: cost of 2 {{.*}} %D = fdiv
-  ; AVX512: cost of 2 {{.*}} %D = fdiv
-  %D = fdiv <2 x double> undef, undef
-  ; SSE2: cost of 4 {{.*}} %E = fdiv
-  ; SSE42: cost of 4 {{.*}} %E = fdiv
-  ; AVX: cost of 2 {{.*}} %E = fdiv
-  ; AVX2: cost of 2 {{.*}} %E = fdiv
-  ; AVX512: cost of 2 {{.*}} %E = fdiv
-  %E = fdiv <4 x double> undef, undef
-  ; SSE2: cost of 8 {{.*}} %F = fdiv
-  ; SSE42: cost of 8 {{.*}} %F = fdiv
-  ; AVX: cost of 4 {{.*}} %F = fdiv
-  ; AVX2: cost of 4 {{.*}} %F = fdiv
-  ; AVX512: cost of 2 {{.*}} %F = fdiv
-  %F = fdiv <8 x double> undef, undef
+  ; SSE2: cost of 2 {{.*}} %F32 = fdiv
+  ; SSE42: cost of 2 {{.*}} %F32 = fdiv
+  ; AVX: cost of 2 {{.*}} %F32 = fdiv
+  ; AVX2: cost of 2 {{.*}} %F32 = fdiv
+  ; AVX512: cost of 2 {{.*}} %F32 = fdiv
+  %F32 = fdiv float undef, undef
+  ; SSE2: cost of 2 {{.*}} %V4F32 = fdiv
+  ; SSE42: cost of 2 {{.*}} %V4F32 = fdiv
+  ; AVX: cost of 2 {{.*}} %V4F32 = fdiv
+  ; AVX2: cost of 2 {{.*}} %V4F32 = fdiv
+  ; AVX512: cost of 2 {{.*}} %V4F32 = fdiv
+  %V4F32 = fdiv <4 x float> undef, undef
+  ; SSE2: cost of 4 {{.*}} %V8F32 = fdiv
+  ; SSE42: cost of 4 {{.*}} %V8F32 = fdiv
+  ; AVX: cost of 2 {{.*}} %V8F32 = fdiv
+  ; AVX2: cost of 2 {{.*}} %V8F32 = fdiv
+  ; AVX512: cost of 2 {{.*}} %V8F32 = fdiv
+  %V8F32 = fdiv <8 x float> undef, undef
+  ; SSE2: cost of 8 {{.*}} %V16F32 = fdiv
+  ; SSE42: cost of 8 {{.*}} %V16F32 = fdiv
+  ; AVX: cost of 4 {{.*}} %V16F32 = fdiv
+  ; AVX2: cost of 4 {{.*}} %V16F32 = fdiv
+  ; AVX512: cost of 2 {{.*}} %V16F32 = fdiv
+  %V16F32 = fdiv <16 x float> undef, undef
+
+  ; SSE2: cost of 2 {{.*}} %F64 = fdiv
+  ; SSE42: cost of 2 {{.*}} %F64 = fdiv
+  ; AVX: cost of 2 {{.*}} %F64 = fdiv
+  ; AVX2: cost of 2 {{.*}} %F64 = fdiv
+  ; AVX512: cost of 2 {{.*}} %F64 = fdiv
+  %F64 = fdiv double undef, undef
+  ; SSE2: cost of 2 {{.*}} %V2F64 = fdiv
+  ; SSE42: cost of 2 {{.*}} %V2F64 = fdiv
+  ; AVX: cost of 2 {{.*}} %V2F64 = fdiv
+  ; AVX2: cost of 2 {{.*}} %V2F64 = fdiv
+  ; AVX512: cost of 2 {{.*}} %V2F64 = fdiv
+  %V2F64 = fdiv <2 x double> undef, undef
+  ; SSE2: cost of 4 {{.*}} %V4F64 = fdiv
+  ; SSE42: cost of 4 {{.*}} %V4F64 = fdiv
+  ; AVX: cost of 2 {{.*}} %V4F64 = fdiv
+  ; AVX2: cost of 2 {{.*}} %V4F64 = fdiv
+  ; AVX512: cost of 2 {{.*}} %V4F64 = fdiv
+  %V4F64 = fdiv <4 x double> undef, undef
+  ; SSE2: cost of 8 {{.*}} %V8F64 = fdiv
+  ; SSE42: cost of 8 {{.*}} %V8F64 = fdiv
+  ; AVX: cost of 4 {{.*}} %V8F64 = fdiv
+  ; AVX2: cost of 4 {{.*}} %V8F64 = fdiv
+  ; AVX512: cost of 2 {{.*}} %V8F64 = fdiv
+  %V8F64 = fdiv <8 x double> undef, undef
 
   ret i32 undef
 }
 
 ; CHECK-LABEL: 'frem'
 define i32 @frem(i32 %arg) {
-  ; SSE2: cost of 14 {{.*}} %A = frem
-  ; SSE42: cost of 14 {{.*}} %A = frem
-  ; AVX: cost of 14 {{.*}} %A = frem
-  ; AVX2: cost of 14 {{.*}} %A = frem
-  ; AVX512: cost of 14 {{.*}} %A = frem
-  %A = frem <4 x float> undef, undef
-  ; SSE2: cost of 28 {{.*}} %B = frem
-  ; SSE42: cost of 28 {{.*}} %B = frem
-  ; AVX: cost of 30 {{.*}} %B = frem
-  ; AVX2: cost of 30 {{.*}} %B = frem
-  ; AVX512: cost of 30 {{.*}} %B = frem
-  %B = frem <8 x float> undef, undef
-  ; SSE2: cost of 56 {{.*}} %C = frem
-  ; SSE42: cost of 56 {{.*}} %C = frem
-  ; AVX: cost of 60 {{.*}} %C = frem
-  ; AVX2: cost of 60 {{.*}} %C = frem
-  ; AVX512: cost of 62 {{.*}} %C = frem
-  %C = frem <16 x float> undef, undef
-
-  ; SSE2: cost of 6 {{.*}} %D = frem
-  ; SSE42: cost of 6 {{.*}} %D = frem
-  ; AVX: cost of 6 {{.*}} %D = frem
-  ; AVX2: cost of 6 {{.*}} %D = frem
-  ; AVX512: cost of 6 {{.*}} %D = frem
-  %D = frem <2 x double> undef, undef
-  ; SSE2: cost of 12 {{.*}} %E = frem
-  ; SSE42: cost of 12 {{.*}} %E = frem
-  ; AVX: cost of 14 {{.*}} %E = frem
-  ; AVX2: cost of 14 {{.*}} %E = frem
-  ; AVX512: cost of 14 {{.*}} %E = frem
-  %E = frem <4 x double> undef, undef
-  ; SSE2: cost of 24 {{.*}} %F = frem
-  ; SSE42: cost of 24 {{.*}} %F = frem
-  ; AVX: cost of 28 {{.*}} %F = frem
-  ; AVX2: cost of 28 {{.*}} %F = frem
-  ; AVX512: cost of 30 {{.*}} %F = frem
-  %F = frem <8 x double> undef, undef
+  ; SSE2: cost of 2 {{.*}} %F32 = frem
+  ; SSE42: cost of 2 {{.*}} %F32 = frem
+  ; AVX: cost of 2 {{.*}} %F32 = frem
+  ; AVX2: cost of 2 {{.*}} %F32 = frem
+  ; AVX512: cost of 2 {{.*}} %F32 = frem
+  %F32 = frem float undef, undef
+  ; SSE2: cost of 14 {{.*}} %V4F32 = frem
+  ; SSE42: cost of 14 {{.*}} %V4F32 = frem
+  ; AVX: cost of 14 {{.*}} %V4F32 = frem
+  ; AVX2: cost of 14 {{.*}} %V4F32 = frem
+  ; AVX512: cost of 14 {{.*}} %V4F32 = frem
+  %V4F32 = frem <4 x float> undef, undef
+  ; SSE2: cost of 28 {{.*}} %V8F32 = frem
+  ; SSE42: cost of 28 {{.*}} %V8F32 = frem
+  ; AVX: cost of 30 {{.*}} %V8F32 = frem
+  ; AVX2: cost of 30 {{.*}} %V8F32 = frem
+  ; AVX512: cost of 30 {{.*}} %V8F32 = frem
+  %V8F32 = frem <8 x float> undef, undef
+  ; SSE2: cost of 56 {{.*}} %V16F32 = frem
+  ; SSE42: cost of 56 {{.*}} %V16F32 = frem
+  ; AVX: cost of 60 {{.*}} %V16F32 = frem
+  ; AVX2: cost of 60 {{.*}} %V16F32 = frem
+  ; AVX512: cost of 62 {{.*}} %V16F32 = frem
+  %V16F32 = frem <16 x float> undef, undef
+
+  ; SSE2: cost of 2 {{.*}} %F64 = frem
+  ; SSE42: cost of 2 {{.*}} %F64 = frem
+  ; AVX: cost of 2 {{.*}} %F64 = frem
+  ; AVX2: cost of 2 {{.*}} %F64 = frem
+  ; AVX512: cost of 2 {{.*}} %F64 = frem
+  %F64 = frem double undef, undef
+  ; SSE2: cost of 6 {{.*}} %V2F64 = frem
+  ; SSE42: cost of 6 {{.*}} %V2F64 = frem
+  ; AVX: cost of 6 {{.*}} %V2F64 = frem
+  ; AVX2: cost of 6 {{.*}} %V2F64 = frem
+  ; AVX512: cost of 6 {{.*}} %V2F64 = frem
+  %V2F64 = frem <2 x double> undef, undef
+  ; SSE2: cost of 12 {{.*}} %V4F64 = frem
+  ; SSE42: cost of 12 {{.*}} %V4F64 = frem
+  ; AVX: cost of 14 {{.*}} %V4F64 = frem
+  ; AVX2: cost of 14 {{.*}} %V4F64 = frem
+  ; AVX512: cost of 14 {{.*}} %V4F64 = frem
+  %V4F64 = frem <4 x double> undef, undef
+  ; SSE2: cost of 24 {{.*}} %V8F64 = frem
+  ; SSE42: cost of 24 {{.*}} %V8F64 = frem
+  ; AVX: cost of 28 {{.*}} %V8F64 = frem
+  ; AVX2: cost of 28 {{.*}} %V8F64 = frem
+  ; AVX512: cost of 30 {{.*}} %V8F64 = frem
+  %V8F64 = frem <8 x double> undef, undef
 
   ret i32 undef
 }
 
 ; CHECK-LABEL: 'fsqrt'
 define i32 @fsqrt(i32 %arg) {
-  ; SSE2: cost of 1 {{.*}} %A = call <4 x float> @llvm.sqrt.v4f32
-  ; SSE42: cost of 1 {{.*}} %A = call <4 x float> @llvm.sqrt.v4f32
-  ; AVX: cost of 1 {{.*}} %A = call <4 x float> @llvm.sqrt.v4f32
-  ; AVX2: cost of 1 {{.*}} %A = call <4 x float> @llvm.sqrt.v4f32
-  ; AVX512: cost of 1 {{.*}} %A = call <4 x float> @llvm.sqrt.v4f32
-  %A = call <4 x float> @llvm.sqrt.v4f32(<4 x float> undef)
-  ; SSE2: cost of 4 {{.*}} %B = call <8 x float> @llvm.sqrt.v8f32
-  ; SSE42: cost of 4 {{.*}} %B = call <8 x float> @llvm.sqrt.v8f32
-  ; AVX: cost of 1 {{.*}} %B = call <8 x float> @llvm.sqrt.v8f32
-  ; AVX2: cost of 1 {{.*}} %B = call <8 x float> @llvm.sqrt.v8f32
-  ; AVX512: cost of 1 {{.*}} %B = call <8 x float> @llvm.sqrt.v8f32
-  %B = call <8 x float> @llvm.sqrt.v8f32(<8 x float> undef)
-  ; SSE2: cost of 8 {{.*}} %C = call <16 x float> @llvm.sqrt.v16f32
-  ; SSE42: cost of 8 {{.*}} %C = call <16 x float> @llvm.sqrt.v16f32
-  ; AVX: cost of 4 {{.*}} %C = call <16 x float> @llvm.sqrt.v16f32
-  ; AVX2: cost of 4 {{.*}} %C = call <16 x float> @llvm.sqrt.v16f32
-  ; AVX512: cost of 1 {{.*}} %C = call <16 x float> @llvm.sqrt.v16f32
-  %C = call <16 x float> @llvm.sqrt.v16f32(<16 x float> undef)
-
-  ; SSE2: cost of 1 {{.*}} %D = call <2 x double> @llvm.sqrt.v2f64
-  ; SSE42: cost of 1 {{.*}} %D = call <2 x double> @llvm.sqrt.v2f64
-  ; AVX: cost of 1 {{.*}} %D = call <2 x double> @llvm.sqrt.v2f64
-  ; AVX2: cost of 1 {{.*}} %D = call <2 x double> @llvm.sqrt.v2f64
-  ; AVX512: cost of 1 {{.*}} %D = call <2 x double> @llvm.sqrt.v2f64
-  %D = call <2 x double> @llvm.sqrt.v2f64(<2 x double> undef)
-  ; SSE2: cost of 4 {{.*}} %E = call <4 x double> @llvm.sqrt.v4f64
-  ; SSE42: cost of 4 {{.*}} %E = call <4 x double> @llvm.sqrt.v4f64
-  ; AVX: cost of 1 {{.*}} %E = call <4 x double> @llvm.sqrt.v4f64
-  ; AVX2: cost of 1 {{.*}} %E = call <4 x double> @llvm.sqrt.v4f64
-  ; AVX512: cost of 1 {{.*}} %E = call <4 x double> @llvm.sqrt.v4f64
-  %E = call <4 x double> @llvm.sqrt.v4f64(<4 x double> undef)
-  ; SSE2: cost of 8 {{.*}} %F = call <8 x double> @llvm.sqrt.v8f64
-  ; SSE42: cost of 8 {{.*}} %F = call <8 x double> @llvm.sqrt.v8f64
-  ; AVX: cost of 4 {{.*}} %F = call <8 x double> @llvm.sqrt.v8f64
-  ; AVX2: cost of 4 {{.*}} %F = call <8 x double> @llvm.sqrt.v8f64
-  ; AVX512: cost of 1 {{.*}} %F = call <8 x double> @llvm.sqrt.v8f64
-  %F = call <8 x double> @llvm.sqrt.v8f64(<8 x double> undef)
+  ; SSE2: cost of 1 {{.*}} %F32 = call float @llvm.sqrt.f32
+  ; SSE42: cost of 1 {{.*}} %F32 = call float @llvm.sqrt.f32
+  ; AVX: cost of 1 {{.*}} %F32 = call float @llvm.sqrt.f32
+  ; AVX2: cost of 1 {{.*}} %F32 = call float @llvm.sqrt.f32
+  ; AVX512: cost of 1 {{.*}} %F32 = call float @llvm.sqrt.f32
+  %F32 = call float @llvm.sqrt.f32(float undef)
+  ; SSE2: cost of 1 {{.*}} %V4F32 = call <4 x float> @llvm.sqrt.v4f32
+  ; SSE42: cost of 1 {{.*}} %V4F32 = call <4 x float> @llvm.sqrt.v4f32
+  ; AVX: cost of 1 {{.*}} %V4F32 = call <4 x float> @llvm.sqrt.v4f32
+  ; AVX2: cost of 1 {{.*}} %V4F32 = call <4 x float> @llvm.sqrt.v4f32
+  ; AVX512: cost of 1 {{.*}} %V4F32 = call <4 x float> @llvm.sqrt.v4f32
+  %V4F32 = call <4 x float> @llvm.sqrt.v4f32(<4 x float> undef)
+  ; SSE2: cost of 4 {{.*}} %V8F32 = call <8 x float> @llvm.sqrt.v8f32
+  ; SSE42: cost of 4 {{.*}} %V8F32 = call <8 x float> @llvm.sqrt.v8f32
+  ; AVX: cost of 1 {{.*}} %V8F32 = call <8 x float> @llvm.sqrt.v8f32
+  ; AVX2: cost of 1 {{.*}} %V8F32 = call <8 x float> @llvm.sqrt.v8f32
+  ; AVX512: cost of 1 {{.*}} %V8F32 = call <8 x float> @llvm.sqrt.v8f32
+  %V8F32 = call <8 x float> @llvm.sqrt.v8f32(<8 x float> undef)
+  ; SSE2: cost of 8 {{.*}} %V16F32 = call <16 x float> @llvm.sqrt.v16f32
+  ; SSE42: cost of 8 {{.*}} %V16F32 = call <16 x float> @llvm.sqrt.v16f32
+  ; AVX: cost of 4 {{.*}} %V16F32 = call <16 x float> @llvm.sqrt.v16f32
+  ; AVX2: cost of 4 {{.*}} %V16F32 = call <16 x float> @llvm.sqrt.v16f32
+  ; AVX512: cost of 1 {{.*}} %V16F32 = call <16 x float> @llvm.sqrt.v16f32
+  %V16F32 = call <16 x float> @llvm.sqrt.v16f32(<16 x float> undef)
+
+  ; SSE2: cost of 1 {{.*}} %F64 = call double @llvm.sqrt.f64
+  ; SSE42: cost of 1 {{.*}} %F64 = call double @llvm.sqrt.f64
+  ; AVX: cost of 1 {{.*}} %F64 = call double @llvm.sqrt.f64
+  ; AVX2: cost of 1 {{.*}} %F64 = call double @llvm.sqrt.f64
+  ; AVX512: cost of 1 {{.*}} %F64 = call double @llvm.sqrt.f64
+  %F64 = call double @llvm.sqrt.f64(double undef)
+  ; SSE2: cost of 1 {{.*}} %V2F64 = call <2 x double> @llvm.sqrt.v2f64
+  ; SSE42: cost of 1 {{.*}} %V2F64 = call <2 x double> @llvm.sqrt.v2f64
+  ; AVX: cost of 1 {{.*}} %V2F64 = call <2 x double> @llvm.sqrt.v2f64
+  ; AVX2: cost of 1 {{.*}} %V2F64 = call <2 x double> @llvm.sqrt.v2f64
+  ; AVX512: cost of 1 {{.*}} %V2F64 = call <2 x double> @llvm.sqrt.v2f64
+  %V2F64 = call <2 x double> @llvm.sqrt.v2f64(<2 x double> undef)
+  ; SSE2: cost of 4 {{.*}} %V4F64 = call <4 x double> @llvm.sqrt.v4f64
+  ; SSE42: cost of 4 {{.*}} %V4F64 = call <4 x double> @llvm.sqrt.v4f64
+  ; AVX: cost of 1 {{.*}} %V4F64 = call <4 x double> @llvm.sqrt.v4f64
+  ; AVX2: cost of 1 {{.*}} %V4F64 = call <4 x double> @llvm.sqrt.v4f64
+  ; AVX512: cost of 1 {{.*}} %V4F64 = call <4 x double> @llvm.sqrt.v4f64
+  %V4F64 = call <4 x double> @llvm.sqrt.v4f64(<4 x double> undef)
+  ; SSE2: cost of 8 {{.*}} %V8F64 = call <8 x double> @llvm.sqrt.v8f64
+  ; SSE42: cost of 8 {{.*}} %V8F64 = call <8 x double> @llvm.sqrt.v8f64
+  ; AVX: cost of 4 {{.*}} %V8F64 = call <8 x double> @llvm.sqrt.v8f64
+  ; AVX2: cost of 4 {{.*}} %V8F64 = call <8 x double> @llvm.sqrt.v8f64
+  ; AVX512: cost of 1 {{.*}} %V8F64 = call <8 x double> @llvm.sqrt.v8f64
+  %V8F64 = call <8 x double> @llvm.sqrt.v8f64(<8 x double> undef)
 
   ret i32 undef
 }
 
 ; CHECK-LABEL: 'fma'
 define i32 @fma(i32 %arg) {
-  ; SSE2: cost of 52 {{.*}} %A = call <4 x float> @llvm.fma.v4f32
-  ; SSE42: cost of 52 {{.*}} %A = call <4 x float> @llvm.fma.v4f32
-  ; AVX: cost of 1 {{.*}} %A = call <4 x float> @llvm.fma.v4f32
-  ; AVX2: cost of 1 {{.*}} %A = call <4 x float> @llvm.fma.v4f32
-  ; AVX512: cost of 1 {{.*}} %A = call <4 x float> @llvm.fma.v4f32
-  %A = call <4 x float> @llvm.fma.v4f32(<4 x float> undef, <4 x float> undef, <4 x float> undef)
-  ; SSE2: cost of 104 {{.*}} %B = call <8 x float> @llvm.fma.v8f32
-  ; SSE42: cost of 104 {{.*}} %B = call <8 x float> @llvm.fma.v8f32
-  ; AVX: cost of 1 {{.*}} %B = call <8 x float> @llvm.fma.v8f32
-  ; AVX2: cost of 1 {{.*}} %B = call <8 x float> @llvm.fma.v8f32
-  ; AVX512: cost of 1 {{.*}} %B = call <8 x float> @llvm.fma.v8f32
-  %B = call <8 x float> @llvm.fma.v8f32(<8 x float> undef, <8 x float> undef, <8 x float> undef)
-  ; SSE2: cost of 208 {{.*}} %C = call <16 x float> @llvm.fma.v16f32
-  ; SSE42: cost of 208 {{.*}} %C = call <16 x float> @llvm.fma.v16f32
-  ; AVX: cost of 4 {{.*}} %C = call <16 x float> @llvm.fma.v16f32
-  ; AVX2: cost of 4 {{.*}} %C = call <16 x float> @llvm.fma.v16f32
-  ; AVX512: cost of 1 {{.*}} %C = call <16 x float> @llvm.fma.v16f32
-  %C = call <16 x float> @llvm.fma.v16f32(<16 x float> undef, <16 x float> undef, <16 x float> undef)
-
-  ; SSE2: cost of 24 {{.*}} %D = call <2 x double> @llvm.fma.v2f64
-  ; SSE42: cost of 24 {{.*}} %D = call <2 x double> @llvm.fma.v2f64
-  ; AVX: cost of 1 {{.*}} %D = call <2 x double> @llvm.fma.v2f64
-  ; AVX2: cost of 1 {{.*}} %D = call <2 x double> @llvm.fma.v2f64
-  ; AVX512: cost of 1 {{.*}} %D = call <2 x double> @llvm.fma.v2f64
-  %D = call <2 x double> @llvm.fma.v2f64(<2 x double> undef, <2 x double> undef, <2 x double> undef)
-  ; SSE2: cost of 48 {{.*}} %E = call <4 x double> @llvm.fma.v4f64
-  ; SSE42: cost of 48 {{.*}} %E = call <4 x double> @llvm.fma.v4f64
-  ; AVX: cost of 1 {{.*}} %E = call <4 x double> @llvm.fma.v4f64
-  ; AVX2: cost of 1 {{.*}} %E = call <4 x double> @llvm.fma.v4f64
-  ; AVX512: cost of 1 {{.*}} %E = call <4 x double> @llvm.fma.v4f64
-  %E = call <4 x double> @llvm.fma.v4f64(<4 x double> undef, <4 x double> undef, <4 x double> undef)
-  ; SSE2: cost of 96 {{.*}} %F = call <8 x double> @llvm.fma.v8f64
-  ; SSE42: cost of 96 {{.*}} %F = call <8 x double> @llvm.fma.v8f64
-  ; AVX: cost of 4 {{.*}} %F = call <8 x double> @llvm.fma.v8f64
-  ; AVX2: cost of 4 {{.*}} %F = call <8 x double> @llvm.fma.v8f64
-  ; AVX512: cost of 1 {{.*}} %F = call <8 x double> @llvm.fma.v8f64
-  %F = call <8 x double> @llvm.fma.v8f64(<8 x double> undef, <8 x double> undef, <8 x double> undef)
+  ; SSE2: cost of 10 {{.*}} %F32 = call float @llvm.fma.f32
+  ; SSE42: cost of 10 {{.*}} %F32 = call float @llvm.fma.f32
+  ; AVX: cost of 1 {{.*}} %F32 = call float @llvm.fma.f32
+  ; AVX2: cost of 1 {{.*}} %F32 = call float @llvm.fma.f32
+  ; AVX512: cost of 1 {{.*}} %F32 = call float @llvm.fma.f32
+  %F32 = call float @llvm.fma.f32(float undef, float undef, float undef)
+  ; SSE2: cost of 52 {{.*}} %V4F32 = call <4 x float> @llvm.fma.v4f32
+  ; SSE42: cost of 52 {{.*}} %V4F32 = call <4 x float> @llvm.fma.v4f32
+  ; AVX: cost of 1 {{.*}} %V4F32 = call <4 x float> @llvm.fma.v4f32
+  ; AVX2: cost of 1 {{.*}} %V4F32 = call <4 x float> @llvm.fma.v4f32
+  ; AVX512: cost of 1 {{.*}} %V4F32 = call <4 x float> @llvm.fma.v4f32
+  %V4F32 = call <4 x float> @llvm.fma.v4f32(<4 x float> undef, <4 x float> undef, <4 x float> undef)
+  ; SSE2: cost of 104 {{.*}} %V8F32 = call <8 x float> @llvm.fma.v8f32
+  ; SSE42: cost of 104 {{.*}} %V8F32 = call <8 x float> @llvm.fma.v8f32
+  ; AVX: cost of 1 {{.*}} %V8F32 = call <8 x float> @llvm.fma.v8f32
+  ; AVX2: cost of 1 {{.*}} %V8F32 = call <8 x float> @llvm.fma.v8f32
+  ; AVX512: cost of 1 {{.*}} %V8F32 = call <8 x float> @llvm.fma.v8f32
+  %V8F32 = call <8 x float> @llvm.fma.v8f32(<8 x float> undef, <8 x float> undef, <8 x float> undef)
+  ; SSE2: cost of 208 {{.*}} %V16F32 = call <16 x float> @llvm.fma.v16f32
+  ; SSE42: cost of 208 {{.*}} %V16F32 = call <16 x float> @llvm.fma.v16f32
+  ; AVX: cost of 4 {{.*}} %V16F32 = call <16 x float> @llvm.fma.v16f32
+  ; AVX2: cost of 4 {{.*}} %V16F32 = call <16 x float> @llvm.fma.v16f32
+  ; AVX512: cost of 1 {{.*}} %V16F32 = call <16 x float> @llvm.fma.v16f32
+  %V16F32 = call <16 x float> @llvm.fma.v16f32(<16 x float> undef, <16 x float> undef, <16 x float> undef)
+
+  ; SSE2: cost of 10 {{.*}} %F64 = call double @llvm.fma.f64
+  ; SSE42: cost of 10 {{.*}} %F64 = call double @llvm.fma.f64
+  ; AVX: cost of 1 {{.*}} %F64 = call double @llvm.fma.f64
+  ; AVX2: cost of 1 {{.*}} %F64 = call double @llvm.fma.f64
+  ; AVX512: cost of 1 {{.*}} %F64 = call double @llvm.fma.f64
+  %F64 = call double @llvm.fma.f64(double undef, double undef, double undef)
+  ; SSE2: cost of 24 {{.*}} %V2F64 = call <2 x double> @llvm.fma.v2f64
+  ; SSE42: cost of 24 {{.*}} %V2F64 = call <2 x double> @llvm.fma.v2f64
+  ; AVX: cost of 1 {{.*}} %V2F64 = call <2 x double> @llvm.fma.v2f64
+  ; AVX2: cost of 1 {{.*}} %V2F64 = call <2 x double> @llvm.fma.v2f64
+  ; AVX512: cost of 1 {{.*}} %V2F64 = call <2 x double> @llvm.fma.v2f64
+  %V2F64 = call <2 x double> @llvm.fma.v2f64(<2 x double> undef, <2 x double> undef, <2 x double> undef)
+  ; SSE2: cost of 48 {{.*}} %V4F64 = call <4 x double> @llvm.fma.v4f64
+  ; SSE42: cost of 48 {{.*}} %V4F64 = call <4 x double> @llvm.fma.v4f64
+  ; AVX: cost of 1 {{.*}} %V4F64 = call <4 x double> @llvm.fma.v4f64
+  ; AVX2: cost of 1 {{.*}} %V4F64 = call <4 x double> @llvm.fma.v4f64
+  ; AVX512: cost of 1 {{.*}} %V4F64 = call <4 x double> @llvm.fma.v4f64
+  %V4F64 = call <4 x double> @llvm.fma.v4f64(<4 x double> undef, <4 x double> undef, <4 x double> undef)
+  ; SSE2: cost of 96 {{.*}} %V8F64 = call <8 x double> @llvm.fma.v8f64
+  ; SSE42: cost of 96 {{.*}} %V8F64 = call <8 x double> @llvm.fma.v8f64
+  ; AVX: cost of 4 {{.*}} %V8F64 = call <8 x double> @llvm.fma.v8f64
+  ; AVX2: cost of 4 {{.*}} %V8F64 = call <8 x double> @llvm.fma.v8f64
+  ; AVX512: cost of 1 {{.*}} %V8F64 = call <8 x double> @llvm.fma.v8f64
+  %V8F64 = call <8 x double> @llvm.fma.v8f64(<8 x double> undef, <8 x double> undef, <8 x double> undef)
 
   ret i32 undef
 }
 
+declare float @llvm.sqrt.f32(float)
 declare <4 x float> @llvm.sqrt.v4f32(<4 x float>)
 declare <8 x float> @llvm.sqrt.v8f32(<8 x float>)
 declare <16 x float> @llvm.sqrt.v16f32(<16 x float>)
 
+declare double @llvm.sqrt.f64(double)
 declare <2 x double> @llvm.sqrt.v2f64(<2 x double>)
 declare <4 x double> @llvm.sqrt.v4f64(<4 x double>)
 declare <8 x double> @llvm.sqrt.v8f64(<8 x double>)
 
+declare float @llvm.fma.f32(float, float, float)
 declare <4 x float> @llvm.fma.v4f32(<4 x float>, <4 x float>, <4 x float>)
 declare <8 x float> @llvm.fma.v8f32(<8 x float>, <8 x float>, <8 x float>)
 declare <16 x float> @llvm.fma.v16f32(<16 x float>, <16 x float>, <16 x float>)
 
+declare double @llvm.fma.f64(double, double, double)
 declare <2 x double> @llvm.fma.v2f64(<2 x double>, <2 x double>, <2 x double>)
 declare <4 x double> @llvm.fma.v4f64(<4 x double>, <4 x double>, <4 x double>)
 declare <8 x double> @llvm.fma.v8f64(<8 x double>, <8 x double>, <8 x double>)




More information about the llvm-commits mailing list