[llvm] 60a5799 - [X86] Fix intrinsic names in strict fp80 tests to use f80 in their names instead of x86_fp80.
Craig Topper via llvm-commits
llvm-commits at lists.llvm.org
Sat Jul 25 13:31:08 PDT 2020
Author: Craig Topper
Date: 2020-07-25T13:12:49-07:00
New Revision: 60a5799e6e8b1206640a2321eab60dcf1b65975d
URL: https://github.com/llvm/llvm-project/commit/60a5799e6e8b1206640a2321eab60dcf1b65975d
DIFF: https://github.com/llvm/llvm-project/commit/60a5799e6e8b1206640a2321eab60dcf1b65975d.diff
LOG: [X86] Fix intrinsic names in strict fp80 tests to use f80 in their names instead of x86_fp80.
The type is called x86_fp80, but when it is printed in the intrinsic
name it should be f80. The parser doesn't seem to care that the
name was wrong.
Added:
Modified:
llvm/test/CodeGen/X86/fp80-strict-scalar-cmp.ll
llvm/test/CodeGen/X86/fp80-strict-scalar.ll
Removed:
################################################################################
diff --git a/llvm/test/CodeGen/X86/fp80-strict-scalar-cmp.ll b/llvm/test/CodeGen/X86/fp80-strict-scalar-cmp.ll
index 1e38b6744f3c..a3951f826fe0 100644
--- a/llvm/test/CodeGen/X86/fp80-strict-scalar-cmp.ll
+++ b/llvm/test/CodeGen/X86/fp80-strict-scalar-cmp.ll
@@ -33,7 +33,7 @@ define i32 @test_oeq_q(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: cmovnel %esi, %eax
; X87-64-NEXT: cmovpl %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmp.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmp.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"oeq",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -70,7 +70,7 @@ define i32 @test_ogt_q(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovbel %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmp.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmp.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"ogt",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -107,7 +107,7 @@ define i32 @test_oge_q(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovbl %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmp.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmp.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"oge",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -144,7 +144,7 @@ define i32 @test_olt_q(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovbel %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmp.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmp.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"olt",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -181,7 +181,7 @@ define i32 @test_ole_q(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovbl %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmp.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmp.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"ole",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -218,7 +218,7 @@ define i32 @test_one_q(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovel %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmp.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmp.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"one",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -255,7 +255,7 @@ define i32 @test_ord_q(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovpl %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmp.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmp.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"ord",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -292,7 +292,7 @@ define i32 @test_ueq_q(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovnel %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmp.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmp.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"ueq",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -329,7 +329,7 @@ define i32 @test_ugt_q(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovael %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmp.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmp.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"ugt",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -366,7 +366,7 @@ define i32 @test_uge_q(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmoval %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmp.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmp.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"uge",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -403,7 +403,7 @@ define i32 @test_ult_q(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovael %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmp.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmp.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"ult",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -440,7 +440,7 @@ define i32 @test_ule_q(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmoval %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmp.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmp.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"ule",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -478,7 +478,7 @@ define i32 @test_une_q(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: cmovnel %edi, %eax
; X87-64-NEXT: cmovpl %edi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmp.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmp.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"une",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -515,7 +515,7 @@ define i32 @test_uno_q(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovnpl %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmp.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmp.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"uno",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -553,7 +553,7 @@ define i32 @test_oeq_s(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: cmovnel %esi, %eax
; X87-64-NEXT: cmovpl %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmps.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmps.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"oeq",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -590,7 +590,7 @@ define i32 @test_ogt_s(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovbel %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmps.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmps.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"ogt",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -627,7 +627,7 @@ define i32 @test_oge_s(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovbl %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmps.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmps.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"oge",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -664,7 +664,7 @@ define i32 @test_olt_s(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovbel %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmps.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmps.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"olt",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -701,7 +701,7 @@ define i32 @test_ole_s(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovbl %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmps.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmps.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"ole",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -738,7 +738,7 @@ define i32 @test_one_s(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovel %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmps.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmps.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"one",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -775,7 +775,7 @@ define i32 @test_ord_s(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovpl %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmps.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmps.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"ord",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -812,7 +812,7 @@ define i32 @test_ueq_s(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovnel %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmps.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmps.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"ueq",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -849,7 +849,7 @@ define i32 @test_ugt_s(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovael %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmps.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmps.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"ugt",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -886,7 +886,7 @@ define i32 @test_uge_s(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmoval %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmps.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmps.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"uge",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -923,7 +923,7 @@ define i32 @test_ult_s(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovael %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmps.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmps.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"ult",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -960,7 +960,7 @@ define i32 @test_ule_s(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmoval %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmps.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmps.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"ule",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -998,7 +998,7 @@ define i32 @test_une_s(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: cmovnel %edi, %eax
; X87-64-NEXT: cmovpl %edi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmps.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmps.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"une",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -1035,7 +1035,7 @@ define i32 @test_uno_s(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
; X87-64-NEXT: wait
; X87-64-NEXT: cmovnpl %esi, %eax
; X87-64-NEXT: retq
- %cond = call i1 @llvm.experimental.constrained.fcmps.x86_fp80(
+ %cond = call i1 @llvm.experimental.constrained.fcmps.f80(
x86_fp80 %f1, x86_fp80 %f2, metadata !"uno",
metadata !"fpexcept.strict") #0
%res = select i1 %cond, i32 %a, i32 %b
@@ -1044,5 +1044,5 @@ define i32 @test_uno_s(i32 %a, i32 %b, x86_fp80 %f1, x86_fp80 %f2) #0 {
attributes #0 = { strictfp }
-declare i1 @llvm.experimental.constrained.fcmp.x86_fp80(x86_fp80, x86_fp80, metadata, metadata)
-declare i1 @llvm.experimental.constrained.fcmps.x86_fp80(x86_fp80, x86_fp80, metadata, metadata)
+declare i1 @llvm.experimental.constrained.fcmp.f80(x86_fp80, x86_fp80, metadata, metadata)
+declare i1 @llvm.experimental.constrained.fcmps.f80(x86_fp80, x86_fp80, metadata, metadata)
diff --git a/llvm/test/CodeGen/X86/fp80-strict-scalar.ll b/llvm/test/CodeGen/X86/fp80-strict-scalar.ll
index cf4a51fd6920..221bebea2957 100644
--- a/llvm/test/CodeGen/X86/fp80-strict-scalar.ll
+++ b/llvm/test/CodeGen/X86/fp80-strict-scalar.ll
@@ -2,35 +2,35 @@
; RUN: llc < %s -mtriple=i686-unknown-unknown -O3 | FileCheck %s --check-prefixes=CHECK,X86
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -O3 | FileCheck %s --check-prefixes=CHECK,X64
-declare x86_fp80 @llvm.experimental.constrained.fadd.x86_fp80(x86_fp80, x86_fp80, metadata, metadata)
-declare x86_fp80 @llvm.experimental.constrained.fsub.x86_fp80(x86_fp80, x86_fp80, metadata, metadata)
-declare x86_fp80 @llvm.experimental.constrained.fmul.x86_fp80(x86_fp80, x86_fp80, metadata, metadata)
-declare x86_fp80 @llvm.experimental.constrained.fdiv.x86_fp80(x86_fp80, x86_fp80, metadata, metadata)
-declare x86_fp80 @llvm.experimental.constrained.fpext.x86_fp80.f32(float, metadata)
-declare x86_fp80 @llvm.experimental.constrained.fpext.x86_fp80.f64(double, metadata)
-declare x86_fp80 @llvm.experimental.constrained.sqrt.x86_fp80(x86_fp80, metadata, metadata)
-declare float @llvm.experimental.constrained.fptrunc.f32.x86_fp80(x86_fp80, metadata, metadata)
-declare double @llvm.experimental.constrained.fptrunc.f64.x86_fp80(x86_fp80, metadata, metadata)
-declare i1 @llvm.experimental.constrained.fptosi.i1.x86_fp80(x86_fp80, metadata)
-declare i8 @llvm.experimental.constrained.fptosi.i8.x86_fp80(x86_fp80, metadata)
-declare i16 @llvm.experimental.constrained.fptosi.i16.x86_fp80(x86_fp80, metadata)
-declare i32 @llvm.experimental.constrained.fptosi.i32.x86_fp80(x86_fp80, metadata)
-declare i64 @llvm.experimental.constrained.fptosi.i64.x86_fp80(x86_fp80, metadata)
-declare i1 @llvm.experimental.constrained.fptoui.i1.x86_fp80(x86_fp80, metadata)
-declare i8 @llvm.experimental.constrained.fptoui.i8.x86_fp80(x86_fp80, metadata)
-declare i16 @llvm.experimental.constrained.fptoui.i16.x86_fp80(x86_fp80, metadata)
-declare i32 @llvm.experimental.constrained.fptoui.i32.x86_fp80(x86_fp80, metadata)
-declare i64 @llvm.experimental.constrained.fptoui.i64.x86_fp80(x86_fp80, metadata)
-declare x86_fp80 @llvm.experimental.constrained.sitofp.x86_fp80.i1(i1, metadata, metadata)
-declare x86_fp80 @llvm.experimental.constrained.sitofp.x86_fp80.i8(i8, metadata, metadata)
-declare x86_fp80 @llvm.experimental.constrained.sitofp.x86_fp80.i16(i16, metadata, metadata)
-declare x86_fp80 @llvm.experimental.constrained.sitofp.x86_fp80.i32(i32, metadata, metadata)
-declare x86_fp80 @llvm.experimental.constrained.sitofp.x86_fp80.i64(i64, metadata, metadata)
-declare x86_fp80 @llvm.experimental.constrained.uitofp.x86_fp80.i1(i1, metadata, metadata)
-declare x86_fp80 @llvm.experimental.constrained.uitofp.x86_fp80.i8(i8, metadata, metadata)
-declare x86_fp80 @llvm.experimental.constrained.uitofp.x86_fp80.i16(i16, metadata, metadata)
-declare x86_fp80 @llvm.experimental.constrained.uitofp.x86_fp80.i32(i32, metadata, metadata)
-declare x86_fp80 @llvm.experimental.constrained.uitofp.x86_fp80.i64(i64, metadata, metadata)
+declare x86_fp80 @llvm.experimental.constrained.fadd.f80(x86_fp80, x86_fp80, metadata, metadata)
+declare x86_fp80 @llvm.experimental.constrained.fsub.f80(x86_fp80, x86_fp80, metadata, metadata)
+declare x86_fp80 @llvm.experimental.constrained.fmul.f80(x86_fp80, x86_fp80, metadata, metadata)
+declare x86_fp80 @llvm.experimental.constrained.fdiv.f80(x86_fp80, x86_fp80, metadata, metadata)
+declare x86_fp80 @llvm.experimental.constrained.fpext.f80.f32(float, metadata)
+declare x86_fp80 @llvm.experimental.constrained.fpext.f80.f64(double, metadata)
+declare x86_fp80 @llvm.experimental.constrained.sqrt.f80(x86_fp80, metadata, metadata)
+declare float @llvm.experimental.constrained.fptrunc.f32.f80(x86_fp80, metadata, metadata)
+declare double @llvm.experimental.constrained.fptrunc.f64.f80(x86_fp80, metadata, metadata)
+declare i1 @llvm.experimental.constrained.fptosi.i1.f80(x86_fp80, metadata)
+declare i8 @llvm.experimental.constrained.fptosi.i8.f80(x86_fp80, metadata)
+declare i16 @llvm.experimental.constrained.fptosi.i16.f80(x86_fp80, metadata)
+declare i32 @llvm.experimental.constrained.fptosi.i32.f80(x86_fp80, metadata)
+declare i64 @llvm.experimental.constrained.fptosi.i64.f80(x86_fp80, metadata)
+declare i1 @llvm.experimental.constrained.fptoui.i1.f80(x86_fp80, metadata)
+declare i8 @llvm.experimental.constrained.fptoui.i8.f80(x86_fp80, metadata)
+declare i16 @llvm.experimental.constrained.fptoui.i16.f80(x86_fp80, metadata)
+declare i32 @llvm.experimental.constrained.fptoui.i32.f80(x86_fp80, metadata)
+declare i64 @llvm.experimental.constrained.fptoui.i64.f80(x86_fp80, metadata)
+declare x86_fp80 @llvm.experimental.constrained.sitofp.f80.i1(i1, metadata, metadata)
+declare x86_fp80 @llvm.experimental.constrained.sitofp.f80.i8(i8, metadata, metadata)
+declare x86_fp80 @llvm.experimental.constrained.sitofp.f80.i16(i16, metadata, metadata)
+declare x86_fp80 @llvm.experimental.constrained.sitofp.f80.i32(i32, metadata, metadata)
+declare x86_fp80 @llvm.experimental.constrained.sitofp.f80.i64(i64, metadata, metadata)
+declare x86_fp80 @llvm.experimental.constrained.uitofp.f80.i1(i1, metadata, metadata)
+declare x86_fp80 @llvm.experimental.constrained.uitofp.f80.i8(i8, metadata, metadata)
+declare x86_fp80 @llvm.experimental.constrained.uitofp.f80.i16(i16, metadata, metadata)
+declare x86_fp80 @llvm.experimental.constrained.uitofp.f80.i32(i32, metadata, metadata)
+declare x86_fp80 @llvm.experimental.constrained.uitofp.f80.i64(i64, metadata, metadata)
define x86_fp80 @fadd_fp80(x86_fp80 %a, x86_fp80 %b) nounwind strictfp {
; X86-LABEL: fadd_fp80:
@@ -48,7 +48,7 @@ define x86_fp80 @fadd_fp80(x86_fp80 %a, x86_fp80 %b) nounwind strictfp {
; X64-NEXT: faddp %st, %st(1)
; X64-NEXT: wait
; X64-NEXT: retq
- %ret = call x86_fp80 @llvm.experimental.constrained.fadd.x86_fp80(x86_fp80 %a, x86_fp80 %b,
+ %ret = call x86_fp80 @llvm.experimental.constrained.fadd.f80(x86_fp80 %a, x86_fp80 %b,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret x86_fp80 %ret
@@ -70,7 +70,7 @@ define x86_fp80 @fsub_fp80(x86_fp80 %a, x86_fp80 %b) nounwind strictfp {
; X64-NEXT: fsubp %st, %st(1)
; X64-NEXT: wait
; X64-NEXT: retq
- %ret = call x86_fp80 @llvm.experimental.constrained.fsub.x86_fp80(x86_fp80 %a, x86_fp80 %b,
+ %ret = call x86_fp80 @llvm.experimental.constrained.fsub.f80(x86_fp80 %a, x86_fp80 %b,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret x86_fp80 %ret
@@ -92,7 +92,7 @@ define x86_fp80 @fmul_fp80(x86_fp80 %a, x86_fp80 %b) nounwind strictfp {
; X64-NEXT: fmulp %st, %st(1)
; X64-NEXT: wait
; X64-NEXT: retq
- %ret = call x86_fp80 @llvm.experimental.constrained.fmul.x86_fp80(x86_fp80 %a, x86_fp80 %b,
+ %ret = call x86_fp80 @llvm.experimental.constrained.fmul.f80(x86_fp80 %a, x86_fp80 %b,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret x86_fp80 %ret
@@ -114,7 +114,7 @@ define x86_fp80 @fdiv_fp80(x86_fp80 %a, x86_fp80 %b) nounwind strictfp {
; X64-NEXT: fdivp %st, %st(1)
; X64-NEXT: wait
; X64-NEXT: retq
- %ret = call x86_fp80 @llvm.experimental.constrained.fdiv.x86_fp80(x86_fp80 %a, x86_fp80 %b,
+ %ret = call x86_fp80 @llvm.experimental.constrained.fdiv.f80(x86_fp80 %a, x86_fp80 %b,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret x86_fp80 %ret
@@ -133,7 +133,7 @@ define x86_fp80 @fpext_f32_to_fp80(float %a) nounwind strictfp {
; X64-NEXT: flds -{{[0-9]+}}(%rsp)
; X64-NEXT: wait
; X64-NEXT: retq
- %ret = call x86_fp80 @llvm.experimental.constrained.fpext.x86_fp80.f32(float %a,
+ %ret = call x86_fp80 @llvm.experimental.constrained.fpext.f80.f32(float %a,
metadata !"fpexcept.strict") #0
ret x86_fp80 %ret
}
@@ -151,7 +151,7 @@ define x86_fp80 @fpext_f64_to_fp80(double %a) nounwind strictfp {
; X64-NEXT: fldl -{{[0-9]+}}(%rsp)
; X64-NEXT: wait
; X64-NEXT: retq
- %ret = call x86_fp80 @llvm.experimental.constrained.fpext.x86_fp80.f64(double %a,
+ %ret = call x86_fp80 @llvm.experimental.constrained.fpext.f80.f64(double %a,
metadata !"fpexcept.strict") #0
ret x86_fp80 %ret
}
@@ -174,7 +174,7 @@ define float @fptrunc_fp80_to_f32(x86_fp80 %a) nounwind strictfp {
; X64-NEXT: wait
; X64-NEXT: movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
; X64-NEXT: retq
- %ret = call float @llvm.experimental.constrained.fptrunc.f32.x86_fp80(x86_fp80 %a,
+ %ret = call float @llvm.experimental.constrained.fptrunc.f32.f80(x86_fp80 %a,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret float %ret
@@ -202,7 +202,7 @@ define double @fptrunc_fp80_to_f64(x86_fp80 %a) nounwind strictfp {
; X64-NEXT: wait
; X64-NEXT: movsd {{.*#+}} xmm0 = mem[0],zero
; X64-NEXT: retq
- %ret = call double @llvm.experimental.constrained.fptrunc.f64.x86_fp80(x86_fp80 %a,
+ %ret = call double @llvm.experimental.constrained.fptrunc.f64.f80(x86_fp80 %a,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret double %ret
@@ -222,7 +222,7 @@ define x86_fp80 @fsqrt_fp80(x86_fp80 %a) nounwind strictfp {
; X64-NEXT: fsqrt
; X64-NEXT: wait
; X64-NEXT: retq
- %ret = call x86_fp80 @llvm.experimental.constrained.sqrt.x86_fp80(x86_fp80 %a,
+ %ret = call x86_fp80 @llvm.experimental.constrained.sqrt.f80(x86_fp80 %a,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret x86_fp80 %ret
@@ -260,7 +260,7 @@ define i1 @fp80_to_sint1(x86_fp80 %x) #0 {
; X64-NEXT: fldcw -{{[0-9]+}}(%rsp)
; X64-NEXT: movb -{{[0-9]+}}(%rsp), %al
; X64-NEXT: retq
- %result = call i1 @llvm.experimental.constrained.fptosi.i1.x86_fp80(x86_fp80 %x,
+ %result = call i1 @llvm.experimental.constrained.fptosi.i1.f80(x86_fp80 %x,
metadata !"fpexcept.strict") #0
ret i1 %result
}
@@ -297,7 +297,7 @@ define i8 @fp80_to_sint8(x86_fp80 %x) #0 {
; X64-NEXT: fldcw -{{[0-9]+}}(%rsp)
; X64-NEXT: movb -{{[0-9]+}}(%rsp), %al
; X64-NEXT: retq
- %result = call i8 @llvm.experimental.constrained.fptosi.i8.x86_fp80(x86_fp80 %x,
+ %result = call i8 @llvm.experimental.constrained.fptosi.i8.f80(x86_fp80 %x,
metadata !"fpexcept.strict") #0
ret i8 %result
}
@@ -334,7 +334,7 @@ define i16 @fp80_to_sint16(x86_fp80 %x) #0 {
; X64-NEXT: fldcw -{{[0-9]+}}(%rsp)
; X64-NEXT: movzwl -{{[0-9]+}}(%rsp), %eax
; X64-NEXT: retq
- %result = call i16 @llvm.experimental.constrained.fptosi.i16.x86_fp80(x86_fp80 %x,
+ %result = call i16 @llvm.experimental.constrained.fptosi.i16.f80(x86_fp80 %x,
metadata !"fpexcept.strict") #0
ret i16 %result
}
@@ -372,7 +372,7 @@ define i32 @fp80_to_sint32(x86_fp80 %x) #0 {
; X64-NEXT: movl -{{[0-9]+}}(%rsp), %eax
; X64-NEXT: retq
entry:
- %result = call i32 @llvm.experimental.constrained.fptosi.i32.x86_fp80(x86_fp80 %x,
+ %result = call i32 @llvm.experimental.constrained.fptosi.i32.f80(x86_fp80 %x,
metadata !"fpexcept.strict") #0
ret i32 %result
}
@@ -416,7 +416,7 @@ define i64 @fp80_to_sint64(x86_fp80 %x) #0 {
; X64-NEXT: fldcw -{{[0-9]+}}(%rsp)
; X64-NEXT: movq -{{[0-9]+}}(%rsp), %rax
; X64-NEXT: retq
- %result = call i64 @llvm.experimental.constrained.fptosi.i64.x86_fp80(x86_fp80 %x,
+ %result = call i64 @llvm.experimental.constrained.fptosi.i64.f80(x86_fp80 %x,
metadata !"fpexcept.strict") #0
ret i64 %result
}
@@ -453,7 +453,7 @@ define i1 @fp80_to_uint1(x86_fp80 %x) #0 {
; X64-NEXT: fldcw -{{[0-9]+}}(%rsp)
; X64-NEXT: movb -{{[0-9]+}}(%rsp), %al
; X64-NEXT: retq
- %result = call i1 @llvm.experimental.constrained.fptoui.i1.x86_fp80(x86_fp80 %x,
+ %result = call i1 @llvm.experimental.constrained.fptoui.i1.f80(x86_fp80 %x,
metadata !"fpexcept.strict") #0
ret i1 %result
}
@@ -490,7 +490,7 @@ define i8 @fp80_to_uint8(x86_fp80 %x) #0 {
; X64-NEXT: fldcw -{{[0-9]+}}(%rsp)
; X64-NEXT: movb -{{[0-9]+}}(%rsp), %al
; X64-NEXT: retq
- %result = call i8 @llvm.experimental.constrained.fptoui.i8.x86_fp80(x86_fp80 %x,
+ %result = call i8 @llvm.experimental.constrained.fptoui.i8.f80(x86_fp80 %x,
metadata !"fpexcept.strict") #0
ret i8 %result
}
@@ -529,7 +529,7 @@ define i16 @fp80_to_uint16(x86_fp80 %x) #0 {
; X64-NEXT: movl -{{[0-9]+}}(%rsp), %eax
; X64-NEXT: # kill: def $ax killed $ax killed $eax
; X64-NEXT: retq
- %result = call i16 @llvm.experimental.constrained.fptoui.i16.x86_fp80(x86_fp80 %x,
+ %result = call i16 @llvm.experimental.constrained.fptoui.i16.f80(x86_fp80 %x,
metadata !"fpexcept.strict") #0
ret i16 %result
}
@@ -572,7 +572,7 @@ define i32 @fp80_to_uint32(x86_fp80 %x) #0 {
; X64-NEXT: fldcw -{{[0-9]+}}(%rsp)
; X64-NEXT: movl -{{[0-9]+}}(%rsp), %eax
; X64-NEXT: retq
- %result = call i32 @llvm.experimental.constrained.fptoui.i32.x86_fp80(x86_fp80 %x,
+ %result = call i32 @llvm.experimental.constrained.fptoui.i32.f80(x86_fp80 %x,
metadata !"fpexcept.strict") #0
ret i32 %result
}
@@ -647,7 +647,7 @@ define i64 @fp80_to_uint64(x86_fp80 %x) #0 {
; X64-NEXT: shlq $63, %rax
; X64-NEXT: xorq -{{[0-9]+}}(%rsp), %rax
; X64-NEXT: retq
- %result = call i64 @llvm.experimental.constrained.fptoui.i64.x86_fp80(x86_fp80 %x,
+ %result = call i64 @llvm.experimental.constrained.fptoui.i64.f80(x86_fp80 %x,
metadata !"fpexcept.strict") #0
ret i64 %result
}
@@ -677,7 +677,7 @@ define x86_fp80 @sint1_to_fp80(i1 %x) #0 {
; X64-NEXT: filds -{{[0-9]+}}(%rsp)
; X64-NEXT: wait
; X64-NEXT: retq
- %result = call x86_fp80 @llvm.experimental.constrained.sitofp.x86_fp80.i1(i1 %x,
+ %result = call x86_fp80 @llvm.experimental.constrained.sitofp.f80.i1(i1 %x,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret x86_fp80 %result
@@ -703,7 +703,7 @@ define x86_fp80 @sint8_to_fp80(i8 %x) #0 {
; X64-NEXT: filds -{{[0-9]+}}(%rsp)
; X64-NEXT: wait
; X64-NEXT: retq
- %result = call x86_fp80 @llvm.experimental.constrained.sitofp.x86_fp80.i8(i8 %x,
+ %result = call x86_fp80 @llvm.experimental.constrained.sitofp.f80.i8(i8 %x,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret x86_fp80 %result
@@ -728,7 +728,7 @@ define x86_fp80 @sint16_to_fp80(i16 %x) #0 {
; X64-NEXT: filds -{{[0-9]+}}(%rsp)
; X64-NEXT: wait
; X64-NEXT: retq
- %result = call x86_fp80 @llvm.experimental.constrained.sitofp.x86_fp80.i16(i16 %x,
+ %result = call x86_fp80 @llvm.experimental.constrained.sitofp.f80.i16(i16 %x,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret x86_fp80 %result
@@ -753,7 +753,7 @@ define x86_fp80 @sint32_to_fp80(i32 %x) #0 {
; X64-NEXT: fildl -{{[0-9]+}}(%rsp)
; X64-NEXT: wait
; X64-NEXT: retq
- %result = call x86_fp80 @llvm.experimental.constrained.sitofp.x86_fp80.i32(i32 %x,
+ %result = call x86_fp80 @llvm.experimental.constrained.sitofp.f80.i32(i32 %x,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret x86_fp80 %result
@@ -772,7 +772,7 @@ define x86_fp80 @sint64_to_fp80(i64 %x) #0 {
; X64-NEXT: fildll -{{[0-9]+}}(%rsp)
; X64-NEXT: wait
; X64-NEXT: retq
- %result = call x86_fp80 @llvm.experimental.constrained.sitofp.x86_fp80.i64(i64 %x,
+ %result = call x86_fp80 @llvm.experimental.constrained.sitofp.f80.i64(i64 %x,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret x86_fp80 %result
@@ -800,7 +800,7 @@ define x86_fp80 @uint1_to_fp80(i1 %x) #0 {
; X64-NEXT: filds -{{[0-9]+}}(%rsp)
; X64-NEXT: wait
; X64-NEXT: retq
- %result = call x86_fp80 @llvm.experimental.constrained.uitofp.x86_fp80.i1(i1 %x,
+ %result = call x86_fp80 @llvm.experimental.constrained.uitofp.f80.i1(i1 %x,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret x86_fp80 %result
@@ -826,7 +826,7 @@ define x86_fp80 @uint8_to_fp80(i8 %x) #0 {
; X64-NEXT: filds -{{[0-9]+}}(%rsp)
; X64-NEXT: wait
; X64-NEXT: retq
- %result = call x86_fp80 @llvm.experimental.constrained.uitofp.x86_fp80.i8(i8 %x,
+ %result = call x86_fp80 @llvm.experimental.constrained.uitofp.f80.i8(i8 %x,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret x86_fp80 %result
@@ -852,7 +852,7 @@ define x86_fp80 @uint16_to_fp80(i16 %x) #0 {
; X64-NEXT: fildl -{{[0-9]+}}(%rsp)
; X64-NEXT: wait
; X64-NEXT: retq
- %result = call x86_fp80 @llvm.experimental.constrained.uitofp.x86_fp80.i16(i16 %x,
+ %result = call x86_fp80 @llvm.experimental.constrained.uitofp.f80.i16(i16 %x,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret x86_fp80 %result
@@ -885,7 +885,7 @@ define x86_fp80 @uint32_to_fp80(i32 %x) #0 {
; X64-NEXT: fildll -{{[0-9]+}}(%rsp)
; X64-NEXT: wait
; X64-NEXT: retq
- %result = call x86_fp80 @llvm.experimental.constrained.uitofp.x86_fp80.i32(i32 %x,
+ %result = call x86_fp80 @llvm.experimental.constrained.uitofp.f80.i32(i32 %x,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret x86_fp80 %result
@@ -924,7 +924,7 @@ define x86_fp80 @uint64_to_fp80(i64 %x) #0 {
; X64-NEXT: fadds {{\.LCPI.*}}(,%rax,4)
; X64-NEXT: wait
; X64-NEXT: retq
- %result = call x86_fp80 @llvm.experimental.constrained.uitofp.x86_fp80.i64(i64 %x,
+ %result = call x86_fp80 @llvm.experimental.constrained.uitofp.f80.i64(i64 %x,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret x86_fp80 %result
More information about the llvm-commits
mailing list