[llvm] f418344 - Automatically generate aix32-cc-abi-vaarg.ll . NFC

Amaury Séchet via llvm-commits llvm-commits at lists.llvm.org
Sat May 7 06:25:56 PDT 2022


Author: Amaury Séchet
Date: 2022-05-07T13:22:40Z
New Revision: f4183441d45d48edd84a46bafa4deb6c45abc773

URL: https://github.com/llvm/llvm-project/commit/f4183441d45d48edd84a46bafa4deb6c45abc773
DIFF: https://github.com/llvm/llvm-project/commit/f4183441d45d48edd84a46bafa4deb6c45abc773.diff

LOG: Automatically generate aix32-cc-abi-vaarg.ll . NFC

Added: 
    

Modified: 
    llvm/test/CodeGen/PowerPC/aix32-cc-abi-vaarg.ll

Removed: 
    


################################################################################
diff  --git a/llvm/test/CodeGen/PowerPC/aix32-cc-abi-vaarg.ll b/llvm/test/CodeGen/PowerPC/aix32-cc-abi-vaarg.ll
index ce9557ae10edf..990b692715abc 100644
--- a/llvm/test/CodeGen/PowerPC/aix32-cc-abi-vaarg.ll
+++ b/llvm/test/CodeGen/PowerPC/aix32-cc-abi-vaarg.ll
@@ -1,45 +1,57 @@
-; RUN: llc -O2 -mtriple powerpc-ibm-aix-xcoff -stop-after=machine-cp -verify-machineinstrs < %s | \
-; RUN: FileCheck --check-prefix=32BIT %s
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc -O2 -verify-machineinstrs -mcpu=pwr4 -mattr=-altivec -mtriple powerpc-ibm-aix-xcoff < %s | FileCheck --check-prefix=ASM32 %s
+; RUN: llc -O2 -mtriple powerpc-ibm-aix-xcoff -stop-after=machine-cp -verify-machineinstrs < %s | FileCheck --check-prefix=32BIT %s
 
-; RUN: llc -O2 -verify-machineinstrs -mcpu=pwr4 -mattr=-altivec \
-; RUN:  -mtriple powerpc-ibm-aix-xcoff < %s | \
-; RUN: FileCheck --check-prefix=ASM32 %s
-
-  define i32 @int_va_arg(i32 %a, ...) local_unnamed_addr  {
-  entry:
-    %arg1 = alloca i8*, align 4
-    %arg2 = alloca i8*, align 4
-    %0 = bitcast i8** %arg1 to i8*
-    call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %0)
-    %1 = bitcast i8** %arg2 to i8*
-    call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %1)
-    call void @llvm.va_start(i8* nonnull %0)
-    call void @llvm.va_copy(i8* nonnull %1, i8* nonnull %0)
-    %argp.cur = load i8*, i8** %arg1, align 4
-    %argp.next = getelementptr inbounds i8, i8* %argp.cur, i32 4
-    store i8* %argp.next, i8** %arg1, align 4
-    %2 = bitcast i8* %argp.cur to i32*
-    %3 = load i32, i32* %2, align 4
-    %add = add nsw i32 %3, %a
-    %argp.cur2 = load i8*, i8** %arg2, align 4
-    %argp.next3 = getelementptr inbounds i8, i8* %argp.cur2, i32 4
-    store i8* %argp.next3, i8** %arg2, align 4
-    %4 = bitcast i8* %argp.cur2 to i32*
-    %5 = load i32, i32* %4, align 4
-    %mul = shl i32 %5, 1
-    %add4 = add nsw i32 %add, %mul
-    call void @llvm.va_end(i8* nonnull %0)
-    call void @llvm.va_end(i8* nonnull %1)
-    call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %1)
-    call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %0)
-    ret i32 %add4
-  }
-
-  declare void @llvm.lifetime.start.p0i8(i64 immarg, i8* nocapture)
-  declare void @llvm.va_start(i8*)
-  declare void @llvm.va_copy(i8*, i8*)
-  declare void @llvm.va_end(i8*)
-  declare void @llvm.lifetime.end.p0i8(i64 immarg, i8* nocapture)
+define i32 @int_va_arg(i32 %a, ...) local_unnamed_addr  {
+; ASM32-LABEL: int_va_arg:
+; ASM32:       # %bb.0: # %entry
+; ASM32-NEXT:    addi 11, 1, 28
+; ASM32-NEXT:    stw 4, 28(1)
+; ASM32-NEXT:    addi 4, 1, 32
+; ASM32-NEXT:    stw 6, 36(1)
+; ASM32-NEXT:    stw 11, -4(1)
+; ASM32-NEXT:    stw 11, -8(1)
+; ASM32-NEXT:    stw 4, -4(1)
+; ASM32-NEXT:    lwz 6, 28(1)
+; ASM32-NEXT:    stw 4, -8(1)
+; ASM32-NEXT:    add 3, 6, 3
+; ASM32-NEXT:    lwz 4, 28(1)
+; ASM32-NEXT:    slwi 4, 4, 1
+; ASM32-NEXT:    stw 7, 40(1)
+; ASM32-NEXT:    add 3, 3, 4
+; ASM32-NEXT:    stw 8, 44(1)
+; ASM32-NEXT:    stw 9, 48(1)
+; ASM32-NEXT:    stw 10, 52(1)
+; ASM32-NEXT:    stw 5, 32(1)
+; ASM32-NEXT:    blr
+entry:
+  %arg1 = alloca i8*, align 4
+  %arg2 = alloca i8*, align 4
+  %0 = bitcast i8** %arg1 to i8*
+  call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %0)
+  %1 = bitcast i8** %arg2 to i8*
+  call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %1)
+  call void @llvm.va_start(i8* nonnull %0)
+  call void @llvm.va_copy(i8* nonnull %1, i8* nonnull %0)
+  %argp.cur = load i8*, i8** %arg1, align 4
+  %argp.next = getelementptr inbounds i8, i8* %argp.cur, i32 4
+  store i8* %argp.next, i8** %arg1, align 4
+  %2 = bitcast i8* %argp.cur to i32*
+  %3 = load i32, i32* %2, align 4
+  %add = add nsw i32 %3, %a
+  %argp.cur2 = load i8*, i8** %arg2, align 4
+  %argp.next3 = getelementptr inbounds i8, i8* %argp.cur2, i32 4
+  store i8* %argp.next3, i8** %arg2, align 4
+  %4 = bitcast i8* %argp.cur2 to i32*
+  %5 = load i32, i32* %4, align 4
+  %mul = shl i32 %5, 1
+  %add4 = add nsw i32 %add, %mul
+  call void @llvm.va_end(i8* nonnull %0)
+  call void @llvm.va_end(i8* nonnull %1)
+  call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %1)
+  call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %0)
+  ret i32 %add4
+}
 
 ; 32BIT-LABEL:   name:            int_va_arg
 ; 32BIT-LABEL:   liveins:
@@ -80,57 +92,66 @@
 ; 32BIT-DAG:     renamable $r3 = nsw ADD4 killed renamable $r3, killed renamable $r4
 ; 32BIT-DAG:     BLR implicit $lr, implicit $rm, implicit $r3
 
-; ASM32-LABEL:   .int_va_arg:
-; set up fixed stack frame for incoming va_args r4->r10
-; ASM32-DAG:     stw 4, 28(1)
-; ASM32-DAG:     stw 5, 32(1)
-; ASM32-DAG:     stw 6, 36(1)
-; ASM32-DAG:     stw 7, 40(1)
-; ASM32-DAG:     stw 8, 44(1)
-; ASM32-DAG:     stw 9, 48(1)
-; ASM32-DAG:     stw 10, 52(1)
-; load of arg1 from fixed stack offset
-; ASM32-DAG:     lwz [[ARG1:[0-9]+]], 28(1)
-; va_copy load of arg2 from fixed stack offset
-; ASM32-DAG:     lwz [[ARG2:[0-9]+]], 28(1)
-; ASM32-DAG:     blr
+declare void @llvm.lifetime.start.p0i8(i64 immarg, i8* nocapture)
+declare void @llvm.va_start(i8*)
+declare void @llvm.va_copy(i8*, i8*)
+declare void @llvm.va_end(i8*)
+declare void @llvm.lifetime.end.p0i8(i64 immarg, i8* nocapture)
 
-  define i32 @int_stack_va_arg(i32 %one, i32 %two, i32 %three, i32 %four, i32 %five, i32 %six, i32 %seven, i32 %eight, ...) local_unnamed_addr {
-  entry:
-    %arg1 = alloca i8*, align 4
-    %arg2 = alloca i8*, align 4
-    %0 = bitcast i8** %arg1 to i8*
-    call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %0)
-    %1 = bitcast i8** %arg2 to i8*
-    call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %1)
-    call void @llvm.va_start(i8* nonnull %0)
-    call void @llvm.va_copy(i8* nonnull %1, i8* nonnull %0)
-    %add = add nsw i32 %two, %one
-    %add2 = add nsw i32 %add, %three
-    %add3 = add nsw i32 %add2, %four
-    %add4 = add nsw i32 %add3, %five
-    %add5 = add nsw i32 %add4, %six
-    %add6 = add nsw i32 %add5, %seven
-    %add7 = add nsw i32 %add6, %eight
-    %argp.cur = load i8*, i8** %arg1, align 4
-    %argp.next = getelementptr inbounds i8, i8* %argp.cur, i32 4
-    store i8* %argp.next, i8** %arg1, align 4
-    %2 = bitcast i8* %argp.cur to i32*
-    %3 = load i32, i32* %2, align 4
-    %add8 = add nsw i32 %add7, %3
-    %argp.cur9 = load i8*, i8** %arg2, align 4
-    %argp.next10 = getelementptr inbounds i8, i8* %argp.cur9, i32 4
-    store i8* %argp.next10, i8** %arg2, align 4
-    %4 = bitcast i8* %argp.cur9 to i32*
-    %5 = load i32, i32* %4, align 4
-    %mul = shl i32 %5, 1
-    %add11 = add nsw i32 %add8, %mul
-    call void @llvm.va_end(i8* nonnull %0)
-    call void @llvm.va_end(i8* nonnull %1)
-    call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %1)
-    call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %0)
-    ret i32 %add11
-  }
+define i32 @int_stack_va_arg(i32 %one, i32 %two, i32 %three, i32 %four, i32 %five, i32 %six, i32 %seven, i32 %eight, ...) local_unnamed_addr {
+; ASM32-LABEL: int_stack_va_arg:
+; ASM32:       # %bb.0: # %entry
+; ASM32-NEXT:    add 3, 4, 3
+; ASM32-NEXT:    lwz 4, 56(1)
+; ASM32-NEXT:    li 11, 4
+; ASM32-NEXT:    add 3, 3, 5
+; ASM32-NEXT:    addi 12, 1, 56
+; ASM32-NEXT:    add 3, 3, 6
+; ASM32-NEXT:    rlwimi 12, 11, 0, 29, 29
+; ASM32-NEXT:    stw 12, -4(1)
+; ASM32-NEXT:    add 3, 3, 7
+; ASM32-NEXT:    add 3, 3, 8
+; ASM32-NEXT:    add 3, 3, 9
+; ASM32-NEXT:    add 3, 3, 10
+; ASM32-NEXT:    add 3, 3, 4
+; ASM32-NEXT:    slwi 4, 4, 1
+; ASM32-NEXT:    add 3, 3, 4
+; ASM32-NEXT:    blr
+entry:
+  %arg1 = alloca i8*, align 4
+  %arg2 = alloca i8*, align 4
+  %0 = bitcast i8** %arg1 to i8*
+  call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %0)
+  %1 = bitcast i8** %arg2 to i8*
+  call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %1)
+  call void @llvm.va_start(i8* nonnull %0)
+  call void @llvm.va_copy(i8* nonnull %1, i8* nonnull %0)
+  %add = add nsw i32 %two, %one
+  %add2 = add nsw i32 %add, %three
+  %add3 = add nsw i32 %add2, %four
+  %add4 = add nsw i32 %add3, %five
+  %add5 = add nsw i32 %add4, %six
+  %add6 = add nsw i32 %add5, %seven
+  %add7 = add nsw i32 %add6, %eight
+  %argp.cur = load i8*, i8** %arg1, align 4
+  %argp.next = getelementptr inbounds i8, i8* %argp.cur, i32 4
+  store i8* %argp.next, i8** %arg1, align 4
+  %2 = bitcast i8* %argp.cur to i32*
+  %3 = load i32, i32* %2, align 4
+  %add8 = add nsw i32 %add7, %3
+  %argp.cur9 = load i8*, i8** %arg2, align 4
+  %argp.next10 = getelementptr inbounds i8, i8* %argp.cur9, i32 4
+  store i8* %argp.next10, i8** %arg2, align 4
+  %4 = bitcast i8* %argp.cur9 to i32*
+  %5 = load i32, i32* %4, align 4
+  %mul = shl i32 %5, 1
+  %add11 = add nsw i32 %add8, %mul
+  call void @llvm.va_end(i8* nonnull %0)
+  call void @llvm.va_end(i8* nonnull %1)
+  call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %1)
+  call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %0)
+  ret i32 %add11
+}
 
 ; 32BIT-LABEL:   name:            int_stack_va_arg
 ; 32BIT-LABEL:   liveins:
@@ -167,49 +188,56 @@
 ; 32BIT-DAG:     renamable $r11 = LI 4
 ; 32BIT-DAG:     BLR implicit $lr, implicit $rm, implicit $r3
 
-; ASM32-LABEL:   .int_stack_va_arg:
-; ASM32-DAG:     add 3, 4, 3
-; ASM32-DAG:     add 3, 3, 5
-; ASM32-DAG:     add 3, 3, 6
-; ASM32-DAG:     add 3, 3, 7
-; ASM32-DAG:     add 3, 3, 8
-; ASM32-DAG:     add 3, 3, 9
-; ASM32-DAG:     add 3, 3, 10
-; ASM32-DAG:     lwz [[ARG1:[0-9]+]], 56(1)
-; ASM32-DAG:     li [[ARG2:[0-9]+]], [[ARG1]]
-; ASM32-DAG:     add 3, 3, [[ARG1:[0-9]+]]
-; ASM32-DAG:     add 3, 3, [[ARG2:[0-9]+]]
-; ASM32-DAG:     blr
-
-  define double @double_va_arg(double %a, ...) local_unnamed_addr  {
-  entry:
-    %arg1 = alloca i8*, align 4
-    %arg2 = alloca i8*, align 4
-    %0 = bitcast i8** %arg1 to i8*
-    call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %0)
-    %1 = bitcast i8** %arg2 to i8*
-    call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %1)
-    call void @llvm.va_start(i8* nonnull %0)
-    call void @llvm.va_copy(i8* nonnull %1, i8* nonnull %0)
-    %argp.cur = load i8*, i8** %arg1, align 4
-    %argp.next = getelementptr inbounds i8, i8* %argp.cur, i32 8
-    store i8* %argp.next, i8** %arg1, align 4
-    %2 = bitcast i8* %argp.cur to double*
-    %3 = load double, double* %2, align 4
-    %add = fadd double %3, %a
-    %argp.cur2 = load i8*, i8** %arg2, align 4
-    %argp.next3 = getelementptr inbounds i8, i8* %argp.cur2, i32 8
-    store i8* %argp.next3, i8** %arg2, align 4
-    %4 = bitcast i8* %argp.cur2 to double*
-    %5 = load double, double* %4, align 4
-    %mul = fmul double %5, 2.000000e+00
-    %add4 = fadd double %add, %mul
-    call void @llvm.va_end(i8* nonnull %0)
-    call void @llvm.va_end(i8* nonnull %1)
-    call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %1)
-    call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %0)
-    ret double %add4
-  }
+define double @double_va_arg(double %a, ...) local_unnamed_addr  {
+; ASM32-LABEL: double_va_arg:
+; ASM32:       # %bb.0: # %entry
+; ASM32-NEXT:    stw 6, -12(1)
+; ASM32-NEXT:    addi 3, 1, 32
+; ASM32-NEXT:    stw 5, -16(1)
+; ASM32-NEXT:    lfd 0, -16(1)
+; ASM32-NEXT:    stw 6, -20(1)
+; ASM32-NEXT:    fadd 0, 0, 1
+; ASM32-NEXT:    stw 5, -24(1)
+; ASM32-NEXT:    lfd 1, -24(1)
+; ASM32-NEXT:    fadd 1, 1, 1
+; ASM32-NEXT:    stw 7, 40(1)
+; ASM32-NEXT:    fadd 1, 0, 1
+; ASM32-NEXT:    stw 5, 32(1)
+; ASM32-NEXT:    stw 6, 36(1)
+; ASM32-NEXT:    stw 8, 44(1)
+; ASM32-NEXT:    stw 9, 48(1)
+; ASM32-NEXT:    stw 10, 52(1)
+; ASM32-NEXT:    stw 3, -4(1)
+; ASM32-NEXT:    stw 3, -8(1)
+; ASM32-NEXT:    blr
+entry:
+  %arg1 = alloca i8*, align 4
+  %arg2 = alloca i8*, align 4
+  %0 = bitcast i8** %arg1 to i8*
+  call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %0)
+  %1 = bitcast i8** %arg2 to i8*
+  call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %1)
+  call void @llvm.va_start(i8* nonnull %0)
+  call void @llvm.va_copy(i8* nonnull %1, i8* nonnull %0)
+  %argp.cur = load i8*, i8** %arg1, align 4
+  %argp.next = getelementptr inbounds i8, i8* %argp.cur, i32 8
+  store i8* %argp.next, i8** %arg1, align 4
+  %2 = bitcast i8* %argp.cur to double*
+  %3 = load double, double* %2, align 4
+  %add = fadd double %3, %a
+  %argp.cur2 = load i8*, i8** %arg2, align 4
+  %argp.next3 = getelementptr inbounds i8, i8* %argp.cur2, i32 8
+  store i8* %argp.next3, i8** %arg2, align 4
+  %4 = bitcast i8* %argp.cur2 to double*
+  %5 = load double, double* %4, align 4
+  %mul = fmul double %5, 2.000000e+00
+  %add4 = fadd double %add, %mul
+  call void @llvm.va_end(i8* nonnull %0)
+  call void @llvm.va_end(i8* nonnull %1)
+  call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %1)
+  call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %0)
+  ret double %add4
+}
 
 ; 32BIT-LABEL:   name:            double_va_arg
 ; 32BIT-LABEL:   liveins:
@@ -241,59 +269,69 @@
 ; 32BIT-DAG:     STW killed renamable $r3, 0, %stack.1.arg2 :: (store (s32) into %ir.1)
 ; 32BIT-DAG:     BLR implicit $lr, implicit $rm, implicit $f1
 
-; ASM32-LABEL:  .double_va_arg:
-; ASM32-DAG:    stw 5, 32(1)
-; ASM32-DAG:    stw 6, 36(1)
-; ASM32-DAG:    stw 7, 40(1)
-; ASM32-DAG:    stw 8, 44(1)
-; ASM32-DAG:    stw 9, 48(1)
-; ASM32-DAG:    stw 10, 52(1)
-; ASM32-DAG:    stw [[ARG1A:[0-9]+]], -12(1)
-; ASM32-DAG:    stw [[ARG1B:[0-9]+]], -16(1)
-; ASM32-DAG:    stw [[ARG2A:[0-9]+]], -20(1)
-; ASM32-DAG:    stw [[ARG2B:[0-9]+]], -24(1)
-; ASM32-DAG:    lfd [[ARG1:[0-9]+]], -16(1)
-; ASM32-DAG:    fadd 0, [[ARG1]], 1
-; ASM32-DAG:    fadd 1, 1, 1
-; ASM32-DAG:    lfd [[ARG2:[0-9]+]], -24(1)
-; ASM32-DAG:    fadd 1, 0, [[ARG2]]
-; ASM32-DAG:    blr
-
-  define double @double_stack_va_arg(double %one, double %two, double %three, double %four, double %five, double %six, double %seven, double %eight, double %nine, double %ten, double %eleven, double %twelve, double %thirteen, ...) local_unnamed_addr  {
-  entry:
-    %arg1 = alloca i8*, align 4
-    %arg2 = alloca i8*, align 4
-    %0 = bitcast i8** %arg1 to i8*
-    call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %0)
-    %1 = bitcast i8** %arg2 to i8*
-    call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %1)
-    call void @llvm.va_start(i8* nonnull %0)
-    call void @llvm.va_copy(i8* nonnull %1, i8* nonnull %0)
-    %add = fadd double %one, %two
-    %add2 = fadd double %add, %three
-    %add3 = fadd double %add2, %four
-    %add4 = fadd double %add3, %five
-    %add5 = fadd double %add4, %six
-    %add6 = fadd double %add5, %seven
-    %add7 = fadd double %add6, %eight
-    %add8 = fadd double %add7, %nine
-    %add9 = fadd double %add8, %ten
-    %add10 = fadd double %add9, %eleven
-    %add11 = fadd double %add10, %twelve
-    %add12 = fadd double %add11, %thirteen
-    %2 = bitcast i8** %arg1 to double**
-    %argp.cur1 = load double*, double** %2, align 4
-    %3 = load double, double* %argp.cur1, align 4
-    %add13 = fadd double %add12, %3
-    %4 = bitcast i8** %arg2 to double**
-    %argp.cur142 = load double*, double** %4, align 4
-    %5 = load double, double* %argp.cur142, align 4
-    %mul = fmul double %5, 2.000000e+00
-    %add16 = fadd double %add13, %mul
-    call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %1)
-    call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %0)
-    ret double %add16
-  }
+define double @double_stack_va_arg(double %one, double %two, double %three, double %four, double %five, double %six, double %seven, double %eight, double %nine, double %ten, double %eleven, double %twelve, double %thirteen, ...) local_unnamed_addr  {
+; ASM32-LABEL: double_stack_va_arg:
+; ASM32:       # %bb.0: # %entry
+; ASM32-NEXT:    fadd 0, 1, 2
+; ASM32-NEXT:    addi 3, 1, 128
+; ASM32-NEXT:    lwz 4, 132(1)
+; ASM32-NEXT:    fadd 0, 0, 3
+; ASM32-NEXT:    stw 3, -4(1)
+; ASM32-NEXT:    fadd 0, 0, 4
+; ASM32-NEXT:    lwz 3, 128(1)
+; ASM32-NEXT:    fadd 0, 0, 5
+; ASM32-NEXT:    stw 3, -16(1)
+; ASM32-NEXT:    fadd 0, 0, 6
+; ASM32-NEXT:    stw 4, -12(1)
+; ASM32-NEXT:    fadd 0, 0, 7
+; ASM32-NEXT:    lfd 1, -16(1)
+; ASM32-NEXT:    fadd 0, 0, 8
+; ASM32-NEXT:    stw 3, -24(1)
+; ASM32-NEXT:    fadd 0, 0, 9
+; ASM32-NEXT:    stw 4, -20(1)
+; ASM32-NEXT:    fadd 0, 0, 10
+; ASM32-NEXT:    fadd 0, 0, 11
+; ASM32-NEXT:    fadd 0, 0, 12
+; ASM32-NEXT:    fadd 0, 0, 13
+; ASM32-NEXT:    fadd 0, 0, 1
+; ASM32-NEXT:    lfd 1, -24(1)
+; ASM32-NEXT:    fadd 1, 1, 1
+; ASM32-NEXT:    fadd 1, 0, 1
+; ASM32-NEXT:    blr
+entry:
+  %arg1 = alloca i8*, align 4
+  %arg2 = alloca i8*, align 4
+  %0 = bitcast i8** %arg1 to i8*
+  call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %0)
+  %1 = bitcast i8** %arg2 to i8*
+  call void @llvm.lifetime.start.p0i8(i64 4, i8* nonnull %1)
+  call void @llvm.va_start(i8* nonnull %0)
+  call void @llvm.va_copy(i8* nonnull %1, i8* nonnull %0)
+  %add = fadd double %one, %two
+  %add2 = fadd double %add, %three
+  %add3 = fadd double %add2, %four
+  %add4 = fadd double %add3, %five
+  %add5 = fadd double %add4, %six
+  %add6 = fadd double %add5, %seven
+  %add7 = fadd double %add6, %eight
+  %add8 = fadd double %add7, %nine
+  %add9 = fadd double %add8, %ten
+  %add10 = fadd double %add9, %eleven
+  %add11 = fadd double %add10, %twelve
+  %add12 = fadd double %add11, %thirteen
+  %2 = bitcast i8** %arg1 to double**
+  %argp.cur1 = load double*, double** %2, align 4
+  %3 = load double, double* %argp.cur1, align 4
+  %add13 = fadd double %add12, %3
+  %4 = bitcast i8** %arg2 to double**
+  %argp.cur142 = load double*, double** %4, align 4
+  %5 = load double, double* %argp.cur142, align 4
+  %mul = fmul double %5, 2.000000e+00
+  %add16 = fadd double %add13, %mul
+  call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %1)
+  call void @llvm.lifetime.end.p0i8(i64 4, i8* nonnull %0)
+  ret double %add16
+}
 
 ; 32BIT-LABEL:   name:            double_stack_va_arg
 ; 32BIT-LABEL:   liveins:
@@ -349,23 +387,3 @@
 ; 32BIT-DAG:     renamable $f1 = nofpexcept FADD killed renamable $f2, renamable $f2, implicit $rm
 ; 32BIT-DAG:     BLR implicit $lr, implicit $rm, implicit $f1
 
-; ASM32-LABEL:  .double_stack_va_arg:
-; ASM32-DAG:    fadd 0, 1, 2
-; ASM32-DAG:    fadd 0, 0, 3
-; ASM32-DAG:    fadd 0, 0, 4
-; ASM32-DAG:    fadd 0, 0, 5
-; ASM32-DAG:    fadd 0, 0, 6
-; ASM32-DAG:    fadd 0, 0, 7
-; ASM32-DAG:    fadd 0, 0, 8
-; ASM32-DAG:    fadd 0, 0, 9
-; ASM32-DAG:    fadd 0, 0, 10
-; ASM32-DAG:    fadd 0, 0, 11
-; ASM32-DAG:    fadd 0, 0, 12
-; ASM32-DAG:    fadd 0, 0, 13
-; ASM32-DAG:    lwz [[ARG1:[0-9]+]], 128(1)
-; ASM32-DAG:    lwz [[ARG2:[0-9]+]], 132(1)
-; ASM32-DAG:    lfd [[ARG1:[0-9]+]], -16(1)
-; ASM32-DAG:    lfd [[ARG2:[0-9]+]], -24(1)
-; ASM32-DAG:    fadd 0, 0, [[ARG1]]
-; ASM32-DAG:    fadd [[ARG1]], 0, [[ARG2]]
-; ASM32-DAG:    blr


        


More information about the llvm-commits mailing list