/* mculib libgcc routines of Andes NDS32 cpu for GNU compiler
Copyright (C) 2012-2022 Free Software Foundation, Inc.
Contributed by Andes Technology Corporation.
This file is part of GCC.
GCC is free software; you can redistribute it and/or modify it
under the terms of the GNU General Public License as published
by the Free Software Foundation; either version 3, or (at your
option) any later version.
GCC is distributed in the hope that it will be useful, but WITHOUT
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public
License for more details.
Under Section 7 of GPL version 3, you are granted additional
permissions described in the GCC Runtime Library Exception, version
3.1, as published by the Free Software Foundation.
You should have received a copy of the GNU General Public License and
a copy of the GCC Runtime Library Exception along with this program;
see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
. */
.section .mdebug.abi_nds32
.previous
/* ------------------------------------------- */
/* FPBIT floating point operations for libgcc */
/* ------------------------------------------- */
#ifdef L_addsub_sf
.text
.align 2
.global __subsf3
.type __subsf3, @function
__subsf3:
push $lp
pushm $r6, $r9
move $r2, #0x80000000
xor $r1, $r1, $r2
j .Lsfpadd
.global __addsf3
.type __addsf3, @function
__addsf3:
push $lp
pushm $r6, $r9
.Lsfpadd:
srli $r5, $r0, #23
andi $r5, $r5, #0xff
srli $r7, $r1, #23
andi $r7, $r7, #0xff
move $r3, #0x80000000
slli $r4, $r0, #8
or $r4, $r4, $r3
slli $r6, $r1, #8
or $r6, $r6, $r3
addi $r9, $r5, #-1
slti $r15, $r9, #0xfe
beqzs8 .LEspecA
.LElab1:
addi $r9, $r7, #-1
slti $r15, $r9, #0xfe
beqzs8 .LEspecB
.LElab2:
sub $r8, $r5, $r7
sltsi $r15, $r8, #0
bnezs8 .Li1
sltsi $r15, $r8, #0x20
bnezs8 .Li2
move $r6, #2
j .Le1
.Li2:
move $r2, $r6
srl $r6, $r6, $r8
sll $r9, $r6, $r8
beq $r9, $r2, .Le1
ori $r6, $r6, #2
j .Le1
.Li1:
move $r5, $r7
subri $r8, $r8, #0
sltsi $r15, $r8, #0x20
bnezs8 .Li4
move $r4, #2
j .Le1
.Li4:
move $r2, $r4
srl $r4, $r4, $r8
sll $r9, $r4, $r8
beq $r9, $r2, .Le1
ori $r4, $r4, #2
.Le1:
and $r8, $r0, $r3
xor $r9, $r8, $r1
sltsi $r15, $r9, #0
bnezs8 .LEsub1
#ADD($r4, $r6)
add $r4, $r4, $r6
slt $r15, $r4, $r6
beqzs8 .LEres
andi $r9, $r4, #1
beqz $r9, .Li7
ori $r4, $r4, #2
.Li7:
srli $r4, $r4, #1
addi $r5, $r5, #1
subri $r15, $r5, #0xff
bnezs8 .LEres
move $r4, #0
j .LEres
.LEsub1:
#SUB($r4, $r6)
move $r15, $r4
sub $r4, $r4, $r6
slt $r15, $r15, $r4
beqzs8 .Li9
subri $r4, $r4, #0
xor $r8, $r8, $r3
j .Le9
.Li9:
beqz $r4, .LEzer
.Le9:
#ifdef __NDS32_PERF_EXT__
clz $r2, $r4
#else
pushm $r0, $r1
pushm $r3, $r5
move $r0, $r4
bal __clzsi2
move $r2, $r0
popm $r3, $r5
popm $r0, $r1
#endif
sub $r5, $r5, $r2
sll $r4, $r4, $r2
.LEres:
blez $r5, .LEund
.LElab12:
#ADD($r4, $0x80)
move $r15, #0x80
add $r4, $r4, $r15
slt $r15, $r4, $r15
#ADDC($r5, $0x0)
add $r5, $r5, $r15
srli $r9, $r4, #8
andi $r9, $r9, #1
sub $r4, $r4, $r9
slli $r4, $r4, #1
srli $r4, $r4, #9
slli $r9, $r5, #23
or $r4, $r4, $r9
or $r0, $r4, $r8
.LE999:
popm $r6, $r9
pop $lp
ret5 $lp
.LEund:
subri $r2, $r5, #1
slti $r15, $r2, #0x20
beqzs8 .LEzer
move $r9, #0x80000000
or $r4, $r4, $r9
subri $r9, $r2, #0x20
sll $r5, $r4, $r9
srl $r4, $r4, $r2
beqz $r5, .Li10
ori $r4, $r4, #1
.Li10:
move $r5, #0
addi $r9, $r4, #0x80
sltsi $r15, $r9, #0
beqzs8 .LElab12
move $r5, #1
j .LElab12
.LEspecA:
bnez $r5, .Li12
add $r4, $r4, $r4
beqz $r4, .Li13
#ifdef __NDS32_PERF_EXT__
clz $r8, $r4
#else
pushm $r0, $r5
move $r0, $r4
bal __clzsi2
move $r8, $r0
popm $r0, $r5
#endif
sub $r5, $r5, $r8
sll $r4, $r4, $r8
j .LElab1
.Li13:
subri $r15, $r7, #0xff
beqzs8 .LEspecB
move $r9, #0x80000000
bne $r1, $r9, .LEretB
.Li12:
add $r9, $r4, $r4
bnez $r9, .LEnan
subri $r15, $r7, #0xff
bnezs8 .LEretA
xor $r9, $r0, $r1
sltsi $r15, $r9, #0
bnezs8 .LEnan
j .LEretB
.LEspecB:
bnez $r7, .Li15
add $r6, $r6, $r6
beqz $r6, .LEretA
#ifdef __NDS32_PERF_EXT__
clz $r8, $r6
#else
pushm $r0, $r5
move $r0, $r6
bal __clzsi2
move $r8, $r0
popm $r0, $r5
#endif
sub $r7, $r7, $r8
sll $r6, $r6, $r8
j .LElab2
.Li15:
add $r9, $r6, $r6
bnez $r9, .LEnan
.LEretB:
move $r0, $r1
j .LE999
.LEretA:
j .LE999
.LEzer:
move $r0, #0
j .LE999
.LEnan:
move $r0, #0xffc00000
j .LE999
.size __subsf3, .-__subsf3
.size __addsf3, .-__addsf3
#endif /* L_addsub_sf */
#ifdef L_sf_to_si
.text
.align 2
.global __fixsfsi
.type __fixsfsi, @function
__fixsfsi:
push $lp
slli $r1, $r0, #8
move $r3, #0x80000000
or $r1, $r1, $r3
srli $r3, $r0, #23
andi $r3, $r3, #0xff
subri $r2, $r3, #0x9e
blez $r2, .LJspec
sltsi $r15, $r2, #0x20
bnezs8 .Li42
move $r0, #0
j .LJ999
.Li42:
srl $r1, $r1, $r2
sltsi $r15, $r0, #0
beqzs8 .Li43
subri $r1, $r1, #0
.Li43:
move $r0, $r1
.LJ999:
pop $lp
ret5 $lp
.LJspec:
move $r3, #0x7f800000
slt $r15, $r3, $r0
beqzs8 .Li44
move $r0, #0x80000000
j .LJ999
.Li44:
move $r0, #0x7fffffff
j .LJ999
.size __fixsfsi, .-__fixsfsi
#endif /* L_sf_to_si */
#ifdef L_divsi3
.text
.align 2
.globl __divsi3
.type __divsi3, @function
__divsi3:
! ---------------------------------------------------------------------
! neg = 0;
! if (a < 0)
! { a = -a;
! neg = !neg;
! }
! ---------------------------------------------------------------------
sltsi $r5, $r0, 0 ! $r5 <- neg = (a < 0) ? 1 : 0
subri $r4, $r0, 0 ! $r4 <- a = -a
cmovn $r0, $r4, $r5 ! $r0 <- a = neg ? -a : a
.L2:
! ---------------------------------------------------------------------
! if (b < 0)
! ---------------------------------------------------------------------
bgez $r1, .L3 ! if b >= 0, skip
! ---------------------------------------------------------------------
! { b=-b;
! neg=!neg;
! }
! ---------------------------------------------------------------------
subri $r1, $r1, 0 ! $r1 <- b = -b
subri $r5, $r5, 1 ! $r5 <- neg = !neg
.L3:
! ---------------------------------------------------------------------
!!res = udivmodsi4 (a, b, 1);
! res = 0;
! if (den != 0)
! ---------------------------------------------------------------------
movi $r2, 0 ! $r2 <- res = 0
beqz $r1, .L1 ! if den == 0, skip
! ---------------------------------------------------------------------
! bit = 1;
! ---------------------------------------------------------------------
movi $r4, 1 ! $r4 <- bit = 1
#ifndef __OPTIMIZE_SIZE__
.L6:
#endif
! ---------------------------------------------------------------------
! while (den < num && bit && !(den & (1L << 31)))
! ---------------------------------------------------------------------
slt $ta, $r1, $r0 ! $ta <- den < num ?
beqz $ta, .L5 ! if no, skip
! ---------------------------------------------------------------------
! { den << = 1;
! bit << = 1;
! }
! ---------------------------------------------------------------------
#if defined (__OPTIMIZE_SIZE__) && !defined (__NDS32_ISA_V3M__)
clz $r3, $r1 ! $r3 <- leading zero count for den
clz $ta, $r0 ! $ta <- leading zero count for num
sub $r3, $r3, $ta ! $r3 <- number of bits to shift
sll $r1, $r1, $r3 ! $r1 <- den
sll $r4, $r4, $r3 ! $r2 <- bit
#else
slli $r1, $r1, 1 ! $r1 <- den << = 1
slli $r4, $r4, 1 ! $r4 <- bit << = 1
b .L6 ! continue loop
#endif
.L5:
! ---------------------------------------------------------------------
! while (bit)
! { if (num >= den)
! ---------------------------------------------------------------------
slt $ta, $r0, $r1 ! $ta <- num < den ?
bnez $ta, .L9 ! if yes, skip
! ---------------------------------------------------------------------
! { num -= den;
! res |= bit;
! }
! ---------------------------------------------------------------------
sub $r0, $r0, $r1 ! $r0 <- num -= den
or $r2, $r2, $r4 ! $r2 <- res |= bit
.L9:
! ---------------------------------------------------------------------
! bit >> = 1;
! den >> = 1;
! }
!!if (modwanted)
!! return num;
!!return res;
! ---------------------------------------------------------------------
srli $r4, $r4, 1 ! $r4 <- bit >> = 1
srli $r1, $r1, 1 ! $r1 <- den >> = 1
bnez $r4, .L5 ! if bit != 0, continue loop
.L1:
! ---------------------------------------------------------------------
! if (neg)
! res = -res;
! return res;
! ---------------------------------------------------------------------
subri $r0, $r2, 0 ! $r0 <- -res
cmovz $r0, $r2, $r5 ! $r0 <- neg ? -res : res
! ---------------------------------------------------------------------
ret
.size __divsi3, .-__divsi3
#endif /* L_divsi3 */
#ifdef L_divdi3
!--------------------------------------
#ifdef __big_endian__
#define V1H $r0
#define V1L $r1
#define V2H $r2
#define V2L $r3
#else
#define V1H $r1
#define V1L $r0
#define V2H $r3
#define V2L $r2
#endif
!--------------------------------------
.text
.align 2
.globl __divdi3
.type __divdi3, @function
__divdi3:
! prologue
#ifdef __NDS32_ISA_V3M__
push25 $r10, 0
#else
smw.adm $r6, [$sp], $r10, 2
#endif
! end of prologue
move $r8, V1L
move $r9, V1H
move $r6, V2L
move $r7, V2H
movi $r10, 0
bgez V1H, .L80
bal __negdi2
move $r8, V1L
move $r9, V1H
movi $r10, -1
.L80:
bgez $r7, .L81
move V1L, $r6
move V1H, $r7
bal __negdi2
move $r6, V1L
move $r7, V1H
nor $r10, $r10, $r10
.L81:
move V2L, $r6
move V2H, $r7
move V1L, $r8
move V1H, $r9
movi $r4, 0
bal __udivmoddi4
beqz $r10, .L82
bal __negdi2
.L82:
! epilogue
#ifdef __NDS32_ISA_V3M__
pop25 $r10, 0
#else
lmw.bim $r6, [$sp], $r10, 2
ret
#endif
.size __divdi3, .-__divdi3
#endif /* L_divdi3 */
#ifdef L_modsi3
.text
.align 2
.globl __modsi3
.type __modsi3, @function
__modsi3:
! ---------------------------------------------------------------------
! neg=0;
! if (a<0)
! { a=-a;
! neg=1;
! }
! ---------------------------------------------------------------------
sltsi $r5, $r0, 0 ! $r5 <- neg < 0 ? 1 : 0
subri $r4, $r0, 0 ! $r4 <- -a
cmovn $r0, $r4, $r5 ! $r0 <- |a|
! ---------------------------------------------------------------------
! if (b < 0)
#ifndef __NDS32_PERF_EXT__
! ---------------------------------------------------------------------
bgez $r1, .L3 ! if b >= 0, skip
! ---------------------------------------------------------------------
! b = -b;
! ---------------------------------------------------------------------
subri $r1, $r1, 0 ! $r1 <- |b|
.L3:
! ---------------------------------------------------------------------
!!res = udivmodsi4 (a, b, 1);
! if (den != 0)
! ---------------------------------------------------------------------
#else /* __NDS32_PERF_EXT__ */
! b = -b;
!!res = udivmodsi4 (a, b, 1);
! if (den != 0)
! ---------------------------------------------------------------------
abs $r1, $r1 ! $r1 <- |b|
#endif /* __NDS32_PERF_EXT__ */
beqz $r1, .L1 ! if den == 0, skip
! ---------------------------------------------------------------------
! { bit = 1;
! res = 0;
! ---------------------------------------------------------------------
movi $r4, 1 ! $r4 <- bit = 1
#ifndef __OPTIMIZE_SIZE__
.L6:
#endif
! ---------------------------------------------------------------------
! while (den < num&&bit && !(den & (1L << 31)))
! ---------------------------------------------------------------------
slt $ta, $r1, $r0 ! $ta <- den < num ?
beqz $ta, .L5 ! if no, skip
! ---------------------------------------------------------------------
! { den << = 1;
! bit << = 1;
! }
! ---------------------------------------------------------------------
#if defined (__OPTIMIZE_SIZE__) && ! defined (__NDS32_ISA_V3M__)
clz $r3, $r1 ! $r3 <- leading zero count for den
clz $ta, $r0 ! $ta <- leading zero count for num
sub $r3, $r3, $ta ! $r3 <- number of bits to shift
sll $r1, $r1, $r3 ! $r1 <- den
sll $r4, $r4, $r3 ! $r2 <- bit
#else
slli $r1, $r1, 1 ! $r1 <- den << = 1
slli $r4, $r4, 1 ! $r4 <- bit << = 1
b .L6 ! continue loop
#endif
.L5:
! ---------------------------------------------------------------------
! while (bit)
! { if (num >= den)
! { num -= den;
! res |= bit;
! }
! bit >> = 1;
! den >> = 1;
! }
! }
!!if (modwanted)
!! return num;
!!return res;
! ---------------------------------------------------------------------
sub $r2, $r0, $r1 ! $r2 <- num - den
slt $ta, $r0, $r1 ! $ta <- num < den ?
srli $r4, $r4, 1 ! $r4 <- bit >> = 1
cmovz $r0, $r2, $ta ! $r0 <- num = (num < den) ? num : num - den
srli $r1, $r1, 1 ! $r1 <- den >> = 1
bnez $r4, .L5 ! if bit != 0, continue loop
.L1:
! ---------------------------------------------------------------------
! if (neg)
! res = -res;
! return res;
! ---------------------------------------------------------------------
subri $r3, $r0, 0 ! $r3 <- -res
cmovn $r0, $r3, $r5 ! $r0 <- neg ? -res : res
! ---------------------------------------------------------------------
ret
.size __modsi3, .-__modsi3
#endif /* L_modsi3 */
#ifdef L_moddi3
!--------------------------------------
#ifdef __big_endian__
#define V1H $r0
#define V1L $r1
#define V2H $r2
#define V2L $r3
#else
#define V1H $r1
#define V1L $r0
#define V2H $r3
#define V2L $r2
#endif
!--------------------------------------
.text
.align 2
.globl __moddi3
.type __moddi3, @function
__moddi3:
! =====================================================================
! stack allocation:
! sp+32 +-----------------------+
! | $lp |
! sp+28 +-----------------------+
! | $r6 - $r10 |
! sp+8 +-----------------------+
! | |
! sp+4 +-----------------------+
! | |
! sp +-----------------------+
! =====================================================================
! prologue
#ifdef __NDS32_ISA_V3M__
push25 $r10, 8
#else
smw.adm $r6, [$sp], $r10, 2
addi $sp, $sp, -8
#endif
! end of prologue
!------------------------------------------
! __moddi3 (DWtype u, DWtype v)
! {
! word_type c = 0;
! DWunion uu = {.ll = u};
! DWunion vv = {.ll = v};
! DWtype w;
! if (uu.s.high < 0)
! c = ~c,
! uu.ll = -uu.ll;
!---------------------------------------------
move $r8, V1L
move $r9, V1H
move $r6, V2L
move $r7, V2H
movi $r10, 0 ! r10 = c = 0
bgez V1H, .L80 ! if u > 0 , go L80
bal __negdi2
move $r8, V1L
move $r9, V1H
movi $r10, -1 ! r10 = c = ~c
!------------------------------------------------
! if (vv.s.high < 0)
! vv.ll = -vv.ll;
!----------------------------------------------
.L80:
bgez $r7, .L81 ! if v > 0 , go L81
move V1L, $r6
move V1H, $r7
bal __negdi2
move $r6, V1L
move $r7, V1H
!------------------------------------------
! (void) __udivmoddi4 (uu.ll, vv.ll, &w);
! if (c)
! w = -w;
! return w;
!-----------------------------------------
.L81:
move V2L, $r6
move V2H, $r7
move V1L, $r8
move V1H, $r9
addi $r4, $sp, 0
bal __udivmoddi4
lwi $r0, [$sp+(0)] ! le: sp + 0 is low, be: sp + 0 is high
lwi $r1, [$sp+(4)] ! le: sp + 4 is low, be: sp + 4 is high
beqz $r10, .L82
bal __negdi2
.L82:
! epilogue
#ifdef __NDS32_ISA_V3M__
pop25 $r10, 8
#else
addi $sp, $sp, 8
lmw.bim $r6, [$sp], $r10, 2
ret
#endif
.size __moddi3, .-__moddi3
#endif /* L_moddi3 */
#ifdef L_mulsi3
.text
.align 2
.globl __mulsi3
.type __mulsi3, @function
__mulsi3:
! ---------------------------------------------------------------------
! r = 0;
! while (a)
! $r0: r
! $r1: b
! $r2: a
! ---------------------------------------------------------------------
beqz $r0, .L7 ! if a == 0, done
move $r2, $r0 ! $r2 <- a
movi $r0, 0 ! $r0 <- r <- 0
.L8:
! ---------------------------------------------------------------------
! { if (a & 1)
! r += b;
! a >> = 1;
! b << = 1;
! }
! $r0: r
! $r1: b
! $r2: a
! $r3: scratch
! $r4: scratch
! ---------------------------------------------------------------------
andi $r3, $r2, 1 ! $r3 <- a & 1
add $r4, $r0, $r1 ! $r4 <- r += b
cmovn $r0, $r4, $r3 ! $r0 <- r
srli $r2, $r2, 1 ! $r2 <- a >> = 1
slli $r1, $r1, 1 ! $r1 <- b << = 1
bnez $r2, .L8 ! if a != 0, continue loop
.L7:
! ---------------------------------------------------------------------
! $r0: return code
! ---------------------------------------------------------------------
ret
.size __mulsi3, .-__mulsi3
#endif /* L_mulsi3 */
#ifdef L_udivsi3
.text
.align 2
.globl __udivsi3
.type __udivsi3, @function
__udivsi3:
! ---------------------------------------------------------------------
!!res=udivmodsi4(a,b,0);
! res=0;
! if (den!=0)
! ---------------------------------------------------------------------
movi $r2, 0 ! $r2 <- res=0
beqz $r1, .L1 ! if den==0, skip
! ---------------------------------------------------------------------
! { bit=1;
! ---------------------------------------------------------------------
movi $r4, 1 ! $r4 <- bit=1
#ifndef __OPTIMIZE_SIZE__
.L6:
#endif
! ---------------------------------------------------------------------
! while (den=den)
! ---------------------------------------------------------------------
slt $ta, $r0, $r1 ! $ta <- num>=1;
! den>>=1;
! }
! }
!!if (modwanted)
!! return num;
!!return res;
! ---------------------------------------------------------------------
srli $r4, $r4, 1 ! $r4 <- bit>>=1
srli $r1, $r1, 1 ! $r1 <- den>>=1
bnez $r4, .L5 ! if bit!=0, continue loop
.L1:
! ---------------------------------------------------------------------
! return res;
! ---------------------------------------------------------------------
move $r0, $r2 ! $r0 <- return value
! ---------------------------------------------------------------------
! ---------------------------------------------------------------------
ret
.size __udivsi3, .-__udivsi3
#endif /* L_udivsi3 */
#ifdef L_udivdi3
!--------------------------------------
#ifdef __big_endian__
#define V1H $r0
#define V1L $r1
#define V2H $r2
#define V2L $r3
#else
#define V1H $r1
#define V1L $r0
#define V2H $r3
#define V2L $r2
#endif
!--------------------------------------
.text
.align 2
.globl __udivdi3
.type __udivdi3, @function
__udivdi3:
! prologue
#ifdef __NDS32_ISA_V3M__
push25 $r8, 0
#else
smw.adm $r6, [$sp], $r8, 2
#endif
! end of prologue
movi $r4, 0
bal __udivmoddi4
! epilogue
#ifdef __NDS32_ISA_V3M__
pop25 $r8, 0
#else
lmw.bim $r6, [$sp], $r8, 2
ret
#endif
.size __udivdi3, .-__udivdi3
#endif /* L_udivdi3 */
#ifdef L_udivmoddi4
.text
.align 2
.globl fudiv_qrnnd
.type fudiv_qrnnd, @function
#ifdef __big_endian__
#define P1H $r0
#define P1L $r1
#define P2H $r2
#define P2L $r3
#define W6H $r4
#define W6L $r5
#define OFFSET_L 4
#define OFFSET_H 0
#else
#define P1H $r1
#define P1L $r0
#define P2H $r3
#define P2L $r2
#define W6H $r5
#define W6L $r4
#define OFFSET_L 0
#define OFFSET_H 4
#endif
fudiv_qrnnd:
!------------------------------------------------------
! function: fudiv_qrnnd(quotient, remainder, high_numerator, low_numerator, denominator)
! divides a UDWtype, composed by the UWtype integers,HIGH_NUMERATOR (from $r4)
! and LOW_NUMERATOR(from $r5) by DENOMINATOR(from $r6), and places the quotient
! in $r7 and the remainder in $r8.
!------------------------------------------------------
! in reg:$r4(n1), $r5(n0), $r6(d0)
! __d1 = ((USItype) (d) >> ((4 * 8) / 2));
! __d0 = ((USItype) (d) & (((USItype) 1 << ((4 * 8) / 2)) - 1));
! __r1 = (n1) % __d1;
! __q1 = (n1) / __d1;
! __m = (USItype) __q1 * __d0;
! __r1 = __r1 * ((USItype) 1 << ((4 * 8) / 2)) | ((USItype) (n0) >> ((4 * 8) / 2));
! if (__r1 < __m)
! {
!------------------------------------------------------
smw.adm $r0, [$sp], $r4, 2 ! store $lp, when use BASELINE_V1,and must store $r0-$r3
srli $r7, $r6, 16 ! $r7 = d1 =__ll_highpart (d)
movi $ta, 65535
and $r8, $r6, $ta ! $r8 = d0 = __ll_lowpart (d)
divr $r9, $r10, $r4, $r7 ! $r9 = q1, $r10 = r1
and $r4, $r5, $ta ! $r4 = __ll_lowpart (n0)
slli $r10, $r10, 16 ! $r10 = r1 << 16
srli $ta, $r5, 16 ! $ta = __ll_highpart (n0)
or $r10, $r10, $ta ! $r10 <- $r0|$r3=__r1
mul $r5, $r9, $r8 ! $r5 = m = __q1*__d0
slt $ta, $r10, $r5 ! $ta <- __r1<__m
beqz $ta, .L2 !if yes,skip
!------------------------------------------------------
! __q1--, __r1 += (d);
! if (__r1 >= (d))
! {
!------------------------------------------------------
add $r10, $r10, $r6 !$r10 <- __r1+d=__r1
addi $r9, $r9, -1 !$r9 <- __q1--=__q1
slt $ta, $r10, $r6 !$ta <- __r1= (d))
! {
!------------------------------------------------------
add $r10, $r10, $r6 !$r10 <- __r0+d=__r0
addi $r7, $r7, -1 !$r7 <- __q0--=__q0
slt $ta, $r10, $r6 !$ta <- __r0 n1)
! {
!------------------------------------------------------
slt $ta, P1H, P2L !$ta <- n1> ((4 * 8) - bm));
! n0 = n0 << bm;
! }
!------------------------------------------------------
subri $r5, $r7, 32 !$r5 <- 32-bm
srl $r5, P1L, $r5 !$r5 <- n0>>$r5
sll $r6, P1H, $r7 !$r6 <- n1< n1)
!------------------------------------------------------
move $r4,P1H ! give fudiv_qrnnd args
move $r5,P1L !
move $r6,P2L !
bal fudiv_qrnnd !calcaulte q0 n0
movi $r6, 0 !P1L <- 0
swi $r7,[$sp+32] !q0
swi $r6,[$sp+36] !q1
move P1L,$r8 !n0
b .L19
.L10:
!------------------------------------------------------
! else #if (d0 > n1)
! {
! if(d0 == 0)
!------------------------------------------------------
bnez P2L, .L20 !if yes,skip
!------------------------------------------------------
! d0 = 1 / d0;
!------------------------------------------------------
movi $r4, 1 !P1L <- 1
divr P2L, $r4, $r4, P2L !$r9=1/d0,P1L=1%d0
.L20:
#ifndef __NDS32_PERF_EXT__
smw.adm $r0, [$sp], $r5, 0
move $r0, P2L
bal __clzsi2
move $r7, $r0
lmw.bim $r0, [$sp], $r5, 0
#else
clz $r7, P2L
#endif
swi $r7,[$sp+(28)] ! store bm
beqz $r7, .L28 ! if yes,skip
!------------------------------------------------------
! b = (4 * 8) - bm;
! d0 = d0 << bm;
! n2 = n1 >> b;
! n1 = (n1 << bm) | (n0 >> b);
! n0 = n0 << bm;
! fudiv_qrnnd (&q1, &n1, n2, n1, d0);
! }
!------------------------------------------------------
subri $r10, $r7, 32 !$r10 <- 32-bm=b
srl $r4, P1L, $r10 !$r4 <- n0>>b
sll $r5, P1H, $r7 !$r5 <- n1<>b=n2 !for fun
move $r6,P2L !for fun
bal fudiv_qrnnd !caculate q1, n1
swi $r7,[$sp+(36)] ! q1 store
move P1H,$r8 ! n1 store
move $r4,$r8 ! prepare for next fudiv_qrnnd()
move $r5,P1L
move $r6,P2L
b .L29
.L28:
!------------------------------------------------------
! else // bm != 0
! {
! n1 -= d0;
! q1 = 1;
!
!------------------------------------------------------
sub P1H, P1H, P2L !P1L <- n1-d0=n1
movi $ta, 1 !
swi $ta, [$sp+(36)] !1 -> [$sp+(36)]
move $r4,P1H ! give fudiv_qrnnd args
move $r5,P1L
move $r6,P2L
.L29:
!------------------------------------------------------
! fudiv_qrnnd (&q0, &n0, n1, n0, d0);
!------------------------------------------------------
bal fudiv_qrnnd !calcuate q0, n0
swi $r7,[$sp+(32)] !q0 store
move P1L,$r8 !n0
.L19:
!------------------------------------------------------
! if (rp != 0)
! {
!------------------------------------------------------
beqz $fp, .L31 !if yes,skip
!------------------------------------------------------
! rr.s.low = n0 >> bm;
! rr.s.high = 0;
! *rp = rr.ll;
! }
!------------------------------------------------------
movi $r5, 0 !$r5 <- 0
lwi $r7,[$sp+(28)] !load bm
srl $r4, P1L, $r7 !$r4 <- n0>>bm
swi $r4, [$fp+OFFSET_L] !r0 !$r4 -> [$sp+(48)]
swi $r5, [$fp+OFFSET_H] !r1 !0 -> [$sp+(52)]
b .L31
.L9:
!------------------------------------------------------
! else # d1 == 0
! {
! if(d1 > n1)
! {
!------------------------------------------------------
slt $ta, P1H, P2H !$ta <- n1 [$sp+(40)]=q1
swi $r5, [$sp+(36)] !q1 !0 -> [$sp+(32)]=q0
beqz $fp, .L31 !if yes,skip
!------------------------------------------------------
! rr.s.low = n0;
! rr.s.high = n1;
! *rp = rr.ll;
! }
!------------------------------------------------------
swi P1L, [$fp+OFFSET_L] !P1L -> [rp]
swi P1H, [$fp+OFFSET_H] !P1H -> [rp+4]
b .L31
.L32:
#ifndef __NDS32_PERF_EXT__
smw.adm $r0, [$sp], $r5, 0
move $r0, P2H
bal __clzsi2
move $r7, $r0
lmw.bim $r0, [$sp], $r5, 0
#else
clz $r7,P2H
#endif
swi $r7,[$sp+(28)] !$r7=bm store
beqz $r7, .L42 !if yes,skip
!------------------------------------------------------
! USItype m1, m0;
! b = (4 * 8) - bm;
! d1 = (d0 >> b) | (d1 << bm);
! d0 = d0 << bm;
! n2 = n1 >> b;
! n1 = (n0 >> b) | (n1 << bm);
! n0 = n0 << bm;
! fudiv_qrnnd (&q0, &n1, n2, n1, d1);
!------------------------------------------------------
subri $r10, $r7, 32 !$r10 <- 32-bm=b
srl $r5, P2L, $r10 !$r5 <- d0>>b
sll $r6, P2H, $r7 !$r6 <- d1<>b=n2 !!! func
srl $r8, P1L, $r10 !$r8 <- n0>>b !!$r8
sll $r9, P1H, $r7 !$r9 <- n1<> ((4 * 8) / 2));
! __vl = ((USItype) (d0) & (((USItype) 1 << ((4 * 8) / 2)) - 1));
! __vh = ((USItype) (d0) >> ((4 * 8) / 2));
! __x0 = (USItype) __ul * __vl;
! __x1 = (USItype) __ul * __vh;
! __x2 = (USItype) __uh * __vl;
! __x3 = (USItype) __uh * __vh;
! __x1 += ((USItype) (__x0) >> ((4 * 8) / 2));
! __x1 += __x2;
! if (__x1 < __x2)
! __x3 += ((USItype) 1 << ((4 * 8) / 2));
! (m1) = __x3 + ((USItype) (__x1) >> ((4 * 8) / 2));
! (m0) = (USItype)(q0*d0);
! }
! if (m1 > n1)
!---------------------------------------------------
#ifdef __NDS32_ISA_V3M__
!mulr64 $r4, P2L, $r6
smw.adm $r0, [$sp], $r3, 0
move P1L, P2L
move P2L, $r6
movi P1H, 0
movi P2H, 0
bal __muldi3
movd44 $r4, $r0
lmw.bim $r0, [$sp], $r3, 0
move $r8, W6H
move $r5, W6L
#else
mulr64 $r4, P2L, $r6
move $r8, W6H
move $r5, W6L
#endif
slt $ta, P1H, $r8 !$ta <- n1 n0)
!------------------------------------------------------
slt $ta, P1L, $r5 !$ta <- n0 (m0));
! (m0) = __x;
! }
! }
!------------------------------------------------------
sub $r4, $r5, P2L !$r4 <- m0-d0=__x
addi $r6, $r6, -1 !$r6 <- q0--=q0
sub $r8, $r8, P2H !$r8 <- m1-d1
swi $r6, [$sp+(32)] ! q0 !$r6->[$sp+(32)]
slt $ta, $r5, $r4 !$ta <- m0<__x
sub $r8, $r8, $ta !$r8 <- P1H-P1L=m1
move $r5, $r4 !$r5 <- __x=m0
.L45:
!------------------------------------------------------
! q1 = 0;
! if (rp != 0)
! {
!------------------------------------------------------
movi $r4, 0 !$r4 <- 0
swi $r4, [$sp+(36)] !0 -> [$sp+(40)]=q1
beqz $fp, .L31 !if yes,skip
!------------------------------------------------------
! # sub_ddmmss (n1, n0, n1, n0, m1, m0);
! do
! { USItype __x;
! __x = (n0) - (m0);
! (n1) = (n1) - (m1) - (__x > (n0));
! (n0) = __x;
! }
! rr.s.low = (n1 << b) | (n0 >> bm);
! rr.s.high = n1 >> bm;
! *rp = rr.ll;
!------------------------------------------------------
sub $r4, P1H, $r8 !$r4 <- n1-m1
sub $r6, P1L, $r5 !$r6 <- n0-m0=__x=n0
slt $ta, P1L, $r6 !$ta <- n0<__x
sub P1H, $r4, $ta !P1H <- $r4-$ta=n1
move P1L, $r6
lwi $r7,[$sp+(28)] ! load bm
subri $r10,$r7,32
sll $r4, P1H, $r10 !$r4 <- n1<>bm
or $r6, $r5, $r4 !$r6 <- $r5|$r4=rr.s.low
srl $r8, P1H, $r7 !$r8 <- n1>>bm =rr.s.high
swi $r6, [$fp+OFFSET_L] !
swi $r8, [$fp+OFFSET_H] !
b .L31
.L42:
!------------------------------------------------------
! else
! {
! if(n1 > d1)
!------------------------------------------------------
slt $ta, P2H, P1H !$ta <- P2H= d0)
!------------------------------------------------------
slt $ta, P1L, P2L !$ta <- P1L (n0));
! (n0) = __x;
! }
!------------------------------------------------------
.L52:
sub $r4, P1H, P2H !$r4 <- P1H-P2H
sub $r6, P1L, P2L !$r6 <- no-d0=__x=n0
slt $ta, P1L, $r6 !$ta <- no<__x
sub P1H, $r4, $ta !P1H <- $r4-$ta=n1
move P1L, $r6 !n0
movi $r5, 1 !
swi $r5, [$sp+(32)] !1 -> [$sp+(32)]=q0
b .L54
.L51:
!------------------------------------------------------
! q0 = 0;
!------------------------------------------------------
movi $r5,0
swi $r5, [$sp+(32)] !$r5=0 -> [$sp+(32)]
.L54:
!------------------------------------------------------
! q1 = 0;
! if (rp != 0)
! {
!------------------------------------------------------
movi $r5, 0 !
swi $r5, [$sp+(36)] !0 -> [$sp+(36)]
beqz $fp, .L31
!------------------------------------------------------
! rr.s.low = n0;
! rr.s.high = n1;
! *rp = rr.ll;
! }
!------------------------------------------------------
swi P1L, [$fp+OFFSET_L] !remainder
swi P1H, [$fp+OFFSET_H] !
.L31:
!------------------------------------------------------
! const DWunion ww = {{.low = q0, .high = q1}};
! return ww.ll;
!}
!------------------------------------------------------
lwi P1L, [$sp+(32)] !quotient
lwi P1H, [$sp+(36)]
lmw.bim $r6, [$sp], $r10, 10
addi $sp, $sp, 12
ret
.size __udivmoddi4, .-__udivmoddi4
#endif /* L_udivmoddi4 */
#ifdef L_umodsi3
! =====================================================================
.text
.align 2
.globl __umodsi3
.type __umodsi3, @function
__umodsi3:
! ---------------------------------------------------------------------
!!res=udivmodsi4(a,b,1);
! if (den==0)
! return num;
! ---------------------------------------------------------------------
beqz $r1, .L1 ! if den==0, skip
! ---------------------------------------------------------------------
! bit=1;
! res=0;
! ---------------------------------------------------------------------
movi $r4, 1 ! $r4 <- bit=1
#ifndef __OPTIMIZE_SIZE__
.L6:
#endif
! ---------------------------------------------------------------------
! while (den=den)
! { num-=den;
! res|=bit;
! }
! bit>>=1;
! den>>=1;
! }
!!if (modwanted)
!! return num;
!!return res;
! ---------------------------------------------------------------------
sub $r2, $r0, $r1 ! $r2 <- num-den
slt $ta, $r0, $r1 ! $ta <- num>=1
cmovz $r0, $r2, $ta ! $r0 <- num=(num>=1
bnez $r4, .L5 ! if bit!=0, continue loop
.L1:
! ---------------------------------------------------------------------
! return res;
! ---------------------------------------------------------------------
ret
.size __umodsi3, .-__umodsi3
#endif /* L_umodsi3 */
#ifdef L_umoddi3
!--------------------------------------
#ifdef __big_endian__
#define V1H $r0
#define V1L $r1
#define V2H $r2
#define V2L $r3
#else
#define V1H $r1
#define V1L $r0
#define V2H $r3
#define V2L $r2
#endif
!--------------------------------------
.text
.align 2
.globl __umoddi3
.type __umoddi3, @function
__umoddi3:
! prologue
addi $sp, $sp, -12
swi $lp, [$sp+(0)]
! end of prologue
addi $r4, $sp, 4
bal __udivmoddi4
lwi $r0, [$sp+(4)] ! __udivmoddi4 return low when LE mode or return high when BE mode
lwi $r1, [$sp+(8)] !
.L82:
! epilogue
lwi $lp, [$sp+(0)]
addi $sp, $sp, 12
ret
.size __umoddi3, .-__umoddi3
#endif /* L_umoddi3 */
#ifdef L_muldi3
#ifdef __big_endian__
#define P1H $r0
#define P1L $r1
#define P2H $r2
#define P2L $r3
#define V2H $r4
#define V2L $r5
#else
#define P1H $r1
#define P1L $r0
#define P2H $r3
#define P2L $r2
#define V2H $r5
#define V2L $r4
#endif
! ====================================================================
.text
.align 2
.globl __muldi3
.type __muldi3, @function
__muldi3:
! parameter passing for libgcc functions normally involves 2 doubles
!---------------------------------------
#ifdef __NDS32_ISA_V3M__
! There is no mulr64 instruction in Andes ISA V3M.
! So we must provide a sequence of calculations to complete the job.
smw.adm $r6, [$sp], $r9, 0x0
zeh33 $r4, P1L
srli $r7, P1L, 16
zeh33 $r5, P2L
mul $r6, $r5, $r4
mul33 $r5, $r7
srli $r8, P2L, 16
mov55 $r9, $r5
maddr32 $r9, $r8, $r4
srli $r4, $r6, 16
add $r4, $r9, $r4
slt45 $r4, $r5
slli $r5, $r15, 16
maddr32 $r5, $r8, $r7
mul P2L, P1H, P2L
srli $r7, $r4, 16
maddr32 P2L, P2H, P1L
add333 P1H, $r5, $r7
slli $r4, $r4, 16
zeh33 $r6, $r6
add333 P1L, $r4, $r6
add333 P1H, P2L, P1H
lmw.bim $r6, [$sp], $r9, 0x0
ret
#else /* not __NDS32_ISA_V3M__ */
mul $ta, P1L, P2H
mulr64 $r4, P1L, P2L
maddr32 $ta, P1H, P2L
move P1L, V2L
add P1H, $ta, V2H
ret
#endif /* not __NDS32_ISA_V3M__ */
.size __muldi3, .-__muldi3
#endif /* L_muldi3 */
#ifdef L_addsub_df
#ifndef __big_endian__
#define P1L $r0
#define P1H $r1
#define P2L $r2
#define P2H $r3
#define P3L $r4
#define P3H $r5
#define O1L $r7
#define O1H $r8
#else
#define P1H $r0
#define P1L $r1
#define P2H $r2
#define P2L $r3
#define P3H $r4
#define P3L $r5
#define O1H $r7
#define O1L $r8
#endif
.text
.align 2
.global __subdf3
.type __subdf3, @function
__subdf3:
push $lp
pushm $r6, $r10
move $r4, #0x80000000
xor P2H, P2H, $r4
j .Lsdpadd
.global __adddf3
.type __adddf3, @function
__adddf3:
push $lp
pushm $r6, $r10
.Lsdpadd:
slli $r6, P1H, #1
srli $r6, $r6, #21
slli P3H, P1H, #11
srli $r10, P1L, #21
or P3H, P3H, $r10
slli P3L, P1L, #11
move O1L, #0x80000000
or P3H, P3H, O1L
slli $r9, P2H, #1
srli $r9, $r9, #21
slli O1H, P2H, #11
srli $r10, P2L, #21
or O1H, O1H, $r10
or O1H, O1H, O1L
slli O1L, P2L, #11
addi $r10, $r6, #-1
slti $r15, $r10, #0x7fe
beqzs8 .LEspecA
.LElab1:
addi $r10, $r9, #-1
slti $r15, $r10, #0x7fe
beqzs8 .LEspecB
.LElab2:
#NORMd($r4, P2L, P1L)
bnez P3H, .LL1
bnez P3L, .LL2
move $r6, #0
j .LL3
.LL2:
move P3H, P3L
move P3L, #0
move P2L, #32
sub $r6, $r6, P2L
.LL1:
#ifndef __big_endian__
#ifdef __NDS32_PERF_EXT__
clz $r2, $r5
#else
pushm $r0, $r1
pushm $r3, $r5
move $r0, $r5
bal __clzsi2
move $r2, $r0
popm $r3, $r5
popm $r0, $r1
#endif
#else /* __big_endian__ */
#ifdef __NDS32_PERF_EXT__
clz $r3, $r4
#else
pushm $r0, $r2
pushm $r4, $r5
move $r0, $r4
bal __clzsi2
move $r3, $r0
popm $r4, $r5
popm $r0, $r2
#endif
#endif /* __big_endian__ */
beqz P2L, .LL3
sub $r6, $r6, P2L
subri P1L, P2L, #32
srl P1L, P3L, P1L
sll P3L, P3L, P2L
sll P3H, P3H, P2L
or P3H, P3H, P1L
.LL3:
#NORMd End
#NORMd($r7, P2L, P1L)
bnez O1H, .LL4
bnez O1L, .LL5
move $r9, #0
j .LL6
.LL5:
move O1H, O1L
move O1L, #0
move P2L, #32
sub $r9, $r9, P2L
.LL4:
#ifndef __big_endian__
#ifdef __NDS32_PERF_EXT__
clz $r2, O1H
#else
pushm $r0, $r1
pushm $r3, $r5
move $r0, O1H
bal __clzsi2
move $r2, $r0
popm $r3, $r5
popm $r0, $r1
#endif
#else /* __big_endian__ */
#ifdef __NDS32_PERF_EXT__
clz $r3, O1H
#else
pushm $r0, $r2
pushm $r4, $r5
move $r0, O1H
bal __clzsi2
move $r3, $r0
popm $r4, $r5
popm $r0, $r2
#endif
#endif /* __big_endian__ */
beqz P2L, .LL6
sub $r9, $r9, P2L
subri P1L, P2L, #32
srl P1L, O1L, P1L
sll O1L, O1L, P2L
sll O1H, O1H, P2L
or O1H, O1H, P1L
.LL6:
#NORMd End
move $r10, #0x80000000
and P1H, P1H, $r10
beq $r6, $r9, .LEadd3
slts $r15, $r9, $r6
beqzs8 .Li1
sub $r9, $r6, $r9
move P2L, #0
.LL7:
move $r10, #0x20
slt $r15, $r9, $r10
bnezs8 .LL8
or P2L, P2L, O1L
move O1L, O1H
move O1H, #0
addi $r9, $r9, #0xffffffe0
bnez O1L, .LL7
.LL8:
beqz $r9, .LEadd3
move P1L, O1H
move $r10, O1L
srl O1L, O1L, $r9
srl O1H, O1H, $r9
subri $r9, $r9, #0x20
sll P1L, P1L, $r9
or O1L, O1L, P1L
sll $r10, $r10, $r9
or P2L, P2L, $r10
beqz P2L, .LEadd3
ori O1L, O1L, #1
j .LEadd3
.Li1:
move $r15, $r6
move $r6, $r9
sub $r9, $r9, $r15
move P2L, #0
.LL10:
move $r10, #0x20
slt $r15, $r9, $r10
bnezs8 .LL11
or P2L, P2L, P3L
move P3L, P3H
move P3H, #0
addi $r9, $r9, #0xffffffe0
bnez P3L, .LL10
.LL11:
beqz $r9, .LEadd3
move P1L, P3H
move $r10, P3L
srl P3L, P3L, $r9
srl P3H, P3H, $r9
subri $r9, $r9, #0x20
sll P1L, P1L, $r9
or P3L, P3L, P1L
sll $r10, $r10, $r9
or P2L, P2L, $r10
beqz P2L, .LEadd3
ori P3L, P3L, #1
.LEadd3:
xor $r10, P1H, P2H
sltsi $r15, $r10, #0
bnezs8 .LEsub1
#ADD(P3L, O1L)
add P3L, P3L, O1L
slt $r15, P3L, O1L
#ADDCC(P3H, O1H)
beqzs8 .LL13
add P3H, P3H, O1H
slt $r15, P3H, O1H
beqzs8 .LL14
addi P3H, P3H, #0x1
j .LL15
.LL14:
move $r15, #1
add P3H, P3H, $r15
slt $r15, P3H, $r15
j .LL15
.LL13:
add P3H, P3H, O1H
slt $r15, P3H, O1H
.LL15:
beqzs8 .LEres
andi $r10, P3L, #1
beqz $r10, .Li3
ori P3L, P3L, #2
.Li3:
srli P3L, P3L, #1
slli $r10, P3H, #31
or P3L, P3L, $r10
srli P3H, P3H, #1
move $r10, #0x80000000
or P3H, P3H, $r10
addi $r6, $r6, #1
subri $r15, $r6, #0x7ff
bnezs8 .LEres
move $r10, #0x7ff00000
or P1H, P1H, $r10
move P1L, #0
j .LEretA
.LEsub1:
#SUB(P3L, O1L)
move $r15, P3L
sub P3L, P3L, O1L
slt $r15, $r15, P3L
#SUBCC(P3H, O1H)
beqzs8 .LL16
move $r15, P3H
sub P3H, P3H, O1H
slt $r15, $r15, P3H
beqzs8 .LL17
subi333 P3H, P3H, #1
j .LL18
.LL17:
move $r15, P3H
subi333 P3H, P3H, #1
slt $r15, $r15, P3H
j .LL18
.LL16:
move $r15, P3H
sub P3H, P3H, O1H
slt $r15, $r15, P3H
.LL18:
beqzs8 .Li5
move $r10, #0x80000000
xor P1H, P1H, $r10
subri P3H, P3H, #0
beqz P3L, .LL19
subri P3L, P3L, #0
subi45 P3H, #1
.LL19:
.Li5:
#NORMd($r4, $r9, P1L)
bnez P3H, .LL20
bnez P3L, .LL21
move $r6, #0
j .LL22
.LL21:
move P3H, P3L
move P3L, #0
move $r9, #32
sub $r6, $r6, $r9
.LL20:
#ifdef __NDS32_PERF_EXT__
clz $r9, P3H
#else
pushm $r0, $r5
move $r0, P3H
bal __clzsi2
move $r9, $r0
popm $r0, $r5
#endif
beqz $r9, .LL22
sub $r6, $r6, $r9
subri P1L, $r9, #32
srl P1L, P3L, P1L
sll P3L, P3L, $r9
sll P3H, P3H, $r9
or P3H, P3H, P1L
.LL22:
#NORMd End
or $r10, P3H, P3L
bnez $r10, .LEres
move P1H, #0
.LEres:
blez $r6, .LEund
.LElab8:
#ADD(P3L, $0x400)
move $r15, #0x400
add P3L, P3L, $r15
slt $r15, P3L, $r15
#ADDCC(P3H, $0x0)
beqzs8 .LL25
add P3H, P3H, $r15
slt $r15, P3H, $r15
.LL25:
#ADDC($r6, $0x0)
add $r6, $r6, $r15
srli $r10, P3L, #11
andi $r10, $r10, #1
sub P3L, P3L, $r10
srli P1L, P3L, #11
slli $r10, P3H, #21
or P1L, P1L, $r10
slli $r10, P3H, #1
srli $r10, $r10, #12
or P1H, P1H, $r10
slli $r10, $r6, #20
or P1H, P1H, $r10
.LEretA:
.LE999:
popm $r6, $r10
pop $lp
ret5 $lp
.LEspecA:
#ADD(P3L, P3L)
move $r15, P3L
add P3L, P3L, P3L
slt $r15, P3L, $r15
#ADDC(P3H, P3H)
add P3H, P3H, P3H
add P3H, P3H, $r15
bnez $r6, .Li7
or $r10, P3H, P3L
beqz $r10, .Li8
j .LElab1
.Li8:
subri $r15, $r9, #0x7ff
beqzs8 .LEspecB
add P3L, P2H, P2H
or $r10, P3L, P2L
bnez $r10, .LEretB
sltsi $r15, P2H, #0
bnezs8 .LEretA
.LEretB:
move P1L, P2L
move P1H, P2H
j .LE999
.Li7:
or $r10, P3H, P3L
bnez $r10, .LEnan
subri $r15, $r9, #0x7ff
bnezs8 .LEretA
xor $r10, P1H, P2H
sltsi $r15, $r10, #0
bnezs8 .LEnan
j .LEretB
.LEspecB:
#ADD(O1L, O1L)
move $r15, O1L
add O1L, O1L, O1L
slt $r15, O1L, $r15
#ADDC(O1H, O1H)
add O1H, O1H, O1H
add O1H, O1H, $r15
bnez $r9, .Li11
or $r10, O1H, O1L
beqz $r10, .LEretA
j .LElab2
.Li11:
or $r10, O1H, O1L
beqz $r10, .LEretB
.LEnan:
move P1H, #0xfff80000
move P1L, #0
j .LEretA
.LEund:
subri $r9, $r6, #1
move P2L, #0
.LL26:
move $r10, #0x20
slt $r15, $r9, $r10
bnezs8 .LL27
or P2L, P2L, P3L
move P3L, P3H
move P3H, #0
addi $r9, $r9, #0xffffffe0
bnez P3L, .LL26
.LL27:
beqz $r9, .LL28
move P1L, P3H
move $r10, P3L
srl P3L, P3L, $r9
srl P3H, P3H, $r9
subri $r9, $r9, #0x20
sll P1L, P1L, $r9
or P3L, P3L, P1L
sll $r10, $r10, $r9
or P2L, P2L, $r10
beqz P2L, .LL28
ori P3L, P3L, #1
.LL28:
move $r6, #0
j .LElab8
.size __subdf3, .-__subdf3
.size __adddf3, .-__adddf3
#endif /* L_addsub_df */
#ifdef L_mul_sf
#if !defined (__big_endian__)
#define P1L $r0
#define P1H $r1
#define P2L $r2
#define P2H $r3
#else
#define P1H $r0
#define P1L $r1
#define P2H $r2
#define P2L $r3
#endif
.text
.align 2
.global __mulsf3
.type __mulsf3, @function
__mulsf3:
push $lp
pushm $r6, $r10
srli $r3, $r0, #23
andi $r3, $r3, #0xff
srli $r5, $r1, #23
andi $r5, $r5, #0xff
move $r6, #0x80000000
slli $r2, $r0, #8
or $r2, $r2, $r6
slli $r4, $r1, #8
or $r4, $r4, $r6
xor $r8, $r0, $r1
and $r6, $r6, $r8
addi $r8, $r3, #-1
slti $r15, $r8, #0xfe
beqzs8 .LFspecA
.LFlab1:
addi $r8, $r5, #-1
slti $r15, $r8, #0xfe
beqzs8 .LFspecB
.LFlab2:
move $r10, $r3
/* This is a 64-bit multiple. ($r2, $r7) is (high, low). */
#ifndef __NDS32_ISA_V3M__
mulr64 $r2, $r2, $r4
#else
pushm $r0, $r1
pushm $r4, $r5
move P1L, $r2
movi P1H, #0
move P2L, $r4
movi P2H, #0
bal __muldi3
movd44 $r2, $r0
popm $r4, $r5
popm $r0, $r1
#endif
#ifndef __big_endian__
move $r7, $r2
move $r2, $r3
#else
move $r7, $r3
#endif
move $r3, $r10
beqz $r7, .Li17
ori $r2, $r2, #1
.Li17:
sltsi $r15, $r2, #0
bnezs8 .Li18
slli $r2, $r2, #1
addi $r3, $r3, #-1
.Li18:
addi $r8, $r5, #0xffffff82
add $r3, $r3, $r8
addi $r8, $r3, #-1
slti $r15, $r8, #0xfe
beqzs8 .LFoveund
.LFlab8:
#ADD($r2, $0x80)
move $r15, #0x80
add $r2, $r2, $r15
slt $r15, $r2, $r15
#ADDC($r3, $0x0)
add $r3, $r3, $r15
srli $r8, $r2, #8
andi $r8, $r8, #1
sub $r2, $r2, $r8
slli $r2, $r2, #1
srli $r2, $r2, #9
slli $r8, $r3, #23
or $r2, $r2, $r8
or $r0, $r2, $r6
.LF999:
popm $r6, $r10
pop $lp
ret5 $lp
.LFspecA:
bnez $r3, .Li19
add $r2, $r2, $r2
beqz $r2, .Li20
#ifdef __NDS32_PERF_EXT__
clz $r7, $r2
#else
pushm $r0, $r5
move $r0, $r2
bal __clzsi2
move $r7, $r0
popm $r0, $r5
#endif
sub $r3, $r3, $r7
sll $r2, $r2, $r7
j .LFlab1
.Li20:
subri $r15, $r5, #0xff
beqzs8 .LFnan
j .LFzer
.Li19:
add $r8, $r2, $r2
bnez $r8, .LFnan
bnez $r5, .Li21
add $r8, $r4, $r4
beqz $r8, .LFnan
.Li21:
subri $r15, $r5, #0xff
bnezs8 .LFinf
.LFspecB:
bnez $r5, .Li22
add $r4, $r4, $r4
beqz $r4, .LFzer
#ifdef __NDS32_PERF_EXT__
clz $r7, $r4
#else
pushm $r0, $r5
move $r0, $r4
bal __clzsi2
move $r7, $r0
popm $r0, $r5
#endif
sub $r5, $r5, $r7
sll $r4, $r4, $r7
j .LFlab2
.LFzer:
move $r0, $r6
j .LF999
.Li22:
add $r8, $r4, $r4
bnez $r8, .LFnan
.LFinf:
move $r8, #0x7f800000
or $r0, $r6, $r8
j .LF999
.LFnan:
move $r0, #0xffc00000
j .LF999
.LFoveund:
bgtz $r3, .LFinf
subri $r7, $r3, #1
slti $r15, $r7, #0x20
beqzs8 .LFzer
subri $r8, $r7, #0x20
sll $r3, $r2, $r8
srl $r2, $r2, $r7
beqz $r3, .Li25
ori $r2, $r2, #2
.Li25:
move $r3, #0
addi $r8, $r2, #0x80
sltsi $r15, $r8, #0
beqzs8 .LFlab8
move $r3, #1
j .LFlab8
.size __mulsf3, .-__mulsf3
#endif /* L_mul_sf */
#ifdef L_mul_df
#ifndef __big_endian__
#define P1L $r0
#define P1H $r1
#define P2L $r2
#define P2H $r3
#define P3L $r4
#define P3H $r5
#define O1L $r7
#define O1H $r8
#else
#define P1H $r0
#define P1L $r1
#define P2H $r2
#define P2L $r3
#define P3H $r4
#define P3L $r5
#define O1H $r7
#define O1L $r8
#endif
.text
.align 2
.global __muldf3
.type __muldf3, @function
__muldf3:
push $lp
pushm $r6, $r10
slli $r6, P1H, #1
srli $r6, $r6, #21
slli P3H, P1H, #11
srli $r10, P1L, #21
or P3H, P3H, $r10
slli P3L, P1L, #11
move O1L, #0x80000000
or P3H, P3H, O1L
slli $r9, P2H, #1
srli $r9, $r9, #21
slli O1H, P2H, #11
srli $r10, P2L, #21
or O1H, O1H, $r10
or O1H, O1H, O1L
xor P1H, P1H, P2H
and P1H, P1H, O1L
slli O1L, P2L, #11
addi $r10, $r6, #-1
slti $r15, $r10, #0x7fe
beqzs8 .LFspecA
.LFlab1:
addi $r10, $r9, #-1
slti $r15, $r10, #0x7fe
beqzs8 .LFspecB
.LFlab2:
addi $r10, $r9, #0xfffffc02
add $r6, $r6, $r10
move $r10, $r8
/* This is a 64-bit multiple. */
#ifndef __big_endian__
/* For little endian: ($r9, $r3) is (high, low). */
#ifndef __NDS32_ISA_V3M__
mulr64 $r8, $r5, $r8
#else
pushm $r0, $r5
move $r0, $r5
movi $r1, #0
move $r2, $r8
movi $r3, #0
bal __muldi3
movd44 $r8, $r0
popm $r0, $r5
#endif
move $r3, $r8
#else /* __big_endian__ */
/* For big endain: ($r9, $r2) is (high, low). */
#ifndef __NDS32_ISA_V3M__
mulr64 $r8, $r4, $r7
#else
pushm $r0, $r5
move $r1, $r4
movi $r0, #0
move $r3, $r7
movi $r2, #0
bal __muldi3
movd44 $r8, $r0
popm $r0, $r5
#endif
move $r2, $r9
move $r9, $r8
#endif /* __big_endian__ */
move $r8, $r10
move $r10, P1H
/* This is a 64-bit multiple. */
#ifndef __big_endian__
/* For little endian: ($r0, $r2) is (high, low). */
#ifndef __NDS32_ISA_V3M__
mulr64 $r0, $r4, $r8
#else
pushm $r2, $r5
move $r0, $r4
movi $r1, #0
move $r2, $r8
movi $r3, #0
bal __muldi3
popm $r2, $r5
#endif
move $r2, $r0
move $r0, $r1
#else /* __big_endian__ */
/* For big endain: ($r1, $r3) is (high, low). */
#ifndef __NDS32_ISA_V3M__
mulr64 $r0, $r5, $r7
#else
pushm $r2, $r5
move $r1, $r5
movi $r0, #0
move $r3, $r7
movi $r2, #0
bal __muldi3
popm $r2, $r5
#endif
move $r3, $r1
move $r1, $r0
#endif /* __big_endian__ */
move P1H, $r10
#ADD(P2H, P1L)
add P2H, P2H, P1L
slt $r15, P2H, P1L
#ADDC($r9, $0x0)
add $r9, $r9, $r15
move $r10, P1H
/* This is a 64-bit multiple. */
#ifndef __big_endian__
/* For little endian: ($r0, $r8) is (high, low). */
#ifndef __NDS32_ISA_V3M__
mulr64 $r0, $r5, $r7
#else
pushm $r2, $r5
move $r0, $r5
movi $r1, #0
move $r2, $r7
movi $r3, #0
bal __muldi3
popm $r2, $r5
#endif
move $r8, $r0
move $r0, $r1
#else /* __big_endian__ */
/* For big endian: ($r1, $r7) is (high, low). */
#ifndef __NDS32_ISA_V3M__
mulr64 $r0, $r4, $r8
#else
pushm $r2, $r5
move $r1, $r4
movi $r0, #0
move $r3, $r8
movi $r2, #0
bal __muldi3
popm $r2, $r5
#endif
move $r7, $r1
move $r1, $r0
#endif /* __big_endian__ */
move P1H, $r10
#ADD(P2L, O1H)
add P2L, P2L, O1H
slt $r15, P2L, O1H
#ADDCC(P2H, P1L)
beqzs8 .LL29
add P2H, P2H, P1L
slt $r15, P2H, P1L
beqzs8 .LL30
addi P2H, P2H, #0x1
j .LL31
.LL30:
move $r15, #1
add P2H, P2H, $r15
slt $r15, P2H, $r15
j .LL31
.LL29:
add P2H, P2H, P1L
slt $r15, P2H, P1L
.LL31:
#ADDC($r9, $0x0)
add $r9, $r9, $r15
/* This is a 64-bit multiple. */
#ifndef __big_endian__
/* For little endian: ($r8, $r0) is (high, low). */
move $r10, $r9
#ifndef __NDS32_ISA_V3M__
mulr64 $r8, $r4, $r7
#else
pushm $r0, $r5
move $r0, $r4
movi $r1, #0
move $r2, $r7
movi $r3, #0
bal __muldi3
movd44 $r8, $r0
popm $r0, $r5
#endif
move $r0, $r8
move $r8, $r9
move $r9, $r10
#else /* __big_endian__ */
/* For big endian: ($r7, $r1) is (high, low). */
move $r10, $r6
#ifndef __NDS32_ISA_V3M__
mulr64 $r6, $r5, $r8
#else
pushm $r0, $r5
move $r1, $r5
movi $r0, #0
move $r3, $r8
movi $r2, #0
bal __muldi3
movd44 $r6, $r0
popm $r0, $r5
#endif
move $r1, $r7
move $r7, $r6
move $r6, $r10
#endif /* __big_endian__ */
#ADD(P2L, O1H)
add P2L, P2L, O1H
slt $r15, P2L, O1H
#ADDCC(P2H, $0x0)
beqzs8 .LL34
add P2H, P2H, $r15
slt $r15, P2H, $r15
.LL34:
#ADDC($r9, $0x0)
add $r9, $r9, $r15
or $r10, P1L, P2L
beqz $r10, .Li13
ori P2H, P2H, #1
.Li13:
move P3H, $r9
move P3L, P2H
sltsi $r15, P3H, #0
bnezs8 .Li14
move $r15, P3L
add P3L, P3L, P3L
slt $r15, P3L, $r15
add P3H, P3H, P3H
add P3H, P3H, $r15
addi $r6, $r6, #-1
.Li14:
addi $r10, $r6, #-1
slti $r15, $r10, #0x7fe
beqzs8 .LFoveund
#ADD(P3L, $0x400)
move $r15, #0x400
add P3L, P3L, $r15
slt $r15, P3L, $r15
#ADDCC(P3H, $0x0)
beqzs8 .LL37
add P3H, P3H, $r15
slt $r15, P3H, $r15
.LL37:
#ADDC($r6, $0x0)
add $r6, $r6, $r15
.LFlab8:
srli $r10, P3L, #11
andi $r10, $r10, #1
sub P3L, P3L, $r10
srli P1L, P3L, #11
slli $r10, P3H, #21
or P1L, P1L, $r10
slli $r10, P3H, #1
srli $r10, $r10, #12
or P1H, P1H, $r10
slli $r10, $r6, #20
or P1H, P1H, $r10
.LFret:
.LF999:
popm $r6, $r10
pop $lp
ret5 $lp
.LFspecA:
#ADD(P3L, P3L)
move $r15, P3L
add P3L, P3L, P3L
slt $r15, P3L, $r15
#ADDC(P3H, P3H)
add P3H, P3H, P3H
add P3H, P3H, $r15
bnez $r6, .Li15
or $r10, P3H, P3L
beqz $r10, .Li16
#NORMd($r4, P1L, P2H)
bnez P3H, .LL38
bnez P3L, .LL39
move $r6, #0
j .LL40
.LL39:
move P3H, P3L
move P3L, #0
move P1L, #32
sub $r6, $r6, P1L
.LL38:
#ifndef __big_endian__
#ifdef __NDS32_PERF_EXT__
clz $r0, P3H
#else
pushm $r1, P3H
move $r0, P3H
bal __clzsi2
popm $r1, $r5
#endif
#else /* __big_endian__ */
#ifdef __NDS32_PERF_EXT__
clz $r1, $r4
#else
push $r0
pushm $r2, $r5
move $r0, $r4
bal __clzsi2
move $r1, $r0
popm $r2, $r5
pop $r0
#endif
#endif /* __big_endian__ */
beqz P1L, .LL40
sub $r6, $r6, P1L
subri P2H, P1L, #32
srl P2H, P3L, P2H
sll P3L, P3L, P1L
sll P3H, P3H, P1L
or P3H, P3H, P2H
.LL40:
#NORMd End
j .LFlab1
.Li16:
subri $r15, $r9, #0x7ff
beqzs8 .LFnan
j .LFret
.Li15:
or $r10, P3H, P3L
bnez $r10, .LFnan
bnez $r9, .Li17
slli $r10, O1H, #1
or $r10, $r10, O1L
beqz $r10, .LFnan
.Li17:
subri $r15, $r9, #0x7ff
bnezs8 .LFinf
.LFspecB:
#ADD(O1L, O1L)
move $r15, O1L
add O1L, O1L, O1L
slt $r15, O1L, $r15
#ADDC(O1H, O1H)
add O1H, O1H, O1H
add O1H, O1H, $r15
bnez $r9, .Li18
or $r10, O1H, O1L
beqz $r10, .Li19
#NORMd($r7, P2L, P1L)
bnez O1H, .LL41
bnez O1L, .LL42
move $r9, #0
j .LL43
.LL42:
move O1H, O1L
move O1L, #0
move P2L, #32
sub $r9, $r9, P2L
.LL41:
#ifndef __big_endian__
#ifdef __NDS32_PERF_EXT__
clz $r2, $r8
#else
pushm $r0, $r1
pushm $r3, $r5
move $r0, $r8
bal __clzsi2
move $r2, $r0
popm $r3, $r5
popm $r0, $r1
#endif
#else /* __big_endian__ */
#ifdef __NDS32_PERF_EXT__
clz $r3, $r7
#else
pushm $r0, $r2
pushm $r4, $r5
move $r0, $r7
bal __clzsi2
move $r3, $r0
popm $r4, $r5
popm $r0, $r2
#endif
#endif /* __big_endian__ */
beqz P2L, .LL43
sub $r9, $r9, P2L
subri P1L, P2L, #32
srl P1L, O1L, P1L
sll O1L, O1L, P2L
sll O1H, O1H, P2L
or O1H, O1H, P1L
.LL43:
#NORMd End
j .LFlab2
.Li19:
move P1L, #0
j .LFret
.Li18:
or $r10, O1H, O1L
bnez $r10, .LFnan
.LFinf:
move $r10, #0x7ff00000
or P1H, P1H, $r10
move P1L, #0
j .LFret
.LFnan:
move P1H, #0xfff80000
move P1L, #0
j .LFret
.LFoveund:
bgtz $r6, .LFinf
subri P1L, $r6, #1
move P2L, #0
.LL44:
move $r10, #0x20
slt $r15, P1L, $r10
bnezs8 .LL45
or P2L, P2L, P3L
move P3L, P3H
move P3H, #0
addi P1L, P1L, #0xffffffe0
bnez P3L, .LL44
.LL45:
beqz P1L, .LL46
move P2H, P3H
move $r10, P3L
srl P3L, P3L, P1L
srl P3H, P3H, P1L
subri P1L, P1L, #0x20
sll P2H, P2H, P1L
or P3L, P3L, P2H
sll $r10, $r10, P1L
or P2L, P2L, $r10
beqz P2L, .LL46
ori P3L, P3L, #1
.LL46:
#ADD(P3L, $0x400)
move $r15, #0x400
add P3L, P3L, $r15
slt $r15, P3L, $r15
#ADDC(P3H, $0x0)
add P3H, P3H, $r15
srli $r6, P3H, #31
j .LFlab8
.size __muldf3, .-__muldf3
#endif /* L_mul_df */
#ifdef L_div_sf
.text
.align 2
.global __divsf3
.type __divsf3, @function
__divsf3:
push $lp
pushm $r6, $r10
move $r7, #0x80000000
srli $r4, $r0, #23
andi $r4, $r4, #0xff
srli $r6, $r1, #23
andi $r6, $r6, #0xff
slli $r3, $r0, #8
or $r3, $r3, $r7
slli $r5, $r1, #8
or $r5, $r5, $r7
xor $r10, $r0, $r1
and $r7, $r7, $r10
addi $r10, $r4, #-1
slti $r15, $r10, #0xfe
beqzs8 .LGspecA
.LGlab1:
addi $r10, $r6, #-1
slti $r15, $r10, #0xfe
beqzs8 .LGspecB
.LGlab2:
slt $r15, $r3, $r5
bnezs8 .Li27
srli $r3, $r3, #1
addi $r4, $r4, #1
.Li27:
srli $r8, $r5, #14
divr $r0, $r2, $r3, $r8
andi $r9, $r5, #0x3fff
mul $r1, $r9, $r0
slli $r2, $r2, #14
#SUB($r2, $r1)
move $r15, $r2
sub $r2, $r2, $r1
slt $r15, $r15, $r2
beqzs8 .Li28
addi $r0, $r0, #-1
#ADD($r2, $r5)
add $r2, $r2, $r5
slt $r15, $r2, $r5
.Li28:
divr $r3, $r2, $r2, $r8
mul $r1, $r9, $r3
slli $r2, $r2, #14
#SUB($r2, $r1)
move $r15, $r2
sub $r2, $r2, $r1
slt $r15, $r15, $r2
beqzs8 .Li29
addi $r3, $r3, #-1
#ADD($r2, $r5)
add $r2, $r2, $r5
slt $r15, $r2, $r5
.Li29:
slli $r10, $r0, #14
add $r3, $r3, $r10
slli $r3, $r3, #4
beqz $r2, .Li30
ori $r3, $r3, #1
.Li30:
subri $r10, $r6, #0x7e
add $r4, $r4, $r10
addi $r10, $r4, #-1
slti $r15, $r10, #0xfe
beqzs8 .LGoveund
.LGlab8:
#ADD($r3, $0x80)
move $r15, #0x80
add $r3, $r3, $r15
slt $r15, $r3, $r15
#ADDC($r4, $0x0)
add $r4, $r4, $r15
srli $r10, $r3, #8
andi $r10, $r10, #1
sub $r3, $r3, $r10
slli $r3, $r3, #1
srli $r3, $r3, #9
slli $r10, $r4, #23
or $r3, $r3, $r10
or $r0, $r3, $r7
.LG999:
popm $r6, $r10
pop $lp
ret5 $lp
.LGspecA:
bnez $r4, .Li31
add $r3, $r3, $r3
beqz $r3, .Li31
#ifdef __NDS32_PERF_EXT__
clz $r8, $r3
#else
pushm $r0, $r5
move $r0, $r3
bal __clzsi2
move $r8, $r0
popm $r0, $r5
#endif
sub $r4, $r4, $r8
sll $r3, $r3, $r8
j .LGlab1
.Li31:
bne $r6, $r4, .Li33
add $r10, $r5, $r5
beqz $r10, .LGnan
.Li33:
subri $r15, $r6, #0xff
beqzs8 .LGspecB
beqz $r4, .LGzer
add $r10, $r3, $r3
bnez $r10, .LGnan
j .LGinf
.LGspecB:
bnez $r6, .Li34
add $r5, $r5, $r5
beqz $r5, .LGinf
#ifdef __NDS32_PERF_EXT__
clz $r8, $r5
#else
pushm $r0, $r5
move $r0, $r5
bal __clzsi2
move $r8, $r0
popm $r0, $r5
#endif
sub $r6, $r6, $r8
sll $r5, $r5, $r8
j .LGlab2
.Li34:
add $r10, $r5, $r5
bnez $r10, .LGnan
.LGzer:
move $r0, $r7
j .LG999
.LGoveund:
bgtz $r4, .LGinf
subri $r8, $r4, #1
slti $r15, $r8, #0x20
beqzs8 .LGzer
subri $r10, $r8, #0x20
sll $r4, $r3, $r10
srl $r3, $r3, $r8
beqz $r4, .Li37
ori $r3, $r3, #2
.Li37:
move $r4, #0
addi $r10, $r3, #0x80
sltsi $r15, $r10, #0
beqzs8 .LGlab8
move $r4, #1
j .LGlab8
.LGinf:
move $r10, #0x7f800000
or $r0, $r7, $r10
j .LG999
.LGnan:
move $r0, #0xffc00000
j .LG999
.size __divsf3, .-__divsf3
#endif /* L_div_sf */
#ifdef L_div_df
#ifndef __big_endian__
#define P1L $r0
#define P1H $r1
#define P2L $r2
#define P2H $r3
#define P3L $r4
#define P3H $r5
#define O1L $r7
#define O1H $r8
#else
#define P1H $r0
#define P1L $r1
#define P2H $r2
#define P2L $r3
#define P3H $r4
#define P3L $r5
#define O1H $r7
#define O1L $r8
#endif
.text
.align 2
.global __divdf3
.type __divdf3, @function
__divdf3:
push $lp
pushm $r6, $r10
slli $r6, P1H, #1
srli $r6, $r6, #21
slli P3H, P1H, #11
srli $r10, P1L, #21
or P3H, P3H, $r10
slli P3L, P1L, #11
move O1L, #0x80000000
or P3H, P3H, O1L
slli $r9, P2H, #1
srli $r9, $r9, #21
slli O1H, P2H, #11
srli $r10, P2L, #21
or O1H, O1H, $r10
or O1H, O1H, O1L
xor P1H, P1H, P2H
and P1H, P1H, O1L
slli O1L, P2L, #11
addi $r10, $r6, #-1
slti $r15, $r10, #0x7fe
beqzs8 .LGspecA
.LGlab1:
addi $r10, $r9, #-1
slti $r15, $r10, #0x7fe
beqzs8 .LGspecB
.LGlab2:
sub $r6, $r6, $r9
addi $r6, $r6, #0x3ff
srli P3L, P3L, #1
slli $r10, P3H, #31
or P3L, P3L, $r10
srli P3H, P3H, #1
srli $r9, O1H, #16
divr P2H, P3H, P3H, $r9
move $r10, #0xffff
and P2L, O1H, $r10
mul P1L, P2L, P2H
slli P3H, P3H, #16
srli $r10, P3L, #16
or P3H, P3H, $r10
#SUB(P3H, P1L)
move $r15, P3H
sub P3H, P3H, P1L
slt $r15, $r15, P3H
beqzs8 .Li20
.Lb21:
addi P2H, P2H, #-1
add P3H, P3H, O1H
slt $r15, P3H, O1H
beqzs8 .Lb21
.Li20:
divr $r9, P3H, P3H, $r9
mul P1L, P2L, $r9
slli P3H, P3H, #16
move $r15, #0xffff
and $r10, P3L, $r15
or P3H, P3H, $r10
#SUB(P3H, P1L)
move $r15, P3H
sub P3H, P3H, P1L
slt $r15, $r15, P3H
beqzs8 .Li22
.Lb23:
addi $r9, $r9, #-1
add P3H, P3H, O1H
slt $r15, P3H, O1H
beqzs8 .Lb23
.Li22:
slli P2H, P2H, #16
add P2H, P2H, $r9
/* This is a 64-bit multiple. */
#ifndef __big_endian__
/* For little endian: ($r0, $r9) is (high, low). */
move $r10, $r1
#ifndef __NDS32_ISA_V3M__
mulr64 $r0, $r3, $r7
#else
pushm $r2, $r5
move $r0, $r3
movi $r1, #0
move $r2, $r7
movi $r3, #0
bal __muldi3
popm $r2, $r5
#endif
move $r9, $r0
move $r0, $r1
move $r1, $r10
#else /* __big_endian__ */
/* For big endian: ($r1, $r9) is (high, low). */
move $r10, $r0
#ifndef __NDS32_ISA_V3M__
mulr64 $r0, $r2, $r8
#else
pushm $r2, $r5
move $r1, $r2
movi $r0, #0
move $r3, $r8
movi $r2, #0
bal __muldi3
popm $r2, $r5
#endif
move $r9, $r1
move $r1, $r0
move $r0, $r10
#endif /* __big_endian__ */
move P3L, #0
#SUB(P3L, $r9)
move $r15, P3L
sub P3L, P3L, $r9
slt $r15, $r15, P3L
#SUBCC(P3H, P1L)
beqzs8 .LL47
move $r15, P3H
sub P3H, P3H, P1L
slt $r15, $r15, P3H
beqzs8 .LL48
subi333 P3H, P3H, #1
j .LL49
.LL48:
move $r15, P3H
subi333 P3H, P3H, #1
slt $r15, $r15, P3H
j .LL49
.LL47:
move $r15, P3H
sub P3H, P3H, P1L
slt $r15, $r15, P3H
.LL49:
beqzs8 .Li24
.LGlab3:
addi P2H, P2H, #-1
#ADD(P3L, O1L)
add P3L, P3L, O1L
slt $r15, P3L, O1L
#ADDCC(P3H, O1H)
beqzs8 .LL50
add P3H, P3H, O1H
slt $r15, P3H, O1H
beqzs8 .LL51
addi P3H, P3H, #0x1
j .LL52
.LL51:
move $r15, #1
add P3H, P3H, $r15
slt $r15, P3H, $r15
j .LL52
.LL50:
add P3H, P3H, O1H
slt $r15, P3H, O1H
.LL52:
beqzs8 .LGlab3
.Li24:
bne P3H, O1H, .Li25
move P1L, O1L
move P3H, P3L
move $r9, #0
move P2L, $r9
j .Le25
.Li25:
srli P2L, O1H, #16
divr $r9, P3H, P3H, P2L
move $r10, #0xffff
and $r10, O1H, $r10
mul P1L, $r10, $r9
slli P3H, P3H, #16
srli $r15, P3L, #16
or P3H, P3H, $r15
#SUB(P3H, P1L)
move $r15, P3H
sub P3H, P3H, P1L
slt $r15, $r15, P3H
beqzs8 .Li26
.Lb27:
addi $r9, $r9, #-1
add P3H, P3H, O1H
slt $r15, P3H, O1H
beqzs8 .Lb27
.Li26:
divr P2L, P3H, P3H, P2L
mul P1L, $r10, P2L
slli P3H, P3H, #16
move $r10, #0xffff
and $r10, P3L, $r10
or P3H, P3H, $r10
#SUB(P3H, P1L)
move $r15, P3H
sub P3H, P3H, P1L
slt $r15, $r15, P3H
beqzs8 .Li28
.Lb29:
addi P2L, P2L, #-1
add P3H, P3H, O1H
slt $r15, P3H, O1H
beqzs8 .Lb29
.Li28:
slli $r9, $r9, #16
add $r9, $r9, P2L
/* This is a 64-bit multiple. */
#ifndef __big_endian__
/* For little endian: ($r0, $r2) is (high, low). */
move $r10, $r1
#ifndef __NDS32_ISA_V3M__
mulr64 $r0, $r9, $r7
#else
pushm $r2, $r5
move $r0, $r9
movi $r1, #0
move $r2, $r7
movi $r3, #0
bal __muldi3
popm $r2, $r5
#endif
move $r2, $r0
move $r0, $r1
move $r1, $r10
#else /* __big_endian__ */
/* For big endian: ($r1, $r3) is (high, low). */
move $r10, $r0
#ifndef __NDS32_ISA_V3M__
mulr64 $r0, $r9, $r8
#else
pushm $r2, $r5
move $r0, $r9
movi $r1, #0
move $r2, $r7
movi $r3, #0
bal __muldi3
popm $r2, $r5
#endif
move $r3, $r1
move $r1, $r0
move $r0, $r10
#endif /* __big_endian__ */
.Le25:
move P3L, #0
#SUB(P3L, P2L)
move $r15, P3L
sub P3L, P3L, P2L
slt $r15, $r15, P3L
#SUBCC(P3H, P1L)
beqzs8 .LL53
move $r15, P3H
sub P3H, P3H, P1L
slt $r15, $r15, P3H
beqzs8 .LL54
subi333 P3H, P3H, #1
j .LL55
.LL54:
move $r15, P3H
subi333 P3H, P3H, #1
slt $r15, $r15, P3H
j .LL55
.LL53:
move $r15, P3H
sub P3H, P3H, P1L
slt $r15, $r15, P3H
.LL55:
beqzs8 .Li30
.LGlab4:
addi $r9, $r9, #-1
#ADD(P3L, O1L)
add P3L, P3L, O1L
slt $r15, P3L, O1L
#ADDCC(P3H, O1H)
beqzs8 .LL56
add P3H, P3H, O1H
slt $r15, P3H, O1H
beqzs8 .LL57
addi P3H, P3H, #0x1
j .LL58
.LL57:
move $r15, #1
add P3H, P3H, $r15
slt $r15, P3H, $r15
j .LL58
.LL56:
add P3H, P3H, O1H
slt $r15, P3H, O1H
.LL58:
beqzs8 .LGlab4
.Li30:
sltsi $r15, P2H, #0
bnezs8 .Li31
#ADD($r9, $r9)
move $r15, $r9
add $r9, $r9, $r9
slt $r15, $r9, $r15
#ADDC(P2H, P2H)
add P2H, P2H, P2H
add P2H, P2H, $r15
addi $r6, $r6, #-1
.Li31:
or $r10, P3H, P3L
beqz $r10, .Li32
ori $r9, $r9, #1
.Li32:
move P3H, P2H
move P3L, $r9
addi $r10, $r6, #-1
slti $r15, $r10, #0x7fe
beqzs8 .LGoveund
#ADD(P3L, $0x400)
move $r15, #0x400
add P3L, P3L, $r15
slt $r15, P3L, $r15
#ADDCC(P3H, $0x0)
beqzs8 .LL61
add P3H, P3H, $r15
slt $r15, P3H, $r15
.LL61:
#ADDC($r6, $0x0)
add $r6, $r6, $r15
.LGlab8:
srli $r10, P3L, #11
andi $r10, $r10, #1
sub P3L, P3L, $r10
srli P1L, P3L, #11
slli $r10, P3H, #21
or P1L, P1L, $r10
slli $r10, P3H, #1
srli $r10, $r10, #12
or P1H, P1H, $r10
slli $r10, $r6, #20
or P1H, P1H, $r10
.LGret:
.LG999:
popm $r6, $r10
pop $lp
ret5 $lp
.LGoveund:
bgtz $r6, .LGinf
subri P2H, $r6, #1
move P1L, #0
.LL62:
move $r10, #0x20
slt $r15, P2H, $r10
bnezs8 .LL63
or P1L, P1L, P3L
move P3L, P3H
move P3H, #0
addi P2H, P2H, #0xffffffe0
bnez P3L, .LL62
.LL63:
beqz P2H, .LL64
move P2L, P3H
move $r10, P3L
srl P3L, P3L, P2H
srl P3H, P3H, P2H
subri P2H, P2H, #0x20
sll P2L, P2L, P2H
or P3L, P3L, P2L
sll $r10, $r10, P2H
or P1L, P1L, $r10
beqz P1L, .LL64
ori P3L, P3L, #1
.LL64:
#ADD(P3L, $0x400)
move $r15, #0x400
add P3L, P3L, $r15
slt $r15, P3L, $r15
#ADDC(P3H, $0x0)
add P3H, P3H, $r15
srli $r6, P3H, #31
j .LGlab8
.LGspecA:
#ADD(P3L, P3L)
move $r15, P3L
add P3L, P3L, P3L
slt $r15, P3L, $r15
#ADDC(P3H, P3H)
add P3H, P3H, P3H
add P3H, P3H, $r15
bnez $r6, .Li33
or $r10, P3H, P3L
beqz $r10, .Li33
#NORMd($r4, P2H, P2L)
bnez P3H, .LL65
bnez P3L, .LL66
move $r6, #0
j .LL67
.LL66:
move P3H, P3L
move P3L, #0
move P2H, #32
sub $r6, $r6, P2H
.LL65:
#ifndef __big_endian__
#ifdef __NDS32_PERF_EXT__
clz $r3, $r5
#else
pushm $r0, $r2
pushm $r4, $r5
move $r0, $r5
bal __clzsi2
move $r3, $r0
popm $r4, $r5
popm $r0, $r2
#endif
#else /* __big_endian__ */
#ifdef __NDS32_PERF_EXT__
clz $r2, $r4
#else
pushm $r0, $r1
pushm $r3, $r5
move $r0, $r4
bal __clzsi2
move $r2, $r0
popm $r3, $r5
popm $r0, $r1
#endif
#endif /* __big_endian_ */
beqz P2H, .LL67
sub $r6, $r6, P2H
subri P2L, P2H, #32
srl P2L, P3L, P2L
sll P3L, P3L, P2H
sll P3H, P3H, P2H
or P3H, P3H, P2L
.LL67:
#NORMd End
j .LGlab1
.Li33:
bne $r6, $r9, .Li35
slli $r10, O1H, #1
or $r10, $r10, O1L
beqz $r10, .LGnan
.Li35:
subri $r15, $r9, #0x7ff
beqzs8 .LGspecB
beqz $r6, .LGret
or $r10, P3H, P3L
bnez $r10, .LGnan
.LGinf:
move $r10, #0x7ff00000
or P1H, P1H, $r10
move P1L, #0
j .LGret
.LGspecB:
#ADD(O1L, O1L)
move $r15, O1L
add O1L, O1L, O1L
slt $r15, O1L, $r15
#ADDC(O1H, O1H)
add O1H, O1H, O1H
add O1H, O1H, $r15
bnez $r9, .Li36
or $r10, O1H, O1L
beqz $r10, .LGinf
#NORMd($r7, P2H, P2L)
bnez O1H, .LL68
bnez O1L, .LL69
move $r9, #0
j .LL70
.LL69:
move O1H, O1L
move O1L, #0
move P2H, #32
sub $r9, $r9, P2H
.LL68:
#ifndef __big_endian__
#ifdef __NDS32_PERF_EXT__
clz $r3, $r8
#else
pushm $r0, $r2
pushm $r4, $r5
move $r0, $r8
bal __clzsi2
move $r3, $r0
popm $r4, $r5
popm $r0, $r2
#endif
#else /* __big_endian__ */
#ifdef __NDS32_PERF_EXT__
clz $r2, $r7
#else
pushm $r0, $r1
pushm $r3, $r5
move $r0, $r7
bal __clzsi2
move $r2, $r0
popm $r3, $r5
popm $r0, $r1
#endif
#endif /* __big_endian__ */
beqz P2H, .LL70
sub $r9, $r9, P2H
subri P2L, P2H, #32
srl P2L, O1L, P2L
sll O1L, O1L, P2H
sll O1H, O1H, P2H
or O1H, O1H, P2L
.LL70:
#NORMd End
j .LGlab2
.Li36:
or $r10, O1H, O1L
beqz $r10, .Li38
.LGnan:
move P1H, #0xfff80000
.Li38:
move P1L, #0
j .LGret
.size __divdf3, .-__divdf3
#endif /* L_div_df */
#ifdef L_negate_sf
.text
.align 2
.global __negsf2
.type __negsf2, @function
__negsf2:
push $lp
move $r1, #0x80000000
xor $r0, $r0, $r1
.LN999:
pop $lp
ret5 $lp
.size __negsf2, .-__negsf2
#endif /* L_negate_sf */
#ifdef L_negate_df
#ifndef __big_endian__
#define P1H $r1
#else
#define P1H $r0
#endif
.text
.align 2
.global __negdf2
.type __negdf2, @function
__negdf2:
push $lp
move $r2, #0x80000000
xor P1H, P1H, $r2
.LP999:
pop $lp
ret5 $lp
.size __negdf2, .-__negdf2
#endif /* L_negate_df */
#ifdef L_sf_to_df
#ifndef __big_endian__
#define O1L $r1
#define O1H $r2
#else
#define O1H $r1
#define O1L $r2
#endif
.text
.align 2
.global __extendsfdf2
.type __extendsfdf2, @function
__extendsfdf2:
push $lp
srli $r3, $r0, #23
andi $r3, $r3, #0xff
move $r5, #0x80000000
and O1H, $r0, $r5
addi $r5, $r3, #-1
slti $r15, $r5, #0xfe
beqzs8 .LJspec
.LJlab1:
addi $r3, $r3, #0x380
slli $r5, $r0, #9
srli $r5, $r5, #12
or O1H, O1H, $r5
slli O1L, $r0, #29
.LJret:
slli $r5, $r3, #20
or O1H, O1H, $r5
move $r0, $r1
move $r1, $r2
.LJ999:
pop $lp
ret5 $lp
.LJspec:
move O1L, #0
add $r0, $r0, $r0
beqz $r0, .LJret
bnez $r3, .Li42
.Lb43:
addi $r3, $r3, #-1
add $r0, $r0, $r0
move $r5, #0x800000
slt $r15, $r0, $r5
bnezs8 .Lb43
j .LJlab1
.Li42:
move $r3, #0x7ff
move $r5, #0xff000000
slt $r15, $r5, $r0
beqzs8 .LJret
move O1H, #0xfff80000
j .LJret
.size __extendsfdf2, .-__extendsfdf2
#endif /* L_sf_to_df */
#ifdef L_df_to_sf
#ifndef __big_endian__
#define P1L $r0
#define P1H $r1
#define P2L $r2
#define P2H $r3
#else
#define P1H $r0
#define P1L $r1
#define P2H $r2
#define P2L $r3
#endif
.text
.align 2
.global __truncdfsf2
.type __truncdfsf2, @function
__truncdfsf2:
push $lp
pushm $r6, $r8
slli P2H, P1H, #11
srli $r7, P1L, #21
or P2H, P2H, $r7
slli P2L, P1L, #11
move $r7, #0x80000000
or P2H, P2H, $r7
and $r5, P1H, $r7
slli $r4, P1H, #1
srli $r4, $r4, #21
addi $r4, $r4, #0xfffffc80
addi $r7, $r4, #-1
slti $r15, $r7, #0xfe
beqzs8 .LKspec
.LKlab1:
beqz P2L, .Li45
ori P2H, P2H, #1
.Li45:
#ADD(P2H, $0x80)
move $r15, #0x80
add P2H, P2H, $r15
slt $r15, P2H, $r15
#ADDC($r4, $0x0)
add $r4, $r4, $r15
srli $r7, P2H, #8
andi $r7, $r7, #1
sub P2H, P2H, $r7
slli P2H, P2H, #1
srli P2H, P2H, #9
slli $r7, $r4, #23
or P2H, P2H, $r7
or $r0, P2H, $r5
.LK999:
popm $r6, $r8
pop $lp
ret5 $lp
.LKspec:
subri $r15, $r4, #0x47f
bnezs8 .Li46
slli $r7, P2H, #1
or $r7, $r7, P2L
beqz $r7, .Li46
move $r0, #0xffc00000
j .LK999
.Li46:
sltsi $r15, $r4, #0xff
bnezs8 .Li48
move $r7, #0x7f800000
or $r0, $r5, $r7
j .LK999
.Li48:
subri $r6, $r4, #1
move $r7, #0x20
slt $r15, $r6, $r7
bnezs8 .Li49
move $r0, $r5
j .LK999
.Li49:
subri $r8, $r6, #0x20
sll $r7, P2H, $r8
or P2L, P2L, $r7
srl P2H, P2H, $r6
move $r4, #0
move $r7, #0x80000000
or P2H, P2H, $r7
j .LKlab1
.size __truncdfsf2, .-__truncdfsf2
#endif /* L_df_to_sf */
#ifdef L_df_to_si
#ifndef __big_endian__
#define P1L $r0
#define P1H $r1
#else
#define P1H $r0
#define P1L $r1
#endif
.global __fixdfsi
.type __fixdfsi, @function
__fixdfsi:
push $lp
pushm $r6, $r6
slli $r3, P1H, #11
srli $r6, P1L, #21
or $r3, $r3, $r6
move $r6, #0x80000000
or $r3, $r3, $r6
slli $r6, P1H, #1
srli $r6, $r6, #21
subri $r2, $r6, #0x41e
blez $r2, .LLnaninf
move $r6, #0x20
slt $r15, $r2, $r6
bnezs8 .LL72
move $r3, #0
.LL72:
srl $r3, $r3, $r2
sltsi $r15, P1H, #0
beqzs8 .Li50
subri $r3, $r3, #0
.Li50:
move $r0, $r3
.LL999:
popm $r6, $r6
pop $lp
ret5 $lp
.LLnaninf:
beqz P1L, .Li51
ori P1H, P1H, #1
.Li51:
move $r6, #0x7ff00000
slt $r15, $r6, P1H
beqzs8 .Li52
move $r0, #0x80000000
j .LL999
.Li52:
move $r0, #0x7fffffff
j .LL999
.size __fixdfsi, .-__fixdfsi
#endif /* L_df_to_si */
#ifdef L_fixsfdi
#ifndef __big_endian__
#define O1L $r1
#define O1H $r2
#else
#define O1H $r1
#define O1L $r2
#endif
.text
.align 2
.global __fixsfdi
.type __fixsfdi, @function
__fixsfdi:
push $lp
srli $r3, $r0, #23
andi $r3, $r3, #0xff
slli O1H, $r0, #8
move $r5, #0x80000000
or O1H, O1H, $r5
move O1L, #0
sltsi $r15, $r3, #0xbe
beqzs8 .LCinfnan
subri $r3, $r3, #0xbe
.LL8:
move $r5, #0x20
slt $r15, $r3, $r5
bnezs8 .LL9
move O1L, O1H
move O1H, #0
addi $r3, $r3, #0xffffffe0
bnez O1L, .LL8
.LL9:
beqz $r3, .LL10
move $r4, O1H
srl O1L, O1L, $r3
srl O1H, O1H, $r3
subri $r3, $r3, #0x20
sll $r4, $r4, $r3
or O1L, O1L, $r4
.LL10:
sltsi $r15, $r0, #0
beqzs8 .LCret
subri O1H, O1H, #0
beqz O1L, .LL11
subri O1L, O1L, #0
subi45 O1H, #1
.LL11:
.LCret:
move $r0, $r1
move $r1, $r2
.LC999:
pop $lp
ret5 $lp
.LCinfnan:
sltsi $r15, $r0, #0
bnezs8 .LCret3
subri $r15, $r3, #0xff
bnezs8 .Li7
slli $r5, O1H, #1
beqz $r5, .Li7
.LCret3:
move O1H, #0x80000000
j .LCret
.Li7:
move O1H, #0x7fffffff
move O1L, #-1
j .LCret
.size __fixsfdi, .-__fixsfdi
#endif /* L_fixsfdi */
#ifdef L_fixdfdi
#ifndef __big_endian__
#define P1L $r0
#define P1H $r1
#define O1L $r3
#define O1H $r4
#else
#define P1H $r0
#define P1L $r1
#define O1H $r3
#define O1L $r4
#endif
.text
.align 2
.global __fixdfdi
.type __fixdfdi, @function
__fixdfdi:
push $lp
pushm $r6, $r6
slli $r5, P1H, #1
srli $r5, $r5, #21
slli O1H, P1H, #11
srli $r6, P1L, #21
or O1H, O1H, $r6
slli O1L, P1L, #11
move $r6, #0x80000000
or O1H, O1H, $r6
slti $r15, $r5, #0x43e
beqzs8 .LCnaninf
subri $r2, $r5, #0x43e
.LL14:
move $r6, #0x20
slt $r15, $r2, $r6
bnezs8 .LL15
move O1L, O1H
move O1H, #0
addi $r2, $r2, #0xffffffe0
bnez O1L, .LL14
.LL15:
beqz $r2, .LL16
move P1L, O1H
srl O1L, O1L, $r2
srl O1H, O1H, $r2
subri $r2, $r2, #0x20
sll P1L, P1L, $r2
or O1L, O1L, P1L
.LL16:
sltsi $r15, P1H, #0
beqzs8 .LCret
subri O1H, O1H, #0
beqz O1L, .LL17
subri O1L, O1L, #0
subi45 O1H, #1
.LL17:
.LCret:
move P1L, O1L
move P1H, O1H
.LC999:
popm $r6, $r6
pop $lp
ret5 $lp
.LCnaninf:
sltsi $r15, P1H, #0
bnezs8 .LCret3
subri $r15, $r5, #0x7ff
bnezs8 .Li5
slli $r6, O1H, #1
or $r6, $r6, O1L
beqz $r6, .Li5
.LCret3:
move O1H, #0x80000000
move O1L, #0
j .LCret
.Li5:
move O1H, #0x7fffffff
move O1L, #-1
j .LCret
.size __fixdfdi, .-__fixdfdi
#endif /* L_fixdfdi */
#ifdef L_fixunssfsi
.global __fixunssfsi
.type __fixunssfsi, @function
__fixunssfsi:
push $lp
slli $r1, $r0, #8
move $r3, #0x80000000
or $r1, $r1, $r3
srli $r3, $r0, #23
andi $r3, $r3, #0xff
subri $r2, $r3, #0x9e
sltsi $r15, $r2, #0
bnezs8 .LLspec
sltsi $r15, $r2, #0x20
bnezs8 .Li45
move $r0, #0
j .LL999
.Li45:
srl $r1, $r1, $r2
sltsi $r15, $r0, #0
beqzs8 .Li46
subri $r1, $r1, #0
.Li46:
move $r0, $r1
.LL999:
pop $lp
ret5 $lp
.LLspec:
move $r3, #0x7f800000
slt $r15, $r3, $r0
beqzs8 .Li47
move $r0, #0x80000000
j .LL999
.Li47:
move $r0, #-1
j .LL999
.size __fixunssfsi, .-__fixunssfsi
#endif /* L_fixunssfsi */
#ifdef L_fixunsdfsi
#ifndef __big_endian__
#define P1L $r0
#define P1H $r1
#else
#define P1H $r0
#define P1L $r1
#endif
.text
.align 2
.global __fixunsdfsi
.type __fixunsdfsi, @function
__fixunsdfsi:
push $lp
pushm $r6, $r6
slli $r3, P1H, #11
srli $r6, P1L, #21
or $r3, $r3, $r6
move $r6, #0x80000000
or $r3, $r3, $r6
slli $r6, P1H, #1
srli $r6, $r6, #21
subri $r2, $r6, #0x41e
sltsi $r15, $r2, #0
bnezs8 .LNnaninf
move $r6, #0x20
slt $r15, $r2, $r6
bnezs8 .LL73
move $r3, #0
.LL73:
srl $r3, $r3, $r2
sltsi $r15, P1H, #0
beqzs8 .Li53
subri $r3, $r3, #0
.Li53:
move $r0, $r3
.LN999:
popm $r6, $r6
pop $lp
ret5 $lp
.LNnaninf:
beqz P1L, .Li54
ori P1H, P1H, #1
.Li54:
move $r6, #0x7ff00000
slt $r15, $r6, P1H
beqzs8 .Li55
move $r0, #0x80000000
j .LN999
.Li55:
move $r0, #-1
j .LN999
.size __fixunsdfsi, .-__fixunsdfsi
#endif /* L_fixunsdfsi */
#ifdef L_fixunssfdi
#ifndef __big_endian__
#define O1L $r1
#define O1H $r2
#else
#define O1H $r1
#define O1L $r2
#endif
.text
.align 2
.global __fixunssfdi
.type __fixunssfdi, @function
__fixunssfdi:
push $lp
srli $r3, $r0, #23
andi $r3, $r3, #0xff
slli O1H, $r0, #8
move $r5, #0x80000000
or O1H, O1H, $r5
move O1L, #0
sltsi $r15, $r3, #0xbe
beqzs8 .LDinfnan
subri $r3, $r3, #0xbe
.LL12:
move $r5, #0x20
slt $r15, $r3, $r5
bnezs8 .LL13
move O1L, O1H
move O1H, #0
addi $r3, $r3, #0xffffffe0
bnez O1L, .LL12
.LL13:
beqz $r3, .LL14
move $r4, O1H
srl O1L, O1L, $r3
srl O1H, O1H, $r3
subri $r3, $r3, #0x20
sll $r4, $r4, $r3
or O1L, O1L, $r4
.LL14:
sltsi $r15, $r0, #0
beqzs8 .LDret
subri O1H, O1H, #0
beqz O1L, .LL15
subri O1L, O1L, #0
subi45 O1H, #1
.LL15:
.LDret:
move $r0, $r1
move $r1, $r2
.LD999:
pop $lp
ret5 $lp
.LDinfnan:
move O1H, #0x80000000
move O1L, #0
j .LDret
.size __fixunssfdi, .-__fixunssfdi
#endif /* L_fixunssfdi */
#ifdef L_fixunsdfdi
#ifndef __big_endian__
#define P1L $r0
#define P1H $r1
#define O1L $r3
#define O1H $r4
#else
#define P1H $r0
#define P1L $r1
#define O1H $r3
#define O1L $r4
#endif
.text
.align 2
.global __fixunsdfdi
.type __fixunsdfdi, @function
__fixunsdfdi:
push $lp
pushm $r6, $r6
slli $r5, P1H, #1
srli $r5, $r5, #21
slli O1H, P1H, #11
srli $r6, P1L, #21
or O1H, O1H, $r6
slli O1L, P1L, #11
move $r6, #0x80000000
or O1H, O1H, $r6
slti $r15, $r5, #0x43e
beqzs8 .LDnaninf
subri $r2, $r5, #0x43e
.LL18:
move $r6, #0x20
slt $r15, $r2, $r6
bnezs8 .LL19
move O1L, O1H
move O1H, #0
addi $r2, $r2, #0xffffffe0
bnez O1L, .LL18
.LL19:
beqz $r2, .LL20
move P1L, O1H
srl O1L, O1L, $r2
srl O1H, O1H, $r2
subri $r2, $r2, #0x20
sll P1L, P1L, $r2
or O1L, O1L, P1L
.LL20:
sltsi $r15, P1H, #0
beqzs8 .LDret
subri O1H, O1H, #0
beqz O1L, .LL21
subri O1L, O1L, #0
subi45 O1H, #1
.LL21:
.LDret:
move P1L, O1L
move P1H, O1H
.LD999:
popm $r6, $r6
pop $lp
ret5 $lp
.LDnaninf:
move O1H, #0x80000000
move O1L, #0
j .LDret
.size __fixunsdfdi, .-__fixunsdfdi
#endif /* L_fixunsdfdi */
#ifdef L_si_to_sf
.text
.align 2
.global __floatsisf
.type __floatsisf, @function
__floatsisf:
push $lp
move $r4, #0x80000000
and $r2, $r0, $r4
beqz $r0, .Li39
sltsi $r15, $r0, #0
beqzs8 .Li40
subri $r0, $r0, #0
.Li40:
move $r1, #0x9e
#ifdef __NDS32_PERF_EXT__
clz $r3, $r0
#else
pushm $r0, $r2
pushm $r4, $r5
bal __clzsi2
move $r3, $r0
popm $r4, $r5
popm $r0, $r2
#endif
sub $r1, $r1, $r3
sll $r0, $r0, $r3
#ADD($r0, $0x80)
move $r15, #0x80
add $r0, $r0, $r15
slt $r15, $r0, $r15
#ADDC($r1, $0x0)
add $r1, $r1, $r15
srai $r4, $r0, #8
andi $r4, $r4, #1
sub $r0, $r0, $r4
slli $r0, $r0, #1
srli $r0, $r0, #9
slli $r4, $r1, #23
or $r0, $r0, $r4
.Li39:
or $r0, $r0, $r2
.LH999:
pop $lp
ret5 $lp
.size __floatsisf, .-__floatsisf
#endif /* L_si_to_sf */
#ifdef L_si_to_df
#ifndef __big_endian__
#define O1L $r1
#define O1H $r2
#define O2L $r4
#define O2H $r5
#else
#define O1H $r1
#define O1L $r2
#define O2H $r4
#define O2L $r5
#endif
.text
.align 2
.global __floatsidf
.type __floatsidf, @function
__floatsidf:
push $lp
pushm $r6, $r6
move O1L, #0
move O2H, O1L
move $r3, O1L
move O1H, $r0
beqz O1H, .Li39
sltsi $r15, O1H, #0
beqzs8 .Li40
move O2H, #0x80000000
subri O1H, O1H, #0
beqz O1L, .LL71
subri O1L, O1L, #0
subi45 O1H, #1
.LL71:
.Li40:
move $r3, #0x41e
#ifndef __big_endian__
#ifdef __NDS32_PERF_EXT__
clz $r4, $r2
#else
pushm $r0, $r3
push $r5
move $r0, $r2
bal __clzsi2
move $r4, $r0
pop $r5
popm $r0, $r3
#endif
#else /* __big_endian__ */
#ifdef __NDS32_PERF_EXT__
clz $r5, $r1
#else
pushm $r0, $r4
move $r0, $r1
bal __clzsi2
move $r5, $r0
popm $r0, $r4
#endif
#endif /* __big_endian__ */
sub $r3, $r3, O2L
sll O1H, O1H, O2L
.Li39:
srli O2L, O1L, #11
slli $r6, O1H, #21
or O2L, O2L, $r6
slli $r6, O1H, #1
srli $r6, $r6, #12
or O2H, O2H, $r6
slli $r6, $r3, #20
or O2H, O2H, $r6
move $r0, $r4
move $r1, $r5
.LH999:
popm $r6, $r6
pop $lp
ret5 $lp
.size __floatsidf, .-__floatsidf
#endif /* L_si_to_df */
#ifdef L_floatdisf
#ifndef __big_endian__
#define P1L $r0
#define P1H $r1
#define P2L $r2
#define P2H $r3
#else
#define P1H $r0
#define P1L $r1
#define P2H $r2
#define P2L $r3
#endif
.text
.align 2
.global __floatdisf
.type __floatdisf, @function
__floatdisf:
push $lp
pushm $r6, $r7
move $r7, #0x80000000
and $r5, P1H, $r7
move P2H, P1H
move P2L, P1L
or $r7, P1H, P1L
beqz $r7, .Li1
sltsi $r15, P1H, #0
beqzs8 .Li2
subri P2H, P2H, #0
beqz P2L, .LL1
subri P2L, P2L, #0
subi45 P2H, #1
.LL1:
.Li2:
move $r4, #0xbe
#NORMd($r2, $r6, P1L)
bnez P2H, .LL2
bnez P2L, .LL3
move $r4, #0
j .LL4
.LL3:
move P2H, P2L
move P2L, #0
move $r6, #32
sub $r4, $r4, $r6
.LL2:
#ifdef __NDS32_PERF_EXT__
clz $r6, P2H
#else
pushm $r0, $r5
move $r0, P2H
bal __clzsi2
move $r6, $r0
popm $r0, $r5
#endif
beqz $r6, .LL4
sub $r4, $r4, $r6
subri P1L, $r6, #32
srl P1L, P2L, P1L
sll P2L, P2L, $r6
sll P2H, P2H, $r6
or P2H, P2H, P1L
.LL4:
#NORMd End
beqz P2L, .Li3
ori P2H, P2H, #1
.Li3:
#ADD(P2H, $0x80)
move $r15, #0x80
add P2H, P2H, $r15
slt $r15, P2H, $r15
#ADDC($r4, $0x0)
add $r4, $r4, $r15
srli $r7, P2H, #8
andi $r7, $r7, #1
sub P2H, P2H, $r7
slli P2H, P2H, #1
srli P2H, P2H, #9
slli $r7, $r4, #23
or P2H, P2H, $r7
.Li1:
or $r0, P2H, $r5
.LA999:
popm $r6, $r7
pop $lp
ret5 $lp
.size __floatdisf, .-__floatdisf
#endif /* L_floatdisf */
#ifdef L_floatdidf
#ifndef __big_endian__
#define P1L $r0
#define P1H $r1
#define P2L $r2
#define P2H $r3
#define O1L $r5
#define O1H $r6
#else
#define P1H $r0
#define P1L $r1
#define P2H $r2
#define P2L $r3
#define O1H $r5
#define O1L $r6
#endif
.text
.align 2
.global __floatdidf
.type __floatdidf, @function
__floatdidf:
push $lp
pushm $r6, $r8
move $r4, #0
move $r7, $r4
move P2H, P1H
move P2L, P1L
or $r8, P1H, P1L
beqz $r8, .Li1
move $r4, #0x43e
sltsi $r15, P1H, #0
beqzs8 .Li2
move $r7, #0x80000000
subri P2H, P2H, #0
beqz P2L, .LL1
subri P2L, P2L, #0
subi45 P2H, #1
.LL1:
.Li2:
#NORMd($r2, O1H, O1L)
bnez P2H, .LL2
bnez P2L, .LL3
move $r4, #0
j .LL4
.LL3:
move P2H, P2L
move P2L, #0
move O1H, #32
sub $r4, $r4, O1H
.LL2:
#ifdef __NDS32_PERF_EXT__
clz O1H, P2H
#else /* not __NDS32_PERF_EXT__ */
/*
Replace clz with function call.
clz O1H, P2H
EL: clz $r6, $r3
EB: clz $r5, $r2
*/
#ifndef __big_endian__
pushm $r0, $r5
move $r0, $r3
bal __clzsi2
move $r6, $r0
popm $r0, $r5
#else
pushm $r0, $r4
move $r0, $r2
bal __clzsi2
move $r5, $r0
popm $r0, $r4
#endif
#endif /* not __NDS32_PERF_EXT__ */
beqz O1H, .LL4
sub $r4, $r4, O1H
subri O1L, O1H, #32
srl O1L, P2L, O1L
sll P2L, P2L, O1H
sll P2H, P2H, O1H
or P2H, P2H, O1L
.LL4:
#NORMd End
#ADD(P2L, $0x400)
move $r15, #0x400
add P2L, P2L, $r15
slt $r15, P2L, $r15
#ADDCC(P2H, $0x0)
beqzs8 .LL7
add P2H, P2H, $r15
slt $r15, P2H, $r15
.LL7:
#ADDC($r4, $0x0)
add $r4, $r4, $r15
srli $r8, P2L, #11
andi $r8, $r8, #1
sub P2L, P2L, $r8
.Li1:
srli O1L, P2L, #11
slli $r8, P2H, #21
or O1L, O1L, $r8
slli O1H, P2H, #1
srli O1H, O1H, #12
slli $r8, $r4, #20
or O1H, O1H, $r8
or O1H, O1H, $r7
move P1L, O1L
move P1H, O1H
.LA999:
popm $r6, $r8
pop $lp
ret5 $lp
.size __floatdidf, .-__floatdidf
#endif /* L_floatdidf */
#ifdef L_floatunsisf
.text
.align 2
.global __floatunsisf
.type __floatunsisf, @function
__floatunsisf:
push $lp
beqz $r0, .Li41
move $r2, #0x9e
#ifdef __NDS32_PERF_EXT__
clz $r1, $r0
#else
push $r0
pushm $r2, $r5
bal __clzsi2
move $r1, $r0
popm $r2, $r5
pop $r0
#endif
sub $r2, $r2, $r1
sll $r0, $r0, $r1
#ADD($r0, $0x80)
move $r15, #0x80
add $r0, $r0, $r15
slt $r15, $r0, $r15
#ADDC($r2, $0x0)
add $r2, $r2, $r15
srli $r3, $r0, #8
andi $r3, $r3, #1
sub $r0, $r0, $r3
slli $r0, $r0, #1
srli $r0, $r0, #9
slli $r3, $r2, #23
or $r0, $r0, $r3
.Li41:
.LI999:
pop $lp
ret5 $lp
.size __floatunsisf, .-__floatunsisf
#endif /* L_floatunsisf */
#ifdef L_floatunsidf
#ifndef __big_endian__
#define O1L $r1
#define O1H $r2
#define O2L $r4
#define O2H $r5
#else
#define O1H $r1
#define O1L $r2
#define O2H $r4
#define O2L $r5
#endif
.text
.align 2
.global __floatunsidf
.type __floatunsidf, @function
__floatunsidf:
push $lp
pushm $r6, $r6
move O1L, #0
move $r3, O1L
move O1H, $r0
beqz O1H, .Li41
move $r3, #0x41e
#ifndef __big_endian__
#ifdef __NDS32_PERF_EXT__
clz $r5, $r2
#else
pushm $r0, $r4
move $r0, $r2
bal __clzsi2
move $r5, $r0
popm $r0, $r4
#endif
#else /* __big_endian__ */
#ifdef __NDS32_PERF_EXT__
clz $r4, $r1
#else
pushm $r0, $r3
push $r5
move $r0, $r1
bal __clzsi2
move $r4, $r0
pop $r5
popm $r0, $r3
#endif
#endif /* __big_endian__ */
sub $r3, $r3, O2H
sll O1H, O1H, O2H
.Li41:
srli O2L, O1L, #11
slli $r6, O1H, #21
or O2L, O2L, $r6
slli O2H, O1H, #1
srli O2H, O2H, #12
slli $r6, $r3, #20
or O2H, O2H, $r6
move $r0, $r4
move $r1, $r5
.LI999:
popm $r6, $r6
pop $lp
ret5 $lp
.size __floatunsidf, .-__floatunsidf
#endif /* L_floatunsidf */
#ifdef L_floatundisf
#ifndef __big_endian__
#define P1L $r0
#define P1H $r1
#define P2L $r2
#define P2H $r3
#else
#define P1H $r0
#define P1L $r1
#define P2H $r2
#define P2L $r3
#endif
.text
.align 2
.global __floatundisf
.type __floatundisf, @function
__floatundisf:
push $lp
pushm $r6, $r6
move P2H, P1H
move P2L, P1L
or $r6, P1H, P1L
beqz $r6, .Li4
move $r4, #0xbe
#NORMd($r2, $r5, P1L)
bnez P2H, .LL5
bnez P2L, .LL6
move $r4, #0
j .LL7
.LL6:
move P2H, P2L
move P2L, #0
move $r5, #32
sub $r4, $r4, $r5
.LL5:
#ifdef __NDS32_PERF_EXT__
clz $r5, P2H
#else
pushm $r0, $r4
move $r0, P2H
bal __clzsi2
move $r5, $r0
popm $r0, $r4
#endif
beqz $r5, .LL7
sub $r4, $r4, $r5
subri P1L, $r5, #32
srl P1L, P2L, P1L
sll P2L, P2L, $r5
sll P2H, P2H, $r5
or P2H, P2H, P1L
.LL7:
#NORMd End
beqz P2L, .Li5
ori P2H, P2H, #1
.Li5:
#ADD(P2H, $0x80)
move $r15, #0x80
add P2H, P2H, $r15
slt $r15, P2H, $r15
#ADDC($r4, $0x0)
add $r4, $r4, $r15
srli $r6, P2H, #8
andi $r6, $r6, #1
sub P2H, P2H, $r6
slli P2H, P2H, #1
srli P2H, P2H, #9
slli $r6, $r4, #23
or P2H, P2H, $r6
.Li4:
move $r0, P2H
.LB999:
popm $r6, $r6
pop $lp
ret5 $lp
.size __floatundisf, .-__floatundisf
#endif /* L_floatundisf */
#ifdef L_floatundidf
#ifndef __big_endian__
#define P1L $r0
#define P1H $r1
#define P2L $r2
#define P2H $r3
#define O1L $r5
#define O1H $r6
#else
#define P1H $r0
#define P1L $r1
#define P2H $r2
#define P2L $r3
#define O1H $r5
#define O1L $r6
#endif
.text
.align 2
.global __floatundidf
.type __floatundidf, @function
__floatundidf:
push $lp
pushm $r6, $r7
move $r4, #0
move P2H, P1H
move P2L, P1L
or $r7, P1H, P1L
beqz $r7, .Li3
move $r4, #0x43e
#NORMd($r2, O1H, O1L)
bnez P2H, .LL8
bnez P2L, .LL9
move $r4, #0
j .LL10
.LL9:
move P2H, P2L
move P2L, #0
move O1H, #32
sub $r4, $r4, O1H
.LL8:
#ifdef __NDS32_PERF_EXT__
clz O1H, P2H
#else /* not __NDS32_PERF_EXT__ */
/*
Replace clz with function call.
clz O1H, P2H
EL: clz $r6, $r3
EB: clz $r5, $r2
*/
#ifndef __big_endian__
pushm $r0, $r5
move $r0, $r3
bal __clzsi2
move $r6, $r0
popm $r0, $r5
#else
pushm $r0, $r4
move $r0, $r2
bal __clzsi2
move $r5, $r0
popm $r0, $r4
#endif
#endif /* not __NDS32_PERF_EXT__ */
beqz O1H, .LL10
sub $r4, $r4, O1H
subri O1L, O1H, #32
srl O1L, P2L, O1L
sll P2L, P2L, O1H
sll P2H, P2H, O1H
or P2H, P2H, O1L
.LL10:
#NORMd End
#ADD(P2L, $0x400)
move $r15, #0x400
add P2L, P2L, $r15
slt $r15, P2L, $r15
#ADDCC(P2H, $0x0)
beqzs8 .LL13
add P2H, P2H, $r15
slt $r15, P2H, $r15
.LL13:
#ADDC($r4, $0x0)
add $r4, $r4, $r15
srli $r7, P2L, #11
andi $r7, $r7, #1
sub P2L, P2L, $r7
.Li3:
srli O1L, P2L, #11
slli $r7, P2H, #21
or O1L, O1L, $r7
slli O1H, P2H, #1
srli O1H, O1H, #12
slli $r7, $r4, #20
or O1H, O1H, $r7
move P1L, O1L
move P1H, O1H
.LB999:
popm $r6, $r7
pop $lp
ret5 $lp
.size __floatundidf, .-__floatundidf
#endif /* L_floatundidf */
#ifdef L_compare_sf
.text
.align 2
.global __cmpsf2
.type __cmpsf2, @function
__cmpsf2:
.global __eqsf2
.type __eqsf2, @function
__eqsf2:
.global __ltsf2
.type __ltsf2, @function
__ltsf2:
.global __lesf2
.type __lesf2, @function
__lesf2:
.global __nesf2
.type __nesf2, @function
__nesf2:
move $r4, #1
j .LA
.global __gesf2
.type __gesf2, @function
__gesf2:
.global __gtsf2
.type __gtsf2, @function
__gtsf2:
move $r4, #-1
.LA:
push $lp
slli $r2, $r0, #1
slli $r3, $r1, #1
or $r5, $r2, $r3
beqz $r5, .LMequ
move $r5, #0xff000000
slt $r15, $r5, $r2
bnezs8 .LMnan
slt $r15, $r5, $r3
bnezs8 .LMnan
srli $r2, $r2, #1
sltsi $r15, $r0, #0
beqzs8 .Li48
subri $r2, $r2, #0
.Li48:
srli $r3, $r3, #1
sltsi $r15, $r1, #0
beqzs8 .Li49
subri $r3, $r3, #0
.Li49:
slts $r15, $r2, $r3
beqzs8 .Li50
move $r0, #-1
j .LM999
.Li50:
slts $r15, $r3, $r2
beqzs8 .LMequ
move $r0, #1
j .LM999
.LMequ:
move $r0, #0
.LM999:
pop $lp
ret5 $lp
.LMnan:
move $r0, $r4
j .LM999
.size __cmpsf2, .-__cmpsf2
.size __eqsf2, .-__eqsf2
.size __ltsf2, .-__ltsf2
.size __lesf2, .-__lesf2
.size __nesf2, .-__nesf2
.size __gesf2, .-__gesf2
.size __gtsf2, .-__gtsf2
#endif /* L_compare_sf */
#ifdef L_compare_df
#ifdef __big_endian__
#define P1H $r0
#define P1L $r1
#define P2H $r2
#define P2L $r3
#else
#define P1H $r1
#define P1L $r0
#define P2H $r3
#define P2L $r2
#endif
.align 2
.globl __gtdf2
.globl __gedf2
.globl __ltdf2
.globl __ledf2
.globl __eqdf2
.globl __nedf2
.globl __cmpdf2
.type __gtdf2, @function
.type __gedf2, @function
.type __ltdf2, @function
.type __ledf2, @function
.type __eqdf2, @function
.type __nedf2, @function
.type __cmpdf2, @function
__gtdf2:
__gedf2:
movi $r4, -1
b .L1
__ltdf2:
__ledf2:
__cmpdf2:
__nedf2:
__eqdf2:
movi $r4, 1
.L1:
#if defined (__NDS32_ISA_V3M__)
push25 $r10, 0
#else
smw.adm $r6, [$sp], $r9, 0
#endif
sethi $r5, 0x7ff00
and $r6, P1H, $r5 ! r6=aExp
and $r7, P2H, $r5 ! r7=bExp
slli $r8, P1H, 12 ! r8=aSig0
slli $r9, P2H, 12 ! r9=bSig0
beq $r6, $r5, .L11 ! aExp==0x7ff
beq $r7, $r5, .L12 ! bExp==0x7ff
.L2:
slli $ta, P1H, 1 ! ta=ahigh<<1
or $ta, P1L, $ta !
xor $r5, P1H, P2H ! r5=ahigh^bhigh
beqz $ta, .L3 ! if(ahigh<<1)==0,go .L3
!-------------------------------
! (ahigh<<1)!=0 || (bhigh<<1)!=0
!-------------------------------
.L4:
beqz $r5, .L5 ! ahigh==bhigh, go .L5
!--------------------
! a != b
!--------------------
.L6:
bltz $r5, .L7 ! if(aSign!=bSign), go .L7
!--------------------
! aSign==bSign
!--------------------
slt $ta, $r6, $r7 ! ta=(aExp|b|), go .L10
nor $r0, P2H, P2H ! if(|a|<|b|),return (~yh)
.L14:
#if defined (__NDS32_ISA_V3M__)
pop25 $r10, 0
#else
lmw.bim $r6, [$sp], $r9, 0
ret
#endif
.L10:
ori $r0, P2H, 1 ! return (yh|1)
b .L14
!--------------------
! (ahigh<<1)=0
!--------------------
.L3:
slli $ta, P2H, 1 ! ta=bhigh<<1
or $ta, P2L, $ta !
bnez $ta, .L4 ! ta=(bhigh<<1)!=0,go .L4
.L5:
xor $ta, P1L, P2L ! ta=alow^blow
bnez $ta, .L6 ! alow!=blow,go .L6
movi $r0, 0 ! a==b, return 0
b .L14
!--------------------
! aExp=0x7ff;
!--------------------
.L11:
or P1L, P1L, $r8 ! x1=(aSig0|aSig1)
bnez P1L, .L13 ! if(a=nan), go.L13
xor $ta, $r7, $r5 ! ta=(bExp^0x7ff)
bnez $ta, .L2 ! if(bExp!=0x7ff), go .L2
!--------------------
! bExp=0x7ff;
!--------------------
.L12:
or $ta, P2L, $r9 ! ta=(bSig0|bSig1)
beqz $ta, .L2 ! if(b!=nan), go .L2
.L13:
move $r0, $r4
b .L14
!--------------------
! aSign!=bSign
!--------------------
.L7:
ori $r0, P1H, 1 ! if(aSign!=bSign), return (ahigh|1)
b .L14
.size __gtdf2, .-__gtdf2
.size __gedf2, .-__gedf2
.size __ltdf2, .-__ltdf2
.size __ledf2, .-__ledf2
.size __eqdf2, .-__eqdf2
.size __nedf2, .-__nedf2
.size __cmpdf2, .-__cmpdf2
#endif /* L_compare_df */
#ifdef L_unord_sf
.text
.align 2
.global __unordsf2
.type __unordsf2, @function
__unordsf2:
push $lp
slli $r2, $r0, #1
move $r3, #0xff000000
slt $r15, $r3, $r2
beqzs8 .Li52
move $r0, #1
j .LP999
.Li52:
slli $r2, $r1, #1
move $r3, #0xff000000
slt $r15, $r3, $r2
beqzs8 .Li53
move $r0, #1
j .LP999
.Li53:
move $r0, #0
.LP999:
pop $lp
ret5 $lp
.size __unordsf2, .-__unordsf2
#endif /* L_unord_sf */
#ifdef L_unord_df
#ifndef __big_endian__
#define P1L $r0
#define P1H $r1
#define P2L $r2
#define P2H $r3
#else
#define P1H $r0
#define P1L $r1
#define P2H $r2
#define P2L $r3
#endif
.text
.align 2
.global __unorddf2
.type __unorddf2, @function
__unorddf2:
push $lp
slli $r4, P1H, #1
beqz P1L, .Li66
addi $r4, $r4, #1
.Li66:
move $r5, #0xffe00000
slt $r15, $r5, $r4
beqzs8 .Li67
move $r0, #1
j .LR999
.Li67:
slli $r4, P2H, #1
beqz P2L, .Li68
addi $r4, $r4, #1
.Li68:
move $r5, #0xffe00000
slt $r15, $r5, $r4
beqzs8 .Li69
move $r0, #1
j .LR999
.Li69:
move $r0, #0
.LR999:
pop $lp
ret5 $lp
.size __unorddf2, .-__unorddf2
#endif /* L_unord_df */
/* ------------------------------------------- */
/* DPBIT floating point operations for libgcc */
/* ------------------------------------------- */