From edf30d84cc0e8299d61c98f5bb40a6428d1576bc Mon Sep 17 00:00:00 2001
From: Jannik Vogel <email@jannikvogel.de>
Date: Wed, 18 May 2016 15:12:31 +0200
Subject: [PATCH] Remove `exceptions` parameter from `normaliseround` VFP
 functions

---
 src/core/arm/skyeye_common/vfp/vfp_helper.h  |  4 +-
 src/core/arm/skyeye_common/vfp/vfpdouble.cpp | 40 ++++++++++++-------
 src/core/arm/skyeye_common/vfp/vfpsingle.cpp | 41 +++++++++++++-------
 3 files changed, 57 insertions(+), 28 deletions(-)

diff --git a/src/core/arm/skyeye_common/vfp/vfp_helper.h b/src/core/arm/skyeye_common/vfp/vfp_helper.h
index c66a3fad4b..68714800c8 100644
--- a/src/core/arm/skyeye_common/vfp/vfp_helper.h
+++ b/src/core/arm/skyeye_common/vfp/vfp_helper.h
@@ -304,7 +304,7 @@ inline s32 vfp_single_pack(const vfp_single* s)
 }
 
 
-u32 vfp_single_normaliseround(ARMul_State* state, int sd, vfp_single* vs, u32 fpscr, u32 exceptions, const char* func);
+u32 vfp_single_normaliseround(ARMul_State* state, int sd, vfp_single* vs, u32 fpscr, const char* func);
 
 // Double-precision
 struct vfp_double {
@@ -451,4 +451,4 @@ inline u32 fls(u32 x)
 
 u32 vfp_double_multiply(vfp_double* vdd, vfp_double* vdn, vfp_double* vdm, u32 fpscr);
 u32 vfp_double_add(vfp_double* vdd, vfp_double* vdn, vfp_double *vdm, u32 fpscr);
-u32 vfp_double_normaliseround(ARMul_State* state, int dd, vfp_double* vd, u32 fpscr, u32 exceptions, const char* func);
+u32 vfp_double_normaliseround(ARMul_State* state, int dd, vfp_double* vd, u32 fpscr, const char* func);
diff --git a/src/core/arm/skyeye_common/vfp/vfpdouble.cpp b/src/core/arm/skyeye_common/vfp/vfpdouble.cpp
index 0182b90b81..580e60c85f 100644
--- a/src/core/arm/skyeye_common/vfp/vfpdouble.cpp
+++ b/src/core/arm/skyeye_common/vfp/vfpdouble.cpp
@@ -85,11 +85,12 @@ static void vfp_double_normalise_denormal(struct vfp_double *vd)
     vfp_double_dump("normalise_denormal: out", vd);
 }
 
-u32 vfp_double_normaliseround(ARMul_State* state, int dd, struct vfp_double *vd, u32 fpscr, u32 exceptions, const char *func)
+u32 vfp_double_normaliseround(ARMul_State* state, int dd, struct vfp_double *vd, u32 fpscr, const char *func)
 {
     u64 significand, incr;
     int exponent, shift, underflow;
     u32 rmode;
+    u32 exceptions = 0;
 
     vfp_double_dump("pack: in", vd);
 
@@ -370,8 +371,7 @@ sqrt_invalid:
     }
     vdd.significand = vfp_shiftright64jamming(vdd.significand, 1);
 
-    exceptions |= vfp_double_normaliseround(state, dd, &vdd, fpscr, 0, "fsqrt");
-
+    exceptions |= vfp_double_normaliseround(state, dd, &vdd, fpscr, "fsqrt");
     return exceptions;
 }
 
@@ -507,7 +507,8 @@ static u32 vfp_double_fcvts(ARMul_State* state, int sd, int unused, int dm, u32
     else
         vsd.exponent = vdm.exponent - (1023 - 127);
 
-    return vfp_single_normaliseround(state, sd, &vsd, fpscr, exceptions, "fcvts");
+    exceptions |= vfp_single_normaliseround(state, sd, &vsd, fpscr, "fcvts");
+    return exceptions;
 
 pack_nan:
     vfp_put_float(state, vfp_single_pack(&vsd), sd);
@@ -517,6 +518,7 @@ pack_nan:
 static u32 vfp_double_fuito(ARMul_State* state, int dd, int unused, int dm, u32 fpscr)
 {
     struct vfp_double vdm;
+    u32 exceptions = 0;
     u32 m = vfp_get_float(state, dm);
 
     LOG_TRACE(Core_ARM11, "In %s", __FUNCTION__);
@@ -524,12 +526,14 @@ static u32 vfp_double_fuito(ARMul_State* state, int dd, int unused, int dm, u32
     vdm.exponent = 1023 + 63 - 1;
     vdm.significand = (u64)m;
 
-    return vfp_double_normaliseround(state, dd, &vdm, fpscr, 0, "fuito");
+    exceptions |= vfp_double_normaliseround(state, dd, &vdm, fpscr, "fuito");
+    return exceptions;
 }
 
 static u32 vfp_double_fsito(ARMul_State* state, int dd, int unused, int dm, u32 fpscr)
 {
     struct vfp_double vdm;
+    u32 exceptions = 0;
     u32 m = vfp_get_float(state, dm);
 
     LOG_TRACE(Core_ARM11, "In %s", __FUNCTION__);
@@ -537,7 +541,8 @@ static u32 vfp_double_fsito(ARMul_State* state, int dd, int unused, int dm, u32
     vdm.exponent = 1023 + 63 - 1;
     vdm.significand = vdm.sign ? (~m + 1) : m;
 
-    return vfp_double_normaliseround(state, dd, &vdm, fpscr, 0, "fsito");
+    exceptions |= vfp_double_normaliseround(state, dd, &vdm, fpscr, "fsito");
+    return exceptions;
 }
 
 static u32 vfp_double_ftoui(ARMul_State* state, int sd, int unused, int dm, u32 fpscr)
@@ -918,7 +923,8 @@ vfp_double_multiply_accumulate(ARMul_State* state, int dd, int dn, int dm, u32 f
 
     exceptions |= vfp_double_add(&vdd, &vdn, &vdp, fpscr);
 
-    return vfp_double_normaliseround(state, dd, &vdd, fpscr, exceptions, func);
+    exceptions |= vfp_double_normaliseround(state, dd, &vdd, fpscr, func);
+    return exceptions;
 }
 
 /*
@@ -979,7 +985,9 @@ static u32 vfp_double_fmul(ARMul_State* state, int dd, int dn, int dm, u32 fpscr
         vfp_double_normalise_denormal(&vdm);
 
     exceptions |= vfp_double_multiply(&vdd, &vdn, &vdm, fpscr);
-    return vfp_double_normaliseround(state, dd, &vdd, fpscr, exceptions, "fmul");
+
+    exceptions |= vfp_double_normaliseround(state, dd, &vdd, fpscr, "fmul");
+    return exceptions;
 }
 
 /*
@@ -1002,7 +1010,8 @@ static u32 vfp_double_fnmul(ARMul_State* state, int dd, int dn, int dm, u32 fpsc
     exceptions |= vfp_double_multiply(&vdd, &vdn, &vdm, fpscr);
     vdd.sign = vfp_sign_negate(vdd.sign);
 
-    return vfp_double_normaliseround(state, dd, &vdd, fpscr, exceptions, "fnmul");
+    exceptions |= vfp_double_normaliseround(state, dd, &vdd, fpscr, "fnmul");
+    return exceptions;
 }
 
 /*
@@ -1024,7 +1033,8 @@ static u32 vfp_double_fadd(ARMul_State* state, int dd, int dn, int dm, u32 fpscr
 
     exceptions |= vfp_double_add(&vdd, &vdn, &vdm, fpscr);
 
-    return vfp_double_normaliseround(state, dd, &vdd, fpscr, exceptions, "fadd");
+    exceptions |= vfp_double_normaliseround(state, dd, &vdd, fpscr, "fadd");
+    return exceptions;
 }
 
 /*
@@ -1051,7 +1061,8 @@ static u32 vfp_double_fsub(ARMul_State* state, int dd, int dn, int dm, u32 fpscr
 
     exceptions |= vfp_double_add(&vdd, &vdn, &vdm, fpscr);
 
-    return vfp_double_normaliseround(state, dd, &vdd, fpscr, exceptions, "fsub");
+    exceptions |= vfp_double_normaliseround(state, dd, &vdd, fpscr, "fsub");
+    return exceptions;
 }
 
 /*
@@ -1134,7 +1145,9 @@ static u32 vfp_double_fdiv(ARMul_State* state, int dd, int dn, int dm, u32 fpscr
         }
         vdd.significand |= (reml != 0);
     }
-    return vfp_double_normaliseround(state, dd, &vdd, fpscr, 0, "fdiv");
+
+    exceptions |= vfp_double_normaliseround(state, dd, &vdd, fpscr, "fdiv");
+    return exceptions;
 
 vdn_nan:
     exceptions |= vfp_propagate_nan(&vdd, &vdn, &vdm, fpscr);
@@ -1160,7 +1173,8 @@ infinity:
 
 invalid:
     vfp_put_double(state, vfp_double_pack(&vfp_double_default_qnan), dd);
-    return FPSCR_IOC;
+    exceptions |= FPSCR_IOC;
+    return exceptions;
 }
 
 static struct op fops[] = {
diff --git a/src/core/arm/skyeye_common/vfp/vfpsingle.cpp b/src/core/arm/skyeye_common/vfp/vfpsingle.cpp
index 5fb6b51ec7..23e0cdf261 100644
--- a/src/core/arm/skyeye_common/vfp/vfpsingle.cpp
+++ b/src/core/arm/skyeye_common/vfp/vfpsingle.cpp
@@ -89,10 +89,11 @@ static void vfp_single_normalise_denormal(struct vfp_single *vs)
 }
 
 
-u32 vfp_single_normaliseround(ARMul_State* state, int sd, struct vfp_single *vs, u32 fpscr, u32 exceptions, const char *func)
+u32 vfp_single_normaliseround(ARMul_State* state, int sd, struct vfp_single *vs, u32 fpscr, const char *func)
 {
     u32 significand, incr, rmode;
     int exponent, shift, underflow;
+    u32 exceptions = 0;
 
     vfp_single_dump("pack: in", vs);
 
@@ -409,8 +410,7 @@ sqrt_invalid:
     }
     vsd.significand = vfp_shiftright32jamming(vsd.significand, 1);
 
-    exceptions |= vfp_single_normaliseround(state, sd, &vsd, fpscr, 0, "fsqrt");
-
+    exceptions |= vfp_single_normaliseround(state, sd, &vsd, fpscr, "fsqrt");
     return exceptions;
 }
 
@@ -535,7 +535,8 @@ static u32 vfp_single_fcvtd(ARMul_State* state, int dd, int unused, s32 m, u32 f
     else
         vdd.exponent = vsm.exponent + (1023 - 127);
 
-    return vfp_double_normaliseround(state, dd, &vdd, fpscr, exceptions, "fcvtd");
+    exceptions |= vfp_double_normaliseround(state, dd, &vdd, fpscr, "fcvtd");
+    return exceptions;
 
 pack_nan:
     vfp_put_double(state, vfp_double_pack(&vdd), dd);
@@ -545,23 +546,27 @@ pack_nan:
 static u32 vfp_single_fuito(ARMul_State* state, int sd, int unused, s32 m, u32 fpscr)
 {
     struct vfp_single vs;
+    u32 exceptions = 0;
 
     vs.sign = 0;
     vs.exponent = 127 + 31 - 1;
     vs.significand = (u32)m;
 
-    return vfp_single_normaliseround(state, sd, &vs, fpscr, 0, "fuito");
+    exceptions |= vfp_single_normaliseround(state, sd, &vs, fpscr, "fuito");
+    return exceptions;
 }
 
 static u32 vfp_single_fsito(ARMul_State* state, int sd, int unused, s32 m, u32 fpscr)
 {
     struct vfp_single vs;
+    u32 exceptions = 0;
 
     vs.sign = (m & 0x80000000) >> 16;
     vs.exponent = 127 + 31 - 1;
     vs.significand = vs.sign ? -m : m;
 
-    return vfp_single_normaliseround(state, sd, &vs, fpscr, 0, "fsito");
+    exceptions |= vfp_single_normaliseround(state, sd, &vs, fpscr, "fsito");
+    return exceptions;
 }
 
 static u32 vfp_single_ftoui(ARMul_State* state, int sd, int unused, s32 m, u32 fpscr)
@@ -953,7 +958,8 @@ vfp_single_multiply_accumulate(ARMul_State* state, int sd, int sn, s32 m, u32 fp
 
     exceptions |= vfp_single_add(&vsd, &vsn, &vsp, fpscr);
 
-    return vfp_single_normaliseround(state, sd, &vsd, fpscr, exceptions, func);
+    exceptions |= vfp_single_normaliseround(state, sd, &vsd, fpscr, func);
+    return exceptions;
 }
 
 /*
@@ -965,8 +971,10 @@ vfp_single_multiply_accumulate(ARMul_State* state, int sd, int sn, s32 m, u32 fp
  */
 static u32 vfp_single_fmac(ARMul_State* state, int sd, int sn, s32 m, u32 fpscr)
 {
+    u32 exceptions = 0;
     LOG_TRACE(Core_ARM11, "s%u = %08x", sn, sd);
-    return vfp_single_multiply_accumulate(state, sd, sn, m, fpscr, 0, "fmac");
+    exceptions |= vfp_single_multiply_accumulate(state, sd, sn, m, fpscr, 0, "fmac");
+    return exceptions;
 }
 
 /*
@@ -1017,7 +1025,9 @@ static u32 vfp_single_fmul(ARMul_State* state, int sd, int sn, s32 m, u32 fpscr)
         vfp_single_normalise_denormal(&vsm);
 
     exceptions |= vfp_single_multiply(&vsd, &vsn, &vsm, fpscr);
-    return vfp_single_normaliseround(state, sd, &vsd, fpscr, exceptions, "fmul");
+
+    exceptions |= vfp_single_normaliseround(state, sd, &vsd, fpscr, "fmul");
+    return exceptions;
 }
 
 /*
@@ -1041,7 +1051,9 @@ static u32 vfp_single_fnmul(ARMul_State* state, int sd, int sn, s32 m, u32 fpscr
 
     exceptions |= vfp_single_multiply(&vsd, &vsn, &vsm, fpscr);
     vsd.sign = vfp_sign_negate(vsd.sign);
-    return vfp_single_normaliseround(state, sd, &vsd, fpscr, exceptions, "fnmul");
+
+    exceptions |= vfp_single_normaliseround(state, sd, &vsd, fpscr, "fnmul");
+    return exceptions;
 }
 
 /*
@@ -1068,7 +1080,8 @@ static u32 vfp_single_fadd(ARMul_State* state, int sd, int sn, s32 m, u32 fpscr)
 
     exceptions |= vfp_single_add(&vsd, &vsn, &vsm, fpscr);
 
-    return vfp_single_normaliseround(state, sd, &vsd, fpscr, exceptions, "fadd");
+    exceptions |= vfp_single_normaliseround(state, sd, &vsd, fpscr, "fadd");
+    return exceptions;
 }
 
 /*
@@ -1165,7 +1178,8 @@ static u32 vfp_single_fdiv(ARMul_State* state, int sd, int sn, s32 m, u32 fpscr)
     if ((vsd.significand & 0x3f) == 0)
         vsd.significand |= ((u64)vsm.significand * vsd.significand != (u64)vsn.significand << 32);
 
-    return vfp_single_normaliseround(state, sd, &vsd, fpscr, 0, "fdiv");
+    exceptions |= vfp_single_normaliseround(state, sd, &vsd, fpscr, "fdiv");
+    return exceptions;
 
 vsn_nan:
     exceptions |= vfp_propagate_nan(&vsd, &vsn, &vsm, fpscr);
@@ -1191,7 +1205,8 @@ infinity:
 
 invalid:
     vfp_put_float(state, vfp_single_pack(&vfp_single_default_qnan), sd);
-    return FPSCR_IOC;
+    exceptions |= FPSCR_IOC;
+    return exceptions;
 }
 
 static struct op fops[] = {
-- 
GitLab