From: "Alex Bennée" <alex.bennee@linaro.org>
To: Richard Henderson <richard.henderson@linaro.org>
Cc: qemu-devel@nongnu.org, peter.maydell@linaro.org
Subject: Re: [Qemu-devel] [PATCH v5 07/28] fpu/softfloat: Replace float_class_dnan with parts_default_nan
Date: Tue, 15 May 2018 12:46:10 +0100 [thread overview]
Message-ID: <87lgclf8fx.fsf@linaro.org> (raw)
In-Reply-To: <20180514221219.7091-8-richard.henderson@linaro.org>
Richard Henderson <richard.henderson@linaro.org> writes:
> With a canonical representation of NaNs, we can return the
> default nan directly rather than delay the expansion until
> the final format is known.
>
> Note one case where we uselessly assigned to a.sign, which was
> overwritten/ignored later when expanding float_class_dnan.
>
> Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
> ---
> fpu/softfloat-specialize.h | 37 +++++++++++++++++++++++++++++++++++++
> fpu/softfloat.c | 38 +++++++++++---------------------------
> 2 files changed, 48 insertions(+), 27 deletions(-)
>
> diff --git a/fpu/softfloat-specialize.h b/fpu/softfloat-specialize.h
> index 515cb12cfa..0d3d81a52b 100644
> --- a/fpu/softfloat-specialize.h
> +++ b/fpu/softfloat-specialize.h
> @@ -101,6 +101,43 @@ static bool parts_is_snan_frac(uint64_t frac, float_status *status)
> #endif
> }
>
> +/*----------------------------------------------------------------------------
> +| The pattern for a default generated deconstructed floating-point NaN.
> +*----------------------------------------------------------------------------*/
> +
> +static FloatParts parts_default_nan(float_status *status)
> +{
> + bool sign = 0;
> + uint64_t frac;
> +
> +#if defined(TARGET_SPARC) || defined(TARGET_M68K)
> + frac = (1ULL << DECOMPOSED_BINARY_POINT) - 1;
> +#elif defined(TARGET_PPC) || defined(TARGET_ARM) || defined(TARGET_ALPHA) || \
> + defined(TARGET_S390X) || defined(TARGET_RISCV)
> + frac = 1ULL << (DECOMPOSED_BINARY_POINT - 1);
> +#elif defined(TARGET_HPPA)
> + frac = 1ULL << (DECOMPOSED_BINARY_POINT - 2);
> +#else
> + if (status->snan_bit_is_one) {
> + frac = (1ULL << (DECOMPOSED_BINARY_POINT - 1)) - 1;
> + } else {
> +#if defined(TARGET_MIPS)
> + frac = 1ULL << (DECOMPOSED_BINARY_POINT - 1);
> +#else
> + frac = 1ULL << (DECOMPOSED_BINARY_POINT - 1);
> + sign = 1;
> +#endif
> + }
> +#endif
> +
> + return (FloatParts) {
> + .cls = float_class_qnan,
> + .sign = sign,
> + .exp = INT_MAX,
> + .frac = frac
> + };
> +}
> +
> /*----------------------------------------------------------------------------
> | The pattern for a default generated half-precision NaN.
> *----------------------------------------------------------------------------*/
> diff --git a/fpu/softfloat.c b/fpu/softfloat.c
> index 19f40d6932..51780b718f 100644
> --- a/fpu/softfloat.c
> +++ b/fpu/softfloat.c
> @@ -188,7 +188,6 @@ typedef enum __attribute__ ((__packed__)) {
> float_class_inf,
> float_class_qnan, /* all NaNs from here */
> float_class_snan,
> - float_class_dnan,
> float_class_msnan, /* maybe silenced */
> } FloatClass;
>
> @@ -494,8 +493,6 @@ static FloatParts float16_unpack_canonical(float16 f, float_status *s)
> static float16 float16_round_pack_canonical(FloatParts p, float_status *s)
> {
> switch (p.cls) {
> - case float_class_dnan:
> - return float16_default_nan(s);
> case float_class_msnan:
> p.frac >>= float16_params.frac_shift;
> return float16_maybe_silence_nan(float16_pack_raw(p), s);
> @@ -513,8 +510,6 @@ static FloatParts float32_unpack_canonical(float32 f, float_status *s)
> static float32 float32_round_pack_canonical(FloatParts p, float_status *s)
> {
> switch (p.cls) {
> - case float_class_dnan:
> - return float32_default_nan(s);
> case float_class_msnan:
> p.frac >>= float32_params.frac_shift;
> return float32_maybe_silence_nan(float32_pack_raw(p), s);
> @@ -532,8 +527,6 @@ static FloatParts float64_unpack_canonical(float64 f, float_status *s)
> static float64 float64_round_pack_canonical(FloatParts p, float_status *s)
> {
> switch (p.cls) {
> - case float_class_dnan:
> - return float64_default_nan(s);
> case float_class_msnan:
> p.frac >>= float64_params.frac_shift;
> return float64_maybe_silence_nan(float64_pack_raw(p), s);
> @@ -566,7 +559,7 @@ static FloatParts return_nan(FloatParts a, float_status *s)
> /* fall through */
> case float_class_qnan:
> if (s->default_nan_mode) {
> - a.cls = float_class_dnan;
> + return parts_default_nan(s);
> }
> break;
>
> @@ -583,7 +576,7 @@ static FloatParts pick_nan(FloatParts a, FloatParts b, float_status *s)
> }
>
> if (s->default_nan_mode) {
> - a.cls = float_class_dnan;
> + return parts_default_nan(s);
> } else {
> if (pickNaN(is_qnan(a.cls), is_snan(a.cls),
> is_qnan(b.cls), is_snan(b.cls),
> @@ -614,8 +607,7 @@ static FloatParts pick_nan_muladd(FloatParts a, FloatParts b, FloatParts c,
> /* Note that this check is after pickNaNMulAdd so that function
> * has an opportunity to set the Invalid flag.
> */
> - a.cls = float_class_dnan;
> - return a;
> + which = 3;
> }
>
> switch (which) {
> @@ -628,8 +620,7 @@ static FloatParts pick_nan_muladd(FloatParts a, FloatParts b, FloatParts c,
> a = c;
> break;
> case 3:
> - a.cls = float_class_dnan;
> - return a;
> + return parts_default_nan(s);
> default:
> g_assert_not_reached();
> }
> @@ -682,7 +673,7 @@ static FloatParts addsub_floats(FloatParts a, FloatParts b, bool subtract,
> if (a.cls == float_class_inf) {
> if (b.cls == float_class_inf) {
> float_raise(float_flag_invalid, s);
> - a.cls = float_class_dnan;
> + return parts_default_nan(s);
> }
> return a;
> }
> @@ -828,9 +819,7 @@ static FloatParts mul_floats(FloatParts a, FloatParts b, float_status *s)
> if ((a.cls == float_class_inf && b.cls == float_class_zero) ||
> (a.cls == float_class_zero && b.cls == float_class_inf)) {
> s->float_exception_flags |= float_flag_invalid;
> - a.cls = float_class_dnan;
> - a.sign = sign;
> - return a;
> + return parts_default_nan(s);
> }
> /* Multiply by 0 or Inf */
> if (a.cls == float_class_inf || a.cls == float_class_zero) {
> @@ -908,8 +897,7 @@ static FloatParts muladd_floats(FloatParts a, FloatParts b, FloatParts c,
>
> if (inf_zero) {
> s->float_exception_flags |= float_flag_invalid;
> - a.cls = float_class_dnan;
> - return a;
> + return parts_default_nan(s);
> }
>
> if (flags & float_muladd_negate_c) {
> @@ -933,12 +921,12 @@ static FloatParts muladd_floats(FloatParts a, FloatParts b, FloatParts c,
> if (c.cls == float_class_inf) {
> if (p_class == float_class_inf && p_sign != c.sign) {
> s->float_exception_flags |= float_flag_invalid;
> - a.cls = float_class_dnan;
> + return parts_default_nan(s);
> } else {
> a.cls = float_class_inf;
> a.sign = c.sign ^ sign_flip;
> + return a;
> }
> - return a;
> }
>
> if (p_class == float_class_inf) {
> @@ -1148,8 +1136,7 @@ static FloatParts div_floats(FloatParts a, FloatParts b, float_status *s)
> &&
> (a.cls == float_class_inf || a.cls == float_class_zero)) {
> s->float_exception_flags |= float_flag_invalid;
> - a.cls = float_class_dnan;
> - return a;
> + return parts_default_nan(s);
> }
> /* Inf / x or 0 / x */
> if (a.cls == float_class_inf || a.cls == float_class_zero) {
> @@ -1347,7 +1334,6 @@ static int64_t round_to_int_and_pack(FloatParts in, int rmode,
> switch (p.cls) {
> case float_class_snan:
> case float_class_qnan:
> - case float_class_dnan:
> case float_class_msnan:
> s->float_exception_flags = orig_flags | float_flag_invalid;
> return max;
> @@ -1439,7 +1425,6 @@ static uint64_t round_to_uint_and_pack(FloatParts in, int rmode, uint64_t max,
> switch (p.cls) {
> case float_class_snan:
> case float_class_qnan:
> - case float_class_dnan:
> case float_class_msnan:
> s->float_exception_flags = orig_flags | float_flag_invalid;
> return max;
> @@ -1940,8 +1925,7 @@ static FloatParts sqrt_float(FloatParts a, float_status *s, const FloatFmt *p)
> }
> if (a.sign) {
> s->float_exception_flags |= float_flag_invalid;
> - a.cls = float_class_dnan;
> - return a;
> + return parts_default_nan(s);
> }
> if (a.cls == float_class_inf) {
> return a; /* sqrt(+inf) = +inf */
--
Alex Bennée
next prev parent reply other threads:[~2018-05-15 11:46 UTC|newest]
Thread overview: 55+ messages / expand[flat|nested] mbox.gz Atom feed top
2018-05-14 22:11 [Qemu-devel] [PATCH v5 00/28] softfloat patch roundup Richard Henderson
2018-05-14 22:11 ` [Qemu-devel] [PATCH v5 01/28] fpu/softfloat: Fix conversion from uint64 to float128 Richard Henderson
2018-05-15 10:52 ` Alex Bennée
2018-05-14 22:11 ` [Qemu-devel] [PATCH v5 02/28] fpu/softfloat: Merge NO_SIGNALING_NANS definitions Richard Henderson
2018-05-14 22:11 ` [Qemu-devel] [PATCH v5 03/28] fpu/softfloat: Split floatXX_silence_nan from floatXX_maybe_silence_nan Richard Henderson
2018-05-14 22:11 ` [Qemu-devel] [PATCH v5 04/28] fpu/softfloat: Move softfloat-specialize.h below FloatParts definition Richard Henderson
2018-05-14 22:11 ` [Qemu-devel] [PATCH v5 05/28] fpu/softfloat: Canonicalize NaN fraction Richard Henderson
2018-05-15 10:50 ` Peter Maydell
2018-05-15 11:44 ` Alex Bennée
2018-05-14 22:11 ` [Qemu-devel] [PATCH v5 06/28] fpu/softfloat: Introduce parts_is_snan_frac Richard Henderson
2018-05-15 11:44 ` Alex Bennée
2018-05-14 22:11 ` [Qemu-devel] [PATCH v5 07/28] fpu/softfloat: Replace float_class_dnan with parts_default_nan Richard Henderson
2018-05-15 11:46 ` Alex Bennée [this message]
2018-05-14 22:11 ` [Qemu-devel] [PATCH v5 08/28] fpu/softfloat: Replace float_class_msnan with parts_silence_nan Richard Henderson
2018-05-15 11:51 ` Alex Bennée
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 09/28] target/arm: Use floatX_silence_nan when we have already checked for SNaN Richard Henderson
2018-05-15 13:19 ` Alex Bennée
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 10/28] target/arm: convert conversion helpers to fpst/ahp_flag Richard Henderson
2018-05-15 10:08 ` Peter Maydell
2018-05-15 15:01 ` Richard Henderson
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 11/28] target/arm: squash FZ16 behaviour for conversions Richard Henderson
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 12/28] target/arm: Remove floatX_maybe_silence_nan from conversions Richard Henderson
2018-05-15 13:34 ` Alex Bennée
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 13/28] fpu/softfloat: Partial support for ARM Alternative half-precision Richard Henderson
2018-05-15 10:52 ` Peter Maydell
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 14/28] fpu/softfloat: re-factor float to float conversions Richard Henderson
2018-05-15 10:53 ` Peter Maydell
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 15/28] target/hppa: Remove floatX_maybe_silence_nan from conversions Richard Henderson
2018-05-15 13:39 ` Alex Bennée
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 16/28] target/m68k: Use floatX_silence_nan when we have already checked for SNaN Richard Henderson
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 17/28] target/mips: Remove floatX_maybe_silence_nan from conversions Richard Henderson
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 18/28] target/riscv: " Richard Henderson
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 19/28] target/s390x: " Richard Henderson
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 20/28] fpu/softfloat: Use float*_silence_nan in propagateFloat*NaN Richard Henderson
2018-05-15 13:40 ` Alex Bennée
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 21/28] fpu/softfloat: Remove floatX_maybe_silence_nan Richard Henderson
2018-05-15 13:41 ` Alex Bennée
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 22/28] fpu/softfloat: Specialize on snan_bit_is_one Richard Henderson
2018-05-15 10:57 ` Peter Maydell
2018-05-15 13:51 ` Alex Bennée
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 23/28] fpu/softfloat: Make is_nan et al available to softfloat-specialize.h Richard Henderson
2018-05-15 10:59 ` Peter Maydell
2018-05-15 13:51 ` Alex Bennée
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 24/28] fpu/softfloat: Pass FloatClass to pickNaN Richard Henderson
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 25/28] fpu/softfloat: Pass FloatClass to pickNaNMulAdd Richard Henderson
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 26/28] fpu/softfloat: Define floatN_default_nan in terms of parts_default_nan Richard Henderson
2018-05-15 10:38 ` Peter Maydell
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 27/28] fpu/softfloat: Clean up parts_default_nan Richard Henderson
2018-05-15 13:42 ` Alex Bennée
2018-05-14 22:12 ` [Qemu-devel] [PATCH v5 28/28] fpu/softfloat: Define floatN_silence_nan in terms of parts_silence_nan Richard Henderson
2018-05-15 13:45 ` Alex Bennée
2018-05-15 15:41 ` Richard Henderson
2018-05-15 16:14 ` Richard Henderson
2018-05-15 11:43 ` [Qemu-devel] [PATCH v5 00/28] softfloat patch roundup Alex Bennée
2018-05-15 13:57 ` Alex Bennée
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=87lgclf8fx.fsf@linaro.org \
--to=alex.bennee@linaro.org \
--cc=peter.maydell@linaro.org \
--cc=qemu-devel@nongnu.org \
--cc=richard.henderson@linaro.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).