From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from mail-pl1-f176.google.com (mail-pl1-f176.google.com [209.85.214.176]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 7D15920E02E for ; Tue, 18 Mar 2025 16:45:30 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=209.85.214.176 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1742316335; cv=none; b=kZjaWbgeGgHbuTXIQMaurVqoBYhsEzfEtvZlBeRkfrPBxIsszsvTTw3/z73Z9t3qIwpAs/saF7X6rIUx0fzM/6l7VL1FJeCx3or9TiNKtft1mZQ9UjgTc4EmoGhNtRgJz43JDq6OF0xBzgv7RVwCMMJ+rmTdzYj7QWEHIENnKtM= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1742316335; c=relaxed/simple; bh=GX9Hhpr4Cf4kaSZFj7ZaqrW8sQpYr3Qo5Sn/Z+LN98A=; h=Date:From:To:Cc:Subject:Message-ID:References:MIME-Version: Content-Type:Content-Disposition:In-Reply-To; b=OE+ZpCDyzUNxOv+Ed6YQxKQ5DrZG7YfpSTwV/Rg6OmpyL6C290LkRKvsr0qYeB8rb203LGchGrdudRqHEr9UM3sNlSnZMP4o7V+7geule0Wp8btgZ4VrJOLbFGH7kznk21r5jO/QRRGP5enIT489GBebZOgGDW+UVYnVCAe2KW8= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=gmail.com; spf=pass smtp.mailfrom=gmail.com; dkim=pass (2048-bit key) header.d=gmail.com header.i=@gmail.com header.b=eiTb4f99; arc=none smtp.client-ip=209.85.214.176 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=gmail.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=gmail.com Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=gmail.com header.i=@gmail.com header.b="eiTb4f99" Received: by mail-pl1-f176.google.com with SMTP id d9443c01a7336-2240b4de12bso17505565ad.2 for ; Tue, 18 Mar 2025 09:45:30 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20230601; t=1742316330; x=1742921130; darn=vger.kernel.org; h=in-reply-to:content-disposition:mime-version:references:message-id :subject:cc:to:from:date:from:to:cc:subject:date:message-id:reply-to; bh=zXmFAGpWhU+jGQxea1D54Lg56ApGGJjtidCnAmzD5IQ=; b=eiTb4f999GUc8aEebbU1OElepSn480KPvXbMU3KpZ6D1bhPPqMzw4vjZf7Rj/FU8r8 b7w6cqrA5YGgHlLhxobZByx1RhJ5aE0xQUWI+gTAP43JQYDW30C1EfpTlQde3h3D4DMj 5M/8C2aDmimzzXFP2UiEmGcCI5D2GqeB3ZEsNfgIxrlXW/JsmUd39y6e9obqqeqIZLN8 3rOPWlhDx/BA518gscKkOwxnuQKdX0E+ingj2YQlT6xwUvg09zvWO27SoudJZ9PwGmey 7GqYcW/t6kWL2N9TFWJ13zCo31zB2w6Em/UlavtB2mE1V3Byfsauzihhqvi1jBS8gwzK 2syg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1742316330; x=1742921130; h=in-reply-to:content-disposition:mime-version:references:message-id :subject:cc:to:from:date:x-gm-message-state:from:to:cc:subject:date :message-id:reply-to; bh=zXmFAGpWhU+jGQxea1D54Lg56ApGGJjtidCnAmzD5IQ=; b=fIruzuo+U7cXpo+xxF1snBT9nXltY4ploiR0WM3TolzF+P/9NIaLi0VAoNFNKqYadG ihG22WlMfy80nV6RGITldmgnmO+wRWQ89YwgNwlMKsPobbluhYCPCrN8rZJqSfxc2FQr iOvDlgGG0FlNP3UALW2d/E67WEvrR+zf4g3bF1TZ8njuwH04n3i8dJIutKH3NMhRKF42 sQNM08kbv7iL/g1T9dY3F4oJWkLfEj/LHcJs4gSjYkgdgzEl0X8pMirXMcsodKmx42C+ AKFTmJmhfkNtNZWIErp0s2rwoXf/k7WAiGQVg2ONN4U8b3xe8hAEkMJZJrBBVsC7ZJPi c7ew== X-Forwarded-Encrypted: i=1; AJvYcCVgPz431JHA3nHA1Z6wNj/Dp59Z3My16nfVL7+I7wo52JcIQsMaZzcd02cMmaI/QwWSp9vgE37xtny00gs=@vger.kernel.org X-Gm-Message-State: AOJu0Yzv9CYIRNgwcYeB8va7zjf14ny/gRlxT7Vwlc0708pz/IK0bs6W dkHDDzusKpftuwDwhwaHStDY4qkRC8YIW5EMTzNbucejzsVAjBbd X-Gm-Gg: ASbGncunZ7yNwJ7rCoXxNRpv2ofpwNAbsytukrVQKgQEF2Uqsp5gLPdDEEcVBf6CfB1 nxSyD4KZWpTZYv9JboRbXI5t7d1npDTiFPoIyl+H+7kd0J3YJ9jE4z9hX1b3QcnUXNzOmywt5rT jETyoR9VeEIzzWeAMO9KismgdNgDXeeEpUauLSR/nw0IZEWkGDHR5biKgx0Oaql7Yh7OQ6MV43P 4XQsgAEwXAG4NxVbcqK5GwNtBnv2qveIzOGmSJpPI5jSVHM+eSmQaumSuvwPd/QkevJxubLOACe 0yAWvt/TgdvJbhujU4ymgCgbzSWqiyxDAUy+gYGFM8G+ZaRW/w1F/r3hw8p49vxWX6Gs X-Google-Smtp-Source: AGHT+IHCxAIWfvRZL39c//ZSCAbKfEdVL95M2CO4F3JUYMCR0tKZOs+seM5sZPP1v97nJ2K2V93M/g== X-Received: by 2002:a17:902:f785:b0:221:7b4a:475a with SMTP id d9443c01a7336-225e0b62d31mr227139965ad.52.1742316329526; Tue, 18 Mar 2025 09:45:29 -0700 (PDT) Received: from localhost (maglev-oncall.nvidia.com. [216.228.125.128]) by smtp.gmail.com with ESMTPSA id d2e1a72fcca58-737116af372sm9796545b3a.160.2025.03.18.09.45.28 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 18 Mar 2025 09:45:28 -0700 (PDT) Date: Tue, 18 Mar 2025 12:45:27 -0400 From: Yury Norov To: mailhol.vincent@wanadoo.fr Cc: Lucas De Marchi , Rasmus Villemoes , Jani Nikula , Joonas Lahtinen , Rodrigo Vivi , Tvrtko Ursulin , David Airlie , Simona Vetter , Andrew Morton , linux-kernel@vger.kernel.org, intel-gfx@lists.freedesktop.org, dri-devel@lists.freedesktop.org, Andi Shyti , David Laight , Dmitry Baryshkov , Andy Shevchenko , Jani Nikula Subject: Re: [PATCH v6 2/7] bits: introduce fixed-type genmasks Message-ID: References: <20250308-fixed-type-genmasks-v6-0-f59315e73c29@wanadoo.fr> <20250308-fixed-type-genmasks-v6-2-f59315e73c29@wanadoo.fr> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20250308-fixed-type-genmasks-v6-2-f59315e73c29@wanadoo.fr> On Sat, Mar 08, 2025 at 01:48:49AM +0900, Vincent Mailhol via B4 Relay wrote: > From: Yury Norov > > Add GENMASK_TYPE() which generalizes __GENMASK() to support different > types, and implement fixed-types versions of GENMASK() based on it. > The fixed-type version allows more strict checks to the min/max values > accepted, which is useful for defining registers like implemented by > i915 and xe drivers with their REG_GENMASK*() macros. > > The strict checks rely on shift-count-overflow compiler check to fail > the build if a number outside of the range allowed is passed. > Example: > > #define FOO_MASK GENMASK_U32(33, 4) > > will generate a warning like: > > include/linux/bits.h:51:27: error: right shift count >= width of type [-Werror=shift-count-overflow] > 51 | type_max(t) >> (BITS_PER_TYPE(t) - 1 - (h))))) > | ^~ > > Signed-off-by: Yury Norov > Signed-off-by: Lucas De Marchi > Acked-by: Jani Nikula > Co-developed-by: Vincent Mailhol > Signed-off-by: Vincent Mailhol > --- > Changelog: > > v5 -> v6: > > - No changes. > > v4 -> v5: > > - Rename GENMASK_t() to GENMASK_TYPE(). > > - Fix typo in patch description. > > - Use tab indentations instead of single space to separate the > macro name from its body. > > - s/__GENMASK_U*()/GENMASK_U*()/g in the comment. > > - Add a tag to credit myself as Co-developer. Keep Yury as the > main author. > > - Modify GENMASK_TYPE() to match the changes made to __GENMASK() > in: https://github.com/norov/linux/commit/1e7933a575ed > > - Replace (t)~_ULL(0) with type_max(t). This is OK because > GENMASK_TYPE() is not available in asm. > > - linux/const.h and asm/bitsperlong.h are not used anymore. Remove > them. > > - Apply GENMASK_TYPE() to GENMASK_U128(). > > - Remove the unsigned int cast for the U8 and U16 variants. Cast > to the target type instead. Do that cast directly in > GENMASK_TYPE(). > > v3 -> v4: > > - The v3 is one year old. Meanwhile people started using > __GENMASK() directly. So instead of generalizing __GENMASK() to > support different types, add a new GENMASK_t(). > > - replace ~0ULL by ~_ULL(0). Otherwise, GENMASK_t() would fail in > asm code. > > - Make GENMASK_U8() and GENMASK_U16() return an unsigned int. In > v3, due to the integer promotion rules, these were returning a > signed integer. By casting these to unsigned int, at least the > signedness is kept. > --- > include/linux/bitops.h | 1 - > include/linux/bits.h | 55 ++++++++++++++++++++++++++++++++------------------ > 2 files changed, 35 insertions(+), 21 deletions(-) > > diff --git a/include/linux/bitops.h b/include/linux/bitops.h > index c1cb53cf2f0f8662ed3e324578f74330e63f935d..9be2d50da09a417966b3d11c84092bb2f4cd0bef 100644 > --- a/include/linux/bitops.h > +++ b/include/linux/bitops.h > @@ -8,7 +8,6 @@ > > #include > > -#define BITS_PER_TYPE(type) (sizeof(type) * BITS_PER_BYTE) > #define BITS_TO_LONGS(nr) __KERNEL_DIV_ROUND_UP(nr, BITS_PER_TYPE(long)) > #define BITS_TO_U64(nr) __KERNEL_DIV_ROUND_UP(nr, BITS_PER_TYPE(u64)) > #define BITS_TO_U32(nr) __KERNEL_DIV_ROUND_UP(nr, BITS_PER_TYPE(u32)) > diff --git a/include/linux/bits.h b/include/linux/bits.h > index 9c1c7ce0bba6bb09490d891904c143a5394fd512..b690611c769be61ab2b5ced43c8302ba5693308b 100644 > --- a/include/linux/bits.h > +++ b/include/linux/bits.h > @@ -2,16 +2,15 @@ > #ifndef __LINUX_BITS_H > #define __LINUX_BITS_H > > -#include > #include > #include > -#include > > #define BIT_MASK(nr) (UL(1) << ((nr) % BITS_PER_LONG)) > #define BIT_WORD(nr) ((nr) / BITS_PER_LONG) > #define BIT_ULL_MASK(nr) (ULL(1) << ((nr) % BITS_PER_LONG_LONG)) > #define BIT_ULL_WORD(nr) ((nr) / BITS_PER_LONG_LONG) > #define BITS_PER_BYTE 8 > +#define BITS_PER_TYPE(type) (sizeof(type) * BITS_PER_BYTE) > > /* > * Create a contiguous bitmask starting at bit position @l and ending at > @@ -20,28 +19,44 @@ > */ > #if !defined(__ASSEMBLY__) > > -#include > -#include > - > -#define GENMASK_INPUT_CHECK(h, l) BUILD_BUG_ON_ZERO(const_true((l) > (h))) > - > -#define GENMASK(h, l) \ > - (GENMASK_INPUT_CHECK(h, l) + __GENMASK(h, l)) > -#define GENMASK_ULL(h, l) \ > - (GENMASK_INPUT_CHECK(h, l) + __GENMASK_ULL(h, l)) > - > /* > * Missing asm support > * > - * __GENMASK_U128() depends on _BIT128() which would not work > - * in the asm code, as it shifts an 'unsigned __int128' data > - * type instead of direct representation of 128 bit constants > - * such as long and unsigned long. The fundamental problem is > - * that a 128 bit constant will get silently truncated by the > - * gcc compiler. > + * GENMASK_U*() depends on BITS_PER_TYPE() which relies on sizeof(), > + * something not available in asm. Nethertheless, fixed width integers > + * is a C concept. Assembly code can rely on the long and long long > + * versions instead. > */ > -#define GENMASK_U128(h, l) \ > - (GENMASK_INPUT_CHECK(h, l) + __GENMASK_U128(h, l)) > + > +#include > +#include > +#include > + > +#define GENMASK_INPUT_CHECK(h, l) BUILD_BUG_ON_ZERO(const_true((l) > (h))) > + > +/* > + * Generate a mask for the specified type @t. Additional checks are made to > + * guarantee the value returned fits in that type, relying on > + * shift-count-overflow compiler check to detect incompatible arguments. > + * For example, all these create build errors or warnings: > + * > + * - GENMASK(15, 20): wrong argument order > + * - GENMASK(72, 15): doesn't fit unsigned long > + * - GENMASK_U32(33, 15): doesn't fit in a u32 > + */ > +#define GENMASK_TYPE(t, h, l) \ > + ((t)(GENMASK_INPUT_CHECK(h, l) + \ > + (type_max(t) << (l) & \ > + type_max(t) >> (BITS_PER_TYPE(t) - 1 - (h))))) > + > +#define GENMASK(h, l) GENMASK_TYPE(unsigned long, h, l) > +#define GENMASK_ULL(h, l) GENMASK_TYPE(unsigned long long, h, l) I like everything except this part. We switch GENMASK() from a well tested implementation, including an asm code, and we split uapi and non-uapi users, with no functionality changes. Unification is a solid point, however. Let's make it a 2-step procedure? Adding fixed-width GENMASKs is a non-questionable improvement. Switching an existing API from one implementation to another should be a separate patch, and probably even a separate series. And we should be very clear that __GENMASK() is uapi-only thing from now. If we decide to switch GENMASK() in a separate series, we'll have some extra time to think about unification... > +#define GENMASK_U8(h, l) GENMASK_TYPE(u8, h, l) > +#define GENMASK_U16(h, l) GENMASK_TYPE(u16, h, l) > +#define GENMASK_U32(h, l) GENMASK_TYPE(u32, h, l) > +#define GENMASK_U64(h, l) GENMASK_TYPE(u64, h, l) > +#define GENMASK_U128(h, l) GENMASK_TYPE(u128, h, l) > > #else /* defined(__ASSEMBLY__) */ > > > -- > 2.45.3 >