From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from gabe.freedesktop.org (gabe.freedesktop.org [131.252.210.177]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.lore.kernel.org (Postfix) with ESMTPS id 27D16C433F5 for ; Wed, 1 Jun 2022 08:19:13 +0000 (UTC) Received: from gabe.freedesktop.org (localhost [127.0.0.1]) by gabe.freedesktop.org (Postfix) with ESMTP id B39711134E0; Wed, 1 Jun 2022 08:19:12 +0000 (UTC) Received: from mga07.intel.com (mga07.intel.com [134.134.136.100]) by gabe.freedesktop.org (Postfix) with ESMTPS id DC5111134E0; Wed, 1 Jun 2022 08:19:11 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1654071551; x=1685607551; h=date:from:to:cc:subject:message-id:references: content-transfer-encoding:in-reply-to:mime-version; bh=ekhjadUi5MThPheaK2PhUx9B8SyCiZki6M33kt09RPw=; b=F/cgpOtR/rUVCOfqbDyG3jdd2e4wqo7FFF3H/RDbxNkQ078S9mMiWV07 VC/4lZCqBBviUSGBzb9tbM2lShdgBjGXaF4cToArKkqgoeQPW/BovSSQF Dv8SUMzw8LuXXo27ItcwfE7cC/FyKL7G5qWDU36tcWLOjvPID6RT+P4Dc OvVNGD4cdPXl/sFClPfd5swer0aWIpr4ObMo2amkhT0k3/Vj6UTZhcD/r hbuK+heZaED4CBAUt37QjFoNWWRCUQ9d34nbhLIRdnwrXrtid2zbQD4PX jfWjN3oROSq05EfXftXv3EkPqbOaZvWrvHMufKH71m3zQ1N9lQKKfrmxo A==; X-IronPort-AV: E=McAfee;i="6400,9594,10364"; a="338559717" X-IronPort-AV: E=Sophos;i="5.91,266,1647327600"; d="scan'208";a="338559717" Received: from fmsmga007.fm.intel.com ([10.253.24.52]) by orsmga105.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 01 Jun 2022 01:19:11 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.91,266,1647327600"; d="scan'208";a="581476284" Received: from fmsmsx604.amr.corp.intel.com ([10.18.126.84]) by fmsmga007.fm.intel.com with ESMTP; 01 Jun 2022 01:19:10 -0700 Received: from fmsmsx612.amr.corp.intel.com (10.18.126.92) by fmsmsx604.amr.corp.intel.com (10.18.126.84) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256) id 15.1.2308.27; Wed, 1 Jun 2022 01:19:10 -0700 Received: from FMSEDG603.ED.cps.intel.com (10.1.192.133) by fmsmsx612.amr.corp.intel.com (10.18.126.92) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256) id 15.1.2308.27 via Frontend Transport; Wed, 1 Jun 2022 01:19:10 -0700 Received: from NAM11-CO1-obe.outbound.protection.outlook.com (104.47.56.174) by edgegateway.intel.com (192.55.55.68) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.1.2308.27; Wed, 1 Jun 2022 01:19:09 -0700 ARC-Seal: i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none; b=alGgYHhVVEdxNtK3Yzj31xiWeJnBsX5wUWjQLi0p0Pvkb/XDWn0YkH1kbBoP8Z0yxj1ZA8++RUK43ONXZzx1YVySAduVp1o20Yp3pDjHuTCbDB5DcPsdgkmSx+c1YYh+q1YBI3mNY4PmaXD1056EEL2mmoqpvhw39zC5INGzyu2qnwlROyWlyxMsKuDTm79NwOQMuXiRLfLvojuXglIfYfggyKbSOEfu18fYA/VBs37nUAyAI97h/WVJSVg6GlS4FZS5D1nKuhqijrKoxgP5V+4ns1gH+5GRvliz/MfCgBfR7Fbn3enJ99DeeLUz7AdqfS+btEOHmlgf821u+CUTAQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector9901; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-AntiSpam-MessageData-ChunkCount:X-MS-Exchange-AntiSpam-MessageData-0:X-MS-Exchange-AntiSpam-MessageData-1; bh=lPjxY0pfmg29XXYG5QCeLFauO9vf5YXyKMJSvI6u1ow=; b=FoRagiOQ0GzBDNy6Y9XDsjtYjLaq5eojzQrf2985+ZE7Ag0GgDC+5PSbeCPH7X6/kdBM+93PsKLgQ03JkoVDGFw3r/56EDV5xwfBmRx72Fk6mwxAB8P2RlfzNIMfjDjUQc3lW7XPyX8E0nnmamINue0vt4hFCz1wC2FyUPqdPLM5p+mYEFv6PePtZNihfQrTV14lvkqhmHQT7ccH2DV1TiQKtI055Y+D+cfgIsmeNTrPMy7rt5Cr28tSzqY9q8bFe1Z0rDT9uJiyffSmqLwXhgj1OVGNyGgMXpqSbmDJqebFHNHO+hKDsmvwqAep0fbAWO2iICC1E4GK4oRgHu+9Ow== ARC-Authentication-Results: i=1; mx.microsoft.com 1; spf=pass smtp.mailfrom=intel.com; dmarc=pass action=none header.from=intel.com; dkim=pass header.d=intel.com; arc=none Authentication-Results: dkim=none (message not signed) header.d=none;dmarc=none action=none header.from=intel.com; Received: from DM5PR11MB1897.namprd11.prod.outlook.com (2603:10b6:3:112::9) by BL0PR11MB3233.namprd11.prod.outlook.com (2603:10b6:208:68::17) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.5293.13; Wed, 1 Jun 2022 08:19:07 +0000 Received: from DM5PR11MB1897.namprd11.prod.outlook.com ([fe80::29a0:67ad:4b77:ce13]) by DM5PR11MB1897.namprd11.prod.outlook.com ([fe80::29a0:67ad:4b77:ce13%8]) with mapi id 15.20.5293.019; Wed, 1 Jun 2022 08:19:07 +0000 Date: Wed, 1 Jun 2022 13:48:56 +0530 From: Balasubramani Vivekanandan To: Matt Roper , Message-ID: References: <20220520230408.3787166-6-matthew.d.roper@intel.com> <20220523204525.4084150-1-matthew.d.roper@intel.com> Content-Type: text/plain; charset="utf-8" Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: <20220523204525.4084150-1-matthew.d.roper@intel.com> X-ClientProxiedBy: PN3PR01CA0171.INDPRD01.PROD.OUTLOOK.COM (2603:1096:c01:de::14) To DM5PR11MB1897.namprd11.prod.outlook.com (2603:10b6:3:112::9) MIME-Version: 1.0 X-MS-PublicTrafficType: Email X-MS-Office365-Filtering-Correlation-Id: b9a71a2d-bf86-48de-9c3c-08da43a76598 X-MS-TrafficTypeDiagnostic: BL0PR11MB3233:EE_ X-LD-Processed: 46c98d88-e344-4ed4-8496-4ed7712e255d,ExtAddr X-Microsoft-Antispam-PRVS: X-MS-Exchange-SenderADCheck: 1 X-MS-Exchange-AntiSpam-Relay: 0 X-Microsoft-Antispam: BCL:0; X-Microsoft-Antispam-Message-Info: Lng9mQbNSueVCbkCODT6guC9iJbUAJLu83UCSXB2eVNYZnAtH+KjFku/PJq0elK2JerfvLp+BXKhhOP0pF/LdOrEWJt//lAamoicu6U1FnkCNul1lgnJBXeuPa6HbkXnuzOFbwl6zmls3HZUI2b9nxYOFqRQ/6oawv72fyiAIJSn/9k8Pzupu23zuBEm2xQjTyuYHI1c/EpsASBWE7DI8VMRbRHxIkMbucFI3cLv1CfrTm6cI/Alx2lzktZC2oE9IViZ1G6W/GsWGXLcX+U8+Qbq7GfUxmkOcnJ1HyGDB3yjbgT41n/Yj7Ifaqr98ZhTNlNrZ3RI/IO0J0R1GD/GUNljHB/pFSaNCEDdzVymNKCXyRKNs/y+mpiEqBare/+0oMya8coTsPzv2xFwz0K+c6+MV49hcs3rjhPuCifK27jkCY5TEtFFE6u6y1W6u7KWZX3b/CMETVZlcL3f/+2vsDreaDPRGNz5T9w2CI33oK36vi6NdBxy/CNNvAVEo1NVBFzx0pQ6odE99oIE+El5dm/ugGgtE+ebkrD7swsrEiY936ZqnMPXLmZy5lUDA5W0wTpqTR43etOT67UGSgr3bKAOzezyZTeWLzMceT9UgdNZK+tNQbrkwL7n4RG9bwhg3dBXLPBHUWw+lPBvbsqsZw== X-Forefront-Antispam-Report: CIP:255.255.255.255; CTRY:; LANG:en; SCL:1; SRV:; IPV:NLI; SFV:NSPM; H:DM5PR11MB1897.namprd11.prod.outlook.com; PTR:; CAT:NONE; SFS:(13230001)(366004)(186003)(8936002)(38100700002)(8676002)(316002)(4326008)(66476007)(83380400001)(508600001)(6506007)(66946007)(86362001)(6486002)(53546011)(5660300002)(44832011)(30864003)(66556008)(82960400001)(6666004)(26005)(6512007)(2906002); DIR:OUT; SFP:1102; X-MS-Exchange-AntiSpam-MessageData-ChunkCount: 1 X-MS-Exchange-AntiSpam-MessageData-0: =?utf-8?B?MGtoQ1VZR1pDcjBrTXBqM0pGNWc2UnFraVc5dkFUSUR6Z2hLWkhmWUY3cUZO?= =?utf-8?B?TDJzcHMwSE5qdFpPcVM1WlhnM25tODBpWHJpSCtXNEdNaHhNaFNvUUhSY2U4?= =?utf-8?B?YURZWHNyRFlmOXltQk9tSEVEZ3NQMFdib3psTUhEa2RmcmlRMjVUZGExRVdt?= =?utf-8?B?TWV3R2Fjby9PbUtYKzJuZW1STTFHaEZob2NaamRVdlJsUG9EejZoZDAxaHhu?= =?utf-8?B?b3BrZGxrSjErekpLTlJFSkUyUzV0cVdlOFFGSit2VWhuV25vU1dYOWRNT1BY?= =?utf-8?B?QlNpNGNvUGN3anR1b25XeTNYbmpuekRkMlVWY2hudEdNRFVsaWdjVDUvZmYx?= =?utf-8?B?UGVpMlpzWWR2SHJNUE1EbTBPRHNsQVRpYVAwQ2lHT25Vbld5V2szTWdqb3JD?= =?utf-8?B?aWY4Q1NKbzBxS3EyS3FOYlIyWDFPYlhxT3R4Mm92eUltZEtDOXhWaUZwK2pk?= =?utf-8?B?a1lIZmxDWHFUd2VGNXhCS0pVQU5ibTBaNGNNaStWY3dxbSt4Z2kwaWpzQlNZ?= =?utf-8?B?VUoxVTRZUmQ1MytMT1ExOGwwMjJmeUdJSmV1TFQrRDFPUkRoYjU4ZEtweS9t?= =?utf-8?B?VlFndTlTR0RlSnRMMDF3Q2tzREtSUnhYUkJFZzcrcmxuRjFCRlpBekFUaXp2?= =?utf-8?B?cHNOajhaYVdrY21SMFY5clpVQmxlUkM5L2dFUUxpK3hDblZBL09WTnU4blVW?= =?utf-8?B?bkdsUXdDOXIvSVFtVjBOTXZwdFhyTE9xZURTcjl3em12MnZaSjB0N041YWc0?= =?utf-8?B?QitmNk9OMFppK1E2bEpQbkw3cW5aZjhvSDFsQ2ZjeTFnVEoybFZpWEh1RXdR?= =?utf-8?B?dXRiRm5wV3ZUN3pkZnpmU2dBL3J2Z2JjYXNVZHFhYWlUQVZaTUx6NkZ6aGVY?= =?utf-8?B?Mi90ckhncU5QajNJaC9hZW5VRmNtME5DUWwvaW9QK0JJejRPNEMrc1lSMk9N?= =?utf-8?B?U0U5ZS9ZaU5Ydm0rRm13RXdhWmFpeW1oZGlmTUhkcjl0ZVpFbDFaWjhOOHBZ?= =?utf-8?B?cjFXUGFZNXZFbmxrS1dzLzJHVVpmMEtxUE1vZlVGaWFFY0xyeUIwQkdzZm5u?= =?utf-8?B?a1UxOEk4Vm5ka1g5WHBGMFRMSGpmWURIYVo0bTNWaXNMYjA1TWdxSjNYam90?= =?utf-8?B?cXVNbitKcnJmOFpoNnlBUDFPQnlxdGF6THRMYzRLa0I4VlY3ZkpVSzlBRzMw?= =?utf-8?B?VXE1Z0Jmc1ZiWWoyMHg3TSttSXp4bU03NHZiaDRPZmcvK1IwVFY2Z2hhRU9i?= =?utf-8?B?RFpGUGRnMmpVRG51c0ZoV1V0dlpycFFVc3JGbXpSdExjSkhlZ3Mya0lKY2k1?= =?utf-8?B?ak1YMk8rUi8vbkYwTnQrTlRWTE1CczR3allQMnRZM1h6RUVheUZFODZqK1d1?= =?utf-8?B?dUZEeGhKUlYreGh4UFhqVVdISzcxYUV3QUxreE5YY0k4YnBLUExBSHFobHBV?= =?utf-8?B?R25wbExXN1pRSWJKTGNBT3NIQTFuVGlOTzVSSjhMK0d6cldvdThET1hXNVRh?= =?utf-8?B?Q1g1U09yYmVNS2lQa1F6VDFqZ0twSVVLcFJDdmFYYkc4dFBYbFdpaHhtM1Ix?= =?utf-8?B?TXBqbSs5dU9mREtDUkpFZi81YXlVU0hZYXhHQ2FldzViYjNMZkNhQVgrU3dv?= =?utf-8?B?WHBQSGwyaWJBdGg3WXROWndiV2kzNkp1SEE1L0wvSFZZRitSdFlZZkw1elpZ?= =?utf-8?B?NUhld0JxaVBEUmptODhrQVRhdGtRajhYbUJ4YktGRExMc3o5K0xWUllNZVFj?= =?utf-8?B?VDNYN0JQRytVL0Fqd3hLMzhHZ0VyOXdzNlY2dW9tU011Q3hIQkZKbUpyeDVZ?= =?utf-8?B?ejZOeEp4K3U0WEk4WTV2UnJ0S2ZkanJrL2xmRTNmenUrNm05UmFpeVVzUklu?= =?utf-8?B?RzZnOWF5a3FPbVZFNUhuNW5TR3U5ZTNiL29jWGp5dzh4bnVvMEZxZnJ0YlB2?= =?utf-8?B?RGcxZW5HMGJXN0JoK2ZHczNlQ0NHT2JNbU1QWkVkeWtVTDBUY0tkaDMzM1ZF?= =?utf-8?B?OFpmZmJaQmZmTzNIMCthVG9qNWxNeDFBL3BDMUJ6OXVOZWZMT1NNeTdRbm9k?= =?utf-8?B?MGM4MVQ4MlpuazAxMGhTTU9CU3h1dWRmQ2plMXRVTUtLWnFBUlJQN2lKQ1ZS?= =?utf-8?B?c0JmR3FhSjRMczQ2Z0pMcGtBbTlIM21IUFdUd1daQlRlZlp5YjhJRUhGbVZ4?= =?utf-8?B?aFNKeERMZUcvQk1CNjNsUk5oQU5ZaGIxbjAvQndkMG0wbnJodGJWYWM1aUZt?= =?utf-8?B?dUZiUnVONW45VkhYeFR6U1ZRdElQRkF0WDdTOVI3bm9DMjRrREtLcnJFZVBR?= =?utf-8?B?T0dRRXZIWWpIRXhFdUxWMit0ZURNR0V2dS9sdkFrYXREaEFnV3FnMVBRQ3Rt?= =?utf-8?Q?G1J1OyPpFFagRpgxGzuwsttFolJDPK0UYtWy/?= X-MS-Exchange-CrossTenant-Network-Message-Id: b9a71a2d-bf86-48de-9c3c-08da43a76598 X-MS-Exchange-CrossTenant-AuthSource: DM5PR11MB1897.namprd11.prod.outlook.com X-MS-Exchange-CrossTenant-AuthAs: Internal X-MS-Exchange-CrossTenant-OriginalArrivalTime: 01 Jun 2022 08:19:07.4914 (UTC) X-MS-Exchange-CrossTenant-FromEntityHeader: Hosted X-MS-Exchange-CrossTenant-Id: 46c98d88-e344-4ed4-8496-4ed7712e255d X-MS-Exchange-CrossTenant-MailboxType: HOSTED X-MS-Exchange-CrossTenant-UserPrincipalName: oatjKBPB2K+bZ2jPzS55R3fkFe+bxrH+PdUwGhBTJLdd9tQtsAa8X3toOxwVPzefDsSKSxyetZ2OHBeLLXm2Spuy8iTxca0TDrj/YE/hgOAnxtsl4VnwFTbPjjYsv+lt X-MS-Exchange-Transport-CrossTenantHeadersStamped: BL0PR11MB3233 X-OriginatorOrg: intel.com Subject: Re: [Intel-gfx] [PATCH v5 5/6] drm/i915/sseu: Disassociate internal subslice mask representation from uapi X-BeenThere: intel-gfx@lists.freedesktop.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: Intel graphics driver community testing & development List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: dri-devel@lists.freedesktop.org Errors-To: intel-gfx-bounces@lists.freedesktop.org Sender: "Intel-gfx" On 23.05.2022 13:45, Matt Roper wrote: > As with EU masks, it's easier to store subslice/DSS masks internally in > a format that's more natural for the driver to work with, and then only > covert into the u8[] uapi form when the query ioctl is invoked. Since > the hardware design changed significantly with Xe_HP, we'll use a union > to choose between the old "hsw-style" subslice masks or the newer xehp > mask. HSW-style masks will be stored in an array of u8's, indexed by > slice (there's never more than 6 subslices per slice on older > platforms). For Xe_HP and beyond where slices no longer exist, we only > need a single bitmask. However we already know that this mask is > eventually going to grow too large for a simple u64 to hold, so we'll > represent it in a manner that can be operated on by the utilities in > linux/bitmap.h. > > v2: > - Fix typo: BIT(s) -> BIT(ss) in gen9_sseu_device_status() > > v3: > - Eliminate sseu->ss_stride and just calculate the stride while > specifically handling uapi. (Tvrtko) > - Use BITMAP_BITS() macro to refer to size of masks rather than > passing I915_MAX_SS_FUSE_BITS directly. (Tvrtko) > - Report compute/geometry DSS masks separately when dumping Xe_HP SSEU > info. (Tvrtko) > - Restore dropped range checks to intel_sseu_has_subslice(). (Tvrtko) > > v4: > - Make the bitmap size macro check the size of the .xehp field rather > than the containing union. (Tvrtko) > - Don't add GEM_BUG_ON() intel_sseu_has_subslice()'s check for whether > slice or subslice ID exceed sseu->max_[sub]slices; various loops > in the driver are expected to exceed these, so we should just > silently return 'false.' > > Cc: Tvrtko Ursulin > Signed-off-by: Matt Roper > --- > drivers/gpu/drm/i915/gem/i915_gem_context.c | 5 +- > drivers/gpu/drm/i915/gt/intel_engine_cs.c | 4 +- > drivers/gpu/drm/i915/gt/intel_gt.c | 12 +- > drivers/gpu/drm/i915/gt/intel_sseu.c | 261 +++++++++++-------- > drivers/gpu/drm/i915/gt/intel_sseu.h | 76 ++++-- > drivers/gpu/drm/i915/gt/intel_sseu_debugfs.c | 30 +-- > drivers/gpu/drm/i915/gt/intel_workarounds.c | 24 +- > drivers/gpu/drm/i915/i915_getparam.c | 3 +- > drivers/gpu/drm/i915/i915_query.c | 13 +- > 9 files changed, 241 insertions(+), 187 deletions(-) > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_context.c b/drivers/gpu/drm/i915/gem/i915_gem_context.c > index ab4c5ab28e4d..a3bb73f5d53b 100644 > --- a/drivers/gpu/drm/i915/gem/i915_gem_context.c > +++ b/drivers/gpu/drm/i915/gem/i915_gem_context.c > @@ -1875,6 +1875,7 @@ i915_gem_user_to_context_sseu(struct intel_gt *gt, > { > const struct sseu_dev_info *device = >->info.sseu; > struct drm_i915_private *i915 = gt->i915; > + unsigned int dev_subslice_mask = intel_sseu_get_hsw_subslices(device, 0); > > /* No zeros in any field. */ > if (!user->slice_mask || !user->subslice_mask || > @@ -1901,7 +1902,7 @@ i915_gem_user_to_context_sseu(struct intel_gt *gt, > if (user->slice_mask & ~device->slice_mask) > return -EINVAL; > > - if (user->subslice_mask & ~device->subslice_mask[0]) > + if (user->subslice_mask & ~dev_subslice_mask) > return -EINVAL; > > if (user->max_eus_per_subslice > device->max_eus_per_subslice) > @@ -1915,7 +1916,7 @@ i915_gem_user_to_context_sseu(struct intel_gt *gt, > /* Part specific restrictions. */ > if (GRAPHICS_VER(i915) == 11) { > unsigned int hw_s = hweight8(device->slice_mask); > - unsigned int hw_ss_per_s = hweight8(device->subslice_mask[0]); > + unsigned int hw_ss_per_s = hweight8(dev_subslice_mask); > unsigned int req_s = hweight8(context->slice_mask); > unsigned int req_ss = hweight8(context->subslice_mask); > > diff --git a/drivers/gpu/drm/i915/gt/intel_engine_cs.c b/drivers/gpu/drm/i915/gt/intel_engine_cs.c > index 1adbf34c3632..f0acf8518a51 100644 > --- a/drivers/gpu/drm/i915/gt/intel_engine_cs.c > +++ b/drivers/gpu/drm/i915/gt/intel_engine_cs.c > @@ -674,8 +674,8 @@ static void engine_mask_apply_compute_fuses(struct intel_gt *gt) > if (GRAPHICS_VER_FULL(i915) < IP_VER(12, 50)) > return; > > - ccs_mask = intel_slicemask_from_dssmask(intel_sseu_get_compute_subslices(&info->sseu), > - ss_per_ccs); > + ccs_mask = intel_slicemask_from_xehp_dssmask(info->sseu.compute_subslice_mask, > + ss_per_ccs); > /* > * If all DSS in a quadrant are fused off, the corresponding CCS > * engine is not available for use. > diff --git a/drivers/gpu/drm/i915/gt/intel_gt.c b/drivers/gpu/drm/i915/gt/intel_gt.c > index 034182f85501..2921f510642f 100644 > --- a/drivers/gpu/drm/i915/gt/intel_gt.c > +++ b/drivers/gpu/drm/i915/gt/intel_gt.c > @@ -133,13 +133,6 @@ static const struct intel_mmio_range dg2_lncf_steering_table[] = { > {}, > }; > > -static u16 slicemask(struct intel_gt *gt, int count) > -{ > - u64 dss_mask = intel_sseu_get_subslices(>->info.sseu, 0); > - > - return intel_slicemask_from_dssmask(dss_mask, count); > -} > - > int intel_gt_init_mmio(struct intel_gt *gt) > { > struct drm_i915_private *i915 = gt->i915; > @@ -155,9 +148,12 @@ int intel_gt_init_mmio(struct intel_gt *gt) > */ > if (HAS_MSLICES(i915)) { > gt->info.mslice_mask = > - slicemask(gt, GEN_DSS_PER_MSLICE) | > + intel_slicemask_from_xehp_dssmask(gt->info.sseu.subslice_mask, > + GEN_DSS_PER_MSLICE); > + gt->info.mslice_mask |= > (intel_uncore_read(gt->uncore, GEN10_MIRROR_FUSE3) & > GEN12_MEML3_EN_MASK); > + > if (!gt->info.mslice_mask) /* should be impossible! */ > drm_warn(&i915->drm, "mslice mask all zero!\n"); > } > diff --git a/drivers/gpu/drm/i915/gt/intel_sseu.c b/drivers/gpu/drm/i915/gt/intel_sseu.c > index 285cfd758bdc..3a721db20290 100644 > --- a/drivers/gpu/drm/i915/gt/intel_sseu.c > +++ b/drivers/gpu/drm/i915/gt/intel_sseu.c > @@ -10,15 +10,14 @@ > #include "intel_gt_regs.h" > #include "intel_sseu.h" > > +#define XEHP_BITMAP_BITS(mask) ((int)BITS_PER_TYPE(typeof(mask.xehp))) > + > void intel_sseu_set_info(struct sseu_dev_info *sseu, u8 max_slices, > u8 max_subslices, u8 max_eus_per_subslice) > { > sseu->max_slices = max_slices; > sseu->max_subslices = max_subslices; > sseu->max_eus_per_subslice = max_eus_per_subslice; > - > - sseu->ss_stride = GEN_SSEU_STRIDE(sseu->max_subslices); > - GEM_BUG_ON(sseu->ss_stride > GEN_MAX_SUBSLICE_STRIDE); > } > > unsigned int > @@ -26,54 +25,24 @@ intel_sseu_subslice_total(const struct sseu_dev_info *sseu) > { > unsigned int i, total = 0; > > - for (i = 0; i < ARRAY_SIZE(sseu->subslice_mask); i++) > - total += hweight8(sseu->subslice_mask[i]); > - > - return total; > -} > - > -static u32 > -sseu_get_subslices(const struct sseu_dev_info *sseu, > - const u8 *subslice_mask, u8 slice) > -{ > - int i, offset = slice * sseu->ss_stride; > - u32 mask = 0; > - > - GEM_BUG_ON(slice >= sseu->max_slices); > - > - for (i = 0; i < sseu->ss_stride; i++) > - mask |= (u32)subslice_mask[offset + i] << i * BITS_PER_BYTE; > - > - return mask; > -} > - > -u32 intel_sseu_get_subslices(const struct sseu_dev_info *sseu, u8 slice) > -{ > - return sseu_get_subslices(sseu, sseu->subslice_mask, slice); > -} > + if (sseu->has_xehp_dss) > + return bitmap_weight(sseu->subslice_mask.xehp, > + XEHP_BITMAP_BITS(sseu->subslice_mask)); > > -static u32 sseu_get_geometry_subslices(const struct sseu_dev_info *sseu) > -{ > - return sseu_get_subslices(sseu, sseu->geometry_subslice_mask, 0); > -} > + for (i = 0; i < ARRAY_SIZE(sseu->subslice_mask.hsw); i++) > + total += hweight8(sseu->subslice_mask.hsw[i]); > > -u32 intel_sseu_get_compute_subslices(const struct sseu_dev_info *sseu) > -{ > - return sseu_get_subslices(sseu, sseu->compute_subslice_mask, 0); > -} > - > -void intel_sseu_set_subslices(struct sseu_dev_info *sseu, int slice, > - u8 *subslice_mask, u32 ss_mask) > -{ > - int offset = slice * sseu->ss_stride; > - > - memcpy(&subslice_mask[offset], &ss_mask, sseu->ss_stride); > + return total; > } > > unsigned int > -intel_sseu_subslices_per_slice(const struct sseu_dev_info *sseu, u8 slice) > +intel_sseu_get_hsw_subslices(const struct sseu_dev_info *sseu, u8 slice) > { > - return hweight32(intel_sseu_get_subslices(sseu, slice)); > + WARN_ON(sseu->has_xehp_dss); > + if (WARN_ON(slice >= sseu->max_slices)) > + return 0; > + > + return sseu->subslice_mask.hsw[slice]; > } > > static u16 sseu_get_eus(const struct sseu_dev_info *sseu, int slice, > @@ -147,32 +116,66 @@ int intel_sseu_copy_eumask_to_user(void __user *to, > return copy_to_user(to, eu_mask, len); > } > > +/** > + * intel_sseu_copy_ssmask_to_user - Copy subslice mask into a userspace buffer > + * @to: Pointer to userspace buffer to copy to > + * @sseu: SSEU structure containing subslice mask to copy > + * > + * Copies the subslice mask to a userspace buffer in the format expected by > + * the query ioctl's topology queries. > + * > + * Returns the result of the copy_to_user() operation. > + */ > +int intel_sseu_copy_ssmask_to_user(void __user *to, > + const struct sseu_dev_info *sseu) > +{ > + u8 ss_mask[GEN_SS_MASK_SIZE] = {}; > + int ss_stride = GEN_SSEU_STRIDE(sseu->max_subslices); > + int len = sseu->max_slices * ss_stride; > + int s, ss, i; > + > + for (s = 0; s < sseu->max_slices; s++) { > + for (ss = 0; ss < sseu->max_subslices; ss++) { > + i = s * ss_stride * BITS_PER_BYTE + ss; > + > + if (!intel_sseu_has_subslice(sseu, s, ss)) > + continue; > + > + ss_mask[i / BITS_PER_BYTE] |= BIT(i % BITS_PER_BYTE); > + } > + } > + > + return copy_to_user(to, ss_mask, len); > +} > + > static void gen11_compute_sseu_info(struct sseu_dev_info *sseu, > - u32 g_ss_en, u32 c_ss_en, u16 eu_en) > + u32 ss_en, u16 eu_en) > { > u32 valid_ss_mask = GENMASK(sseu->max_subslices - 1, 0); > int ss; > > - /* g_ss_en/c_ss_en represent entire subslice mask across all slices */ > - GEM_BUG_ON(sseu->max_slices * sseu->max_subslices > > - sizeof(g_ss_en) * BITS_PER_BYTE); > + sseu->slice_mask |= BIT(0); > + sseu->subslice_mask.hsw[0] = ss_en & valid_ss_mask; > + > + for (ss = 0; ss < sseu->max_subslices; ss++) > + if (intel_sseu_has_subslice(sseu, 0, ss)) > + sseu_set_eus(sseu, 0, ss, eu_en); > + > + sseu->eu_per_subslice = hweight16(eu_en); > + sseu->eu_total = compute_eu_total(sseu); > +} > + > +static void xehp_compute_sseu_info(struct sseu_dev_info *sseu, > + u16 eu_en) > +{ > + int ss; > > sseu->slice_mask |= BIT(0); > > - /* > - * XeHP introduces the concept of compute vs geometry DSS. To reduce > - * variation between GENs around subslice usage, store a mask for both > - * the geometry and compute enabled masks since userspace will need to > - * be able to query these masks independently. Also compute a total > - * enabled subslice count for the purposes of selecting subslices to > - * use in a particular GEM context. > - */ > - intel_sseu_set_subslices(sseu, 0, sseu->compute_subslice_mask, > - c_ss_en & valid_ss_mask); > - intel_sseu_set_subslices(sseu, 0, sseu->geometry_subslice_mask, > - g_ss_en & valid_ss_mask); > - intel_sseu_set_subslices(sseu, 0, sseu->subslice_mask, > - (g_ss_en | c_ss_en) & valid_ss_mask); > + bitmap_or(sseu->subslice_mask.xehp, > + sseu->compute_subslice_mask.xehp, > + sseu->geometry_subslice_mask.xehp, > + XEHP_BITMAP_BITS(sseu->subslice_mask)); > > for (ss = 0; ss < sseu->max_subslices; ss++) > if (intel_sseu_has_subslice(sseu, 0, ss)) > @@ -182,11 +185,31 @@ static void gen11_compute_sseu_info(struct sseu_dev_info *sseu, > sseu->eu_total = compute_eu_total(sseu); > } > > +static void > +xehp_load_dss_mask(struct intel_uncore *uncore, > + intel_sseu_ss_mask_t *ssmask, > + int numregs, > + ...) > +{ > + va_list argp; > + u32 fuse_val[I915_MAX_SS_FUSE_REGS] = {}; > + int i; > + > + if (WARN_ON(numregs > I915_MAX_SS_FUSE_REGS)) > + numregs = I915_MAX_SS_FUSE_REGS; > + > + va_start(argp, numregs); > + for (i = 0; i < numregs; i++) > + fuse_val[i] = intel_uncore_read(uncore, va_arg(argp, i915_reg_t)); > + va_end(argp); > + > + bitmap_from_arr32(ssmask->xehp, fuse_val, numregs * 32); > +} > + > static void xehp_sseu_info_init(struct intel_gt *gt) > { > struct sseu_dev_info *sseu = >->info.sseu; > struct intel_uncore *uncore = gt->uncore; > - u32 g_dss_en, c_dss_en = 0; > u16 eu_en = 0; > u8 eu_en_fuse; > int eu; > @@ -200,8 +223,10 @@ static void xehp_sseu_info_init(struct intel_gt *gt) > intel_sseu_set_info(sseu, 1, 32, 16); > sseu->has_xehp_dss = 1; > > - g_dss_en = intel_uncore_read(uncore, GEN12_GT_GEOMETRY_DSS_ENABLE); > - c_dss_en = intel_uncore_read(uncore, GEN12_GT_COMPUTE_DSS_ENABLE); > + xehp_load_dss_mask(uncore, &sseu->geometry_subslice_mask, 1, > + GEN12_GT_GEOMETRY_DSS_ENABLE); > + xehp_load_dss_mask(uncore, &sseu->compute_subslice_mask, 1, > + GEN12_GT_COMPUTE_DSS_ENABLE); > > eu_en_fuse = intel_uncore_read(uncore, XEHP_EU_ENABLE) & XEHP_EU_ENA_MASK; > > @@ -209,7 +234,7 @@ static void xehp_sseu_info_init(struct intel_gt *gt) > if (eu_en_fuse & BIT(eu)) > eu_en |= BIT(eu * 2) | BIT(eu * 2 + 1); > > - gen11_compute_sseu_info(sseu, g_dss_en, c_dss_en, eu_en); > + xehp_compute_sseu_info(sseu, eu_en); > } > > static void gen12_sseu_info_init(struct intel_gt *gt) > @@ -247,7 +272,7 @@ static void gen12_sseu_info_init(struct intel_gt *gt) > if (eu_en_fuse & BIT(eu)) > eu_en |= BIT(eu * 2) | BIT(eu * 2 + 1); > > - gen11_compute_sseu_info(sseu, g_dss_en, 0, eu_en); > + gen11_compute_sseu_info(sseu, g_dss_en, eu_en); > > /* TGL only supports slice-level power gating */ > sseu->has_slice_pg = 1; > @@ -279,7 +304,7 @@ static void gen11_sseu_info_init(struct intel_gt *gt) > eu_en = ~(intel_uncore_read(uncore, GEN11_EU_DISABLE) & > GEN11_EU_DIS_MASK); > > - gen11_compute_sseu_info(sseu, ss_en, 0, eu_en); > + gen11_compute_sseu_info(sseu, ss_en, eu_en); > > /* ICL has no power gating restrictions. */ > sseu->has_slice_pg = 1; > @@ -291,7 +316,6 @@ static void cherryview_sseu_info_init(struct intel_gt *gt) > { > struct sseu_dev_info *sseu = >->info.sseu; > u32 fuse; > - u8 subslice_mask = 0; > > fuse = intel_uncore_read(gt->uncore, CHV_FUSE_GT); > > @@ -305,7 +329,7 @@ static void cherryview_sseu_info_init(struct intel_gt *gt) > (((fuse & CHV_FGT_EU_DIS_SS0_R1_MASK) >> > CHV_FGT_EU_DIS_SS0_R1_SHIFT) << 4); > > - subslice_mask |= BIT(0); > + sseu->subslice_mask.hsw[0] |= BIT(0); > sseu_set_eus(sseu, 0, 0, ~disabled_mask & 0xFF); > } > > @@ -316,12 +340,10 @@ static void cherryview_sseu_info_init(struct intel_gt *gt) > (((fuse & CHV_FGT_EU_DIS_SS1_R1_MASK) >> > CHV_FGT_EU_DIS_SS1_R1_SHIFT) << 4); > > - subslice_mask |= BIT(1); > + sseu->subslice_mask.hsw[0] |= BIT(1); > sseu_set_eus(sseu, 0, 1, ~disabled_mask & 0xFF); > } > > - intel_sseu_set_subslices(sseu, 0, sseu->subslice_mask, subslice_mask); > - > sseu->eu_total = compute_eu_total(sseu); > > /* > @@ -376,8 +398,7 @@ static void gen9_sseu_info_init(struct intel_gt *gt) > /* skip disabled slice */ > continue; > > - intel_sseu_set_subslices(sseu, s, sseu->subslice_mask, > - subslice_mask); > + sseu->subslice_mask.hsw[s] = subslice_mask; > > eu_disable = intel_uncore_read(uncore, GEN9_EU_DISABLE(s)); > for (ss = 0; ss < sseu->max_subslices; ss++) { > @@ -434,8 +455,8 @@ static void gen9_sseu_info_init(struct intel_gt *gt) > sseu->has_eu_pg = sseu->eu_per_subslice > 2; > > if (IS_GEN9_LP(i915)) { > -#define IS_SS_DISABLED(ss) (!(sseu->subslice_mask[0] & BIT(ss))) > - info->has_pooled_eu = hweight8(sseu->subslice_mask[0]) == 3; > +#define IS_SS_DISABLED(ss) (!(sseu->subslice_mask.hsw[0] & BIT(ss))) > + info->has_pooled_eu = hweight8(sseu->subslice_mask.hsw[0]) == 3; > > sseu->min_eu_in_pool = 0; > if (info->has_pooled_eu) { > @@ -489,8 +510,7 @@ static void bdw_sseu_info_init(struct intel_gt *gt) > /* skip disabled slice */ > continue; > > - intel_sseu_set_subslices(sseu, s, sseu->subslice_mask, > - subslice_mask); > + sseu->subslice_mask.hsw[s] = subslice_mask; > > for (ss = 0; ss < sseu->max_subslices; ss++) { > u8 eu_disabled_mask; > @@ -587,8 +607,7 @@ static void hsw_sseu_info_init(struct intel_gt *gt) > sseu->eu_per_subslice); > > for (s = 0; s < sseu->max_slices; s++) { > - intel_sseu_set_subslices(sseu, s, sseu->subslice_mask, > - subslice_mask); > + sseu->subslice_mask.hsw[s] = subslice_mask; > > for (ss = 0; ss < sseu->max_subslices; ss++) { > sseu_set_eus(sseu, s, ss, > @@ -677,7 +696,7 @@ u32 intel_sseu_make_rpcs(struct intel_gt *gt, > */ > if (GRAPHICS_VER(i915) == 11 && > slices == 1 && > - subslices > min_t(u8, 4, hweight8(sseu->subslice_mask[0]) / 2)) { > + subslices > min_t(u8, 4, hweight8(sseu->subslice_mask.hsw[0]) / 2)) { > GEM_BUG_ON(subslices & 1); > > subslice_pg = false; > @@ -743,14 +762,29 @@ void intel_sseu_dump(const struct sseu_dev_info *sseu, struct drm_printer *p) > { > int s; > > - drm_printf(p, "slice total: %u, mask=%04x\n", > - hweight8(sseu->slice_mask), sseu->slice_mask); > - drm_printf(p, "subslice total: %u\n", intel_sseu_subslice_total(sseu)); > - for (s = 0; s < sseu->max_slices; s++) { > - drm_printf(p, "slice%d: %u subslices, mask=%08x\n", > - s, intel_sseu_subslices_per_slice(sseu, s), > - intel_sseu_get_subslices(sseu, s)); > + if (sseu->has_xehp_dss) { > + drm_printf(p, "subslice total: %u\n", > + intel_sseu_subslice_total(sseu)); > + drm_printf(p, "geometry dss mask=%*pb\n", > + XEHP_BITMAP_BITS(sseu->geometry_subslice_mask), > + sseu->geometry_subslice_mask.xehp); > + drm_printf(p, "compute dss mask=%*pb\n", > + XEHP_BITMAP_BITS(sseu->compute_subslice_mask), > + sseu->compute_subslice_mask.xehp); > + } else { > + drm_printf(p, "slice total: %u, mask=%04x\n", > + hweight8(sseu->slice_mask), sseu->slice_mask); > + drm_printf(p, "subslice total: %u\n", > + intel_sseu_subslice_total(sseu)); > + > + for (s = 0; s < sseu->max_slices; s++) { > + u8 ss_mask = sseu->subslice_mask.hsw[s]; > + > + drm_printf(p, "slice%d: %u subslices, mask=%08x\n", > + s, hweight8(ss_mask), ss_mask); > + } > } > + > drm_printf(p, "EU total: %u\n", sseu->eu_total); > drm_printf(p, "EU per subslice: %u\n", sseu->eu_per_subslice); > drm_printf(p, "has slice power gating: %s\n", > @@ -767,9 +801,10 @@ static void sseu_print_hsw_topology(const struct sseu_dev_info *sseu, > int s, ss; > > for (s = 0; s < sseu->max_slices; s++) { > + u8 ss_mask = sseu->subslice_mask.hsw[s]; > + > drm_printf(p, "slice%d: %u subslice(s) (0x%08x):\n", > - s, intel_sseu_subslices_per_slice(sseu, s), > - intel_sseu_get_subslices(sseu, s)); > + s, hweight8(ss_mask), ss_mask); > > for (ss = 0; ss < sseu->max_subslices; ss++) { > u16 enabled_eus = sseu_get_eus(sseu, s, ss); > @@ -783,16 +818,14 @@ static void sseu_print_hsw_topology(const struct sseu_dev_info *sseu, > static void sseu_print_xehp_topology(const struct sseu_dev_info *sseu, > struct drm_printer *p) > { > - u32 g_dss_mask = sseu_get_geometry_subslices(sseu); > - u32 c_dss_mask = intel_sseu_get_compute_subslices(sseu); > int dss; > > for (dss = 0; dss < sseu->max_subslices; dss++) { > u16 enabled_eus = sseu_get_eus(sseu, 0, dss); > > drm_printf(p, "DSS_%02d: G:%3s C:%3s, %2u EUs (0x%04hx)\n", dss, > - str_yes_no(g_dss_mask & BIT(dss)), > - str_yes_no(c_dss_mask & BIT(dss)), > + str_yes_no(test_bit(dss, sseu->geometry_subslice_mask.xehp)), > + str_yes_no(test_bit(dss, sseu->compute_subslice_mask.xehp)), > hweight16(enabled_eus), enabled_eus); > } > } > @@ -810,20 +843,44 @@ void intel_sseu_print_topology(struct drm_i915_private *i915, > } > } > > -u16 intel_slicemask_from_dssmask(u64 dss_mask, int dss_per_slice) > +void intel_sseu_print_ss_info(const char* type, > + const struct sseu_dev_info *sseu, > + struct seq_file *m) > +{ > + int s; > + > + if (sseu->has_xehp_dss) { > + seq_printf(m, " %s Geometry DSS: %u\n", type, > + bitmap_weight(sseu->geometry_subslice_mask.xehp, > + XEHP_BITMAP_BITS(sseu->geometry_subslice_mask))); > + seq_printf(m, " %s Compute DSS: %u\n", type, > + bitmap_weight(sseu->compute_subslice_mask.xehp, > + XEHP_BITMAP_BITS(sseu->compute_subslice_mask))); > + } else { > + for (s = 0; s < fls(sseu->slice_mask); s++) > + seq_printf(m, " %s Slice%i subslices: %u\n", type, > + s, hweight8(sseu->subslice_mask.hsw[s])); > + } > +} > + > +u16 intel_slicemask_from_xehp_dssmask(intel_sseu_ss_mask_t dss_mask, > + int dss_per_slice) > { > + intel_sseu_ss_mask_t per_slice_mask = {}; > u16 slice_mask = 0; > int i; > > - WARN_ON(sizeof(dss_mask) * 8 / dss_per_slice > 8 * sizeof(slice_mask)); > + WARN_ON(DIV_ROUND_UP(XEHP_BITMAP_BITS(dss_mask), dss_per_slice) > > + 8 * sizeof(slice_mask)); > > - for (i = 0; dss_mask; i++) { > - if (dss_mask & GENMASK(dss_per_slice - 1, 0)) > + bitmap_fill(per_slice_mask.xehp, dss_per_slice); > + for (i = 0; !bitmap_empty(dss_mask.xehp, XEHP_BITMAP_BITS(dss_mask)); i++) { > + if (bitmap_intersects(dss_mask.xehp, per_slice_mask.xehp, dss_per_slice)) > slice_mask |= BIT(i); > > - dss_mask >>= dss_per_slice; > + bitmap_shift_right(dss_mask.xehp, dss_mask.xehp, dss_per_slice, > + XEHP_BITMAP_BITS(dss_mask)); > } > > return slice_mask; > } > - > diff --git a/drivers/gpu/drm/i915/gt/intel_sseu.h b/drivers/gpu/drm/i915/gt/intel_sseu.h > index ffa375e68959..e7f336a7ffbb 100644 > --- a/drivers/gpu/drm/i915/gt/intel_sseu.h > +++ b/drivers/gpu/drm/i915/gt/intel_sseu.h > @@ -25,12 +25,16 @@ struct drm_printer; > /* > * Maximum number of subslices that can exist within a HSW-style slice. This > * is only relevant to pre-Xe_HP platforms (Xe_HP and beyond use the > - * GEN_MAX_DSS value below). > + * I915_MAX_SS_FUSE_BITS value below). > */ > #define GEN_MAX_SS_PER_HSW_SLICE 6 > > -/* Maximum number of DSS on newer platforms (Xe_HP and beyond). */ > -#define GEN_MAX_DSS 32 > +/* > + * Maximum number of 32-bit registers used by hardware to express the > + * enabled/disabled subslices. > + */ > +#define I915_MAX_SS_FUSE_REGS 1 > +#define I915_MAX_SS_FUSE_BITS (I915_MAX_SS_FUSE_REGS * 32) > > /* Maximum number of EUs that can exist within a subslice or DSS. */ > #define GEN_MAX_EUS_PER_SS 16 > @@ -38,7 +42,7 @@ struct drm_printer; > #define SSEU_MAX(a, b) ((a) > (b) ? (a) : (b)) > > /* The maximum number of bits needed to express each subslice/DSS independently */ > -#define GEN_SS_MASK_SIZE SSEU_MAX(GEN_MAX_DSS, \ > +#define GEN_SS_MASK_SIZE SSEU_MAX(I915_MAX_SS_FUSE_BITS, \ > GEN_MAX_HSW_SLICES * GEN_MAX_SS_PER_HSW_SLICE) > > #define GEN_SSEU_STRIDE(max_entries) DIV_ROUND_UP(max_entries, BITS_PER_BYTE) > @@ -49,17 +53,24 @@ struct drm_printer; > #define GEN_DSS_PER_CSLICE 8 > #define GEN_DSS_PER_MSLICE 8 > > -#define GEN_MAX_GSLICES (GEN_MAX_DSS / GEN_DSS_PER_GSLICE) > -#define GEN_MAX_CSLICES (GEN_MAX_DSS / GEN_DSS_PER_CSLICE) > +#define GEN_MAX_GSLICES (I915_MAX_SS_FUSE_BITS / GEN_DSS_PER_GSLICE) > +#define GEN_MAX_CSLICES (I915_MAX_SS_FUSE_BITS / GEN_DSS_PER_CSLICE) > + > +typedef union { > + u8 hsw[GEN_MAX_HSW_SLICES]; > + > + /* Bitmap compatible with linux/bitmap.h; may exceed size of u64 */ > + unsigned long xehp[BITS_TO_LONGS(I915_MAX_SS_FUSE_BITS)]; > +} intel_sseu_ss_mask_t; > > struct sseu_dev_info { > u8 slice_mask; > - u8 subslice_mask[GEN_SS_MASK_SIZE]; > - u8 geometry_subslice_mask[GEN_SS_MASK_SIZE]; > - u8 compute_subslice_mask[GEN_SS_MASK_SIZE]; > + intel_sseu_ss_mask_t subslice_mask; > + intel_sseu_ss_mask_t geometry_subslice_mask; > + intel_sseu_ss_mask_t compute_subslice_mask; > union { > u16 hsw[GEN_MAX_HSW_SLICES][GEN_MAX_SS_PER_HSW_SLICE]; > - u16 xehp[GEN_MAX_DSS]; > + u16 xehp[I915_MAX_SS_FUSE_BITS]; > } eu_mask; > > u16 eu_total; > @@ -80,8 +91,6 @@ struct sseu_dev_info { > u8 max_slices; > u8 max_subslices; > u8 max_eus_per_subslice; > - > - u8 ss_stride; > }; > > /* > @@ -99,7 +108,7 @@ intel_sseu_from_device_info(const struct sseu_dev_info *sseu) > { > struct intel_sseu value = { > .slice_mask = sseu->slice_mask, > - .subslice_mask = sseu->subslice_mask[0], > + .subslice_mask = sseu->subslice_mask.hsw[0], > .min_eus_per_subslice = sseu->max_eus_per_subslice, > .max_eus_per_subslice = sseu->max_eus_per_subslice, > }; > @@ -111,18 +120,27 @@ static inline bool > intel_sseu_has_subslice(const struct sseu_dev_info *sseu, int slice, > int subslice) > { > - u8 mask; > - int ss_idx = subslice / BITS_PER_BYTE; > - > if (slice >= sseu->max_slices || > subslice >= sseu->max_subslices) > return false; > > - GEM_BUG_ON(ss_idx >= sseu->ss_stride); > - > - mask = sseu->subslice_mask[slice * sseu->ss_stride + ss_idx]; > + if (sseu->has_xehp_dss) > + return test_bit(subslice, sseu->subslice_mask.xehp); > + else > + return sseu->subslice_mask.hsw[slice] & BIT(subslice); > +} > > - return mask & BIT(subslice % BITS_PER_BYTE); > +/* > + * Used to obtain the index of the first DSS. Can start searching from the > + * beginning of a specific dss group (e.g., gslice, cslice, etc.) if > + * groupsize and groupnum are non-zero. > + */ > +static inline unsigned int > +intel_sseu_find_first_xehp_dss(const struct sseu_dev_info *sseu, int groupsize, > + int groupnum) > +{ > + return find_next_bit(sseu->subslice_mask.xehp, I915_MAX_SS_FUSE_BITS, > + groupnum * groupsize); Any specific reason for not replacing I915_MAX_SS_FUSE_BITS with XEHP_BITMAP_BITS here? Regards, Bala > } > > void intel_sseu_set_info(struct sseu_dev_info *sseu, u8 max_slices, > @@ -132,14 +150,10 @@ unsigned int > intel_sseu_subslice_total(const struct sseu_dev_info *sseu); > > unsigned int > -intel_sseu_subslices_per_slice(const struct sseu_dev_info *sseu, u8 slice); > - > -u32 intel_sseu_get_subslices(const struct sseu_dev_info *sseu, u8 slice); > +intel_sseu_get_hsw_subslices(const struct sseu_dev_info *sseu, u8 slice); > > -u32 intel_sseu_get_compute_subslices(const struct sseu_dev_info *sseu); > - > -void intel_sseu_set_subslices(struct sseu_dev_info *sseu, int slice, > - u8 *subslice_mask, u32 ss_mask); > +intel_sseu_ss_mask_t > +intel_sseu_get_compute_subslices(const struct sseu_dev_info *sseu); > > void intel_sseu_info_init(struct intel_gt *gt); > > @@ -151,9 +165,15 @@ void intel_sseu_print_topology(struct drm_i915_private *i915, > const struct sseu_dev_info *sseu, > struct drm_printer *p); > > -u16 intel_slicemask_from_dssmask(u64 dss_mask, int dss_per_slice); > +u16 intel_slicemask_from_xehp_dssmask(intel_sseu_ss_mask_t dss_mask, int dss_per_slice); > > int intel_sseu_copy_eumask_to_user(void __user *to, > const struct sseu_dev_info *sseu); > +int intel_sseu_copy_ssmask_to_user(void __user *to, > + const struct sseu_dev_info *sseu); > + > +void intel_sseu_print_ss_info(const char* type, > + const struct sseu_dev_info *sseu, > + struct seq_file *m); > > #endif /* __INTEL_SSEU_H__ */ > diff --git a/drivers/gpu/drm/i915/gt/intel_sseu_debugfs.c b/drivers/gpu/drm/i915/gt/intel_sseu_debugfs.c > index 2d5d011e01db..c2ee5e1826b5 100644 > --- a/drivers/gpu/drm/i915/gt/intel_sseu_debugfs.c > +++ b/drivers/gpu/drm/i915/gt/intel_sseu_debugfs.c > @@ -4,6 +4,7 @@ > * Copyright © 2020 Intel Corporation > */ > > +#include > #include > > #include "i915_drv.h" > @@ -11,14 +12,6 @@ > #include "intel_gt_regs.h" > #include "intel_sseu_debugfs.h" > > -static void sseu_copy_subslices(const struct sseu_dev_info *sseu, > - int slice, u8 *to_mask) > -{ > - int offset = slice * sseu->ss_stride; > - > - memcpy(&to_mask[offset], &sseu->subslice_mask[offset], sseu->ss_stride); > -} > - > static void cherryview_sseu_device_status(struct intel_gt *gt, > struct sseu_dev_info *sseu) > { > @@ -41,7 +34,7 @@ static void cherryview_sseu_device_status(struct intel_gt *gt, > continue; > > sseu->slice_mask = BIT(0); > - sseu->subslice_mask[0] |= BIT(ss); > + sseu->subslice_mask.hsw[0] |= BIT(ss); > eu_cnt = ((sig1[ss] & CHV_EU08_PG_ENABLE) ? 0 : 2) + > ((sig1[ss] & CHV_EU19_PG_ENABLE) ? 0 : 2) + > ((sig1[ss] & CHV_EU210_PG_ENABLE) ? 0 : 2) + > @@ -92,7 +85,7 @@ static void gen11_sseu_device_status(struct intel_gt *gt, > continue; > > sseu->slice_mask |= BIT(s); > - sseu_copy_subslices(&info->sseu, s, sseu->subslice_mask); > + sseu->subslice_mask.hsw[s] = info->sseu.subslice_mask.hsw[s]; > > for (ss = 0; ss < info->sseu.max_subslices; ss++) { > unsigned int eu_cnt; > @@ -147,21 +140,17 @@ static void gen9_sseu_device_status(struct intel_gt *gt, > sseu->slice_mask |= BIT(s); > > if (IS_GEN9_BC(gt->i915)) > - sseu_copy_subslices(&info->sseu, s, > - sseu->subslice_mask); > + sseu->subslice_mask.hsw[s] = info->sseu.subslice_mask.hsw[s]; > > for (ss = 0; ss < info->sseu.max_subslices; ss++) { > unsigned int eu_cnt; > - u8 ss_idx = s * info->sseu.ss_stride + > - ss / BITS_PER_BYTE; > > if (IS_GEN9_LP(gt->i915)) { > if (!(s_reg[s] & (GEN9_PGCTL_SS_ACK(ss)))) > /* skip disabled subslice */ > continue; > > - sseu->subslice_mask[ss_idx] |= > - BIT(ss % BITS_PER_BYTE); > + sseu->subslice_mask.hsw[s] |= BIT(ss); > } > > eu_cnt = eu_reg[2 * s + ss / 2] & eu_mask[ss % 2]; > @@ -188,8 +177,7 @@ static void bdw_sseu_device_status(struct intel_gt *gt, > if (sseu->slice_mask) { > sseu->eu_per_subslice = info->sseu.eu_per_subslice; > for (s = 0; s < fls(sseu->slice_mask); s++) > - sseu_copy_subslices(&info->sseu, s, > - sseu->subslice_mask); > + sseu->subslice_mask.hsw[s] = info->sseu.subslice_mask.hsw[s]; > sseu->eu_total = sseu->eu_per_subslice * > intel_sseu_subslice_total(sseu); > > @@ -208,7 +196,6 @@ static void i915_print_sseu_info(struct seq_file *m, > const struct sseu_dev_info *sseu) > { > const char *type = is_available_info ? "Available" : "Enabled"; > - int s; > > seq_printf(m, " %s Slice Mask: %04x\n", type, > sseu->slice_mask); > @@ -216,10 +203,7 @@ static void i915_print_sseu_info(struct seq_file *m, > hweight8(sseu->slice_mask)); > seq_printf(m, " %s Subslice Total: %u\n", type, > intel_sseu_subslice_total(sseu)); > - for (s = 0; s < fls(sseu->slice_mask); s++) { > - seq_printf(m, " %s Slice%i subslices: %u\n", type, > - s, intel_sseu_subslices_per_slice(sseu, s)); > - } > + intel_sseu_print_ss_info(type, sseu, m); > seq_printf(m, " %s EU Total: %u\n", type, > sseu->eu_total); > seq_printf(m, " %s EU Per Subslice: %u\n", type, > diff --git a/drivers/gpu/drm/i915/gt/intel_workarounds.c b/drivers/gpu/drm/i915/gt/intel_workarounds.c > index 73b59ea6fd3b..1bf62273e02d 100644 > --- a/drivers/gpu/drm/i915/gt/intel_workarounds.c > +++ b/drivers/gpu/drm/i915/gt/intel_workarounds.c > @@ -948,8 +948,8 @@ gen9_wa_init_mcr(struct drm_i915_private *i915, struct i915_wa_list *wal) > * on s/ss combo, the read should be done with read_subslice_reg. > */ > slice = ffs(sseu->slice_mask) - 1; > - GEM_BUG_ON(slice >= ARRAY_SIZE(sseu->subslice_mask)); > - subslice = ffs(intel_sseu_get_subslices(sseu, slice)); > + GEM_BUG_ON(slice >= ARRAY_SIZE(sseu->subslice_mask.hsw)); > + subslice = ffs(intel_sseu_get_hsw_subslices(sseu, slice)); > GEM_BUG_ON(!subslice); > subslice--; > > @@ -1087,11 +1087,10 @@ static void > icl_wa_init_mcr(struct intel_gt *gt, struct i915_wa_list *wal) > { > const struct sseu_dev_info *sseu = >->info.sseu; > - unsigned int slice, subslice; > + unsigned int subslice; > > GEM_BUG_ON(GRAPHICS_VER(gt->i915) < 11); > GEM_BUG_ON(hweight8(sseu->slice_mask) > 1); > - slice = 0; > > /* > * Although a platform may have subslices, we need to always steer > @@ -1102,7 +1101,7 @@ icl_wa_init_mcr(struct intel_gt *gt, struct i915_wa_list *wal) > * one of the higher subslices, we run the risk of reading back 0's or > * random garbage. > */ > - subslice = __ffs(intel_sseu_get_subslices(sseu, slice)); > + subslice = __ffs(intel_sseu_get_hsw_subslices(sseu, 0)); > > /* > * If the subslice we picked above also steers us to a valid L3 bank, > @@ -1112,7 +1111,7 @@ icl_wa_init_mcr(struct intel_gt *gt, struct i915_wa_list *wal) > if (gt->info.l3bank_mask & BIT(subslice)) > gt->steering_table[L3BANK] = NULL; > > - __add_mcr_wa(gt, wal, slice, subslice); > + __add_mcr_wa(gt, wal, 0, subslice); > } > > static void > @@ -1120,7 +1119,6 @@ xehp_init_mcr(struct intel_gt *gt, struct i915_wa_list *wal) > { > const struct sseu_dev_info *sseu = >->info.sseu; > unsigned long slice, subslice = 0, slice_mask = 0; > - u64 dss_mask = 0; > u32 lncf_mask = 0; > int i; > > @@ -1151,8 +1149,8 @@ xehp_init_mcr(struct intel_gt *gt, struct i915_wa_list *wal) > */ > > /* Find the potential gslice candidates */ > - dss_mask = intel_sseu_get_subslices(sseu, 0); > - slice_mask = intel_slicemask_from_dssmask(dss_mask, GEN_DSS_PER_GSLICE); > + slice_mask = intel_slicemask_from_xehp_dssmask(sseu->subslice_mask, > + GEN_DSS_PER_GSLICE); > > /* > * Find the potential LNCF candidates. Either LNCF within a valid > @@ -1177,9 +1175,8 @@ xehp_init_mcr(struct intel_gt *gt, struct i915_wa_list *wal) > } > > slice = __ffs(slice_mask); > - subslice = __ffs(dss_mask >> (slice * GEN_DSS_PER_GSLICE)); > + subslice = intel_sseu_find_first_xehp_dss(sseu, GEN_DSS_PER_GSLICE, slice); > WARN_ON(subslice > GEN_DSS_PER_GSLICE); > - WARN_ON(dss_mask >> (slice * GEN_DSS_PER_GSLICE) == 0); > > __add_mcr_wa(gt, wal, slice, subslice); > > @@ -2030,9 +2027,8 @@ engine_fake_wa_init(struct intel_engine_cs *engine, struct i915_wa_list *wal) > > static bool needs_wa_1308578152(struct intel_engine_cs *engine) > { > - u64 dss_mask = intel_sseu_get_subslices(&engine->gt->info.sseu, 0); > - > - return (dss_mask & GENMASK(GEN_DSS_PER_GSLICE - 1, 0)) == 0; > + return intel_sseu_find_first_xehp_dss(&engine->gt->info.sseu, 0, 0) > > + GEN_DSS_PER_GSLICE; > } > > static void > diff --git a/drivers/gpu/drm/i915/i915_getparam.c b/drivers/gpu/drm/i915/i915_getparam.c > index ac9767c56619..6fd15b39570c 100644 > --- a/drivers/gpu/drm/i915/i915_getparam.c > +++ b/drivers/gpu/drm/i915/i915_getparam.c > @@ -162,8 +162,7 @@ int i915_getparam_ioctl(struct drm_device *dev, void *data, > return -EINVAL; > > /* Only copy bits from the first slice */ > - memcpy(&value, sseu->subslice_mask, > - min(sseu->ss_stride, (u8)sizeof(value))); > + value = intel_sseu_get_hsw_subslices(sseu, 0); > if (!value) > return -ENODEV; > break; > diff --git a/drivers/gpu/drm/i915/i915_query.c b/drivers/gpu/drm/i915/i915_query.c > index 89c475d525b8..0094f67c63f2 100644 > --- a/drivers/gpu/drm/i915/i915_query.c > +++ b/drivers/gpu/drm/i915/i915_query.c > @@ -31,10 +31,11 @@ static int copy_query_item(void *query_hdr, size_t query_sz, > > static int fill_topology_info(const struct sseu_dev_info *sseu, > struct drm_i915_query_item *query_item, > - const u8 *subslice_mask) > + intel_sseu_ss_mask_t subslice_mask) > { > struct drm_i915_query_topology_info topo; > u32 slice_length, subslice_length, eu_length, total_length; > + int ss_stride = GEN_SSEU_STRIDE(sseu->max_subslices); > int eu_stride = GEN_SSEU_STRIDE(sseu->max_eus_per_subslice); > int ret; > > @@ -44,7 +45,7 @@ static int fill_topology_info(const struct sseu_dev_info *sseu, > return -ENODEV; > > slice_length = sizeof(sseu->slice_mask); > - subslice_length = sseu->max_slices * sseu->ss_stride; > + subslice_length = sseu->max_slices * ss_stride; > eu_length = sseu->max_slices * sseu->max_subslices * eu_stride; > total_length = sizeof(topo) + slice_length + subslice_length + > eu_length; > @@ -60,7 +61,7 @@ static int fill_topology_info(const struct sseu_dev_info *sseu, > topo.max_eus_per_subslice = sseu->max_eus_per_subslice; > > topo.subslice_offset = slice_length; > - topo.subslice_stride = sseu->ss_stride; > + topo.subslice_stride = ss_stride; > topo.eu_offset = slice_length + subslice_length; > topo.eu_stride = eu_stride; > > @@ -72,9 +73,9 @@ static int fill_topology_info(const struct sseu_dev_info *sseu, > &sseu->slice_mask, slice_length)) > return -EFAULT; > > - if (copy_to_user(u64_to_user_ptr(query_item->data_ptr + > - sizeof(topo) + slice_length), > - subslice_mask, subslice_length)) > + if (intel_sseu_copy_ssmask_to_user(u64_to_user_ptr(query_item->data_ptr + > + sizeof(topo) + slice_length), > + sseu)) > return -EFAULT; > > if (intel_sseu_copy_eumask_to_user(u64_to_user_ptr(query_item->data_ptr + > -- > 2.35.3 >