From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from mx0b-00082601.pphosted.com (mx0b-00082601.pphosted.com [67.231.153.30]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 76F29246BBA for ; Mon, 9 Feb 2026 02:42:41 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=fail smtp.client-ip=67.231.153.30 ARC-Seal:i=2; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1770604961; cv=fail; b=ZjUCulmNosUbt9cFsSN2QjENieK186LUq48LAH3Ubn8cLFxTagsjeYxKFZJYLjLxBQPuKh2kRT6gN+K46urn4+KqjWT2sWrxNbRK2/GJ3xs1sBpJuOV5PiYpzYH3emIaJ374ImiCBQNdYjupXSZJQckO6JxUZKSIdiEhH6mwpa8= ARC-Message-Signature:i=2; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1770604961; c=relaxed/simple; bh=YGjqBBRluYfxV5bqFwa8uZg0WpX6TSkWkhv6fdZJqo0=; h=Message-ID:Date:Subject:To:Cc:References:From:In-Reply-To: Content-Type:MIME-Version; b=dcc7rG0UA4Vj2vFq88oX/VxwY2nXLbzKsIYooyN91JKugMuX7OHgEdol6JQ8Nc3CznkTZOz4tnuXBDNrGco8NiUL7Uo8ny+/ImLFLejuF5cR/NaspYe7z8kHbt9+gFp7Y4705xl7TnTmhoz4XjTLD1te5wErG7Dr/0wFlDaEj04= ARC-Authentication-Results:i=2; smtp.subspace.kernel.org; dmarc=pass (p=reject dis=none) header.from=meta.com; spf=pass smtp.mailfrom=meta.com; dkim=pass (2048-bit key) header.d=meta.com header.i=@meta.com header.b=PzuqrO63; arc=fail smtp.client-ip=67.231.153.30 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=reject dis=none) header.from=meta.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=meta.com Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=meta.com header.i=@meta.com header.b="PzuqrO63" Received: from pps.filterd (m0109331.ppops.net [127.0.0.1]) by mx0a-00082601.pphosted.com (8.18.1.11/8.18.1.11) with ESMTP id 6192SInE2507955; Sun, 8 Feb 2026 18:42:33 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=meta.com; h=cc :content-transfer-encoding:content-type:date:from:in-reply-to :message-id:mime-version:references:subject:to; s=s2048-2025-q2; bh=hrI/Y2Fk5Kyv/zs5ml2gwI4faJmLvvUJGYwd7SViW6E=; b=PzuqrO635+C6 rbioGJWW5I62J9WbvdW1XUC6NPoNCfvfKuZaRfWE8sggIQeA0pf5Yvu4weNQQNj2 7pdsCU+O8VyLpYR4GEZsPJPiCgInQ8F+sQDKcZIS8Gra56OEU8fMshpWx2QpPuKY Lx3fJZFsuAJxhI5ImFgVRJOQFAtm9zP2cQ50iT7nmPrsuoyb3mONdt2r96q1USIp xz2cH1XXaQiv9A8l0DCarnYUmBU4LbzE53DEAw1AVJFbeGI6YyMqzpv7+wI62KXi KL3u/W9nhkOBetLV08U+129PolDZaTh2JntoL90RkQt2wO8m2Ne4bNQ239DZ7wYc OEKfXm2cuw== Received: from byapr05cu005.outbound.protection.outlook.com (mail-westusazon11010019.outbound.protection.outlook.com [52.101.85.19]) by mx0a-00082601.pphosted.com (PPS) with ESMTPS id 4c6vm1488d-2 (version=TLSv1.3 cipher=TLS_AES_256_GCM_SHA384 bits=256 verify=NOT); Sun, 08 Feb 2026 18:42:32 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; s=arcselector10001; d=microsoft.com; cv=none; b=pBI17vOk5lTUOVmo/xNHqRywxUeXOSF76zpKqi9KrkH3bdYcXEz0scsrokEfkVf0IzktsKuPrxyBI51pXIWeEuz2rOTF8AQAvPmPlSWIY9XNeI3wd4I7LVNwxWSiDSdl+Aic78XmXlT+ZdPTeCqa+3yR54YgDegyYa+h8bYqgAuWGoQ3jNPDLpYDqB/OmB6ieykaXtvrX496coV9dwv4/m1h7WYx8DIk7P/9kWWOLpfD3KYemz3US3rArHWaGRFooLKsL9bzjpxDleKdObZRtIbJxK58iHid0N5AWWF8tWYn7IYIaCa6HlrZ6EN4nIbRQs6nYRMHa9nWZRtrsMkRyA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector10001; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-AntiSpam-MessageData-ChunkCount:X-MS-Exchange-AntiSpam-MessageData-0:X-MS-Exchange-AntiSpam-MessageData-1; bh=hrI/Y2Fk5Kyv/zs5ml2gwI4faJmLvvUJGYwd7SViW6E=; b=C5WU9Lu0hwfXJ5UZbbAzRsB3W+stcL6kL85sYmEKwa8Hz0vlJ8kcioUTovsIzYlOMTUn97Ri5qH18oypwYbjz+p8U7U5IiFQPnD/mLkaIb+yYbXuXZmCjCgAoERAMcwgb3StdVS7figV+czauXEJ7wIoDXcuHTVnIYf2iViJji4svol1q4tPjo8tObGz7hSE19nu01SNtdg4Lzi+GUlPGqtm8F6lzrC+qTxXAFz7GRWgY1rHC5XfrNnPOw/W501h9BPRwaD2dKn7Zo9fzs9hd58eMiedLZS5pP1CRC5kbE+pUbhbpdIN54vLnojYGVZhnDve7zjQ8W+xelBLdVBVxg== ARC-Authentication-Results: i=1; mx.microsoft.com 1; spf=pass smtp.mailfrom=meta.com; dmarc=pass action=none header.from=meta.com; dkim=pass header.d=meta.com; arc=none Received: from BLAPR15MB3889.namprd15.prod.outlook.com (2603:10b6:208:27a::11) by PH0PR15MB4232.namprd15.prod.outlook.com (2603:10b6:510:29::13) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.9587.18; Mon, 9 Feb 2026 02:42:29 +0000 Received: from BLAPR15MB3889.namprd15.prod.outlook.com ([fe80::d5d7:f18c:a916:6044]) by BLAPR15MB3889.namprd15.prod.outlook.com ([fe80::d5d7:f18c:a916:6044%5]) with mapi id 15.20.9587.013; Mon, 9 Feb 2026 02:42:29 +0000 Message-ID: Date: Sun, 8 Feb 2026 18:42:26 -0800 User-Agent: Mozilla Thunderbird Subject: Re: [PATCH v6] virtio_net: add page_pool support for buffer allocation To: Xuan Zhuo Cc: =?UTF-8?Q?Eugenio_P=C3=A9rez?= , Andrew Lunn , "David S . Miller" , Eric Dumazet , Jakub Kicinski , Paolo Abeni , David Wei , Matteo Croce , Ilias Apalodimas , netdev@vger.kernel.org, virtualization@lists.linux.dev, linux-kernel@vger.kernel.org, kernel-team@meta.com, "Michael S . Tsirkin" , Jason Wang References: <20260208175410.1910001-1-vishs@meta.com> <1770602432.6451533-1-xuanzhuo@linux.alibaba.com> Content-Language: en-US From: Vishwanath Seshagiri In-Reply-To: <1770602432.6451533-1-xuanzhuo@linux.alibaba.com> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit X-ClientProxiedBy: BYAPR11CA0101.namprd11.prod.outlook.com (2603:10b6:a03:f4::42) To BLAPR15MB3889.namprd15.prod.outlook.com (2603:10b6:208:27a::11) Precedence: bulk X-Mailing-List: virtualization@lists.linux.dev List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 X-MS-PublicTrafficType: Email X-MS-TrafficTypeDiagnostic: BLAPR15MB3889:EE_|PH0PR15MB4232:EE_ X-MS-Office365-Filtering-Correlation-Id: d711deff-2d6f-4424-5365-08de6784de0a X-LD-Processed: 8ae927fe-1255-47a7-a2af-5f3a069daaa2,ExtAddr X-FB-Source: Internal X-MS-Exchange-SenderADCheck: 1 X-MS-Exchange-AntiSpam-Relay: 0 X-Microsoft-Antispam: BCL:0;ARA:13230040|366016|1800799024|10070799003|7416014|376014; X-Microsoft-Antispam-Message-Info: =?utf-8?B?VnI0cXJQMnhUaGZKYURPck5QaXB5MlRoc0R1NnhUZHBSb3JnZGplK1Nld1p0?= =?utf-8?B?T2JlcGlhUVZkVzVHcFlFaEo0aExQRWsycTNteFUxWERBN0h4RThiZzh4dHJv?= =?utf-8?B?Mlg0STc0VmdLN1oyM2RHRWhFdzdic0VNTnJDWjJCKzZHYktmN3V4b2Rad2Ir?= =?utf-8?B?Q3FGekd2TlFsRExJTWVGYmtTMW5RMktBWUpBZTlndGtRQTVFdjY0dmZ1cUpm?= =?utf-8?B?emE0cjIwTVMrbFlXRHJHeE5IaDJLT0EzaGtDbENQZUFrWUY0aDc0ZVNjUkor?= =?utf-8?B?RWNDbEtvVGZGQUlFUUM3SUhZVWh5UUVNWVdQejNHWXpOWnhhT05pYzBlS3dw?= =?utf-8?B?OXIvRmRjMEtyanVEOVJVdkJCQTI5VURXTDhGb1hRTFIrcVVzcVlWQWhXZUI5?= =?utf-8?B?RmYxYTZHV2xxdzR6S0lJdk44ZDkrQk9nczFiQngxZ1NYWTdiRTZHc2JSUUt6?= =?utf-8?B?NEVZbWVYL01tNzNQcWxaSEZPZ0hqT0NiSTg3azJPTThVczJzMXAxNVViM1Nr?= =?utf-8?B?Vjc4aWVSNTA4R0thVlg5UEVnRVFjT2dNZFJvMUxubHNYT1o0K3N4SzBNdlcz?= =?utf-8?B?OEwvRzZpNHd0cjN5OERRVHJQL2hhS2pLR1VHN1Q1WnU4RFVaeCtCOVBPUjVT?= =?utf-8?B?MitweFZPY2ZjKzNRYlA0dWVlMG5PWm9BTTNLdFQ0b1hzWW05UmQ1MTV0WFZ0?= =?utf-8?B?azRaaTVKV1RSTGd0MFJBbjcxSG5SSGZwaWZwVDdjRnhGdW9Yd2JIR2xvQmRH?= =?utf-8?B?YlQrWmlYNTB0QTB5dWtWWFgxVFAwQ093Mlp4Z0lhaXdiRHdBVFE1NytBN0NH?= =?utf-8?B?Y05GS2x4cnRHNTFRcmhpeDVOTjJwME8rUnVKUEsyMHpnU3ZtRWVoUW11R3dq?= =?utf-8?B?eXNvSXJPNHlUSlIxUFRZQXIwYldpZGhNeEU2SEpEclRoZVJWME5xM2FIeUxo?= =?utf-8?B?NjdkblN4UVp3WWdPZWpiVlA3cFZlZzdWak96RFV2cE8zRkFwd2NVUTBGcFZs?= =?utf-8?B?azJ0b3JNYndGWDJkVjBSWXlPR1dXaVZsaEk1QXAyZ0MraklRbnFvWFRXZ1pq?= =?utf-8?B?dXg0TVY4aFVabUZvd3BBKzROVlkxZGJwaVRlWWU4T0tSTkg4OUVSUVpPTDJL?= =?utf-8?B?WU11OUp2WEpBMlY2SnYxNXdRcnhaczlNWVp3SCs3eStmNlpqc3BaY0dRQTAw?= =?utf-8?B?RGpEMXhVQXJjL3NlUXRhaUtUdStvRU5ic3hJRWNLSVRGcVB1NURIaEhqeGZ0?= =?utf-8?B?QUtjMlN5VEZjaUQydDVFVHE3QXV4V0p6bE1xYmd1QlBmNmtsUE94MVQwMmEv?= =?utf-8?B?VVFtTVRnaTRnaTZKRXZQM2Q2L3VrS2tXZHdEYWpJVDBoYVlqRTRFQ3l1TmV5?= =?utf-8?B?QmYxR3ZSNkVBSVVmNktMa3c3UnZacDZpWEdzUlJzZFBSdC80SWwwbjU1bmx5?= =?utf-8?B?NTY2a2VZWHBtb0VqeWdFUUV4Ympkb3JIczE3OExpUTJHOHhnditEdkNuVkxS?= =?utf-8?B?VUlPUHRNM3grYllrejU1ZkJSR1NuQkdtTjhZM3d3SHNVZ2g2K2dPT2daYlZl?= =?utf-8?B?MXAwaWN4WDhiQWlaVVZFZ2dvcFA1ZFVDMXZRTC9ZdnNxY1F0NlVLdlJDNEFC?= =?utf-8?B?UmQyWS8yeVE2YXkrbVlrekFjSkQ2eGF1SVhrOFpGaVdqR0ZkYmJ1bmRVVEp4?= =?utf-8?B?L1Q2aDFUYk5ISlBGTXJQdlNFTnAvbDJ0OHpLQ20vRDdzb2pPNjlvaWFibUdQ?= =?utf-8?B?MGIvVTJqMVVmZW1YSzdJZWQxRXI5SWNKNHZiSTJtRFBWSGxuVmo3Q2ZDMXFm?= =?utf-8?B?bG91QjV3a3dBUEtuSGpSTmhla28xelpIMlM5T1V3YjRvQjNMUk1sM29NZ3NC?= =?utf-8?B?NEdRTVRtOE81b29icUFPdHNiWW5CL1c4NytYUVRkdFArVXp6cGVBV3M3NlRH?= =?utf-8?B?dWN5b1liS2Y2SEZPQ3RSc1BCd3lnc0lpdXZreWhmNDZ4V3E3OWhocW40Znpo?= =?utf-8?B?N1hTL1Z5RXVoWDZMYnBxOGRxRE1NbEl2N2tPWE4waTYzNmExNTNNWXBiUzlX?= =?utf-8?B?Qm15a0FQWEd2WnUwTFlqajVtMjMzaSsydmhHYXZ3cTBjc3UxVFc0VHdMZkJU?= =?utf-8?Q?mN+4=3D?= X-Forefront-Antispam-Report: CIP:255.255.255.255;CTRY:;LANG:en;SCL:1;SRV:;IPV:NLI;SFV:NSPM;H:BLAPR15MB3889.namprd15.prod.outlook.com;PTR:;CAT:NONE;SFS:(13230040)(366016)(1800799024)(10070799003)(7416014)(376014);DIR:OUT;SFP:1101; X-MS-Exchange-AntiSpam-MessageData-ChunkCount: 1 X-MS-Exchange-AntiSpam-MessageData-0: =?utf-8?B?Z3BybklmYjZXZUw0S2l2d2x6SzJEeHRiQ1dCZnVTL05iemlkYW1GcG1RWDV5?= =?utf-8?B?V2JiZmFUeTJkVHp1WVVxcnNCbzhhem9JY0s4KzRmTkt1bThXTHhiWDVvZEpa?= =?utf-8?B?cEdodVBtT2F1N1JFTlRsdkszUHQ0Y1RWVzZxeUtJOG9CemNqRm9TZHNmZXRX?= =?utf-8?B?T1IyNUNUa3ZuOWkvYm8zTSt4eUNlaVJDU0VQQjB0NjY5Q0pEcmo0akN1SCs1?= =?utf-8?B?blQ1d0Y1ZW1wMjhQOStNeVNBY1Z0RTNmNFdTeTZhM0lab0tZSit4VkRDYkhn?= =?utf-8?B?QmRpMy8wcFh3M20rcG9ITFZXczNiay8zclltODJKUjZXb0QzL3VCT0RyaDVn?= =?utf-8?B?cnA5S3c2NGJXUmtFcWdjMmx3bFdJMHlrak5hVHBwN3pZUHNsc1d3ZkZhSFZS?= =?utf-8?B?QnovZmQ2V0Z1SWlmd1BWOFlkb1p5b2RGUjJYbDdoT2J3YjRvdTFFT0k2YStN?= =?utf-8?B?cG9nNCtiM0pWZ3AzY0JKRUd1WHZkd0Qya1pOU0dTT2MxbXNjQ0dtckhubGNI?= =?utf-8?B?a04zOHdIMzVRbmQ2TnpaQ1hxazh2WnJLMTlhbnZqbmhORDBTZnlKYk00dHZo?= =?utf-8?B?d3VvZXRqTGZaVkk3SFVoUlN4c2tYWEF2dHViM3BEcG1kQjZYTlAzY2FvV1RO?= =?utf-8?B?dGRCUE50OE5EM2hnYzlwTkMvZEtucU9BVnVhWDE4alUyeW5xeXkzZ3VkeUJv?= =?utf-8?B?cUF3UkpiSXhCYzNiVWkybE4ydGRKQnVNRERlTFpabU1JMzc3a2M1RnIxOE1J?= =?utf-8?B?b3FmK2V5cTlXdHpZbkZFM09BWFc4UGxQcElwYVpsU09maG1GWEVzSjVSc3N3?= =?utf-8?B?WUFsSDJJNS9FZjYvbDBMeHNxbHlERDhVb1JCZy9oa3Vsek1vUXY5UUpIakNO?= =?utf-8?B?RVdtTUc4aUM2Zm5DSEZsMnozUkNCSGI2VVR4Y0F0LzNtbC8vRWNqaGxZRlNz?= =?utf-8?B?ME9uWi95dUUyUjlnVjkraFNwb09VRVdDclBSM21oc2tYUnFUU1F4R0RtTGYv?= =?utf-8?B?Yy9Oa0M5czR4U1dENlhoNVdPaDRaK0VwMHNEOUZtNCtJSkU2REVCNTlPazVl?= =?utf-8?B?Smx3REpWWDhnWlpWWTVsd0hxM0ZxOGhLY2hNWjhiTDF3Zmd5djAremFWcHlq?= =?utf-8?B?b1pBbTdMNFlsNkdVMC81TzNFQkkxR3JKWGhnak5VM0Nsb2F0cTl2dEJBUDJt?= =?utf-8?B?MmxsMjgzK0RLWDlaQUZKRjNWYzdDclFacXZIRkhjcmFySkN6QmUxQnBLMGFa?= =?utf-8?B?bFRXWlJGVGkya1VxRzBmUnhOZUFMNVlxVGQrK1NjZ2lwSGxmVXpOcU5hVTdr?= =?utf-8?B?SU0rS2RCV21KRzRGNnE0MmpqMDZFSG0rRURkV2tqcFZPQkpZaTg5QmQ0SWk2?= =?utf-8?B?aXZDQ3lTd2tjR0VpZXdvb1MvTmU5a3c2b0gxMUVRS0ZGaDlPbERvQWhhYVg4?= =?utf-8?B?b2lBNGJ4SS8yUUMzZlRKNHpMUkN3ZVE2c044TXdvSzIzVm1IbDRHLzVuVUp5?= =?utf-8?B?T3laTVNjQkhrVEpJTnpRNEpJalRHdGpyeGdlOEdvRGFKQVBhUXRMVDN6UXFs?= =?utf-8?B?VVRPSExtMWJ1WnBoS0V2MGhqU0ZaL0kyc3VDY1V3TllrUTY3eEF3bzNpQmVt?= =?utf-8?B?OGFQeEIvVTY5U3d0TnM3Z09rSTdzUHNvL1BoeGNnRnRiMFhqcDNQTFg3dDVi?= =?utf-8?B?alZQVTFKWWRLWG1wYjFsOStoWVhsNTg2VThIT0VvRjAwQ2ZjUnVRc21RazZv?= =?utf-8?B?WU9pRytHYXlXdmNEcHZoVlIyL2g2S3Rxb3VhemdxOW1HNXY1c0lNNnJtdm1W?= =?utf-8?B?THFvVUxRaEpzM0Jqc3pPZmtuSkh1aFI2c3JHeVFjdytKSG9YRURCenUzdnla?= =?utf-8?B?bllpdkhJU01TZUFpNXJDS0h3Z1NQTGJMQnl0aU9ZS0xNZ1JmbmRPVkhMR2lv?= =?utf-8?B?eGtlVXVCSVVxR3ZPZEhhWHQ0T0lJTXZ0WTFrODlzb011SGR3eGhVZUNtV0tx?= =?utf-8?B?bzZ5dmdFMjl6RTVKNDFjOEJzQUU0R1BoNm9EemluTWR0S0RFbDNGSzJwaHR1?= =?utf-8?B?cWJGZmhDZWZ4b0VPQTFOVDFNVXVZOEk0VXlPRE5QYzdhemRldUxBOFg5azhG?= =?utf-8?B?MktaeTZiMWR5K1lRb0hBcXNUMFh5MllZY0JFSnhXNXFnYXBzTDR4Q01rOWhZ?= =?utf-8?B?R3A4ODFtU1c5dmREc1dST0plSkdxZjVJWVpLYkxlaDE1YmZMbkNFLzd4NzlU?= =?utf-8?B?UUlEUCtMMFk5ZFpOZFR0a3ZobGtPK2pVNjNSNXR5aFBnaFR4RTJrdS9LdTN3?= =?utf-8?B?blhxeFhzbklTMktENUF6Z0FQYVU5V1dMR0JmZjRiZHV1ZS9yNTdScFUvaFdr?= =?utf-8?Q?ayDItp5T8CTb3C/N2ywwYCWbgUpEzc2IVxp36?= X-OriginatorOrg: meta.com X-MS-Exchange-CrossTenant-Network-Message-Id: d711deff-2d6f-4424-5365-08de6784de0a X-MS-Exchange-CrossTenant-AuthSource: BLAPR15MB3889.namprd15.prod.outlook.com X-MS-Exchange-CrossTenant-AuthAs: Internal X-MS-Exchange-CrossTenant-OriginalArrivalTime: 09 Feb 2026 02:42:29.4679 (UTC) X-MS-Exchange-CrossTenant-FromEntityHeader: Hosted X-MS-Exchange-CrossTenant-Id: 8ae927fe-1255-47a7-a2af-5f3a069daaa2 X-MS-Exchange-CrossTenant-MailboxType: HOSTED X-MS-Exchange-CrossTenant-UserPrincipalName: 4nUreR3pCFFXzJMkOjP/J6WnpjGSUGP5+XvTDfHKJXZQtqSyXWHfspELT5hFBtuY X-MS-Exchange-Transport-CrossTenantHeadersStamped: PH0PR15MB4232 X-Authority-Analysis: v=2.4 cv=ILUPywvG c=1 sm=1 tr=0 ts=69894998 cx=c_pps a=fdYRkKHAJehZsuGPaltaPA==:117 a=6eWqkTHjU83fiwn7nKZWdM+Sl24=:19 a=z/mQ4Ysz8XfWz/Q5cLBRGdckG28=:19 a=lCpzRmAYbLLaTzLvsPZ7Mbvzbb8=:19 a=xqWC_Br6kY4A:10 a=IkcTkHD0fZMA:10 a=HzLeVaNsDn8A:10 a=VkNPw1HP01LnGYTKEx00:22 a=Mpw57Om8IfrbqaoTuvik:22 a=GgsMoib0sEa3-_RKJdDe:22 a=VwQbUJbxAAAA:8 a=VabnemYjAAAA:8 a=LQxVJagRmHXr5uwAUZgA:9 a=QEXdDO2ut3YA:10 a=gKebqoRLp9LExxC7YDUY:22 X-Proofpoint-ORIG-GUID: IAx-N6qRTIAKmCLINtRzXK4mf02PkPkW X-Proofpoint-Spam-Details-Enc: AW1haW4tMjYwMjA5MDAyMSBTYWx0ZWRfX+hYFWiU3nkCE w+i2IlhIO2eW+aldmgKnaZDswLvG3Tldo3PtK+N8CuTc3X6IXKPrgNkQwkDnCiWO4a9rGZe0qGq p/2oyxDrJ+FumVxvW/MQ17Ha/BkzdQHmTASvBi1x++bpdYqWgy8JceV9i5TRMQFYnAWGa4GyTOe WEFiYnrFikbPq6df3qf1lbzAGpMilrpzgDVZsWIDi6RXT+v/WNAFeH4ONc5gnt+a3bjjJpN4jxL fSP46j7F412Gxk2a9M9wGIuMuX7EFf8HiZe3zG2rSAWY5z3PGZrxFRTFwIzhJKttJfZzPgbQe4z S+RZv7by8ea+v1LvXaJMM2X3EGxhahTRhPZzLzHYLr258jJ68284PfJiM1UnImfE6T+Ys3EG28h 6CNHqWjiiKDekB/JQCYJZz+4vvzOENe7SR5WugFmFdgjXysz0kx9VB+8R2fq3SABsba7YlOQxkm rslAbkMR+RJO/CGmfIg== X-Proofpoint-GUID: IAx-N6qRTIAKmCLINtRzXK4mf02PkPkW X-Proofpoint-Virus-Version: vendor=baseguard engine=ICAP:2.0.293,Aquarius:18.0.1121,Hydra:6.1.51,FMLib:17.12.100.49 definitions=2026-02-08_05,2026-02-05_03,2025-10-01_01 On 2/8/26 6:00 PM, Xuan Zhuo wrote: > On Sun, 8 Feb 2026 09:54:10 -0800, Vishwanath Seshagiri wrote: >> Use page_pool for RX buffer allocation in mergeable and small buffer >> modes to enable page recycling and avoid repeated page allocator calls. >> skb_mark_for_recycle() enables page reuse in the network stack. >> >> Big packets mode is unchanged because it uses page->private for linked >> list chaining of multiple pages per buffer, which conflicts with >> page_pool's internal use of page->private. >> >> Implement conditional DMA premapping using virtqueue_dma_dev(): >> - When non-NULL (vhost, virtio-pci): use PP_FLAG_DMA_MAP with page_pool >> handling DMA mapping, submit via virtqueue_add_inbuf_premapped() >> - When NULL (VDUSE, direct physical): page_pool handles allocation only, >> submit via virtqueue_add_inbuf_ctx() >> >> This preserves the DMA premapping optimization from commit 31f3cd4e5756b >> ("virtio-net: rq submits premapped per-buffer") while adding page_pool >> support as a prerequisite for future zero-copy features (devmem TCP, >> io_uring ZCRX). >> >> Page pools are created in probe and destroyed in remove (not open/close), >> following existing driver behavior where RX buffers remain in virtqueues >> across interface state changes. >> >> Signed-off-by: Vishwanath Seshagiri >> --- >> Changes in v6: >> - Drop page_pool_frag_offset_add() helper and switch to page_pool_alloc_va(); >> page_pool_alloc_netmem() already handles internal fragmentation internally >> (Jakub Kicinski) >> - v5: >> https://lore.kernel.org/virtualization/20260206002715.1885869-1-vishs@meta.com/ >> >> Benchmark results: >> >> Configuration: pktgen TX -> tap -> vhost-net | virtio-net RX -> XDP_DROP >> >> Small packets (64 bytes, mrg_rxbuf=off): >> 1Q: 853,493 -> 868,923 pps (+1.8%) >> 2Q: 1,655,793 -> 1,696,707 pps (+2.5%) >> 4Q: 3,143,375 -> 3,302,511 pps (+5.1%) >> 8Q: 6,082,590 -> 6,156,894 pps (+1.2%) >> >> Mergeable RX (64 bytes): >> 1Q: 766,168 -> 814,493 pps (+6.3%) >> 2Q: 1,384,871 -> 1,670,639 pps (+20.6%) >> 4Q: 2,773,081 -> 3,080,574 pps (+11.1%) >> 8Q: 5,600,615 -> 6,043,891 pps (+7.9%) >> >> Mergeable RX (1500 bytes): >> 1Q: 741,579 -> 785,442 pps (+5.9%) >> 2Q: 1,310,043 -> 1,534,554 pps (+17.1%) >> 4Q: 2,748,700 -> 2,890,582 pps (+5.2%) >> 8Q: 5,348,589 -> 5,618,664 pps (+5.0%) >> >> drivers/net/Kconfig | 1 + >> drivers/net/virtio_net.c | 434 +++++++++++++++++++-------------------- >> 2 files changed, 217 insertions(+), 218 deletions(-) >> >> diff --git a/drivers/net/Kconfig b/drivers/net/Kconfig >> index ac12eaf11755..f1e6b6b0a86f 100644 >> --- a/drivers/net/Kconfig >> +++ b/drivers/net/Kconfig >> @@ -450,6 +450,7 @@ config VIRTIO_NET >> depends on VIRTIO >> select NET_FAILOVER >> select DIMLIB >> + select PAGE_POOL >> help >> This is the virtual network driver for virtio. It can be used with >> QEMU based VMMs (like KVM or Xen). Say Y or M. >> diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c >> index db88dcaefb20..5055df56e4a7 100644 >> --- a/drivers/net/virtio_net.c >> +++ b/drivers/net/virtio_net.c >> @@ -26,6 +26,7 @@ >> #include >> #include >> #include >> +#include >> >> static int napi_weight = NAPI_POLL_WEIGHT; >> module_param(napi_weight, int, 0444); >> @@ -290,14 +291,6 @@ struct virtnet_interrupt_coalesce { >> u32 max_usecs; >> }; >> >> -/* The dma information of pages allocated at a time. */ >> -struct virtnet_rq_dma { >> - dma_addr_t addr; >> - u32 ref; >> - u16 len; >> - u16 need_sync; >> -}; >> - >> /* Internal representation of a send virtqueue */ >> struct send_queue { >> /* Virtqueue associated with this send _queue */ >> @@ -356,8 +349,10 @@ struct receive_queue { >> /* Average packet length for mergeable receive buffers. */ >> struct ewma_pkt_len mrg_avg_pkt_len; >> >> - /* Page frag for packet buffer allocation. */ >> - struct page_frag alloc_frag; >> + struct page_pool *page_pool; >> + >> + /* True if page_pool handles DMA mapping via PP_FLAG_DMA_MAP */ >> + bool use_page_pool_dma; >> >> /* RX: fragments + linear part + virtio header */ >> struct scatterlist sg[MAX_SKB_FRAGS + 2]; >> @@ -370,9 +365,6 @@ struct receive_queue { >> >> struct xdp_rxq_info xdp_rxq; >> >> - /* Record the last dma info to free after new pages is allocated. */ >> - struct virtnet_rq_dma *last_dma; >> - >> struct xsk_buff_pool *xsk_pool; >> >> /* xdp rxq used by xsk */ >> @@ -521,11 +513,13 @@ static int virtnet_xdp_handler(struct bpf_prog *xdp_prog, struct xdp_buff *xdp, >> struct virtnet_rq_stats *stats); >> static void virtnet_receive_done(struct virtnet_info *vi, struct receive_queue *rq, >> struct sk_buff *skb, u8 flags); >> -static struct sk_buff *virtnet_skb_append_frag(struct sk_buff *head_skb, >> +static struct sk_buff *virtnet_skb_append_frag(struct receive_queue *rq, >> + struct sk_buff *head_skb, >> struct sk_buff *curr_skb, >> struct page *page, void *buf, >> int len, int truesize); >> static void virtnet_xsk_completed(struct send_queue *sq, int num); >> +static void free_unused_bufs(struct virtnet_info *vi); >> >> enum virtnet_xmit_type { >> VIRTNET_XMIT_TYPE_SKB, >> @@ -706,15 +700,24 @@ static struct page *get_a_page(struct receive_queue *rq, gfp_t gfp_mask) >> return p; >> } >> >> +static void virtnet_put_page(struct receive_queue *rq, struct page *page, >> + bool allow_direct) >> +{ >> + if (page_pool_page_is_pp(page)) >> + page_pool_put_page(rq->page_pool, page, -1, allow_direct); >> + else >> + put_page(page); >> +} > > Why we need this? > For the caller, we should know which one should be used? > This was after some feedback to unify the alloc/free path checks in v4. But you raise a valid point - callers already know the mode via virtnet_no_page_pool(). I can simplify this to just call page_pool_put_page() directly, since virtnet_put_page() is only called from paths that already checked we're using page_pool. Would you prefer that? > >> + >> static void virtnet_rq_free_buf(struct virtnet_info *vi, >> struct receive_queue *rq, void *buf) >> { >> if (vi->mergeable_rx_bufs) >> - put_page(virt_to_head_page(buf)); >> + virtnet_put_page(rq, virt_to_head_page(buf), false); >> else if (vi->big_packets) >> give_pages(rq, buf); >> else >> - put_page(virt_to_head_page(buf)); >> + virtnet_put_page(rq, virt_to_head_page(buf), false); >> } >> >> static void enable_rx_mode_work(struct virtnet_info *vi) >> @@ -876,10 +879,16 @@ static struct sk_buff *page_to_skb(struct virtnet_info *vi, >> skb = virtnet_build_skb(buf, truesize, p - buf, len); >> if (unlikely(!skb)) >> return NULL; >> + /* Big packets mode chains pages via page->private, which is >> + * incompatible with the way page_pool uses page->private. >> + * Currently, big packets mode doesn't use page pools. >> + */ >> + if (vi->big_packets && !vi->mergeable_rx_bufs) { >> + page = (struct page *)page->private; >> + if (page) >> + give_pages(rq, page); >> + } >> >> - page = (struct page *)page->private; >> - if (page) >> - give_pages(rq, page); >> goto ok; >> } >> >> @@ -925,133 +934,18 @@ static struct sk_buff *page_to_skb(struct virtnet_info *vi, >> hdr = skb_vnet_common_hdr(skb); >> memcpy(hdr, hdr_p, hdr_len); >> if (page_to_free) >> - put_page(page_to_free); >> + virtnet_put_page(rq, page_to_free, true); >> >> return skb; >> } >> >> -static void virtnet_rq_unmap(struct receive_queue *rq, void *buf, u32 len) >> -{ >> - struct virtnet_info *vi = rq->vq->vdev->priv; >> - struct page *page = virt_to_head_page(buf); >> - struct virtnet_rq_dma *dma; >> - void *head; >> - int offset; >> - >> - BUG_ON(vi->big_packets && !vi->mergeable_rx_bufs); >> - >> - head = page_address(page); >> - >> - dma = head; >> - >> - --dma->ref; >> - >> - if (dma->need_sync && len) { >> - offset = buf - (head + sizeof(*dma)); >> - >> - virtqueue_map_sync_single_range_for_cpu(rq->vq, dma->addr, >> - offset, len, >> - DMA_FROM_DEVICE); >> - } >> - >> - if (dma->ref) >> - return; >> - >> - virtqueue_unmap_single_attrs(rq->vq, dma->addr, dma->len, >> - DMA_FROM_DEVICE, DMA_ATTR_SKIP_CPU_SYNC); >> - put_page(page); >> -} >> - >> static void *virtnet_rq_get_buf(struct receive_queue *rq, u32 *len, void **ctx) >> { >> struct virtnet_info *vi = rq->vq->vdev->priv; >> - void *buf; >> >> BUG_ON(vi->big_packets && !vi->mergeable_rx_bufs); >> >> - buf = virtqueue_get_buf_ctx(rq->vq, len, ctx); >> - if (buf) >> - virtnet_rq_unmap(rq, buf, *len); >> - >> - return buf; >> -} >> - >> -static void virtnet_rq_init_one_sg(struct receive_queue *rq, void *buf, u32 len) >> -{ >> - struct virtnet_info *vi = rq->vq->vdev->priv; >> - struct virtnet_rq_dma *dma; >> - dma_addr_t addr; >> - u32 offset; >> - void *head; >> - >> - BUG_ON(vi->big_packets && !vi->mergeable_rx_bufs); >> - >> - head = page_address(rq->alloc_frag.page); >> - >> - offset = buf - head; >> - >> - dma = head; >> - >> - addr = dma->addr - sizeof(*dma) + offset; >> - >> - sg_init_table(rq->sg, 1); >> - sg_fill_dma(rq->sg, addr, len); >> -} >> - >> -static void *virtnet_rq_alloc(struct receive_queue *rq, u32 size, gfp_t gfp) >> -{ >> - struct page_frag *alloc_frag = &rq->alloc_frag; >> - struct virtnet_info *vi = rq->vq->vdev->priv; >> - struct virtnet_rq_dma *dma; >> - void *buf, *head; >> - dma_addr_t addr; >> - >> - BUG_ON(vi->big_packets && !vi->mergeable_rx_bufs); >> - >> - head = page_address(alloc_frag->page); >> - >> - dma = head; >> - >> - /* new pages */ >> - if (!alloc_frag->offset) { >> - if (rq->last_dma) { >> - /* Now, the new page is allocated, the last dma >> - * will not be used. So the dma can be unmapped >> - * if the ref is 0. >> - */ >> - virtnet_rq_unmap(rq, rq->last_dma, 0); >> - rq->last_dma = NULL; >> - } >> - >> - dma->len = alloc_frag->size - sizeof(*dma); >> - >> - addr = virtqueue_map_single_attrs(rq->vq, dma + 1, >> - dma->len, DMA_FROM_DEVICE, 0); >> - if (virtqueue_map_mapping_error(rq->vq, addr)) >> - return NULL; >> - >> - dma->addr = addr; >> - dma->need_sync = virtqueue_map_need_sync(rq->vq, addr); >> - >> - /* Add a reference to dma to prevent the entire dma from >> - * being released during error handling. This reference >> - * will be freed after the pages are no longer used. >> - */ >> - get_page(alloc_frag->page); >> - dma->ref = 1; >> - alloc_frag->offset = sizeof(*dma); >> - >> - rq->last_dma = dma; >> - } >> - >> - ++dma->ref; >> - >> - buf = head + alloc_frag->offset; >> - >> - get_page(alloc_frag->page); >> - alloc_frag->offset += size; >> - >> - return buf; >> + return virtqueue_get_buf_ctx(rq->vq, len, ctx); >> } >> >> static void virtnet_rq_unmap_free_buf(struct virtqueue *vq, void *buf) >> @@ -1067,9 +961,6 @@ static void virtnet_rq_unmap_free_buf(struct virtqueue *vq, void *buf) >> return; >> } >> >> - if (!vi->big_packets || vi->mergeable_rx_bufs) >> - virtnet_rq_unmap(rq, buf, 0); >> - >> virtnet_rq_free_buf(vi, rq, buf); >> } >> >> @@ -1335,7 +1226,7 @@ static int xsk_append_merge_buffer(struct virtnet_info *vi, >> >> truesize = len; >> >> - curr_skb = virtnet_skb_append_frag(head_skb, curr_skb, page, >> + curr_skb = virtnet_skb_append_frag(rq, head_skb, curr_skb, page, >> buf, len, truesize); >> if (!curr_skb) { >> put_page(page); >> @@ -1771,7 +1662,7 @@ static int virtnet_xdp_xmit(struct net_device *dev, >> return ret; >> } >> >> -static void put_xdp_frags(struct xdp_buff *xdp) >> +static void put_xdp_frags(struct receive_queue *rq, struct xdp_buff *xdp) >> { >> struct skb_shared_info *shinfo; >> struct page *xdp_page; >> @@ -1781,7 +1672,7 @@ static void put_xdp_frags(struct xdp_buff *xdp) >> shinfo = xdp_get_shared_info_from_buff(xdp); >> for (i = 0; i < shinfo->nr_frags; i++) { >> xdp_page = skb_frag_page(&shinfo->frags[i]); >> - put_page(xdp_page); >> + virtnet_put_page(rq, xdp_page, true); >> } >> } >> } >> @@ -1873,7 +1764,7 @@ static struct page *xdp_linearize_page(struct net_device *dev, >> if (page_off + *len + tailroom > PAGE_SIZE) >> return NULL; >> >> - page = alloc_page(GFP_ATOMIC); >> + page = page_pool_alloc_pages(rq->page_pool, GFP_ATOMIC); >> if (!page) >> return NULL; >> >> @@ -1897,7 +1788,7 @@ static struct page *xdp_linearize_page(struct net_device *dev, >> off = buf - page_address(p); >> >> if (check_mergeable_len(dev, ctx, buflen)) { >> - put_page(p); >> + virtnet_put_page(rq, p, true); >> goto err_buf; >> } >> >> @@ -1905,21 +1796,21 @@ static struct page *xdp_linearize_page(struct net_device *dev, >> * is sending packet larger than the MTU. >> */ >> if ((page_off + buflen + tailroom) > PAGE_SIZE) { >> - put_page(p); >> + virtnet_put_page(rq, p, true); >> goto err_buf; >> } >> >> memcpy(page_address(page) + page_off, >> page_address(p) + off, buflen); >> page_off += buflen; >> - put_page(p); >> + virtnet_put_page(rq, p, true); >> } >> >> /* Headroom does not contribute to packet length */ >> *len = page_off - XDP_PACKET_HEADROOM; >> return page; >> err_buf: >> - __free_pages(page, 0); >> + page_pool_put_page(rq->page_pool, page, -1, true); >> return NULL; >> } >> >> @@ -1969,6 +1860,12 @@ static struct sk_buff *receive_small_xdp(struct net_device *dev, >> unsigned int metasize = 0; >> u32 act; >> >> + if (rq->use_page_pool_dma) { >> + int off = buf - page_address(page); >> + >> + page_pool_dma_sync_for_cpu(rq->page_pool, page, off, len); >> + } >> + >> if (unlikely(hdr->hdr.gso_type)) >> goto err_xdp; >> >> @@ -1996,7 +1893,7 @@ static struct sk_buff *receive_small_xdp(struct net_device *dev, >> goto err_xdp; >> >> buf = page_address(xdp_page); >> - put_page(page); >> + virtnet_put_page(rq, page, true); >> page = xdp_page; >> } >> >> @@ -2028,13 +1925,15 @@ static struct sk_buff *receive_small_xdp(struct net_device *dev, >> if (metasize) >> skb_metadata_set(skb, metasize); >> >> + skb_mark_for_recycle(skb); >> + >> return skb; >> >> err_xdp: >> u64_stats_inc(&stats->xdp_drops); >> err: >> u64_stats_inc(&stats->drops); >> - put_page(page); >> + virtnet_put_page(rq, page, true); >> xdp_xmit: >> return NULL; >> } >> @@ -2056,6 +1955,12 @@ static struct sk_buff *receive_small(struct net_device *dev, >> */ >> buf -= VIRTNET_RX_PAD + xdp_headroom; >> >> + if (rq->use_page_pool_dma) { >> + int offset = buf - page_address(page); >> + >> + page_pool_dma_sync_for_cpu(rq->page_pool, page, offset, len); >> + } >> + >> len -= vi->hdr_len; >> u64_stats_add(&stats->bytes, len); >> >> @@ -2082,12 +1987,14 @@ static struct sk_buff *receive_small(struct net_device *dev, >> } >> >> skb = receive_small_build_skb(vi, xdp_headroom, buf, len); >> - if (likely(skb)) >> + if (likely(skb)) { >> + skb_mark_for_recycle(skb); >> return skb; >> + } >> >> err: >> u64_stats_inc(&stats->drops); >> - put_page(page); >> + virtnet_put_page(rq, page, true); >> return NULL; >> } >> >> @@ -2142,7 +2049,7 @@ static void mergeable_buf_free(struct receive_queue *rq, int num_buf, >> } >> u64_stats_add(&stats->bytes, len); >> page = virt_to_head_page(buf); >> - put_page(page); >> + virtnet_put_page(rq, page, true); >> } >> } >> >> @@ -2253,7 +2160,7 @@ static int virtnet_build_xdp_buff_mrg(struct net_device *dev, >> offset = buf - page_address(page); >> >> if (check_mergeable_len(dev, ctx, len)) { >> - put_page(page); >> + virtnet_put_page(rq, page, true); >> goto err; >> } >> >> @@ -2272,7 +2179,7 @@ static int virtnet_build_xdp_buff_mrg(struct net_device *dev, >> return 0; >> >> err: >> - put_xdp_frags(xdp); >> + put_xdp_frags(rq, xdp); >> return -EINVAL; >> } >> >> @@ -2337,7 +2244,7 @@ static void *mergeable_xdp_get_buf(struct virtnet_info *vi, >> if (*len + xdp_room > PAGE_SIZE) >> return NULL; >> >> - xdp_page = alloc_page(GFP_ATOMIC); >> + xdp_page = page_pool_alloc_pages(rq->page_pool, GFP_ATOMIC); >> if (!xdp_page) >> return NULL; >> >> @@ -2347,7 +2254,7 @@ static void *mergeable_xdp_get_buf(struct virtnet_info *vi, >> >> *frame_sz = PAGE_SIZE; >> >> - put_page(*page); >> + virtnet_put_page(rq, *page, true); >> >> *page = xdp_page; >> >> @@ -2393,6 +2300,8 @@ static struct sk_buff *receive_mergeable_xdp(struct net_device *dev, >> head_skb = build_skb_from_xdp_buff(dev, vi, &xdp, xdp_frags_truesz); >> if (unlikely(!head_skb)) >> break; >> + >> + skb_mark_for_recycle(head_skb); >> return head_skb; >> >> case XDP_TX: >> @@ -2403,10 +2312,10 @@ static struct sk_buff *receive_mergeable_xdp(struct net_device *dev, >> break; >> } >> >> - put_xdp_frags(&xdp); >> + put_xdp_frags(rq, &xdp); >> >> err_xdp: >> - put_page(page); >> + virtnet_put_page(rq, page, true); >> mergeable_buf_free(rq, num_buf, dev, stats); >> >> u64_stats_inc(&stats->xdp_drops); >> @@ -2414,7 +2323,8 @@ static struct sk_buff *receive_mergeable_xdp(struct net_device *dev, >> return NULL; >> } >> >> -static struct sk_buff *virtnet_skb_append_frag(struct sk_buff *head_skb, >> +static struct sk_buff *virtnet_skb_append_frag(struct receive_queue *rq, >> + struct sk_buff *head_skb, >> struct sk_buff *curr_skb, >> struct page *page, void *buf, >> int len, int truesize) >> @@ -2446,7 +2356,7 @@ static struct sk_buff *virtnet_skb_append_frag(struct sk_buff *head_skb, >> >> offset = buf - page_address(page); >> if (skb_can_coalesce(curr_skb, num_skb_frags, page, offset)) { >> - put_page(page); >> + virtnet_put_page(rq, page, true); >> skb_coalesce_rx_frag(curr_skb, num_skb_frags - 1, >> len, truesize); >> } else { >> @@ -2475,6 +2385,10 @@ static struct sk_buff *receive_mergeable(struct net_device *dev, >> unsigned int headroom = mergeable_ctx_to_headroom(ctx); >> >> head_skb = NULL; >> + >> + if (rq->use_page_pool_dma) >> + page_pool_dma_sync_for_cpu(rq->page_pool, page, offset, len); >> + >> u64_stats_add(&stats->bytes, len - vi->hdr_len); >> >> if (check_mergeable_len(dev, ctx, len)) >> @@ -2499,6 +2413,8 @@ static struct sk_buff *receive_mergeable(struct net_device *dev, >> >> if (unlikely(!curr_skb)) >> goto err_skb; >> + >> + skb_mark_for_recycle(head_skb); >> while (--num_buf) { >> buf = virtnet_rq_get_buf(rq, &len, &ctx); >> if (unlikely(!buf)) { >> @@ -2517,7 +2433,7 @@ static struct sk_buff *receive_mergeable(struct net_device *dev, >> goto err_skb; >> >> truesize = mergeable_ctx_to_truesize(ctx); >> - curr_skb = virtnet_skb_append_frag(head_skb, curr_skb, page, >> + curr_skb = virtnet_skb_append_frag(rq, head_skb, curr_skb, page, >> buf, len, truesize); >> if (!curr_skb) >> goto err_skb; >> @@ -2527,7 +2443,7 @@ static struct sk_buff *receive_mergeable(struct net_device *dev, >> return head_skb; >> >> err_skb: >> - put_page(page); >> + virtnet_put_page(rq, page, true); >> mergeable_buf_free(rq, num_buf, dev, stats); >> >> err_buf: >> @@ -2666,32 +2582,41 @@ static void receive_buf(struct virtnet_info *vi, struct receive_queue *rq, >> static int add_recvbuf_small(struct virtnet_info *vi, struct receive_queue *rq, >> gfp_t gfp) >> { >> - char *buf; >> unsigned int xdp_headroom = virtnet_get_headroom(vi); >> void *ctx = (void *)(unsigned long)xdp_headroom; >> - int len = vi->hdr_len + VIRTNET_RX_PAD + GOOD_PACKET_LEN + xdp_headroom; >> + unsigned int len = vi->hdr_len + VIRTNET_RX_PAD + GOOD_PACKET_LEN + xdp_headroom; >> + struct page *page; >> + dma_addr_t addr; >> + char *buf; >> int err; >> >> len = SKB_DATA_ALIGN(len) + >> SKB_DATA_ALIGN(sizeof(struct skb_shared_info)); >> >> - if (unlikely(!skb_page_frag_refill(len, &rq->alloc_frag, gfp))) >> - return -ENOMEM; >> - >> - buf = virtnet_rq_alloc(rq, len, gfp); >> + buf = page_pool_alloc_va(rq->page_pool, &len, gfp); >> if (unlikely(!buf)) >> return -ENOMEM; >> >> buf += VIRTNET_RX_PAD + xdp_headroom; >> >> - virtnet_rq_init_one_sg(rq, buf, vi->hdr_len + GOOD_PACKET_LEN); >> + if (rq->use_page_pool_dma) { >> + page = virt_to_head_page(buf); >> + addr = page_pool_get_dma_addr(page) + >> + (buf - (char *)page_address(page)); >> >> - err = virtqueue_add_inbuf_premapped(rq->vq, rq->sg, 1, buf, ctx, gfp); >> - if (err < 0) { >> - virtnet_rq_unmap(rq, buf, 0); >> - put_page(virt_to_head_page(buf)); >> + sg_init_table(rq->sg, 1); >> + sg_fill_dma(rq->sg, addr, vi->hdr_len + GOOD_PACKET_LEN); >> + err = virtqueue_add_inbuf_premapped(rq->vq, rq->sg, 1, >> + buf, ctx, gfp); >> + } else { >> + sg_init_one(rq->sg, buf, vi->hdr_len + GOOD_PACKET_LEN); >> + err = virtqueue_add_inbuf_ctx(rq->vq, rq->sg, 1, >> + buf, ctx, gfp); >> } >> >> + if (err < 0) >> + page_pool_put_page(rq->page_pool, virt_to_head_page(buf), >> + -1, false); >> return err; >> } >> >> @@ -2764,13 +2689,14 @@ static unsigned int get_mergeable_buf_len(struct receive_queue *rq, >> static int add_recvbuf_mergeable(struct virtnet_info *vi, >> struct receive_queue *rq, gfp_t gfp) >> { >> - struct page_frag *alloc_frag = &rq->alloc_frag; >> unsigned int headroom = virtnet_get_headroom(vi); >> unsigned int tailroom = headroom ? sizeof(struct skb_shared_info) : 0; >> unsigned int room = SKB_DATA_ALIGN(headroom + tailroom); >> - unsigned int len, hole; >> - void *ctx; >> + unsigned int len, alloc_len; >> + struct page *page; >> + dma_addr_t addr; >> char *buf; >> + void *ctx; >> int err; >> >> /* Extra tailroom is needed to satisfy XDP's assumption. This >> @@ -2779,39 +2705,36 @@ static int add_recvbuf_mergeable(struct virtnet_info *vi, >> */ >> len = get_mergeable_buf_len(rq, &rq->mrg_avg_pkt_len, room); >> >> - if (unlikely(!skb_page_frag_refill(len + room, alloc_frag, gfp))) >> - return -ENOMEM; >> - >> - if (!alloc_frag->offset && len + room + sizeof(struct virtnet_rq_dma) > alloc_frag->size) >> - len -= sizeof(struct virtnet_rq_dma); >> - >> - buf = virtnet_rq_alloc(rq, len + room, gfp); >> + alloc_len = len + room; >> + buf = page_pool_alloc_va(rq->page_pool, &alloc_len, gfp); >> if (unlikely(!buf)) >> return -ENOMEM; >> >> buf += headroom; /* advance address leaving hole at front of pkt */ >> - hole = alloc_frag->size - alloc_frag->offset; >> - if (hole < len + room) { >> - /* To avoid internal fragmentation, if there is very likely not >> - * enough space for another buffer, add the remaining space to >> - * the current buffer. >> - * XDP core assumes that frame_size of xdp_buff and the length >> - * of the frag are PAGE_SIZE, so we disable the hole mechanism. >> - */ >> - if (!headroom) >> - len += hole; >> - alloc_frag->offset += hole; >> - } >> >> - virtnet_rq_init_one_sg(rq, buf, len); >> + if (!headroom) >> + len = alloc_len - room; >> >> ctx = mergeable_len_to_ctx(len + room, headroom); >> - err = virtqueue_add_inbuf_premapped(rq->vq, rq->sg, 1, buf, ctx, gfp); >> - if (err < 0) { >> - virtnet_rq_unmap(rq, buf, 0); >> - put_page(virt_to_head_page(buf)); >> + >> + if (rq->use_page_pool_dma) { >> + page = virt_to_head_page(buf); >> + addr = page_pool_get_dma_addr(page) + >> + (buf - (char *)page_address(page)); >> + >> + sg_init_table(rq->sg, 1); >> + sg_fill_dma(rq->sg, addr, len); >> + err = virtqueue_add_inbuf_premapped(rq->vq, rq->sg, 1, >> + buf, ctx, gfp); >> + } else { >> + sg_init_one(rq->sg, buf, len); >> + err = virtqueue_add_inbuf_ctx(rq->vq, rq->sg, 1, >> + buf, ctx, gfp); >> } >> >> + if (err < 0) >> + page_pool_put_page(rq->page_pool, virt_to_head_page(buf), >> + -1, false); >> return err; >> } >> >> @@ -3128,7 +3051,10 @@ static int virtnet_enable_queue_pair(struct virtnet_info *vi, int qp_index) >> return err; >> >> err = xdp_rxq_info_reg_mem_model(&vi->rq[qp_index].xdp_rxq, >> - MEM_TYPE_PAGE_SHARED, NULL); >> + vi->rq[qp_index].page_pool ? >> + MEM_TYPE_PAGE_POOL : >> + MEM_TYPE_PAGE_SHARED, >> + vi->rq[qp_index].page_pool); >> if (err < 0) >> goto err_xdp_reg_mem_model; >> >> @@ -3168,6 +3094,81 @@ static void virtnet_update_settings(struct virtnet_info *vi) >> vi->duplex = duplex; >> } >> >> +static int virtnet_create_page_pools(struct virtnet_info *vi) >> +{ >> + int i, err; >> + >> + if (!vi->mergeable_rx_bufs && vi->big_packets) >> + return 0; >> + >> + for (i = 0; i < vi->max_queue_pairs; i++) { >> + struct receive_queue *rq = &vi->rq[i]; >> + struct page_pool_params pp_params = { 0 }; >> + struct device *dma_dev; >> + >> + if (rq->page_pool) >> + continue; >> + >> + if (rq->xsk_pool) >> + continue; >> + >> + pp_params.order = 0; >> + pp_params.pool_size = virtqueue_get_vring_size(rq->vq); >> + pp_params.nid = dev_to_node(vi->vdev->dev.parent); >> + pp_params.netdev = vi->dev; >> + pp_params.napi = &rq->napi; >> + >> + /* Check if backend supports DMA API (e.g., vhost, virtio-pci). >> + * If so, use page_pool's DMA mapping for premapped buffers. >> + * Otherwise (e.g., VDUSE), page_pool only handles allocation. >> + */ >> + dma_dev = virtqueue_dma_dev(rq->vq); >> + if (dma_dev) { >> + pp_params.dev = dma_dev; >> + pp_params.flags = PP_FLAG_DMA_MAP; >> + pp_params.dma_dir = DMA_FROM_DEVICE; >> + rq->use_page_pool_dma = true; >> + } else { >> + pp_params.dev = vi->vdev->dev.parent; >> + pp_params.flags = 0; >> + rq->use_page_pool_dma = false; > > Can the page pool handles dma with vi->vdev->dev.parent? No, we cannot use the page_pool DMA with vi->vdev->dev.parent in VDUSE case because VDUSE uses its own address translation. virtqueue_dma_dev() returns NULL, virtio doesn't use standard DMA API at all. Now that I think about it, setting pp_params.dev in this branch is unnecessary since it is never accessed. I can remove it, if you prefer. > > Thanks. > >> + } >> + >> + rq->page_pool = page_pool_create(&pp_params); >> + if (IS_ERR(rq->page_pool)) { >> + err = PTR_ERR(rq->page_pool); >> + rq->page_pool = NULL; >> + goto err_cleanup; >> + } >> + } >> + return 0; >> + >> +err_cleanup: >> + while (--i >= 0) { >> + struct receive_queue *rq = &vi->rq[i]; >> + >> + if (rq->page_pool) { >> + page_pool_destroy(rq->page_pool); >> + rq->page_pool = NULL; >> + } >> + } >> + return err; >> +} >> + >> +static void virtnet_destroy_page_pools(struct virtnet_info *vi) >> +{ >> + int i; >> + >> + for (i = 0; i < vi->max_queue_pairs; i++) { >> + struct receive_queue *rq = &vi->rq[i]; >> + >> + if (rq->page_pool) { >> + page_pool_destroy(rq->page_pool); >> + rq->page_pool = NULL; >> + } >> + } >> +} >> + >> static int virtnet_open(struct net_device *dev) >> { >> struct virtnet_info *vi = netdev_priv(dev); >> @@ -6287,17 +6288,6 @@ static void free_receive_bufs(struct virtnet_info *vi) >> rtnl_unlock(); >> } >> >> -static void free_receive_page_frags(struct virtnet_info *vi) >> -{ >> - int i; >> - for (i = 0; i < vi->max_queue_pairs; i++) >> - if (vi->rq[i].alloc_frag.page) { >> - if (vi->rq[i].last_dma) >> - virtnet_rq_unmap(&vi->rq[i], vi->rq[i].last_dma, 0); >> - put_page(vi->rq[i].alloc_frag.page); >> - } >> -} >> - >> static void virtnet_sq_free_unused_buf(struct virtqueue *vq, void *buf) >> { >> struct virtnet_info *vi = vq->vdev->priv; >> @@ -6441,10 +6431,8 @@ static int virtnet_find_vqs(struct virtnet_info *vi) >> vi->rq[i].min_buf_len = mergeable_min_buf_len(vi, vi->rq[i].vq); >> vi->sq[i].vq = vqs[txq2vq(i)]; >> } >> - >> /* run here: ret == 0. */ >> >> - >> err_find: >> kfree(ctx); >> err_ctx: >> @@ -6945,6 +6933,14 @@ static int virtnet_probe(struct virtio_device *vdev) >> goto free; >> } >> >> + /* Create page pools for receive queues. >> + * Page pools are created at probe time so they can be used >> + * with premapped DMA addresses throughout the device lifetime. >> + */ >> + err = virtnet_create_page_pools(vi); >> + if (err) >> + goto free_irq_moder; >> + >> #ifdef CONFIG_SYSFS >> if (vi->mergeable_rx_bufs) >> dev->sysfs_rx_queue_group = &virtio_net_mrg_rx_group; >> @@ -6958,7 +6954,7 @@ static int virtnet_probe(struct virtio_device *vdev) >> vi->failover = net_failover_create(vi->dev); >> if (IS_ERR(vi->failover)) { >> err = PTR_ERR(vi->failover); >> - goto free_vqs; >> + goto free_page_pools; >> } >> } >> >> @@ -7075,9 +7071,11 @@ static int virtnet_probe(struct virtio_device *vdev) >> unregister_netdev(dev); >> free_failover: >> net_failover_destroy(vi->failover); >> -free_vqs: >> +free_page_pools: >> + virtnet_destroy_page_pools(vi); >> +free_irq_moder: >> + virtnet_free_irq_moder(vi); >> virtio_reset_device(vdev); >> - free_receive_page_frags(vi); >> virtnet_del_vqs(vi); >> free: >> free_netdev(dev); >> @@ -7102,7 +7100,7 @@ static void remove_vq_common(struct virtnet_info *vi) >> >> free_receive_bufs(vi); >> >> - free_receive_page_frags(vi); >> + virtnet_destroy_page_pools(vi); >> >> virtnet_del_vqs(vi); >> } >> -- >> 2.47.3 >>