From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from smtp.kernel.org (aws-us-west-2-korg-mail-1.web.codeaurora.org [10.30.226.201]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 9F39913A26D; Fri, 14 Mar 2025 18:49:16 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=10.30.226.201 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1741978157; cv=none; b=Rfwp00l7YChLwnfv2D+qvaSMdEVw1AreDkDDDNwQzwyC2a345rvLAWq8TG9Dywqhqyw98QPazUfpYgeY9Al0JzksAc4Mwk44OIfneIJeJUKTsNPLYy4w1zRBPftr5mKlafeLRX7/Qjz5VX8o7Dj9ndpXiatN1qg96M52qDxwdRQ= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1741978157; c=relaxed/simple; bh=O6TQHcNZOIDORUQUex99gbjkbXRSAsx+LziCN6QYCcQ=; h=Date:From:To:Cc:Subject:Message-ID:References:MIME-Version: Content-Type:Content-Disposition:In-Reply-To; b=XRiC/ALWiGdDRnu3t9+TNmp2WE4FzEkuuQpnHJWJ1ue3RLEWuflPvd+mKPiw+yQRyhtZvud2ba2W8pNmOZxDsOJtB8ilWir8dyaU/ycLGwPBAqtl/3Y3JXfH9BKmC1ybq0dn3NNPjGFSytbWyfTDFZFpjRluh1ptVui85nhswDE= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b=rQVmjcjZ; arc=none smtp.client-ip=10.30.226.201 Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b="rQVmjcjZ" Received: by smtp.kernel.org (Postfix) with ESMTPSA id 6C4A3C4CEE9; Fri, 14 Mar 2025 18:49:15 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1741978156; bh=O6TQHcNZOIDORUQUex99gbjkbXRSAsx+LziCN6QYCcQ=; h=Date:From:To:Cc:Subject:References:In-Reply-To:From; b=rQVmjcjZhe/sj5yONxV2qVxL98CsQ/XLSI8I1aZI9aDjryLW+sVGMqcfI9Bca9bOS BjeRuzyi8Km8hWOU7pWxLh0zSj3Q5xmQc8mdYFqkiDMoJN2jPM4GlDtxa49auZ9+wv o4Atu8WInpcl8PAepC/fXyU6iiNJ15uue0LdQRbYK1nJzWh83j90x9tnoa0sRasVK1 38iJcCj/HUFJNu/wqR3UGB0O36uPS2R51CHZ5YoiakTWuHP1tXd0Pvob4wtZn293CI sVep3kKdHjc3A4MK0/PKcdqXaMCX1hnu3s63bSmSw7dabMqNw6WwX2YJoxC1T+nsVF 6r+AVkQEbO0Yg== Date: Fri, 14 Mar 2025 20:49:11 +0200 From: Leon Romanovsky To: Marek Szyprowski Cc: Robin Murphy , Christoph Hellwig , Jason Gunthorpe , Jens Axboe , Joerg Roedel , Will Deacon , Sagi Grimberg , Keith Busch , Bjorn Helgaas , Logan Gunthorpe , Yishai Hadas , Shameer Kolothum , Kevin Tian , Alex Williamson , =?iso-8859-1?B?Suly9G1l?= Glisse , Andrew Morton , Jonathan Corbet , linux-doc@vger.kernel.org, linux-kernel@vger.kernel.org, linux-block@vger.kernel.org, linux-rdma@vger.kernel.org, iommu@lists.linux.dev, linux-nvme@lists.infradead.org, linux-pci@vger.kernel.org, kvm@vger.kernel.org, linux-mm@kvack.org, Randy Dunlap Subject: Re: [PATCH v7 00/17] Provide a new two step DMA mapping API Message-ID: <20250314184911.GR1322339@unreal> References: <20250220124827.GR53094@unreal> <1166a5f5-23cc-4cce-ba40-5e10ad2606de@arm.com> <20250312193249.GI1322339@unreal> Precedence: bulk X-Mailing-List: linux-pci@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Type: text/plain; charset=iso-8859-1 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: On Fri, Mar 14, 2025 at 11:52:58AM +0100, Marek Szyprowski wrote: > On 12.03.2025 20:32, Leon Romanovsky wrote: > > On Wed, Mar 12, 2025 at 10:28:32AM +0100, Marek Szyprowski wrote: > >> Hi Robin > >> > >> On 28.02.2025 20:54, Robin Murphy wrote: > >>> On 20/02/2025 12:48 pm, Leon Romanovsky wrote: > >>>> On Wed, Feb 05, 2025 at 04:40:20PM +0200, Leon Romanovsky wrote: > >>>>> From: Leon Romanovsky > >>>>> > >>>>> Changelog: > >>>>> v7: > >>>>>   * Rebased to v6.14-rc1 > >>>> <...> > >>>> > >>>>> Christoph Hellwig (6): > >>>>>    PCI/P2PDMA: Refactor the p2pdma mapping helpers > >>>>>    dma-mapping: move the PCI P2PDMA mapping helpers to pci-p2pdma.h > >>>>>    iommu: generalize the batched sync after map interface > >>>>>    iommu/dma: Factor out a iommu_dma_map_swiotlb helper > >>>>>    dma-mapping: add a dma_need_unmap helper > >>>>>    docs: core-api: document the IOVA-based API > >>>>> > >>>>> Leon Romanovsky (11): > >>>>>    iommu: add kernel-doc for iommu_unmap and iommu_unmap_fast > >>>>>    dma-mapping: Provide an interface to allow allocate IOVA > >>>>>    dma-mapping: Implement link/unlink ranges API > >>>>>    mm/hmm: let users to tag specific PFN with DMA mapped bit > >>>>>    mm/hmm: provide generic DMA managing logic > >>>>>    RDMA/umem: Store ODP access mask information in PFN > >>>>>    RDMA/core: Convert UMEM ODP DMA mapping to caching IOVA and page > >>>>>      linkage > >>>>>    RDMA/umem: Separate implicit ODP initialization from explicit ODP > >>>>>    vfio/mlx5: Explicitly use number of pages instead of allocated > >>>>> length > >>>>>    vfio/mlx5: Rewrite create mkey flow to allow better code reuse > >>>>>    vfio/mlx5: Enable the DMA link API > >>>>> > >>>>>   Documentation/core-api/dma-api.rst   |  70 ++++ > >>>>   drivers/infiniband/core/umem_odp.c   | 250 +++++--------- > >>>>>   drivers/infiniband/hw/mlx5/mlx5_ib.h |  12 +- > >>>>>   drivers/infiniband/hw/mlx5/odp.c     |  65 ++-- > >>>>>   drivers/infiniband/hw/mlx5/umr.c     |  12 +- > >>>>>   drivers/iommu/dma-iommu.c            | 468 > >>>>> +++++++++++++++++++++++---- > >>>>>   drivers/iommu/iommu.c                |  84 ++--- > >>>>>   drivers/pci/p2pdma.c                 |  38 +-- > >>>>>   drivers/vfio/pci/mlx5/cmd.c          | 375 +++++++++++---------- > >>>>>   drivers/vfio/pci/mlx5/cmd.h          |  35 +- > >>>>>   drivers/vfio/pci/mlx5/main.c         |  87 +++-- > >>>>>   include/linux/dma-map-ops.h          |  54 ---- > >>>>>   include/linux/dma-mapping.h          |  85 +++++ > >>>>>   include/linux/hmm-dma.h              |  33 ++ > >>>>>   include/linux/hmm.h                  |  21 ++ > >>>>>   include/linux/iommu.h                |   4 + > >>>>>   include/linux/pci-p2pdma.h           |  84 +++++ > >>>>>   include/rdma/ib_umem_odp.h           |  25 +- > >>>>>   kernel/dma/direct.c                  |  44 +-- > >>>>>   kernel/dma/mapping.c                 |  18 ++ > >>>>>   mm/hmm.c                             | 264 +++++++++++++-- > >>>>>   21 files changed, 1435 insertions(+), 693 deletions(-) > >>>>>   create mode 100644 include/linux/hmm-dma.h > >>>> Kind reminder. > > <...> > > > >> Removing the need for scatterlists was advertised as the main goal of > >> this new API, but it looks that similar effects can be achieved with > >> just iterating over the pages and calling page-based DMA API directly. > > Such iteration can't be enough because P2P pages don't have struct pages, > > so you can't use reliably and efficiently dma_map_page_attrs() call. > > > > The only way to do so is to use dma_map_sg_attrs(), which relies on SG > > (the one that we want to remove) to map P2P pages. > > That's something I don't get yet. How P2P pages can be used with > dma_map_sg_attrs(), but not with dma_map_page_attrs()? Both operate > internally on struct page pointer. Yes, and no. See users of is_pci_p2pdma_page(...) function. In dma_*_sg() APIs, there is a real check and support for p2p. In dma_map_page_attrs() variants, this support is missing (ignored, or error is returned). > > >> Maybe I missed something. I still see some advantages in this DMA API > >> extension, but I would also like to see the clear benefits from > >> introducing it, like perf logs or other benchmark summary. > > We didn't focus yet on performance, however Christoph mentioned in his > > block RFC [1] that even simple conversion should improve performance as > > we are performing one P2P lookup per-bio and not per-SG entry as was > > before [2]. In addition it decreases memory [3] too. > > > > [1] https://lore.kernel.org/all/cover.1730037261.git.leon@kernel.org/ > > [2] https://lore.kernel.org/all/34d44537a65aba6ede215a8ad882aeee028b423a.1730037261.git.leon@kernel.org/ > > [3] https://lore.kernel.org/all/383557d0fa1aa393dbab4e1daec94b6cced384ab.1730037261.git.leon@kernel.org/ > > > > So clear benefits are: > > 1. Ability to use native for subsystem structure, e.g. bio for block, > > umem for RDMA, dmabuf for DRM, e.t.c. It removes current wasteful > > conversions from and to SG in order to work with DMA API. > > 2. Batched request and iotlb sync optimizations (perform only once). > > 3. Avoid very expensive call to pgmap pointer. > > 4. Expose MMIO over VFIO without hacks (PCI BAR doesn't have struct pages). > > See this series for such a hack > > https://lore.kernel.org/all/20250307052248.405803-1-vivek.kasireddy@intel.com/ > > I see those benefits and I admit that for typical DMA-with-IOMMU case it > would improve some things. I think that main concern from Robin was how > to handle it for the cases without an IOMMU. In such case, we fallback to non-IOMMU flow (old, well-established one). See this HMM patch as an example https://lore.kernel.org/all/a796da065fa8a9cb35d591ce6930400619572dcc.1738765879.git.leonro@nvidia.com/ +dma_addr_t hmm_dma_map_pfn(struct device *dev, struct hmm_dma_map *map, + size_t idx, + struct pci_p2pdma_map_state *p2pdma_state) ... + if (dma_use_iova(state)) { ... + } else { ... + dma_addr = dma_map_page(dev, page, 0, map->dma_entry_size, + DMA_BIDIRECTIONAL); Thanks > > Best regards > -- > Marek Szyprowski, PhD > Samsung R&D Institute Poland >