From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1751953AbbIQTHx (ORCPT ); Thu, 17 Sep 2015 15:07:53 -0400 Received: from mx1.redhat.com ([209.132.183.28]:48894 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751478AbbIQTHw (ORCPT ); Thu, 17 Sep 2015 15:07:52 -0400 Date: Thu, 17 Sep 2015 15:07:47 -0400 From: Jerome Glisse To: Konrad Rzeszutek Wilk Cc: iommu@lists.linux-foundation.org, Joerg Roedel , Alex Deucher , Dave Airlie , linux-kernel@vger.kernel.org Subject: Re: [RFC PATCH] dma/swiotlb: Add helper for device driver to opt-out from swiotlb. Message-ID: <20150917190746.GA6699@redhat.com> References: <1442514158-30281-1-git-send-email-jglisse@redhat.com> <20150917190251.GE20952@x230.dumpdata.com> MIME-Version: 1.0 Content-Type: text/plain; charset=iso-8859-1 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: <20150917190251.GE20952@x230.dumpdata.com> User-Agent: Mutt/1.5.23 (2014-03-12) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, Sep 17, 2015 at 03:02:51PM -0400, Konrad Rzeszutek Wilk wrote: > On Thu, Sep 17, 2015 at 02:22:38PM -0400, jglisse@redhat.com wrote: > > From: Jérôme Glisse > > > > The swiotlb dma backend is not appropriate for some devices like > > GPU where bounce buffer or slow dma page allocations is just not > > acceptable. With that helper device drivers can opt-out from the > > swiotlb and just do sane things without wasting CPU cycles inside > > the swiotlb code. > > What if SWIOTLB is the only one available? On x86 no_mmu is always available and we assume that device driver that would use this knows that their device can access all memory with no restriction or at very least use DMA32 gfp flag. > And what can't the devices use the TTM DMA backend which sets up > buffers which don't need bounce buffer or slow dma page allocations? We want to get rid of this TTM code path for radeon and likely nouveau. This is the motivation for that patch. Benchmark shows that the TTM DMA backend is much much much slower (20% on some benchmark) that the regular page allocation and going through no_mmu. So this is all about allowing to directly allocate page through regular kernel page alloc code and not through specialize dma allocator. Cheers, Jérôme