From mboxrd@z Thu Jan 1 00:00:00 1970 From: Ingo Molnar Subject: Re: [PATCH 03/10] x86: add initialization code for DMA-API debugging Date: Sun, 23 Nov 2008 12:35:43 +0100 Message-ID: <20081123113543.GD29663@elte.hu> References: <1227284770-19215-1-git-send-email-joerg.roedel@amd.com> <1227284770-19215-4-git-send-email-joerg.roedel@amd.com> <20081121174348.GB4336@elte.hu> <20081122094807.GK29705@8bytes.org> <20081123112818.GC29663@elte.hu> Mime-Version: 1.0 Content-Type: text/plain; charset=us-ascii Cc: Joerg Roedel , netdev@vger.kernel.org, Thomas Gleixner , Ingo Molnar , linux-kernel@vger.kernel.org, iommu@lists.linux-foundation.org To: Joerg Roedel Return-path: Received: from mx2.mail.elte.hu ([157.181.151.9]:45001 "EHLO mx2.mail.elte.hu" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1758763AbYKWLf6 (ORCPT ); Sun, 23 Nov 2008 06:35:58 -0500 Content-Disposition: inline In-Reply-To: <20081123112818.GC29663@elte.hu> Sender: netdev-owner@vger.kernel.org List-ID: * Ingo Molnar wrote: > > * Joerg Roedel wrote: > > > On Fri, Nov 21, 2008 at 06:43:48PM +0100, Ingo Molnar wrote: > > > > > > * Joerg Roedel wrote: > > > > > > > +static struct list_head dma_entry_hash[HASH_SIZE]; > > > > + > > > > +/* A slab cache to allocate dma_map_entries fast */ > > > > +static struct kmem_cache *dma_entry_cache; > > > > + > > > > +/* lock to protect the data structures */ > > > > +static DEFINE_SPINLOCK(dma_lock); > > > > > > some more generic comments about the data structure: it's main purpose > > > is to provide a mapping based on (dev,addr). There's little if any > > > cross-entry interaction - same-address+same-dev DMA is checked. > > > > > > 1) > > > > > > the hash: > > > > > > + return (entry->dev_addr >> HASH_FN_SHIFT) & HASH_FN_MASK; > > > > > > should mix in entry->dev as well - that way we get not just per > > > address but per device hash space separation as well. > > > > > > 2) > > > > > > HASH_FN_SHIFT is 1MB chunks right now - that's probably fine in > > > practice albeit perhaps a bit too small. There's seldom any coherency > > > between the physical addresses of DMA - we rarely have any real > > > (performance-relevant) physical co-location of DMA addresses beyond 4K > > > granularity. So using 1MB chunking here will discard a good deal of > > > random low bits we should be hashing on. > > > > > > 3) > > > > > > And the most scalable locking would be per hash bucket locking - no > > > global lock is needed. The bucket hash heads should probably be > > > cacheline sized - so we'd get one lock per bucket. > > > > Hmm, I just had the idea of saving this data in struct device. How > > about that? The locking should scale too and we can extend it > > easier. For example it simplifys a per-device disable function for > > the checking. Or another future feature might be leak tracing. > > that will help with spreading the hash across devices, but brings in > lifetime issues: you must be absolutely sure all DMA has drained at > the point a device is deinitialized. Note that obviously proper DMA quiescence is a must-have during device dinit anyway, but still, it's an extra complication to init/deinit the hashes, etc. Ingo