public inbox for linux-arch@vger.kernel.org
 help / color / mirror / Atom feed
From: Ralf Baechle <ralf@linux-mips.org>
To: Haavard Skinnemoen <haavard.skinnemoen@atmel.com>,
	linux-arch@vger.kernel.org, Atsushi Nemoto <anemo@mba.ocn.ne.jp>,
	dan.j.williams@intel.com, linux-kernel@vger.kernel.org,
	maciej.sos
Subject: Re: [PATCH] dmatest: flush and invalidate destination buffer before DMA
Date: Sun, 11 Jan 2009 18:44:02 +0000	[thread overview]
Message-ID: <20090111184402.GA3810@linux-mips.org> (raw)
In-Reply-To: <20090109222721.GA3618@flint.arm.linux.org.uk>

On Fri, Jan 09, 2009 at 10:27:21PM +0000, Russell King wrote:

> On Fri, Jan 09, 2009 at 11:19:36AM +0000, Ralf Baechle wrote:
> > On Thu, Jan 08, 2009 at 09:36:03AM +0100, Haavard Skinnemoen wrote:
> > > In the general case, however, I think MIPS has a bug: I've seen drivers
> > > DMA to/from tiny buffers stored inside another struct. This is legal
> > > because the driver can guarantee that the other fields in the struct
> > > aren't accessed in the mean time, but any fields sharing a cacheline
> > > with the buffer must be written back before the lines are invalidated.
> > 
> > Depending on the implementation details, the use of such a struct might be
> > relying on implementation-specific behaviour.  This is what
> > Documentation/DMA-API.txt has to say:
> > 
> > [...]
> > int
> > dma_get_cache_alignment(void)
> > 
> > Returns the processor cache alignment.  This is the absolute minimum
> > alignment *and* width that you must observe when either mapping
> > memory or doing partial flushes.
> > 
> > Notes: This API may return a number *larger* than the actual cache
> > line, but it will guarantee that one or more cache lines fit exactly
> > into the width returned by this call.  It will also always be a power
> > of two for easy alignment.
> > [...]
> > 
> > Since dma_get_cache_alignment() is a function not a constant its result
> > can't be used in the definition of a struct unless possibly excessive
> > padding is used.
> > 
> > The debate has shown that we problably need BUG_ON() assertions in the
> > DMA API implementations to catch this sort of dangerous use.
> 
> I really don't think that's a realistic option.  You're asking for
> every call to the DMA API to ensure that the buffer and length are
> a multiple of the cache line size.
> 
> So, what happens if, eg, SPI wants to send a 16 byte buffer, and your
> cache lines are 64 bytes?  Does the SPI driver have to kmalloc a new
> chunk of memory 64 bytes long and copy the data into that before
> passing it into the DMA API?
> 
> If you start enforcing that kind of thing, I think the cache coherent
> people will take violent exception and refuse to play such games - and
> quite rightly so.

I only want to force people to be aware of what they're doing.  So far I've
seen cache lines of up to 256 bytes in size on non-coherent systems.  Be
paranoid, very paranoid ...

Below patch should solve Dan William's concerns.  It will peform a writeback
and invalidation operation on the first and last cacheline worth of data.
The instruction costs around a dozen cycles so I won't even try to optimize
possible double cache operations away; that'd probably be more expensive.

  Ralf

Signed-off-by: Ralf Baechle <ralf@linux-mips.org>

diff --git a/arch/mips/mm/c-r4k.c b/arch/mips/mm/c-r4k.c
index 6e99665..56290a7 100644
--- a/arch/mips/mm/c-r4k.c
+++ b/arch/mips/mm/c-r4k.c
@@ -618,8 +618,11 @@ static void r4k_dma_cache_inv(unsigned long addr, unsigned long size)
 	if (cpu_has_inclusive_pcaches) {
 		if (size >= scache_size)
 			r4k_blast_scache();
-		else
+		else {
+			cache_op(Hit_Writeback_Inv_SD, addr);
+			cache_op(Hit_Writeback_Inv_SD, addr + size - 1);
 			blast_inv_scache_range(addr, addr + size);
+		}
 		return;
 	}
 
@@ -627,6 +630,8 @@ static void r4k_dma_cache_inv(unsigned long addr, unsigned long size)
 		r4k_blast_dcache();
 	} else {
 		R4600_HIT_CACHEOP_WAR_IMPL;
+		cache_op(Hit_Writeback_Inv_D, addr);
+		cache_op(Hit_Writeback_Inv_D, addr + size - 1);
 		blast_inv_dcache_range(addr, addr + size);
 	}
 

WARNING: multiple messages have this Message-ID (diff)
From: Ralf Baechle <ralf@linux-mips.org>
To: Haavard Skinnemoen <haavard.skinnemoen@atmel.com>,
	linux-arch@vger.kernel.org, Atsushi Nemoto <anemo@mba.ocn.ne.jp>,
	dan.j.williams@intel.com, linux-kernel@vger.kernel.org,
	maciej.sosnowski@intel.com
Subject: Re: [PATCH] dmatest: flush and invalidate destination buffer before DMA
Date: Sun, 11 Jan 2009 18:44:02 +0000	[thread overview]
Message-ID: <20090111184402.GA3810@linux-mips.org> (raw)
Message-ID: <20090111184402.fZCTdVnx3JUrW4zFjKDUTk_Z3ZPH_kwcul5L6TgrqK0@z> (raw)
In-Reply-To: <20090109222721.GA3618@flint.arm.linux.org.uk>

On Fri, Jan 09, 2009 at 10:27:21PM +0000, Russell King wrote:

> On Fri, Jan 09, 2009 at 11:19:36AM +0000, Ralf Baechle wrote:
> > On Thu, Jan 08, 2009 at 09:36:03AM +0100, Haavard Skinnemoen wrote:
> > > In the general case, however, I think MIPS has a bug: I've seen drivers
> > > DMA to/from tiny buffers stored inside another struct. This is legal
> > > because the driver can guarantee that the other fields in the struct
> > > aren't accessed in the mean time, but any fields sharing a cacheline
> > > with the buffer must be written back before the lines are invalidated.
> > 
> > Depending on the implementation details, the use of such a struct might be
> > relying on implementation-specific behaviour.  This is what
> > Documentation/DMA-API.txt has to say:
> > 
> > [...]
> > int
> > dma_get_cache_alignment(void)
> > 
> > Returns the processor cache alignment.  This is the absolute minimum
> > alignment *and* width that you must observe when either mapping
> > memory or doing partial flushes.
> > 
> > Notes: This API may return a number *larger* than the actual cache
> > line, but it will guarantee that one or more cache lines fit exactly
> > into the width returned by this call.  It will also always be a power
> > of two for easy alignment.
> > [...]
> > 
> > Since dma_get_cache_alignment() is a function not a constant its result
> > can't be used in the definition of a struct unless possibly excessive
> > padding is used.
> > 
> > The debate has shown that we problably need BUG_ON() assertions in the
> > DMA API implementations to catch this sort of dangerous use.
> 
> I really don't think that's a realistic option.  You're asking for
> every call to the DMA API to ensure that the buffer and length are
> a multiple of the cache line size.
> 
> So, what happens if, eg, SPI wants to send a 16 byte buffer, and your
> cache lines are 64 bytes?  Does the SPI driver have to kmalloc a new
> chunk of memory 64 bytes long and copy the data into that before
> passing it into the DMA API?
> 
> If you start enforcing that kind of thing, I think the cache coherent
> people will take violent exception and refuse to play such games - and
> quite rightly so.

I only want to force people to be aware of what they're doing.  So far I've
seen cache lines of up to 256 bytes in size on non-coherent systems.  Be
paranoid, very paranoid ...

Below patch should solve Dan William's concerns.  It will peform a writeback
and invalidation operation on the first and last cacheline worth of data.
The instruction costs around a dozen cycles so I won't even try to optimize
possible double cache operations away; that'd probably be more expensive.

  Ralf

Signed-off-by: Ralf Baechle <ralf@linux-mips.org>

diff --git a/arch/mips/mm/c-r4k.c b/arch/mips/mm/c-r4k.c
index 6e99665..56290a7 100644
--- a/arch/mips/mm/c-r4k.c
+++ b/arch/mips/mm/c-r4k.c
@@ -618,8 +618,11 @@ static void r4k_dma_cache_inv(unsigned long addr, unsigned long size)
 	if (cpu_has_inclusive_pcaches) {
 		if (size >= scache_size)
 			r4k_blast_scache();
-		else
+		else {
+			cache_op(Hit_Writeback_Inv_SD, addr);
+			cache_op(Hit_Writeback_Inv_SD, addr + size - 1);
 			blast_inv_scache_range(addr, addr + size);
+		}
 		return;
 	}
 
@@ -627,6 +630,8 @@ static void r4k_dma_cache_inv(unsigned long addr, unsigned long size)
 		r4k_blast_dcache();
 	} else {
 		R4600_HIT_CACHEOP_WAR_IMPL;
+		cache_op(Hit_Writeback_Inv_D, addr);
+		cache_op(Hit_Writeback_Inv_D, addr + size - 1);
 		blast_inv_dcache_range(addr, addr + size);
 	}
 

  reply	other threads:[~2009-01-11 18:44 UTC|newest]

Thread overview: 4+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
     [not found] <20081227111037.3bd13adc@hskinnemoen-d830>
     [not found] ` <20081229.025352.01917409.anemo@mba.ocn.ne.jp>
     [not found]   ` <e9c3a7c20901051031y528d0d31r18d44c5096c59e0@mail.gmail.com>
     [not found]     ` <20090108.134336.127659765.nemoto@toshiba-tops.co.jp>
     [not found]       ` <20090108093603.691c1200@hskinnemoen-d830>
2009-01-09 11:19         ` [PATCH] dmatest: flush and invalidate destination buffer before DMA Ralf Baechle
2009-01-09 22:27           ` Russell King
2009-01-11 18:44             ` Ralf Baechle [this message]
2009-01-11 18:44               ` Ralf Baechle

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20090111184402.GA3810@linux-mips.org \
    --to=ralf@linux-mips.org \
    --cc=anemo@mba.ocn.ne.jp \
    --cc=dan.j.williams@intel.com \
    --cc=haavard.skinnemoen@atmel.com \
    --cc=linux-arch@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox