DMA Engine development
 help / color / mirror / Atom feed
From: Jonathan Cameron <jic23@kernel.org>
To: Paul Cercueil <paul@crapouillou.net>
Cc: "Lars-Peter Clausen" <lars@metafoo.de>,
	"Vinod Koul" <vkoul@kernel.org>,
	"Michael Hennerich" <Michael.Hennerich@analog.com>,
	"Nuno Sá" <noname.nuno@gmail.com>,
	"Sumit Semwal" <sumit.semwal@linaro.org>,
	"Christian König" <christian.koenig@amd.com>,
	linux-kernel@vger.kernel.org, dmaengine@vger.kernel.org,
	linux-iio@vger.kernel.org, linux-media@vger.kernel.org,
	dri-devel@lists.freedesktop.org, linaro-mm-sig@lists.linaro.org
Subject: Re: [PATCH v3 10/11] iio: buffer-dmaengine: Support new DMABUF based userspace API
Date: Sun, 16 Apr 2023 16:15:27 +0100	[thread overview]
Message-ID: <20230416161527.6e7021f1@jic23-huawei> (raw)
In-Reply-To: <20230403154955.216148-1-paul@crapouillou.net>

On Mon,  3 Apr 2023 17:49:54 +0200
Paul Cercueil <paul@crapouillou.net> wrote:

> Use the functions provided by the buffer-dma core to implement the
> DMABUF userspace API in the buffer-dmaengine IIO buffer implementation.
> 
> Since we want to be able to transfer an arbitrary number of bytes and
> not necesarily the full DMABUF, the associated scatterlist is converted
> to an array of DMA addresses + lengths, which is then passed to
> dmaengine_prep_slave_dma_array().
> 
> Signed-off-by: Paul Cercueil <paul@crapouillou.net>
A few things inline.

Thanks,

Jonathan

> 
> ---
> v3: Use the new dmaengine_prep_slave_dma_array(), and adapt the code to
>     work with the new functions introduced in industrialio-buffer-dma.c.
> ---
>  .../buffer/industrialio-buffer-dmaengine.c    | 69 ++++++++++++++++---
>  include/linux/iio/buffer-dma.h                |  2 +
>  2 files changed, 60 insertions(+), 11 deletions(-)
> 
> diff --git a/drivers/iio/buffer/industrialio-buffer-dmaengine.c b/drivers/iio/buffer/industrialio-buffer-dmaengine.c
> index 866c8b84bb24..faed9c2b089c 100644
> --- a/drivers/iio/buffer/industrialio-buffer-dmaengine.c
> +++ b/drivers/iio/buffer/industrialio-buffer-dmaengine.c
> @@ -65,25 +65,68 @@ static int iio_dmaengine_buffer_submit_block(struct iio_dma_buffer_queue *queue,
>  		iio_buffer_to_dmaengine_buffer(&queue->buffer);
>  	struct dma_async_tx_descriptor *desc;
>  	enum dma_transfer_direction dma_dir;
> +	unsigned int i, nents, *lenghts;
> +	struct scatterlist *sgl;
> +	unsigned long flags;
> +	dma_addr_t *addrs;
>  	size_t max_size;
>  	dma_cookie_t cookie;
> +	size_t len_total;
>  
> -	max_size = min(block->size, dmaengine_buffer->max_size);
> -	max_size = round_down(max_size, dmaengine_buffer->align);
> +	if (!block->bytes_used)
> +		return -EINVAL;
>  
> -	if (queue->buffer.direction == IIO_BUFFER_DIRECTION_IN) {
> -		block->bytes_used = max_size;
> +	if (queue->buffer.direction == IIO_BUFFER_DIRECTION_IN)
>  		dma_dir = DMA_DEV_TO_MEM;
> -	} else {
> +	else
>  		dma_dir = DMA_MEM_TO_DEV;
> -	}
>  
> -	if (!block->bytes_used || block->bytes_used > max_size)
> -		return -EINVAL;

Ah this is dropping the code I moaned about earlier.  I'll probably
forget though so maybe add a note to that patch saying it goes
away later anyway so I don't keep moaning about it in future versions.

> +	if (block->sg_table) {
> +		sgl = block->sg_table->sgl;
> +		nents = sg_nents_for_len(sgl, block->bytes_used);
> +
> +		addrs = kmalloc_array(nents, sizeof(*addrs), GFP_KERNEL);
> +		if (!addrs)
> +			return -ENOMEM;
> +
> +		lenghts = kmalloc_array(nents, sizeof(*lenghts), GFP_KERNEL);

lengths?

> +		if (!lenghts) {
> +			kfree(addrs);
> +			return -ENOMEM;
> +		}
> +
> +		len_total = block->bytes_used;
>  
> -	desc = dmaengine_prep_slave_single(dmaengine_buffer->chan,
> -		block->phys_addr, block->bytes_used, dma_dir,
> -		DMA_PREP_INTERRUPT);
> +		for (i = 0; i < nents; i++) {
> +			addrs[i] = sg_dma_address(sgl);
> +			lenghts[i] = min(sg_dma_len(sgl), len_total);
> +			len_total -= lenghts[i];
> +
> +			sgl = sg_next(sgl);
> +		}
> +
> +		flags = block->cyclic ? DMA_PREP_REPEAT : DMA_PREP_INTERRUPT;
> +
> +		desc = dmaengine_prep_slave_dma_array(dmaengine_buffer->chan,
> +						      addrs, lenghts, nents,
> +						      dma_dir, flags);
> +		kfree(addrs);
> +		kfree(lenghts);
> +	} else {
> +		max_size = min(block->size, dmaengine_buffer->max_size);
> +		max_size = round_down(max_size, dmaengine_buffer->align);
> +
> +		if (queue->buffer.direction == IIO_BUFFER_DIRECTION_IN)
> +			block->bytes_used = max_size;
> +
> +		if (block->bytes_used > max_size)
> +			return -EINVAL;
> +
> +		desc = dmaengine_prep_slave_single(dmaengine_buffer->chan,
> +						   block->phys_addr,
> +						   block->bytes_used, dma_dir,
> +						   DMA_PREP_INTERRUPT);
> +	}


> diff --git a/include/linux/iio/buffer-dma.h b/include/linux/iio/buffer-dma.h
> index e5e5817e99db..48f7ffaf0867 100644
> --- a/include/linux/iio/buffer-dma.h
> +++ b/include/linux/iio/buffer-dma.h
> @@ -43,6 +43,7 @@ enum iio_block_state {
>   * @queue: Parent DMA buffer queue
>   * @kref: kref used to manage the lifetime of block
>   * @state: Current state of the block
> + * @cyclic: True if this is a cyclic buffer
>   * @fileio: True if this buffer is used for fileio mode

I might have commented on it earlier (I've lost track) but
attach should be documented as well.   Worth sanity checking
by either building with W=1 or running kernel-doc over
the files and fixing the warnings.

>   */
>  struct iio_dma_buffer_block {
> @@ -67,6 +68,7 @@ struct iio_dma_buffer_block {
>  	 */
>  	enum iio_block_state state;
>  
> +	bool cyclic;
>  	bool fileio;
>  
>  	struct dma_buf_attachment *attach;


  parent reply	other threads:[~2023-04-16 15:15 UTC|newest]

Thread overview: 32+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-04-03 15:47 [PATCH v3 00/11] iio: new DMABUF based API, v3 Paul Cercueil
2023-04-03 15:47 ` [PATCH v3 01/11] dmaengine: Add API function dmaengine_prep_slave_dma_array() Paul Cercueil
2023-04-12 17:23   ` Vinod Koul
2023-04-13  7:59     ` Paul Cercueil
2023-04-03 15:47 ` [PATCH v3 02/11] dmaengine: dma-axi-dmac: Implement device_prep_slave_dma_array Paul Cercueil
2023-04-03 15:47 ` [PATCH v3 03/11] iio: buffer-dma: Get rid of outgoing queue Paul Cercueil
2023-04-16 14:24   ` Jonathan Cameron
2023-04-18  8:08     ` Paul Cercueil
2023-05-01 16:25       ` Jonathan Cameron
2023-04-03 15:47 ` [PATCH v3 04/11] iio: buffer-dma: Enable buffer write support Paul Cercueil
2023-04-16 14:30   ` Jonathan Cameron
2023-04-03 15:47 ` [PATCH v3 05/11] iio: buffer-dmaengine: Support specifying buffer direction Paul Cercueil
2023-04-16 14:35   ` Jonathan Cameron
2023-04-03 15:47 ` [PATCH v3 06/11] iio: buffer-dmaengine: Enable write support Paul Cercueil
2023-04-16 14:37   ` Jonathan Cameron
2023-04-03 15:47 ` [PATCH v3 07/11] iio: core: Add new DMABUF interface infrastructure Paul Cercueil
2023-04-04  7:32   ` Nuno Sá
2023-04-04  7:55     ` Paul Cercueil
2023-04-04  8:21       ` Nuno Sá
2023-04-04 13:22       ` Lars-Peter Clausen
2023-04-16 15:04   ` Jonathan Cameron
2023-04-03 15:47 ` [PATCH v3 08/11] iio: buffer-dma: split iio_dma_buffer_fileio_free() function Paul Cercueil
2023-04-03 15:47 ` [PATCH v3 09/11] iio: buffer-dma: Enable support for DMABUFs Paul Cercueil
2023-04-16 15:10   ` Jonathan Cameron
2023-04-03 15:49 ` [PATCH v3 10/11] iio: buffer-dmaengine: Support new DMABUF based userspace API Paul Cercueil
2023-04-03 15:49   ` [PATCH v3 11/11] Documentation: iio: Document high-speed DMABUF based API Paul Cercueil
2023-04-03 16:05     ` Jonathan Corbet
2023-04-03 18:37       ` Paul Cercueil
2023-04-16 15:15   ` Jonathan Cameron [this message]
     [not found] ` <20230404015944.502-1-hdanton@sina.com>
2023-04-04  7:42   ` [PATCH v3 01/11] dmaengine: Add API function dmaengine_prep_slave_dma_array() Paul Cercueil
2023-04-04  8:54     ` Christian König
2023-04-04  7:44 ` [PATCH v3 00/11] iio: new DMABUF based API, v3 Nuno Sá

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20230416161527.6e7021f1@jic23-huawei \
    --to=jic23@kernel.org \
    --cc=Michael.Hennerich@analog.com \
    --cc=christian.koenig@amd.com \
    --cc=dmaengine@vger.kernel.org \
    --cc=dri-devel@lists.freedesktop.org \
    --cc=lars@metafoo.de \
    --cc=linaro-mm-sig@lists.linaro.org \
    --cc=linux-iio@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-media@vger.kernel.org \
    --cc=noname.nuno@gmail.com \
    --cc=paul@crapouillou.net \
    --cc=sumit.semwal@linaro.org \
    --cc=vkoul@kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox