From: Alexandru Ardelean <alexandru.ardelean@analog.com>
To: "vkoul@kernel.org" <vkoul@kernel.org>
Cc: "dmaengine@vger.kernel.org" <dmaengine@vger.kernel.org>
Subject: dmaengine: dmatest: wrap src & dst data into a struct
Date: Mon, 21 Jan 2019 07:54:55 +0000 [thread overview]
Message-ID: <de4d0ed1f5c9923afe19f31f87b2c7ed3192d005.camel@analog.com> (raw)
On Sat, 2019-01-19 at 18:37 +0530, Vinod Koul wrote:
>
>
> Hi Alexandru,
>
> On 26-11-18, 10:43, Alexandru Ardelean wrote:
> > There's a bit too much un-wrapped & duplicated code that can be
> > organized
> > into some common logic/functions.
> >
> > This change wraps all the common parts between srcs & dsts into a
> > `dmatest_data` struct. The purpose is to split the `dmatestfunc` into
> > smaller chunks that are easier to follow & extend.
>
> Thanks for the patch, this looks good but somehow seems to have slipped
> by..
Hey,
Replies inline
>
> >
> > Signed-off-by: Alexandru Ardelean <alexandru.ardelean@analog.com>
> > ---
> > drivers/dma/dmatest.c | 257 ++++++++++++++++++++++--------------------
> > 1 file changed, 133 insertions(+), 124 deletions(-)
> >
> > diff --git a/drivers/dma/dmatest.c b/drivers/dma/dmatest.c
> > index e71aa1e3451c..c37c643e7d29 100644
> > --- a/drivers/dma/dmatest.c
> > +++ b/drivers/dma/dmatest.c
> > @@ -166,15 +166,20 @@ struct dmatest_done {
> > wait_queue_head_t *wait;
> > };
> >
> > +struct dmatest_data {
> > + u8 **raw;
> > + u8 **aligned;
> > + unsigned int cnt;
> > + unsigned int off;
> > +};
> > +
> > struct dmatest_thread {
> > struct list_head node;
> > struct dmatest_info *info;
> > struct task_struct *task;
> > struct dma_chan *chan;
> > - u8 **srcs;
> > - u8 **usrcs;
> > - u8 **dsts;
> > - u8 **udsts;
> > + struct dmatest_data src;
> > + struct dmatest_data dst;
> > enum dma_transaction_type type;
> > wait_queue_head_t done_wait;
> > struct dmatest_done test_done;
> > @@ -428,6 +433,53 @@ static unsigned long long dmatest_KBs(s64 runtime,
> > unsigned long long len)
> > return dmatest_persec(runtime, len >> 10);
> > }
> >
> > +static void __dmatest_free_test_data(struct dmatest_data *d, unsigned
> > int cnt)
> > +{
> > + unsigned int i;
> > +
> > + for (i = 0; i < cnt; i++)
> > + kfree(d->raw[i]);
> > +
> > + kfree(d->aligned);
> > + kfree(d->raw);
> > +}
> > +
> > +static void dmatest_free_test_data(struct dmatest_data *d)
> > +{
> > + __dmatest_free_test_data(d, d->cnt);
> > +}
>
> why do we need a wrapper here?
see <comment1> below
>
> > +
> > +static int dmatest_alloc_test_data(struct dmatest_data *d,
> > + unsigned int buf_size, u8 align)
> > +{
> > + unsigned int i = 0;
>
> superfluous initialization
ack, will remove
>
> > + d->raw = kcalloc(d->cnt + 1, sizeof(u8 *), GFP_KERNEL);
> > + if (!d->raw)
> > + return -ENOMEM;
> > +
> > + d->aligned = kcalloc(d->cnt + 1, sizeof(u8 *), GFP_KERNEL);
> > + if (!d->aligned)
> > + goto err;
> > +
> > + for (i = 0; i < d->cnt; i++) {
> > + d->raw[i] = kmalloc(buf_size + align, GFP_KERNEL);
> > + if (!d->raw[i])
> > + goto err;
> > +
> > + /* align to alignment restriction */
> > + if (align)
> > + d->aligned[i] = PTR_ALIGN(d->raw[i], align);
> > + else
> > + d->aligned[i] = d->raw[i];
> > + }
> > +
> > + return 0;
> > +err:
> > + __dmatest_free_test_data(d, i);
<comment1>: the __dmatest_free_test_data() will be used with the `i` index
here in case a kmalloc() has failed while allocating `d->cnt` memory
locations
> > + return -ENOMEM;
> > +}
> > +
> > /*
> > * This function repeatedly tests DMA transfers of various lengths and
> > * offsets for a given operation type until it is told to exit by
> > @@ -458,8 +510,9 @@ static int dmatest_func(void *data)
> > enum dma_ctrl_flags flags;
> > u8 *pq_coefs = NULL;
> > int ret;
> > - int src_cnt;
> > - int dst_cnt;
> > + unsigned int buf_size;
> > + struct dmatest_data *src;
> > + struct dmatest_data *dst;
> > int i;
> > ktime_t ktime, start, diff;
> > ktime_t filltime = 0;
> > @@ -480,99 +533,64 @@ static int dmatest_func(void *data)
> > params = &info->params;
> > chan = thread->chan;
> > dev = chan->device;
> > + src = &thread->src;
> > + dst = &thread->dst;
> > if (thread->type == DMA_MEMCPY) {
> > align = dev->copy_align;
> > - src_cnt = dst_cnt = 1;
> > + src->cnt = dst->cnt = 1;
> > } else if (thread->type == DMA_MEMSET) {
> > align = dev->fill_align;
> > - src_cnt = dst_cnt = 1;
> > + src->cnt = dst->cnt = 1;
> > is_memset = true;
> > } else if (thread->type == DMA_XOR) {
> > /* force odd to ensure dst = src */
> > - src_cnt = min_odd(params->xor_sources | 1, dev->max_xor);
> > - dst_cnt = 1;
> > + src->cnt = min_odd(params->xor_sources | 1, dev-
> > >max_xor);
> > + dst->cnt = 1;
> > align = dev->xor_align;
> > } else if (thread->type == DMA_PQ) {
> > /* force odd to ensure dst = src */
> > - src_cnt = min_odd(params->pq_sources | 1, dma_maxpq(dev,
> > 0));
> > - dst_cnt = 2;
> > + src->cnt = min_odd(params->pq_sources | 1, dma_maxpq(dev,
> > 0));
> > + dst->cnt = 2;
> > align = dev->pq_align;
> >
> > pq_coefs = kmalloc(params->pq_sources + 1, GFP_KERNEL);
> > if (!pq_coefs)
> > goto err_thread_type;
> >
> > - for (i = 0; i < src_cnt; i++)
> > + for (i = 0; i < src->cnt; i++)
> > pq_coefs[i] = 1;
> > } else
> > goto err_thread_type;
> >
> > /* Check if buffer count fits into map count variable (u8) */
> > - if ((src_cnt + dst_cnt) >= 255) {
> > + if ((src->cnt + dst->cnt) >= 255) {
> > pr_err("too many buffers (%d of 255 supported)\n",
> > - src_cnt + dst_cnt);
> > + src->cnt + dst->cnt);
> > goto err_thread_type;
> > }
> >
> > + buf_size = params->buf_size;
> > if (1 << align > params->buf_size) {
> > pr_err("%u-byte buffer too small for %d-byte
> > alignment\n",
> > params->buf_size, 1 << align);
> > goto err_thread_type;
> > }
> >
> > - thread->srcs = kcalloc(src_cnt + 1, sizeof(u8 *), GFP_KERNEL);
> > - if (!thread->srcs)
> > - goto err_srcs;
> > -
> > - thread->usrcs = kcalloc(src_cnt + 1, sizeof(u8 *), GFP_KERNEL);
> > - if (!thread->usrcs)
> > - goto err_usrcs;
> > + if (dmatest_alloc_test_data(src, buf_size, align) < 0)
> > + goto err_src;
> >
> > - for (i = 0; i < src_cnt; i++) {
> > - thread->usrcs[i] = kmalloc(params->buf_size + align,
> > - GFP_KERNEL);
> > - if (!thread->usrcs[i])
> > - goto err_srcbuf;
> > -
> > - /* align srcs to alignment restriction */
> > - if (align)
> > - thread->srcs[i] = PTR_ALIGN(thread->usrcs[i],
> > align);
> > - else
> > - thread->srcs[i] = thread->usrcs[i];
> > - }
> > - thread->srcs[i] = NULL;
> > -
> > - thread->dsts = kcalloc(dst_cnt + 1, sizeof(u8 *), GFP_KERNEL);
> > - if (!thread->dsts)
> > - goto err_dsts;
> > -
> > - thread->udsts = kcalloc(dst_cnt + 1, sizeof(u8 *), GFP_KERNEL);
> > - if (!thread->udsts)
> > - goto err_udsts;
> > -
> > - for (i = 0; i < dst_cnt; i++) {
> > - thread->udsts[i] = kmalloc(params->buf_size + align,
> > - GFP_KERNEL);
> > - if (!thread->udsts[i])
> > - goto err_dstbuf;
> > -
> > - /* align dsts to alignment restriction */
> > - if (align)
> > - thread->dsts[i] = PTR_ALIGN(thread->udsts[i],
> > align);
> > - else
> > - thread->dsts[i] = thread->udsts[i];
> > - }
> > - thread->dsts[i] = NULL;
> > + if (dmatest_alloc_test_data(dst, buf_size, align) < 0)
> > + goto err_dst;
> >
> > set_user_nice(current, 10);
> >
> > - srcs = kcalloc(src_cnt, sizeof(dma_addr_t), GFP_KERNEL);
> > + srcs = kcalloc(src->cnt, sizeof(dma_addr_t), GFP_KERNEL);
> > if (!srcs)
> > - goto err_dstbuf;
> > + goto err_srcs;
> >
> > - dma_pq = kcalloc(dst_cnt, sizeof(dma_addr_t), GFP_KERNEL);
> > + dma_pq = kcalloc(dst->cnt, sizeof(dma_addr_t), GFP_KERNEL);
> > if (!dma_pq)
> > - goto err_srcs_array;
> > + goto err_dma_pq;
> >
> > /*
> > * src and dst buffers are freed by ourselves below
> > @@ -585,7 +603,7 @@ static int dmatest_func(void *data)
> > struct dma_async_tx_descriptor *tx = NULL;
> > struct dmaengine_unmap_data *um;
> > dma_addr_t *dsts;
> > - unsigned int src_off, dst_off, len;
> > + unsigned int len;
> >
> > total_tests++;
> >
> > @@ -601,59 +619,59 @@ static int dmatest_func(void *data)
> > total_len += len;
> >
> > if (params->norandom) {
> > - src_off = 0;
> > - dst_off = 0;
> > + src->off = 0;
> > + dst->off = 0;
> > } else {
> > - src_off = dmatest_random() % (params->buf_size -
> > len + 1);
> > - dst_off = dmatest_random() % (params->buf_size -
> > len + 1);
> > + src->off = dmatest_random() % (params->buf_size -
> > len + 1);
> > + dst->off = dmatest_random() % (params->buf_size -
> > len + 1);
> >
> > - src_off = (src_off >> align) << align;
> > - dst_off = (dst_off >> align) << align;
> > + src->off = (src->off >> align) << align;
> > + dst->off = (dst->off >> align) << align;
> > }
> >
> > if (!params->noverify) {
> > start = ktime_get();
> > - dmatest_init_srcs(thread->srcs, src_off, len,
> > + dmatest_init_srcs(src->aligned, src->off, len,
> > params->buf_size, is_memset);
> > - dmatest_init_dsts(thread->dsts, dst_off, len,
> > + dmatest_init_dsts(dst->aligned, dst->off, len,
> > params->buf_size, is_memset);
> >
> > diff = ktime_sub(ktime_get(), start);
> > filltime = ktime_add(filltime, diff);
> > }
> >
> > - um = dmaengine_get_unmap_data(dev->dev, src_cnt +
> > dst_cnt,
> > + um = dmaengine_get_unmap_data(dev->dev, src->cnt + dst-
> > >cnt,
> > GFP_KERNEL);
> > if (!um) {
> > failed_tests++;
> > result("unmap data NULL", total_tests,
> > - src_off, dst_off, len, ret);
> > + src->off, dst->off, len, ret);
> > continue;
> > }
> >
> > um->len = params->buf_size;
> > - for (i = 0; i < src_cnt; i++) {
> > - void *buf = thread->srcs[i];
> > + for (i = 0; i < src->cnt; i++) {
> > + void *buf = src->aligned[i];
> > struct page *pg = virt_to_page(buf);
> > unsigned long pg_off = offset_in_page(buf);
> >
> > um->addr[i] = dma_map_page(dev->dev, pg, pg_off,
> > um->len,
> > DMA_TO_DEVICE);
> > - srcs[i] = um->addr[i] + src_off;
> > + srcs[i] = um->addr[i] + src->off;
> > ret = dma_mapping_error(dev->dev, um->addr[i]);
> > if (ret) {
> > dmaengine_unmap_put(um);
> > result("src mapping error", total_tests,
> > - src_off, dst_off, len, ret);
> > + src->off, dst->off, len, ret);
> > failed_tests++;
> > continue;
> > }
> > um->to_cnt++;
> > }
> > /* map with DMA_BIDIRECTIONAL to force
> > writeback/invalidate */
> > - dsts = &um->addr[src_cnt];
> > - for (i = 0; i < dst_cnt; i++) {
> > - void *buf = thread->dsts[i];
> > + dsts = &um->addr[src->cnt];
> > + for (i = 0; i < dst->cnt; i++) {
> > + void *buf = dst->aligned[i];
> > struct page *pg = virt_to_page(buf);
> > unsigned long pg_off = offset_in_page(buf);
> >
> > @@ -663,7 +681,7 @@ static int dmatest_func(void *data)
> > if (ret) {
> > dmaengine_unmap_put(um);
> > result("dst mapping error", total_tests,
> > - src_off, dst_off, len, ret);
> > + src->off, dst->off, len, ret);
> > failed_tests++;
> > continue;
> > }
> > @@ -672,30 +690,30 @@ static int dmatest_func(void *data)
> >
> > if (thread->type == DMA_MEMCPY)
> > tx = dev->device_prep_dma_memcpy(chan,
> > - dsts[0] +
> > dst_off,
> > + dsts[0] + dst-
> > >off,
> > srcs[0], len,
> > flags);
> > else if (thread->type == DMA_MEMSET)
> > tx = dev->device_prep_dma_memset(chan,
> > - dsts[0] + dst_off,
> > - *(thread->srcs[0] +
> > src_off),
> > + dsts[0] + dst->off,
> > + *(src->aligned[0] + src-
> > >off),
> > len, flags);
> > else if (thread->type == DMA_XOR)
> > tx = dev->device_prep_dma_xor(chan,
> > - dsts[0] + dst_off,
> > - srcs, src_cnt,
> > + dsts[0] + dst->off,
> > + srcs, src->cnt,
> > len, flags);
> > else if (thread->type == DMA_PQ) {
> > - for (i = 0; i < dst_cnt; i++)
> > - dma_pq[i] = dsts[i] + dst_off;
> > + for (i = 0; i < dst->cnt; i++)
> > + dma_pq[i] = dsts[i] + dst->off;
> > tx = dev->device_prep_dma_pq(chan, dma_pq, srcs,
> > - src_cnt, pq_coefs,
> > + src->cnt, pq_coefs,
> > len, flags);
> > }
> >
> > if (!tx) {
> > dmaengine_unmap_put(um);
> > - result("prep error", total_tests, src_off,
> > - dst_off, len, ret);
> > + result("prep error", total_tests, src->off,
> > + dst->off, len, ret);
> > msleep(100);
> > failed_tests++;
> > continue;
> > @@ -708,8 +726,8 @@ static int dmatest_func(void *data)
> >
> > if (dma_submit_error(cookie)) {
> > dmaengine_unmap_put(um);
> > - result("submit error", total_tests, src_off,
> > - dst_off, len, ret);
> > + result("submit error", total_tests, src->off,
> > + dst->off, len, ret);
> > msleep(100);
> > failed_tests++;
> > continue;
> > @@ -724,58 +742,58 @@ static int dmatest_func(void *data)
> > dmaengine_unmap_put(um);
> >
> > if (!done->done) {
> > - result("test timed out", total_tests, src_off,
> > dst_off,
> > + result("test timed out", total_tests, src->off,
> > dst->off,
> > len, 0);
> > failed_tests++;
> > continue;
> > } else if (status != DMA_COMPLETE) {
> > result(status == DMA_ERROR ?
> > "completion error status" :
> > - "completion busy status", total_tests,
> > src_off,
> > - dst_off, len, ret);
> > + "completion busy status", total_tests,
> > src->off,
> > + dst->off, len, ret);
> > failed_tests++;
> > continue;
> > }
> >
> > if (params->noverify) {
> > - verbose_result("test passed", total_tests,
> > src_off,
> > - dst_off, len, 0);
> > + verbose_result("test passed", total_tests, src-
> > >off,
> > + dst->off, len, 0);
> > continue;
> > }
> >
> > start = ktime_get();
> > pr_debug("%s: verifying source buffer...\n", current-
> > >comm);
> > - error_count = dmatest_verify(thread->srcs, 0, src_off,
> > + error_count = dmatest_verify(src->aligned, 0, src->off,
> > 0, PATTERN_SRC, true, is_memset);
> > - error_count += dmatest_verify(thread->srcs, src_off,
> > - src_off + len, src_off,
> > + error_count += dmatest_verify(src->aligned, src->off,
> > + src->off + len, src->off,
> > PATTERN_SRC | PATTERN_COPY, true,
> > is_memset);
> > - error_count += dmatest_verify(thread->srcs, src_off +
> > len,
> > - params->buf_size, src_off + len,
> > + error_count += dmatest_verify(src->aligned, src->off +
> > len,
> > + params->buf_size, src->off + len,
> > PATTERN_SRC, true, is_memset);
> >
> > pr_debug("%s: verifying dest buffer...\n", current-
> > >comm);
> > - error_count += dmatest_verify(thread->dsts, 0, dst_off,
> > + error_count += dmatest_verify(dst->aligned, 0, dst->off,
> > 0, PATTERN_DST, false, is_memset);
> >
> > - error_count += dmatest_verify(thread->dsts, dst_off,
> > - dst_off + len, src_off,
> > + error_count += dmatest_verify(dst->aligned, dst->off,
> > + dst->off + len, src->off,
> > PATTERN_SRC | PATTERN_COPY, false,
> > is_memset);
> >
> > - error_count += dmatest_verify(thread->dsts, dst_off +
> > len,
> > - params->buf_size, dst_off + len,
> > + error_count += dmatest_verify(dst->aligned, dst->off +
> > len,
> > + params->buf_size, dst->off + len,
> > PATTERN_DST, false, is_memset);
> >
> > diff = ktime_sub(ktime_get(), start);
> > comparetime = ktime_add(comparetime, diff);
> >
> > if (error_count) {
> > - result("data error", total_tests, src_off,
> > dst_off,
> > + result("data error", total_tests, src->off, dst-
> > >off,
> > len, error_count);
> > failed_tests++;
> > } else {
> > - verbose_result("test passed", total_tests,
> > src_off,
> > - dst_off, len, 0);
> > + verbose_result("test passed", total_tests, src-
> > >off,
> > + dst->off, len, 0);
> > }
> > }
> > ktime = ktime_sub(ktime_get(), ktime);
> > @@ -785,22 +803,13 @@ static int dmatest_func(void *data)
> >
> > ret = 0;
> > kfree(dma_pq);
> > -err_srcs_array:
> > +err_dma_pq:
> > kfree(srcs);
> > -err_dstbuf:
> > - for (i = 0; thread->udsts[i]; i++)
> > - kfree(thread->udsts[i]);
> > - kfree(thread->udsts);
> > -err_udsts:
> > - kfree(thread->dsts);
> > -err_dsts:
> > -err_srcbuf:
> > - for (i = 0; thread->usrcs[i]; i++)
> > - kfree(thread->usrcs[i]);
> > - kfree(thread->usrcs);
> > -err_usrcs:
> > - kfree(thread->srcs);
> > err_srcs:
> > + dmatest_free_test_data(dst);
> > +err_dst:
> > + dmatest_free_test_data(src);
> > +err_src:
> > kfree(pq_coefs);
> > err_thread_type:
> > pr_info("%s: summary %u tests, %u failures %llu iops %llu KB/s
> > (%d)\n",
> > --
> > 2.17.1
>
> It would also help review if things are moved in smaller chunks. I can
> think of creating the common mem alloc and free routine by moving
> existing code and then adding new structs..
I'll take a look about splitting this into smaller chunks.
I think this patch may need to be re-applied ; not sure if it applies now.
Thanks
Alex
>
> --
> ~Vinod
next reply other threads:[~2019-01-21 7:54 UTC|newest]
Thread overview: 3+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-01-21 7:54 Alexandru Ardelean [this message]
-- strict thread matches above, loose matches on Subject: below --
2019-01-19 13:07 dmaengine: dmatest: wrap src & dst data into a struct Vinod Koul
2018-11-26 8:43 Alexandru Ardelean
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=de4d0ed1f5c9923afe19f31f87b2c7ed3192d005.camel@analog.com \
--to=alexandru.ardelean@analog.com \
--cc=dmaengine@vger.kernel.org \
--cc=vkoul@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox