public inbox for linux-s390@vger.kernel.org
 help / color / mirror / Atom feed
From: Guangguan Wang <guangguan.wang@linux.alibaba.com>
To: Wenjia Zhang <wenjia@linux.ibm.com>,
	jaka@linux.ibm.com, davem@davemloft.net, edumazet@google.com,
	kuba@kernel.org, pabeni@redhat.com
Cc: alibuda@linux.alibaba.com, tonylu@linux.alibaba.com,
	guwen@linux.alibaba.com, linux-s390@vger.kernel.org,
	netdev@vger.kernel.org, linux-kernel@vger.kernel.org
Subject: Re: [PATCH net-next] net/smc: introduce autosplit for smc
Date: Mon, 15 Jul 2024 10:53:44 +0800	[thread overview]
Message-ID: <63862dcc-33fd-4757-8daf-e0a018a1c7a3@linux.alibaba.com> (raw)
In-Reply-To: <cf07ec76-9d48-4bff-99f6-0842b5127c81@linux.ibm.com>



On 2024/7/11 23:57, Wenjia Zhang wrote:
> 
> 
> On 09.07.24 18:05, Guangguan Wang wrote:
>> When sending large size data in TCP, the data will be split into
>> several segments(packets) to transfer due to MTU config. And in
>> the receive side, application can be woken up to recv data every
>> packet arrived, the data transmission and data recv copy are
>> pipelined.
>>
>> But for SMC-R, it will transmit as many data as possible in one
>> RDMA WRITE and a CDC msg follows the RDMA WRITE, in the receive
>> size, the application only be woken up to recv data when all RDMA
>> WRITE data and the followed CDC msg arrived. The data transmission
>> and data recv copy are sequential.
>>
>> This patch introduce autosplit for SMC, which can automatic split
>> data into several segments and every segment transmitted by one RDMA
>> WRITE when sending large size data in SMC. Because of the split, the
>> data transmission and data send copy can be pipelined in the send side,
>> and the data transmission and data recv copy can be pipelined in the
>> receive side. Thus autosplit helps improving latency performance when
>> sending large size data. The autosplit also works for SMC-D.
>>
>> This patch also introduce a sysctl names autosplit_size for configure
>> the max size of the split segment, whose default value is 128KiB
>> (128KiB perform best in my environment).
>>
>> The sockperf benchmark shows 17%-28% latency improvement when msgsize
>>> = 256KB for SMC-R, 15%-32% latency improvement when msgsize >= 256KB
>> for SMC-D with smc-loopback.
>>
>> Test command:
>> sockperf sr --tcp -m 1048575
>> sockperf pp --tcp -i <server ip> -m <msgsize> -t 20
>>
>> Test config:
>> sysctl -w net.smc.wmem=524288
>> sysctl -w net.smc.rmem=524288
>>
>> Test results:
>> SMC-R
>> msgsize   noautosplit    autosplit
>> 128KB       55.546 us     55.763 us
>> 256KB       83.537 us     69.743 us (17% improve)
>> 512KB      138.306 us    100.313 us (28% improve)
>> 1MB        273.702 us    197.222 us (28% improve)
>>
>> SMC-D with smc-loopback
>> msgsize   noautosplit    autosplit
>> 128KB       14.672 us     14.690 us
>> 256KB       28.277 us     23.958 us (15% improve)
>> 512KB       63.047 us     45.339 us (28% improve)
>> 1MB        129.306 us     87.278 us (32% improve)
>>
>> Signed-off-by: Guangguan Wang <guangguan.wang@linux.alibaba.com>
>> ---
>>   Documentation/networking/smc-sysctl.rst | 11 +++++++++++
>>   include/net/netns/smc.h                 |  1 +
>>   net/smc/smc_sysctl.c                    | 12 ++++++++++++
>>   net/smc/smc_tx.c                        | 19 ++++++++++++++++++-
>>   4 files changed, 42 insertions(+), 1 deletion(-)
>>
> 
> Hi Guangguan,
> 
> If I remember correctly, the intention to use one RDMA-write for a possible large data is to reduce possible many partial stores. Since many year has gone, I'm not that sure if it would still be an issue. I need some time to check on it.
> 

Did you mean too many partial stores will result in some issue? What's the issue?


> BTW, I don't really like the idea to use sysctl to set the autosplit_size in any value at will. That makes no sense to improve the performance.

Although 128KB autosplit_size have a good performance in most scenario, I still found some better autosplit_size for some specific network configurations.
For example, 128KB autosplit_size have a good performance whether the MTU is 1500 or 8500, but for 8500 MTU, 64KB autosplit_size performs better.

Maybe the sysctl is not the best way, but I think it should have a way to set the value of autosplit_size for possible performance tuning.

Thanks,
Guangguan Wang

> 
> Thanks,
> Wenjia

  reply	other threads:[~2024-07-15  2:53 UTC|newest]

Thread overview: 7+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2024-07-09 16:05 [PATCH net-next] net/smc: introduce autosplit for smc Guangguan Wang
2024-07-11 15:57 ` Wenjia Zhang
2024-07-15  2:53   ` Guangguan Wang [this message]
2024-08-08  6:26     ` Guangguan Wang
2024-08-09 21:07       ` Wenjia Zhang
2024-08-12  2:06         ` Guangguan Wang
2024-11-27 11:02         ` Guangguan Wang

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=63862dcc-33fd-4757-8daf-e0a018a1c7a3@linux.alibaba.com \
    --to=guangguan.wang@linux.alibaba.com \
    --cc=alibuda@linux.alibaba.com \
    --cc=davem@davemloft.net \
    --cc=edumazet@google.com \
    --cc=guwen@linux.alibaba.com \
    --cc=jaka@linux.ibm.com \
    --cc=kuba@kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-s390@vger.kernel.org \
    --cc=netdev@vger.kernel.org \
    --cc=pabeni@redhat.com \
    --cc=tonylu@linux.alibaba.com \
    --cc=wenjia@linux.ibm.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox