public inbox for linux-rdma@vger.kernel.org
 help / color / mirror / Atom feed
* RDMA reads/writes per second
@ 2013-10-29  0:28 Anuj Kalia
  2013-10-30 16:01 ` Ido Shamai
  0 siblings, 1 reply; 3+ messages in thread
From: Anuj Kalia @ 2013-10-29  0:28 UTC (permalink / raw)
  To: linux-rdma-u79uwXL29TY76Z2rM5mHXA

Hi.

I'm measuring the number of RDMA reads and writes per second. In my
experimental setup I have one server connected to several clients and
I want to extract the maximum IOs from the server. I had two questions
regarding this:

1. What is the expected number of small (16 byte values) RDMA reads
for ConnectX 3 cards? Currently, I've seen a maximum of 9 million
reads per second with my code. However, several websites report much
higher messages per second. For
example,http://www.marketwatch.com/story/mellanox-fdr-56gbs-infiniband-solutions-deliver-leading-application-performance-and-scalability-2013-06-17talks
about 137 messages per second.
http://www.mellanox.com/pdf/products/oem/RG_HP.pdf reports 40 million
MPI messages per second. What sort optimizations could I do to reach
similar numbers?

2. The number of IOPS drops when the size of the registered region
increases. For a 1 KB registered region, the maximum random reads per
second that the server can provide is around 9 million. It drops to 2
million when I increase the registered size to 1 GB.
What is the reason behind this? Does the HCA perform caching for
reads? That could be a possible explanation. Another possible reason
is TLB misses in the HCA.
Further, I'm seeing even greater variation with writes. I can think of
2 possible explanations for that:
a. As my writes are to random locations, there could be more TLB
misses for larger registered regions.
b. The HCA buffers writes locally and does not transfer them into the
CPU memory immediately (this can be done only for small registered
regions).

Thanks for your time!
I'm sorry if the list receives more than one copy of this email. I've
been running into a HTML rejection error.

Anuj Kalia,
Carnegie Mellon University
--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majordomo-u79uwXL29TY76Z2rM5mHXA@public.gmane.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 3+ messages in thread
* RDMA reads/writes per second
@ 2013-10-29  0:05 akalia-OM76b2Iv3yLQjUSlxSEPGw
  0 siblings, 0 replies; 3+ messages in thread
From: akalia-OM76b2Iv3yLQjUSlxSEPGw @ 2013-10-29  0:05 UTC (permalink / raw)
  To: linux-rdma-u79uwXL29TY76Z2rM5mHXA

Hi.

I'm measuring the number of RDMA reads and writes per second. In my
experimental setup I have one server connected to several clients and I
want to extract the maximum IOs from the server. I had two questions
regarding this:

1. What is the expected number of small (16 byte values) RDMA reads for
ConnectX 3 cards? Currently, I've seen a maximum of 9 million reads per
second with my code. However, several websites report much higher messages
per second. For example,
http://www.marketwatch.com/story/mellanox-fdr-56gbs-infiniband-solutions-deliver-leading-application-performance-and-scalability-2013-06-17
talks about 137 messages per second.
http://www.mellanox.com/pdf/products/oem/RG_HP.pdf reports 40 million MPI
messages per second. What sort optimizations could I do to reach similar
numbers?

2. The number of IOPS drops when the size of the registered region
increases. For a 1 KB registered region, the maximum random reads per
second that the server can provide is around 9 million. It drops to 2
million when I increase the registered size to 1 GB.
What is the reason behind this? Does the HCA perform caching for reads?
That could be a possible explanation. Another possible reason is TLB
misses in the HCA.
Further, I'm seeing even greater variation with writes. I can think of 2
possible explanations for that:
a. As my writes are to random locations, there could be more TLB misses
for larger registered regions.
b. The HCA buffers writes locally and does not transfer them into the CPU
memory immediately (this can be done only for small registered regions).

Thanks for your time!

Anuj Kalia,
Carnegie Mellon University

--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majordomo-u79uwXL29TY76Z2rM5mHXA@public.gmane.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 3+ messages in thread

end of thread, other threads:[~2013-10-30 16:01 UTC | newest]

Thread overview: 3+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2013-10-29  0:28 RDMA reads/writes per second Anuj Kalia
2013-10-30 16:01 ` Ido Shamai
  -- strict thread matches above, loose matches on Subject: below --
2013-10-29  0:05 akalia-OM76b2Iv3yLQjUSlxSEPGw

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox