Flexible I/O Tester development
 help / color / mirror / Atom feed
* Re: [ceph-users] Poor libRBD write performance
       [not found] <034AAD465C6CBE4F96D9FB98573A79A62C2C92E0@FMSMSX108.amr.corp.intel.com>
@ 2017-11-20 16:10 ` Jason Dillaman
  2017-11-20 17:00   ` Moreno, Orlando
  2017-11-20 16:16 ` [Cbt] " Mark Nelson
  1 sibling, 1 reply; 4+ messages in thread
From: Jason Dillaman @ 2017-11-20 16:10 UTC (permalink / raw)
  To: Moreno, Orlando
  Cc: fio@vger.kernel.org, ceph-users@lists.ceph.com,
	cbt@lists.ceph.com

I suspect you are seeing this issue [1]. TL;DR: never use "numjobs" >
1 against an RBD image that has the exclusive-lock feature enabled.

[1] http://lists.ceph.com/pipermail/ceph-users-ceph.com/2016-August/012123.html

On Mon, Nov 20, 2017 at 11:06 AM, Moreno, Orlando
<orlando.moreno@intel.com> wrote:
> Hi all,
>
>
>
> I’ve been experiencing weird performance behavior when using FIO RBD engine
> directly to an RBD volume with numjobs > 1. For a 4KB random write test at
> 32 QD and 1 numjob, I can get about 40K IOPS, but when I increase the
> numjobs to 4, it plummets to 2800 IOPS. I tried running the same exact test
> on a VM using FIO libaio targeting a block device (volume) attached through
> QEMU/RBD and I get ~35K-40K IOPS in both situations. In all cases, CPU was
> not fully utilized and there were no signs of any hardware bottlenecks. I
> did not disable any RBD features and most of the Ceph parameters are default
> (besides auth, debug, pool size, etc).
>
>
>
> My Ceph cluster is running on 6 nodes, all-NVMe, 22-core, 376GB mem,
> Luminous 12.2.1, Ubuntu 16.04, and clients running FIO job/VM on similar
> HW/SW spec. The VM has 16 vCPU, 64GB mem, and the root disk is locally
> stored while the persistent disk comes from an RBD volume serviced by the
> Ceph cluster.
>
>
>
> If anyone has seen this issue or have any suggestions please let me know.
>
>
>
> Thanks,
>
> Orlando
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>



-- 
Jason

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [Cbt] Poor libRBD write performance
       [not found] <034AAD465C6CBE4F96D9FB98573A79A62C2C92E0@FMSMSX108.amr.corp.intel.com>
  2017-11-20 16:10 ` [ceph-users] Poor libRBD write performance Jason Dillaman
@ 2017-11-20 16:16 ` Mark Nelson
  1 sibling, 0 replies; 4+ messages in thread
From: Mark Nelson @ 2017-11-20 16:16 UTC (permalink / raw)
  To: Moreno, Orlando, fio@vger.kernel.org, ceph-users@lists.ceph.com
  Cc: 'cbt@lists.ceph.com'

On 11/20/2017 10:06 AM, Moreno, Orlando wrote:
> Hi all,
>
>
>
> I�ve been experiencing weird performance behavior when using FIO RBD
> engine directly to an RBD volume with numjobs > 1. For a 4KB random
> write test at 32 QD and 1 numjob, I can get about 40K IOPS, but when I
> increase the numjobs to 4, it plummets to 2800 IOPS. I tried running the
> same exact test on a VM using FIO libaio targeting a block device
> (volume) attached through QEMU/RBD and I get ~35K-40K IOPS in both
> situations. In all cases, CPU was not fully utilized and there were no
> signs of any hardware bottlenecks. I did not disable any RBD features
> and most of the Ceph parameters are default (besides auth, debug, pool
> size, etc).
>
>
>
> My Ceph cluster is running on 6 nodes, all-NVMe, 22-core, 376GB mem,
> Luminous 12.2.1, Ubuntu 16.04, and clients running FIO job/VM on similar
> HW/SW spec. The VM has 16 vCPU, 64GB mem, and the root disk is locally
> stored while the persistent disk comes from an RBD volume serviced by
> the Ceph cluster.
>
>
>
> If anyone has seen this issue or have any suggestions please let me know.

Hi Orlando,

Try seeing if disabling the RBD image exclusive lock helps (if only to 
confirm that's what's going on).  I usually test with numjobs=1 and run 
multiple fio instances with higher iodepth values instead to avoid this. 
  See:

https://www.spinics.net/lists/ceph-devel/msg30468.html

and

http://lists.ceph.com/pipermail/ceph-users-ceph.com/2015-September/004872.html

Mark

>
>
>
> Thanks,
>
> Orlando
>
>
>
> _______________________________________________
> Cbt mailing list
> Cbt@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/cbt-ceph.com
>

^ permalink raw reply	[flat|nested] 4+ messages in thread

* RE: [ceph-users] Poor libRBD write performance
  2017-11-20 16:10 ` [ceph-users] Poor libRBD write performance Jason Dillaman
@ 2017-11-20 17:00   ` Moreno, Orlando
  2017-11-20 17:01     ` Jason Dillaman
  0 siblings, 1 reply; 4+ messages in thread
From: Moreno, Orlando @ 2017-11-20 17:00 UTC (permalink / raw)
  To: dillaman@redhat.com
  Cc: fio@vger.kernel.org, ceph-users@lists.ceph.com,
	cbt@lists.ceph.com

Hi Jason,

You're right, thanks for pointing that out. I could've sworn I saw the same problem with exclusive-lock disabled, but after trying it again, disabling the feature does fix the write performance :)

So does this mean that when an RBD is attached to a VM, it is considered a single client connection?

Thanks,
Orlando

-----Original Message-----
From: Jason Dillaman [mailto:jdillama@redhat.com] 
Sent: Monday, November 20, 2017 9:10 AM
To: Moreno, Orlando <orlando.moreno@intel.com>
Cc: fio@vger.kernel.org; ceph-users@lists.ceph.com; cbt@lists.ceph.com
Subject: Re: [ceph-users] Poor libRBD write performance

I suspect you are seeing this issue [1]. TL;DR: never use "numjobs" >
1 against an RBD image that has the exclusive-lock feature enabled.

[1] http://lists.ceph.com/pipermail/ceph-users-ceph.com/2016-August/012123.html

On Mon, Nov 20, 2017 at 11:06 AM, Moreno, Orlando <orlando.moreno@intel.com> wrote:
> Hi all,
>
>
>
> I’ve been experiencing weird performance behavior when using FIO RBD 
> engine directly to an RBD volume with numjobs > 1. For a 4KB random 
> write test at
> 32 QD and 1 numjob, I can get about 40K IOPS, but when I increase the 
> numjobs to 4, it plummets to 2800 IOPS. I tried running the same exact 
> test on a VM using FIO libaio targeting a block device (volume) 
> attached through QEMU/RBD and I get ~35K-40K IOPS in both situations. 
> In all cases, CPU was not fully utilized and there were no signs of 
> any hardware bottlenecks. I did not disable any RBD features and most 
> of the Ceph parameters are default (besides auth, debug, pool size, etc).
>
>
>
> My Ceph cluster is running on 6 nodes, all-NVMe, 22-core, 376GB mem, 
> Luminous 12.2.1, Ubuntu 16.04, and clients running FIO job/VM on 
> similar HW/SW spec. The VM has 16 vCPU, 64GB mem, and the root disk is 
> locally stored while the persistent disk comes from an RBD volume 
> serviced by the Ceph cluster.
>
>
>
> If anyone has seen this issue or have any suggestions please let me know.
>
>
>
> Thanks,
>
> Orlando
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>



--
Jason

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [ceph-users] Poor libRBD write performance
  2017-11-20 17:00   ` Moreno, Orlando
@ 2017-11-20 17:01     ` Jason Dillaman
  0 siblings, 0 replies; 4+ messages in thread
From: Jason Dillaman @ 2017-11-20 17:01 UTC (permalink / raw)
  To: Moreno, Orlando
  Cc: fio@vger.kernel.org, ceph-users@lists.ceph.com,
	cbt@lists.ceph.com

On Mon, Nov 20, 2017 at 12:00 PM, Moreno, Orlando
<orlando.moreno@intel.com> wrote:
> Hi Jason,
>
> You're right, thanks for pointing that out. I could've sworn I saw the same problem with exclusive-lock disabled, but after trying it again, disabling the feature does fix the write performance :)
>
> So does this mean that when an RBD is attached to a VM, it is considered a single client connection?

Yes, the VM (QEMU) is a single client to librbd.

> Thanks,
> Orlando
>
> -----Original Message-----
> From: Jason Dillaman [mailto:jdillama@redhat.com]
> Sent: Monday, November 20, 2017 9:10 AM
> To: Moreno, Orlando <orlando.moreno@intel.com>
> Cc: fio@vger.kernel.org; ceph-users@lists.ceph.com; cbt@lists.ceph.com
> Subject: Re: [ceph-users] Poor libRBD write performance
>
> I suspect you are seeing this issue [1]. TL;DR: never use "numjobs" >
> 1 against an RBD image that has the exclusive-lock feature enabled.
>
> [1] http://lists.ceph.com/pipermail/ceph-users-ceph.com/2016-August/012123.html
>
> On Mon, Nov 20, 2017 at 11:06 AM, Moreno, Orlando <orlando.moreno@intel.com> wrote:
>> Hi all,
>>
>>
>>
>> I’ve been experiencing weird performance behavior when using FIO RBD
>> engine directly to an RBD volume with numjobs > 1. For a 4KB random
>> write test at
>> 32 QD and 1 numjob, I can get about 40K IOPS, but when I increase the
>> numjobs to 4, it plummets to 2800 IOPS. I tried running the same exact
>> test on a VM using FIO libaio targeting a block device (volume)
>> attached through QEMU/RBD and I get ~35K-40K IOPS in both situations.
>> In all cases, CPU was not fully utilized and there were no signs of
>> any hardware bottlenecks. I did not disable any RBD features and most
>> of the Ceph parameters are default (besides auth, debug, pool size, etc).
>>
>>
>>
>> My Ceph cluster is running on 6 nodes, all-NVMe, 22-core, 376GB mem,
>> Luminous 12.2.1, Ubuntu 16.04, and clients running FIO job/VM on
>> similar HW/SW spec. The VM has 16 vCPU, 64GB mem, and the root disk is
>> locally stored while the persistent disk comes from an RBD volume
>> serviced by the Ceph cluster.
>>
>>
>>
>> If anyone has seen this issue or have any suggestions please let me know.
>>
>>
>>
>> Thanks,
>>
>> Orlando
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>
>
> --
> Jason



-- 
Jason

^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2017-11-20 17:01 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
     [not found] <034AAD465C6CBE4F96D9FB98573A79A62C2C92E0@FMSMSX108.amr.corp.intel.com>
2017-11-20 16:10 ` [ceph-users] Poor libRBD write performance Jason Dillaman
2017-11-20 17:00   ` Moreno, Orlando
2017-11-20 17:01     ` Jason Dillaman
2017-11-20 16:16 ` [Cbt] " Mark Nelson

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox