* Reducing the size of the dump file/speeding up collection
@ 2015-09-16 8:30 Nikolay Borisov
2015-09-17 3:27 ` qiaonuohan
0 siblings, 1 reply; 8+ messages in thread
From: Nikolay Borisov @ 2015-09-16 8:30 UTC (permalink / raw)
To: kexec; +Cc: SiteGround Operations
Hello,
I've been using makedumpfile as the crash collector with the -d31
parameter. The machine this is being run on usually have 128-256GB of
ram and the resulting crash dumps are in the range of 14-20gb which is
very bug for the type of analysis I'm usually performing on crashed
machine. I was wondering whether there is a way to further reduce the
size and the time to take the dump (now it takes around 25 minutes to
collect such a dump). I've seen reports where people with TBs of ram
take that long, meaning for a machine with 256gb it should be even
faster. I've been running this configuration on kernels 3.12.28 and 4.1
where mmap for the vmcore file is supported.
Please advise
Regards,
Nikolay
_______________________________________________
kexec mailing list
kexec@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/kexec
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: Reducing the size of the dump file/speeding up collection
2015-09-16 8:30 Reducing the size of the dump file/speeding up collection Nikolay Borisov
@ 2015-09-17 3:27 ` qiaonuohan
2015-09-17 6:32 ` Nikolay Borisov
0 siblings, 1 reply; 8+ messages in thread
From: qiaonuohan @ 2015-09-17 3:27 UTC (permalink / raw)
To: Nikolay Borisov, kexec; +Cc: SiteGround Operations
On 09/16/2015 04:30 PM, Nikolay Borisov wrote:
> Hello,
>
> I've been using makedumpfile as the crash collector with the -d31
> parameter. The machine this is being run on usually have 128-256GB of
> ram and the resulting crash dumps are in the range of 14-20gb which is
> very bug for the type of analysis I'm usually performing on crashed
> machine. I was wondering whether there is a way to further reduce the
> size and the time to take the dump (now it takes around 25 minutes to
> collect such a dump). I've seen reports where people with TBs of ram
> take that long, meaning for a machine with 256gb it should be even
> faster. I've been running this configuration on kernels 3.12.28 and 4.1
> where mmap for the vmcore file is supported.
>
> Please advise
Hi nikolay,
Yes, this issue is what we are concerning a lot.
About the current situation, try --split, it will save time.
And lzo/snappy instead of zlib, these two compression format are faster
but need more space to save. Or if you still want zlib (to save space),
try multiple threads, check the following site, it will help you:
https://lists.fedoraproject.org/pipermail/kexec/2015-September/002322.html
--
Regards
Qiao Nuohan
>
> Regards,
> Nikolay
>
> _______________________________________________
> kexec mailing list
> kexec@lists.infradead.org
> http://lists.infradead.org/mailman/listinfo/kexec
>
_______________________________________________
kexec mailing list
kexec@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/kexec
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: Reducing the size of the dump file/speeding up collection
2015-09-17 3:27 ` qiaonuohan
@ 2015-09-17 6:32 ` Nikolay Borisov
2015-09-17 7:08 ` Nikolay Borisov
2015-09-18 2:38 ` qiaonuohan
0 siblings, 2 replies; 8+ messages in thread
From: Nikolay Borisov @ 2015-09-17 6:32 UTC (permalink / raw)
To: qiaonuohan, kexec; +Cc: SiteGround Operations
Hi Qiao,
Thanks for the reply. So far I haven't been using the the compression
feature of makedumpfile. But I want to ask if anything wouldn't
compression make the dump process slower since in addition to having to
write the dump to disk it also has to compress it which would put more
strain on the cpu. Also, which part of the dump process is the bottleneck:
- Reading from /proc/vmcore - that has mmap support so should be fairly
fast?
- Discarding unnecessary pages as memory is being scanned?
- Writing/compressing content to disk?
Regards,
Nikolay
On 09/17/2015 06:27 AM, qiaonuohan wrote:
> On 09/16/2015 04:30 PM, Nikolay Borisov wrote:
>> Hello,
>>
>> I've been using makedumpfile as the crash collector with the -d31
>> parameter. The machine this is being run on usually have 128-256GB of
>> ram and the resulting crash dumps are in the range of 14-20gb which is
>> very bug for the type of analysis I'm usually performing on crashed
>> machine. I was wondering whether there is a way to further reduce the
>> size and the time to take the dump (now it takes around 25 minutes to
>> collect such a dump). I've seen reports where people with TBs of ram
>> take that long, meaning for a machine with 256gb it should be even
>> faster. I've been running this configuration on kernels 3.12.28 and 4.1
>> where mmap for the vmcore file is supported.
>>
>> Please advise
>
> Hi nikolay,
>
> Yes, this issue is what we are concerning a lot.
> About the current situation, try --split, it will save time.
>
>
> And lzo/snappy instead of zlib, these two compression format are faster
> but need more space to save. Or if you still want zlib (to save space),
> try multiple threads, check the following site, it will help you:
>
> https://lists.fedoraproject.org/pipermail/kexec/2015-September/002322.html
>
>
_______________________________________________
kexec mailing list
kexec@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/kexec
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: Reducing the size of the dump file/speeding up collection
2015-09-17 6:32 ` Nikolay Borisov
@ 2015-09-17 7:08 ` Nikolay Borisov
2015-09-18 2:38 ` qiaonuohan
1 sibling, 0 replies; 8+ messages in thread
From: Nikolay Borisov @ 2015-09-17 7:08 UTC (permalink / raw)
To: qiaonuohan, kexec; +Cc: SiteGround Operations
Just to follow up: which version of kexec/makedumpfile are required to
have the parallel dump feature? The thread you referenced adds
documentation how to use that and I grepped the source of makedumpfile
1.5.8 and couldn't find any references to the num-threads option? So
what version of the code do I need to test that ?
On 09/17/2015 09:32 AM, Nikolay Borisov wrote:
> Hi Qiao,
>
> Thanks for the reply. So far I haven't been using the the compression
> feature of makedumpfile. But I want to ask if anything wouldn't
> compression make the dump process slower since in addition to having to
> write the dump to disk it also has to compress it which would put more
> strain on the cpu. Also, which part of the dump process is the bottleneck:
>
> - Reading from /proc/vmcore - that has mmap support so should be fairly
> fast?
> - Discarding unnecessary pages as memory is being scanned?
> - Writing/compressing content to disk?
>
> Regards,
> Nikolay
>
> On 09/17/2015 06:27 AM, qiaonuohan wrote:
>> On 09/16/2015 04:30 PM, Nikolay Borisov wrote:
>>> Hello,
>>>
>>> I've been using makedumpfile as the crash collector with the -d31
>>> parameter. The machine this is being run on usually have 128-256GB of
>>> ram and the resulting crash dumps are in the range of 14-20gb which is
>>> very bug for the type of analysis I'm usually performing on crashed
>>> machine. I was wondering whether there is a way to further reduce the
>>> size and the time to take the dump (now it takes around 25 minutes to
>>> collect such a dump). I've seen reports where people with TBs of ram
>>> take that long, meaning for a machine with 256gb it should be even
>>> faster. I've been running this configuration on kernels 3.12.28 and 4.1
>>> where mmap for the vmcore file is supported.
>>>
>>> Please advise
>>
>> Hi nikolay,
>>
>> Yes, this issue is what we are concerning a lot.
>> About the current situation, try --split, it will save time.
>>
>>
>> And lzo/snappy instead of zlib, these two compression format are faster
>> but need more space to save. Or if you still want zlib (to save space),
>> try multiple threads, check the following site, it will help you:
>>
>> https://lists.fedoraproject.org/pipermail/kexec/2015-September/002322.html
>>
>>
_______________________________________________
kexec mailing list
kexec@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/kexec
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: Reducing the size of the dump file/speeding up collection
2015-09-17 6:32 ` Nikolay Borisov
2015-09-17 7:08 ` Nikolay Borisov
@ 2015-09-18 2:38 ` qiaonuohan
2015-09-18 12:45 ` Nikolay Borisov
1 sibling, 1 reply; 8+ messages in thread
From: qiaonuohan @ 2015-09-18 2:38 UTC (permalink / raw)
To: Nikolay Borisov, kexec; +Cc: SiteGround Operations
On 09/17/2015 02:32 PM, Nikolay Borisov wrote:
> Hi Qiao,
>
> Thanks for the reply. So far I haven't been using the the compression
> feature of makedumpfile. But I want to ask if anything wouldn't
> compression make the dump process slower since in addition to having to
> write the dump to disk it also has to compress it which would put more
> strain on the cpu. Also, which part of the dump process is the bottleneck:
>
> - Reading from /proc/vmcore - that has mmap support so should be fairly
> fast?
> - Discarding unnecessary pages as memory is being scanned?
> - Writing/compressing content to disk?
I cannot recall percentage of each part. But writing/compression takes most
of the time
1. mmap is used for reading faster
2. --split is used to split the dump task into several processes, so compressing
and writing will be speeded up.
3. multiple-thread is another option for speeding up compressing, it is a recently
committed patch, so you cannot find it in the master branch, checkout devel branch
or find it here:
http://sourceforge.net/p/makedumpfile/code/commit_browser
Make makedumpfile available to read and compress pages parallelly.
>
> Regards,
> Nikolay
>
> On 09/17/2015 06:27 AM, qiaonuohan wrote:
>> On 09/16/2015 04:30 PM, Nikolay Borisov wrote:
>>> Hello,
>>>
>>> I've been using makedumpfile as the crash collector with the -d31
>>> parameter. The machine this is being run on usually have 128-256GB of
>>> ram and the resulting crash dumps are in the range of 14-20gb which is
>>> very bug for the type of analysis I'm usually performing on crashed
>>> machine. I was wondering whether there is a way to further reduce the
>>> size and the time to take the dump (now it takes around 25 minutes to
>>> collect such a dump). I've seen reports where people with TBs of ram
>>> take that long, meaning for a machine with 256gb it should be even
>>> faster. I've been running this configuration on kernels 3.12.28 and 4.1
>>> where mmap for the vmcore file is supported.
>>>
>>> Please advise
>>
>> Hi nikolay,
>>
>> Yes, this issue is what we are concerning a lot.
>> About the current situation, try --split, it will save time.
>>
>>
>> And lzo/snappy instead of zlib, these two compression format are faster
>> but need more space to save. Or if you still want zlib (to save space),
>> try multiple threads, check the following site, it will help you:
>>
>> https://lists.fedoraproject.org/pipermail/kexec/2015-September/002322.html
>>
>>
> .
>
--
Regards
Qiao Nuohan
_______________________________________________
kexec mailing list
kexec@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/kexec
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: Reducing the size of the dump file/speeding up collection
2015-09-18 2:38 ` qiaonuohan
@ 2015-09-18 12:45 ` Nikolay Borisov
2015-09-21 6:27 ` qiaonuohan
2015-09-23 1:44 ` Baoquan He
0 siblings, 2 replies; 8+ messages in thread
From: Nikolay Borisov @ 2015-09-18 12:45 UTC (permalink / raw)
To: qiaonuohan, kexec; +Cc: SiteGround Operations
Yeah, I did see the commit browser. But in my case I haven't even tested
the split option so I guess there are things to try. Am I correct in my
understanding as to how --split is supposed to work ( i tried that to no
avail though):
My core_collector line is this:
core_collector makedumpfile --message-level 1 -d 3 --split dump1 dump2
dump3 dump4 dump5 dump6
And then in /etc/sysconfig/kdump I have:
KDUMP_COMMANDLINE_APPEND="irqpoll nr_cpus=6 reset_devices
cgroup_disable=memory mce=off"
(the machine I'm testing on has 4 cores x2 hyperthreads so 8 logical
cores in total). Do I need to do something else to utilize the --split
option?
On 09/18/2015 05:38 AM, qiaonuohan wrote:
> On 09/17/2015 02:32 PM, Nikolay Borisov wrote:
>> Hi Qiao,
>>
>> Thanks for the reply. So far I haven't been using the the compression
>> feature of makedumpfile. But I want to ask if anything wouldn't
>> compression make the dump process slower since in addition to having to
>> write the dump to disk it also has to compress it which would put more
>> strain on the cpu. Also, which part of the dump process is the
>> bottleneck:
>>
>> - Reading from /proc/vmcore - that has mmap support so should be fairly
>> fast?
>> - Discarding unnecessary pages as memory is being scanned?
>> - Writing/compressing content to disk?
>
> I cannot recall percentage of each part. But writing/compression takes most
> of the time
>
> 1. mmap is used for reading faster
> 2. --split is used to split the dump task into several processes, so
> compressing
> and writing will be speeded up.
> 3. multiple-thread is another option for speeding up compressing, it is
> a recently
> committed patch, so you cannot find it in the master branch, checkout
> devel branch
> or find it here:
>
> http://sourceforge.net/p/makedumpfile/code/commit_browser
>
> Make makedumpfile available to read and compress pages parallelly.
>
>>
>> Regards,
>> Nikolay
>>
>> On 09/17/2015 06:27 AM, qiaonuohan wrote:
>>> On 09/16/2015 04:30 PM, Nikolay Borisov wrote:
>>>> Hello,
>>>>
>>>> I've been using makedumpfile as the crash collector with the -d31
>>>> parameter. The machine this is being run on usually have 128-256GB of
>>>> ram and the resulting crash dumps are in the range of 14-20gb which is
>>>> very bug for the type of analysis I'm usually performing on crashed
>>>> machine. I was wondering whether there is a way to further reduce the
>>>> size and the time to take the dump (now it takes around 25 minutes to
>>>> collect such a dump). I've seen reports where people with TBs of ram
>>>> take that long, meaning for a machine with 256gb it should be even
>>>> faster. I've been running this configuration on kernels 3.12.28 and 4.1
>>>> where mmap for the vmcore file is supported.
>>>>
>>>> Please advise
>>>
>>> Hi nikolay,
>>>
>>> Yes, this issue is what we are concerning a lot.
>>> About the current situation, try --split, it will save time.
>>>
>>>
>>> And lzo/snappy instead of zlib, these two compression format are faster
>>> but need more space to save. Or if you still want zlib (to save space),
>>> try multiple threads, check the following site, it will help you:
>>>
>>> https://lists.fedoraproject.org/pipermail/kexec/2015-September/002322.html
>>>
>>>
>>>
>> .
>>
>
>
_______________________________________________
kexec mailing list
kexec@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/kexec
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: Reducing the size of the dump file/speeding up collection
2015-09-18 12:45 ` Nikolay Borisov
@ 2015-09-21 6:27 ` qiaonuohan
2015-09-23 1:44 ` Baoquan He
1 sibling, 0 replies; 8+ messages in thread
From: qiaonuohan @ 2015-09-21 6:27 UTC (permalink / raw)
To: Nikolay Borisov, kexec; +Cc: SiteGround Operations
On 09/18/2015 08:45 PM, Nikolay Borisov wrote:
> Yeah, I did see the commit browser. But in my case I haven't even tested
> the split option so I guess there are things to try. Am I correct in my
> understanding as to how --split is supposed to work ( i tried that to no
> avail though):
>
> My core_collector line is this:
>
> core_collector makedumpfile --message-level 1 -d 3 --split dump1 dump2
> dump3 dump4 dump5 dump6
>
> And then in /etc/sysconfig/kdump I have:
>
> KDUMP_COMMANDLINE_APPEND="irqpoll nr_cpus=6 reset_devices
> cgroup_disable=memory mce=off"
>
> (the machine I'm testing on has 4 cores x2 hyperthreads so 8 logical
> cores in total). Do I need to do something else to utilize the --split
> option?
As I recall, it is OK to go on the test.
>
> On 09/18/2015 05:38 AM, qiaonuohan wrote:
>> On 09/17/2015 02:32 PM, Nikolay Borisov wrote:
>>> Hi Qiao,
>>>
>>> Thanks for the reply. So far I haven't been using the the compression
>>> feature of makedumpfile. But I want to ask if anything wouldn't
>>> compression make the dump process slower since in addition to having to
>>> write the dump to disk it also has to compress it which would put more
>>> strain on the cpu. Also, which part of the dump process is the
>>> bottleneck:
>>>
>>> - Reading from /proc/vmcore - that has mmap support so should be fairly
>>> fast?
>>> - Discarding unnecessary pages as memory is being scanned?
>>> - Writing/compressing content to disk?
>>
>> I cannot recall percentage of each part. But writing/compression takes most
>> of the time
>>
>> 1. mmap is used for reading faster
>> 2. --split is used to split the dump task into several processes, so
>> compressing
>> and writing will be speeded up.
>> 3. multiple-thread is another option for speeding up compressing, it is
>> a recently
>> committed patch, so you cannot find it in the master branch, checkout
>> devel branch
>> or find it here:
>>
>> http://sourceforge.net/p/makedumpfile/code/commit_browser
>>
>> Make makedumpfile available to read and compress pages parallelly.
>>
>>>
>>> Regards,
>>> Nikolay
>>>
>>> On 09/17/2015 06:27 AM, qiaonuohan wrote:
>>>> On 09/16/2015 04:30 PM, Nikolay Borisov wrote:
>>>>> Hello,
>>>>>
>>>>> I've been using makedumpfile as the crash collector with the -d31
>>>>> parameter. The machine this is being run on usually have 128-256GB of
>>>>> ram and the resulting crash dumps are in the range of 14-20gb which is
>>>>> very bug for the type of analysis I'm usually performing on crashed
>>>>> machine. I was wondering whether there is a way to further reduce the
>>>>> size and the time to take the dump (now it takes around 25 minutes to
>>>>> collect such a dump). I've seen reports where people with TBs of ram
>>>>> take that long, meaning for a machine with 256gb it should be even
>>>>> faster. I've been running this configuration on kernels 3.12.28 and 4.1
>>>>> where mmap for the vmcore file is supported.
>>>>>
>>>>> Please advise
>>>>
>>>> Hi nikolay,
>>>>
>>>> Yes, this issue is what we are concerning a lot.
>>>> About the current situation, try --split, it will save time.
>>>>
>>>>
>>>> And lzo/snappy instead of zlib, these two compression format are faster
>>>> but need more space to save. Or if you still want zlib (to save space),
>>>> try multiple threads, check the following site, it will help you:
>>>>
>>>> https://lists.fedoraproject.org/pipermail/kexec/2015-September/002322.html
>>>>
>>>>
>>>>
>>> .
>>>
>>
>>
> .
>
--
Regards
Qiao Nuohan
_______________________________________________
kexec mailing list
kexec@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/kexec
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: Reducing the size of the dump file/speeding up collection
2015-09-18 12:45 ` Nikolay Borisov
2015-09-21 6:27 ` qiaonuohan
@ 2015-09-23 1:44 ` Baoquan He
1 sibling, 0 replies; 8+ messages in thread
From: Baoquan He @ 2015-09-23 1:44 UTC (permalink / raw)
To: Nikolay Borisov; +Cc: qiaonuohan, kexec, SiteGround Operations
On 09/18/15 at 03:45pm, Nikolay Borisov wrote:
> Yeah, I did see the commit browser. But in my case I haven't even tested
> the split option so I guess there are things to try. Am I correct in my
> understanding as to how --split is supposed to work ( i tried that to no
> avail though):
>
> My core_collector line is this:
>
> core_collector makedumpfile --message-level 1 -d 3 --split dump1 dump2
> dump3 dump4 dump5 dump6
>
> And then in /etc/sysconfig/kdump I have:
>
> KDUMP_COMMANDLINE_APPEND="irqpoll nr_cpus=6 reset_devices
> cgroup_disable=memory mce=off"
disable_cpu_apicid need be specified if nr_cpus > 1. Otherwise it could
be hang.
>
> (the machine I'm testing on has 4 cores x2 hyperthreads so 8 logical
> cores in total). Do I need to do something else to utilize the --split
> option?
>
> On 09/18/2015 05:38 AM, qiaonuohan wrote:
> > On 09/17/2015 02:32 PM, Nikolay Borisov wrote:
> >> Hi Qiao,
> >>
> >> Thanks for the reply. So far I haven't been using the the compression
> >> feature of makedumpfile. But I want to ask if anything wouldn't
> >> compression make the dump process slower since in addition to having to
> >> write the dump to disk it also has to compress it which would put more
> >> strain on the cpu. Also, which part of the dump process is the
> >> bottleneck:
> >>
> >> - Reading from /proc/vmcore - that has mmap support so should be fairly
> >> fast?
> >> - Discarding unnecessary pages as memory is being scanned?
> >> - Writing/compressing content to disk?
> >
> > I cannot recall percentage of each part. But writing/compression takes most
> > of the time
> >
> > 1. mmap is used for reading faster
> > 2. --split is used to split the dump task into several processes, so
> > compressing
> > and writing will be speeded up.
> > 3. multiple-thread is another option for speeding up compressing, it is
> > a recently
> > committed patch, so you cannot find it in the master branch, checkout
> > devel branch
> > or find it here:
> >
> > http://sourceforge.net/p/makedumpfile/code/commit_browser
> >
> > Make makedumpfile available to read and compress pages parallelly.
> >
> >>
> >> Regards,
> >> Nikolay
> >>
> >> On 09/17/2015 06:27 AM, qiaonuohan wrote:
> >>> On 09/16/2015 04:30 PM, Nikolay Borisov wrote:
> >>>> Hello,
> >>>>
> >>>> I've been using makedumpfile as the crash collector with the -d31
> >>>> parameter. The machine this is being run on usually have 128-256GB of
> >>>> ram and the resulting crash dumps are in the range of 14-20gb which is
> >>>> very bug for the type of analysis I'm usually performing on crashed
> >>>> machine. I was wondering whether there is a way to further reduce the
> >>>> size and the time to take the dump (now it takes around 25 minutes to
> >>>> collect such a dump). I've seen reports where people with TBs of ram
> >>>> take that long, meaning for a machine with 256gb it should be even
> >>>> faster. I've been running this configuration on kernels 3.12.28 and 4.1
> >>>> where mmap for the vmcore file is supported.
> >>>>
> >>>> Please advise
> >>>
> >>> Hi nikolay,
> >>>
> >>> Yes, this issue is what we are concerning a lot.
> >>> About the current situation, try --split, it will save time.
> >>>
> >>>
> >>> And lzo/snappy instead of zlib, these two compression format are faster
> >>> but need more space to save. Or if you still want zlib (to save space),
> >>> try multiple threads, check the following site, it will help you:
> >>>
> >>> https://lists.fedoraproject.org/pipermail/kexec/2015-September/002322.html
> >>>
> >>>
> >>>
> >> .
> >>
> >
> >
>
> _______________________________________________
> kexec mailing list
> kexec@lists.infradead.org
> http://lists.infradead.org/mailman/listinfo/kexec
_______________________________________________
kexec mailing list
kexec@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/kexec
^ permalink raw reply [flat|nested] 8+ messages in thread
end of thread, other threads:[~2015-09-23 1:44 UTC | newest]
Thread overview: 8+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2015-09-16 8:30 Reducing the size of the dump file/speeding up collection Nikolay Borisov
2015-09-17 3:27 ` qiaonuohan
2015-09-17 6:32 ` Nikolay Borisov
2015-09-17 7:08 ` Nikolay Borisov
2015-09-18 2:38 ` qiaonuohan
2015-09-18 12:45 ` Nikolay Borisov
2015-09-21 6:27 ` qiaonuohan
2015-09-23 1:44 ` Baoquan He
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox