* Reducing the size of the dump file/speeding up collection @ 2015-09-16 8:30 Nikolay Borisov 2015-09-17 3:27 ` qiaonuohan 0 siblings, 1 reply; 8+ messages in thread From: Nikolay Borisov @ 2015-09-16 8:30 UTC (permalink / raw) To: kexec; +Cc: SiteGround Operations Hello, I've been using makedumpfile as the crash collector with the -d31 parameter. The machine this is being run on usually have 128-256GB of ram and the resulting crash dumps are in the range of 14-20gb which is very bug for the type of analysis I'm usually performing on crashed machine. I was wondering whether there is a way to further reduce the size and the time to take the dump (now it takes around 25 minutes to collect such a dump). I've seen reports where people with TBs of ram take that long, meaning for a machine with 256gb it should be even faster. I've been running this configuration on kernels 3.12.28 and 4.1 where mmap for the vmcore file is supported. Please advise Regards, Nikolay _______________________________________________ kexec mailing list kexec@lists.infradead.org http://lists.infradead.org/mailman/listinfo/kexec ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: Reducing the size of the dump file/speeding up collection 2015-09-16 8:30 Reducing the size of the dump file/speeding up collection Nikolay Borisov @ 2015-09-17 3:27 ` qiaonuohan 2015-09-17 6:32 ` Nikolay Borisov 0 siblings, 1 reply; 8+ messages in thread From: qiaonuohan @ 2015-09-17 3:27 UTC (permalink / raw) To: Nikolay Borisov, kexec; +Cc: SiteGround Operations On 09/16/2015 04:30 PM, Nikolay Borisov wrote: > Hello, > > I've been using makedumpfile as the crash collector with the -d31 > parameter. The machine this is being run on usually have 128-256GB of > ram and the resulting crash dumps are in the range of 14-20gb which is > very bug for the type of analysis I'm usually performing on crashed > machine. I was wondering whether there is a way to further reduce the > size and the time to take the dump (now it takes around 25 minutes to > collect such a dump). I've seen reports where people with TBs of ram > take that long, meaning for a machine with 256gb it should be even > faster. I've been running this configuration on kernels 3.12.28 and 4.1 > where mmap for the vmcore file is supported. > > Please advise Hi nikolay, Yes, this issue is what we are concerning a lot. About the current situation, try --split, it will save time. And lzo/snappy instead of zlib, these two compression format are faster but need more space to save. Or if you still want zlib (to save space), try multiple threads, check the following site, it will help you: https://lists.fedoraproject.org/pipermail/kexec/2015-September/002322.html -- Regards Qiao Nuohan > > Regards, > Nikolay > > _______________________________________________ > kexec mailing list > kexec@lists.infradead.org > http://lists.infradead.org/mailman/listinfo/kexec > _______________________________________________ kexec mailing list kexec@lists.infradead.org http://lists.infradead.org/mailman/listinfo/kexec ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: Reducing the size of the dump file/speeding up collection 2015-09-17 3:27 ` qiaonuohan @ 2015-09-17 6:32 ` Nikolay Borisov 2015-09-17 7:08 ` Nikolay Borisov 2015-09-18 2:38 ` qiaonuohan 0 siblings, 2 replies; 8+ messages in thread From: Nikolay Borisov @ 2015-09-17 6:32 UTC (permalink / raw) To: qiaonuohan, kexec; +Cc: SiteGround Operations Hi Qiao, Thanks for the reply. So far I haven't been using the the compression feature of makedumpfile. But I want to ask if anything wouldn't compression make the dump process slower since in addition to having to write the dump to disk it also has to compress it which would put more strain on the cpu. Also, which part of the dump process is the bottleneck: - Reading from /proc/vmcore - that has mmap support so should be fairly fast? - Discarding unnecessary pages as memory is being scanned? - Writing/compressing content to disk? Regards, Nikolay On 09/17/2015 06:27 AM, qiaonuohan wrote: > On 09/16/2015 04:30 PM, Nikolay Borisov wrote: >> Hello, >> >> I've been using makedumpfile as the crash collector with the -d31 >> parameter. The machine this is being run on usually have 128-256GB of >> ram and the resulting crash dumps are in the range of 14-20gb which is >> very bug for the type of analysis I'm usually performing on crashed >> machine. I was wondering whether there is a way to further reduce the >> size and the time to take the dump (now it takes around 25 minutes to >> collect such a dump). I've seen reports where people with TBs of ram >> take that long, meaning for a machine with 256gb it should be even >> faster. I've been running this configuration on kernels 3.12.28 and 4.1 >> where mmap for the vmcore file is supported. >> >> Please advise > > Hi nikolay, > > Yes, this issue is what we are concerning a lot. > About the current situation, try --split, it will save time. > > > And lzo/snappy instead of zlib, these two compression format are faster > but need more space to save. Or if you still want zlib (to save space), > try multiple threads, check the following site, it will help you: > > https://lists.fedoraproject.org/pipermail/kexec/2015-September/002322.html > > _______________________________________________ kexec mailing list kexec@lists.infradead.org http://lists.infradead.org/mailman/listinfo/kexec ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: Reducing the size of the dump file/speeding up collection 2015-09-17 6:32 ` Nikolay Borisov @ 2015-09-17 7:08 ` Nikolay Borisov 2015-09-18 2:38 ` qiaonuohan 1 sibling, 0 replies; 8+ messages in thread From: Nikolay Borisov @ 2015-09-17 7:08 UTC (permalink / raw) To: qiaonuohan, kexec; +Cc: SiteGround Operations Just to follow up: which version of kexec/makedumpfile are required to have the parallel dump feature? The thread you referenced adds documentation how to use that and I grepped the source of makedumpfile 1.5.8 and couldn't find any references to the num-threads option? So what version of the code do I need to test that ? On 09/17/2015 09:32 AM, Nikolay Borisov wrote: > Hi Qiao, > > Thanks for the reply. So far I haven't been using the the compression > feature of makedumpfile. But I want to ask if anything wouldn't > compression make the dump process slower since in addition to having to > write the dump to disk it also has to compress it which would put more > strain on the cpu. Also, which part of the dump process is the bottleneck: > > - Reading from /proc/vmcore - that has mmap support so should be fairly > fast? > - Discarding unnecessary pages as memory is being scanned? > - Writing/compressing content to disk? > > Regards, > Nikolay > > On 09/17/2015 06:27 AM, qiaonuohan wrote: >> On 09/16/2015 04:30 PM, Nikolay Borisov wrote: >>> Hello, >>> >>> I've been using makedumpfile as the crash collector with the -d31 >>> parameter. The machine this is being run on usually have 128-256GB of >>> ram and the resulting crash dumps are in the range of 14-20gb which is >>> very bug for the type of analysis I'm usually performing on crashed >>> machine. I was wondering whether there is a way to further reduce the >>> size and the time to take the dump (now it takes around 25 minutes to >>> collect such a dump). I've seen reports where people with TBs of ram >>> take that long, meaning for a machine with 256gb it should be even >>> faster. I've been running this configuration on kernels 3.12.28 and 4.1 >>> where mmap for the vmcore file is supported. >>> >>> Please advise >> >> Hi nikolay, >> >> Yes, this issue is what we are concerning a lot. >> About the current situation, try --split, it will save time. >> >> >> And lzo/snappy instead of zlib, these two compression format are faster >> but need more space to save. Or if you still want zlib (to save space), >> try multiple threads, check the following site, it will help you: >> >> https://lists.fedoraproject.org/pipermail/kexec/2015-September/002322.html >> >> _______________________________________________ kexec mailing list kexec@lists.infradead.org http://lists.infradead.org/mailman/listinfo/kexec ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: Reducing the size of the dump file/speeding up collection 2015-09-17 6:32 ` Nikolay Borisov 2015-09-17 7:08 ` Nikolay Borisov @ 2015-09-18 2:38 ` qiaonuohan 2015-09-18 12:45 ` Nikolay Borisov 1 sibling, 1 reply; 8+ messages in thread From: qiaonuohan @ 2015-09-18 2:38 UTC (permalink / raw) To: Nikolay Borisov, kexec; +Cc: SiteGround Operations On 09/17/2015 02:32 PM, Nikolay Borisov wrote: > Hi Qiao, > > Thanks for the reply. So far I haven't been using the the compression > feature of makedumpfile. But I want to ask if anything wouldn't > compression make the dump process slower since in addition to having to > write the dump to disk it also has to compress it which would put more > strain on the cpu. Also, which part of the dump process is the bottleneck: > > - Reading from /proc/vmcore - that has mmap support so should be fairly > fast? > - Discarding unnecessary pages as memory is being scanned? > - Writing/compressing content to disk? I cannot recall percentage of each part. But writing/compression takes most of the time 1. mmap is used for reading faster 2. --split is used to split the dump task into several processes, so compressing and writing will be speeded up. 3. multiple-thread is another option for speeding up compressing, it is a recently committed patch, so you cannot find it in the master branch, checkout devel branch or find it here: http://sourceforge.net/p/makedumpfile/code/commit_browser Make makedumpfile available to read and compress pages parallelly. > > Regards, > Nikolay > > On 09/17/2015 06:27 AM, qiaonuohan wrote: >> On 09/16/2015 04:30 PM, Nikolay Borisov wrote: >>> Hello, >>> >>> I've been using makedumpfile as the crash collector with the -d31 >>> parameter. The machine this is being run on usually have 128-256GB of >>> ram and the resulting crash dumps are in the range of 14-20gb which is >>> very bug for the type of analysis I'm usually performing on crashed >>> machine. I was wondering whether there is a way to further reduce the >>> size and the time to take the dump (now it takes around 25 minutes to >>> collect such a dump). I've seen reports where people with TBs of ram >>> take that long, meaning for a machine with 256gb it should be even >>> faster. I've been running this configuration on kernels 3.12.28 and 4.1 >>> where mmap for the vmcore file is supported. >>> >>> Please advise >> >> Hi nikolay, >> >> Yes, this issue is what we are concerning a lot. >> About the current situation, try --split, it will save time. >> >> >> And lzo/snappy instead of zlib, these two compression format are faster >> but need more space to save. Or if you still want zlib (to save space), >> try multiple threads, check the following site, it will help you: >> >> https://lists.fedoraproject.org/pipermail/kexec/2015-September/002322.html >> >> > . > -- Regards Qiao Nuohan _______________________________________________ kexec mailing list kexec@lists.infradead.org http://lists.infradead.org/mailman/listinfo/kexec ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: Reducing the size of the dump file/speeding up collection 2015-09-18 2:38 ` qiaonuohan @ 2015-09-18 12:45 ` Nikolay Borisov 2015-09-21 6:27 ` qiaonuohan 2015-09-23 1:44 ` Baoquan He 0 siblings, 2 replies; 8+ messages in thread From: Nikolay Borisov @ 2015-09-18 12:45 UTC (permalink / raw) To: qiaonuohan, kexec; +Cc: SiteGround Operations Yeah, I did see the commit browser. But in my case I haven't even tested the split option so I guess there are things to try. Am I correct in my understanding as to how --split is supposed to work ( i tried that to no avail though): My core_collector line is this: core_collector makedumpfile --message-level 1 -d 3 --split dump1 dump2 dump3 dump4 dump5 dump6 And then in /etc/sysconfig/kdump I have: KDUMP_COMMANDLINE_APPEND="irqpoll nr_cpus=6 reset_devices cgroup_disable=memory mce=off" (the machine I'm testing on has 4 cores x2 hyperthreads so 8 logical cores in total). Do I need to do something else to utilize the --split option? On 09/18/2015 05:38 AM, qiaonuohan wrote: > On 09/17/2015 02:32 PM, Nikolay Borisov wrote: >> Hi Qiao, >> >> Thanks for the reply. So far I haven't been using the the compression >> feature of makedumpfile. But I want to ask if anything wouldn't >> compression make the dump process slower since in addition to having to >> write the dump to disk it also has to compress it which would put more >> strain on the cpu. Also, which part of the dump process is the >> bottleneck: >> >> - Reading from /proc/vmcore - that has mmap support so should be fairly >> fast? >> - Discarding unnecessary pages as memory is being scanned? >> - Writing/compressing content to disk? > > I cannot recall percentage of each part. But writing/compression takes most > of the time > > 1. mmap is used for reading faster > 2. --split is used to split the dump task into several processes, so > compressing > and writing will be speeded up. > 3. multiple-thread is another option for speeding up compressing, it is > a recently > committed patch, so you cannot find it in the master branch, checkout > devel branch > or find it here: > > http://sourceforge.net/p/makedumpfile/code/commit_browser > > Make makedumpfile available to read and compress pages parallelly. > >> >> Regards, >> Nikolay >> >> On 09/17/2015 06:27 AM, qiaonuohan wrote: >>> On 09/16/2015 04:30 PM, Nikolay Borisov wrote: >>>> Hello, >>>> >>>> I've been using makedumpfile as the crash collector with the -d31 >>>> parameter. The machine this is being run on usually have 128-256GB of >>>> ram and the resulting crash dumps are in the range of 14-20gb which is >>>> very bug for the type of analysis I'm usually performing on crashed >>>> machine. I was wondering whether there is a way to further reduce the >>>> size and the time to take the dump (now it takes around 25 minutes to >>>> collect such a dump). I've seen reports where people with TBs of ram >>>> take that long, meaning for a machine with 256gb it should be even >>>> faster. I've been running this configuration on kernels 3.12.28 and 4.1 >>>> where mmap for the vmcore file is supported. >>>> >>>> Please advise >>> >>> Hi nikolay, >>> >>> Yes, this issue is what we are concerning a lot. >>> About the current situation, try --split, it will save time. >>> >>> >>> And lzo/snappy instead of zlib, these two compression format are faster >>> but need more space to save. Or if you still want zlib (to save space), >>> try multiple threads, check the following site, it will help you: >>> >>> https://lists.fedoraproject.org/pipermail/kexec/2015-September/002322.html >>> >>> >>> >> . >> > > _______________________________________________ kexec mailing list kexec@lists.infradead.org http://lists.infradead.org/mailman/listinfo/kexec ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: Reducing the size of the dump file/speeding up collection 2015-09-18 12:45 ` Nikolay Borisov @ 2015-09-21 6:27 ` qiaonuohan 2015-09-23 1:44 ` Baoquan He 1 sibling, 0 replies; 8+ messages in thread From: qiaonuohan @ 2015-09-21 6:27 UTC (permalink / raw) To: Nikolay Borisov, kexec; +Cc: SiteGround Operations On 09/18/2015 08:45 PM, Nikolay Borisov wrote: > Yeah, I did see the commit browser. But in my case I haven't even tested > the split option so I guess there are things to try. Am I correct in my > understanding as to how --split is supposed to work ( i tried that to no > avail though): > > My core_collector line is this: > > core_collector makedumpfile --message-level 1 -d 3 --split dump1 dump2 > dump3 dump4 dump5 dump6 > > And then in /etc/sysconfig/kdump I have: > > KDUMP_COMMANDLINE_APPEND="irqpoll nr_cpus=6 reset_devices > cgroup_disable=memory mce=off" > > (the machine I'm testing on has 4 cores x2 hyperthreads so 8 logical > cores in total). Do I need to do something else to utilize the --split > option? As I recall, it is OK to go on the test. > > On 09/18/2015 05:38 AM, qiaonuohan wrote: >> On 09/17/2015 02:32 PM, Nikolay Borisov wrote: >>> Hi Qiao, >>> >>> Thanks for the reply. So far I haven't been using the the compression >>> feature of makedumpfile. But I want to ask if anything wouldn't >>> compression make the dump process slower since in addition to having to >>> write the dump to disk it also has to compress it which would put more >>> strain on the cpu. Also, which part of the dump process is the >>> bottleneck: >>> >>> - Reading from /proc/vmcore - that has mmap support so should be fairly >>> fast? >>> - Discarding unnecessary pages as memory is being scanned? >>> - Writing/compressing content to disk? >> >> I cannot recall percentage of each part. But writing/compression takes most >> of the time >> >> 1. mmap is used for reading faster >> 2. --split is used to split the dump task into several processes, so >> compressing >> and writing will be speeded up. >> 3. multiple-thread is another option for speeding up compressing, it is >> a recently >> committed patch, so you cannot find it in the master branch, checkout >> devel branch >> or find it here: >> >> http://sourceforge.net/p/makedumpfile/code/commit_browser >> >> Make makedumpfile available to read and compress pages parallelly. >> >>> >>> Regards, >>> Nikolay >>> >>> On 09/17/2015 06:27 AM, qiaonuohan wrote: >>>> On 09/16/2015 04:30 PM, Nikolay Borisov wrote: >>>>> Hello, >>>>> >>>>> I've been using makedumpfile as the crash collector with the -d31 >>>>> parameter. The machine this is being run on usually have 128-256GB of >>>>> ram and the resulting crash dumps are in the range of 14-20gb which is >>>>> very bug for the type of analysis I'm usually performing on crashed >>>>> machine. I was wondering whether there is a way to further reduce the >>>>> size and the time to take the dump (now it takes around 25 minutes to >>>>> collect such a dump). I've seen reports where people with TBs of ram >>>>> take that long, meaning for a machine with 256gb it should be even >>>>> faster. I've been running this configuration on kernels 3.12.28 and 4.1 >>>>> where mmap for the vmcore file is supported. >>>>> >>>>> Please advise >>>> >>>> Hi nikolay, >>>> >>>> Yes, this issue is what we are concerning a lot. >>>> About the current situation, try --split, it will save time. >>>> >>>> >>>> And lzo/snappy instead of zlib, these two compression format are faster >>>> but need more space to save. Or if you still want zlib (to save space), >>>> try multiple threads, check the following site, it will help you: >>>> >>>> https://lists.fedoraproject.org/pipermail/kexec/2015-September/002322.html >>>> >>>> >>>> >>> . >>> >> >> > . > -- Regards Qiao Nuohan _______________________________________________ kexec mailing list kexec@lists.infradead.org http://lists.infradead.org/mailman/listinfo/kexec ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: Reducing the size of the dump file/speeding up collection 2015-09-18 12:45 ` Nikolay Borisov 2015-09-21 6:27 ` qiaonuohan @ 2015-09-23 1:44 ` Baoquan He 1 sibling, 0 replies; 8+ messages in thread From: Baoquan He @ 2015-09-23 1:44 UTC (permalink / raw) To: Nikolay Borisov; +Cc: qiaonuohan, kexec, SiteGround Operations On 09/18/15 at 03:45pm, Nikolay Borisov wrote: > Yeah, I did see the commit browser. But in my case I haven't even tested > the split option so I guess there are things to try. Am I correct in my > understanding as to how --split is supposed to work ( i tried that to no > avail though): > > My core_collector line is this: > > core_collector makedumpfile --message-level 1 -d 3 --split dump1 dump2 > dump3 dump4 dump5 dump6 > > And then in /etc/sysconfig/kdump I have: > > KDUMP_COMMANDLINE_APPEND="irqpoll nr_cpus=6 reset_devices > cgroup_disable=memory mce=off" disable_cpu_apicid need be specified if nr_cpus > 1. Otherwise it could be hang. > > (the machine I'm testing on has 4 cores x2 hyperthreads so 8 logical > cores in total). Do I need to do something else to utilize the --split > option? > > On 09/18/2015 05:38 AM, qiaonuohan wrote: > > On 09/17/2015 02:32 PM, Nikolay Borisov wrote: > >> Hi Qiao, > >> > >> Thanks for the reply. So far I haven't been using the the compression > >> feature of makedumpfile. But I want to ask if anything wouldn't > >> compression make the dump process slower since in addition to having to > >> write the dump to disk it also has to compress it which would put more > >> strain on the cpu. Also, which part of the dump process is the > >> bottleneck: > >> > >> - Reading from /proc/vmcore - that has mmap support so should be fairly > >> fast? > >> - Discarding unnecessary pages as memory is being scanned? > >> - Writing/compressing content to disk? > > > > I cannot recall percentage of each part. But writing/compression takes most > > of the time > > > > 1. mmap is used for reading faster > > 2. --split is used to split the dump task into several processes, so > > compressing > > and writing will be speeded up. > > 3. multiple-thread is another option for speeding up compressing, it is > > a recently > > committed patch, so you cannot find it in the master branch, checkout > > devel branch > > or find it here: > > > > http://sourceforge.net/p/makedumpfile/code/commit_browser > > > > Make makedumpfile available to read and compress pages parallelly. > > > >> > >> Regards, > >> Nikolay > >> > >> On 09/17/2015 06:27 AM, qiaonuohan wrote: > >>> On 09/16/2015 04:30 PM, Nikolay Borisov wrote: > >>>> Hello, > >>>> > >>>> I've been using makedumpfile as the crash collector with the -d31 > >>>> parameter. The machine this is being run on usually have 128-256GB of > >>>> ram and the resulting crash dumps are in the range of 14-20gb which is > >>>> very bug for the type of analysis I'm usually performing on crashed > >>>> machine. I was wondering whether there is a way to further reduce the > >>>> size and the time to take the dump (now it takes around 25 minutes to > >>>> collect such a dump). I've seen reports where people with TBs of ram > >>>> take that long, meaning for a machine with 256gb it should be even > >>>> faster. I've been running this configuration on kernels 3.12.28 and 4.1 > >>>> where mmap for the vmcore file is supported. > >>>> > >>>> Please advise > >>> > >>> Hi nikolay, > >>> > >>> Yes, this issue is what we are concerning a lot. > >>> About the current situation, try --split, it will save time. > >>> > >>> > >>> And lzo/snappy instead of zlib, these two compression format are faster > >>> but need more space to save. Or if you still want zlib (to save space), > >>> try multiple threads, check the following site, it will help you: > >>> > >>> https://lists.fedoraproject.org/pipermail/kexec/2015-September/002322.html > >>> > >>> > >>> > >> . > >> > > > > > > _______________________________________________ > kexec mailing list > kexec@lists.infradead.org > http://lists.infradead.org/mailman/listinfo/kexec _______________________________________________ kexec mailing list kexec@lists.infradead.org http://lists.infradead.org/mailman/listinfo/kexec ^ permalink raw reply [flat|nested] 8+ messages in thread
end of thread, other threads:[~2015-09-23 1:44 UTC | newest] Thread overview: 8+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2015-09-16 8:30 Reducing the size of the dump file/speeding up collection Nikolay Borisov 2015-09-17 3:27 ` qiaonuohan 2015-09-17 6:32 ` Nikolay Borisov 2015-09-17 7:08 ` Nikolay Borisov 2015-09-18 2:38 ` qiaonuohan 2015-09-18 12:45 ` Nikolay Borisov 2015-09-21 6:27 ` qiaonuohan 2015-09-23 1:44 ` Baoquan He
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox