[PATCH RFC 00/11] makedumpfile: parallel processing

Atsushi Kumagai ats-kumagai at wm.jp.nec.com
Thu Dec 24 00:20:42 PST 2015


>> >> >> >> Could you provide the information of your cpu ?
>> >> >> >> I will do some further investigation later.
>> >> >> >>
>> >> >> >
>> >> >> > OK, of course, here is the information of cpu:
>> >> >> >
>> >> >> > # lscpu
>> >> >> > Architecture:          x86_64
>> >> >> > CPU op-mode(s):        32-bit, 64-bit
>> >> >> > Byte Order:            Little Endian
>> >> >> > CPU(s):                48
>> >> >> > On-line CPU(s) list:   0-47
>> >> >> > Thread(s) per core:    1
>> >> >> > Core(s) per socket:    6
>> >> >> > Socket(s):             8
>> >> >> > NUMA node(s):          8
>> >> >> > Vendor ID:             AuthenticAMD
>> >> >> > CPU family:            16
>> >> >> > Model:                 8
>> >> >> > Model name:            Six-Core AMD Opteron(tm) Processor 8439 SE
>> >> >> > Stepping:              0
>> >> >> > CPU MHz:               2793.040
>> >> >> > BogoMIPS:              5586.22
>> >> >> > Virtualization:        AMD-V
>> >> >> > L1d cache:             64K
>> >> >> > L1i cache:             64K
>> >> >> > L2 cache:              512K
>> >> >> > L3 cache:              5118K
>> >> >> > NUMA node0 CPU(s):     0,8,16,24,32,40
>> >> >> > NUMA node1 CPU(s):     1,9,17,25,33,41
>> >> >> > NUMA node2 CPU(s):     2,10,18,26,34,42
>> >> >> > NUMA node3 CPU(s):     3,11,19,27,35,43
>> >> >> > NUMA node4 CPU(s):     4,12,20,28,36,44
>> >> >> > NUMA node5 CPU(s):     5,13,21,29,37,45
>> >> >> > NUMA node6 CPU(s):     6,14,22,30,38,46
>> >> >> > NUMA node7 CPU(s):     7,15,23,31,39,47
>> >> >>
>> >> >> This CPU assignment on NUMA nodes looks interesting. Is it possible
>> >> >> that this affects performance of makedumpfile? This is just a guess.
>> >> >>
>> >> >> Could you check whether the performance gets imporoved if you run each
>> >> >> thread on the same NUMA node? For example:
>> >> >>
>> >> >>   # taskset -c 0,8,16,24 makedumpfile --num-threads 4 -c -d 0 vmcore
>> >> >>   vmcore-cd0
>> >> >>
>> >> > Hi HATAYAMA,
>> >> >
>> >> > I think your guess is right, but maybe your command has a little
>> >> > problem.
>> >> >
>> >> > From my test, the NUMA did affect the performance, but not too much.
>> >> > The average time of cpus in the same NUMA node:
>> >> > # taskset -c 0,8,16,24,32 makedumpfile --num-threads 4 -c -d 0 vmcore
>> >> > vmcore-cd0
>> >> > is 314s
>> >> > The average time of cpus in different NUMA node:
>> >> > # taskset -c 2,3,5,6,7 makedumpfile --num-threads 4 -c -d 0 vmcore
>> >> > vmcore-cd0
>> >> > is 354s
>> >> >
>> >>
>> >> Hmm, according to some previous discussion, what we should see here is
>> >> whether it affects performance of makedumpfile with --num-threads 1
>> >> and -d 31. So you should need to compare:
>> >>
>> >>     # taskset 0,8 makedumpfile --num-threads 1 -c -d 31 vmcore vmcore-d31
>> >>
>> >> with:
>> >>
>> >>     # taskset 0 makedumpfile -c -d 0 vmcore vmcore-d31
>>
>> I removed -c option wrongly. What I wanted to write is:
>>
>>     # taskset -c 0,8 makedumpfile --num-threads 1 -d 31 vmcore vmcore-d31
>>
>> and:
>>
>>     # taskset -c 0 makedumpfile -d 31 vmcore vmcore-d31
>>
>> just in case...

Why did you remove -c option from makedumpfile ?
We are discussing the performance with compression.
I think the below is correct:

       # taskset -c 0,8 makedumpfile --num-threads 1 [-c|-l|-p] -d 31 vmcore vmcore-d31

and:

       # taskset -c 0 makedumpfile [-c|-l|-p] -d 31 vmcore vmcore-d31


Thanks,
Atsushi Kumagai

>Hi HATAYAMA,
>
>the average time of
># taskset -c 0,8 makedumpfile --num-threads 1 -d 31 vmcore vmcore-d31
>is 33s.
>the average time of
># taskset -c 0 makedumpfile -d 31 vmcore vmcore-d31
>is 18s.
>
>My test steps:
>1. change /etc/kdump/conf with
>"core_collector makedumpfile -l --message-level 1 -d 31"
>2. make a crash
>3. cd into the directory of the vmcore made by kdump
>4. in the directory of vmcore do
># taskset -c 0,8 makedumpfile --num-threads 1 -d 31 vmcore vmcore-d31
>or
># taskset -c 0 makedumpfile -d 31 vmcore vmcore-d31
>
>if there are there any problems, please tell me.
>
>Thanks,
>Chao Fan
>
>> >>
>> >> Also, I'm assuming that you've done these benchmark on kdump 1st
>> >> kernel, not kdump 2nd kernel. Is this correct?
>> >>
>> > Hi HATAYAMA,
>> >
>> > I test in the first kernel, not in the kdump second kernel.
>> >
>>
>> I see.
>>
>> --
>> Thanks.
>> HATAYAMA, Daisuke
>> _______________________________________________
>> kexec mailing list
>> kexec at lists.infradead.org
>> http://lists.infradead.org/mailman/listinfo/kexec
>>
>
>_______________________________________________
>kexec mailing list
>kexec at lists.infradead.org
>http://lists.infradead.org/mailman/listinfo/kexec


More information about the kexec mailing list