MPI_wait problem in cp2k 4.1 with openmpi_2.0.0
Alfio Lazzaro
alfio.... at gmail.com
Wed Mar 29 11:44:36 UTC 2017
I'm sorry, my request was to run CP2K 2.1 with OpenMPI 2.0.0. This test
would tell us if there is a problem in CP2K 4.1 test, or if it is something
related to OpenMPI version.
In any case, let's first analyze your logs. I see that there are a lot of
differences.
First of all, the energy result:
cp2k-4.1:
ENERGY| Total FORCE_EVAL ( QS ) energy (a.u.):
-1306.152561776134235
cp2k-2.1:
ENERGY| Total FORCE_EVAL ( QS ) energy (a.u.):
-1306.153425733494032
It is a quite large difference... By looking at the SCF steps, I see that
4.1 does
*** SCF run converged in 13 steps ***
while the 2.1 does
*** SCF run converged in 7 steps ***
That's a large difference! The 2.1 does a completely different set of
calculations (much less than 4.1). For instance, the number of MP_Wait
calls are:
cp2k-2.1: 27263
cp2k-4.1: 198040
So, definitely I would conclude that the problem is not in OpenMPI. At this
point, I will stop to compare with cp2k-2.1. Assuming that you are
interested in using the newer CP2K, I suggest to check first your
installation by running the regression tests:
https://www.cp2k.org/dev:regtesting
A couple of other suggestions:
1) I know that x86-64 arch file is pretty complicated, but this is the
right arch file for your tests. To get a simplified version, you can see
the arch file used in our regression tests dashboard, for
example: http://cp2k-www.epcc.ed.ac.uk/phi/popt/regtest-0 (the arch file is
on the top)
2) you can get the best performance when using a minimal *square* number of
MPI ranks. In your case, I see that you are requesting 24 ranks, I think
the best would be to use 4 ranks and 6 threads (for which you have to use
PSMP version, example at http://cp2k-www.epcc.ed.ac.uk/phi/psmp/regtest-0)
3) you can improve the overall performance by using libxsmm library. Take a
look at INSTALL file on how to use this library with CP2K
Alfio
Il giorno mercoledì 29 marzo 2017 10:36:43 UTC+2, jim wang ha scritto:
>
> I did what you said and checked the x86-64 arhc files but they are really
> complicated with so many command lines. I think they are not useful for my
> compilation demands. As for the libraries, I am sure that the libraries
> linked to cp2k-4.1 and cp2k-2.1 are nearly the same except for FFT
> libs(Intel FFT for cp2k-4.1 and FFTW3 for cp2k-2.1). My tests showed that
> cp2k-4.1 runned at exactly slow manner with Intel FFT libs and FFTW3 lib.
>
> The two output files are attached to my reply. I hope we can find out what
> is the ultimate reason for the bad performance of cp2k-4.1.
>
> 在 2017年3月29日星期三 UTC+8下午4:14:08,Alfio Lazzaro写道:
>>
>> OK, I answered to another email related to your problem, where I said
>> that Intel Xeon is a x86-64 architecture. IA64 is the Intel Itanium.
>> Therefore, please us the x86-64 arch file as a template. Anyway, this is
>> not really related to your problem with OpenMPI (I hope so!)...
>>
>> Concerning your last email, yes, please attach the CP2K logs.
>> Then, have you tried to compile CP2K 4.1 with the same CP2K 2.1 libraries
>> (or vice versa)?
>>
>> Alfio
>>
>>
>>
>> Il giorno lunedì 27 marzo 2017 11:38:52 UTC+2, jim wang ha scritto:
>>>
>>> Hi, everybody!
>>>
>>> I am using cp2k 4.1 for the testing in our new cluster. But strangly,
>>> the result showed that the cp2k 4.1 version is 3 to 4 times slower than
>>> cp2k 2.1 version built on the same cluster. After examining the output file
>>> genertated by both binary file running the same job, I found out that the
>>> MPI_wait function may be the key problem.
>>>
>>> Here is the result of time consumed by MPI_wait function:
>>> 1. cp2k 4.1: MPI_wait time:1131(s) , Total run time: 1779(s)
>>> 2. cp2k 2.1: MPI_wait time:68(s), Total run time: 616(s)
>>>
>>> How can I determine whether the problem should be with our cluster or
>>> the compilation?
>>> Hope you guys can give me some hints on the version comparison.
>>>
>>> THANKS!!!
>>>
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.cp2k.org/archives/cp2k-user/attachments/20170329/f6e81659/attachment.htm>
More information about the CP2K-user
mailing list