[CP2K-user] [CP2K:19612] Re: Error: Parameter ‘mpi_comm_null’ at (1) has not been declared

Mikhail Povarnitsyn povarnitsynme at gmail.com
Fri Dec 1 00:21:09 UTC 2023


Dear Frederick,

Thank you for the message. I've read the manual and found out the processor 
type: Intel(R) Xeon(R) CPU E5-2640 v4 @ 2.40GHz. By using gcc-9.3 and the 
command gcc -march=native -E -v - </dev/null 2>&1 | grep cc1, I obtained on 
the particular Intel-based node that the actual value of native is 
broadwell.

Then, I used the toolchain with the option --target-cpu=broadwell. 
Installation of libs and cp2k compilation passed successfully; however, I 
received a warning message: 
*** WARNING in fm/cp_fm_elpa.F:522 :: Setting real_kernel for ELPA failed 
***
 when running VIBRATIONAL_ANALYSIS. 

It seems that ELPA is not working for me with the option 
--target-cpu=broadwell. 
What steps can I take to resolve this issue?

Best regards, 
Mikhail

On Wednesday, November 29, 2023 at 8:36:23 PM UTC+1 Frederick Stein wrote:

> Dear Mikhail,
>
> It is to be expected that both arch files have the -march=native option. 
> My idea is mostly about ensuring that the compilation of the toolchain and 
> CP2K was performed for the correct architecture. The actual values of 
> -march and -mtune are determined by the compiler automatically. What 
> options work in your case depends on your compiler and the CPU. Have you 
> already consulted the manual of your supercomputing center? If it is about 
> fine-tuning, ask your supercomputing center for help or consult the CPU 
> specs and the manual of your compiler or just with leave it like that as 
> the main part of the performance comes from libraries. In case of doubt, 
> you may also check the timing report of CP2K at the very end of the output 
> file or ask here for further help regarding performance tuning (in a 
> different thread as that is feature-dependent).
>
> HTH
> Frederick
>
> Mikhail Povarnitsyn schrieb am Mittwoch, 29. November 2023 um 19:14:25 
> UTC+1:
>
>> Dear Frederick,
>>
>> Thank you for this. Following your suggestion, I ran the toolchain job on 
>> a node of interest, and upon successful completion, I observed that the 
>> 'local.psmp' arch file is identical to one obtained by compilation on the 
>> head node. In both cases, the -march=native option is present.
>>
>> Did I correctly understand your idea of achieving node-dependent 
>> compilation?  Where I can find the actual values of -march, -mtune, and 
>> other relevant parameters during the toolchain step? 
>>
>> Best regards, 
>> Mikhail
>>
>> On Wednesday, November 29, 2023 at 1:05:39 PM UTC+1 Frederick Stein wrote:
>>
>>> Dear Mikhail,
>>>
>>> Did you try to compile the code as part of a job on the given machine? 
>>> Then, the compiler should be able grep the correct flags.
>>>
>>> Best,
>>> Frederick 
>>>
>>> Mikhail Povarnitsyn schrieb am Mittwoch, 29. November 2023 um 13:00:52 
>>> UTC+1:
>>>
>>>> Dear Frederick,
>>>>
>>>> I appreciate your continued assistance. Given that we have a mixture of 
>>>> processor types (Intel Xeon and AMD EPYC), determining the optimal -march 
>>>> and -mtune options (now 'native', by default) is currently not 
>>>> straightforward.
>>>>
>>>> Best regards,
>>>> Mikhail
>>>> On Wednesday, November 29, 2023 at 9:52:35 AM UTC+1 Frederick Stein 
>>>> wrote:
>>>>
>>>>> Dear Mikhail,
>>>>>
>>>>> I am not quite the expert for these machine-dependent options. I 
>>>>> personally ignore this error message. Machine-related optimizations 
>>>>> dependent on the actual setup (CPU, cross-compilation, ...) and the 
>>>>> compiler. If you compile for a supercomputing cluster, it is recommended to 
>>>>> make use of the center-provided compiler wrappers as they may have better 
>>>>> optimized MPI libraries or setup for their machine. You may check the 
>>>>> manual of your compiler for further information on machine-dependent 
>>>>> options in case you know the CPU and its instructions set (For Gfortran: 
>>>>> https://gcc.gnu.org/onlinedocs/gcc/x86-Options.html).
>>>>>
>>>>> Maybe, some of the other folks are of better help here.
>>>>>
>>>>> Best,
>>>>> Frederick
>>>>> Mikhail Povarnitsyn schrieb am Mittwoch, 29. November 2023 um 00:41:41 
>>>>> UTC+1:
>>>>>
>>>>>> Dear Frederick,
>>>>>>
>>>>>> I wanted to express my gratitude for your advice on removing 
>>>>>> -D__MPI_F08; it was immensely helpful.
>>>>>>
>>>>>> Upon comparing the performance of the 'cp2k.popt' code across 
>>>>>> versions 2023.2, 7.1, and 9.1, I observed a consistent runtime of 
>>>>>> approximately 25 minutes, with minor variations within a few seconds for 
>>>>>> all versions. However, in the output of version 2023.2, I noticed a new 
>>>>>> message:
>>>>>>
>>>>>> *** HINT in environment.F:904 :: The compiler target flags (generic) 
>>>>>> used *** *** to build this binary cannot exploit all extensions of this CPU 
>>>>>> model *** *** (x86_avx2). Consider compiler target flags as part of FCFLAGS 
>>>>>> and *** *** CFLAGS (ARCH file). ***
>>>>>>
>>>>>> I would greatly appreciate your advice on how I can enhance the 
>>>>>> performance of the parallel version.
>>>>>>
>>>>>> Thank you in advance for your assistance.
>>>>>>
>>>>>> Best regards, Mikhail
>>>>>>
>>>>>>
>>>>>> On Tuesday, November 28, 2023 at 9:43:14 PM UTC+1 Frederick Stein 
>>>>>> wrote:
>>>>>>
>>>>>>> Dear Mikhail,
>>>>>>>
>>>>>>> Can you remove the -D__MPI_F08 and recompile. I think it might be 
>>>>>>> related to an unsufficient support of the mpi_f08 module which CP2K uses by 
>>>>>>> default with OpenMPI and which is not tested with older versions of 
>>>>>>> compiler and library. Alternatively, if you have access to a later version 
>>>>>>> of the library (also with the CP2K Toolchain, add the flags 
>>>>>>> --with-openmpi=install or --with-mpich=install or additionally with 
>>>>>>> --with-gcc=install to install GCC 13).
>>>>>>>
>>>>>>> Best regards,
>>>>>>> Frederick
>>>>>>>
>>>>>>> Mikhail Povarnitsyn schrieb am Dienstag, 28. November 2023 um 
>>>>>>> 18:45:51 UTC+1:
>>>>>>>
>>>>>>>> Dear Frederick,
>>>>>>>>
>>>>>>>> Thank you very much for the reply!
>>>>>>>>
>>>>>>>> 1) Yes, I mean OpenMPI 3.1.6.
>>>>>>>>
>>>>>>>> 2) 'local.psmp' file is attached, hope that is what you asked for.
>>>>>>>>
>>>>>>>> 3) Yes, I did the command  'source 
>>>>>>>> /user/povar/cp2k-2023.2/tools/toolchain/install/setup' after the toolchain.
>>>>>>>>
>>>>>>>> 4)  mpifort --version
>>>>>>>> GNU Fortran (GCC) 9.3.0
>>>>>>>> Copyright (C) 2019 Free Software Foundation, Inc.
>>>>>>>> This is free software; see the source for copying conditions. 
>>>>>>>>  There is NO
>>>>>>>> warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR 
>>>>>>>> PURPOSE.
>>>>>>>>
>>>>>>>> Best regards,
>>>>>>>> Mikhail
>>>>>>>>
>>>>>>>> On Tuesday, November 28, 2023 at 1:08:25 PM UTC+1 Frederick Stein 
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Dear Mikhail,
>>>>>>>>>
>>>>>>>>> I suppose you mean OpenMPI 3.1.6 (MPI is just the standard 
>>>>>>>>> defining an interface for parallel programming). Could you post your arch 
>>>>>>>>> file of the parallel build? Did you source the setup file after the 
>>>>>>>>> toolchain script finished? Could you also post the output of `mpifort 
>>>>>>>>> --version`?
>>>>>>>>>
>>>>>>>>> Best,
>>>>>>>>> Frederick
>>>>>>>>>
>>>>>>>>> Mikhail Povarnitsyn schrieb am Dienstag, 28. November 2023 um 
>>>>>>>>> 11:06:14 UTC+1:
>>>>>>>>>
>>>>>>>>>> Dear Developers and Users,
>>>>>>>>>>
>>>>>>>>>> I am attempting to install the latest version, 2023.2, using the 
>>>>>>>>>> GNU compiler (gcc 9.3.0) along with MPI 3.1.6. I employed the toolchain 
>>>>>>>>>> script as follows: './install_cp2k_toolchain.sh'.
>>>>>>>>>>
>>>>>>>>>> The serial version 'ssmp' has been successfully compiled. 
>>>>>>>>>> However, the compilation of the parallel version 'psmp' failed with the 
>>>>>>>>>> following error: 
>>>>>>>>>>
>>>>>>>>>> /user/povar/cp2k-2023.2/exts/dbcsr/src/mpi/dbcsr_mpiwrap.F:106:53:
>>>>>>>>>>
>>>>>>>>>>   106 |    MPI_COMM_TYPE, PARAMETER :: mp_comm_null_handle = 
>>>>>>>>>> MPI_COMM_NULL
>>>>>>>>>>       |                                                     1
>>>>>>>>>> Error: Parameter ‘mpi_comm_null’ at (1) has not been declared or 
>>>>>>>>>> is a variable, which does not reduce to a constant expression
>>>>>>>>>> /user/povar/cp2k-2023.2/exts/dbcsr/src/mpi/dbcsr_mpiwrap.F:107:53:
>>>>>>>>>>
>>>>>>>>>>   107 |    MPI_COMM_TYPE, PARAMETER :: mp_comm_self_handle = 
>>>>>>>>>> MPI_COMM_SELF
>>>>>>>>>>       |                                                     1
>>>>>>>>>> Error: Parameter ‘mpi_comm_self’ at (1) has not been declared or 
>>>>>>>>>> is a variable, which does not reduce to a constant expression
>>>>>>>>>> /user/povar/cp2k-2023.2/exts/dbcsr/src/mpi/dbcsr_mpiwrap.F:108:54:
>>>>>>>>>>
>>>>>>>>>>   108 |    MPI_COMM_TYPE, PARAMETER :: mp_comm_world_handle = 
>>>>>>>>>> MPI_COMM_WORLD
>>>>>>>>>>
>>>>>>>>>> and other similar errors. 
>>>>>>>>>>
>>>>>>>>>> Could you please help?
>>>>>>>>>>
>>>>>>>>>> Best regards
>>>>>>>>>> Mikhail
>>>>>>>>>>
>>>>>>>>>>

-- 
You received this message because you are subscribed to the Google Groups "cp2k" group.
To unsubscribe from this group and stop receiving emails from it, send an email to cp2k+unsubscribe at googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/cp2k/0619e076-5633-46b3-bdfb-da185e699a6cn%40googlegroups.com.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.cp2k.org/archives/cp2k-user/attachments/20231130/1d1b8546/attachment-0001.htm>


More information about the CP2K-user mailing list