[CP2K-user] [CP2K:15183] Re: Does CP2K allow a multi-GPU run?

Alfio Lazzaro alfio.... at gmail.com
Thu Apr 22 11:22:32 UTC 2021


Could you try what I suggested:

export OMP_NUM_THREADS=10
mpirun -np 4 ./cp2k.psmp -i gold.inp -o gold_pbc.out

Please check the corresponding log.

As I said above, you need an MPI rank per GPU and you told us that you have 
4 GPUs, so you need 4 ranks (or multiple). With 10 you get unbalance.


Il giorno giovedì 22 aprile 2021 alle 10:17:27 UTC+2 ASSIDUO Network ha 
scritto:

> Correction, he told me to use:
>
> mpirun -np 10 cp2k.psmp -i gold.inp -o gold_pbc.out
>
> but it didn't run correctly.
>
> On Thu, Apr 22, 2021 at 9:51 AM Lenard Carroll <len... at gmail.com> 
> wrote:
>
>> He suggested I try out:
>> mpirun -n 10 cp2k.psmp -i gold.inp -o gold_pbc.out
>>
>> as he is hoping that will cause the 1 GPU to use 10 CPUs over the 
>> selected 4 GPUs.
>>
>>
>> On Thu, Apr 22, 2021 at 9:48 AM Alfio Lazzaro <al... at gmail.com> 
>> wrote:
>>
>>> Hi,
>>> Your command to run CP2K doesn't mention MPI (mpirun, mpiexc, ...). Are 
>>> you running with multiple ranks?
>>>
>>> You can check those lines in the output:
>>>
>>>  GLOBAL| Total number of message passing processes                      
>>>       32
>>>  GLOBAL| Number of threads for this process                              
>>>       4
>>>
>>> And check your numbers.
>>> I can guess you have 1 rank and 40 threads.
>>> To use 4 GPUs you need 4 ranks (and less threads per rank), i.e. 
>>> something like
>>>
>>> export OMP_NUM_THREADS=10
>>> mpiexec -n 4 ./cp2k.psmp -i gold.inp -o gold_pbc.out
>>>
>>> Please check with your sysadmin on how to run with multiple MPI ranks.
>>>
>>> Hope it helps.
>>>
>>> Alfio
>>>
>>>
>>>
>>> Il giorno mercoledì 21 aprile 2021 alle 09:26:53 UTC+2 ASSIDUO Network 
>>> ha scritto:
>>>
>>>> This is what my PBS file looks like:
>>>>
>>>> #!/bin/bash
>>>> #PBS -P <PROJECT>
>>>> #PBS -N <JOBNAME>
>>>> #PBS -l select=1:ncpus=40:ngpus=4
>>>> #PBS -l walltime=08:00:00
>>>> #PBS -q gpu_4
>>>> #PBS -m be
>>>> #PBS -M none
>>>>
>>>> module purge
>>>> module load chpc/cp2k/8.1.0/cuda10.1/openmpi-4.0.0/gcc-7.3.0
>>>> source $SETUP
>>>> cd $PBS_O_WORKDIR
>>>>
>>>> cp2k.psmp -i gold.inp -o gold_pbc.out
>>>> ~                                                                      
>>>>                                                  ~                          
>>>>                
>>>>
>>>> On Wed, Apr 21, 2021 at 9:22 AM Alfio Lazzaro <al... at gmail.com> 
>>>> wrote:
>>>>
>>>>> The way to use 4 GPUs per node is to use 4 MPI ranks. How many ranks 
>>>>> are you using?
>>>>>
>>>>> Il giorno martedì 20 aprile 2021 alle 19:44:15 UTC+2 ASSIDUO Network 
>>>>> ha scritto:
>>>>>
>>>>>> I'm asking, since the administrator running my country's HPC is 
>>>>>> saying that although I'm requesting access to 4 GPUs, CP2K is only using 1. 
>>>>>> I checked the following output:
>>>>>>  DBCSR| ACC: Number of devices/node                                  
>>>>>>           4
>>>>>>
>>>>>> And it shows that CP2K is picking up 4 GPUs.
>>>>>>
>>>>>> On Tuesday, April 20, 2021 at 3:00:17 PM UTC+2 ASSIDUO Network wrote:
>>>>>>
>>>>>>> I currently have access to 4 GPUs to run an AIMD simulation, but 
>>>>>>> only one of the GPUs are being used. Is there a way to use the other 3, and 
>>>>>>> if so, can you tell me how to set it up with a PBS job?
>>>>>>
>>>>>> -- 
>>>>> You received this message because you are subscribed to the Google 
>>>>> Groups "cp2k" group.
>>>>> To unsubscribe from this group and stop receiving emails from it, send 
>>>>> an email to cp... at googlegroups.com.
>>>>> To view this discussion on the web visit 
>>>>> https://groups.google.com/d/msgid/cp2k/70ba0fce-8636-4b75-940d-133ce4dbf0can%40googlegroups.com 
>>>>> <https://groups.google.com/d/msgid/cp2k/70ba0fce-8636-4b75-940d-133ce4dbf0can%40googlegroups.com?utm_medium=email&utm_source=footer>
>>>>> .
>>>>>
>>>> -- 
>>> You received this message because you are subscribed to the Google 
>>> Groups "cp2k" group.
>>> To unsubscribe from this group and stop receiving emails from it, send 
>>> an email to cp... at googlegroups.com.
>>> To view this discussion on the web visit 
>>> https://groups.google.com/d/msgid/cp2k/92e4f88d-fde8-4127-ab5f-0b98bbbba8ebn%40googlegroups.com 
>>> <https://groups.google.com/d/msgid/cp2k/92e4f88d-fde8-4127-ab5f-0b98bbbba8ebn%40googlegroups.com?utm_medium=email&utm_source=footer>
>>> .
>>>
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.cp2k.org/archives/cp2k-user/attachments/20210422/746cd175/attachment.htm>


More information about the CP2K-user mailing list