[CP2K-user] [CP2K:20726] Re: out of memory: C3N4 (56 atoms) HES06 + RI-HFX + ADMM

yj jiang rudin.jiang at gmail.com
Sun Sep 22 09:13:50 UTC 2024


Deer Bussy

Your suggestions have been extremely helpful. I am now able to run this 
system on a 512GB HPC.
The running speed is acceptable to me, considering the large amount of 
computation involved. 
Thank you very much for your help.

Best wishes.
Jiang
在2024年9月19日星期四 UTC+8 21:33:29<Augustin Bussy> 写道:

> Hi Jiang,
> I was able to run your input file on a 512G system, but I'd like to add a 
> few comments:
> 1) I somehow missed the fact that you are running RI-HFX with K-points. In 
> such a case, the choice of the RI metric is not crucial for performance (or 
> memory usage), and the best course of action is to keep the default.
> 2) What really matters is the number of atoms in the unit cell, the 
> diffuseness of the basis set, and the range of the HFX potential. There is 
> not much you can do here, especially since you already use ADMM. The range 
> of the HFX potential really matters, because this influences how many 
> periodic images of your system need to be considered. Typically, this  has 
> a limited impact for 2D systems, because there should only be images along 
> the XY axes. However, if the extent of the cell in the Z direction is too 
> short, periodic images get involved as well. *I found that increasing the 
> Z cell parameters from 20 Ang to 30 Ang speeds up the calculations by a 
> factor 2*, because less images are needed.
> 3) Unless you reach a very high number of K-points (~1000), this has 
> practically no impact on cost, and only a little on memory. You might as 
> well use a denser grid than 3x3x1
> 4) I am not sure if your system is a metal. If it is not, you should avoid 
> using Broyden mixing and smearing, as SCF convergence slows down.
> 5) While you can run this calculation on a single HPC node with 512G, this 
> will be extremely slow. I highly recommand using more resources, and 
> increasing the KP_NGROUPS keyword the the HF%RI section. This is not a 
> small calculation you are attempting.
>
> I invite you to read the paper written for the method: 
> https://doi.org/10.1063/5.0189659, this might help you understand some of 
> the points above.
> I hope that helps. Best,
> Augustin
> On Wednesday 18 September 2024 at 16:54:17 UTC+2 yj jiang wrote:
>
>> Thank you all very much for your help. A few days ago, I thought I had 
>> uploaded the output file here, but it seems that due to the large file 
>> size, the upload may not have been successful. I apologize for this.
>>
>> Yesterday, I seemed to have found the reason. The supercomputing job 
>> system I am using has a memory allocation limit for each process, and this 
>> issue is not exclusive to cp2k jobs. I am now planning to set the maximum 
>> memory usage for cp2k to see if it resolves the problem.
>>
>> Thank you again for your help!
>>
>> Jiang
>>
>> 在2024年9月17日星期二 UTC+8 15:33:33<Augustin Bussy> 写道:
>>
>> Hi,
>> everything that Frederick mentioned is very relevant. Additionally, I'd 
>> like to add that the method was developed with small unit cells in mind. 
>> Your cell is quite large, and this might cause memory issues, although 
>> having a 2D material should help. I'll see if I am able to run it.
>> Best,
>> Augustin
>>
>> On Monday 16 September 2024 at 08:18:57 UTC+2 Frederick Stein wrote:
>>
>> Due to my lack of experience, I can just guess from my knowledge on RI 
>> methods, the manual and what I was told by the developer. Please also 
>> consider the respective manual page: 
>> https://manual.cp2k.org/trunk/CP2K_INPUT/FORCE_EVAL/DFT/XC/HF/RI.html#CP2K_INPUT.FORCE_EVAL.DFT.XC.HF.RI 
>> .
>> - Change RI_METRIC to TRUNCATED and CUTOFF_RADIUS in the &RI section to a 
>> value between 1.0 and 2.0 . Lower values improve sparsity, larger values 
>> increase accuracy. The shortrange Coulomb metric has a rather large range. 
>> If you would like to stick to it, set the OMEGA parameter of the &RI 
>> section to a value larger then 0.11 (let's say 1-10).
>> - Just for testing, try a single k-point.
>> - Try other values of MEMORY_CUT in &RI. Larger values (try 10) should 
>> decrease the block sizes. I do not really know how it works but it does 
>> have an effect on memory usage.
>> - Increase EPS_FILTER in &RI to a larger value (double-check later 
>> whether you have to decrease it again).
>> If nothing of my suggestions work and the developer of the feature does 
>> not chime in, it may be that you do not have enough memory.
>> Best,
>> Frederick
>>
>>
>> Frederick Stein schrieb am Sonntag, 15. September 2024 um 18:42:31 UTC+2:
>>
>> Hi,
>> I am not an expert with RI-HFX but I would try to run potentially cheaper 
>> settings first (single k-point, smaller basis sets, larger value of omega, 
>> identity metric). It may also help to have an CP2K output file or 
>> supplementary input/output file if you submit jobs with a scheduler like 
>> Slurm. If there is no function stack in the output file(s) given (either by 
>> the compiler or by CP2K), you could also add the keywords TRACE and 
>> TRACE_MASTER to your GLOBAL section.
>> Best,
>> Frederick
>>
>> yj jiang schrieb am Sonntag, 15. September 2024 um 18:13:42 UTC+2:
>>
>> Hi, I'm calculating band structure for C3N4 with a Ni atom (attachment). 
>> I want to use HES06 + RI-HFX + ADMM. There is a "out of memory" error. How 
>> can I address this? Thx. a lot.
>>
>> I can access a hpc with max memory of 512G.
>>
>>

-- 
You received this message because you are subscribed to the Google Groups "cp2k" group.
To unsubscribe from this group and stop receiving emails from it, send an email to cp2k+unsubscribe at googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/cp2k/1450b32b-3522-4e79-a93f-17b236e093d4n%40googlegroups.com.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.cp2k.org/archives/cp2k-user/attachments/20240922/05c48529/attachment.htm>


More information about the CP2K-user mailing list