Pseudopotential mismatch between pretrained data and fine-tuning data #1277
Replies: 3 comments 12 replies
-
|
Dear @dominicvarghese, There is no problem with fine-tuning to a different level of theory. The change in level of theories is handled by two things:
If you want, you can share with us your log file and we might be able to help you further. |
Beta Was this translation helpful? Give feedback.
-
|
Just to add to Ilyes' comments. MACE fits the atomisation energy, this is why we need the E0s. This also allows us to cope with different pseudo potentials and functionals better, because the atomisation energy is not as different as the total energy when you change pseudopotentials or functionals. your errors look good, especially on the forces. I think your energy errors could perhaps be lowered, but I suspect you may be limited there by your k-point sampling in your original DFT data! "standard" settings actually often lead to a few meV/atom error (which many DFT studies don't actually care about because they benefit from the cancellation of errors which occurs when you use the same unit cell, same k-grid when computing energy differences). So you could try to up your k-point density. |
Beta Was this translation helpful? Give feedback.
-
|
Thanks so much for the clear and helpful replies, @ilyes319 and @gabor1! @ilyes319 I am attaching the log file here as requested: log.txt @gabor1 In my command, I set This is the full command I used for the fine-tuning. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi everyone,
I have a question about the best practices for multi-head fine-tuning. I am currently trying to fine-tune a foundation model on my own VASP dataset and am seeing good preliminary results, but I have a doubt about the underlying methodology.
Here is my setup:
Foundation Model: I am using MATPES-R2SCAN as my foundation model, which is based on the R2SCAN functional.
Replay Dataset: I am using the standard matpes-r2scan-replay-data.extxyz as my pt_train_file.
My Finetuning Dataset: I have a small, specific dataset for BaO that I generated myself using VASP with the PBEsol functional.
My main concern is the mismatch in the DFT functionals. The foundation model and its replay data are based on the R2SCAN potential energy surface (PES), while my new fine-tuning data is based on the PBEsol PES.
I have a few specific questions about this:
How does this mismatch in functionals affect the fine-tuning process? Will the model's shared weights get "confused" or learn an incorrect average of the two different potential energy surfaces?
I have calculated my own e0s.json file using PBEsol for my train_file. The replay data will use the original R2SCAN E0s. Is the multi-head model designed to handle two different sets of E0s (one for each head)?
For the best results, is it a firm requirement to generate the fine-tuning dataset using the exact same functional (R2SCAN) as the foundation model?
My validation errors on the fine-tuning head look very good, but I want to be sure I am following the procedure for these simulations.
From the current implementation my errors as follows:
Any advice or insights on this would be greatly appreciated!
Thanks
Dominic
Beta Was this translation helpful? Give feedback.
All reactions