Fine-Tuning "2023-10-29-mace-16M-pbenner-mptrj-no-conditional-loss.model" with a Subset of Elements #254
-
Hello, I am currently working on a project where I need to fine-tune the "2023-10-29-mace-16M-pbenner-mptrj-no-conditional-loss.model" with my own dataset. My dataset is a subset of the original dataset used for this model, containing only a specific group of elements out of the 89 included in the original dataset. Right now, I am not able to read the previously trained model and continue training with my data. How can this be done? I am seeking advice or guidance on the best practices to approach this fine-tuning process. I have the following questions: Data Preparation: Are there any specific preprocessing steps recommended for a dataset that only includes a subset of the original model's elements? Model Adjustments: Given that my dataset includes fewer elements than the original model was trained on, are there any necessary modifications or considerations I should take into account before starting the fine-tuning? Training Process: Could you provide any tips or recommendations on the fine-tuning procedure itself, such as learning rate adjustments, batch sizes, number of epochs, etc.? Potential Challenges: Are there any common challenges or pitfalls I should be aware of when fine-tuning a model on a dataset that is different in scope from the original training data? Any insights, resources, or examples you could provide would be immensely helpful. Thank you in advance for your time and assistance. Best regards, |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 4 replies
-
If you go to the foundations branch in MACE, I made a way to fine tune a foundation model to your own dataset.
The model will select the species that are contained in your dataset. |
Beta Was this translation helpful? Give feedback.
-
Hello, sorry for the unrelated reply, but do you have the development branch or fine-tuning branch? If yes, could you share me how to get and compile that version? |
Beta Was this translation helpful? Give feedback.
If you go to the foundations branch in MACE, I made a way to fine tune a foundation model to your own dataset.
There is a new arg parser input called
--foundation_model
where you can put the path of the foundation model you want to fine tune. If you just do--foundation_model="use_mp"
you will fine tune the material project model (no need to download anything). For now you will need to use the same settings as the foundation model for your model (probably there are ways to go beyond that). Here are the setting for the mp model. There might checkpointing problems so tell me.