-
Notifications
You must be signed in to change notification settings - Fork 34
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Retraining models #11
Comments
If I understand you correctly, you want to take a model as input, and then initialize training on new data with that model. Is that what you want to do? |
Yes. For example, take http://smartschat.de/downloads/model-pair-train.obj and continue training on it using some new data. |
That's not implemented, but the code can be adapted. Unfortunately, I will not be able to have a closer look at this during this week. If you want to do it by yourself, I can give you some pointers. |
Please do! I'll create a pull request if I am able to successfully implement this. |
The constructor of perceptrons.pyx has You need to make this overwriting optional, for example by adding a boolean parameter which controls whether weights/priors should be initialized or not. Then you also need to adapt the experiments.py-API training/predicting scripts with the new parameters. I hope this helps. if you have any questions, I'm happy to answer them! |
Hi Sebastian, Can you give approximate information on the amount of time it took you to train CORT on CONLL data? Thanks and regards, |
Hi Rakesh, training the ranking model takes around two minutes per epoch. Preprocessing takes ~ 20 minutes if I remember correctly. Due to high memory requirements I train the models on a server with > 100GB RAM, using ~20 2.3GHz CPUs. However, only preprocessing is parallelized. |
Is it possible to retrain models (for example, the one's from https://github.com/smartschat/cort/blob/master/COREFERENCE.md#model-downloads) with new data?
I tried training using-
but I think it overwrites the model.
The text was updated successfully, but these errors were encountered: