Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fine-Tuning process #9

Open
ghost opened this issue Jun 8, 2020 · 3 comments
Open

Fine-Tuning process #9

ghost opened this issue Jun 8, 2020 · 3 comments

Comments

@ghost
Copy link

ghost commented Jun 8, 2020

Hi!
I would like to know the process of fine-tuning UniLM with inverted SQUAD (hardware, training time, number of steps, parameters, etc.)
Would that be possible?
Thanks in advance!

@thusithaC
Copy link

Yes, I have the same question. The repo is extremely useful and provides good quality results and easy to use and setup compared to some purely research based github repos.

However, this might be a naive question, but does this repo even include the code needed to train the .bin file. Would love to recreate this in other languages, so it would be extremely helpful if a re-training guide can be included in the readme, with links to the source datasets.

@artitw
Copy link
Owner

artitw commented Aug 22, 2020

@ugmSorcero please see fine-tuning parameters below

--max_seq_length 512 \
--max_position_embeddings 512 \
--mask_prob 0.7 \
--max_pred 48 \
--train_batch_size 32 \
--gradient_accumulation_steps 2 \
--learning_rate 0.00002 \
--warmup_proportion 0.1 \
--label_smoothing 0.1 \
--num_train_epochs 10

@artitw
Copy link
Owner

artitw commented Aug 22, 2020

@thusithaC I will have to play around and think about how to best incorporate training the model from scratch and get back on this. If you have any ideas about that, feel free to let us know.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants