You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Our finetuning focus has been a bit lacking lately. Let's change that, starting with LoRA!
Run a finetuning baseline with a Pythia model that saturates VRAM, measure its TFLOPs/VRAM usage in wandb and link it here. Choose a small finetuning dataset for this that requires little compute to converge without being trivial.
Add prototype LoRA support to gpt-neox
Compare to baseline in step 1 and ensure TFLOP/VRAM changes make sense
Compare to baseline in step 1 and ensure loss is maintained
Thanks for creating the feature request! As discussed - I'm happy to take this one on :) I expect to start work on it late this week / early next week.
Our finetuning focus has been a bit lacking lately. Let's change that, starting with LoRA!
As discussed in Discord @mkerin
The text was updated successfully, but these errors were encountered: