-
Notifications
You must be signed in to change notification settings - Fork 56
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The ckpt of Quantized OPT model is not be found #53
Comments
We have released the ckpt of OPT. You can find them at https://huggingface.co/ChenMnZ/OmniQuant/tree/main. |
The results by loading the ckpt is not consistent with the reported result in the paper. |
Can you reproduce the results of other models except OPT-30B? |
@linloong Can you provide the training script? |
Sure, CUDA_VISIBLE_DEVICES=0 python main.py \
--model facebook/opt-13b \
--epochs 20 --output_dir ./log/opt-13b-w4a4 \
--wbits 4 --abits 4 --lwc --let --alpha 0.75 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
We have tried to reproduce the quantization of OPT-30B, but it is still difficult for us. Would you mind releasing the details of the procedure?
The text was updated successfully, but these errors were encountered: