Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix maximum seqlen for gptq quantization #1748

Merged
merged 1 commit into from
Mar 18, 2024

Conversation

SunMarc
Copy link
Member

@SunMarc SunMarc commented Mar 7, 2024

What does this PR do

This PR set a maximum seqlen that we use to create our calibration dataset. For models like mistral, we have a seqlen of 32768 which results in issues when creating the calibration dataset. The dataset don't need to have data with a big seqlen too.
Fixes huggingface/transformers#29494

@SunMarc SunMarc requested a review from fxmarty March 7, 2024 20:39
@fxmarty fxmarty merged commit 9ff5ea8 into huggingface:main Mar 18, 2024
37 of 45 checks passed
young-developer pushed a commit to young-developer/optimum that referenced this pull request May 10, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

AutoGPTQ quantization stucks without any progress
2 participants