Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add support for GLM/chatGLM models #20

Open
MarkSchmidty opened this issue Mar 22, 2023 · 1 comment
Open

Add support for GLM/chatGLM models #20

MarkSchmidty opened this issue Mar 22, 2023 · 1 comment

Comments

@MarkSchmidty
Copy link

chatGLM-6B is an open source model based on GLM fine tuned on over 1 trillion tokens of dialogue and RLHF for chat.

It's quickly becoming one of the most popular local models despite no good fast CPU inference support (yet).

Official Repo:
https://github.com/THUDM/ChatGLM-6B/blob/main/README_en.md

@Ayushk4
Copy link
Member

Ayushk4 commented Mar 24, 2023

Are you aware of any differences between GLM's architecture and GPT-NeoX? If not then all we need to do is quantize it.

Also, it's LICENSE seems to have the similar restrictions to LLaMas. Any ideas on what format its int4 quantized version in?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants