Replies: 2 comments 1 reply
-
Hi, LMDeploy does support Llama 2 both in ckpt and huggingface transformers model format. You can try the following command to convert it to LMDeploy required format python lmdeploy/serve/turbomind/deploy.py llama2 /the/path/of/original/llama2/model --model-format llama --tokenizer-path /the/path/of/tokenizer/model/of/original/llama2 |
Beta Was this translation helpful? Give feedback.
1 reply
-
Hi, @realhaik |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I am reading this article and unfortunately I see that the article uses hugging face llama2 model. Does this mean that the original meta llama2 is not supported?
https://openmmlab.medium.com/deploy-llama-2-models-easily-with-lmdeploy-1cb001d70290
This is a real deal breaker, because the hf model is defective. The results that hf model produces are a joke, completely broken. The original meta model works quite nice.
I feel really sorry for anyone who is wasting his time with the hf model.
Beta Was this translation helpful? Give feedback.
All reactions