-
I'm quiet new to text generation and somewhat struggling with model conversions. I could convert the original Meta Llama models, however how does one convert the |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 4 replies
-
I am also struggling with that. I managed to run Llama-7b in 4 bit mode, but Alpaca-7b is not working. Would be glad to have some instructions. |
Beta Was this translation helpful? Give feedback.
-
There is no way to convert the 4bit ggml models without loss because they use a different method for quantization. You'll have to merge the Lora using the alpaca-lora repo then quantize it to 4bit using the GPTQ-for-llama repo. |
Beta Was this translation helpful? Give feedback.
There is no way to convert the 4bit ggml models without loss because they use a different method for quantization. You'll have to merge the Lora using the alpaca-lora repo then quantize it to 4bit using the GPTQ-for-llama repo.