From 4ba09c9fdb069d565b603d8293ed92189ed4e686 Mon Sep 17 00:00:00 2001 From: Ren Xuancheng Date: Mon, 26 Aug 2024 15:39:12 +0800 Subject: [PATCH] Update README.md --- README.md | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/README.md b/README.md index 10e729d3..86cf5d81 100644 --- a/README.md +++ b/README.md @@ -156,9 +156,11 @@ For additional details, please visit [ollama.ai](https://ollama.ai/). Download our provided GGUF files or create them by yourself, and you can directly use them with the latest [`llama.cpp`](https://github.com/ggerganov/llama.cpp) with a one-line command: ```shell -./llama-cli -m -n 512 -co -i -if -f prompts/chat-with-qwen.txt --in-prefix "<|im_start|>user\n" --in-suffix "<|im_end|>\n<|im_start|>assistant\n" +./llama-cli -m -n 512 -co -sp -cnv -f prompts/chat-with-qwen.txt ``` +For additional guides, please refer to [our documentation](https://qwen.readthedocs.io/en/latest/run_locally/llama.cpp.html). + #### MLX-LM If you are running on Apple Silicon, we have also provided checkpoints compatible with [`mlx-lm`](https://github.com/ml-explore/mlx-examples/blob/main/llms/README.md). Look for models ending with MLX on HuggingFace Hub, like [Qwen2-7B-Instruct-MLX](https://huggingface.co/Qwen/Qwen2-7B-Instruct-MLX).