Update llama.cpp.md instructions (#3702)

This commit is contained in:
q5sys (JT) 2023-08-29 16:56:50 -04:00 committed by GitHub
parent a9a1784420
commit cdb854db9e
WARNING! Although there is a key with this ID in the database it does not verify this commit! This commit is SUSPICIOUS.
GPG key ID: 4AEE18F83AFDEB23

View file

@ -9,7 +9,10 @@ llama.cpp is the best backend in two important scenarios:
#### Pre-converted
Download the ggml model directly into your `text-generation-webui/models` folder, making sure that its name contains `ggml` somewhere and ends in `.bin`. It's a single file.
Download the GGUF or GGML models directly into your `text-generation-webui/models` folder. It will be a single file.
* For GGUF models, make sure its name contains `.gguf`.
* For GGML models, make sure its name contains `ggml` and ends in `.bin`.
`q4_K_M` quantization is recommended.