File tree Expand file tree Collapse file tree 1 file changed +16
-0
lines changed
Expand file tree Collapse file tree 1 file changed +16
-0
lines changed Original file line number Diff line number Diff line change @@ -244,6 +244,13 @@ Download `FP16` quantized `Qwen3` .gguf files from:
244244- https://huggingface.co/ggml-org/Qwen3-4B-GGUF
245245- https://huggingface.co/ggml-org/Qwen3-8B-GGUF
246246
247+ Download ` FP16 ` quantized ` Qwen2.5 ` .gguf files from:
248+ - https://huggingface.co/bartowski/Qwen2.5-0.5B-Instruct-GGUF
249+ - https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct-GGUF
250+
251+ Download ` FP16 ` quantized ` DeepSeek-R1-Distill-Qwen ` .gguf files from:
252+ - https://huggingface.co/hdnh2006/DeepSeek-R1-Distill-Qwen-1.5B-GGUF
253+
247254Please be gentle with [ huggingface.co] ( https://huggingface.co ) servers:
248255
249256** Note** FP16 models are first-class citizens for the current version.
@@ -274,6 +281,15 @@ wget https://huggingface.co/ggml-org/Qwen3-0.6B-GGUF/resolve/main/Qwen3-8B-f16.g
274281
275282# Phi-3-mini-4k - FP16
276283wget https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-gguf/resolve/main/Phi-3-mini-4k-instruct-fp16.gguf
284+
285+ # Qwen2.5 (0.5B)
286+ wget https://huggingface.co/bartowski/Qwen2.5-0.5B-Instruct-GGUF/resolve/main/Qwen2.5-0.5B-Instruct-f16.gguf
287+
288+ # Qwen2.5 (1.5B)
289+ wget https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct-GGUF/resolve/main/qwen2.5-1.5b-instruct-fp16.gguf
290+
291+ # DeepSeek-R1-Distill-Qwen (1.5B)
292+ wget https://huggingface.co/hdnh2006/DeepSeek-R1-Distill-Qwen-1.5B-GGUF/resolve/main/DeepSeek-R1-Distill-Qwen-1.5B-F16.gguf
277293```
278294
279295** [ Experimental] ** you can download the Q8 and Q4 used in the original implementation of Llama3.java, but for now are going to be dequanted to FP16 for TornadoVM support:
You can’t perform that action at this time.
0 commit comments