Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,6 @@
|
|
1 |
---
|
2 |
-
base_model:
|
|
|
3 |
language:
|
4 |
- en
|
5 |
license: apache-2.0
|
@@ -12,6 +13,11 @@ tags:
|
|
12 |
- grpo
|
13 |
- llama-cpp
|
14 |
- gguf-my-repo
|
|
|
|
|
|
|
|
|
|
|
15 |
---
|
16 |
|
17 |
# ykarout/Phi4-DeepSeek-16bit-Q4_K_M-GGUF
|
@@ -56,4 +62,4 @@ Step 3: Run inference through the main binary.
|
|
56 |
or
|
57 |
```
|
58 |
./llama-server --hf-repo ykarout/Phi4-DeepSeek-16bit-Q4_K_M-GGUF --hf-file phi4-deepseek-16bit-q4_k_m.gguf -c 2048
|
59 |
-
```
|
|
|
1 |
---
|
2 |
+
base_model:
|
3 |
+
- unsloth/phi-4
|
4 |
language:
|
5 |
- en
|
6 |
license: apache-2.0
|
|
|
13 |
- grpo
|
14 |
- llama-cpp
|
15 |
- gguf-my-repo
|
16 |
+
datasets:
|
17 |
+
- openai/gsm8k
|
18 |
+
- nvidia/Llama-Nemotron-Post-Training-Dataset-v1
|
19 |
+
pipeline_tag: text2text-generation
|
20 |
+
library_name: transformers
|
21 |
---
|
22 |
|
23 |
# ykarout/Phi4-DeepSeek-16bit-Q4_K_M-GGUF
|
|
|
62 |
or
|
63 |
```
|
64 |
./llama-server --hf-repo ykarout/Phi4-DeepSeek-16bit-Q4_K_M-GGUF --hf-file phi4-deepseek-16bit-q4_k_m.gguf -c 2048
|
65 |
+
```
|