Update README.md
Browse files
README.md
CHANGED
|
@@ -17,6 +17,24 @@ tags:
|
|
| 17 |
This model was converted to GGUF format from [`Gryphe/Pantheon-Proto-RP-1.8-30B-A3B`](https://huggingface.co/Gryphe/Pantheon-Proto-RP-1.8-30B-A3B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
| 18 |
Refer to the [original model card](https://huggingface.co/Gryphe/Pantheon-Proto-RP-1.8-30B-A3B) for more details on the model.
|
| 19 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 20 |
## Use with llama.cpp
|
| 21 |
Install llama.cpp through brew (works on Mac and Linux)
|
| 22 |
|
|
|
|
| 17 |
This model was converted to GGUF format from [`Gryphe/Pantheon-Proto-RP-1.8-30B-A3B`](https://huggingface.co/Gryphe/Pantheon-Proto-RP-1.8-30B-A3B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
| 18 |
Refer to the [original model card](https://huggingface.co/Gryphe/Pantheon-Proto-RP-1.8-30B-A3B) for more details on the model.
|
| 19 |
|
| 20 |
+
---
|
| 21 |
+
Ever since Qwen 3 released I've been trying to get MoE finetuning to
|
| 22 |
+
work - After countless frustrating days, much code hacking, etc etc I
|
| 23 |
+
finally got a full finetune to complete with reasonable loss values.
|
| 24 |
+
|
| 25 |
+
|
| 26 |
+
I picked the base model for this since I didn't feel like trying to
|
| 27 |
+
fight a reasoning model's training - Maybe someday I'll make a model
|
| 28 |
+
which uses thinking tags for the character's thoughts or something.
|
| 29 |
+
|
| 30 |
+
|
| 31 |
+
This time the recipe focused on combining as many data sources as I
|
| 32 |
+
possibly could, featuring synthetic data from Sonnet 3.5 + 3.7, ChatGPT
|
| 33 |
+
4o and Deepseek. These then went through an extensive rewriting pipeline
|
| 34 |
+
to eliminate common AI cliches, with the hopeful intent of providing
|
| 35 |
+
you a fresh experience.
|
| 36 |
+
|
| 37 |
+
---
|
| 38 |
## Use with llama.cpp
|
| 39 |
Install llama.cpp through brew (works on Mac and Linux)
|
| 40 |
|