Update README.md
Browse files
README.md
CHANGED
|
@@ -24,7 +24,7 @@ Gemma v2 is a large language model released by Google on Jun 27th 2024.
|
|
| 24 |
- Original model: [google/gemma-2-27b-it](https://huggingface.co/google/gemma-2-27b-it)
|
| 25 |
|
| 26 |
The model is packaged into executable weights, which we call
|
| 27 |
-
[llamafiles](https://github.com/Mozilla-Ocho/llamafile)
|
| 28 |
easy to use the model on Linux, MacOS, Windows, FreeBSD, OpenBSD, and
|
| 29 |
NetBSD for AMD64 and ARM64.
|
| 30 |
|
|
@@ -75,11 +75,9 @@ of the README.
|
|
| 75 |
|
| 76 |
When using the browser GUI, you need to fill out the following fields.
|
| 77 |
|
| 78 |
-
Prompt template:
|
| 79 |
|
| 80 |
```
|
| 81 |
-
<start_of_turn>system
|
| 82 |
-
{{prompt}}<end_of_turn>
|
| 83 |
{{history}}
|
| 84 |
<start_of_turn>{{char}}
|
| 85 |
```
|
|
@@ -100,6 +98,12 @@ The Belobog Academy has discovered a new, invasive species of algae that can dou
|
|
| 100 |
'
|
| 101 |
```
|
| 102 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 103 |
## About llamafile
|
| 104 |
|
| 105 |
llamafile is a new format introduced by Mozilla Ocho on Nov 20th 2023.
|
|
@@ -110,8 +114,10 @@ AMD64.
|
|
| 110 |
## About Quantization Formats
|
| 111 |
|
| 112 |
This model works should work well with any quantization format. Q6\_K is
|
| 113 |
-
the best choice overall
|
| 114 |
-
|
|
|
|
|
|
|
| 115 |
|
| 116 |
---
|
| 117 |
|
|
|
|
| 24 |
- Original model: [google/gemma-2-27b-it](https://huggingface.co/google/gemma-2-27b-it)
|
| 25 |
|
| 26 |
The model is packaged into executable weights, which we call
|
| 27 |
+
[llamafiles](https://github.com/Mozilla-Ocho/llamafile). This makes it
|
| 28 |
easy to use the model on Linux, MacOS, Windows, FreeBSD, OpenBSD, and
|
| 29 |
NetBSD for AMD64 and ARM64.
|
| 30 |
|
|
|
|
| 75 |
|
| 76 |
When using the browser GUI, you need to fill out the following fields.
|
| 77 |
|
| 78 |
+
Prompt template (note: this is for chat; Gemma doesn't have a system role):
|
| 79 |
|
| 80 |
```
|
|
|
|
|
|
|
| 81 |
{{history}}
|
| 82 |
<start_of_turn>{{char}}
|
| 83 |
```
|
|
|
|
| 98 |
'
|
| 99 |
```
|
| 100 |
|
| 101 |
+
## About Upload Limits
|
| 102 |
+
|
| 103 |
+
Files which exceed the Hugging Face 50GB upload limit have a .cat𝑋
|
| 104 |
+
extension. You need to use the `cat` command locally to turn them back
|
| 105 |
+
into a single file, using the same order.
|
| 106 |
+
|
| 107 |
## About llamafile
|
| 108 |
|
| 109 |
llamafile is a new format introduced by Mozilla Ocho on Nov 20th 2023.
|
|
|
|
| 114 |
## About Quantization Formats
|
| 115 |
|
| 116 |
This model works should work well with any quantization format. Q6\_K is
|
| 117 |
+
the best choice overall. We tested that it's able to produce identical
|
| 118 |
+
responses to the Gemma2 27B model that's hosted by Google themselves on
|
| 119 |
+
aistudio.google.com. If you encounter any divergences, then try using
|
| 120 |
+
the BF16 weights, which have the original fidelity.
|
| 121 |
|
| 122 |
---
|
| 123 |
|