Update README.md
Browse files
README.md
CHANGED
|
@@ -30,7 +30,7 @@ pip install causal-conv1d>=1.2.0
|
|
| 30 |
pip install mamba-ssm
|
| 31 |
```
|
| 32 |
|
| 33 |
-
If any of these two is not installed, the "eager" implementation will be used(not recommended). Otherwise the more optimised `
|
| 34 |
|
| 35 |
### Generation
|
| 36 |
|
|
@@ -76,7 +76,7 @@ print(tokenizer.batch_decode(out))
|
|
| 76 |
|
| 77 |
### On-device Inference
|
| 78 |
|
| 79 |
-
Since Mambaoutai is only 1.6B parameters, it can run on a CPU
|
| 80 |
|
| 81 |
Here is an example of how to run it on llama.cpp:
|
| 82 |
|
|
|
|
| 30 |
pip install mamba-ssm
|
| 31 |
```
|
| 32 |
|
| 33 |
+
If any of these two is not installed, the "eager" implementation will be used(not recommended). Otherwise the more optimised `CUDA` kernels will be used.
|
| 34 |
|
| 35 |
### Generation
|
| 36 |
|
|
|
|
| 76 |
|
| 77 |
### On-device Inference
|
| 78 |
|
| 79 |
+
Since Mambaoutai is only 1.6B parameters, it can be run on a CPU with reasonable speed.
|
| 80 |
|
| 81 |
Here is an example of how to run it on llama.cpp:
|
| 82 |
|