Update README.md
Browse files
README.md
CHANGED
|
@@ -210,13 +210,13 @@ print(scores.tolist())
|
|
| 210 |
You can either run / deploy TEI on NVIDIA GPUs as:
|
| 211 |
|
| 212 |
```bash
|
| 213 |
-
docker run -p 8080:80 -v hf_cache:/data --pull always ghcr.io/huggingface/text-embeddings-inference:cpu-1.7.2 --model-id Qwen/Qwen3-Embedding-0.6B
|
| 214 |
```
|
| 215 |
|
| 216 |
Or on CPU devices as:
|
| 217 |
|
| 218 |
```bash
|
| 219 |
-
docker run
|
| 220 |
```
|
| 221 |
|
| 222 |
And then, generate the embeddings sending a HTTP POST request as:
|
|
|
|
| 210 |
You can either run / deploy TEI on NVIDIA GPUs as:
|
| 211 |
|
| 212 |
```bash
|
| 213 |
+
docker run --gpus all -p 8080:80 -v hf_cache:/data --pull always ghcr.io/huggingface/text-embeddings-inference:cpu-1.7.2 --model-id Qwen/Qwen3-Embedding-0.6B --dtype float16
|
| 214 |
```
|
| 215 |
|
| 216 |
Or on CPU devices as:
|
| 217 |
|
| 218 |
```bash
|
| 219 |
+
docker run -p 8080:80 -v hf_cache:/data --pull always ghcr.io/huggingface/text-embeddings-inference:1.7.2 --model-id Qwen/Qwen3-Embedding-0.6B
|
| 220 |
```
|
| 221 |
|
| 222 |
And then, generate the embeddings sending a HTTP POST request as:
|