Update README.md
Browse files
README.md
CHANGED
|
@@ -11,7 +11,7 @@ import torch
|
|
| 11 |
from transformers import BitsAndBytesConfig
|
| 12 |
|
| 13 |
MODEL_PATH = "THUDM/GLM-4.1V-9B-Thinking"
|
| 14 |
-
SAVE_PATH = "
|
| 15 |
|
| 16 |
quant_config = BitsAndBytesConfig(
|
| 17 |
load_in_4bit=True,
|
|
@@ -47,7 +47,7 @@ import io
|
|
| 47 |
|
| 48 |
app = FastAPI()
|
| 49 |
model = Glm4vForConditionalGeneration.from_pretrained(
|
| 50 |
-
"/path/to/the/model/
|
| 51 |
device_map="auto",
|
| 52 |
quantization_config={"load_in_4bit": True, "bnb_4bit_quant_type": "nf4", "bnb_4bit_compute_dtype": torch.bfloat16}
|
| 53 |
)
|
|
|
|
| 11 |
from transformers import BitsAndBytesConfig
|
| 12 |
|
| 13 |
MODEL_PATH = "THUDM/GLM-4.1V-9B-Thinking"
|
| 14 |
+
SAVE_PATH = "/define/the/path/to/save/model"
|
| 15 |
|
| 16 |
quant_config = BitsAndBytesConfig(
|
| 17 |
load_in_4bit=True,
|
|
|
|
| 47 |
|
| 48 |
app = FastAPI()
|
| 49 |
model = Glm4vForConditionalGeneration.from_pretrained(
|
| 50 |
+
"/path/to/the/model/safesensors",
|
| 51 |
device_map="auto",
|
| 52 |
quantization_config={"load_in_4bit": True, "bnb_4bit_quant_type": "nf4", "bnb_4bit_compute_dtype": torch.bfloat16}
|
| 53 |
)
|