mobinln commited on
Commit
c9f46e7
·
verified ·
1 Parent(s): 52dd340

Update start.sh

Browse files
Files changed (1) hide show
  1. start.sh +7 -4
start.sh CHANGED
@@ -3,7 +3,10 @@
3
  export WORK="/home/user/app"
4
  cd $WORK
5
  unzip llama_cpp_avx512.zip > /dev/null 2>&1
6
- echo "Booting up llama server..."
7
- wget -O model.gguf https://huggingface.co/lmstudio-community/Qwen3-4B-Instruct-2507-GGUF/resolve/main/Qwen3-4B-Instruct-2507-Q6_K.gguf?download=true > /dev/null 2>&1
8
- # wget -O model.gguf https://huggingface.co/unsloth/Qwen3-30B-A3B-Instruct-2507-GGUF/resolve/main/Qwen3-30B-A3B-Instruct-2507-UD-TQ1_0.gguf > /dev/null 2>&1
9
- ./llama-server -m model.gguf --port 8000 --host 0.0.0.0 --threads 2 --ctx-size 4096 --mlock --jinja
 
 
 
 
3
  export WORK="/home/user/app"
4
  cd $WORK
5
  unzip llama_cpp_avx512.zip > /dev/null 2>&1
6
+ echo "Booting up llama server..."
7
+
8
+ # wget -O model.gguf https://huggingface.co/lmstudio-community/Qwen3-4B-Instruct-2507-GGUF/resolve/main/Qwen3-4B-Instruct-2507-Q6_K.gguf?download=true > /dev/null 2>&1
9
+ wget -O model.gguf https://huggingface.co/unsloth/gpt-oss-20b-GGUF/resolve/main/gpt-oss-20b-F16.gguf?download=true > /dev/null 2>&1
10
+ ./llama-server -m model.gguf --port 8000 --host 0.0.0.0 --threads 2 --ctx-size 4096 --mlock --jinja
11
+
12
+ echo "✅ llama server running on port 8000"