Keeby-smilyai commited on
Commit
43abbaa
Β·
verified Β·
1 Parent(s): 0f9fe07

Create app.py

Browse files
Files changed (1) hide show
  1. app.py +31 -0
app.py ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ import onnxruntime as ort
3
+ import numpy as np
4
+ from transformers import AutoTokenizer
5
+
6
+ # Load tokenizer (Qwen tokenizer)
7
+ tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen1.5-0.5B")
8
+
9
+ # Load ONNX model
10
+ session = ort.InferenceSession("/models/model.onnx")
11
+
12
+ def chat_fn(user_input, history):
13
+ if not user_input.strip():
14
+ return history
15
+
16
+ # Tokenize input
17
+ inputs = tokenizer(user_input, return_tensors="np", padding=True)
18
+ ort_inputs = {session.get_inputs()[0].name: inputs["input_ids"].astype(np.int64)}
19
+
20
+ # Run inference
21
+ output = session.run(None, ort_inputs)[0]
22
+
23
+ # Decode model output
24
+ text = tokenizer.decode(output[0], skip_special_tokens=True)
25
+ history.append(("πŸ§‘β€πŸ’» You: " + user_input, "πŸ€– Sam (Qwen3): " + text))
26
+ return history
27
+
28
+ demo = gr.ChatInterface(fn=chat_fn, title="πŸ’¬ Qwen3-0.6B-ONNX Demo",
29
+ description="Running ONNX model on a prebuilt Docker Space (SmilyAI Style!)")
30
+
31
+ demo.launch(server_name="0.0.0.0", server_port=7860)