Spaces:
Sleeping
Sleeping
Commit
·
9f8c7c1
1
Parent(s):
11f97a3
Use larger model as default model for non-space and non-openai solution
Browse files
README.md
CHANGED
|
@@ -27,7 +27,7 @@ Check out the configuration reference at https://huggingface.co/docs/hub/spaces-
|
|
| 27 |
python app.py
|
| 28 |
# Streamlit (TODO: able to clear message history)
|
| 29 |
streamlit run streamlit_app.py
|
| 30 |
-
# Streamlit chat_input (TODO: move ReAct process to `st.expander
|
| 31 |
streamlit run streamlit_app_chat.py
|
| 32 |
# Chainlit (TODO: move the ReAct process from agent message to "step")
|
| 33 |
chainlit run chainlit_app.py
|
|
@@ -38,6 +38,9 @@ chainlit run chainlit_app.py
|
|
| 38 |
- `What time is it?`
|
| 39 |
- Expect to use tool `get_current_time_in_timezone`
|
| 40 |
- Result expect the current local time in UTC
|
|
|
|
|
|
|
|
|
|
| 41 |
|
| 42 |
---
|
| 43 |
|
|
|
|
| 27 |
python app.py
|
| 28 |
# Streamlit (TODO: able to clear message history)
|
| 29 |
streamlit run streamlit_app.py
|
| 30 |
+
# Streamlit chat_input (TODO: move ReAct process to `st.expander`; able to render image)
|
| 31 |
streamlit run streamlit_app_chat.py
|
| 32 |
# Chainlit (TODO: move the ReAct process from agent message to "step")
|
| 33 |
chainlit run chainlit_app.py
|
|
|
|
| 38 |
- `What time is it?`
|
| 39 |
- Expect to use tool `get_current_time_in_timezone`
|
| 40 |
- Result expect the current local time in UTC
|
| 41 |
+
- `Draw a bear for me`
|
| 42 |
+
- Expect to use tool `image_generator`
|
| 43 |
+
- Result will be something like `{'path': '/var/folders/g3/gsbp4bsx1gs51t1805r34ztw0000gn/T/tmpch4mrr13/2fa94447-e9e5-4964-b9d2-b4b9a83ccf84.png', 'mime_type': 'image/png'}`
|
| 44 |
|
| 45 |
---
|
| 46 |
|
app.py
CHANGED
|
@@ -62,7 +62,7 @@ if IS_IN_HF_SPACE := os.getenv("SPACE_ID"):
|
|
| 62 |
print("Using HfApiModel with model_id:", model_id)
|
| 63 |
|
| 64 |
model = HfApiModel(
|
| 65 |
-
max_tokens=2096,
|
| 66 |
temperature=0.5,
|
| 67 |
model_id=model_id,
|
| 68 |
custom_role_conversions=None,
|
|
@@ -80,16 +80,29 @@ else:
|
|
| 80 |
|
| 81 |
model = OpenAIServerModel(model_id="gpt-3.5-turbo", api_key=OPENAI_API_KEY)
|
| 82 |
else:
|
| 83 |
-
# NOTE: this model is not good enough for agent
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 84 |
print(
|
| 85 |
-
"Using
|
| 86 |
)
|
| 87 |
|
| 88 |
-
|
| 89 |
-
|
|
|
|
|
|
|
|
|
|
| 90 |
)
|
| 91 |
|
| 92 |
-
|
| 93 |
# Import tool from Hub
|
| 94 |
image_generation_tool = load_tool("agents-course/text-to-image", trust_remote_code=True)
|
| 95 |
|
|
|
|
| 62 |
print("Using HfApiModel with model_id:", model_id)
|
| 63 |
|
| 64 |
model = HfApiModel(
|
| 65 |
+
max_tokens=2096 if IS_IN_HF_SPACE else None,
|
| 66 |
temperature=0.5,
|
| 67 |
model_id=model_id,
|
| 68 |
custom_role_conversions=None,
|
|
|
|
| 80 |
|
| 81 |
model = OpenAIServerModel(model_id="gpt-3.5-turbo", api_key=OPENAI_API_KEY)
|
| 82 |
else:
|
| 83 |
+
# NOTE: this model is not good enough for agent (and still might be too heavy to run on a normal computer)
|
| 84 |
+
# print(
|
| 85 |
+
# "Using TransformersModel with model_id: HuggingFaceTB/SmolLM2-1.7B-Instruct"
|
| 86 |
+
# )
|
| 87 |
+
#
|
| 88 |
+
# model = TransformersModel(
|
| 89 |
+
# model_id="HuggingFaceTB/SmolLM2-1.7B-Instruct", trust_remote_code=True
|
| 90 |
+
# )
|
| 91 |
+
|
| 92 |
+
# NOTE: this model is well enough to use simple tools (but need authentication)
|
| 93 |
+
# https://huggingface.co/meta-llama/Llama-3.2-11B-Vision-Instruct
|
| 94 |
+
# https://huggingface.co/settings/gated-repos
|
| 95 |
print(
|
| 96 |
+
"Using HfApiModel with model_id: meta-llama/Llama-3.2-11B-Vision-Instruct"
|
| 97 |
)
|
| 98 |
|
| 99 |
+
# NOTE: this use InferenceClient under the hood
|
| 100 |
+
model = HfApiModel(
|
| 101 |
+
temperature=0.5,
|
| 102 |
+
model_id="meta-llama/Llama-3.2-11B-Vision-Instruct",
|
| 103 |
+
custom_role_conversions=None,
|
| 104 |
)
|
| 105 |
|
|
|
|
| 106 |
# Import tool from Hub
|
| 107 |
image_generation_tool = load_tool("agents-course/text-to-image", trust_remote_code=True)
|
| 108 |
|