Spaces:
Running
Running
Geevarghese George
commited on
Commit
·
ebe4e2c
1
Parent(s):
8338f39
add HF_INFERENCE_PROVIDER
Browse files- app.py +2 -0
- src/upgrade_advisor/config.py +2 -2
app.py
CHANGED
|
@@ -44,9 +44,11 @@ _monkeypatch_gradio_save_history()
|
|
| 44 |
|
| 45 |
def get_agent_model(model_name: str, oauth_token: gr.OAuthToken = None):
|
| 46 |
token = os.getenv("HF_TOKEN", None) or oauth_token.token if oauth_token else None
|
|
|
|
| 47 |
model = InferenceClientModel(
|
| 48 |
token=token,
|
| 49 |
model_id=model_name,
|
|
|
|
| 50 |
)
|
| 51 |
return model
|
| 52 |
|
|
|
|
| 44 |
|
| 45 |
def get_agent_model(model_name: str, oauth_token: gr.OAuthToken = None):
|
| 46 |
token = os.getenv("HF_TOKEN", None) or oauth_token.token if oauth_token else None
|
| 47 |
+
provider = os.getenv("HF_INFERENCE_PROVIDER", "together")
|
| 48 |
model = InferenceClientModel(
|
| 49 |
token=token,
|
| 50 |
model_id=model_name,
|
| 51 |
+
provider=provider,
|
| 52 |
)
|
| 53 |
return model
|
| 54 |
|
src/upgrade_advisor/config.py
CHANGED
|
@@ -5,8 +5,8 @@ from dotenv import load_dotenv
|
|
| 5 |
load_dotenv()
|
| 6 |
|
| 7 |
HF_INFERENCE_PROVIDER= os.getenv("HF_INFERENCE_PROVIDER", "together")
|
| 8 |
-
AGENT_MODEL = os.getenv("AGENT_MODEL", f"Qwen/Qwen3-Next-80B-A3B-Thinking
|
| 9 |
-
CHAT_MODEL = os.getenv("CHAT_MODEL", f"Qwen/Qwen3-Next-80B-A3B-Thinking
|
| 10 |
|
| 11 |
# GitHub MCP configuration
|
| 12 |
GITHUB_PAT = os.getenv("GITHUB_PAT", None)
|
|
|
|
| 5 |
load_dotenv()
|
| 6 |
|
| 7 |
HF_INFERENCE_PROVIDER= os.getenv("HF_INFERENCE_PROVIDER", "together")
|
| 8 |
+
AGENT_MODEL = os.getenv("AGENT_MODEL", f"Qwen/Qwen3-Next-80B-A3B-Thinking")
|
| 9 |
+
CHAT_MODEL = os.getenv("CHAT_MODEL", f"Qwen/Qwen3-Next-80B-A3B-Thinking")
|
| 10 |
|
| 11 |
# GitHub MCP configuration
|
| 12 |
GITHUB_PAT = os.getenv("GITHUB_PAT", None)
|