Geevarghese George commited on
Commit
ebe4e2c
·
1 Parent(s): 8338f39

add HF_INFERENCE_PROVIDER

Browse files
Files changed (2) hide show
  1. app.py +2 -0
  2. src/upgrade_advisor/config.py +2 -2
app.py CHANGED
@@ -44,9 +44,11 @@ _monkeypatch_gradio_save_history()
44
 
45
  def get_agent_model(model_name: str, oauth_token: gr.OAuthToken = None):
46
  token = os.getenv("HF_TOKEN", None) or oauth_token.token if oauth_token else None
 
47
  model = InferenceClientModel(
48
  token=token,
49
  model_id=model_name,
 
50
  )
51
  return model
52
 
 
44
 
45
  def get_agent_model(model_name: str, oauth_token: gr.OAuthToken = None):
46
  token = os.getenv("HF_TOKEN", None) or oauth_token.token if oauth_token else None
47
+ provider = os.getenv("HF_INFERENCE_PROVIDER", "together")
48
  model = InferenceClientModel(
49
  token=token,
50
  model_id=model_name,
51
+ provider=provider,
52
  )
53
  return model
54
 
src/upgrade_advisor/config.py CHANGED
@@ -5,8 +5,8 @@ from dotenv import load_dotenv
5
  load_dotenv()
6
 
7
  HF_INFERENCE_PROVIDER= os.getenv("HF_INFERENCE_PROVIDER", "together")
8
- AGENT_MODEL = os.getenv("AGENT_MODEL", f"Qwen/Qwen3-Next-80B-A3B-Thinking:{HF_INFERENCE_PROVIDER}")
9
- CHAT_MODEL = os.getenv("CHAT_MODEL", f"Qwen/Qwen3-Next-80B-A3B-Thinking:{HF_INFERENCE_PROVIDER}")
10
 
11
  # GitHub MCP configuration
12
  GITHUB_PAT = os.getenv("GITHUB_PAT", None)
 
5
  load_dotenv()
6
 
7
  HF_INFERENCE_PROVIDER= os.getenv("HF_INFERENCE_PROVIDER", "together")
8
+ AGENT_MODEL = os.getenv("AGENT_MODEL", f"Qwen/Qwen3-Next-80B-A3B-Thinking")
9
+ CHAT_MODEL = os.getenv("CHAT_MODEL", f"Qwen/Qwen3-Next-80B-A3B-Thinking")
10
 
11
  # GitHub MCP configuration
12
  GITHUB_PAT = os.getenv("GITHUB_PAT", None)