Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -2193,14 +2193,14 @@ T4: final_answer_tool("3")
|
|
| 2193 |
# Initialize LLMs
|
| 2194 |
print("Initializing LLMs...")
|
| 2195 |
|
| 2196 |
-
# Primary: Gemini 2.
|
| 2197 |
self.gemini_pro_llm = ChatGoogleGenerativeAI(
|
| 2198 |
-
model="gemini-2.
|
| 2199 |
google_api_key=GOOGLE_API_KEY,
|
| 2200 |
temperature=0,
|
| 2201 |
max_tokens=1024
|
| 2202 |
).bind_tools(self.tools, tool_choice="auto")
|
| 2203 |
-
print("✅ Gemini 2.
|
| 2204 |
|
| 2205 |
# Fallback 1: Gemini 2.5 Flash (5 RPM, 250K TPM, 20 RPD)
|
| 2206 |
self.gemini_flash_llm = ChatGoogleGenerativeAI(
|
|
@@ -2264,7 +2264,7 @@ T4: final_answer_tool("3")
|
|
| 2264 |
self.claude_llm = None
|
| 2265 |
print("ℹ️ Claude fallback unavailable (no ANTHROPIC_API_KEY)")
|
| 2266 |
|
| 2267 |
-
chain = "Gemini 2.
|
| 2268 |
if self.groq_llm:
|
| 2269 |
chain += " → Groq llama-3.3-70b → Groq qwen3-32b"
|
| 2270 |
if self.claude_llm:
|
|
|
|
| 2193 |
# Initialize LLMs
|
| 2194 |
print("Initializing LLMs...")
|
| 2195 |
|
| 2196 |
+
# Primary: Gemini 2.0 Flash (15 RPM, unlimited TPM, 1500 RPD)
|
| 2197 |
self.gemini_pro_llm = ChatGoogleGenerativeAI(
|
| 2198 |
+
model="gemini-2.0-flash",
|
| 2199 |
google_api_key=GOOGLE_API_KEY,
|
| 2200 |
temperature=0,
|
| 2201 |
max_tokens=1024
|
| 2202 |
).bind_tools(self.tools, tool_choice="auto")
|
| 2203 |
+
print("✅ Gemini 2.0 Flash primary initialized")
|
| 2204 |
|
| 2205 |
# Fallback 1: Gemini 2.5 Flash (5 RPM, 250K TPM, 20 RPD)
|
| 2206 |
self.gemini_flash_llm = ChatGoogleGenerativeAI(
|
|
|
|
| 2264 |
self.claude_llm = None
|
| 2265 |
print("ℹ️ Claude fallback unavailable (no ANTHROPIC_API_KEY)")
|
| 2266 |
|
| 2267 |
+
chain = "Gemini 2.0 Flash → Gemini 2.5 Flash"
|
| 2268 |
if self.groq_llm:
|
| 2269 |
chain += " → Groq llama-3.3-70b → Groq qwen3-32b"
|
| 2270 |
if self.claude_llm:
|