Spaces:
Sleeping
feat: Add Gemini/Ollama toggle functionality with API error fixes
Browse files✨ New Features:
- Add interactive toggle switch in UI to choose between Ollama (Local) and Gemini (Cloud)
- Dynamic LLM selection with preference persistence in localStorage
- Support both Ollama and Gemini in research agent with automatic fallbacks
- Real-time status updates showing which AI model is being used
�� Backend Changes:
- Update research agent to support both LLMs with use_gemini parameter
- Add _research_with_gemini_tools method for Gemini-powered research
- Update query request model to include use_gemini field
- Automatic LLM initialization and availability detection
🎨 Frontend Changes:
- Beautiful toggle switch with animated slider and labels
- Persistent user preference storage
- Dynamic status messages based on selected LLM
- Responsive design for header controls
🛠️ Bug Fixes:
- Fix CoinGecko API 401 errors by adding proper fallback to mock data
- Improved error handling for chart_data_tool API failures
- Better session cleanup to prevent connection warnings
🔒 Safety & Performance:
- AI safety validation for both Ollama and Gemini responses
- Proper timeout handling for both LLM types
- Graceful degradation when APIs fail or models unavailable
- app.py +11 -8
- src/agent/research_agent.py +165 -21
- src/tools/chart_data_tool.py +41 -35
- src/utils/config.py +3 -3
- static/app.js +38 -1
- static/styles.css +103 -0
- templates/index.html +15 -3
|
@@ -37,6 +37,7 @@ templates = Jinja2Templates(directory="templates")
|
|
| 37 |
class QueryRequest(BaseModel):
|
| 38 |
query: str
|
| 39 |
chat_history: Optional[List[Dict[str, str]]] = []
|
|
|
|
| 40 |
|
| 41 |
class QueryResponse(BaseModel):
|
| 42 |
success: bool
|
|
@@ -85,7 +86,7 @@ class Web3CoPilotService:
|
|
| 85 |
self.airaa = None
|
| 86 |
self.viz = None
|
| 87 |
|
| 88 |
-
async def process_query(self, query: str) -> QueryResponse:
|
| 89 |
"""Process research query with comprehensive analysis"""
|
| 90 |
logger.info("Processing research request...")
|
| 91 |
|
|
@@ -113,7 +114,7 @@ Configure GEMINI_API_KEY environment variable for full AI analysis."""
|
|
| 113 |
logger.info("🤖 Processing with AI research agent...")
|
| 114 |
logger.info(f"🛠️ Available tools: {[tool.name for tool in self.agent.tools] if self.agent else []}")
|
| 115 |
|
| 116 |
-
result = await self.agent.research_query(query)
|
| 117 |
logger.info(f"🔄 Agent research completed: success={result.get('success')}")
|
| 118 |
|
| 119 |
if result.get("success"):
|
|
@@ -464,22 +465,23 @@ async def process_query_stream(request: QueryRequest):
|
|
| 464 |
yield f"data: {json.dumps({'type': 'status', 'message': 'Executing tools and gathering data...', 'progress': 50})}\n\n"
|
| 465 |
await asyncio.sleep(0.5)
|
| 466 |
|
| 467 |
-
# Send Ollama processing status with heartbeats
|
| 468 |
-
|
|
|
|
| 469 |
await asyncio.sleep(1.0)
|
| 470 |
|
| 471 |
# Send additional heartbeat messages during processing
|
| 472 |
-
yield f"data: {json.dumps({'type': 'status', 'message': '
|
| 473 |
await asyncio.sleep(2.0)
|
| 474 |
|
| 475 |
-
yield f"data: {json.dumps({'type': 'status', 'message': 'Still processing...
|
| 476 |
await asyncio.sleep(3.0)
|
| 477 |
|
| 478 |
# Process the actual query with timeout and periodic heartbeats
|
| 479 |
start_time = datetime.now()
|
| 480 |
|
| 481 |
# Create a task for the query processing
|
| 482 |
-
query_task = asyncio.create_task(service.process_query(request.query))
|
| 483 |
|
| 484 |
try:
|
| 485 |
# Send periodic heartbeats while waiting for Ollama
|
|
@@ -499,7 +501,8 @@ async def process_query_stream(request: QueryRequest):
|
|
| 499 |
raise asyncio.TimeoutError("Hard timeout reached")
|
| 500 |
|
| 501 |
progress = min(85 + (heartbeat_count * 2), 95) # Progress slowly from 85 to 95
|
| 502 |
-
|
|
|
|
| 503 |
|
| 504 |
# If we get here, the query completed successfully
|
| 505 |
result = query_task.result()
|
|
|
|
| 37 |
class QueryRequest(BaseModel):
|
| 38 |
query: str
|
| 39 |
chat_history: Optional[List[Dict[str, str]]] = []
|
| 40 |
+
use_gemini: bool = False
|
| 41 |
|
| 42 |
class QueryResponse(BaseModel):
|
| 43 |
success: bool
|
|
|
|
| 86 |
self.airaa = None
|
| 87 |
self.viz = None
|
| 88 |
|
| 89 |
+
async def process_query(self, query: str, use_gemini: bool = False) -> QueryResponse:
|
| 90 |
"""Process research query with comprehensive analysis"""
|
| 91 |
logger.info("Processing research request...")
|
| 92 |
|
|
|
|
| 114 |
logger.info("🤖 Processing with AI research agent...")
|
| 115 |
logger.info(f"🛠️ Available tools: {[tool.name for tool in self.agent.tools] if self.agent else []}")
|
| 116 |
|
| 117 |
+
result = await self.agent.research_query(query, use_gemini=use_gemini)
|
| 118 |
logger.info(f"🔄 Agent research completed: success={result.get('success')}")
|
| 119 |
|
| 120 |
if result.get("success"):
|
|
|
|
| 465 |
yield f"data: {json.dumps({'type': 'status', 'message': 'Executing tools and gathering data...', 'progress': 50})}\n\n"
|
| 466 |
await asyncio.sleep(0.5)
|
| 467 |
|
| 468 |
+
# Send Ollama/Gemini processing status with heartbeats
|
| 469 |
+
llm_name = "Gemini" if request.use_gemini else "Ollama"
|
| 470 |
+
yield f"data: {json.dumps({'type': 'status', 'message': f'{llm_name} is analyzing data and generating response...', 'progress': 70})}\n\n"
|
| 471 |
await asyncio.sleep(1.0)
|
| 472 |
|
| 473 |
# Send additional heartbeat messages during processing
|
| 474 |
+
yield f"data: {json.dumps({'type': 'status', 'message': f'{llm_name} is thinking deeply about your query...', 'progress': 75})}\n\n"
|
| 475 |
await asyncio.sleep(2.0)
|
| 476 |
|
| 477 |
+
yield f"data: {json.dumps({'type': 'status', 'message': f'Still processing... {llm_name} generates detailed responses', 'progress': 80})}\n\n"
|
| 478 |
await asyncio.sleep(3.0)
|
| 479 |
|
| 480 |
# Process the actual query with timeout and periodic heartbeats
|
| 481 |
start_time = datetime.now()
|
| 482 |
|
| 483 |
# Create a task for the query processing
|
| 484 |
+
query_task = asyncio.create_task(service.process_query(request.query, request.use_gemini))
|
| 485 |
|
| 486 |
try:
|
| 487 |
# Send periodic heartbeats while waiting for Ollama
|
|
|
|
| 501 |
raise asyncio.TimeoutError("Hard timeout reached")
|
| 502 |
|
| 503 |
progress = min(85 + (heartbeat_count * 2), 95) # Progress slowly from 85 to 95
|
| 504 |
+
llm_name = "Gemini" if request.use_gemini else "Ollama"
|
| 505 |
+
yield f"data: {json.dumps({'type': 'status', 'message': f'{llm_name} is still working... ({elapsed:.0f}s elapsed)', 'progress': progress})}\n\n"
|
| 506 |
|
| 507 |
# If we get here, the query completed successfully
|
| 508 |
result = query_task.result()
|
|
@@ -22,41 +22,60 @@ class Web3ResearchAgent:
|
|
| 22 |
self.fallback_llm = None
|
| 23 |
self.tools = []
|
| 24 |
self.enabled = False
|
|
|
|
| 25 |
|
| 26 |
try:
|
| 27 |
-
|
| 28 |
-
|
| 29 |
-
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
|
| 34 |
except Exception as e:
|
| 35 |
logger.error(f"Agent initialization failed: {e}")
|
| 36 |
self.enabled = False
|
| 37 |
|
| 38 |
-
def
|
| 39 |
-
"""Initialize
|
| 40 |
try:
|
| 41 |
self.fallback_llm = Ollama(
|
| 42 |
model=config.OLLAMA_MODEL,
|
| 43 |
base_url=config.OLLAMA_BASE_URL,
|
| 44 |
temperature=0.1
|
| 45 |
)
|
| 46 |
-
|
| 47 |
logger.info(f"✅ Ollama initialized - Model: {config.OLLAMA_MODEL}")
|
| 48 |
-
|
| 49 |
-
self.tools = self._initialize_tools()
|
| 50 |
-
self.enabled = True
|
| 51 |
-
|
| 52 |
except Exception as e:
|
| 53 |
logger.error(f"Ollama initialization failed: {e}")
|
| 54 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 55 |
|
| 56 |
def _init_with_gemini_fallback(self):
|
| 57 |
-
"""Initialize with Gemini primary and Ollama fallback"""
|
| 58 |
-
|
| 59 |
-
|
| 60 |
|
| 61 |
def _initialize_tools(self):
|
| 62 |
tools = []
|
|
@@ -93,8 +112,8 @@ class Web3ResearchAgent:
|
|
| 93 |
|
| 94 |
return tools
|
| 95 |
|
| 96 |
-
async def research_query(self, query: str) -> Dict[str, Any]:
|
| 97 |
-
"""Research query with
|
| 98 |
|
| 99 |
# AI Safety Check 1: Sanitize and validate input
|
| 100 |
sanitized_query, is_safe, safety_reason = ai_safety.sanitize_query(query)
|
|
@@ -140,8 +159,13 @@ class Web3ResearchAgent:
|
|
| 140 |
}
|
| 141 |
|
| 142 |
try:
|
| 143 |
-
|
| 144 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 145 |
|
| 146 |
except Exception as e:
|
| 147 |
logger.error(f"Research failed: {e}")
|
|
@@ -342,6 +366,126 @@ The system successfully gathered data from {len(suggested_tools)} tools:
|
|
| 342 |
logger.error(f"Ollama tools research failed: {e}")
|
| 343 |
raise e
|
| 344 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 345 |
def _extract_sources(self, response: str) -> List[str]:
|
| 346 |
"""Extract sources from response"""
|
| 347 |
# Simple source extraction - can be enhanced
|
|
|
|
| 22 |
self.fallback_llm = None
|
| 23 |
self.tools = []
|
| 24 |
self.enabled = False
|
| 25 |
+
self.gemini_available = False
|
| 26 |
|
| 27 |
try:
|
| 28 |
+
# Always initialize Ollama
|
| 29 |
+
logger.info("🔧 Initializing Ollama as fallback")
|
| 30 |
+
self._init_ollama()
|
| 31 |
+
|
| 32 |
+
# Try to initialize Gemini if API key is available
|
| 33 |
+
if config.GEMINI_API_KEY:
|
| 34 |
+
logger.info("🔧 Initializing Gemini as primary option")
|
| 35 |
+
self._init_gemini()
|
| 36 |
+
|
| 37 |
+
self.tools = self._initialize_tools()
|
| 38 |
+
self.enabled = True
|
| 39 |
|
| 40 |
except Exception as e:
|
| 41 |
logger.error(f"Agent initialization failed: {e}")
|
| 42 |
self.enabled = False
|
| 43 |
|
| 44 |
+
def _init_ollama(self):
|
| 45 |
+
"""Initialize Ollama LLM"""
|
| 46 |
try:
|
| 47 |
self.fallback_llm = Ollama(
|
| 48 |
model=config.OLLAMA_MODEL,
|
| 49 |
base_url=config.OLLAMA_BASE_URL,
|
| 50 |
temperature=0.1
|
| 51 |
)
|
|
|
|
| 52 |
logger.info(f"✅ Ollama initialized - Model: {config.OLLAMA_MODEL}")
|
|
|
|
|
|
|
|
|
|
|
|
|
| 53 |
except Exception as e:
|
| 54 |
logger.error(f"Ollama initialization failed: {e}")
|
| 55 |
+
raise
|
| 56 |
+
|
| 57 |
+
def _init_gemini(self):
|
| 58 |
+
"""Initialize Gemini LLM"""
|
| 59 |
+
try:
|
| 60 |
+
self.llm = ChatGoogleGenerativeAI(
|
| 61 |
+
model="gemini-pro",
|
| 62 |
+
google_api_key=config.GEMINI_API_KEY,
|
| 63 |
+
temperature=0.1
|
| 64 |
+
)
|
| 65 |
+
self.gemini_available = True
|
| 66 |
+
logger.info("✅ Gemini initialized")
|
| 67 |
+
except Exception as e:
|
| 68 |
+
logger.warning(f"Gemini initialization failed: {e}")
|
| 69 |
+
self.gemini_available = False
|
| 70 |
+
|
| 71 |
+
def _init_ollama_only(self):
|
| 72 |
+
"""Initialize with only Ollama LLM (deprecated - kept for compatibility)"""
|
| 73 |
+
self._init_ollama()
|
| 74 |
|
| 75 |
def _init_with_gemini_fallback(self):
|
| 76 |
+
"""Initialize with Gemini primary and Ollama fallback (deprecated - kept for compatibility)"""
|
| 77 |
+
self._init_ollama()
|
| 78 |
+
self._init_gemini()
|
| 79 |
|
| 80 |
def _initialize_tools(self):
|
| 81 |
tools = []
|
|
|
|
| 112 |
|
| 113 |
return tools
|
| 114 |
|
| 115 |
+
async def research_query(self, query: str, use_gemini: bool = False) -> Dict[str, Any]:
|
| 116 |
+
"""Research query with dynamic LLM selection - Enhanced with AI Safety"""
|
| 117 |
|
| 118 |
# AI Safety Check 1: Sanitize and validate input
|
| 119 |
sanitized_query, is_safe, safety_reason = ai_safety.sanitize_query(query)
|
|
|
|
| 159 |
}
|
| 160 |
|
| 161 |
try:
|
| 162 |
+
# Choose LLM based on user preference and availability
|
| 163 |
+
if use_gemini and self.gemini_available:
|
| 164 |
+
logger.info("🤖 Processing with Gemini + Tools (Safety Enhanced)")
|
| 165 |
+
return await self._research_with_gemini_tools(sanitized_query)
|
| 166 |
+
else:
|
| 167 |
+
logger.info("🤖 Processing with Ollama + Tools (Safety Enhanced)")
|
| 168 |
+
return await self._research_with_ollama_tools(sanitized_query)
|
| 169 |
|
| 170 |
except Exception as e:
|
| 171 |
logger.error(f"Research failed: {e}")
|
|
|
|
| 366 |
logger.error(f"Ollama tools research failed: {e}")
|
| 367 |
raise e
|
| 368 |
|
| 369 |
+
async def _research_with_gemini_tools(self, query: str) -> Dict[str, Any]:
|
| 370 |
+
"""Research using Gemini with tools"""
|
| 371 |
+
try:
|
| 372 |
+
# Step 1: Analyze query and suggest tools using Gemini
|
| 373 |
+
tool_analysis_prompt = f"""Based on this Web3/cryptocurrency research query, identify the most relevant tools to use.
|
| 374 |
+
|
| 375 |
+
Query: "{query}"
|
| 376 |
+
|
| 377 |
+
Available tools:
|
| 378 |
+
- cryptocompare_data: Get current cryptocurrency prices and basic info
|
| 379 |
+
- coingecko_data: Comprehensive market data and analytics
|
| 380 |
+
- defillama_data: DeFi protocols, TVL, and yield farming data
|
| 381 |
+
- etherscan_data: Ethereum blockchain data and transactions
|
| 382 |
+
- chart_data_provider: Generate chart data for visualizations
|
| 383 |
+
|
| 384 |
+
If charts/visualizations are mentioned, include chart_data_provider.
|
| 385 |
+
|
| 386 |
+
Examples:
|
| 387 |
+
- "Bitcoin price" → cryptocompare_data, chart_data_provider
|
| 388 |
+
- "DeFi TVL" → defillama_data, chart_data_provider
|
| 389 |
+
- "Ethereum transactions" → etherscan_data
|
| 390 |
+
|
| 391 |
+
Respond with only the tool names, comma-separated (no explanations)."""
|
| 392 |
+
|
| 393 |
+
tool_response = await self.llm.ainvoke(tool_analysis_prompt)
|
| 394 |
+
|
| 395 |
+
logger.info(f"🧠 Gemini tool analysis response: {str(tool_response)[:100]}...")
|
| 396 |
+
|
| 397 |
+
# Parse suggested tools
|
| 398 |
+
suggested_tools = [tool.strip() for tool in str(tool_response).split(',') if tool.strip()]
|
| 399 |
+
suggested_tools = [tool for tool in suggested_tools if tool in {
|
| 400 |
+
'cryptocompare_data', 'coingecko_data', 'defillama_data',
|
| 401 |
+
'etherscan_data', 'chart_data_provider'
|
| 402 |
+
}]
|
| 403 |
+
|
| 404 |
+
logger.info(f"🛠️ Gemini suggested tools: {suggested_tools}")
|
| 405 |
+
|
| 406 |
+
# Step 2: Execute tools (same logic as Ollama version)
|
| 407 |
+
tool_results = []
|
| 408 |
+
for tool_name in suggested_tools:
|
| 409 |
+
tool = next((t for t in self.tools if t.name == tool_name), None)
|
| 410 |
+
if tool:
|
| 411 |
+
try:
|
| 412 |
+
logger.info(f"🔧 Executing {tool_name}")
|
| 413 |
+
|
| 414 |
+
# Handle chart_data_provider with proper parameters
|
| 415 |
+
if tool_name == "chart_data_provider":
|
| 416 |
+
chart_type = "price_chart"
|
| 417 |
+
symbol = "bitcoin"
|
| 418 |
+
|
| 419 |
+
if "defi" in query.lower() or "tvl" in query.lower():
|
| 420 |
+
chart_type = "defi_tvl"
|
| 421 |
+
elif "market" in query.lower() or "overview" in query.lower():
|
| 422 |
+
chart_type = "market_overview"
|
| 423 |
+
elif "gas" in query.lower():
|
| 424 |
+
chart_type = "gas_tracker"
|
| 425 |
+
|
| 426 |
+
if "ethereum" in query.lower() or "eth" in query.lower():
|
| 427 |
+
symbol = "ethereum"
|
| 428 |
+
elif "bitcoin" in query.lower() or "btc" in query.lower():
|
| 429 |
+
symbol = "bitcoin"
|
| 430 |
+
|
| 431 |
+
result = await tool._arun(chart_type=chart_type, symbol=symbol)
|
| 432 |
+
else:
|
| 433 |
+
result = await tool._arun(query)
|
| 434 |
+
|
| 435 |
+
logger.info(f"📊 {tool_name} result preview: {str(result)[:200]}...")
|
| 436 |
+
tool_results.append(f"=== {tool_name} Results ===\n{result}\n")
|
| 437 |
+
except Exception as e:
|
| 438 |
+
logger.error(f"Tool {tool_name} failed: {e}")
|
| 439 |
+
tool_results.append(f"=== {tool_name} Error ===\nTool failed: {str(e)}\n")
|
| 440 |
+
|
| 441 |
+
# Step 3: Generate final response with Gemini
|
| 442 |
+
context = "\n".join(tool_results) if tool_results else "No tool data available - provide general information."
|
| 443 |
+
|
| 444 |
+
final_prompt = ai_safety.create_safe_prompt(query, context)
|
| 445 |
+
|
| 446 |
+
try:
|
| 447 |
+
final_response = await asyncio.wait_for(
|
| 448 |
+
self.llm.ainvoke(final_prompt),
|
| 449 |
+
timeout=30
|
| 450 |
+
)
|
| 451 |
+
logger.info(f"🎯 Gemini final response preview: {str(final_response)[:300]}...")
|
| 452 |
+
|
| 453 |
+
# AI Safety Check: Validate response
|
| 454 |
+
clean_response, response_safe, response_reason = ai_safety.validate_gemini_response(str(final_response))
|
| 455 |
+
if not response_safe:
|
| 456 |
+
ai_safety.log_safety_event("blocked_gemini_response", {
|
| 457 |
+
"reason": response_reason,
|
| 458 |
+
"query": query[:100],
|
| 459 |
+
"timestamp": datetime.now().isoformat()
|
| 460 |
+
})
|
| 461 |
+
clean_response = f"## Cryptocurrency Analysis\n\nBased on the available data:\n\n{context[:1000]}\n\n*Response filtered for safety*"
|
| 462 |
+
|
| 463 |
+
final_response = clean_response
|
| 464 |
+
|
| 465 |
+
except asyncio.TimeoutError:
|
| 466 |
+
logger.warning("⏱️ Gemini final response timed out, using tool data directly")
|
| 467 |
+
final_response = f"## Web3 Research Analysis\n\n{context[:1500]}\n\n*Analysis completed using available tools - Gemini response timed out*"
|
| 468 |
+
|
| 469 |
+
logger.info("✅ Research successful with Gemini + tools")
|
| 470 |
+
|
| 471 |
+
return {
|
| 472 |
+
"success": True,
|
| 473 |
+
"query": query,
|
| 474 |
+
"result": final_response,
|
| 475 |
+
"sources": [],
|
| 476 |
+
"metadata": {
|
| 477 |
+
"llm_used": f"Gemini ({self.llm.model_name if hasattr(self.llm, 'model_name') else 'gemini-pro'})",
|
| 478 |
+
"tools_used": suggested_tools,
|
| 479 |
+
"timestamp": datetime.now().isoformat()
|
| 480 |
+
}
|
| 481 |
+
}
|
| 482 |
+
|
| 483 |
+
except Exception as e:
|
| 484 |
+
logger.error(f"Gemini tools research failed: {e}")
|
| 485 |
+
# Fallback to Ollama if Gemini fails
|
| 486 |
+
logger.info("🔄 Falling back to Ollama due to Gemini error")
|
| 487 |
+
return await self._research_with_ollama_tools(query)
|
| 488 |
+
|
| 489 |
def _extract_sources(self, response: str) -> List[str]:
|
| 490 |
"""Extract sources from response"""
|
| 491 |
# Simple source extraction - can be enhanced
|
|
@@ -84,9 +84,9 @@ class ChartDataTool(BaseTool):
|
|
| 84 |
})
|
| 85 |
|
| 86 |
async def _get_price_chart_data(self, symbol: str, days: int) -> str:
|
| 87 |
-
"""Get
|
| 88 |
try:
|
| 89 |
-
#
|
| 90 |
from src.tools.coingecko_tool import CoinGeckoTool
|
| 91 |
|
| 92 |
coingecko = CoinGeckoTool()
|
|
@@ -107,42 +107,48 @@ class ChartDataTool(BaseTool):
|
|
| 107 |
|
| 108 |
coin_id = symbol_map.get(symbol.lower(), symbol.lower())
|
| 109 |
|
| 110 |
-
|
| 111 |
-
|
| 112 |
-
|
| 113 |
-
|
| 114 |
-
|
| 115 |
-
|
| 116 |
-
|
| 117 |
-
|
| 118 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 119 |
return await self._get_mock_price_data(symbol, days)
|
| 120 |
|
| 121 |
-
# Format the real data
|
| 122 |
-
price_data = data.get("prices", [])
|
| 123 |
-
volume_data = data.get("total_volumes", [])
|
| 124 |
-
|
| 125 |
-
# Get current coin info
|
| 126 |
-
coin_info = await coingecko.make_request(f"https://api.coingecko.com/api/v3/coins/{coin_id}")
|
| 127 |
-
coin_name = coin_info.get("name", symbol.title()) if coin_info else symbol.title()
|
| 128 |
-
|
| 129 |
-
return json.dumps({
|
| 130 |
-
"chart_type": "price_chart",
|
| 131 |
-
"data": {
|
| 132 |
-
"prices": price_data,
|
| 133 |
-
"total_volumes": volume_data,
|
| 134 |
-
"symbol": symbol.upper(),
|
| 135 |
-
"name": coin_name
|
| 136 |
-
},
|
| 137 |
-
"config": {
|
| 138 |
-
"title": f"{coin_name} Price Analysis ({days} days)",
|
| 139 |
-
"timeframe": f"{days}d",
|
| 140 |
-
"currency": "USD"
|
| 141 |
-
}
|
| 142 |
-
})
|
| 143 |
-
|
| 144 |
except Exception as e:
|
| 145 |
-
logger.error(f"
|
|
|
|
| 146 |
return await self._get_mock_price_data(symbol, days)
|
| 147 |
|
| 148 |
async def _get_mock_price_data(self, symbol: str, days: int) -> str:
|
|
|
|
| 84 |
})
|
| 85 |
|
| 86 |
async def _get_price_chart_data(self, symbol: str, days: int) -> str:
|
| 87 |
+
"""Get price chart data with fallback for API failures"""
|
| 88 |
try:
|
| 89 |
+
# First try to get real data from CoinGecko
|
| 90 |
from src.tools.coingecko_tool import CoinGeckoTool
|
| 91 |
|
| 92 |
coingecko = CoinGeckoTool()
|
|
|
|
| 107 |
|
| 108 |
coin_id = symbol_map.get(symbol.lower(), symbol.lower())
|
| 109 |
|
| 110 |
+
try:
|
| 111 |
+
# Use basic API endpoint that doesn't require premium
|
| 112 |
+
url = f"https://api.coingecko.com/api/v3/coins/{coin_id}/market_chart"
|
| 113 |
+
params = {"vs_currency": "usd", "days": days, "interval": "daily"}
|
| 114 |
+
|
| 115 |
+
data = await coingecko.make_request(url, params=params)
|
| 116 |
+
|
| 117 |
+
if data and "prices" in data:
|
| 118 |
+
# Format the real data
|
| 119 |
+
price_data = data.get("prices", [])
|
| 120 |
+
volume_data = data.get("total_volumes", [])
|
| 121 |
+
|
| 122 |
+
# Get current coin info
|
| 123 |
+
coin_info = await coingecko.make_request(f"https://api.coingecko.com/api/v3/coins/{coin_id}")
|
| 124 |
+
coin_name = coin_info.get("name", symbol.title()) if coin_info else symbol.title()
|
| 125 |
+
|
| 126 |
+
return json.dumps({
|
| 127 |
+
"chart_type": "price_chart",
|
| 128 |
+
"data": {
|
| 129 |
+
"prices": price_data,
|
| 130 |
+
"total_volumes": volume_data,
|
| 131 |
+
"symbol": symbol.upper(),
|
| 132 |
+
"name": coin_name
|
| 133 |
+
},
|
| 134 |
+
"config": {
|
| 135 |
+
"title": f"{coin_name} Price Analysis ({days} days)",
|
| 136 |
+
"timeframe": f"{days}d",
|
| 137 |
+
"currency": "USD"
|
| 138 |
+
}
|
| 139 |
+
})
|
| 140 |
+
else:
|
| 141 |
+
raise Exception("No price data in response")
|
| 142 |
+
|
| 143 |
+
except Exception as api_error:
|
| 144 |
+
logger.error(f"Real price data failed: {api_error}")
|
| 145 |
+
# Fallback to mock data on any API error
|
| 146 |
+
logger.info(f"Using fallback mock data for {symbol}")
|
| 147 |
return await self._get_mock_price_data(symbol, days)
|
| 148 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 149 |
except Exception as e:
|
| 150 |
+
logger.error(f"Price chart data generation failed: {e}")
|
| 151 |
+
# Final fallback to mock data
|
| 152 |
return await self._get_mock_price_data(symbol, days)
|
| 153 |
|
| 154 |
async def _get_mock_price_data(self, symbol: str, days: int) -> str:
|
|
@@ -8,9 +8,9 @@ load_dotenv()
|
|
| 8 |
|
| 9 |
@dataclass
|
| 10 |
class Config:
|
| 11 |
-
# LLM Configuration - Ollama
|
| 12 |
-
GEMINI_API_KEY: str = "" #
|
| 13 |
-
USE_OLLAMA_ONLY: bool =
|
| 14 |
|
| 15 |
# Available API Keys
|
| 16 |
COINGECKO_API_KEY: Optional[str] = None # Not available - costs money
|
|
|
|
| 8 |
|
| 9 |
@dataclass
|
| 10 |
class Config:
|
| 11 |
+
# LLM Configuration - Both Ollama and Gemini available
|
| 12 |
+
GEMINI_API_KEY: str = os.getenv("GEMINI_API_KEY", "") # Enable Gemini when API key provided
|
| 13 |
+
USE_OLLAMA_ONLY: bool = not bool(os.getenv("GEMINI_API_KEY")) # Auto-detect based on API key
|
| 14 |
|
| 15 |
# Available API Keys
|
| 16 |
COINGECKO_API_KEY: Optional[str] = None # Not available - costs money
|
|
@@ -1,5 +1,38 @@
|
|
| 1 |
let chatHistory = [];
|
| 2 |
let messageCount = 0;
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
|
| 4 |
async function checkStatus() {
|
| 5 |
try {
|
|
@@ -68,7 +101,11 @@ async function sendQuery() {
|
|
| 68 |
'Accept': 'text/event-stream',
|
| 69 |
'Cache-Control': 'no-cache'
|
| 70 |
},
|
| 71 |
-
body: JSON.stringify({
|
|
|
|
|
|
|
|
|
|
|
|
|
| 72 |
signal: controller.signal,
|
| 73 |
// Disable browser's default timeout behavior
|
| 74 |
keepalive: true
|
|
|
|
| 1 |
let chatHistory = [];
|
| 2 |
let messageCount = 0;
|
| 3 |
+
let useGemini = false; // Track current LLM choice
|
| 4 |
+
|
| 5 |
+
// Initialize Gemini toggle
|
| 6 |
+
document.addEventListener('DOMContentLoaded', function() {
|
| 7 |
+
const geminiToggle = document.getElementById('geminiToggle');
|
| 8 |
+
const toggleLabel = document.querySelector('.toggle-label');
|
| 9 |
+
|
| 10 |
+
// Load saved preference
|
| 11 |
+
useGemini = localStorage.getItem('useGemini') === 'true';
|
| 12 |
+
geminiToggle.checked = useGemini;
|
| 13 |
+
updateToggleLabel();
|
| 14 |
+
|
| 15 |
+
// Handle toggle changes
|
| 16 |
+
geminiToggle.addEventListener('change', function() {
|
| 17 |
+
useGemini = this.checked;
|
| 18 |
+
localStorage.setItem('useGemini', useGemini.toString());
|
| 19 |
+
updateToggleLabel();
|
| 20 |
+
console.log(`Switched to ${useGemini ? 'Gemini' : 'Ollama'} mode`);
|
| 21 |
+
|
| 22 |
+
// Show confirmation
|
| 23 |
+
showStatus(`Switched to ${useGemini ? 'Gemini (Cloud AI)' : 'Ollama (Local AI)'} mode`, 'info');
|
| 24 |
+
|
| 25 |
+
// Refresh status to reflect changes
|
| 26 |
+
checkStatus();
|
| 27 |
+
});
|
| 28 |
+
});
|
| 29 |
+
|
| 30 |
+
function updateToggleLabel() {
|
| 31 |
+
const toggleLabel = document.querySelector('.toggle-label');
|
| 32 |
+
if (toggleLabel) {
|
| 33 |
+
toggleLabel.textContent = `AI Model: ${useGemini ? 'Gemini' : 'Ollama'}`;
|
| 34 |
+
}
|
| 35 |
+
}
|
| 36 |
|
| 37 |
async function checkStatus() {
|
| 38 |
try {
|
|
|
|
| 101 |
'Accept': 'text/event-stream',
|
| 102 |
'Cache-Control': 'no-cache'
|
| 103 |
},
|
| 104 |
+
body: JSON.stringify({
|
| 105 |
+
query,
|
| 106 |
+
chat_history: chatHistory,
|
| 107 |
+
use_gemini: useGemini
|
| 108 |
+
}),
|
| 109 |
signal: controller.signal,
|
| 110 |
// Disable browser's default timeout behavior
|
| 111 |
keepalive: true
|
|
@@ -66,6 +66,109 @@ body {
|
|
| 66 |
flex: 1;
|
| 67 |
text-align: center;
|
| 68 |
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 69 |
.theme-toggle {
|
| 70 |
background: var(--surface);
|
| 71 |
border: 1px solid var(--border);
|
|
|
|
| 66 |
flex: 1;
|
| 67 |
text-align: center;
|
| 68 |
}
|
| 69 |
+
.header-controls {
|
| 70 |
+
display: flex;
|
| 71 |
+
align-items: center;
|
| 72 |
+
gap: 1rem;
|
| 73 |
+
}
|
| 74 |
+
|
| 75 |
+
/* LLM Toggle Switch Styles */
|
| 76 |
+
.llm-toggle {
|
| 77 |
+
display: flex;
|
| 78 |
+
align-items: center;
|
| 79 |
+
gap: 0.5rem;
|
| 80 |
+
}
|
| 81 |
+
.toggle-label {
|
| 82 |
+
font-size: 0.875rem;
|
| 83 |
+
color: var(--text-secondary);
|
| 84 |
+
font-weight: 500;
|
| 85 |
+
}
|
| 86 |
+
|
| 87 |
+
.switch {
|
| 88 |
+
position: relative;
|
| 89 |
+
display: inline-block;
|
| 90 |
+
width: 80px;
|
| 91 |
+
height: 32px;
|
| 92 |
+
}
|
| 93 |
+
|
| 94 |
+
.switch input {
|
| 95 |
+
opacity: 0;
|
| 96 |
+
width: 0;
|
| 97 |
+
height: 0;
|
| 98 |
+
}
|
| 99 |
+
|
| 100 |
+
.slider {
|
| 101 |
+
position: absolute;
|
| 102 |
+
cursor: pointer;
|
| 103 |
+
top: 0;
|
| 104 |
+
left: 0;
|
| 105 |
+
right: 0;
|
| 106 |
+
bottom: 0;
|
| 107 |
+
background-color: var(--surface);
|
| 108 |
+
border: 1px solid var(--border);
|
| 109 |
+
transition: .4s;
|
| 110 |
+
overflow: hidden;
|
| 111 |
+
}
|
| 112 |
+
|
| 113 |
+
.slider:before {
|
| 114 |
+
position: absolute;
|
| 115 |
+
content: "";
|
| 116 |
+
height: 24px;
|
| 117 |
+
width: 24px;
|
| 118 |
+
left: 3px;
|
| 119 |
+
bottom: 3px;
|
| 120 |
+
background-color: var(--primary);
|
| 121 |
+
transition: .4s;
|
| 122 |
+
border-radius: 50%;
|
| 123 |
+
z-index: 2;
|
| 124 |
+
}
|
| 125 |
+
|
| 126 |
+
.slider-text-off, .slider-text-on {
|
| 127 |
+
position: absolute;
|
| 128 |
+
color: var(--text-secondary);
|
| 129 |
+
font-size: 0.7rem;
|
| 130 |
+
font-weight: 500;
|
| 131 |
+
top: 50%;
|
| 132 |
+
transform: translateY(-50%);
|
| 133 |
+
transition: .4s;
|
| 134 |
+
pointer-events: none;
|
| 135 |
+
z-index: 1;
|
| 136 |
+
}
|
| 137 |
+
|
| 138 |
+
.slider-text-off {
|
| 139 |
+
left: 8px;
|
| 140 |
+
}
|
| 141 |
+
|
| 142 |
+
.slider-text-on {
|
| 143 |
+
right: 8px;
|
| 144 |
+
opacity: 0;
|
| 145 |
+
}
|
| 146 |
+
|
| 147 |
+
input:checked + .slider {
|
| 148 |
+
background-color: var(--accent);
|
| 149 |
+
border-color: var(--accent);
|
| 150 |
+
}
|
| 151 |
+
|
| 152 |
+
input:checked + .slider .slider-text-off {
|
| 153 |
+
opacity: 0;
|
| 154 |
+
}
|
| 155 |
+
|
| 156 |
+
input:checked + .slider .slider-text-on {
|
| 157 |
+
opacity: 1;
|
| 158 |
+
}
|
| 159 |
+
|
| 160 |
+
input:checked + .slider:before {
|
| 161 |
+
transform: translateX(48px);
|
| 162 |
+
}
|
| 163 |
+
|
| 164 |
+
.slider.round {
|
| 165 |
+
border-radius: 20px;
|
| 166 |
+
}
|
| 167 |
+
|
| 168 |
+
.slider.round:before {
|
| 169 |
+
border-radius: 50%;
|
| 170 |
+
}
|
| 171 |
+
|
| 172 |
.theme-toggle {
|
| 173 |
background: var(--surface);
|
| 174 |
border: 1px solid var(--border);
|
|
@@ -21,9 +21,21 @@
|
|
| 21 |
<h1><span class="brand">Web3</span> Research Co-Pilot</h1>
|
| 22 |
<p>Professional cryptocurrency analysis and market intelligence</p>
|
| 23 |
</div>
|
| 24 |
-
<
|
| 25 |
-
<
|
| 26 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 27 |
</div>
|
| 28 |
</div>
|
| 29 |
|
|
|
|
| 21 |
<h1><span class="brand">Web3</span> Research Co-Pilot</h1>
|
| 22 |
<p>Professional cryptocurrency analysis and market intelligence</p>
|
| 23 |
</div>
|
| 24 |
+
<div class="header-controls">
|
| 25 |
+
<div class="llm-toggle">
|
| 26 |
+
<span class="toggle-label">AI Model:</span>
|
| 27 |
+
<label class="switch">
|
| 28 |
+
<input type="checkbox" id="geminiToggle" title="Switch between Ollama (Local) and Gemini (Cloud)">
|
| 29 |
+
<span class="slider round">
|
| 30 |
+
<span class="slider-text-off">Ollama</span>
|
| 31 |
+
<span class="slider-text-on">Gemini</span>
|
| 32 |
+
</span>
|
| 33 |
+
</label>
|
| 34 |
+
</div>
|
| 35 |
+
<button id="themeToggle" class="theme-toggle" title="Toggle theme">
|
| 36 |
+
<i class="fas fa-moon"></i>
|
| 37 |
+
</button>
|
| 38 |
+
</div>
|
| 39 |
</div>
|
| 40 |
</div>
|
| 41 |
|