|
|
""" |
|
|
MCP Client implementation for Universal MCP Client - Fixed Version |
|
|
""" |
|
|
import asyncio |
|
|
import json |
|
|
import re |
|
|
import logging |
|
|
import traceback |
|
|
from typing import Dict, Optional, Tuple, List, Any |
|
|
from openai import OpenAI |
|
|
|
|
|
|
|
|
from mcp import ClientSession |
|
|
from mcp.client.sse import sse_client |
|
|
|
|
|
from config import MCPServerConfig, AppConfig, HTTPX_AVAILABLE |
|
|
|
|
|
logger = logging.getLogger(__name__) |
|
|
|
|
|
class UniversalMCPClient: |
|
|
"""Universal MCP Client using HuggingFace Inference Providers instead of Anthropic""" |
|
|
|
|
|
def __init__(self): |
|
|
self.servers: Dict[str, MCPServerConfig] = {} |
|
|
self.enabled_servers: Dict[str, bool] = {} |
|
|
self.hf_client = None |
|
|
self.current_provider = None |
|
|
self.current_model = None |
|
|
self.server_tools = {} |
|
|
self.generation_params: Dict[str, Any] = {} |
|
|
|
|
|
|
|
|
if AppConfig.HF_TOKEN: |
|
|
self.hf_client = OpenAI( |
|
|
base_url="https://router.huggingface.co/v1", |
|
|
api_key=AppConfig.HF_TOKEN |
|
|
) |
|
|
logger.info("β
HuggingFace Inference client initialized") |
|
|
else: |
|
|
logger.warning("β οΈ HF_TOKEN not found") |
|
|
|
|
|
def enable_server(self, server_name: str, enabled: bool = True): |
|
|
"""Enable or disable a server""" |
|
|
if server_name in self.servers: |
|
|
self.enabled_servers[server_name] = enabled |
|
|
logger.info(f"π§ Server {server_name} {'enabled' if enabled else 'disabled'}") |
|
|
|
|
|
def get_enabled_servers(self) -> Dict[str, MCPServerConfig]: |
|
|
"""Get only enabled servers""" |
|
|
return {name: config for name, config in self.servers.items() |
|
|
if self.enabled_servers.get(name, True)} |
|
|
|
|
|
def remove_all_servers(self): |
|
|
"""Remove all servers""" |
|
|
count = len(self.servers) |
|
|
self.servers.clear() |
|
|
self.enabled_servers.clear() |
|
|
self.server_tools.clear() |
|
|
logger.info(f"ποΈ Removed all {count} servers") |
|
|
return count |
|
|
|
|
|
def set_model_and_provider(self, provider_id: str, model_id: str): |
|
|
"""Set the current provider and model""" |
|
|
self.current_provider = provider_id |
|
|
self.current_model = model_id |
|
|
logger.info(f"π§ Set provider: {provider_id}, model: {model_id}") |
|
|
|
|
|
def set_generation_params(self, params: Dict[str, Any]): |
|
|
"""Set generation parameters for chat completions (OpenAI-compatible).""" |
|
|
|
|
|
cleaned = {k: v for k, v in params.items() if v is not None} |
|
|
self.generation_params = cleaned |
|
|
logger.info(f"π§ Updated generation params: {list(self.generation_params.keys())}") |
|
|
|
|
|
def get_model_endpoint(self) -> str: |
|
|
"""Get the current model endpoint for API calls""" |
|
|
if not self.current_provider or not self.current_model: |
|
|
raise ValueError("Provider and model must be set before making API calls") |
|
|
|
|
|
return AppConfig.get_model_endpoint(self.current_model, self.current_provider) |
|
|
|
|
|
async def add_server_async(self, config: MCPServerConfig) -> Tuple[bool, str]: |
|
|
"""Add an MCP server using pure MCP protocol""" |
|
|
try: |
|
|
logger.info(f"π§ Adding MCP server: {config.name} at {config.url}") |
|
|
|
|
|
|
|
|
original_url = config.url.strip() |
|
|
|
|
|
|
|
|
base_url = original_url |
|
|
for endpoint in ["/gradio_api/mcp/sse", "/gradio_api/mcp/", "/gradio_api/mcp"]: |
|
|
if base_url.endswith(endpoint): |
|
|
base_url = base_url[:-len(endpoint)] |
|
|
break |
|
|
|
|
|
|
|
|
base_url = base_url.rstrip("/") |
|
|
|
|
|
|
|
|
mcp_url = f"{base_url}/gradio_api/mcp/sse" |
|
|
|
|
|
logger.info(f"π§ Original URL: {original_url}") |
|
|
logger.info(f"π§ Base URL: {base_url}") |
|
|
logger.info(f"π§ MCP URL: {mcp_url}") |
|
|
|
|
|
|
|
|
if "hf.space" in base_url: |
|
|
space_parts = base_url.split("/") |
|
|
if len(space_parts) >= 1: |
|
|
space_id = space_parts[-1].replace('.hf.space', '').replace('https://', '').replace('http://', '') |
|
|
if '-' in space_id: |
|
|
|
|
|
config.space_id = space_id.replace('-', '/', 1) |
|
|
else: |
|
|
config.space_id = space_id |
|
|
logger.info(f"π Detected HF Space ID: {config.space_id}") |
|
|
|
|
|
|
|
|
config.url = mcp_url |
|
|
|
|
|
|
|
|
success, message = await self._test_mcp_connection(config) |
|
|
|
|
|
if success: |
|
|
self.servers[config.name] = config |
|
|
self.enabled_servers[config.name] = True |
|
|
logger.info(f"β
MCP Server {config.name} added successfully") |
|
|
return True, f"β
Successfully added MCP server: {config.name}\n{message}" |
|
|
else: |
|
|
logger.error(f"β Failed to connect to MCP server {config.name}: {message}") |
|
|
return False, f"β Failed to add server: {config.name}\n{message}" |
|
|
|
|
|
except Exception as e: |
|
|
error_msg = f"Failed to add server {config.name}: {str(e)}" |
|
|
logger.error(error_msg) |
|
|
logger.error(traceback.format_exc()) |
|
|
return False, f"β {error_msg}" |
|
|
|
|
|
async def _test_mcp_connection(self, config: MCPServerConfig) -> Tuple[bool, str]: |
|
|
"""Test MCP server connection with detailed debugging and tool caching""" |
|
|
try: |
|
|
logger.info(f"π Testing MCP connection to {config.url}") |
|
|
|
|
|
async with sse_client(config.url, timeout=20.0) as (read_stream, write_stream): |
|
|
async with ClientSession(read_stream, write_stream) as session: |
|
|
|
|
|
logger.info("π§ Initializing MCP session...") |
|
|
await session.initialize() |
|
|
|
|
|
|
|
|
logger.info("π Listing available tools...") |
|
|
tools = await session.list_tools() |
|
|
|
|
|
|
|
|
server_tools = {} |
|
|
tool_info = [] |
|
|
|
|
|
for tool in tools.tools: |
|
|
server_tools[tool.name] = { |
|
|
'description': tool.description, |
|
|
'schema': tool.inputSchema if hasattr(tool, 'inputSchema') else None |
|
|
} |
|
|
tool_info.append(f" - {tool.name}: {tool.description}") |
|
|
logger.info(f" π Tool: {tool.name}") |
|
|
logger.info(f" Description: {tool.description}") |
|
|
if hasattr(tool, 'inputSchema') and tool.inputSchema: |
|
|
logger.info(f" Input Schema: {tool.inputSchema}") |
|
|
|
|
|
|
|
|
self.server_tools[config.name] = server_tools |
|
|
|
|
|
if len(tools.tools) == 0: |
|
|
return False, "No tools found on MCP server" |
|
|
|
|
|
message = f"Connected successfully!\nFound {len(tools.tools)} tools:\n" + "\n".join(tool_info) |
|
|
return True, message |
|
|
|
|
|
except asyncio.TimeoutError: |
|
|
return False, "Connection timeout - server may be sleeping or unreachable" |
|
|
except Exception as e: |
|
|
logger.error(f"MCP connection failed: {e}") |
|
|
logger.error(traceback.format_exc()) |
|
|
return False, f"Connection failed: {str(e)}" |
|
|
|
|
|
async def call_mcp_tool_async(self, server_name: str, tool_name: str, arguments: dict) -> Tuple[bool, str]: |
|
|
"""Call a tool on a specific MCP server""" |
|
|
logger.info(f"π§ MCP Tool Call - Server: {server_name}, Tool: {tool_name}") |
|
|
logger.info(f"π§ Arguments: {arguments}") |
|
|
|
|
|
if server_name not in self.servers: |
|
|
error_msg = f"Server {server_name} not found. Available servers: {list(self.servers.keys())}" |
|
|
logger.error(f"β {error_msg}") |
|
|
return False, error_msg |
|
|
|
|
|
config = self.servers[server_name] |
|
|
logger.info(f"π§ Using server config: {config.url}") |
|
|
|
|
|
try: |
|
|
logger.info(f"π Connecting to MCP server at {config.url}") |
|
|
async with sse_client(config.url, timeout=30.0) as (read_stream, write_stream): |
|
|
async with ClientSession(read_stream, write_stream) as session: |
|
|
|
|
|
logger.info("π§ Initializing MCP session...") |
|
|
await session.initialize() |
|
|
|
|
|
|
|
|
logger.info(f"π§ Calling tool {tool_name} with arguments: {arguments}") |
|
|
result = await session.call_tool(tool_name, arguments) |
|
|
|
|
|
|
|
|
if result.content: |
|
|
result_text = result.content[0].text if hasattr(result.content[0], 'text') else str(result.content[0]) |
|
|
logger.info(f"β
Tool call successful, result length: {len(result_text)}") |
|
|
logger.info(f"π Result preview: {result_text[:200]}...") |
|
|
return True, result_text |
|
|
else: |
|
|
error_msg = "No content returned from tool" |
|
|
logger.error(f"β {error_msg}") |
|
|
return False, error_msg |
|
|
|
|
|
except asyncio.TimeoutError: |
|
|
error_msg = f"Tool call timeout for {tool_name} on {server_name}" |
|
|
logger.error(f"β {error_msg}") |
|
|
return False, error_msg |
|
|
except Exception as e: |
|
|
error_msg = f"Tool call failed: {str(e)}" |
|
|
logger.error(f"β MCP tool call failed: {e}") |
|
|
logger.error(traceback.format_exc()) |
|
|
return False, error_msg |
|
|
|
|
|
def generate_chat_completion(self, messages: List[Dict[str, Any]], **kwargs) -> Dict[str, Any]: |
|
|
"""Generate chat completion using HuggingFace Inference Providers""" |
|
|
if not self.hf_client: |
|
|
raise ValueError("HuggingFace client not initialized. Please set HF_TOKEN.") |
|
|
|
|
|
if not self.current_provider or not self.current_model: |
|
|
raise ValueError("Provider and model must be set before making API calls") |
|
|
|
|
|
|
|
|
model_endpoint = self.get_model_endpoint() |
|
|
|
|
|
|
|
|
params = { |
|
|
"model": model_endpoint, |
|
|
"messages": messages, |
|
|
"max_tokens": kwargs.pop("max_tokens", 8192), |
|
|
"temperature": kwargs.get("temperature", 0.3), |
|
|
"stream": kwargs.get("stream", False) |
|
|
} |
|
|
|
|
|
|
|
|
for k, v in self.generation_params.items(): |
|
|
if k not in ("model", "messages") and k not in params: |
|
|
params[k] = v |
|
|
|
|
|
|
|
|
params.update(kwargs) |
|
|
|
|
|
|
|
|
if AppConfig.is_gpt_oss_model(self.current_model): |
|
|
reasoning_effort = kwargs.pop("reasoning_effort", self.generation_params.get("reasoning_effort", AppConfig.DEFAULT_REASONING_EFFORT)) |
|
|
if reasoning_effort: |
|
|
|
|
|
system_message = None |
|
|
for msg in messages: |
|
|
if msg.get("role") == "system": |
|
|
system_message = msg |
|
|
break |
|
|
|
|
|
if system_message: |
|
|
system_message["content"] += f"\n\nReasoning: {reasoning_effort}" |
|
|
else: |
|
|
messages.insert(0, { |
|
|
"role": "system", |
|
|
"content": f"You are a helpful AI assistant. Reasoning: {reasoning_effort}" |
|
|
}) |
|
|
else: |
|
|
|
|
|
if "reasoning_effort" in params: |
|
|
params.pop("reasoning_effort", None) |
|
|
|
|
|
try: |
|
|
logger.info(f"π€ Calling {model_endpoint} via {self.current_provider}") |
|
|
response = self.hf_client.chat.completions.create(**params) |
|
|
return response |
|
|
except Exception as e: |
|
|
logger.error(f"HF Inference API call failed: {e}") |
|
|
raise |
|
|
|
|
|
def generate_chat_completion_with_mcp_tools(self, messages: List[Dict[str, Any]], **kwargs) -> Dict[str, Any]: |
|
|
"""Generate chat completion with MCP tool support""" |
|
|
enabled_servers = self.get_enabled_servers() |
|
|
if not enabled_servers: |
|
|
|
|
|
logger.info("π€ No enabled MCP servers available, using regular chat completion") |
|
|
return self.generate_chat_completion(messages, **kwargs) |
|
|
|
|
|
logger.info(f"π§ Processing chat with {len(enabled_servers)} enabled MCP servers available") |
|
|
|
|
|
|
|
|
tool_descriptions = [] |
|
|
server_names = [] |
|
|
exact_tool_mappings = [] |
|
|
|
|
|
for server_name, config in enabled_servers.items(): |
|
|
tool_descriptions.append(f"- **{server_name}**: {config.description}") |
|
|
server_names.append(server_name) |
|
|
|
|
|
|
|
|
if server_name in self.server_tools: |
|
|
for tool_name, tool_info in self.server_tools[server_name].items(): |
|
|
exact_tool_mappings.append(f" * Server '{server_name}' has tool '{tool_name}': {tool_info['description']}") |
|
|
|
|
|
|
|
|
server_list = ", ".join([f'"{name}"' for name in server_names]) |
|
|
|
|
|
tools_system_msg = f""" |
|
|
You have access to the following MCP tools: |
|
|
{chr(10).join(tool_descriptions)} |
|
|
EXACT TOOL MAPPINGS: |
|
|
{chr(10).join(exact_tool_mappings) if exact_tool_mappings else "Loading tool mappings..."} |
|
|
IMPORTANT SERVER NAMES: {server_list} |
|
|
When you need to use a tool, respond with ONLY a JSON object in this EXACT format: |
|
|
{{"use_tool": true, "server": "exact_server_name", "tool": "exact_tool_name", "arguments": {{"param": "value"}}}} |
|
|
CRITICAL INSTRUCTIONS: |
|
|
- Use ONLY the exact server names from this list: {server_list} |
|
|
- Use the exact tool names as shown in the mappings above |
|
|
- Always include all required parameters in the arguments |
|
|
- Do not include any other text before or after the JSON |
|
|
- Make sure the JSON is complete and properly formatted |
|
|
If you don't need to use a tool, respond normally without any JSON. |
|
|
""" |
|
|
|
|
|
|
|
|
enhanced_messages = messages.copy() |
|
|
if enhanced_messages and enhanced_messages[0].get("role") == "system": |
|
|
enhanced_messages[0]["content"] += "\n\n" + tools_system_msg |
|
|
else: |
|
|
enhanced_messages.insert(0, {"role": "system", "content": tools_system_msg}) |
|
|
|
|
|
|
|
|
logger.info("π€ Getting initial response from LLM...") |
|
|
response = self.generate_chat_completion(enhanced_messages, **{"max_tokens": 8192}) |
|
|
response_text = response.choices[0].message.content |
|
|
|
|
|
logger.info(f"π€ LLM Response (length: {len(response_text)}): {response_text}") |
|
|
|
|
|
|
|
|
if '"use_tool": true' in response_text: |
|
|
logger.info("π§ Tool usage detected, parsing JSON...") |
|
|
|
|
|
|
|
|
tool_request = self._extract_tool_json(response_text) |
|
|
|
|
|
if not tool_request: |
|
|
|
|
|
logger.info("π§ JSON parsing failed, trying manual extraction...") |
|
|
tool_request = self._manual_tool_extraction(response_text) |
|
|
|
|
|
if tool_request: |
|
|
server_name = tool_request.get("server") |
|
|
tool_name = tool_request.get("tool") |
|
|
arguments = tool_request.get("arguments", {}) |
|
|
|
|
|
|
|
|
if hasattr(self, 'chat_handler_file_mapping'): |
|
|
for arg_key, arg_value in arguments.items(): |
|
|
if isinstance(arg_value, str) and arg_value.startswith('/tmp/gradio/'): |
|
|
|
|
|
for local_path, uploaded_url in self.chat_handler_file_mapping.items(): |
|
|
if local_path in arg_value or arg_value in local_path: |
|
|
logger.info(f"π Replacing local path {arg_value} with uploaded URL {uploaded_url}") |
|
|
arguments[arg_key] = uploaded_url |
|
|
break |
|
|
|
|
|
logger.info(f"π§ Tool request - Server: {server_name}, Tool: {tool_name}, Args: {arguments}") |
|
|
|
|
|
if server_name not in self.servers: |
|
|
available_servers = list(self.servers.keys()) |
|
|
logger.error(f"β Server '{server_name}' not found. Available servers: {available_servers}") |
|
|
|
|
|
matching_server = None |
|
|
for srv_name, srv_config in self.servers.items(): |
|
|
if (srv_config.space_id and server_name in srv_config.space_id) or server_name in srv_name: |
|
|
matching_server = srv_name |
|
|
logger.info(f"π§ Found matching server: {matching_server}") |
|
|
break |
|
|
|
|
|
if matching_server and self.enabled_servers.get(matching_server, True): |
|
|
server_name = matching_server |
|
|
logger.info(f"π§ Using corrected server name: {server_name}") |
|
|
else: |
|
|
|
|
|
error_msg = f"Server '{server_name}' not found or disabled. Available enabled servers: {[name for name, enabled in self.enabled_servers.items() if enabled]}" |
|
|
response._tool_execution = { |
|
|
"server": server_name, |
|
|
"tool": tool_name, |
|
|
"result": error_msg, |
|
|
"success": False |
|
|
} |
|
|
return response |
|
|
elif not self.enabled_servers.get(server_name, True): |
|
|
logger.error(f"β Server '{server_name}' is disabled") |
|
|
response._tool_execution = { |
|
|
"server": server_name, |
|
|
"tool": tool_name, |
|
|
"result": f"Server '{server_name}' is currently disabled", |
|
|
"success": False |
|
|
} |
|
|
return response |
|
|
|
|
|
if server_name in self.server_tools and tool_name not in self.server_tools[server_name]: |
|
|
available_tools = list(self.server_tools[server_name].keys()) |
|
|
logger.warning(f"β οΈ Tool '{tool_name}' not found for server '{server_name}'. Available tools: {available_tools}") |
|
|
|
|
|
|
|
|
if available_tools: |
|
|
|
|
|
if len(available_tools) == 1: |
|
|
tool_name = available_tools[0] |
|
|
logger.info(f"π§ Using only available tool: {tool_name}") |
|
|
|
|
|
else: |
|
|
for available_tool in available_tools: |
|
|
if tool_name.lower() in available_tool.lower() or available_tool.lower() in tool_name.lower(): |
|
|
tool_name = available_tool |
|
|
logger.info(f"π§ Found similar tool name: {tool_name}") |
|
|
break |
|
|
|
|
|
|
|
|
def run_mcp_tool(): |
|
|
loop = asyncio.new_event_loop() |
|
|
asyncio.set_event_loop(loop) |
|
|
try: |
|
|
return loop.run_until_complete( |
|
|
self.call_mcp_tool_async(server_name, tool_name, arguments) |
|
|
) |
|
|
finally: |
|
|
loop.close() |
|
|
|
|
|
success, result = run_mcp_tool() |
|
|
|
|
|
if success: |
|
|
logger.info(f"β
Tool call successful, result length: {len(str(result))}") |
|
|
|
|
|
|
|
|
enhanced_messages.append({"role": "assistant", "content": response_text}) |
|
|
enhanced_messages.append({"role": "user", "content": f"Tool '{tool_name}' from server '{server_name}' completed successfully. Result: {result}\n\nPlease provide a helpful response based on this tool result. If the result contains media URLs, present them appropriately."}) |
|
|
|
|
|
|
|
|
final_messages = enhanced_messages.copy() |
|
|
if final_messages[0].get("role") == "system": |
|
|
final_messages[0]["content"] = final_messages[0]["content"].split("You have access to the following MCP tools:")[0].strip() |
|
|
|
|
|
logger.info("π€ Getting final response with tool result...") |
|
|
final_response = self.generate_chat_completion(final_messages, **{"max_tokens": 4096}) |
|
|
|
|
|
|
|
|
final_response._tool_execution = { |
|
|
"server": server_name, |
|
|
"tool": tool_name, |
|
|
"result": result, |
|
|
"success": True |
|
|
} |
|
|
|
|
|
return final_response |
|
|
else: |
|
|
logger.error(f"β Tool call failed: {result}") |
|
|
|
|
|
response._tool_execution = { |
|
|
"server": server_name, |
|
|
"tool": tool_name, |
|
|
"result": result, |
|
|
"success": False |
|
|
} |
|
|
return response |
|
|
else: |
|
|
logger.warning("β οΈ Failed to parse tool request JSON") |
|
|
else: |
|
|
logger.info("π¬ No tool usage detected, returning normal response") |
|
|
|
|
|
|
|
|
return response |
|
|
|
|
|
def _extract_tool_json(self, text: str) -> Optional[Dict[str, Any]]: |
|
|
"""Extract JSON from LLM response more robustly""" |
|
|
import json |
|
|
import re |
|
|
|
|
|
logger.info(f"π Full LLM response text: {text}") |
|
|
|
|
|
|
|
|
strategies = [ |
|
|
|
|
|
lambda t: re.search(r'\{[^{}]*"use_tool"[^{}]*"arguments"[^{}]*\{[^{}]*\}[^{}]*\}', t), |
|
|
|
|
|
lambda t: self._reconstruct_json_from_start(t), |
|
|
|
|
|
lambda t: re.search(r'\{(?:[^{}]|\{[^{}]*\})*\}', t), |
|
|
] |
|
|
|
|
|
for i, strategy in enumerate(strategies, 1): |
|
|
try: |
|
|
if i == 2: |
|
|
|
|
|
json_str = strategy(text) |
|
|
if not json_str: |
|
|
continue |
|
|
else: |
|
|
match = strategy(text) |
|
|
if not match: |
|
|
continue |
|
|
json_str = match.group(0) |
|
|
|
|
|
logger.info(f"π JSON extraction strategy {i} found: {json_str}") |
|
|
|
|
|
|
|
|
json_str = json_str.strip() |
|
|
|
|
|
|
|
|
parsed = json.loads(json_str) |
|
|
|
|
|
|
|
|
if parsed.get("use_tool") is True: |
|
|
logger.info(f"β
Valid tool request parsed: {parsed}") |
|
|
return parsed |
|
|
|
|
|
except json.JSONDecodeError as e: |
|
|
logger.warning(f"β οΈ JSON parse error with strategy {i}: {e}") |
|
|
logger.warning(f"β οΈ Problematic JSON: {json_str if 'json_str' in locals() else 'N/A'}") |
|
|
continue |
|
|
except Exception as e: |
|
|
logger.warning(f"β οΈ Strategy {i} failed: {e}") |
|
|
continue |
|
|
|
|
|
logger.error("β Failed to extract valid JSON from response") |
|
|
return None |
|
|
|
|
|
def _manual_tool_extraction(self, text: str) -> Optional[Dict[str, Any]]: |
|
|
"""Manually extract tool information as fallback""" |
|
|
import re |
|
|
|
|
|
logger.info("π§ Attempting manual tool extraction...") |
|
|
|
|
|
try: |
|
|
|
|
|
server_match = re.search(r'"server":\s*"([^"]+)"', text) |
|
|
tool_match = re.search(r'"tool":\s*"([^"]+)"', text) |
|
|
|
|
|
if not server_match or not tool_match: |
|
|
logger.warning("β οΈ Could not find server or tool in manual extraction") |
|
|
return None |
|
|
|
|
|
server_name = server_match.group(1) |
|
|
tool_name = tool_match.group(1) |
|
|
|
|
|
|
|
|
args_match = re.search(r'"arguments":\s*\{([^}]+)\}', text) |
|
|
arguments = {} |
|
|
|
|
|
if args_match: |
|
|
args_content = args_match.group(1) |
|
|
|
|
|
pairs = re.findall(r'"([^"]+)":\s*"([^"]+)"', args_content) |
|
|
arguments = dict(pairs) |
|
|
|
|
|
manual_request = { |
|
|
"use_tool": True, |
|
|
"server": server_name, |
|
|
"tool": tool_name, |
|
|
"arguments": arguments |
|
|
} |
|
|
|
|
|
logger.info(f"π§ Manual extraction successful: {manual_request}") |
|
|
return manual_request |
|
|
|
|
|
except Exception as e: |
|
|
logger.error(f"β Manual extraction failed: {e}") |
|
|
return None |
|
|
|
|
|
def _reconstruct_json_from_start(self, text: str) -> Optional[str]: |
|
|
"""Try to reconstruct JSON if it's truncated""" |
|
|
import re |
|
|
|
|
|
|
|
|
match = re.search(r'\{"use_tool":\s*true[^}]*', text) |
|
|
if not match: |
|
|
return None |
|
|
|
|
|
json_start = match.start() |
|
|
json_part = text[json_start:] |
|
|
|
|
|
logger.info(f"π§ Reconstructing JSON from: {json_part[:200]}...") |
|
|
|
|
|
|
|
|
brace_count = 0 |
|
|
end_pos = 0 |
|
|
in_string = False |
|
|
escape_next = False |
|
|
|
|
|
for i, char in enumerate(json_part): |
|
|
if escape_next: |
|
|
escape_next = False |
|
|
continue |
|
|
|
|
|
if char == '\\': |
|
|
escape_next = True |
|
|
continue |
|
|
|
|
|
if char == '"' and not escape_next: |
|
|
in_string = not in_string |
|
|
continue |
|
|
|
|
|
if not in_string: |
|
|
if char == '{': |
|
|
brace_count += 1 |
|
|
elif char == '}': |
|
|
brace_count -= 1 |
|
|
if brace_count == 0: |
|
|
end_pos = i + 1 |
|
|
break |
|
|
|
|
|
if end_pos > 0: |
|
|
reconstructed = json_part[:end_pos] |
|
|
logger.info(f"π§ Reconstructed JSON: {reconstructed}") |
|
|
return reconstructed |
|
|
else: |
|
|
|
|
|
missing_braces = json_part.count('{') - json_part.count('}') |
|
|
if missing_braces > 0: |
|
|
reconstructed = json_part + '}' * missing_braces |
|
|
logger.info(f"π§ Added {missing_braces} closing braces: {reconstructed}") |
|
|
return reconstructed |
|
|
|
|
|
return None |
|
|
|
|
|
def _extract_media_from_mcp_response(self, result_text: str, config: MCPServerConfig) -> Optional[str]: |
|
|
"""Enhanced media extraction from MCP responses with better URL resolution""" |
|
|
if not isinstance(result_text, str): |
|
|
logger.info(f"π Non-string result: {type(result_text)}") |
|
|
return None |
|
|
|
|
|
base_url = config.url.replace("/gradio_api/mcp/sse", "") |
|
|
logger.info(f"π Processing MCP result for media: {result_text[:300]}...") |
|
|
logger.info(f"π Base URL: {base_url}") |
|
|
|
|
|
|
|
|
try: |
|
|
if result_text.strip().startswith('[') or result_text.strip().startswith('{'): |
|
|
logger.info("π Attempting JSON parse...") |
|
|
data = json.loads(result_text.strip()) |
|
|
logger.info(f"π Parsed JSON structure: {data}") |
|
|
|
|
|
|
|
|
if isinstance(data, list) and len(data) > 0: |
|
|
item = data[0] |
|
|
logger.info(f"π First array item: {item}") |
|
|
|
|
|
if isinstance(item, dict): |
|
|
|
|
|
for media_type in ['image', 'audio', 'video']: |
|
|
if media_type in item and isinstance(item[media_type], dict): |
|
|
media_data = item[media_type] |
|
|
if 'url' in media_data: |
|
|
url = media_data['url'].strip('\'"') |
|
|
logger.info(f"π― Found {media_type} URL: {url}") |
|
|
return self._resolve_media_url(url, base_url) |
|
|
|
|
|
|
|
|
if 'url' in item: |
|
|
url = item['url'].strip('\'"') |
|
|
logger.info(f"π― Found direct URL: {url}") |
|
|
return self._resolve_media_url(url, base_url) |
|
|
|
|
|
|
|
|
elif isinstance(data, dict): |
|
|
logger.info(f"π Processing dict: {data}") |
|
|
|
|
|
|
|
|
for media_type in ['image', 'audio', 'video']: |
|
|
if media_type in data and isinstance(data[media_type], dict): |
|
|
media_data = data[media_type] |
|
|
if 'url' in media_data: |
|
|
url = media_data['url'].strip('\'"') |
|
|
logger.info(f"π― Found {media_type} URL: {url}") |
|
|
return self._resolve_media_url(url, base_url) |
|
|
|
|
|
|
|
|
if 'url' in data: |
|
|
url = data['url'].strip('\'"') |
|
|
logger.info(f"π― Found direct URL: {url}") |
|
|
return self._resolve_media_url(url, base_url) |
|
|
|
|
|
except json.JSONDecodeError: |
|
|
logger.info("π Not valid JSON, trying other formats...") |
|
|
except Exception as e: |
|
|
logger.warning(f"π JSON parsing error: {e}") |
|
|
|
|
|
|
|
|
gradio_file_patterns = [ |
|
|
r'https://[^/]+\.hf\.space/gradio_api/file=/[^/]+/[^/]+/[^"\s\',]+', |
|
|
r'https://[^/]+\.hf\.space/file=[^"\s\',]+', |
|
|
r'/gradio_api/file=/[^"\s\',]+' |
|
|
] |
|
|
|
|
|
for pattern in gradio_file_patterns: |
|
|
match = re.search(pattern, result_text) |
|
|
if match: |
|
|
url = match.group(0).rstrip('\'",:;') |
|
|
logger.info(f"π― Found Gradio file URL: {url}") |
|
|
if url.startswith('/'): |
|
|
url = f"{base_url}{url}" |
|
|
return url |
|
|
|
|
|
|
|
|
http_url_pattern = r'https?://[^\s"<>]+' |
|
|
matches = re.findall(http_url_pattern, result_text) |
|
|
for url in matches: |
|
|
if AppConfig.is_media_file(url): |
|
|
logger.info(f"π― Found HTTP media URL: {url}") |
|
|
return url |
|
|
|
|
|
|
|
|
if result_text.startswith('data:'): |
|
|
logger.info("π― Found data URL") |
|
|
return result_text |
|
|
|
|
|
|
|
|
if AppConfig.is_media_file(result_text): |
|
|
|
|
|
if '/' in result_text: |
|
|
filename = result_text.split('/')[-1] |
|
|
else: |
|
|
filename = result_text.strip() |
|
|
|
|
|
|
|
|
media_url = f"{base_url}/file={filename}" |
|
|
logger.info(f"π― Created media URL from filename: {media_url}") |
|
|
return media_url |
|
|
|
|
|
logger.info("β No media detected in result") |
|
|
return None |
|
|
|
|
|
def _resolve_media_url(self, url: str, base_url: str) -> str: |
|
|
"""Resolve relative URLs to absolute URLs with better handling""" |
|
|
if url.startswith('http') or url.startswith('data:'): |
|
|
return url |
|
|
elif url.startswith('/gradio_api/file='): |
|
|
return f"{base_url}{url}" |
|
|
elif url.startswith('/file='): |
|
|
return f"{base_url}/gradio_api{url}" |
|
|
elif url.startswith('file='): |
|
|
return f"{base_url}/gradio_api/{url}" |
|
|
elif url.startswith('/'): |
|
|
return f"{base_url}/file={url}" |
|
|
else: |
|
|
return f"{base_url}/file={url}" |
|
|
|
|
|
def get_server_status(self) -> Dict[str, str]: |
|
|
"""Get status of all configured servers""" |
|
|
status = {} |
|
|
for name in self.servers: |
|
|
compatibility = self._check_file_upload_compatibility(self.servers[name]) |
|
|
status[name] = f"β
Connected (MCP Protocol) - {compatibility}" |
|
|
return status |
|
|
|
|
|
def _check_file_upload_compatibility(self, config: MCPServerConfig) -> str: |
|
|
"""Check if a server likely supports file uploads""" |
|
|
if "hf.space" in config.url: |
|
|
return "π‘ Hugging Face Space (usually compatible)" |
|
|
elif "gradio" in config.url.lower(): |
|
|
return "π’ Gradio server (likely compatible)" |
|
|
elif "localhost" in config.url or "127.0.0.1" in config.url: |
|
|
return "π’ Local server (file access available)" |
|
|
else: |
|
|
return "π΄ Remote server (may need public URLs)" |
|
|
|