Refactor to use LiteLLM Responses API for automatic MCP tool execution
Major refactoring to properly integrate with LiteLLM's Responses API, which handles MCP tool execution automatically instead of requiring manual tool call loops. Key changes: - Switched from chat.completions.create() to client.responses.create() - Use "server_url": "litellm_proxy" to leverage LiteLLM as MCP gateway - Set "require_approval": "never" for fully automatic tool execution - Simplified get_available_mcp_tools() to get_available_mcp_servers() - Removed manual OpenAI tool format conversion (LiteLLM handles this) - Updated response extraction to use output[0].content[0].text format - Convert system prompts to user role for Responses API compatibility Technical improvements: - LiteLLM now handles the complete tool calling loop automatically - No more placeholder responses - actual MCP tools will execute - Cleaner code with ~100 fewer lines - Better separation between tools-enabled and tools-disabled paths - Proper error handling for Responses API format Responses API benefits: - Single API call returns final response with tool results integrated - Automatic tool discovery, execution, and result formatting - No manual tracking of tool_call_ids or conversation state - Native MCP support via server_label configuration Documentation: - Added comprehensive litellm-mcp-research.md with API examples - Documented Responses API vs chat.completions differences - Included Discord bot migration patterns - Covered authentication, streaming, and tool restrictions Next steps: - Test with actual Discord interactions - Verify GitHub MCP tools execute correctly - Monitor response extraction for edge cases 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com>
This commit is contained in:
408
litellm-mcp-research.md
Normal file
408
litellm-mcp-research.md
Normal file
@@ -0,0 +1,408 @@
|
|||||||
|
# LiteLLM Responses API with MCP tool integration
|
||||||
|
|
||||||
|
LiteLLM's `/v1/responses` endpoint enables automatic MCP tool execution through a single API call, eliminating the manual tool-calling loop required with chat.completions. When configured with `"require_approval": "never"`, LiteLLM handles tool discovery, execution, and response integration automatically—making Discord bot migration straightforward. The key differences from chat.completions are the `input` parameter (replacing `messages`) and native MCP tool support via a `"type": "mcp"` tool specification.
|
||||||
|
|
||||||
|
## Request and response format for /v1/responses
|
||||||
|
|
||||||
|
The Responses API (available in LiteLLM **1.63.8+**) uses `input` instead of `messages`. The `input` parameter accepts either a simple string or an array of message objects:
|
||||||
|
|
||||||
|
```python
|
||||||
|
# Simple string input
|
||||||
|
response = client.responses.create(
|
||||||
|
model="anthropic/claude-3-5-sonnet-latest",
|
||||||
|
input="What is the weather today?"
|
||||||
|
)
|
||||||
|
|
||||||
|
# Array format (for multi-turn conversations)
|
||||||
|
response = client.responses.create(
|
||||||
|
model="anthropic/claude-3-5-sonnet-latest",
|
||||||
|
input=[
|
||||||
|
{"role": "user", "content": "Hello"},
|
||||||
|
{"role": "assistant", "content": "Hi there!"},
|
||||||
|
{"role": "user", "content": "Tell me about Python"}
|
||||||
|
]
|
||||||
|
)
|
||||||
|
```
|
||||||
|
|
||||||
|
**Response structure** differs significantly from chat.completions. Instead of `choices[0].message.content`, responses use an `output` array:
|
||||||
|
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"id": "resp_abc123",
|
||||||
|
"object": "response",
|
||||||
|
"created_at": 1734366691,
|
||||||
|
"status": "completed",
|
||||||
|
"model": "claude-3-5-sonnet-latest",
|
||||||
|
"output": [
|
||||||
|
{
|
||||||
|
"type": "message",
|
||||||
|
"id": "msg_abc123",
|
||||||
|
"status": "completed",
|
||||||
|
"role": "assistant",
|
||||||
|
"content": [
|
||||||
|
{
|
||||||
|
"type": "output_text",
|
||||||
|
"text": "Here is the response text...",
|
||||||
|
"annotations": []
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"usage": {"input_tokens": 18, "output_tokens": 98, "total_tokens": 116}
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
To extract text: `response.output[0].content[0].text`
|
||||||
|
|
||||||
|
## MCP tool specification format
|
||||||
|
|
||||||
|
MCP tools use `"type": "mcp"` with three critical parameters: `server_label`, `server_url`, and `require_approval`. The special value `"server_url": "litellm_proxy"` tells LiteLLM to act as an MCP gateway, handling all tool execution internally:
|
||||||
|
|
||||||
|
```python
|
||||||
|
tools=[
|
||||||
|
{
|
||||||
|
"type": "mcp",
|
||||||
|
"server_label": "my_mcp_server", # Identifier for the MCP server
|
||||||
|
"server_url": "litellm_proxy", # LiteLLM handles MCP bridging
|
||||||
|
"require_approval": "never", # Automatic execution
|
||||||
|
"allowed_tools": ["tool1", "tool2"] # Optional: restrict available tools
|
||||||
|
}
|
||||||
|
]
|
||||||
|
```
|
||||||
|
|
||||||
|
| Parameter | Purpose |
|
||||||
|
|-----------|---------|
|
||||||
|
| `server_label` | Identifies which configured MCP server to use (must match config.yaml) |
|
||||||
|
| `server_url` | `"litellm_proxy"` for LiteLLM gateway, or direct URL like `"https://mcp.example.com/mcp"` |
|
||||||
|
| `require_approval` | `"never"` for automatic execution; omit for approval-based flow |
|
||||||
|
| `allowed_tools` | Whitelist of tool names to make available |
|
||||||
|
|
||||||
|
When `server_url="litellm_proxy"`, LiteLLM performs a **four-step automatic flow**: (1) fetches MCP tools and converts to OpenAI format, (2) sends tools to the LLM with your input, (3) executes any tool calls against MCP servers, and (4) returns the final response with tool results integrated.
|
||||||
|
|
||||||
|
## Streaming versus non-streaming responses
|
||||||
|
|
||||||
|
For **non-streaming**, pass `stream=False` (default) and receive the complete response object:
|
||||||
|
|
||||||
|
```python
|
||||||
|
response = client.responses.create(
|
||||||
|
model="gpt-4o",
|
||||||
|
input="Hello",
|
||||||
|
stream=False
|
||||||
|
)
|
||||||
|
text = response.output[0].content[0].text
|
||||||
|
```
|
||||||
|
|
||||||
|
For **streaming**, set `stream=True` and iterate over events:
|
||||||
|
|
||||||
|
```python
|
||||||
|
stream = client.responses.create(
|
||||||
|
model="gpt-4o",
|
||||||
|
input="Write a poem",
|
||||||
|
stream=True
|
||||||
|
)
|
||||||
|
|
||||||
|
full_text = ""
|
||||||
|
for event in stream:
|
||||||
|
if hasattr(event, 'type'):
|
||||||
|
if event.type == "response.output_text.delta":
|
||||||
|
print(event.delta, end="", flush=True)
|
||||||
|
full_text += event.delta
|
||||||
|
elif event.type == "response.completed":
|
||||||
|
print("\n--- Done ---")
|
||||||
|
```
|
||||||
|
|
||||||
|
Key streaming event types include `response.created`, `response.output_text.delta` (incremental text), `response.output_text.done`, and `response.completed`.
|
||||||
|
|
||||||
|
## Python SDK differences between responses.create() and chat.completions.create()
|
||||||
|
|
||||||
|
| Aspect | `responses.create()` | `chat.completions.create()` |
|
||||||
|
|--------|---------------------|---------------------------|
|
||||||
|
| Input parameter | `input` (string or array) | `messages` (array required) |
|
||||||
|
| Response access | `response.output[0].content[0].text` | `response.choices[0].message.content` |
|
||||||
|
| Conversation history | Built-in via `previous_response_id` | Manual message array management |
|
||||||
|
| MCP tools | Native `"type": "mcp"` support | Standard function calling only |
|
||||||
|
| Endpoint | `/v1/responses` | `/v1/chat/completions` |
|
||||||
|
|
||||||
|
**Client setup** is identical for both APIs:
|
||||||
|
|
||||||
|
```python
|
||||||
|
from openai import OpenAI
|
||||||
|
|
||||||
|
client = OpenAI(
|
||||||
|
base_url="http://localhost:4000", # Your LiteLLM proxy
|
||||||
|
api_key="sk-your-litellm-key"
|
||||||
|
)
|
||||||
|
|
||||||
|
# Responses API
|
||||||
|
response = client.responses.create(model="gpt-4o", input="Hello")
|
||||||
|
|
||||||
|
# Chat Completions API (old way)
|
||||||
|
response = client.chat.completions.create(
|
||||||
|
model="gpt-4o",
|
||||||
|
messages=[{"role": "user", "content": "Hello"}]
|
||||||
|
)
|
||||||
|
```
|
||||||
|
|
||||||
|
## Conversation history with the input parameter
|
||||||
|
|
||||||
|
Unlike chat.completions where you manually pass the full message history each time, the Responses API offers two approaches:
|
||||||
|
|
||||||
|
**Option 1: Use `previous_response_id`** for automatic context (recommended):
|
||||||
|
```python
|
||||||
|
# First message
|
||||||
|
response1 = client.responses.create(model="gpt-4o", input="My name is Alice")
|
||||||
|
|
||||||
|
# Follow-up with context preserved automatically
|
||||||
|
response2 = client.responses.create(
|
||||||
|
model="gpt-4o",
|
||||||
|
input="What's my name?",
|
||||||
|
previous_response_id=response1.id # LiteLLM maintains context
|
||||||
|
)
|
||||||
|
```
|
||||||
|
|
||||||
|
**Option 2: Pass full history in input array** (manual approach):
|
||||||
|
```python
|
||||||
|
response = client.responses.create(
|
||||||
|
model="gpt-4o",
|
||||||
|
input=[
|
||||||
|
{"role": "user", "content": "My name is Alice"},
|
||||||
|
{"role": "assistant", "content": "Nice to meet you, Alice!"},
|
||||||
|
{"role": "user", "content": "What's my name?"}
|
||||||
|
]
|
||||||
|
)
|
||||||
|
```
|
||||||
|
|
||||||
|
The `input` array supports roles: `user`, `assistant`, `developer` (replaces `system` in newer models), and `tool`.
|
||||||
|
|
||||||
|
## The require_approval parameter and MCP options
|
||||||
|
|
||||||
|
**`require_approval: "never"`** enables fully automatic tool execution—LiteLLM returns the final response in a single API call:
|
||||||
|
|
||||||
|
```python
|
||||||
|
response = client.responses.create(
|
||||||
|
model="gpt-4o",
|
||||||
|
input="Search for Python documentation",
|
||||||
|
tools=[{
|
||||||
|
"type": "mcp",
|
||||||
|
"server_label": "search_server",
|
||||||
|
"server_url": "litellm_proxy",
|
||||||
|
"require_approval": "never" # No approval needed
|
||||||
|
}]
|
||||||
|
)
|
||||||
|
# Response includes tool results integrated into final answer
|
||||||
|
```
|
||||||
|
|
||||||
|
**Without `require_approval: "never"`**, you get an approval flow requiring two API calls:
|
||||||
|
|
||||||
|
```python
|
||||||
|
# Step 1: Get approval request
|
||||||
|
response = client.responses.create(
|
||||||
|
model="gpt-4o",
|
||||||
|
input="Search for docs",
|
||||||
|
tools=[{"type": "mcp", "server_label": "search", "server_url": "litellm_proxy"}]
|
||||||
|
)
|
||||||
|
|
||||||
|
# Extract approval request ID from response.output
|
||||||
|
approval_id = None
|
||||||
|
for output in response.output:
|
||||||
|
if output.type == "mcp_approval_request":
|
||||||
|
approval_id = output.id
|
||||||
|
break
|
||||||
|
|
||||||
|
# Step 2: Approve and get final response
|
||||||
|
final_response = client.responses.create(
|
||||||
|
model="gpt-4o",
|
||||||
|
input=[{"type": "mcp_approval_response", "approve": True, "approval_request_id": approval_id}],
|
||||||
|
previous_response_id=response.id,
|
||||||
|
tools=[{"type": "mcp", "server_label": "search", "server_url": "litellm_proxy"}]
|
||||||
|
)
|
||||||
|
```
|
||||||
|
|
||||||
|
## Restricting tools with allowed_tools
|
||||||
|
|
||||||
|
Control which MCP tools are available at **request time** or **server configuration level**:
|
||||||
|
|
||||||
|
**Request-level restriction** (per-call):
|
||||||
|
```python
|
||||||
|
tools=[{
|
||||||
|
"type": "mcp",
|
||||||
|
"server_label": "github_mcp",
|
||||||
|
"server_url": "litellm_proxy",
|
||||||
|
"require_approval": "never",
|
||||||
|
"allowed_tools": ["list_repos", "get_file_contents"] # Only these tools available
|
||||||
|
}]
|
||||||
|
```
|
||||||
|
|
||||||
|
**Server-level restriction** (in config.yaml):
|
||||||
|
```yaml
|
||||||
|
mcp_servers:
|
||||||
|
github_mcp:
|
||||||
|
url: "https://api.github.com/mcp"
|
||||||
|
allowed_tools: ["list_repos", "get_file_contents"] # Whitelist
|
||||||
|
disallowed_tools: ["delete_repo", "force_push"] # Blacklist
|
||||||
|
```
|
||||||
|
|
||||||
|
If both `allowed_tools` and `disallowed_tools` are specified, `allowed_tools` takes priority.
|
||||||
|
|
||||||
|
## Authentication headers
|
||||||
|
|
||||||
|
LiteLLM supports multiple authentication header formats:
|
||||||
|
|
||||||
|
| Header | Use Case |
|
||||||
|
|--------|----------|
|
||||||
|
| `Authorization: Bearer sk-...` | **Standard** - Used by OpenAI SDK automatically |
|
||||||
|
| `x-litellm-api-key: Bearer sk-...` | **MCP connections** and custom scenarios |
|
||||||
|
| `api-key: ...` | Azure OpenAI compatibility |
|
||||||
|
|
||||||
|
**For standard API calls** (Discord bot), use the OpenAI SDK default:
|
||||||
|
```python
|
||||||
|
client = OpenAI(
|
||||||
|
base_url="http://localhost:4000",
|
||||||
|
api_key="sk-your-key" # Sent as "Authorization: Bearer sk-your-key"
|
||||||
|
)
|
||||||
|
```
|
||||||
|
|
||||||
|
**For MCP tool headers** (when calling external MCP servers), use the `headers` parameter:
|
||||||
|
```python
|
||||||
|
tools=[{
|
||||||
|
"type": "mcp",
|
||||||
|
"server_label": "github",
|
||||||
|
"server_url": "litellm_proxy",
|
||||||
|
"require_approval": "never",
|
||||||
|
"headers": {
|
||||||
|
"x-litellm-api-key": "Bearer sk-your-litellm-key",
|
||||||
|
"x-mcp-github-authorization": "Bearer ghp_your_github_token"
|
||||||
|
}
|
||||||
|
}]
|
||||||
|
```
|
||||||
|
|
||||||
|
## Complete Discord bot migration example
|
||||||
|
|
||||||
|
Here's a full implementation pattern for migrating from chat.completions to responses with MCP:
|
||||||
|
|
||||||
|
```python
|
||||||
|
from openai import OpenAI
|
||||||
|
import os
|
||||||
|
|
||||||
|
class LiteLLMResponsesClient:
|
||||||
|
"""Client wrapper for Discord bot using LiteLLM Responses API with MCP."""
|
||||||
|
|
||||||
|
def __init__(self, proxy_url: str, api_key: str):
|
||||||
|
self.client = OpenAI(base_url=proxy_url, api_key=api_key)
|
||||||
|
self.conversations = {} # user_id -> response_id mapping
|
||||||
|
|
||||||
|
def get_mcp_tools(self, server_label: str = "default") -> list:
|
||||||
|
"""Define MCP tools configuration."""
|
||||||
|
return [{
|
||||||
|
"type": "mcp",
|
||||||
|
"server_label": server_label,
|
||||||
|
"server_url": "litellm_proxy",
|
||||||
|
"require_approval": "never",
|
||||||
|
"allowed_tools": ["search", "fetch_data", "analyze"] # Customize as needed
|
||||||
|
}]
|
||||||
|
|
||||||
|
def chat(
|
||||||
|
self,
|
||||||
|
user_id: str,
|
||||||
|
message: str,
|
||||||
|
model: str = "anthropic/claude-3-5-sonnet-latest",
|
||||||
|
use_mcp_tools: bool = True,
|
||||||
|
stream: bool = False
|
||||||
|
):
|
||||||
|
"""Send a message and get response, with optional MCP tools and streaming."""
|
||||||
|
|
||||||
|
previous_id = self.conversations.get(user_id)
|
||||||
|
|
||||||
|
kwargs = {
|
||||||
|
"model": model,
|
||||||
|
"input": message,
|
||||||
|
"stream": stream
|
||||||
|
}
|
||||||
|
|
||||||
|
if previous_id:
|
||||||
|
kwargs["previous_response_id"] = previous_id
|
||||||
|
|
||||||
|
if use_mcp_tools:
|
||||||
|
kwargs["tools"] = self.get_mcp_tools()
|
||||||
|
kwargs["tool_choice"] = "auto"
|
||||||
|
|
||||||
|
if stream:
|
||||||
|
return self._handle_stream(user_id, **kwargs)
|
||||||
|
else:
|
||||||
|
response = self.client.responses.create(**kwargs)
|
||||||
|
self.conversations[user_id] = response.id
|
||||||
|
return self._extract_text(response)
|
||||||
|
|
||||||
|
def _handle_stream(self, user_id: str, **kwargs):
|
||||||
|
"""Generator for streaming responses."""
|
||||||
|
stream = self.client.responses.create(**kwargs)
|
||||||
|
response_id = None
|
||||||
|
|
||||||
|
for event in stream:
|
||||||
|
if hasattr(event, 'type'):
|
||||||
|
if event.type == "response.created":
|
||||||
|
response_id = event.response.id
|
||||||
|
elif event.type == "response.output_text.delta":
|
||||||
|
yield event.delta
|
||||||
|
|
||||||
|
if response_id:
|
||||||
|
self.conversations[user_id] = response_id
|
||||||
|
|
||||||
|
def _extract_text(self, response) -> str:
|
||||||
|
"""Extract text from Responses API response."""
|
||||||
|
for output in response.output:
|
||||||
|
if output.type == "message":
|
||||||
|
for content in output.content:
|
||||||
|
if content.type == "output_text":
|
||||||
|
return content.text
|
||||||
|
return ""
|
||||||
|
|
||||||
|
def clear_history(self, user_id: str):
|
||||||
|
"""Clear conversation history for a user."""
|
||||||
|
self.conversations.pop(user_id, None)
|
||||||
|
|
||||||
|
|
||||||
|
# Discord bot integration example
|
||||||
|
import discord
|
||||||
|
|
||||||
|
bot = discord.Bot()
|
||||||
|
llm_client = LiteLLMResponsesClient(
|
||||||
|
proxy_url=os.environ["LITELLM_PROXY_URL"],
|
||||||
|
api_key=os.environ["LITELLM_API_KEY"]
|
||||||
|
)
|
||||||
|
|
||||||
|
@bot.event
|
||||||
|
async def on_message(message):
|
||||||
|
if message.author.bot:
|
||||||
|
return
|
||||||
|
|
||||||
|
if bot.user.mentioned_in(message):
|
||||||
|
user_id = str(message.author.id)
|
||||||
|
user_message = message.content.replace(f'<@{bot.user.id}>', '').strip()
|
||||||
|
|
||||||
|
# Non-streaming response with MCP tools
|
||||||
|
response_text = llm_client.chat(
|
||||||
|
user_id=user_id,
|
||||||
|
message=user_message,
|
||||||
|
use_mcp_tools=True
|
||||||
|
)
|
||||||
|
await message.reply(response_text)
|
||||||
|
|
||||||
|
# Run: bot.run(os.environ["DISCORD_TOKEN"])
|
||||||
|
```
|
||||||
|
|
||||||
|
## Official documentation links
|
||||||
|
|
||||||
|
- **Responses API documentation**: https://docs.litellm.ai/docs/response_api
|
||||||
|
- **MCP overview**: https://docs.litellm.ai/docs/mcp
|
||||||
|
- **MCP usage guide**: https://docs.litellm.ai/docs/mcp_usage
|
||||||
|
- **MCP permission management**: https://docs.litellm.ai/docs/mcp_control
|
||||||
|
- **OpenAI provider Responses API**: https://docs.litellm.ai/docs/providers/openai/responses_api
|
||||||
|
- **Streaming documentation**: https://docs.litellm.ai/docs/completion/stream
|
||||||
|
- **Virtual keys and auth**: https://docs.litellm.ai/docs/proxy/virtual_keys
|
||||||
|
|
||||||
|
## Key migration considerations
|
||||||
|
|
||||||
|
The Responses API is marked as **BETA** in LiteLLM. Ensure you're running LiteLLM **1.63.8+** and using OpenAI SDK **1.66.1+** for full compatibility. Model names must include the provider prefix (e.g., `openai/gpt-4o`, `anthropic/claude-3-5-sonnet-latest`). Response IDs are encrypted per-user by default for security—users cannot access other users' conversation history unless you disable this with `disable_responses_id_security: true` in config.yaml.
|
||||||
|
|
||||||
|
The primary advantage for Discord bots is the automatic MCP tool execution loop. With chat.completions, you must manually detect tool calls, execute them, and send results back. With Responses API and `require_approval: "never"`, LiteLLM handles this entire flow internally, returning the final integrated response in a single call.
|
||||||
@@ -80,8 +80,8 @@ async def download_image(url: str) -> str | None:
|
|||||||
print(f"Error downloading image from {url}: {e}")
|
print(f"Error downloading image from {url}: {e}")
|
||||||
return None
|
return None
|
||||||
|
|
||||||
async def get_available_mcp_tools():
|
async def get_available_mcp_servers():
|
||||||
"""Query LiteLLM for available MCP servers and tools, convert to OpenAI format"""
|
"""Query LiteLLM for available MCP servers (used with Responses API)"""
|
||||||
try:
|
try:
|
||||||
base_url = LITELLM_API_BASE.rstrip('/')
|
base_url = LITELLM_API_BASE.rstrip('/')
|
||||||
headers = {"x-litellm-api-key": LITELLM_API_KEY}
|
headers = {"x-litellm-api-key": LITELLM_API_KEY}
|
||||||
@@ -95,50 +95,20 @@ async def get_available_mcp_tools():
|
|||||||
|
|
||||||
if server_response.status_code == 200:
|
if server_response.status_code == 200:
|
||||||
server_info = server_response.json()
|
server_info = server_response.json()
|
||||||
debug_log(f"MCP server info: found {len(server_info) if isinstance(server_info, list) else 0} servers")
|
server_count = len(server_info) if isinstance(server_info, list) else 0
|
||||||
|
debug_log(f"MCP server info: found {server_count} servers")
|
||||||
|
|
||||||
# Get available MCP tools
|
if server_count > 0:
|
||||||
tools_response = await http_client.get(
|
# Log server names for visibility
|
||||||
f"{base_url}/v1/mcp/tools",
|
server_names = [s.get("server_name") for s in server_info if isinstance(s, dict) and s.get("server_name")]
|
||||||
headers=headers
|
debug_log(f"Available MCP servers: {server_names}")
|
||||||
)
|
|
||||||
|
|
||||||
if tools_response.status_code == 200:
|
return {"server": server_info}
|
||||||
tools_data = tools_response.json()
|
|
||||||
|
|
||||||
# Tools come in format: {"tools": [...]}
|
|
||||||
mcp_tools = tools_data.get("tools", []) if isinstance(tools_data, dict) else tools_data
|
|
||||||
debug_log(f"Found {len(mcp_tools) if isinstance(mcp_tools, list) else 0} MCP tools")
|
|
||||||
|
|
||||||
# Convert MCP tools to OpenAI function calling format
|
|
||||||
openai_tools = []
|
|
||||||
for tool in mcp_tools[:50]: # Limit to first 50 tools to avoid overwhelming the model
|
|
||||||
if isinstance(tool, dict) and "name" in tool:
|
|
||||||
openai_tool = {
|
|
||||||
"type": "function",
|
|
||||||
"function": {
|
|
||||||
"name": tool["name"],
|
|
||||||
"description": tool.get("description", ""),
|
|
||||||
"parameters": tool.get("inputSchema", {})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
openai_tools.append(openai_tool)
|
|
||||||
|
|
||||||
debug_log(f"Converted {len(openai_tools)} tools to OpenAI format")
|
|
||||||
|
|
||||||
# Return both server info and converted tools
|
|
||||||
return {
|
|
||||||
"server": server_info,
|
|
||||||
"tools": openai_tools,
|
|
||||||
"tool_count": len(openai_tools)
|
|
||||||
}
|
|
||||||
else:
|
|
||||||
debug_log(f"MCP tools endpoint returned {tools_response.status_code}: {tools_response.text}")
|
|
||||||
else:
|
else:
|
||||||
debug_log(f"MCP server endpoint returned {server_response.status_code}: {server_response.text}")
|
debug_log(f"MCP server endpoint returned {server_response.status_code}: {server_response.text}")
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
debug_log(f"Error fetching MCP tools: {e}")
|
debug_log(f"Error fetching MCP servers: {e}")
|
||||||
|
|
||||||
return None
|
return None
|
||||||
|
|
||||||
@@ -286,7 +256,7 @@ async def get_chat_history(channel, bot_user_id: int, limit: int = 50) -> List[D
|
|||||||
|
|
||||||
async def get_ai_response(history_messages: List[Dict[str, Any]], user_message: str, image_urls: List[str] = None) -> str:
|
async def get_ai_response(history_messages: List[Dict[str, Any]], user_message: str, image_urls: List[str] = None) -> str:
|
||||||
"""
|
"""
|
||||||
Get AI response using LiteLLM with proper conversation history and tool calling support.
|
Get AI response using LiteLLM Responses API with automatic MCP tool execution.
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
history_messages: List of previous conversation messages with roles
|
history_messages: List of previous conversation messages with roles
|
||||||
@@ -296,89 +266,114 @@ async def get_ai_response(history_messages: List[Dict[str, Any]], user_message:
|
|||||||
Returns:
|
Returns:
|
||||||
AI response string
|
AI response string
|
||||||
"""
|
"""
|
||||||
# Start with system prompt
|
|
||||||
messages = [{"role": "system", "content": SYSTEM_PROMPT}]
|
|
||||||
|
|
||||||
# Add conversation history
|
|
||||||
messages.extend(history_messages)
|
|
||||||
|
|
||||||
# Build current user message
|
|
||||||
if image_urls:
|
|
||||||
# Multi-modal message with text and images
|
|
||||||
content_parts = [{"type": "text", "text": user_message}]
|
|
||||||
|
|
||||||
for url in image_urls:
|
|
||||||
base64_image = await download_image(url)
|
|
||||||
if base64_image:
|
|
||||||
content_parts.append({
|
|
||||||
"type": "image_url",
|
|
||||||
"image_url": {
|
|
||||||
"url": f"data:image/jpeg;base64,{base64_image}"
|
|
||||||
}
|
|
||||||
})
|
|
||||||
messages.append({"role": "user", "content": content_parts})
|
|
||||||
else:
|
|
||||||
# Text-only message
|
|
||||||
messages.append({"role": "user", "content": user_message})
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
# Build request parameters
|
# When tools are enabled, use Responses API with MCP for automatic tool execution
|
||||||
request_params = {
|
|
||||||
"model": MODEL_NAME,
|
|
||||||
"messages": messages,
|
|
||||||
"temperature": 0.7,
|
|
||||||
}
|
|
||||||
|
|
||||||
# Add MCP tools if enabled
|
|
||||||
if ENABLE_TOOLS:
|
if ENABLE_TOOLS:
|
||||||
debug_log("Tools enabled - fetching and converting MCP tools")
|
debug_log("Tools enabled - using Responses API with MCP auto-execution")
|
||||||
|
|
||||||
# Query and convert MCP tools to OpenAI format
|
# Query MCP server info to get server_label
|
||||||
mcp_info = await get_available_mcp_tools()
|
mcp_info = await get_available_mcp_servers()
|
||||||
if mcp_info and isinstance(mcp_info, dict):
|
|
||||||
openai_tools = mcp_info.get("tools", [])
|
|
||||||
if openai_tools and isinstance(openai_tools, list) and len(openai_tools) > 0:
|
|
||||||
request_params["tools"] = openai_tools
|
|
||||||
request_params["tool_choice"] = "auto"
|
|
||||||
debug_log(f"Added {len(openai_tools)} tools to request")
|
|
||||||
else:
|
|
||||||
debug_log("No tools available to add to request")
|
|
||||||
else:
|
|
||||||
debug_log("Failed to fetch MCP tools")
|
|
||||||
|
|
||||||
debug_log(f"Calling chat completions with {len(request_params.get('tools', []))} tools")
|
# Build input array with system prompt, history, and current message
|
||||||
response = client.chat.completions.create(**request_params)
|
input_messages = []
|
||||||
|
|
||||||
# Handle tool calls if present
|
# Add system prompt as developer role (newer models) or user role
|
||||||
response_message = response.choices[0].message
|
input_messages.append({
|
||||||
tool_calls = getattr(response_message, 'tool_calls', None)
|
"role": "user", # System messages converted to user for Responses API
|
||||||
|
"content": f"[System Instructions]\n{SYSTEM_PROMPT}"
|
||||||
|
})
|
||||||
|
|
||||||
if tool_calls and len(tool_calls) > 0:
|
# Add conversation history
|
||||||
debug_log(f"Model requested {len(tool_calls)} tool calls")
|
for msg in history_messages:
|
||||||
|
input_messages.append({
|
||||||
# Add assistant's response with tool calls to messages
|
"role": msg["role"],
|
||||||
messages.append(response_message)
|
"content": msg["content"]
|
||||||
|
|
||||||
# Execute each tool call - add placeholder responses
|
|
||||||
# TODO: Implement actual MCP tool execution via LiteLLM proxy
|
|
||||||
for tool_call in tool_calls:
|
|
||||||
function_name = tool_call.function.name
|
|
||||||
function_args = tool_call.function.arguments
|
|
||||||
|
|
||||||
debug_log(f"Tool call requested: {function_name} with args: {function_args}")
|
|
||||||
|
|
||||||
# Placeholder response - in production this would execute via MCP
|
|
||||||
messages.append({
|
|
||||||
"role": "tool",
|
|
||||||
"tool_call_id": tool_call.id,
|
|
||||||
"name": function_name,
|
|
||||||
"content": f"Tool execution via MCP is being set up. Tool {function_name} was called with arguments: {function_args}"
|
|
||||||
})
|
})
|
||||||
|
|
||||||
# Get final response from model after tool execution
|
# Build current user message
|
||||||
debug_log("Getting final response after tool execution")
|
if image_urls:
|
||||||
final_response = client.chat.completions.create(**request_params)
|
# Multi-modal message with text and images
|
||||||
return final_response.choices[0].message.content
|
content_parts = [{"type": "text", "text": user_message}]
|
||||||
|
for url in image_urls:
|
||||||
|
base64_image = await download_image(url)
|
||||||
|
if base64_image:
|
||||||
|
content_parts.append({
|
||||||
|
"type": "image_url",
|
||||||
|
"image_url": {
|
||||||
|
"url": f"data:image/jpeg;base64,{base64_image}"
|
||||||
|
}
|
||||||
|
})
|
||||||
|
input_messages.append({"role": "user", "content": content_parts})
|
||||||
|
else:
|
||||||
|
input_messages.append({"role": "user", "content": user_message})
|
||||||
|
|
||||||
|
# Build MCP tools configuration
|
||||||
|
tools_config = []
|
||||||
|
if mcp_info and isinstance(mcp_info, dict):
|
||||||
|
server_list = mcp_info.get("server", [])
|
||||||
|
if isinstance(server_list, list) and len(server_list) > 0:
|
||||||
|
for server_info in server_list:
|
||||||
|
server_name = server_info.get("server_name")
|
||||||
|
if server_name:
|
||||||
|
tools_config.append({
|
||||||
|
"type": "mcp",
|
||||||
|
"server_label": server_name,
|
||||||
|
"server_url": "litellm_proxy", # Use LiteLLM as MCP gateway
|
||||||
|
"require_approval": "never" # Automatic tool execution
|
||||||
|
})
|
||||||
|
debug_log(f"Added MCP server '{server_name}' with auto-execution")
|
||||||
|
|
||||||
|
if not tools_config:
|
||||||
|
debug_log("No MCP servers found, falling back to standard chat completions")
|
||||||
|
# Fall through to standard chat completions below
|
||||||
|
else:
|
||||||
|
# Use Responses API with MCP tools
|
||||||
|
debug_log(f"Calling Responses API with {len(tools_config)} MCP servers")
|
||||||
|
response = client.responses.create(
|
||||||
|
model=MODEL_NAME,
|
||||||
|
input=input_messages,
|
||||||
|
tools=tools_config,
|
||||||
|
stream=False
|
||||||
|
)
|
||||||
|
|
||||||
|
debug_log(f"Response status: {response.status}")
|
||||||
|
|
||||||
|
# Extract text from Responses API format
|
||||||
|
if hasattr(response, 'output') and len(response.output) > 0:
|
||||||
|
for output in response.output:
|
||||||
|
if hasattr(output, 'type') and output.type == "message":
|
||||||
|
if hasattr(output, 'content') and len(output.content) > 0:
|
||||||
|
for content in output.content:
|
||||||
|
if hasattr(content, 'type') and content.type == "output_text":
|
||||||
|
return content.text
|
||||||
|
|
||||||
|
debug_log(f"Unexpected response format: {response}")
|
||||||
|
return "I received a response but couldn't extract the text. Please try again."
|
||||||
|
|
||||||
|
# Standard chat completions (when tools disabled or MCP not available)
|
||||||
|
debug_log("Using standard chat completions")
|
||||||
|
|
||||||
|
messages = [{"role": "system", "content": SYSTEM_PROMPT}]
|
||||||
|
messages.extend(history_messages)
|
||||||
|
|
||||||
|
if image_urls:
|
||||||
|
content_parts = [{"type": "text", "text": user_message}]
|
||||||
|
for url in image_urls:
|
||||||
|
base64_image = await download_image(url)
|
||||||
|
if base64_image:
|
||||||
|
content_parts.append({
|
||||||
|
"type": "image_url",
|
||||||
|
"image_url": {"url": f"data:image/jpeg;base64,{base64_image}"}
|
||||||
|
})
|
||||||
|
messages.append({"role": "user", "content": content_parts})
|
||||||
|
else:
|
||||||
|
messages.append({"role": "user", "content": user_message})
|
||||||
|
|
||||||
|
response = client.chat.completions.create(
|
||||||
|
model=MODEL_NAME,
|
||||||
|
messages=messages,
|
||||||
|
temperature=0.7
|
||||||
|
)
|
||||||
|
|
||||||
return response.choices[0].message.content
|
return response.choices[0].message.content
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user