mirror of
https://github.com/eigent-ai/eigent.git
synced 2026-04-30 20:50:02 +00:00
175 lines
6.6 KiB
Python
175 lines
6.6 KiB
Python
from fastapi import APIRouter, Depends
|
|
from exa_py import Exa
|
|
from app.component.auth import key_must
|
|
from app.component.environment import env_not_empty
|
|
from app.model.mcp.proxy import ExaSearch
|
|
from typing import Any, cast
|
|
import requests
|
|
import traceroot
|
|
|
|
logger = traceroot.get_logger("server_proxy_controller")
|
|
|
|
from app.model.user.key import Key
|
|
|
|
|
|
router = APIRouter(prefix="/proxy", tags=["Mcp Servers"])
|
|
|
|
|
|
@router.post("/exa")
|
|
def exa_search(search: ExaSearch, key: Key = Depends(key_must)):
|
|
EXA_API_KEY = env_not_empty("EXA_API_KEY")
|
|
try:
|
|
exa = Exa(EXA_API_KEY)
|
|
|
|
if search.num_results is not None and not 0 < search.num_results <= 100:
|
|
raise ValueError("num_results must be between 1 and 100")
|
|
|
|
if search.include_text is not None:
|
|
if len(search.include_text) > 1:
|
|
raise ValueError("include_text can only contain 1 string")
|
|
if len(search.include_text[0].split()) > 5:
|
|
raise ValueError("include_text string cannot be longer than 5 words")
|
|
|
|
if search.exclude_text is not None:
|
|
if len(search.exclude_text) > 1:
|
|
raise ValueError("exclude_text can only contain 1 string")
|
|
if len(search.exclude_text[0].split()) > 5:
|
|
raise ValueError("exclude_text string cannot be longer than 5 words")
|
|
|
|
# Call Exa API with direct parameters
|
|
if search.text:
|
|
results = cast(
|
|
dict[str, Any],
|
|
exa.search_and_contents(
|
|
query=search.query,
|
|
type=search.search_type,
|
|
category=search.category,
|
|
num_results=search.num_results,
|
|
include_text=search.include_text,
|
|
exclude_text=search.exclude_text,
|
|
use_autoprompt=search.use_autoprompt,
|
|
text=True,
|
|
),
|
|
)
|
|
else:
|
|
results = cast(
|
|
dict[str, Any],
|
|
exa.search(
|
|
query=search.query,
|
|
type=search.search_type,
|
|
category=search.category,
|
|
num_results=search.num_results,
|
|
include_text=search.include_text,
|
|
exclude_text=search.exclude_text,
|
|
use_autoprompt=search.use_autoprompt,
|
|
),
|
|
)
|
|
|
|
return results
|
|
|
|
except Exception as e:
|
|
return {"error": f"Exa search failed: {e!s}"}
|
|
|
|
|
|
@router.get("/google")
|
|
def google_search(query: str, search_type: str = "web", key: Key = Depends(key_must)):
|
|
# https://developers.google.com/custom-search/v1/overview
|
|
GOOGLE_API_KEY = env_not_empty("GOOGLE_API_KEY")
|
|
# https://cse.google.com/cse/all
|
|
SEARCH_ENGINE_ID = env_not_empty("SEARCH_ENGINE_ID")
|
|
|
|
# Using the first page
|
|
start_page_idx = 1
|
|
# Different language may get different result
|
|
search_language = "en"
|
|
# How many pages to return
|
|
num_result_pages = 10
|
|
# Constructing the URL
|
|
# Doc: https://developers.google.com/custom-search/v1/using_rest
|
|
base_url = (
|
|
f"https://www.googleapis.com/customsearch/v1?"
|
|
f"key={GOOGLE_API_KEY}&cx={SEARCH_ENGINE_ID}&q={query}&start="
|
|
f"{start_page_idx}&lr={search_language}&num={num_result_pages}"
|
|
)
|
|
|
|
if search_type == "image":
|
|
url = base_url + "&searchType=image"
|
|
else:
|
|
url = base_url
|
|
|
|
responses = []
|
|
# Fetch the results given the URL
|
|
try:
|
|
# Make the get
|
|
result = requests.get(url)
|
|
data = result.json()
|
|
|
|
# Get the result items
|
|
if "items" in data:
|
|
search_items = data.get("items")
|
|
|
|
# Iterate over results found
|
|
for i, search_item in enumerate(search_items, start=1):
|
|
if search_type == "image":
|
|
# Process image search results
|
|
title = search_item.get("title")
|
|
image_url = search_item.get("link")
|
|
display_link = search_item.get("displayLink")
|
|
|
|
# Get context URL (page containing the image)
|
|
image_info = search_item.get("image", {})
|
|
context_url = image_info.get("contextLink", "")
|
|
|
|
# Get image dimensions if available
|
|
width = image_info.get("width")
|
|
height = image_info.get("height")
|
|
|
|
response = {
|
|
"result_id": i,
|
|
"title": title,
|
|
"image_url": image_url,
|
|
"display_link": display_link,
|
|
"context_url": context_url,
|
|
}
|
|
|
|
# Add dimensions if available
|
|
if width:
|
|
response["width"] = int(width)
|
|
if height:
|
|
response["height"] = int(height)
|
|
|
|
responses.append(response)
|
|
else:
|
|
# Process web search results (existing logic)
|
|
# Check metatags are present
|
|
if "pagemap" not in search_item:
|
|
continue
|
|
if "metatags" not in search_item["pagemap"]:
|
|
continue
|
|
if "og:description" in search_item["pagemap"]["metatags"][0]:
|
|
long_description = search_item["pagemap"]["metatags"][0]["og:description"]
|
|
else:
|
|
long_description = "N/A"
|
|
# Get the page title
|
|
title = search_item.get("title")
|
|
# Page snippet
|
|
snippet = search_item.get("snippet")
|
|
|
|
# Extract the page url
|
|
link = search_item.get("link")
|
|
response = {
|
|
"result_id": i,
|
|
"title": title,
|
|
"description": snippet,
|
|
"long_description": long_description,
|
|
"url": link,
|
|
}
|
|
responses.append(response)
|
|
else:
|
|
error_info = data.get("error", {})
|
|
logger.error(f"Google search failed - API response: {error_info}")
|
|
responses.append({"error": f"Google search failed - API response: {error_info}"})
|
|
|
|
except Exception as e:
|
|
responses.append({"error": f"google search failed: {e!s}"})
|
|
return responses
|