free-claude-code/tests/api/test_web_server_tools.py
Alishahryar1 6297b48f81 feat(deepseek): use native Anthropic Messages transport
- Point DeepSeek at api.deepseek.com/anthropic with x-api-key headers
- Native request builder, DeepSeek-specific thinking/block sanitization
- Drop deepseek from OpenAI-chat server-tool preflight; update tests and docs
- Default smoke model deepseek-v4-pro; re-export dump_raw_messages_request
2026-04-26 12:03:21 -07:00

621 lines
21 KiB
Python

from typing import Any
from unittest.mock import AsyncMock, MagicMock, patch
import httpx
import pytest
import api.web_tools.constants as web_tool_constants
from api.model_router import ModelRouter, ResolvedModel, RoutedMessagesRequest
from api.models.anthropic import Message, MessagesRequest, Tool
from api.services import ClaudeProxyService
from api.web_tools import egress as web_egress
from api.web_tools.egress import (
WebFetchEgressPolicy,
WebFetchEgressViolation,
enforce_web_fetch_egress,
)
from api.web_tools.outbound import (
_drain_response_body_capped,
_read_response_body_capped,
_run_web_fetch,
)
from api.web_tools.request import is_web_server_tool_request
from api.web_tools.streaming import stream_web_server_tool_response
from config.settings import Settings
from core.anthropic.stream_contracts import (
assert_anthropic_stream_contract,
parse_sse_text,
text_content,
)
from messaging.event_parser import parse_cli_event
from providers.exceptions import InvalidRequestError
_STRICT_EGRESS = WebFetchEgressPolicy(
allow_private_network_targets=False,
allowed_schemes=frozenset({"http", "https"}),
)
class FixedProviderModelRouter(ModelRouter):
"""Test double: pin ``provider_id`` for OpenAI vs native routing assertions."""
def __init__(self, settings: Settings, provider_id: str) -> None:
super().__init__(settings)
self._fixed_provider_id = provider_id
def resolve_messages_request(
self, request: MessagesRequest
) -> RoutedMessagesRequest:
resolved = ResolvedModel(
original_model=request.model,
provider_id=self._fixed_provider_id,
provider_model=request.model,
provider_model_ref=f"{self._fixed_provider_id}/{request.model}",
thinking_enabled=False,
)
routed = request.model_copy(deep=True)
routed.model = resolved.provider_model
return RoutedMessagesRequest(request=routed, resolved=resolved)
def test_web_server_tool_not_detected_when_tool_only_listed():
"""Listing web_search without forcing it must not skip the upstream provider."""
request = MessagesRequest(
model="claude-haiku-4-5-20251001",
max_tokens=100,
messages=[Message(role="user", content="search")],
tools=[Tool(name="web_search", type="web_search_20250305")],
)
assert not is_web_server_tool_request(request)
def test_web_server_tool_detected_when_tool_choice_forces_it():
request = MessagesRequest(
model="claude-haiku-4-5-20251001",
max_tokens=100,
messages=[Message(role="user", content="search")],
tools=[Tool(name="web_search", type="web_search_20250305")],
tool_choice={"type": "tool", "name": "web_search"},
)
assert is_web_server_tool_request(request)
def test_web_server_tool_not_detected_when_forced_name_missing_from_tools():
request = MessagesRequest(
model="claude-haiku-4-5-20251001",
max_tokens=100,
messages=[Message(role="user", content="hi")],
tools=[Tool(name="other", type="function")],
tool_choice={"type": "tool", "name": "web_search"},
)
assert not is_web_server_tool_request(request)
def test_service_rejects_forced_server_tool_on_openai_when_disabled():
"""OpenAI Chat upstreams cannot run forced server tools without the local handler."""
settings = Settings()
assert settings.enable_web_server_tools is False
service = ClaudeProxyService(
settings,
provider_getter=lambda _: MagicMock(),
model_router=FixedProviderModelRouter(settings, "nvidia_nim"),
)
request = MessagesRequest(
model="claude-haiku-4-5-20251001",
max_tokens=100,
messages=[
Message(
role="user",
content="Perform a web search for the query: DeepSeek V4 model release 2026",
)
],
tools=[Tool(name="web_search", type="web_search_20250305")],
tool_choice={"type": "tool", "name": "web_search"},
)
with pytest.raises(InvalidRequestError, match="ENABLE_WEB_SERVER_TOOLS"):
service.create_message(request)
@pytest.mark.parametrize(
"url",
[
"http://127.0.0.1/",
"http://192.168.1.1/",
"http://10.0.0.1/",
"http://[::1]/",
"http://localhost/foo",
"http://mybox.local/",
"file:///etc/passwd",
"http://169.254.169.254/latest/meta-data/",
],
)
def test_enforce_web_fetch_egress_blocks_internal_or_disallowed(url: str):
with pytest.raises(WebFetchEgressViolation):
enforce_web_fetch_egress(url, _STRICT_EGRESS)
def test_enforce_web_fetch_egress_allows_global_literal_ip():
enforce_web_fetch_egress("http://8.8.8.8/", _STRICT_EGRESS)
def test_enforce_web_fetch_egress_skips_private_checks_when_opted_in():
enforce_web_fetch_egress(
"http://127.0.0.1/",
WebFetchEgressPolicy(
allow_private_network_targets=True,
allowed_schemes=frozenset({"http", "https"}),
),
)
def _cm(mock_client: MagicMock) -> MagicMock:
cm = MagicMock()
cm.__aenter__ = AsyncMock(return_value=mock_client)
cm.__aexit__ = AsyncMock(return_value=None)
return cm
def _stream_cm(response: httpx.Response) -> MagicMock:
cm = MagicMock()
cm.__aenter__ = AsyncMock(return_value=response)
cm.__aexit__ = AsyncMock(return_value=None)
return cm
def _aiohttp_response(
status: int,
*,
url: str = "http://8.8.8.8/",
location: str | None = None,
body: bytes = b"hello world",
) -> MagicMock:
r = MagicMock()
r.status = status
r.url = url
hdrs: dict[str, str] = {}
if location is not None:
hdrs["location"] = location
r.headers = hdrs
r.get_encoding = MagicMock(return_value="utf-8")
r.raise_for_status = MagicMock()
r.request_info = MagicMock()
r.history = ()
async def iter_chunked(_n: int) -> Any:
yield body
r.content.iter_chunked = MagicMock(side_effect=iter_chunked)
return r
def _aiohttp_client_session_patch(
*responses: MagicMock,
) -> tuple[MagicMock, MagicMock]:
"""Build ``ClientSession`` mock that serves ``responses`` to successive ``get`` calls."""
queue = list(responses)
n = 0
def get_side(*_a: Any, **_k: Any) -> Any:
nonlocal n
resp = queue[n] if n < len(queue) else queue[-1]
n += 1
cm = MagicMock()
cm.__aenter__ = AsyncMock(return_value=resp)
cm.__aexit__ = AsyncMock(return_value=None)
return cm
session = MagicMock()
session.get = MagicMock(side_effect=get_side)
client_cm = MagicMock()
client_cm.__aenter__ = AsyncMock(return_value=session)
client_cm.__aexit__ = AsyncMock(return_value=None)
return client_cm, session
def test_enforce_web_fetch_egress_documents_connect_time_pinning():
assert enforce_web_fetch_egress.__doc__ and "resolved addresses" in (
enforce_web_fetch_egress.__doc__ or ""
)
assert (
web_egress.get_validated_stream_addrinfos_for_egress.__doc__
and "pinning"
in (web_egress.get_validated_stream_addrinfos_for_egress.__doc__ or "")
)
assert "DNS-pinned" in (_run_web_fetch.__doc__ or "")
@pytest.mark.asyncio
async def test_run_web_fetch_follows_redirect_when_each_hop_is_allowed():
res_redirect = _aiohttp_response(
302, url="http://8.8.8.8/start", location="/final", body=b""
)
res_ok = _aiohttp_response(200, url="http://8.8.8.8/final", body=b"hello world")
client_cm, session = _aiohttp_client_session_patch(res_redirect, res_ok)
with patch("api.web_tools.outbound.ClientSession", return_value=client_cm):
out = await _run_web_fetch("http://8.8.8.8/start", _STRICT_EGRESS)
assert out["data"] == "hello world"
assert session.get.call_count == 2
@pytest.mark.asyncio
async def test_run_web_fetch_truncates_large_body_to_byte_cap(monkeypatch):
huge = b"x" * 5000
res_ok = _aiohttp_response(200, url="http://8.8.8.8/big", body=huge)
client_cm, _ = _aiohttp_client_session_patch(res_ok)
monkeypatch.setattr(web_tool_constants, "_MAX_WEB_FETCH_RESPONSE_BYTES", 100)
with patch("api.web_tools.outbound.ClientSession", return_value=client_cm):
out = await _run_web_fetch("http://8.8.8.8/big", _STRICT_EGRESS)
assert len(out["data"]) <= 100
assert out["data"] == "x" * 100
@pytest.mark.asyncio
async def test_run_web_fetch_redirect_to_blocked_host_raises():
res_redirect = _aiohttp_response(
302,
url="http://8.8.8.8/start",
location="http://127.0.0.1/secret",
body=b"",
)
client_cm, session = _aiohttp_client_session_patch(res_redirect)
with (
patch("api.web_tools.outbound.ClientSession", return_value=client_cm),
pytest.raises(WebFetchEgressViolation),
):
await _run_web_fetch("http://8.8.8.8/start", _STRICT_EGRESS)
session.get.assert_called_once()
@pytest.mark.asyncio
async def test_run_web_fetch_redirect_without_location_raises():
res_bad = _aiohttp_response(302, url="http://8.8.8.8/here", body=b"")
client_cm, _ = _aiohttp_client_session_patch(res_bad)
with (
patch("api.web_tools.outbound.ClientSession", return_value=client_cm),
pytest.raises(WebFetchEgressViolation, match="missing Location"),
):
await _run_web_fetch("http://8.8.8.8/here", _STRICT_EGRESS)
@pytest.mark.asyncio
async def test_run_web_fetch_excess_redirects_raises():
res1 = _aiohttp_response(302, url="http://8.8.8.8/a", location="/b", body=b"")
res2 = _aiohttp_response(302, url="http://8.8.8.8/b", location="/c", body=b"")
client_cm, _ = _aiohttp_client_session_patch(res1, res2)
with (
patch("api.web_tools.constants._MAX_WEB_FETCH_REDIRECTS", 1),
patch("api.web_tools.outbound.ClientSession", return_value=client_cm),
pytest.raises(WebFetchEgressViolation, match="exceeded maximum redirects"),
):
await _run_web_fetch("http://8.8.8.8/a", _STRICT_EGRESS)
@pytest.mark.asyncio
async def test_streams_web_search_server_tool_result(monkeypatch):
async def fake_search(query: str) -> list[dict[str, str]]:
assert query == "DeepSeek V4 model release 2026"
return [{"title": "DeepSeek V4 Released", "url": "https://example.com/v4"}]
monkeypatch.setattr("api.web_tools.outbound._run_web_search", fake_search)
request = MessagesRequest(
model="claude-haiku-4-5-20251001",
max_tokens=100,
messages=[
Message(
role="user",
content=(
"Perform a web search for the query: DeepSeek V4 model release 2026"
),
)
],
tools=[Tool(name="web_search", type="web_search_20250305")],
tool_choice={"type": "tool", "name": "web_search"},
)
raw = "".join(
[
event
async for event in stream_web_server_tool_response(
request, input_tokens=42, web_fetch_egress=_STRICT_EGRESS
)
]
)
events = parse_sse_text(raw)
assert_anthropic_stream_contract(events)
starts = [e for e in events if e.event == "content_block_start"]
assert starts[0].data["content_block"]["type"] == "server_tool_use"
assert starts[0].data["content_block"]["name"] == "web_search"
tool_use_id = starts[0].data["content_block"]["id"]
assert starts[1].data["content_block"]["type"] == "web_search_tool_result"
assert starts[1].data["content_block"]["tool_use_id"] == tool_use_id
assert starts[1].data["content_block"]["content"][0]["url"] == (
"https://example.com/v4"
)
text_deltas = [
e
for e in events
if e.event == "content_block_delta"
and e.data.get("delta", {}).get("type") == "text_delta"
]
assert text_deltas, "summary must be streamed as text_delta"
assert "example.com" in text_content(events)
cli_text: list[str] = []
for ev in events:
cli_text.extend(
str(p.get("text", ""))
for p in parse_cli_event(ev.data)
if p.get("type") == "text_delta"
)
assert "example.com" in "".join(cli_text)
deltas = [e for e in events if e.event == "message_delta"]
assert deltas[-1].data["usage"]["server_tool_use"] == {"web_search_requests": 1}
@pytest.mark.asyncio
async def test_forced_web_fetch_ignores_stale_url_from_prior_user_turns(monkeypatch):
"""Only the latest user message supplies the URL (not earlier transcript text)."""
target = "https://new-only.example.com/page"
async def fake_fetch(url: str, _egress: WebFetchEgressPolicy) -> dict[str, str]:
assert url == target
return {
"url": url,
"title": "T",
"media_type": "text/plain",
"data": "x",
}
monkeypatch.setattr("api.web_tools.outbound._run_web_fetch", fake_fetch)
request = MessagesRequest(
model="claude-haiku-4-5-20251001",
max_tokens=100,
messages=[
Message(
role="user",
content="Earlier turn https://stale.com/old-article ignore this",
),
Message(role="assistant", content="ok"),
Message(
role="user",
content=f"Please fetch {target} for the summary",
),
],
tools=[Tool(name="web_fetch", type="web_fetch_20250910")],
tool_choice={"type": "tool", "name": "web_fetch"},
)
raw = "".join(
[
event
async for event in stream_web_server_tool_response(
request, input_tokens=1, web_fetch_egress=_STRICT_EGRESS
)
]
)
assert target in raw
@pytest.mark.asyncio
async def test_streams_web_fetch_server_tool_result(monkeypatch):
async def fake_fetch(url: str, _egress: WebFetchEgressPolicy) -> dict[str, str]:
assert url == "https://example.com/article"
return {
"url": url,
"title": "Example Article",
"media_type": "text/plain",
"data": "Article body",
}
monkeypatch.setattr("api.web_tools.outbound._run_web_fetch", fake_fetch)
request = MessagesRequest(
model="claude-haiku-4-5-20251001",
max_tokens=100,
messages=[
Message(role="user", content="Fetch https://example.com/article please")
],
tools=[Tool(name="web_fetch", type="web_fetch_20250910")],
tool_choice={"type": "tool", "name": "web_fetch"},
)
raw = "".join(
[
event
async for event in stream_web_server_tool_response(
request, input_tokens=42, web_fetch_egress=_STRICT_EGRESS
)
]
)
events = parse_sse_text(raw)
assert_anthropic_stream_contract(events)
starts = [e for e in events if e.event == "content_block_start"]
assert starts[0].data["content_block"]["type"] == "server_tool_use"
tool_use_id = starts[0].data["content_block"]["id"]
assert starts[1].data["content_block"]["type"] == "web_fetch_tool_result"
assert starts[1].data["content_block"]["tool_use_id"] == tool_use_id
assert starts[1].data["content_block"]["content"]["content"]["title"] == (
"Example Article"
)
assert any(
e.event == "content_block_delta"
and e.data.get("delta", {}).get("type") == "text_delta"
for e in events
)
assert "Article body" in text_content(events)
cli_text: list[str] = []
for ev in events:
cli_text.extend(
str(p.get("text", ""))
for p in parse_cli_event(ev.data)
if p.get("type") == "text_delta"
)
assert "Article body" in "".join(cli_text)
deltas = [e for e in events if e.event == "message_delta"]
assert deltas[-1].data["usage"]["server_tool_use"] == {"web_fetch_requests": 1}
@pytest.mark.asyncio
async def test_streams_web_fetch_error_summary_generic_by_default(monkeypatch):
secret = "sensitive-upstream-token"
async def boom(_url: str, _egress: WebFetchEgressPolicy) -> dict[str, str]:
raise ValueError(secret)
monkeypatch.setattr("api.web_tools.outbound._run_web_fetch", boom)
request = MessagesRequest(
model="claude-haiku-4-5-20251001",
max_tokens=100,
messages=[
Message(
role="user",
content="Fetch https://example.com/sensitive-path?x=1 please",
)
],
tools=[Tool(name="web_fetch", type="web_fetch_20250910")],
tool_choice={"type": "tool", "name": "web_fetch"},
)
with patch("api.web_tools.outbound.logger.warning") as log_warn:
raw = "".join(
[
event
async for event in stream_web_server_tool_response(
request,
input_tokens=1,
web_fetch_egress=_STRICT_EGRESS,
verbose_client_errors=False,
)
]
)
assert secret not in raw
assert "ValueError" not in raw
assert "Web tool request failed." in raw
err_events = parse_sse_text(raw)
assert_anthropic_stream_contract(err_events)
assert any(
e.event == "content_block_delta"
and e.data.get("delta", {}).get("type") == "text_delta"
for e in err_events
)
cli_err_text: list[str] = []
for ev in err_events:
cli_err_text.extend(
str(p.get("text", ""))
for p in parse_cli_event(ev.data)
if p.get("type") == "text_delta"
)
assert "Web tool request failed." in "".join(cli_err_text)
log_blob = " ".join(str(a) for c in log_warn.call_args_list for a in c.args)
assert secret not in log_blob
assert "example.com" in log_blob
assert "/sensitive-path" not in log_blob
@pytest.mark.asyncio
async def test_streams_web_fetch_error_summary_verbose_includes_exception_class(
monkeypatch,
):
async def boom(_url: str, _egress: WebFetchEgressPolicy) -> dict[str, str]:
raise OSError(5, "oops")
monkeypatch.setattr("api.web_tools.outbound._run_web_fetch", boom)
request = MessagesRequest(
model="claude-haiku-4-5-20251001",
max_tokens=100,
messages=[Message(role="user", content="Fetch https://example.com/x")],
tools=[Tool(name="web_fetch", type="web_fetch_20250910")],
tool_choice={"type": "tool", "name": "web_fetch"},
)
raw = "".join(
[
event
async for event in stream_web_server_tool_response(
request,
input_tokens=1,
web_fetch_egress=_STRICT_EGRESS,
verbose_client_errors=True,
)
]
)
assert "OSError" in raw
@pytest.mark.asyncio
async def test_read_response_body_capped_truncates_single_oversized_chunk():
cap = 500
async def aiter_bytes(chunk_size=None):
yield b"z" * (cap * 20)
response = MagicMock()
response.aiter_bytes = aiter_bytes
out = await _read_response_body_capped(response, cap)
assert len(out) == cap
assert out == b"z" * cap
@pytest.mark.asyncio
async def test_drain_response_body_capped_stops_after_first_chunk_when_oversized():
cap = 300
chunk_calls = {"n": 0}
async def aiter_bytes(chunk_size=None):
chunk_calls["n"] += 1
yield b"y" * (cap * 10)
response = MagicMock()
response.aiter_bytes = aiter_bytes
await _drain_response_body_capped(response, cap)
assert chunk_calls["n"] == 1
def test_service_rejects_listed_server_tools_on_openai_chat() -> None:
settings = Settings()
service = ClaudeProxyService(
settings,
provider_getter=lambda _: MagicMock(),
model_router=FixedProviderModelRouter(settings, "nvidia_nim"),
)
request = MessagesRequest(
model="m",
max_tokens=20,
messages=[Message(role="user", content="q")],
tools=[Tool(name="web_search", type="web_search_20250305")],
)
with pytest.raises(InvalidRequestError, match="OpenAI Chat upstreams"):
service.create_message(request)
def test_listed_server_tools_routed_on_open_router() -> None:
"""Native Anthropic transport may receive listed server tool definitions."""
settings = Settings()
async def fake_stream(*_a, **_k):
yield 'event: message_start\ndata: {"type":"message_start"}\n\n'
yield 'event: message_stop\ndata: {"type":"message_stop"}\n\n'
mock_provider = MagicMock()
mock_provider.stream_response = fake_stream
service = ClaudeProxyService(
settings,
provider_getter=lambda _: mock_provider,
model_router=FixedProviderModelRouter(settings, "open_router"),
)
request = MessagesRequest(
model="m",
max_tokens=20,
messages=[Message(role="user", content="q")],
tools=[Tool(name="web_search", type="web_search_20250305")],
)
service.create_message(request)
mock_provider.preflight_stream.assert_called()