mirror of
https://github.com/Alishahryar1/free-claude-code.git
synced 2026-05-02 05:10:35 +00:00
- Point DeepSeek at api.deepseek.com/anthropic with x-api-key headers - Native request builder, DeepSeek-specific thinking/block sanitization - Drop deepseek from OpenAI-chat server-tool preflight; update tests and docs - Default smoke model deepseek-v4-pro; re-export dump_raw_messages_request
621 lines
21 KiB
Python
621 lines
21 KiB
Python
from typing import Any
|
|
from unittest.mock import AsyncMock, MagicMock, patch
|
|
|
|
import httpx
|
|
import pytest
|
|
|
|
import api.web_tools.constants as web_tool_constants
|
|
from api.model_router import ModelRouter, ResolvedModel, RoutedMessagesRequest
|
|
from api.models.anthropic import Message, MessagesRequest, Tool
|
|
from api.services import ClaudeProxyService
|
|
from api.web_tools import egress as web_egress
|
|
from api.web_tools.egress import (
|
|
WebFetchEgressPolicy,
|
|
WebFetchEgressViolation,
|
|
enforce_web_fetch_egress,
|
|
)
|
|
from api.web_tools.outbound import (
|
|
_drain_response_body_capped,
|
|
_read_response_body_capped,
|
|
_run_web_fetch,
|
|
)
|
|
from api.web_tools.request import is_web_server_tool_request
|
|
from api.web_tools.streaming import stream_web_server_tool_response
|
|
from config.settings import Settings
|
|
from core.anthropic.stream_contracts import (
|
|
assert_anthropic_stream_contract,
|
|
parse_sse_text,
|
|
text_content,
|
|
)
|
|
from messaging.event_parser import parse_cli_event
|
|
from providers.exceptions import InvalidRequestError
|
|
|
|
_STRICT_EGRESS = WebFetchEgressPolicy(
|
|
allow_private_network_targets=False,
|
|
allowed_schemes=frozenset({"http", "https"}),
|
|
)
|
|
|
|
|
|
class FixedProviderModelRouter(ModelRouter):
|
|
"""Test double: pin ``provider_id`` for OpenAI vs native routing assertions."""
|
|
|
|
def __init__(self, settings: Settings, provider_id: str) -> None:
|
|
super().__init__(settings)
|
|
self._fixed_provider_id = provider_id
|
|
|
|
def resolve_messages_request(
|
|
self, request: MessagesRequest
|
|
) -> RoutedMessagesRequest:
|
|
resolved = ResolvedModel(
|
|
original_model=request.model,
|
|
provider_id=self._fixed_provider_id,
|
|
provider_model=request.model,
|
|
provider_model_ref=f"{self._fixed_provider_id}/{request.model}",
|
|
thinking_enabled=False,
|
|
)
|
|
routed = request.model_copy(deep=True)
|
|
routed.model = resolved.provider_model
|
|
return RoutedMessagesRequest(request=routed, resolved=resolved)
|
|
|
|
|
|
def test_web_server_tool_not_detected_when_tool_only_listed():
|
|
"""Listing web_search without forcing it must not skip the upstream provider."""
|
|
request = MessagesRequest(
|
|
model="claude-haiku-4-5-20251001",
|
|
max_tokens=100,
|
|
messages=[Message(role="user", content="search")],
|
|
tools=[Tool(name="web_search", type="web_search_20250305")],
|
|
)
|
|
|
|
assert not is_web_server_tool_request(request)
|
|
|
|
|
|
def test_web_server_tool_detected_when_tool_choice_forces_it():
|
|
request = MessagesRequest(
|
|
model="claude-haiku-4-5-20251001",
|
|
max_tokens=100,
|
|
messages=[Message(role="user", content="search")],
|
|
tools=[Tool(name="web_search", type="web_search_20250305")],
|
|
tool_choice={"type": "tool", "name": "web_search"},
|
|
)
|
|
|
|
assert is_web_server_tool_request(request)
|
|
|
|
|
|
def test_web_server_tool_not_detected_when_forced_name_missing_from_tools():
|
|
request = MessagesRequest(
|
|
model="claude-haiku-4-5-20251001",
|
|
max_tokens=100,
|
|
messages=[Message(role="user", content="hi")],
|
|
tools=[Tool(name="other", type="function")],
|
|
tool_choice={"type": "tool", "name": "web_search"},
|
|
)
|
|
|
|
assert not is_web_server_tool_request(request)
|
|
|
|
|
|
def test_service_rejects_forced_server_tool_on_openai_when_disabled():
|
|
"""OpenAI Chat upstreams cannot run forced server tools without the local handler."""
|
|
settings = Settings()
|
|
assert settings.enable_web_server_tools is False
|
|
service = ClaudeProxyService(
|
|
settings,
|
|
provider_getter=lambda _: MagicMock(),
|
|
model_router=FixedProviderModelRouter(settings, "nvidia_nim"),
|
|
)
|
|
request = MessagesRequest(
|
|
model="claude-haiku-4-5-20251001",
|
|
max_tokens=100,
|
|
messages=[
|
|
Message(
|
|
role="user",
|
|
content="Perform a web search for the query: DeepSeek V4 model release 2026",
|
|
)
|
|
],
|
|
tools=[Tool(name="web_search", type="web_search_20250305")],
|
|
tool_choice={"type": "tool", "name": "web_search"},
|
|
)
|
|
with pytest.raises(InvalidRequestError, match="ENABLE_WEB_SERVER_TOOLS"):
|
|
service.create_message(request)
|
|
|
|
|
|
@pytest.mark.parametrize(
|
|
"url",
|
|
[
|
|
"http://127.0.0.1/",
|
|
"http://192.168.1.1/",
|
|
"http://10.0.0.1/",
|
|
"http://[::1]/",
|
|
"http://localhost/foo",
|
|
"http://mybox.local/",
|
|
"file:///etc/passwd",
|
|
"http://169.254.169.254/latest/meta-data/",
|
|
],
|
|
)
|
|
def test_enforce_web_fetch_egress_blocks_internal_or_disallowed(url: str):
|
|
with pytest.raises(WebFetchEgressViolation):
|
|
enforce_web_fetch_egress(url, _STRICT_EGRESS)
|
|
|
|
|
|
def test_enforce_web_fetch_egress_allows_global_literal_ip():
|
|
enforce_web_fetch_egress("http://8.8.8.8/", _STRICT_EGRESS)
|
|
|
|
|
|
def test_enforce_web_fetch_egress_skips_private_checks_when_opted_in():
|
|
enforce_web_fetch_egress(
|
|
"http://127.0.0.1/",
|
|
WebFetchEgressPolicy(
|
|
allow_private_network_targets=True,
|
|
allowed_schemes=frozenset({"http", "https"}),
|
|
),
|
|
)
|
|
|
|
|
|
def _cm(mock_client: MagicMock) -> MagicMock:
|
|
cm = MagicMock()
|
|
cm.__aenter__ = AsyncMock(return_value=mock_client)
|
|
cm.__aexit__ = AsyncMock(return_value=None)
|
|
return cm
|
|
|
|
|
|
def _stream_cm(response: httpx.Response) -> MagicMock:
|
|
cm = MagicMock()
|
|
cm.__aenter__ = AsyncMock(return_value=response)
|
|
cm.__aexit__ = AsyncMock(return_value=None)
|
|
return cm
|
|
|
|
|
|
def _aiohttp_response(
|
|
status: int,
|
|
*,
|
|
url: str = "http://8.8.8.8/",
|
|
location: str | None = None,
|
|
body: bytes = b"hello world",
|
|
) -> MagicMock:
|
|
r = MagicMock()
|
|
r.status = status
|
|
r.url = url
|
|
hdrs: dict[str, str] = {}
|
|
if location is not None:
|
|
hdrs["location"] = location
|
|
r.headers = hdrs
|
|
r.get_encoding = MagicMock(return_value="utf-8")
|
|
r.raise_for_status = MagicMock()
|
|
r.request_info = MagicMock()
|
|
r.history = ()
|
|
|
|
async def iter_chunked(_n: int) -> Any:
|
|
yield body
|
|
|
|
r.content.iter_chunked = MagicMock(side_effect=iter_chunked)
|
|
return r
|
|
|
|
|
|
def _aiohttp_client_session_patch(
|
|
*responses: MagicMock,
|
|
) -> tuple[MagicMock, MagicMock]:
|
|
"""Build ``ClientSession`` mock that serves ``responses`` to successive ``get`` calls."""
|
|
queue = list(responses)
|
|
n = 0
|
|
|
|
def get_side(*_a: Any, **_k: Any) -> Any:
|
|
nonlocal n
|
|
resp = queue[n] if n < len(queue) else queue[-1]
|
|
n += 1
|
|
cm = MagicMock()
|
|
cm.__aenter__ = AsyncMock(return_value=resp)
|
|
cm.__aexit__ = AsyncMock(return_value=None)
|
|
return cm
|
|
|
|
session = MagicMock()
|
|
session.get = MagicMock(side_effect=get_side)
|
|
|
|
client_cm = MagicMock()
|
|
client_cm.__aenter__ = AsyncMock(return_value=session)
|
|
client_cm.__aexit__ = AsyncMock(return_value=None)
|
|
return client_cm, session
|
|
|
|
|
|
def test_enforce_web_fetch_egress_documents_connect_time_pinning():
|
|
assert enforce_web_fetch_egress.__doc__ and "resolved addresses" in (
|
|
enforce_web_fetch_egress.__doc__ or ""
|
|
)
|
|
assert (
|
|
web_egress.get_validated_stream_addrinfos_for_egress.__doc__
|
|
and "pinning"
|
|
in (web_egress.get_validated_stream_addrinfos_for_egress.__doc__ or "")
|
|
)
|
|
assert "DNS-pinned" in (_run_web_fetch.__doc__ or "")
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_run_web_fetch_follows_redirect_when_each_hop_is_allowed():
|
|
res_redirect = _aiohttp_response(
|
|
302, url="http://8.8.8.8/start", location="/final", body=b""
|
|
)
|
|
res_ok = _aiohttp_response(200, url="http://8.8.8.8/final", body=b"hello world")
|
|
client_cm, session = _aiohttp_client_session_patch(res_redirect, res_ok)
|
|
with patch("api.web_tools.outbound.ClientSession", return_value=client_cm):
|
|
out = await _run_web_fetch("http://8.8.8.8/start", _STRICT_EGRESS)
|
|
|
|
assert out["data"] == "hello world"
|
|
assert session.get.call_count == 2
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_run_web_fetch_truncates_large_body_to_byte_cap(monkeypatch):
|
|
huge = b"x" * 5000
|
|
res_ok = _aiohttp_response(200, url="http://8.8.8.8/big", body=huge)
|
|
client_cm, _ = _aiohttp_client_session_patch(res_ok)
|
|
monkeypatch.setattr(web_tool_constants, "_MAX_WEB_FETCH_RESPONSE_BYTES", 100)
|
|
with patch("api.web_tools.outbound.ClientSession", return_value=client_cm):
|
|
out = await _run_web_fetch("http://8.8.8.8/big", _STRICT_EGRESS)
|
|
|
|
assert len(out["data"]) <= 100
|
|
assert out["data"] == "x" * 100
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_run_web_fetch_redirect_to_blocked_host_raises():
|
|
res_redirect = _aiohttp_response(
|
|
302,
|
|
url="http://8.8.8.8/start",
|
|
location="http://127.0.0.1/secret",
|
|
body=b"",
|
|
)
|
|
client_cm, session = _aiohttp_client_session_patch(res_redirect)
|
|
with (
|
|
patch("api.web_tools.outbound.ClientSession", return_value=client_cm),
|
|
pytest.raises(WebFetchEgressViolation),
|
|
):
|
|
await _run_web_fetch("http://8.8.8.8/start", _STRICT_EGRESS)
|
|
|
|
session.get.assert_called_once()
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_run_web_fetch_redirect_without_location_raises():
|
|
res_bad = _aiohttp_response(302, url="http://8.8.8.8/here", body=b"")
|
|
client_cm, _ = _aiohttp_client_session_patch(res_bad)
|
|
with (
|
|
patch("api.web_tools.outbound.ClientSession", return_value=client_cm),
|
|
pytest.raises(WebFetchEgressViolation, match="missing Location"),
|
|
):
|
|
await _run_web_fetch("http://8.8.8.8/here", _STRICT_EGRESS)
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_run_web_fetch_excess_redirects_raises():
|
|
res1 = _aiohttp_response(302, url="http://8.8.8.8/a", location="/b", body=b"")
|
|
res2 = _aiohttp_response(302, url="http://8.8.8.8/b", location="/c", body=b"")
|
|
client_cm, _ = _aiohttp_client_session_patch(res1, res2)
|
|
with (
|
|
patch("api.web_tools.constants._MAX_WEB_FETCH_REDIRECTS", 1),
|
|
patch("api.web_tools.outbound.ClientSession", return_value=client_cm),
|
|
pytest.raises(WebFetchEgressViolation, match="exceeded maximum redirects"),
|
|
):
|
|
await _run_web_fetch("http://8.8.8.8/a", _STRICT_EGRESS)
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_streams_web_search_server_tool_result(monkeypatch):
|
|
async def fake_search(query: str) -> list[dict[str, str]]:
|
|
assert query == "DeepSeek V4 model release 2026"
|
|
return [{"title": "DeepSeek V4 Released", "url": "https://example.com/v4"}]
|
|
|
|
monkeypatch.setattr("api.web_tools.outbound._run_web_search", fake_search)
|
|
request = MessagesRequest(
|
|
model="claude-haiku-4-5-20251001",
|
|
max_tokens=100,
|
|
messages=[
|
|
Message(
|
|
role="user",
|
|
content=(
|
|
"Perform a web search for the query: DeepSeek V4 model release 2026"
|
|
),
|
|
)
|
|
],
|
|
tools=[Tool(name="web_search", type="web_search_20250305")],
|
|
tool_choice={"type": "tool", "name": "web_search"},
|
|
)
|
|
|
|
raw = "".join(
|
|
[
|
|
event
|
|
async for event in stream_web_server_tool_response(
|
|
request, input_tokens=42, web_fetch_egress=_STRICT_EGRESS
|
|
)
|
|
]
|
|
)
|
|
events = parse_sse_text(raw)
|
|
assert_anthropic_stream_contract(events)
|
|
starts = [e for e in events if e.event == "content_block_start"]
|
|
assert starts[0].data["content_block"]["type"] == "server_tool_use"
|
|
assert starts[0].data["content_block"]["name"] == "web_search"
|
|
tool_use_id = starts[0].data["content_block"]["id"]
|
|
assert starts[1].data["content_block"]["type"] == "web_search_tool_result"
|
|
assert starts[1].data["content_block"]["tool_use_id"] == tool_use_id
|
|
assert starts[1].data["content_block"]["content"][0]["url"] == (
|
|
"https://example.com/v4"
|
|
)
|
|
text_deltas = [
|
|
e
|
|
for e in events
|
|
if e.event == "content_block_delta"
|
|
and e.data.get("delta", {}).get("type") == "text_delta"
|
|
]
|
|
assert text_deltas, "summary must be streamed as text_delta"
|
|
assert "example.com" in text_content(events)
|
|
cli_text: list[str] = []
|
|
for ev in events:
|
|
cli_text.extend(
|
|
str(p.get("text", ""))
|
|
for p in parse_cli_event(ev.data)
|
|
if p.get("type") == "text_delta"
|
|
)
|
|
assert "example.com" in "".join(cli_text)
|
|
deltas = [e for e in events if e.event == "message_delta"]
|
|
assert deltas[-1].data["usage"]["server_tool_use"] == {"web_search_requests": 1}
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_forced_web_fetch_ignores_stale_url_from_prior_user_turns(monkeypatch):
|
|
"""Only the latest user message supplies the URL (not earlier transcript text)."""
|
|
target = "https://new-only.example.com/page"
|
|
|
|
async def fake_fetch(url: str, _egress: WebFetchEgressPolicy) -> dict[str, str]:
|
|
assert url == target
|
|
return {
|
|
"url": url,
|
|
"title": "T",
|
|
"media_type": "text/plain",
|
|
"data": "x",
|
|
}
|
|
|
|
monkeypatch.setattr("api.web_tools.outbound._run_web_fetch", fake_fetch)
|
|
request = MessagesRequest(
|
|
model="claude-haiku-4-5-20251001",
|
|
max_tokens=100,
|
|
messages=[
|
|
Message(
|
|
role="user",
|
|
content="Earlier turn https://stale.com/old-article ignore this",
|
|
),
|
|
Message(role="assistant", content="ok"),
|
|
Message(
|
|
role="user",
|
|
content=f"Please fetch {target} for the summary",
|
|
),
|
|
],
|
|
tools=[Tool(name="web_fetch", type="web_fetch_20250910")],
|
|
tool_choice={"type": "tool", "name": "web_fetch"},
|
|
)
|
|
|
|
raw = "".join(
|
|
[
|
|
event
|
|
async for event in stream_web_server_tool_response(
|
|
request, input_tokens=1, web_fetch_egress=_STRICT_EGRESS
|
|
)
|
|
]
|
|
)
|
|
assert target in raw
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_streams_web_fetch_server_tool_result(monkeypatch):
|
|
async def fake_fetch(url: str, _egress: WebFetchEgressPolicy) -> dict[str, str]:
|
|
assert url == "https://example.com/article"
|
|
return {
|
|
"url": url,
|
|
"title": "Example Article",
|
|
"media_type": "text/plain",
|
|
"data": "Article body",
|
|
}
|
|
|
|
monkeypatch.setattr("api.web_tools.outbound._run_web_fetch", fake_fetch)
|
|
request = MessagesRequest(
|
|
model="claude-haiku-4-5-20251001",
|
|
max_tokens=100,
|
|
messages=[
|
|
Message(role="user", content="Fetch https://example.com/article please")
|
|
],
|
|
tools=[Tool(name="web_fetch", type="web_fetch_20250910")],
|
|
tool_choice={"type": "tool", "name": "web_fetch"},
|
|
)
|
|
|
|
raw = "".join(
|
|
[
|
|
event
|
|
async for event in stream_web_server_tool_response(
|
|
request, input_tokens=42, web_fetch_egress=_STRICT_EGRESS
|
|
)
|
|
]
|
|
)
|
|
events = parse_sse_text(raw)
|
|
assert_anthropic_stream_contract(events)
|
|
starts = [e for e in events if e.event == "content_block_start"]
|
|
assert starts[0].data["content_block"]["type"] == "server_tool_use"
|
|
tool_use_id = starts[0].data["content_block"]["id"]
|
|
assert starts[1].data["content_block"]["type"] == "web_fetch_tool_result"
|
|
assert starts[1].data["content_block"]["tool_use_id"] == tool_use_id
|
|
assert starts[1].data["content_block"]["content"]["content"]["title"] == (
|
|
"Example Article"
|
|
)
|
|
assert any(
|
|
e.event == "content_block_delta"
|
|
and e.data.get("delta", {}).get("type") == "text_delta"
|
|
for e in events
|
|
)
|
|
assert "Article body" in text_content(events)
|
|
cli_text: list[str] = []
|
|
for ev in events:
|
|
cli_text.extend(
|
|
str(p.get("text", ""))
|
|
for p in parse_cli_event(ev.data)
|
|
if p.get("type") == "text_delta"
|
|
)
|
|
assert "Article body" in "".join(cli_text)
|
|
deltas = [e for e in events if e.event == "message_delta"]
|
|
assert deltas[-1].data["usage"]["server_tool_use"] == {"web_fetch_requests": 1}
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_streams_web_fetch_error_summary_generic_by_default(monkeypatch):
|
|
secret = "sensitive-upstream-token"
|
|
|
|
async def boom(_url: str, _egress: WebFetchEgressPolicy) -> dict[str, str]:
|
|
raise ValueError(secret)
|
|
|
|
monkeypatch.setattr("api.web_tools.outbound._run_web_fetch", boom)
|
|
request = MessagesRequest(
|
|
model="claude-haiku-4-5-20251001",
|
|
max_tokens=100,
|
|
messages=[
|
|
Message(
|
|
role="user",
|
|
content="Fetch https://example.com/sensitive-path?x=1 please",
|
|
)
|
|
],
|
|
tools=[Tool(name="web_fetch", type="web_fetch_20250910")],
|
|
tool_choice={"type": "tool", "name": "web_fetch"},
|
|
)
|
|
|
|
with patch("api.web_tools.outbound.logger.warning") as log_warn:
|
|
raw = "".join(
|
|
[
|
|
event
|
|
async for event in stream_web_server_tool_response(
|
|
request,
|
|
input_tokens=1,
|
|
web_fetch_egress=_STRICT_EGRESS,
|
|
verbose_client_errors=False,
|
|
)
|
|
]
|
|
)
|
|
|
|
assert secret not in raw
|
|
assert "ValueError" not in raw
|
|
assert "Web tool request failed." in raw
|
|
err_events = parse_sse_text(raw)
|
|
assert_anthropic_stream_contract(err_events)
|
|
assert any(
|
|
e.event == "content_block_delta"
|
|
and e.data.get("delta", {}).get("type") == "text_delta"
|
|
for e in err_events
|
|
)
|
|
cli_err_text: list[str] = []
|
|
for ev in err_events:
|
|
cli_err_text.extend(
|
|
str(p.get("text", ""))
|
|
for p in parse_cli_event(ev.data)
|
|
if p.get("type") == "text_delta"
|
|
)
|
|
assert "Web tool request failed." in "".join(cli_err_text)
|
|
log_blob = " ".join(str(a) for c in log_warn.call_args_list for a in c.args)
|
|
assert secret not in log_blob
|
|
assert "example.com" in log_blob
|
|
assert "/sensitive-path" not in log_blob
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_streams_web_fetch_error_summary_verbose_includes_exception_class(
|
|
monkeypatch,
|
|
):
|
|
async def boom(_url: str, _egress: WebFetchEgressPolicy) -> dict[str, str]:
|
|
raise OSError(5, "oops")
|
|
|
|
monkeypatch.setattr("api.web_tools.outbound._run_web_fetch", boom)
|
|
request = MessagesRequest(
|
|
model="claude-haiku-4-5-20251001",
|
|
max_tokens=100,
|
|
messages=[Message(role="user", content="Fetch https://example.com/x")],
|
|
tools=[Tool(name="web_fetch", type="web_fetch_20250910")],
|
|
tool_choice={"type": "tool", "name": "web_fetch"},
|
|
)
|
|
|
|
raw = "".join(
|
|
[
|
|
event
|
|
async for event in stream_web_server_tool_response(
|
|
request,
|
|
input_tokens=1,
|
|
web_fetch_egress=_STRICT_EGRESS,
|
|
verbose_client_errors=True,
|
|
)
|
|
]
|
|
)
|
|
assert "OSError" in raw
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_read_response_body_capped_truncates_single_oversized_chunk():
|
|
cap = 500
|
|
|
|
async def aiter_bytes(chunk_size=None):
|
|
yield b"z" * (cap * 20)
|
|
|
|
response = MagicMock()
|
|
response.aiter_bytes = aiter_bytes
|
|
|
|
out = await _read_response_body_capped(response, cap)
|
|
assert len(out) == cap
|
|
assert out == b"z" * cap
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_drain_response_body_capped_stops_after_first_chunk_when_oversized():
|
|
cap = 300
|
|
chunk_calls = {"n": 0}
|
|
|
|
async def aiter_bytes(chunk_size=None):
|
|
chunk_calls["n"] += 1
|
|
yield b"y" * (cap * 10)
|
|
|
|
response = MagicMock()
|
|
response.aiter_bytes = aiter_bytes
|
|
|
|
await _drain_response_body_capped(response, cap)
|
|
assert chunk_calls["n"] == 1
|
|
|
|
|
|
def test_service_rejects_listed_server_tools_on_openai_chat() -> None:
|
|
settings = Settings()
|
|
service = ClaudeProxyService(
|
|
settings,
|
|
provider_getter=lambda _: MagicMock(),
|
|
model_router=FixedProviderModelRouter(settings, "nvidia_nim"),
|
|
)
|
|
request = MessagesRequest(
|
|
model="m",
|
|
max_tokens=20,
|
|
messages=[Message(role="user", content="q")],
|
|
tools=[Tool(name="web_search", type="web_search_20250305")],
|
|
)
|
|
with pytest.raises(InvalidRequestError, match="OpenAI Chat upstreams"):
|
|
service.create_message(request)
|
|
|
|
|
|
def test_listed_server_tools_routed_on_open_router() -> None:
|
|
"""Native Anthropic transport may receive listed server tool definitions."""
|
|
settings = Settings()
|
|
|
|
async def fake_stream(*_a, **_k):
|
|
yield 'event: message_start\ndata: {"type":"message_start"}\n\n'
|
|
yield 'event: message_stop\ndata: {"type":"message_stop"}\n\n'
|
|
|
|
mock_provider = MagicMock()
|
|
mock_provider.stream_response = fake_stream
|
|
service = ClaudeProxyService(
|
|
settings,
|
|
provider_getter=lambda _: mock_provider,
|
|
model_router=FixedProviderModelRouter(settings, "open_router"),
|
|
)
|
|
request = MessagesRequest(
|
|
model="m",
|
|
max_tokens=20,
|
|
messages=[Message(role="user", content="q")],
|
|
tools=[Tool(name="web_search", type="web_search_20250305")],
|
|
)
|
|
service.create_message(request)
|
|
mock_provider.preflight_stream.assert_called()
|