fix(openrouter): revert custom_llm_provider, always apply gateway prefix

This commit is contained in:
Xubin Ren
2026-03-15 02:43:50 +00:00
committed by Xubin Ren
parent 350d110fb9
commit 196e0ddbb6
3 changed files with 61 additions and 20 deletions

View File

@@ -1,4 +1,10 @@
"""Regression tests for PR #2026 — litellm_kwargs injection from ProviderSpec."""
"""Regression tests for PR #2026 — litellm_kwargs injection from ProviderSpec.
Validates that:
- OpenRouter uses litellm_prefix (NOT custom_llm_provider) to avoid LiteLLM double-prefixing.
- The litellm_kwargs mechanism works correctly for providers that declare it.
- Non-gateway providers are unaffected.
"""
from __future__ import annotations
@@ -9,6 +15,7 @@ from unittest.mock import AsyncMock, patch
import pytest
from nanobot.providers.litellm_provider import LiteLLMProvider
from nanobot.providers.registry import find_by_name
def _fake_response(content: str = "ok") -> SimpleNamespace:
@@ -24,9 +31,23 @@ def _fake_response(content: str = "ok") -> SimpleNamespace:
return SimpleNamespace(choices=[choice], usage=usage)
def test_openrouter_spec_uses_prefix_not_custom_llm_provider() -> None:
"""OpenRouter must rely on litellm_prefix, not custom_llm_provider kwarg.
LiteLLM internally adds a provider/ prefix when custom_llm_provider is set,
which double-prefixes models (openrouter/anthropic/model) and breaks the API.
"""
spec = find_by_name("openrouter")
assert spec is not None
assert spec.litellm_prefix == "openrouter"
assert "custom_llm_provider" not in spec.litellm_kwargs, (
"custom_llm_provider causes LiteLLM to double-prefix the model name"
)
@pytest.mark.asyncio
async def test_openrouter_injects_litellm_kwargs() -> None:
"""OpenRouter gateway must inject custom_llm_provider into acompletion call."""
async def test_openrouter_prefixes_model_correctly() -> None:
"""OpenRouter should prefix model as openrouter/vendor/model for LiteLLM routing."""
mock_acompletion = AsyncMock(return_value=_fake_response())
with patch("nanobot.providers.litellm_provider.acompletion", mock_acompletion):
@@ -42,16 +63,14 @@ async def test_openrouter_injects_litellm_kwargs() -> None:
)
call_kwargs = mock_acompletion.call_args.kwargs
assert call_kwargs.get("custom_llm_provider") == "openrouter", (
"OpenRouter gateway should pass custom_llm_provider='openrouter' to acompletion"
)
assert call_kwargs["model"] == "anthropic/claude-sonnet-4-5", (
"Model name must NOT get an 'openrouter/' prefix — routing is via custom_llm_provider"
assert call_kwargs["model"] == "openrouter/anthropic/claude-sonnet-4-5", (
"LiteLLM needs openrouter/ prefix to detect the provider and strip it before API call"
)
assert "custom_llm_provider" not in call_kwargs
@pytest.mark.asyncio
async def test_non_gateway_provider_does_not_inject_litellm_kwargs() -> None:
async def test_non_gateway_provider_no_extra_kwargs() -> None:
"""Standard (non-gateway) providers must NOT inject any litellm_kwargs."""
mock_acompletion = AsyncMock(return_value=_fake_response())
@@ -89,9 +108,7 @@ async def test_gateway_without_litellm_kwargs_injects_nothing_extra() -> None:
)
call_kwargs = mock_acompletion.call_args.kwargs
assert "custom_llm_provider" not in call_kwargs, (
"AiHubMix gateway has no litellm_kwargs, should not add custom_llm_provider"
)
assert "custom_llm_provider" not in call_kwargs
@pytest.mark.asyncio
@@ -110,9 +127,35 @@ async def test_openrouter_autodetect_by_key_prefix() -> None:
)
call_kwargs = mock_acompletion.call_args.kwargs
assert call_kwargs.get("custom_llm_provider") == "openrouter", (
"Auto-detected OpenRouter (by sk-or- prefix) should still inject custom_llm_provider"
assert call_kwargs["model"] == "openrouter/anthropic/claude-sonnet-4-5", (
"Auto-detected OpenRouter should prefix model for LiteLLM routing"
)
assert call_kwargs["model"] == "anthropic/claude-sonnet-4-5", (
"Auto-detected OpenRouter must preserve native model name without openrouter/ prefix"
@pytest.mark.asyncio
async def test_openrouter_native_model_id_gets_double_prefixed() -> None:
"""Models like openrouter/free must be double-prefixed so LiteLLM strips one layer.
openrouter/free is an actual OpenRouter model ID. LiteLLM strips the first
openrouter/ for routing, so we must send openrouter/openrouter/free to ensure
the API receives openrouter/free.
"""
mock_acompletion = AsyncMock(return_value=_fake_response())
with patch("nanobot.providers.litellm_provider.acompletion", mock_acompletion):
provider = LiteLLMProvider(
api_key="sk-or-test-key",
api_base="https://openrouter.ai/api/v1",
default_model="openrouter/free",
provider_name="openrouter",
)
await provider.chat(
messages=[{"role": "user", "content": "hello"}],
model="openrouter/free",
)
call_kwargs = mock_acompletion.call_args.kwargs
assert call_kwargs["model"] == "openrouter/openrouter/free", (
"openrouter/free must become openrouter/openrouter/free — "
"LiteLLM strips one layer so the API receives openrouter/free"
)