feat: add MiniMax provider support via LiteLLM
This commit is contained in:
@@ -80,6 +80,7 @@ class ProvidersConfig(BaseModel):
|
||||
vllm: ProviderConfig = Field(default_factory=ProviderConfig)
|
||||
gemini: ProviderConfig = Field(default_factory=ProviderConfig)
|
||||
moonshot: ProviderConfig = Field(default_factory=ProviderConfig)
|
||||
minimax: ProviderConfig = Field(default_factory=ProviderConfig)
|
||||
aihubmix: ProviderConfig = Field(default_factory=ProviderConfig) # AiHubMix API gateway
|
||||
|
||||
|
||||
@@ -139,14 +140,15 @@ class Config(BaseSettings):
|
||||
"openai": p.openai, "gpt": p.openai, "gemini": p.gemini,
|
||||
"zhipu": p.zhipu, "glm": p.zhipu, "zai": p.zhipu,
|
||||
"dashscope": p.dashscope, "qwen": p.dashscope,
|
||||
"groq": p.groq, "moonshot": p.moonshot, "kimi": p.moonshot, "vllm": p.vllm,
|
||||
"groq": p.groq, "moonshot": p.moonshot, "kimi": p.moonshot,
|
||||
"minimax": p.minimax, "vllm": p.vllm,
|
||||
}
|
||||
for kw, provider in keyword_map.items():
|
||||
if kw in model and provider.api_key:
|
||||
return provider
|
||||
# Fallback: gateways first (can serve any model), then specific providers
|
||||
all_providers = [p.openrouter, p.aihubmix, p.anthropic, p.openai, p.deepseek,
|
||||
p.gemini, p.zhipu, p.dashscope, p.moonshot, p.vllm, p.groq]
|
||||
p.gemini, p.zhipu, p.dashscope, p.moonshot, p.minimax, p.vllm, p.groq]
|
||||
return next((pr for pr in all_providers if pr.api_key), None)
|
||||
|
||||
def get_api_key(self, model: str | None = None) -> str | None:
|
||||
|
||||
Reference in New Issue
Block a user