An MCP server that gives your AI assistant access to other LLMs. Query hundreds of models across OpenAI, Google, and OpenRouter through a single interface — with model discovery, usage-based favourites, and automatic enrichment (Elo ratings, pricing, knowledge cutoffs).
- Install uv (see Installation Guide for Linux/Windows):
brew install uv- Add to your Claude Desktop config (
~/Library/Application Support/Claude/claude_desktop_config.json):
{
"mcpServers": {
"ask-another": {
"command": "uvx",
"args": ["--from", "git+https://github.com/matthewgjohnson/ask-another", "ask-another"],
"env": {
"PROVIDER_OPENAI": "openai;sk-your-openai-key",
"PROVIDER_GEMINI": "gemini;your-google-key",
"PROVIDER_OPENROUTER": "openrouter;sk-or-your-openrouter-key"
}
}
}
}At least one provider is required. Remove any you don't have keys for.
- Restart Claude Desktop.
- "Ask GPT-5.2 what it thinks about this architecture" →
completion - "What Gemini models are available?" →
search_models - "Research the state of WebAssembly in 2026" →
start_research - "Generate a logo for my project" →
generate_image - "Note that DeepSeek is good for creative writing" →
annotate_models
See all 10 tools in the Reference.
- Installation Guide — Linux/Windows setup, Claude Code, Cursor, Windsurf, and other MCP clients, all configuration options
- Reference — tools, annotations & enrichment, architecture, development