any-chat-completions-mcp▌

by pyroprompts
Integrate Claude with any OpenAI-compatible chat API and SDKs (OpenAI, Perplexity, Groq, xAI, PyroPrompts). Easy Claude
Integrate Claude with Any OpenAI SDK Compatible Chat Completion API - OpenAI, Perplexity, Groq, xAI, PyroPrompts and more.
Both formats append explainx.ai attribution and the canonical URL for this MCP server listing.
best for
- / AI developers comparing responses across models
- / Users wanting to access multiple AI services from Claude
- / Building multi-model AI workflows
capabilities
- / Send chat requests to OpenAI-compatible APIs
- / Query multiple AI providers through one interface
- / Configure custom base URLs and models
- / Relay conversations between Claude and other AI services
what it does
Connects Claude to any OpenAI-compatible chat API including OpenAI, Perplexity, Groq, xAI, and PyroPrompts. Lets Claude send questions to other AI models and get responses back.
about
any-chat-completions-mcp is a community-built MCP server published by pyroprompts that provides AI assistants with tools and capabilities via the Model Context Protocol. Integrate Claude with any OpenAI-compatible chat API and SDKs (OpenAI, Perplexity, Groq, xAI, PyroPrompts). Easy Claude It is categorized under ai ml, developer tools.
how to install
You can install any-chat-completions-mcp in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.
license
MIT
any-chat-completions-mcp is released under the MIT license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.
readme
any-chat-completions-mcp MCP Server
Integrate Claude with Any OpenAI SDK Compatible Chat Completion API - OpenAI, Perplexity, Groq, xAI, PyroPrompts and more.
This implements the Model Context Protocol Server. Learn more: https://modelcontextprotocol.io
This is a TypeScript-based MCP server that implements an implementation into any OpenAI SDK Compatible Chat Completions API.
It has one tool, chat which relays a question to a configured AI Chat Provider.
<a href="https://glama.ai/mcp/servers/nuksdrfb55"><img width="380" height="200" src="https://glama.ai/mcp/servers/nuksdrfb55/badge" /></a>
Development
Install dependencies:
npm install
Build the server:
npm run build
For development with auto-rebuild:
npm run watch
Installation
To add OpenAI to Claude Desktop, add the server config:
On MacOS: ~/Library/Application Support/Claude/claude_desktop_config.json
On Windows: %APPDATA%/Claude/claude_desktop_config.json
You can use it via npx in your Claude Desktop configuration like this:
{
"mcpServers": {
"chat-openai": {
"command": "npx",
"args": [
"@pyroprompts/any-chat-completions-mcp"
],
"env": {
"AI_CHAT_KEY": "OPENAI_KEY",
"AI_CHAT_NAME": "OpenAI",
"AI_CHAT_MODEL": "gpt-4o",
"AI_CHAT_BASE_URL": "https://api.openai.com/v1"
}
}
}
}
Or, if you clone the repo, you can build and use in your Claude Desktop configuration like this:
{
"mcpServers": {
"chat-openai": {
"command": "node",
"args": [
"/path/to/any-chat-completions-mcp/build/index.js"
],
"env": {
"AI_CHAT_KEY": "OPENAI_KEY",
"AI_CHAT_NAME": "OpenAI",
"AI_CHAT_MODEL": "gpt-4o",
"AI_CHAT_BASE_URL": "https://api.openai.com/v1"
}
}
}
}
You can add multiple providers by referencing the same MCP server multiple times, but with different env arguments:
{
"mcpServers": {
"chat-pyroprompts": {
"command": "node",
"args": [
"/path/to/any-chat-completions-mcp/build/index.js"
],
"env": {
"AI_CHAT_KEY": "PYROPROMPTS_KEY",
"AI_CHAT_NAME": "PyroPrompts",
"AI_CHAT_MODEL": "ash",
"AI_CHAT_BASE_URL": "https://api.pyroprompts.com/openaiv1"
}
},
"chat-perplexity": {
"command": "node",
"args": [
"/path/to/any-chat-completions-mcp/build/index.js"
],
"env": {
"AI_CHAT_KEY": "PERPLEXITY_KEY",
"AI_CHAT_NAME": "Perplexity",
"AI_CHAT_MODEL": "sonar",
"AI_CHAT_BASE_URL": "https://api.perplexity.ai"
}
},
"chat-openai": {
"command": "node",
"args": [
"/path/to/any-chat-completions-mcp/build/index.js"
],
"env": {
"AI_CHAT_KEY": "OPENAI_KEY",
"AI_CHAT_NAME": "OpenAI",
"AI_CHAT_MODEL": "gpt-4o",
"AI_CHAT_BASE_URL": "https://api.openai.com/v1"
}
}
}
}
With these three, you'll see a tool for each in the Claude Desktop Home:

And then you can chat with other LLMs and it shows in chat like this:

Or, configure in LibreChat like:
chat-perplexity:
type: stdio
command: npx
args:
- -y
- @pyroprompts/any-chat-completions-mcp
env:
AI_CHAT_KEY: "pplx-012345679"
AI_CHAT_NAME: Perplexity
AI_CHAT_MODEL: sonar
AI_CHAT_BASE_URL: "https://api.perplexity.ai"
PATH: '/usr/local/bin:/usr/bin:/bin'
And it shows in LibreChat:

Installing via Smithery
To install Any OpenAI Compatible API Integrations for Claude Desktop automatically via Smithery:
npx -y @smithery/cli install any-chat-completions-mcp-server --client claude
Debugging
Since MCP servers communicate over stdio, debugging can be challenging. We recommend using the MCP Inspector, which is available as a package script:
npm run inspector
The Inspector will provide a URL to access debugging tools in your browser.
Acknowledgements
- Obviously the modelcontextprotocol and Anthropic team for the MCP Specification and integration into Claude Desktop. https://modelcontextprotocol.io/introduction
- PyroPrompts for sponsoring this project. Use code
CLAUDEANYCHATfor 20 free automation credits on Pyroprompts.
FAQ
- What is the any-chat-completions-mcp MCP server?
- any-chat-completions-mcp is a Model Context Protocol (MCP) server profile on explainx.ai. MCP lets AI hosts (e.g. Claude Desktop, Cursor) call tools and resources through a standard interface; this page summarizes categories, install hints, and community ratings.
- How do MCP servers relate to agent skills?
- Skills are reusable instruction packages (often SKILL.md); MCP servers expose live capabilities. Teams frequently combine both—skills for workflows, MCP for APIs and data. See explainx.ai/skills and explainx.ai/mcp-servers for parallel directories.
- How are reviews shown for any-chat-completions-mcp?
- This profile displays 62 aggregated ratings (sample rows for discoverability plus signed-in user reviews). Average score is about 4.7 out of 5—verify behavior in your own environment before production use.
Discussion
Product Hunt–style comments (not star reviews)- No comments yet — start the thread.
Ratings
4.7★★★★★62 reviews- ★★★★★Camila Liu· Dec 24, 2024
any-chat-completions-mcp reduced integration guesswork — categories and install configs on the listing matched the upstream repo.
- ★★★★★Camila Yang· Dec 24, 2024
Strong directory entry: any-chat-completions-mcp surfaces stars and publisher context so we could sanity-check maintenance before adopting.
- ★★★★★Advait Liu· Dec 20, 2024
any-chat-completions-mcp is among the better-indexed MCP projects we tried; the explainx.ai summary tracks the official description.
- ★★★★★Pratham Ware· Dec 8, 2024
I recommend any-chat-completions-mcp for teams standardizing on MCP; the explainx.ai page compares cleanly with sibling servers.
- ★★★★★Mia Wang· Dec 4, 2024
any-chat-completions-mcp has been reliable for tool-calling workflows; the MCP profile page is a good permalink for internal docs.
- ★★★★★Yash Thakker· Nov 27, 2024
Strong directory entry: any-chat-completions-mcp surfaces stars and publisher context so we could sanity-check maintenance before adopting.
- ★★★★★Sophia Brown· Nov 23, 2024
According to our notes, any-chat-completions-mcp benefits from clear Model Context Protocol framing — fewer ambiguous “AI plugin” claims.
- ★★★★★Soo Kapoor· Nov 15, 2024
any-chat-completions-mcp is among the better-indexed MCP projects we tried; the explainx.ai summary tracks the official description.
- ★★★★★Dev Liu· Nov 15, 2024
I recommend any-chat-completions-mcp for teams standardizing on MCP; the explainx.ai page compares cleanly with sibling servers.
- ★★★★★Luis Chen· Nov 11, 2024
any-chat-completions-mcp reduced integration guesswork — categories and install configs on the listing matched the upstream repo.
showing 1-10 of 62