Senior Consult MCP
Let your AI agent consult top-tier models for expert technical guidance.
An MCP server that connects your AI agent to expert models (Claude, GPT, Gemini, DeepSeek, Z.ai) for architectural advice, code reviews, and complex problem-solving.
Why?
Your AI agent is great, but sometimes it needs a second opinion. This MCP server lets it consult specialized models for high-stakes decisions without switching context.
Tools
| Tool | Description |
|---|---|
| ask_senior | General technical consultation |
| code_review | Security, performance & quality checks |
| architecture_advice | System design guidance |
| reset_history | Clear conversation memory |
Quick Start
1. Configure MCP Client
Add to your MCP settings (Claude Desktop, Cursor, Windsurf, etc.):
{
"mcpServers": {
"senior-consult": {
"command": "npx",
"args": ["-y", "senior-consult-mcp"],
"env": {
"ANTHROPIC_API_KEY": "your-key"
}
}
}
}
2. Use It
- "Ask senior to review this authentication flow for security issues"
- "Get architecture advice on microservices vs monolith for our scale"
- "Code review this function for potential memory leaks"
Installation
Option A: npx (Recommended)
# No install needed - runs directly
npx -y senior-consult-mcp
Option B: Global
npm install -g senior-consult-mcp
Option C: Local Development
git clone https://github.com/silkyland/senior-consult-mcp.git
cd senior-consult-mcp
npm install
npm run build
npm start
Environment Variables
API Keys
Set at least one provider key:
ANTHROPIC_API_KEY=sk-ant-... # Claude
OPENAI_API_KEY=sk-... # GPT
GEMINI_API_KEY=... # Gemini
DEEPSEEK_API_KEY=sk-... # DeepSeek
ZAI_API_KEY=... # Z.ai (or ZHIPU_API_KEY)
OPENAI_COMPATIBLE_API_KEY=... # Custom endpoint
Model Configuration (Optional)
Override default models:
# Claude
CLAUDE_MODEL=claude-sonnet-4-5
CLAUDE_URL=https://api.anthropic.com/v1/messages
# OpenAI
OPENAI_MODEL=gpt-4o
OPENAI_URL=https://api.openai.com/v1/chat/completions
# Gemini
GEMINI_MODEL=gemini-2.5-flash
# DeepSeek
DEEPSEEK_MODEL=deepseek-chat
DEEPSEEK_URL=https://api.deepseek.com/chat/completions
# Z.ai
ZAI_MODEL=glm-4-plus
ZAI_URL=https://open.bigmodel.cn/api/paas/v4/chat/completions
# Custom OpenAI-compatible
OPENAI_COMPATIBLE_MODEL=your-model
OPENAI_COMPATIBLE_URL=https://your-endpoint/v1/chat/completions
Full Example Config
{
"mcpServers": {
"senior-consult": {
"command": "npx",
"args": ["-y", "senior-consult-mcp"],
"env": {
"ANTHROPIC_API_KEY": "sk-ant-xxx",
"DEEPSEEK_API_KEY": "sk-xxx",
"GEMINI_API_KEY": "xxx",
"CLAUDE_MODEL": "claude-sonnet-4-5",
"DEEPSEEK_MODEL": "deepseek-reasoner"
}
}
}
}
Features
- Multi-Provider - Claude, GPT, Gemini, DeepSeek, Z.ai, custom endpoints
- Smart Memory - Token-aware history with auto-trimming
- Optimized Prompts - Tuned for concise, actionable advice
- Zero Config - Works out of the box with sensible defaults
Contributing
We welcome contributions! Here's how you can help:
Ways to Contribute
- Report bugs - Open an issue with reproduction steps
- Suggest features - Share ideas in discussions
- Improve docs - Fix typos, add examples
- Submit PRs - Bug fixes, new providers, optimizations
Development Setup
# Clone & install
git clone https://github.com/silkyland/senior-consult-mcp.git
cd senior-consult-mcp
npm install
# Build & test
npm run build
# Run locally
npm start
Pull Request Process
- Fork the repo
- Create feature branch (git checkout -b feature/awesome)
- Make changes & add tests
- Commit (git commit -m 'Add awesome feature')
- Push (git push origin feature/awesome)
- Open PR with clear description
Code Style
- TypeScript strict mode
- ESLint + Prettier
- Meaningful commit messages
- Tests for new features
License
MIT
Questions? Open an issue or start a discussion.
Like this project? Give it a star to show support!