llm-fusion-mcp/.env.example
Ryan Malloy c335ba0e1e
Some checks are pending
🚀 LLM Fusion MCP - CI/CD Pipeline / 🔍 Code Quality & Testing (3.10) (push) Waiting to run
🚀 LLM Fusion MCP - CI/CD Pipeline / 🔍 Code Quality & Testing (3.11) (push) Waiting to run
🚀 LLM Fusion MCP - CI/CD Pipeline / 🔍 Code Quality & Testing (3.12) (push) Waiting to run
🚀 LLM Fusion MCP - CI/CD Pipeline / 🛡️ Security Scanning (push) Blocked by required conditions
🚀 LLM Fusion MCP - CI/CD Pipeline / 🐳 Docker Build & Push (push) Blocked by required conditions
🚀 LLM Fusion MCP - CI/CD Pipeline / 🎉 Create Release (push) Blocked by required conditions
🚀 LLM Fusion MCP - CI/CD Pipeline / 📢 Deployment Notification (push) Blocked by required conditions
Initial commit: LLM Fusion MCP Server
- Unified access to 4 major LLM providers (Gemini, OpenAI, Anthropic, Grok)
- Real-time streaming support across all providers
- Multimodal capabilities (text, images, audio)
- Intelligent document processing with smart chunking
- Production-ready with health monitoring and error handling
- Full OpenAI ecosystem integration (Assistants, DALL-E, Whisper)
- Vector embeddings and semantic similarity
- Session-based API key management
- Built with FastMCP and modern Python tooling

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-09-05 05:47:51 -06:00

52 lines
1.6 KiB
Plaintext

# LLM Fusion MCP - Environment Configuration Example
# Copy this file to .env and add your API keys
# =============================================================================
# LLM PROVIDER API KEYS (Add at least one)
# =============================================================================
# Google Gemini (Recommended - Primary Provider)
# Get your key from: https://aistudio.google.com/app/apikey
GOOGLE_API_KEY=your_google_api_key_here
# OpenAI (Optional - GPT models, DALL-E, Whisper)
# Get your key from: https://platform.openai.com/api-keys
OPENAI_API_KEY=your_openai_api_key_here
# Anthropic (Optional - Claude models)
# Get your key from: https://console.anthropic.com/
ANTHROPIC_API_KEY=your_anthropic_api_key_here
# xAI Grok (Optional - Grok models)
# Get your key from: https://console.x.ai/
XAI_API_KEY=your_xai_api_key_here
# =============================================================================
# SERVER CONFIGURATION (Optional)
# =============================================================================
# Server mode (development, production)
SERVER_MODE=development
# Logging level (DEBUG, INFO, WARNING, ERROR)
LOG_LEVEL=INFO
# Maximum file size for analysis (in MB)
MAX_FILE_SIZE_MB=50
# Request timeout (in seconds)
REQUEST_TIMEOUT=300
# =============================================================================
# PERFORMANCE SETTINGS (Optional)
# =============================================================================
# Model cache timeout (in minutes)
MODEL_CACHE_TIMEOUT=5
# Maximum concurrent requests
MAX_CONCURRENT_REQUESTS=10
# Rate limiting (requests per minute per provider)
RATE_LIMIT_PER_MINUTE=60