1
0
forked from MCP/llm-fusion-mcp
llm-fusion-mcp/.env.example
Ryan Malloy c335ba0e1e Initial commit: LLM Fusion MCP Server
- Unified access to 4 major LLM providers (Gemini, OpenAI, Anthropic, Grok)
- Real-time streaming support across all providers
- Multimodal capabilities (text, images, audio)
- Intelligent document processing with smart chunking
- Production-ready with health monitoring and error handling
- Full OpenAI ecosystem integration (Assistants, DALL-E, Whisper)
- Vector embeddings and semantic similarity
- Session-based API key management
- Built with FastMCP and modern Python tooling

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-09-05 05:47:51 -06:00

52 lines
1.6 KiB
Plaintext

# LLM Fusion MCP - Environment Configuration Example
# Copy this file to .env and add your API keys
# =============================================================================
# LLM PROVIDER API KEYS (Add at least one)
# =============================================================================
# Google Gemini (Recommended - Primary Provider)
# Get your key from: https://aistudio.google.com/app/apikey
GOOGLE_API_KEY=your_google_api_key_here
# OpenAI (Optional - GPT models, DALL-E, Whisper)
# Get your key from: https://platform.openai.com/api-keys
OPENAI_API_KEY=your_openai_api_key_here
# Anthropic (Optional - Claude models)
# Get your key from: https://console.anthropic.com/
ANTHROPIC_API_KEY=your_anthropic_api_key_here
# xAI Grok (Optional - Grok models)
# Get your key from: https://console.x.ai/
XAI_API_KEY=your_xai_api_key_here
# =============================================================================
# SERVER CONFIGURATION (Optional)
# =============================================================================
# Server mode (development, production)
SERVER_MODE=development
# Logging level (DEBUG, INFO, WARNING, ERROR)
LOG_LEVEL=INFO
# Maximum file size for analysis (in MB)
MAX_FILE_SIZE_MB=50
# Request timeout (in seconds)
REQUEST_TIMEOUT=300
# =============================================================================
# PERFORMANCE SETTINGS (Optional)
# =============================================================================
# Model cache timeout (in minutes)
MODEL_CACHE_TIMEOUT=5
# Maximum concurrent requests
MAX_CONCURRENT_REQUESTS=10
# Rate limiting (requests per minute per provider)
RATE_LIMIT_PER_MINUTE=60