Supported Providers
Google Gemini
Fast, reliable, generous free tier
OpenAI GPT
High-quality, detailed documentation
OpenRouter
Access to 100+ models through one API
Azure OpenAI
Enterprise-grade with enhanced security
AWS Bedrock
AWS-hosted models with enterprise features
Ollama
Local, private, cost-free AI models
Google Gemini
Google’s Gemini models offer excellent performance with generous free tiers, making them ideal for getting started.Setup
Get API Key
- Visit Google AI Studio
- Sign in with your Google account
- Click “Create API Key”
- Copy the generated key (starts with
AIza)
API key generated and copied
Available Models
- gemini-2.0-flash (Recommended)
- gemini-1.5-flash
- gemini-1.0-pro
Best for: Most documentation tasks
- Speed: Very fast (1-3 seconds per request)
- Quality: Excellent for code analysis
- Context: 1M+ tokens input, 8K output
- Cost: Free tier: 15 RPM, 1M TPM
- General repository documentation
- Quick prototyping and testing
- Regular development workflows
- Small to medium repositories
Optimization Tips
Rate Limit Management
Rate Limit Management
Free tier limits:
- 15 requests per minute (Flash models)
- 60 requests per minute (Pro models)
- 32,000 tokens per minute
Context Window Optimization
Context Window Optimization
Gemini models have large context windows. Optimize usage:
- Large repositories: Use full context for better understanding
- Complex files: Include more surrounding context
- API documentation: Include related endpoints together
OpenAI
OpenAI’s GPT models provide exceptional quality documentation with advanced reasoning capabilities.Setup
Create Account & Get Credits
- Sign up at OpenAI Platform
- Add payment method (required for API access)
- Purchase credits or set up billing
- Navigate to API Keys
Generate API Key
- Click “Create new secret key”
- Add a name (e.g., “DeepWiki-Development”)
- Copy the key (starts with
sk-) - Store securely (you won’t see it again)
API key generated and stored securely
Available Models
- gpt-5 (Latest - Default)
- gpt-4o (Previous Default)
- gpt-4.1
- o1 Series (Reasoning Models)
- o4-mini (Cost-Effective)
Best for: State-of-the-art documentation generation with advanced reasoning
- Speed: Fast to moderate (3-8 seconds per request)
- Quality: Next-generation AI capabilities with superior understanding
- Context: 256K tokens input/output (estimated)
- Temperature: 1.0 (default for creative yet accurate responses)
- Availability: Rolling out to API users (check availability in your region)
- Cutting-edge documentation projects
- Complex architectural documentation
- Multi-language codebases
- Advanced technical analysis
- Projects requiring latest AI capabilities
GPT-5 is now the default model in DeepWiki as of commit 05693d5. Ensure your OpenAI account has access to GPT-5 API.
Cost Optimization
Token Usage Management
Token Usage Management
Monitor and optimize token consumption:Cost calculation example:
- Large repository: ~200K input tokens, 8K output tokens
- GPT-5 cost: Pricing to be announced (expected similar or slightly higher than GPT-4o)
- GPT-4o cost: 0.48 output = $3.48 per generation
- Monthly usage (10 repos): ~$35-50/month (estimated)
Model Selection Strategy
Model Selection Strategy
Match model to task complexity:
- Simple projects: Use o4-mini for cost savings
- Standard projects: Use gpt-5 for latest capabilities or gpt-4o for proven reliability
- Complex analysis: Use gpt-5 for advanced reasoning or o1 series for deep insights
- Budget constraints: Start with o4-mini, upgrade if needed
- Cutting-edge needs: Use gpt-5 for state-of-the-art performance
OpenRouter
OpenRouter provides access to 100+ AI models through a single API, perfect for comparison and specialized needs.Setup
Create Account
- Sign up at OpenRouter
- Verify your email address
- Add payment method for paid models
- Navigate to the Keys section
Some models are free, others require credits. Check individual model pricing.
Generate API Key
- Click “Create Key”
- Name your key (e.g., “DeepWiki-Prod”)
- Copy the key (starts with
sk-or-) - Optionally set spending limits
OpenRouter API key generated with spending limits configured
Popular Models
- Anthropic Claude
- Google Models
- Open Source Models
- Specialized Models
Models:
anthropic/claude-3.5-sonnet, anthropic/claude-3-haikuBest for:- Excellent code analysis and explanation
- Clear, structured documentation
- Complex reasoning tasks
- Safe, helpful responses
- API documentation generation
- Code architecture explanation
- Security-focused analysis
Model Comparison Strategy
Azure OpenAI
Enterprise-grade OpenAI models with enhanced security, compliance, and control.Setup
Create Azure OpenAI Resource
- Sign in to Azure Portal
- Create new Azure OpenAI resource
- Choose region (check model availability)
- Configure pricing tier and network settings
- Wait for deployment completion
Azure OpenAI may require approval for access. Check the application status.
Deploy Models
- Go to Azure OpenAI Studio
- Navigate to Deployments
- Deploy required models (GPT-4, GPT-3.5-turbo, etc.)
- Note deployment names and endpoints
Models deployed and endpoints configured
Enterprise Features
Data Privacy & Compliance
Data Privacy & Compliance
Key benefits:
- Data processed within your Azure tenant
- No data used for model training
- GDPR, SOC 2, HIPAA compliance available
- Private networking with VNet integration
Content Filtering
Content Filtering
Built-in safety features:
- Automatic content filtering for harmful content
- Customizable filter levels
- Compliance with organizational policies
Scale & Performance
Scale & Performance
Enterprise-grade performance:
- Dedicated capacity options
- Predictable performance
- Custom rate limits
- Multi-region deployment
AWS Bedrock
AWS-hosted AI models with enterprise features and AWS service integration.Setup
AWS Account Setup
- Ensure you have an AWS account
- Enable AWS Bedrock in your region
- Request access to required models (may require approval)
- Create IAM user with Bedrock permissions
Available Models
- Anthropic Claude
- Amazon Titan
- AI21 Labs
Models:
anthropic.claude-3-sonnet-20240229-v1:0anthropic.claude-3-haiku-20240307-v1:0anthropic.claude-3-opus-20240229-v1:0
Ollama (Local Models)
Run AI models locally for complete privacy, cost control, and offline capability.Setup
Model Selection
- Code-Focused Models
- General Purpose Models
- Lightweight Options
qwen3:8b (Recommended)
- Size: 4.8GB download
- RAM: 8GB required
- Strengths: Excellent code understanding, multilingual
- Best for: Most documentation tasks
- Size: 3.8GB download
- RAM: 6GB required
- Strengths: Specialized for code generation and analysis
- Best for: Technical documentation, API docs
Performance Optimization
Hardware Requirements
Hardware Requirements
Minimum specs by model size:
- 1B-3B models: 4GB RAM, any modern CPU
- 7B-8B models: 8GB RAM, modern CPU (preferably 8+ cores)
- 13B models: 16GB RAM, high-performance CPU
- 70B+ models: 64GB+ RAM, server-grade hardware
Memory Management
Memory Management
Optimize memory usage:Model configuration:
Multi-Provider Strategy
Provider Selection Matrix
- By Project Type
- By Repository Size
- By Use Case
| Project Type | Primary | Fallback | Reason |
|---|---|---|---|
| Open Source | Google Gemini | OpenRouter | Free tier, good quality |
| Enterprise | Azure OpenAI | OpenAI | Security, compliance |
| Startup | OpenRouter | Cost optimization | |
| Research | OpenAI GPT-4o | Claude via OpenRouter | Highest quality |
| Personal | Ollama | Privacy, no cost |
Auto-Failover Configuration
Next Steps
Authorization Mode
Set up access control for your DeepWiki deployment
Generate First Wiki
Create your first repository documentation
Production Setup
Deploy with multiple providers for production use
API Integration
Integrate provider selection into your workflows