Supported Providers
Google Gemini
Fast, reliable, generous free tier
OpenAI GPT
High-quality, detailed documentation
OpenRouter
Access to 100+ models through one API
Azure OpenAI
Enterprise-grade with enhanced security
AWS Bedrock
AWS-hosted models with enterprise features
Ollama
Local, private, cost-free AI models
Google Gemini
Google’s Gemini models offer excellent performance with generous free tiers, making them ideal for getting started.Setup
1
Get API Key
- Visit Google AI Studio
- Sign in with your Google account
- Click “Create API Key”
- Copy the generated key (starts with
AIza
)
API key generated and copied
2
Configure Environment
Add to your
.env
file:Never commit API keys to version control. Add
.env
to your .gitignore
.3
Verify Setup
Test the configuration by starting DeepWiki:
Available Models
Best for: Most documentation tasks
- Speed: Very fast (1-3 seconds per request)
- Quality: Excellent for code analysis
- Context: 1M+ tokens input, 8K output
- Cost: Free tier: 15 RPM, 1M TPM
- General repository documentation
- Quick prototyping and testing
- Regular development workflows
- Small to medium repositories
Optimization Tips
Rate Limit Management
Rate Limit Management
Free tier limits:
- 15 requests per minute (Flash models)
- 60 requests per minute (Pro models)
- 32,000 tokens per minute
Context Window Optimization
Context Window Optimization
Gemini models have large context windows. Optimize usage:
- Large repositories: Use full context for better understanding
- Complex files: Include more surrounding context
- API documentation: Include related endpoints together
OpenAI
OpenAI’s GPT models provide exceptional quality documentation with advanced reasoning capabilities.Setup
1
Create Account & Get Credits
- Sign up at OpenAI Platform
- Add payment method (required for API access)
- Purchase credits or set up billing
- Navigate to API Keys
OpenAI requires a paid account. Free ChatGPT accounts cannot access the API.
2
Generate API Key
- Click “Create new secret key”
- Add a name (e.g., “DeepWiki-Development”)
- Copy the key (starts with
sk-
) - Store securely (you won’t see it again)
API key generated and stored securely
3
Configure Environment
Available Models
Best for: State-of-the-art documentation generation with advanced reasoning
- Speed: Fast to moderate (3-8 seconds per request)
- Quality: Next-generation AI capabilities with superior understanding
- Context: 256K tokens input/output (estimated)
- Temperature: 1.0 (default for creative yet accurate responses)
- Availability: Rolling out to API users (check availability in your region)
- Cutting-edge documentation projects
- Complex architectural documentation
- Multi-language codebases
- Advanced technical analysis
- Projects requiring latest AI capabilities
GPT-5 is now the default model in DeepWiki as of commit 05693d5. Ensure your OpenAI account has access to GPT-5 API.
Cost Optimization
Token Usage Management
Token Usage Management
Monitor and optimize token consumption:Cost calculation example:
- Large repository: ~200K input tokens, 8K output tokens
- GPT-5 cost: Pricing to be announced (expected similar or slightly higher than GPT-4o)
- GPT-4o cost: 0.48 output = $3.48 per generation
- Monthly usage (10 repos): ~$35-50/month (estimated)
Model Selection Strategy
Model Selection Strategy
Match model to task complexity:
- Simple projects: Use o4-mini for cost savings
- Standard projects: Use gpt-5 for latest capabilities or gpt-4o for proven reliability
- Complex analysis: Use gpt-5 for advanced reasoning or o1 series for deep insights
- Budget constraints: Start with o4-mini, upgrade if needed
- Cutting-edge needs: Use gpt-5 for state-of-the-art performance
OpenRouter
OpenRouter provides access to 100+ AI models through a single API, perfect for comparison and specialized needs.Setup
1
Create Account
- Sign up at OpenRouter
- Verify your email address
- Add payment method for paid models
- Navigate to the Keys section
Some models are free, others require credits. Check individual model pricing.
2
Generate API Key
- Click “Create Key”
- Name your key (e.g., “DeepWiki-Prod”)
- Copy the key (starts with
sk-or-
) - Optionally set spending limits
OpenRouter API key generated with spending limits configured
3
Configure Environment
Popular Models
Models:
anthropic/claude-3.5-sonnet
, anthropic/claude-3-haiku
Best for:- Excellent code analysis and explanation
- Clear, structured documentation
- Complex reasoning tasks
- Safe, helpful responses
- API documentation generation
- Code architecture explanation
- Security-focused analysis
Model Comparison Strategy
1
Baseline Generation
Start with a reliable, fast model:
2
A/B Testing
Compare models for your specific use case:
3
Optimization
Select the best model based on results:
Azure OpenAI
Enterprise-grade OpenAI models with enhanced security, compliance, and control.Setup
1
Create Azure OpenAI Resource
- Sign in to Azure Portal
- Create new Azure OpenAI resource
- Choose region (check model availability)
- Configure pricing tier and network settings
- Wait for deployment completion
Azure OpenAI may require approval for access. Check the application status.
2
Deploy Models
- Go to Azure OpenAI Studio
- Navigate to Deployments
- Deploy required models (GPT-4, GPT-3.5-turbo, etc.)
- Note deployment names and endpoints
Models deployed and endpoints configured
3
Get Configuration Details
Collect the required information:
- Endpoint:
https://your-resource.openai.azure.com
- API Key: From resource keys section
- API Version: e.g.,
2024-02-15-preview
Enterprise Features
Data Privacy & Compliance
Data Privacy & Compliance
Key benefits:
- Data processed within your Azure tenant
- No data used for model training
- GDPR, SOC 2, HIPAA compliance available
- Private networking with VNet integration
Content Filtering
Content Filtering
Built-in safety features:
- Automatic content filtering for harmful content
- Customizable filter levels
- Compliance with organizational policies
Scale & Performance
Scale & Performance
Enterprise-grade performance:
- Dedicated capacity options
- Predictable performance
- Custom rate limits
- Multi-region deployment
AWS Bedrock
AWS-hosted AI models with enterprise features and AWS service integration.Setup
1
AWS Account Setup
- Ensure you have an AWS account
- Enable AWS Bedrock in your region
- Request access to required models (may require approval)
- Create IAM user with Bedrock permissions
Bedrock is not available in all AWS regions. Check regional availability.
2
Configure IAM Permissions
Create IAM policy for Bedrock access:
3
Configure Environment
AWS credentials configured and Bedrock access verified
Available Models
Models:
anthropic.claude-3-sonnet-20240229-v1:0
anthropic.claude-3-haiku-20240307-v1:0
anthropic.claude-3-opus-20240229-v1:0
Ollama (Local Models)
Run AI models locally for complete privacy, cost control, and offline capability.Setup
1
Install Ollama
2
Pull Models
Download models you want to use:
Models downloaded and verified
3
Configure DeepWiki
Model Selection
qwen3:8b (Recommended)
- Size: 4.8GB download
- RAM: 8GB required
- Strengths: Excellent code understanding, multilingual
- Best for: Most documentation tasks
- Size: 3.8GB download
- RAM: 6GB required
- Strengths: Specialized for code generation and analysis
- Best for: Technical documentation, API docs
Performance Optimization
Hardware Requirements
Hardware Requirements
Minimum specs by model size:
- 1B-3B models: 4GB RAM, any modern CPU
- 7B-8B models: 8GB RAM, modern CPU (preferably 8+ cores)
- 13B models: 16GB RAM, high-performance CPU
- 70B+ models: 64GB+ RAM, server-grade hardware
Memory Management
Memory Management
Optimize memory usage:Model configuration:
Multi-Provider Strategy
Provider Selection Matrix
Project Type | Primary | Fallback | Reason |
---|---|---|---|
Open Source | Google Gemini | OpenRouter | Free tier, good quality |
Enterprise | Azure OpenAI | OpenAI | Security, compliance |
Startup | OpenRouter | Cost optimization | |
Research | OpenAI GPT-4o | Claude via OpenRouter | Highest quality |
Personal | Ollama | Privacy, no cost |