provider-configuration
Provider Configuration
This skill helps you configure LLM providers for Letta, supporting both:
- Self-hosted setup - Environment variables for Docker/local deployments
- BYOK via API - Programmatic provider management for Cloud and self-hosted
Quick Start
Add Provider via API
python scripts/setup_provider.py --type openai --api-key sk-...
Generate .env File for Docker
python scripts/generate_env.py --providers openai,anthropic,ollama
Validate Provider Credentials
python scripts/validate_provider.py --provider-id provider-xxx
Common Workflows
Workflow 1: Add OpenAI BYOK Provider
# Using the setup script
python scripts/setup_provider.py \
--type openai \
--name "My OpenAI" \
--api-key sk-your-key-here
# Or via REST API directly
curl -X POST http://localhost:8283/v1/providers \
-H "Content-Type: application/json" \
-d '{
"name": "My OpenAI",
"provider_type": "openai",
"api_key": "sk-your-key-here"
}'
Workflow 2: Configure Self-Hosted Server
# Generate environment file
python scripts/generate_env.py --providers openai,anthropic,ollama --output .env
# Use with Docker
docker run -p 8283:8283 --env-file .env letta/letta:latest
Workflow 3: Validate and Refresh Provider
# Check credentials
python scripts/validate_provider.py --provider-id provider-abc123
# Refresh available models
curl -X PATCH http://localhost:8283/v1/providers/provider-abc123/refresh
Reference Documentation
Common Providers
For detailed OpenAI, Anthropic, Azure OpenAI, and Google AI configuration: Read references/common_providers.md
Self-Hosted Providers
For Ollama, vLLM, LM Studio, and other local providers: Read references/self_hosted_providers.md
All Provider Types
For comprehensive reference covering all 20+ supported providers: Read references/all_providers.md
Environment Variables
For Docker setup and multi-provider configuration: Read references/environment_variables.md
Scripts
setup_provider.py
Add providers programmatically via REST API.
Required flags:
--type- Provider type (openai, anthropic, azure, etc.)--api-key- API key for the provider
Optional flags:
--name- Display name (defaults to provider type)--base-url- Custom endpoint URL--api-version- API version (for Azure)--region- Region (for Bedrock)--access-key- Access key (for AWS)--server- Letta server URL (default: http://localhost:8283)
validate_provider.py
Check if provider credentials are valid and list available models.
Required flags:
--provider-id- Provider ID to validate
Optional flags:
--server- Letta server URL (default: http://localhost:8283)
generate_env.py
Generate .env file for Docker deployments.
Required flags:
--providers- Comma-separated list of providers (openai,anthropic,ollama)
Optional flags:
--output- Output file path (default: .env)--ollama-url- Ollama endpoint URL (default: http://host.docker.internal:11434)
Provider Types
Supported provider types:
openai- OpenAI (GPT models)anthropic- Anthropic (Claude models)azure- Azure OpenAIgoogle_ai- Google AI (Gemini)google_vertex- Google Vertex AIollama- Ollama (local models)groq- Groqdeepseek- DeepSeekxai- xAI (Grok)zai- Z.aitogether- Together AImistral- Mistral AIcerebras- Cerebrasbedrock- AWS Bedrockvllm- vLLMsglang- SGLanghugging_face- Hugging Facelmstudio_openai- LM Studio (OpenAI compatible)minimax- MiniMaxchatgpt_oauth- ChatGPT OAuth
For detailed configuration of each provider, see the reference files.
Notes
- All API-created providers are automatically BYOK (bring your own key)
- Provider IDs are prefixed with
provider- - After creating a provider, models are auto-discovered
- Use
PATCH /v1/providers/{provider_id}/refreshto rescan models - Environment variables are only for self-hosted deployments
- Letta Cloud uses BYOK providers exclusively (no env vars)