Provider Setup
Configure LLM providers for use with InferXgate.
This guide covers setting up each LLM provider with InferXgate.
Anthropic
- Get an API key from console.anthropic.com
- Add to environment:
ANTHROPIC_API_KEY=sk-ant-api03-...
- Test:
curl -X POST http://localhost:3000/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{"model": "claude-3-haiku-20240307", "messages": [{"role": "user", "content": "Hi"}]}'
OpenAI
- Get an API key from platform.openai.com
- Add to environment:
OPENAI_API_KEY=sk-...
OPENAI_ORG_ID=org-... # Optional
- Test:
curl -X POST http://localhost:3000/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{"model": "gpt-3.5-turbo", "messages": [{"role": "user", "content": "Hi"}]}'
Google Gemini
- Get an API key from makersuite.google.com
- Add to environment:
GEMINI_API_KEY=...
- Test:
curl -X POST http://localhost:3000/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{"model": "gemini-1.5-flash", "messages": [{"role": "user", "content": "Hi"}]}'
Azure OpenAI
- Create a resource in Azure Portal
- Deploy models to your resource
- Add to environment:
AZURE_OPENAI_API_KEY=...
AZURE_OPENAI_ENDPOINT=https://your-resource.openai.azure.com
AZURE_OPENAI_API_VERSION=2024-02-01
AZURE_OPENAI_DEPLOYMENTS=gpt-4:my-gpt4-deployment
- Test with
azure-prefix:
curl -X POST http://localhost:3000/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{"model": "azure-gpt-4", "messages": [{"role": "user", "content": "Hi"}]}'
Verify All Providers
Check which providers are configured:
curl http://localhost:3000/health
{
"providers": {
"anthropic": "healthy",
"openai": "healthy",
"gemini": "healthy",
"azure": "healthy"
}
}
Multiple Keys Per Provider
For load balancing:
ANTHROPIC_API_KEYS=key1,key2,key3
OPENAI_API_KEYS=key1,key2