Setting OPENAI_REVERSE_PROXY in LibreChat | Generated by AI
Got it š Youāre editing LibreChatās .env.
That file is exactly where you set OPENAI_REVERSE_PROXY.
1. What the variable does
OPENAI_REVERSE_PROXY is used if you want LibreChat to route OpenAI API requests through a proxy server (for example, if you run a self-hosted reverse proxy to add logging, caching, regional routing, or to bypass restrictions).
2. How to set it
In your .env, just assign it the base URL of your reverse proxy. For example:
OPENAI_REVERSE_PROXY=https://your-proxy.example.com/v1
Notes:
- The URL should point to your proxy endpoint that mimics OpenAIās
/v1API. - Donāt include trailing
/unless your proxy requires it. - Keep it empty (
OPENAI_REVERSE_PROXY=) if you donāt use a proxy.
3. Apply the changes
After editing .env, restart your LibreChat containers so they pick up the new value:
docker compose down
docker compose up -d --force-recreate
4. Verify
You can check inside the running container:
docker compose exec librechat env | grep OPENAI_REVERSE_PROXY
It should show the value you set.
Do you already have a reverse proxy URL (like https://openrouter.ai/api/v1 or your own Nginx instance), or do you want me to show you how to set one up for LibreChat?