Setup
Config
Use it
Auto-Discovery
localhost:8080 for a running TGI instance.
Provider Details
| Provider ID | tgi (custom) |
| Default port | 8080 |
| API type | openai-completions |
| Base URL | http://localhost:8080/v1 |
| API key | Not required |
| Cost | Free (self-hosted) |
Notes
- TGI is optimized for production serving with features like continuous batching and speculative decoding.
- Docker with
--gpus allis the recommended deployment method. - TGI supports the OpenAI-compatible endpoint at
/v1/chat/completions, which Polpo uses via theopenai-completionsAPI mode.