Skip to content

docs: add Text Generation Inference (TGI) section to using_different_models#2209

Open
santoshray02 wants to merge 1 commit intohuggingface:mainfrom
santoshray02:docs/add-tgi-section
Open

docs: add Text Generation Inference (TGI) section to using_different_models#2209
santoshray02 wants to merge 1 commit intohuggingface:mainfrom
santoshray02:docs/add-tgi-section

Conversation

@santoshray02
Copy link
Copy Markdown

Closes #1567

Adds a "Using Text Generation Inference (TGI) Models" section to the "Using different models" guide, following the same pattern as the existing Gemini / OpenRouter / Grok sections.

The section covers three common scenarios:

  • Pointing LiteLLMModel at a Hugging Face Inference Endpoint
  • Authenticating with HF_TOKEN for private endpoints
  • Running TGI locally via Docker and connecting to http://localhost:8080/v1/

Details verified against:

  • LiteLLM's huggingface provider docs (model_id prefix huggingface/tgi, api_base with /v1/ suffix for the OpenAI-compatible Messages API)

  • TGI's quicktour (Docker image tag 3.3.5, port mapping 8080:80)

    Docs-only change; no code touched.

…models

Document how to point LiteLLMModel at a TGI endpoint — covering Hugging Face
Inference Endpoints, authenticated private endpoints, and running TGI locally
via Docker. Verified against LiteLLM's huggingface provider docs and TGI's
quicktour (image tag 3.3.5, /v1/ Messages API path).

Closes huggingface#1567
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

DOC: text generation interface (TGI) documentation is missing.

1 participant