What is FastChat
FastChat is a way to easily host LLMs on cli, using their web client, or as an API server. For the RECAP use case we will focus on interfacing with the model through the API server. See here for more information: https://github.com/lm-sys/FastChat/blob/main/docs/openai_api.md In this case, we use LiteLLM’s custom model server option. See here for more information: https://litellm.vercel.app/docs/providers/custom_openai_proxySet RECAP to use FastChat Server
On theLLM
page in the Admin Panel add a Custom LLM Provider
with the following settings.
Note that the Provider Name
is OpenAI, since FastChat provides an OpenAI compatible.
Hints:
- To point to other Docker containers running locally (e.g. accessible at http://localhost), use http://host.docker.internal.
Don’t forget to include the /v1 in the API base.

