FastChat
FastChat
Configure Hymalaia to use FastChat model servers.
Refer to Model Configs for how to set the environment variables for your particular deployment.
Note: While we support self hosted LLMs, you will get significantly better responses with a more powerful model like GPT-4.
What is FastChat
FastChat is a way to easily host LLMs on cli, using their web client, or as an API server. For the Hymalaia use case we will focus on interfacing with the model through the API server. See here for more information: FastChat OpenAI API Documentation
In this case, we use LiteLLM’s custom model server option. See here for more information: LiteLLM Custom OpenAI Proxy Documentation
Set Hymalaia to use FastChat Server
On the LLM page in the Admin Panel add a Custom LLM Provider with the following settings. Note that the Provider Name is OpenAI, since FastChat provides an OpenAI compatible API.
Hints:
- To point to other Docker containers running locally (e.g. accessible at http://localhost), use
http://host.docker.internal
. - Don’t forget to include the
/v1
in the API base.
Environment Variables
You may also want to update some of the environment variables depending on your model choice / how you’re running FastChat (e.g. on CPU vs GPU):