- Quickly start using a local or Hugging Face model
- Chat with a locally running model
- Test or interact with a model directly from chat
π Tips & Best Practices
* The model is served locally using vLLM.
* The exposed endpoint follows the OpenAI API format.
* The server must be started before sending chat requests.