litellm_config.yaml
Detailed docs on how to setup litellm config - here
http://localhost:4000
Settings > Model Providers > OpenAI-API-compatible
, fill in:
gpt-4
http://localhost:4000
Enter the base URL where the LiteLLM service is accessible.
Chat
4096
The maximum context length of the model. If unsure, use the default value of 4096.
4096
The maximum number of tokens returned by the model. If there are no specific requirements for the model, this can be consistent with the model context length.
Yes
Check this option if the model supports image understanding (multimodal), like gpt4-o
.