ggml-gpt4all-j
serves as the default LLM model, and all-MiniLM-L6-v2
serves as the default Embedding model, for quick local deployment.
.env
doesn’t exceed the number of CPU cores on your machine.
ggml-gpt4all-j
External access name: gpt-3.5-turbo
(This name is customizable and can be configured in models/gpt-3.5-turbo.yaml
).
all-MiniLM-L6-v2
External access name: text-embedding-ada-002
(This name is customizable and can be configured in models/embeddings.yaml
).
If you use the Dify Docker deployment method, you need to pay attention to the network configuration to ensure that the Dify container can access the endpoint of LocalAI. The Dify container cannot access localhost inside, and you need to use the host IP address.
Settings > Model Providers > LocalAI
and fill in:
Model 1: ggml-gpt4all-j
gpt-3.5-turbo
http://<your-LocalAI-endpoint-domain>:8080
, which can be a LAN IP address, like: http://192.168.1.100:8080
all-MiniLM-L6-v2
text-embedding-ada-002
If Dify is deployed via docker, fill in the host domain:http://<your-LocalAI-endpoint-domain>:8080
, which can be a LAN IP address, like:http://192.168.1.100:8080