Currently in the field of open-source large models, Llama3 is undoubtedly the strongest! This time Meta not only released two powerful large models, 8B and 70B, for free, but also 400B is about to be released, which can compete with GPT-4! Today we will introduce 3 local deployment methods, simple and easy to understand, very suitable for beginners!
1. GPT4All: Suitable for low configuration users, can run on CPU/GPU Download here
2. LMstudio, supports downloading multiple models 8B/70B, etc., more model choices! Download here
Reminder: If you are not overseas and cannot download the model, Llama 3 large models will be uploaded to a cloud drive later [Download here] to be updated later...
3. Ollama, supports multiple platforms! Can run on Windows/Mac/Linux Download here
And open the WSL feature on Windows
Install Docker Desktop: Go to download here
Then install the webUI locally
- Run on CPU:
docker run -d -p 3000:8080 --add-host=host.docker.internal:host-gateway -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:main
- Support GPU operation:
docker run -d -p 3000:8080 --gpus=all -v ollama:/root/.ollama -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:ollama