Thanks to visit codestin.com
Credit goes to github.com

Skip to content

Commit f24e7a7

Browse files
committed
Updated instructions
1 parent cf4931a commit f24e7a7

File tree

1 file changed

+9
-8
lines changed

1 file changed

+9
-8
lines changed

docker/README.md

+9-8
Original file line numberDiff line numberDiff line change
@@ -18,14 +18,15 @@
1818
- `docker run -e USE_MLOCK=0 -e MODEL=/var/model/<model-path> -v <model-root-path>:/var/model -t cuda_simple`
1919
where `<model-root-path>/<model-path>` is the full path to the model file on the Docker host system.
2020

21-
# "Bot-in-a-box" - a method to build a Docker image by choosing a model to be downloaded and loading into a Docker image
22-
- `cd ./auto_docker`:
23-
- `hug_model.py` - a Python utility for interactively choosing and downloading the latest `5_1` quantized models from [huggingface.co/TheBloke]( https://huggingface.co/TheBloke)
24-
- `Dockerfile` - a single OpenBLAS and CuBLAS combined Dockerfile that automatically installs a previously downloaded model `model.bin`
25-
26-
## Download a Llama Model from Hugging Face
27-
- To download a MIT licensed Llama model you can run: `python3 ./hug_model.py -a vihangd -s open_llama_7b_700bt_ggml -f ggml-model-q5_1.bin`
28-
- To select and install a restricted license Llama model run: `python3 ./hug_model.py -a TheBloke -t llama`
21+
# "Open-Llama-in-a-box" - Download a MIT licensed Open Llama model and install into a Docker image that runs an OpenBLAS-enabled llama-cpp-python server
22+
```
23+
$ cd ./open_llama
24+
./build.sh
25+
./start.sh
26+
```
27+
28+
# Manually choose your own Llama model from Hugging Face
29+
- `python3 ./hug_model.py -a TheBloke -t llama`
2930
- You should now have a model in the current directory and `model.bin` symlinked to it for the subsequent Docker build and copy step. e.g.
3031
```
3132
docker $ ls -lh *.bin

0 commit comments

Comments
 (0)