mirror of
https://github.com/vimagick/dockerfiles.git
synced 2024-11-21 17:56:53 +02:00
add llama.cpp
This commit is contained in:
parent
60f8d9f257
commit
ef1887567c
@ -404,6 +404,7 @@ A collection of delicious docker recipes.
|
||||
- [x] ipfs/kubo
|
||||
- [x] heartexlabs/label-studio
|
||||
- [x] martialblog/limesurvey
|
||||
- [x] ghcr.io/ggerganov/llama.cpp
|
||||
- [x] lldap/lldap
|
||||
- [x] mailhog/mailhog
|
||||
- [x] linuxserver/mastodon
|
||||
|
21
llama.cpp/README.md
Normal file
21
llama.cpp/README.md
Normal file
@ -0,0 +1,21 @@
|
||||
llama.cpp
|
||||
=========
|
||||
|
||||
[llama.cpp][1] is to enable LLM inference with minimal setup and
|
||||
state-of-the-art performance on a wide variety of hardware - locally and in the
|
||||
cloud.
|
||||
|
||||
```bash
|
||||
$ mkdir -p data
|
||||
|
||||
$ wget -P data https://huggingface.co/TheBloke/TinyLlama-1.1B-Chat-v1.0-GGUF/resolve/main/tinyllama-1.1b-chat-v1.0.Q2_K.gguf
|
||||
|
||||
$ docker compose up -d
|
||||
|
||||
$ curl --request POST \
|
||||
--url http://localhost:8080/completion \
|
||||
--header "Content-Type: application/json" \
|
||||
--data '{"prompt": "Building a website can be done in 10 simple steps:","n_predict": 128}'
|
||||
```
|
||||
|
||||
[1]: https://github.com/ggerganov/llama.cpp
|
9
llama.cpp/docker-compose.yml
Normal file
9
llama.cpp/docker-compose.yml
Normal file
@ -0,0 +1,9 @@
|
||||
services:
|
||||
llama.cpp:
|
||||
image: ghcr.io/ggerganov/llama.cpp:server
|
||||
command: -m /models/tinyllama-1.1b-chat-v1.0.Q2_K.gguf -c 512 --host 0.0.0.0 --port 8080
|
||||
ports:
|
||||
- "8080:8080"
|
||||
volumes:
|
||||
- ./data:/models
|
||||
restart: unless-stopped
|
Loading…
Reference in New Issue
Block a user