mirror of
https://github.com/vimagick/dockerfiles.git
synced 2025-12-21 13:23:02 +01:00
add llama.cpp
This commit is contained in:
@@ -404,6 +404,7 @@ A collection of delicious docker recipes.
|
|||||||
- [x] ipfs/kubo
|
- [x] ipfs/kubo
|
||||||
- [x] heartexlabs/label-studio
|
- [x] heartexlabs/label-studio
|
||||||
- [x] martialblog/limesurvey
|
- [x] martialblog/limesurvey
|
||||||
|
- [x] ghcr.io/ggerganov/llama.cpp
|
||||||
- [x] lldap/lldap
|
- [x] lldap/lldap
|
||||||
- [x] mailhog/mailhog
|
- [x] mailhog/mailhog
|
||||||
- [x] linuxserver/mastodon
|
- [x] linuxserver/mastodon
|
||||||
|
|||||||
21
llama.cpp/README.md
Normal file
21
llama.cpp/README.md
Normal file
@@ -0,0 +1,21 @@
|
|||||||
|
llama.cpp
|
||||||
|
=========
|
||||||
|
|
||||||
|
[llama.cpp][1] is to enable LLM inference with minimal setup and
|
||||||
|
state-of-the-art performance on a wide variety of hardware - locally and in the
|
||||||
|
cloud.
|
||||||
|
|
||||||
|
```bash
|
||||||
|
$ mkdir -p data
|
||||||
|
|
||||||
|
$ wget -P data https://huggingface.co/TheBloke/TinyLlama-1.1B-Chat-v1.0-GGUF/resolve/main/tinyllama-1.1b-chat-v1.0.Q2_K.gguf
|
||||||
|
|
||||||
|
$ docker compose up -d
|
||||||
|
|
||||||
|
$ curl --request POST \
|
||||||
|
--url http://localhost:8080/completion \
|
||||||
|
--header "Content-Type: application/json" \
|
||||||
|
--data '{"prompt": "Building a website can be done in 10 simple steps:","n_predict": 128}'
|
||||||
|
```
|
||||||
|
|
||||||
|
[1]: https://github.com/ggerganov/llama.cpp
|
||||||
9
llama.cpp/docker-compose.yml
Normal file
9
llama.cpp/docker-compose.yml
Normal file
@@ -0,0 +1,9 @@
|
|||||||
|
services:
|
||||||
|
llama.cpp:
|
||||||
|
image: ghcr.io/ggerganov/llama.cpp:server
|
||||||
|
command: -m /models/tinyllama-1.1b-chat-v1.0.Q2_K.gguf -c 512 --host 0.0.0.0 --port 8080
|
||||||
|
ports:
|
||||||
|
- "8080:8080"
|
||||||
|
volumes:
|
||||||
|
- ./data:/models
|
||||||
|
restart: unless-stopped
|
||||||
Reference in New Issue
Block a user