1
0
mirror of https://github.com/vimagick/dockerfiles.git synced 2025-01-06 03:54:01 +02:00
dockerfiles/ollama/README.md
2024-08-09 12:19:01 +08:00

47 lines
1.3 KiB
Markdown

ollama
======
[Ollama][1] is an open-source project that serves as a powerful and
user-friendly platform for running LLMs on your local machine.
```bash
$ docker compose up -d
$ docker compose exec ollama bash
>>> ollama pull tinyllama
pulling manifest
pulling 2af3b81862c6... 100% ▕████████████████▏ 637 MB
pulling af0ddbdaaa26... 100% ▕████████████████▏ 70 B
pulling c8472cd9daed... 100% ▕████████████████▏ 31 B
pulling fa956ab37b8c... 100% ▕████████████████▏ 98 B
pulling 6331358be52a... 100% ▕████████████████▏ 483 B
verifying sha256 digest
writing manifest
removing any unused layers
success
>>> exit
$ curl http://localhost:11434/api/show -d '{"model": "tinyllama"}'
$ curl -s http://localhost:11434/api/generate -d '{
"model": "tinyllama",
"prompt": "What is 1+1?",
"stream": false
}' | jq -r .response
$ curl -s http://localhost:11434/api/chat -d '{
"model": "tinyllama",
"messages": [
{
"role": "user",
"content": "why is the sky blue?"
}
],
"stream": false
}' | jq -r .message.content
```
Read API Doc: https://github.com/ollama/ollama/blob/main/docs/api.md
[1]: https://github.com/ollama/ollama