mirror of
https://github.com/vimagick/dockerfiles.git
synced 2024-12-23 01:39:27 +02:00
…
|
||
---|---|---|
.. | ||
docker-compose.yml | ||
README.md |
ollama
Ollama is an open-source project that serves as a powerful and user-friendly platform for running LLMs on your local machine.
$ docker compose up -d
$ docker compose exec ollama bash
>>> ollama pull tinyllama
pulling manifest
pulling 2af3b81862c6... 100% ▕████████████████▏ 637 MB
pulling af0ddbdaaa26... 100% ▕████████████████▏ 70 B
pulling c8472cd9daed... 100% ▕████████████████▏ 31 B
pulling fa956ab37b8c... 100% ▕████████████████▏ 98 B
pulling 6331358be52a... 100% ▕████████████████▏ 483 B
verifying sha256 digest
writing manifest
removing any unused layers
success
>>> exit
$ curl http://localhost:11434/api/show -d '{"model": "tinyllama"}'
$ curl -s http://localhost:11434/api/generate -d '{
"model": "tinyllama",
"prompt": "What is 1+1?",
"stream": false
}' | jq -r .response
$ curl -s http://localhost:11434/api/chat -d '{
"model": "tinyllama",
"messages": [
{
"role": "user",
"content": "why is the sky blue?"
}
],
"stream": false
}' | jq -r .message.content
Read API Doc: https://github.com/ollama/ollama/blob/main/docs/api.md