Ollama Template
Image: ollama/ollama:latest
Min VRAM: 8 GB | Port: 11434
Run open-source LLMs locally with a simple API.
What’s Included
- Ollama runtime
- Easy model pulling (
ollama pull llama2) - REST API on port 11434
- GPU acceleration
Launch
curl -X POST https://api.pulserun.dev/v1/instances \
-H "Authorization: Bearer pr_live_xxxxxxxxxxxxx" \
-H "Content-Type: application/json" \
-d '{"gpu": "rtx_4090", "template": "ollama"}'Usage
# Pull a model
curl http://<instance-ip>:11434/api/pull -d '{"name": "llama2"}'
# Chat
curl http://<instance-ip>:11434/api/chat -d '{
"model": "llama2",
"messages": [{"role": "user", "content": "Hello!"}]
}'