17 Commits

Author SHA1 Message Date
55ac2e5568 feat(llama-swap): fix contexts, add flash-attn/jinja, tune sampling params 2026-04-10 15:42:19 +02:00
2d852879b6 feat: add new models 2026-04-10 15:10:38 +02:00
c6e4901dee refactor(ollama): simplify model names in llama-swap configuration
Replace verbose Hugging Face model references with concise local aliases for
Qwen3.5 models to improve readability and maintainability.
2026-04-10 11:42:39 +02:00
8ab4213b62 feat(ollama): add persistence in Ollama container
Re-enable qwen3.5:9b and qwen3.5:9bctxSmall using fresh unsloth/Qwen3.5-9B-GGUF
quantization, which uses the correct rope.dimension_sections format (4 elements)
compatible with this llama.cpp build. Both models include the mmproj for
multimodal support. The old Ollama-extracted GGUF (mrope_sections, 3 elements)
has been removed.
2026-04-10 10:57:34 +02:00
ebc71492c3 fix(ollama): restrict to RX 9070 XT, restore mmproj
- Set HIP_VISIBLE_DEVICES=0 to use only the discrete GPU (gfx1201).
  llama.cpp was trying to split layers across the iGPU (gfx1036) which
  caused segfaults when loading the multimodal projector.
- Restore --mmproj for both HF models (multimodal works correctly with
  single GPU).
- Keep qwen3.5:9b disabled (Ollama-extracted GGUF uses old mrope_sections
  key format incompatible with this llama.cpp build).
2026-04-10 00:09:12 +02:00
3034f987d7 feat(ollama): migrate from Ollama to llama.cpp + llama-swap
Replace the Ollama service with a custom ROCm image combining
ghcr.io/ggml-org/llama.cpp:server-rocm and llama-swap v199.

Main motivations:
- Unblock qwen35 HF GGUFs (qwen35 architecture not supported in
  Ollama 0.20.4 for HF-imported models)
- Stay current with llama.cpp upstream without waiting for Ollama releases

Changes:
- ollama/Dockerfile: build llama-swap on top of llama.cpp:server-rocm
- ollama/llama-swap.yaml: define 4 models with full sampler config,
  GPU offload, and mmproj for the two multimodal HF fine-tunes
- ollama/docker-compose.yml: replace Ollama image with local build;
  fix broken volume mount (was /ubuntu/.ollama, now explicit /models)
- ollama/Caddyfile: update upstream port 11434→8080 (llama-swap default)
- ai/docker-compose.yml: switch Open WebUI from OLLAMA_BASE_URL to
  OPENAI_API_BASE_URL pointing at llama-swap /v1 endpoint
2026-04-09 23:14:43 +02:00
299d712400 fix: add persistence in Ollama container 2026-04-09 16:49:45 +02:00
d856a8704b feat(ai): add Open WebUI with Tailscale and Caddy reverse proxy
Set up a complete AI services stack with Open WebUI as the main interface,
secured behind Caddy reverse proxy with automatic HTTPS via Namecheap DNS
challenge. Tailscale integration provides secure remote access.

- Configure Open WebUI to connect to Ollama backend
- Set up Caddy with Namecheap DNS plugin for wildcard SSL certificates
- Add Tailscale for secure networking with health checks
- Configure reverse proxy to forward requests to Open WebUI
- Enable proper header forwarding for client IP preservation
- Set up persistent volumes for data, config, and SSL certificates
- Configure JSON logging with rotation for all services
2026-03-18 22:16:58 +01:00
7fdd996f29 feat: add first ollama stack 2026-03-11 14:27:02 +01:00
b1641c46b1 feat(jellyfin): add sport directory mount
Add /sport volume mount for sports content from backup NAS, allowing Jellyfin to
access and serve sports-related media files
2026-03-08 15:42:23 +01:00
4548e56eef feat: add Jellyfin service
- tailscale + local deployment
- automatic TLS
2026-01-23 19:26:00 +01:00
64cc8b181d fix(navidrome): pin app image to version 0.59.0
Replace latest tag with specific version 0.59.0 for better
reproducibility and stability in deployments.
2025-12-23 12:20:25 +01:00
996f7a184d refactor(navidrome): replace custom Caddy build with pre-built image
Remove custom Dockerfile.caddy and use caddybuilds/caddy-namecheap:2-alpine
image directly in docker-compose.yml to simplify deployment
2025-08-23 15:18:56 +02:00
d7efa69de6 feat(navidrome): enhance caddy container configuration
Add persistent volume mounts for caddy config, data, and share directories, and
improve healthcheck settings with longer intervals, timeout, and start period
for better reliability
2025-08-23 15:01:23 +02:00
d5f9890dc8 feat(navidrome): migrate to DNS challenge with automated SSL
- Add caddy.env to gitignore for sensitive DNS credentials
- Replace manual SSL certificates with Let's Encrypt DNS challenge using
Namecheap
- Build custom Caddy image with namecheap DNS plugin
- Configure wildcard SSL for *.lan.poldebra.me domain
- Update docker-compose to use custom Caddy build and environment file

Note: we had to downgrade to Caddy 2.9, see
https://github.com/caddy-dns/namecheap/issues/14 for more information
2025-08-23 14:12:38 +02:00
9ec5361e40 feat: add navidrome with manual certificates renewal 2025-08-23 11:40:10 +02:00
49cd259955 Initial commit 2025-08-23 11:39:24 +02:00