|
|
0c1989c678
|
ai-llm: serialize rollout for RWO pvc
|
2026-01-01 14:48:54 -03:00 |
|
|
|
5093f77c0a
|
monitoring: per-panel namespace share filters
|
2026-01-01 14:44:33 -03:00 |
|
|
|
6a76fc0fa3
|
gpu: enable time-slicing and refresh dashboards
|
2026-01-01 14:16:08 -03:00 |
|
|
|
c6bae35bc6
|
chore(ai-llm): annotate pod with model and gpu
|
2025-12-21 00:47:57 -03:00 |
|
|
|
9162f5789f
|
ai-llm: GPU qwen2.5-coder on titan-24; add chat.ai host
|
2025-12-20 15:19:03 -03:00 |
|
|
|
39a914effd
|
ai-llm: use phi3 mini model
|
2025-12-20 14:24:52 -03:00 |
|
|
|
16ab7a963d
|
ai: allow ollama to share titan-24 gpu
|
2025-12-20 14:16:22 -03:00 |
|
|
|
c8adca5a5b
|
ai: add ollama service and wire chat backend
|
2025-12-20 14:10:34 -03:00 |
|