|
|
6ac5a0ac46
|
chore(ai-llm): annotate pod with model and gpu
|
2025-12-21 00:47:57 -03:00 |
|
|
|
fb6e71a62a
|
ai-llm: GPU qwen2.5-coder on titan-24; add chat.ai host
|
2025-12-20 15:19:03 -03:00 |
|
|
|
497ac90858
|
ai-llm: use phi3 mini model
|
2025-12-20 14:24:52 -03:00 |
|
|
|
b50977c5a0
|
ai: allow ollama to share titan-24 gpu
|
2025-12-20 14:16:22 -03:00 |
|
|
|
95ebdce813
|
ai: add ollama service and wire chat backend
|
2025-12-20 14:10:34 -03:00 |
|