7.0 KiB
7.0 KiB
Repository Guidelines
Project Structure & Module Organization
infrastructure/: cluster-scoped building blocks (core, flux-system, traefik, longhorn). Add new platform features by mirroring this layout.services/: workload manifests per app (services/gitea/, etc.) withkustomization.yamlplus one file per kind; keep diffs small and focused.dockerfiles/hosts bespoke images, whilescripts/stores operational Fish/Bash helpers—extend these directories instead of relying on ad-hoc commands.
Build, Test, and Development Commands
kustomize build services/<app>(orkubectl kustomize ...) renders manifests exactly as Flux will.kubectl apply --server-side --dry-run=client -k services/<app>checks schema compatibility without touching the cluster.flux reconcile kustomization <name> --namespace flux-system --with-sourcepulls the latest Git state after merges or hotfixes.fish scripts/flux_hammer.fish --helpexplains the recovery tool; read it before running against production workloads.
Coding Style & Naming Conventions
- YAML uses two-space indents; retain the leading path comment (e.g.
# services/gitea/deployment.yaml) to speed code review. - Keep resource names lowercase kebab-case, align labels/selectors, and mirror namespaces with directory names.
- List resources in
kustomization.yamlfrom namespace/config, through storage, then workloads and networking for predictable diffs. - Scripts start with
#!/usr/bin/env fishor bash, stay executable, and follow snake_case names such asflux_hammer.fish.
Testing Guidelines
- Run
kustomize buildand the dry-run apply for every service you touch; capture failures before opening a PR. flux diff kustomization <name> --path services/<app>previews reconciliations—link notable output when behavior shifts.- Docker edits:
docker build -f dockerfiles/Dockerfile.monerod .(swap the file you changed) to verify image builds.
Commit & Pull Request Guidelines
- Keep commit subjects short, present-tense, and optionally scoped (
gpu(titan-24): add RuntimeClass); squash fixups before review. - Describe linked issues, affected services, and required operator steps (e.g.
flux reconcile kustomization services-gitea) in the PR body. - Focus each PR on one kustomization or service and update
infrastructure/flux-systemwhen Flux must track new folders. - Record the validation you ran (dry-runs, diffs, builds) and add screenshots only when ingress or UI behavior changes.
Security & Configuration Tips
- Never commit credentials; use Vault workflows (
services/vault/) or SOPS-encrypted manifests wired throughinfrastructure/flux-system. - Node selectors and tolerations gate workloads to hardware like
hardware: rpi4; confirm labels before scaling or renaming nodes. - Pin external images by digest or rely on Flux image automation to follow approved tags and avoid drift.
Dashboard roadmap / context (2025-12-02)
- Atlas dashboards are generated via
scripts/dashboards_render_atlas.py --build, which writes JSON underservices/monitoring/dashboards/and ConfigMaps underservices/monitoring/. Keep the Grafana manifests in sync by regenerating after edits. - Atlas Overview panels are paired with internal dashboards (pods, nodes, storage, network, GPU). A new
atlas-gpuinternal dashboard holds the detailed GPU metrics that feed the overview share pie. - Old Grafana folders (
Atlas Storage,Atlas SRE,Atlas Public,Atlas Nodes) should be removed in Grafana UI when convenient; onlyAtlas OverviewandAtlas Internalshould remain provisioned. - Future work: add a separate generator (e.g.,
dashboards_render_oceanus.py) for SUI/oceanus validation dashboards, mirroring the atlas pattern of internal dashboards feeding a public overview.
Monitoring state (2025-12-03)
- dcgm-exporter DaemonSet pulls
registry.bstein.dev/monitoring/dcgm-exporter:4.4.2-4.7.0-ubuntu22.04with nvidia runtime/imagePullSecret; titan-24 exports metrics, titan-22 remains NotReady. - Atlas Overview is the Grafana home (1h range, 1m refresh), Overview folder UID
overview, internal folderatlas-internal(oceanus-internal stub). - Panels standardized via generator; hottest row compressed, worker/control rows taller, root disk row taller and top12 bar gauge with labels. GPU share pie uses 1h avg_over_time to persist idle activity.
- Internal dashboards are provisioned without Viewer role; if anonymous still sees them, restart Grafana and tighten auth if needed.
- GPU share panel updated (feature/sso) to use
max_over_time(…[$__range]), so longer ranges (e.g., 12h) keep recent activity visible. Flux trackingfeature/sso.
Upcoming priorities (SSO/storage/mail)
- Establish SSO (Keycloak or similar) and federate Grafana, Gitea, Zot, Nextcloud, Pegasus/Jellyfin; keep Vaultwarden separate until safe.
- Add Nextcloud (limit to rpi5 workers) with office suite; integrate with SSO; plan storage class and ingress.
- Plan mail: mostly self-hosted, relay through trusted provider for outbound; integrate with services (Nextcloud, Vaultwarden, etc.) for notifications and account flows.
SSO plan sketch (2025-12-03)
- IdP: use Keycloak (preferred) in a new
ssonamespace, Bitnami or codecentric chart with Postgres backing store (single PVC), ingresssso.bstein.dev, admin user bound to brad@bstein.dev; stick with local DB initially (no external IdP). - Auth flow goals: Grafana (OIDC), Gitea (OAuth2/Keycloak), Zot (via Traefik forward-auth/oauth2-proxy), Jellyfin/Pegasus via Jellyfin OAuth/OpenID plugin (map existing usernames; run migration to pre-create users in Keycloak with same usernames/emails and temporary passwords), Pegasus keeps using Jellyfin tokens.
- Steps to implement:
- Add service folder
services/keycloak/(namespace, PVC, HelmRelease, ingress, secret for admin creds). Verify with kustomize + Flux reconcile. - Seed realm
atlaswith users (import CSV/realm). Create client for Grafana (public/implicit), Gitea (confidential), and a “jellyfin” client for the OAuth plugin; set email for brad@bstein.dev as admin. - Reconfigure Grafana to OIDC (disable anonymous to internal folders, leave Overview public via folder permissions). Reconfigure Gitea to OIDC (app.ini).
- Add Traefik forward-auth (oauth2-proxy) in front of Zot and any other services needing headers-based auth.
- Deploy Jellyfin OpenID plugin; map Keycloak users to existing Jellyfin usernames; communicate password reset path.
- Add service folder
- Migration caution: do not delete existing local creds until SSO validated; keep Pegasus working via Jellyfin tokens during transition.
Postgres centralization (2025-12-03)
- Prefer a shared in-cluster Postgres deployment with per-service databases to reduce resource sprawl on Pi nodes. Use it for services that can easily point at an external DB.
- Candidates to migrate to shared Postgres: Keycloak (realm DB), Gitea (git DB), Nextcloud (app DB), possibly Grafana (if persistence needed beyond current provisioner), Jitsi prosody/JVB state (if external DB supported). Keep tightly-coupled or lightweight embedded DBs as-is when migration is painful or not supported.