# Known Issues ## BUG-6: Validator logging not configured, only stdout available **Observed:** 2026-03-03 The validator only logs to stdout. kubectl logs retains ~2 minutes of history at current log volume before the buffer fills. When diagnosing a replay stall, the startup logs (snapshot load, initial replay, error conditions) were gone. **Impact:** Cannot determine why the validator replay stage stalled — the startup logs that would show the root cause are not available. **Fix:** Configure the `--log` flag in the validator start script to write to a persistent volume, so logs survive container restarts and aren't limited to the kubectl buffer. ## BUG-7: Metrics endpoint unreachable from validator pod **Observed:** 2026-03-03 ``` WARN solana_metrics::metrics submit error: error sending request for url (http://localhost:8086/write?db=agave_metrics&u=admin&p=admin&precision=n) ``` The validator is configured with `SOLANA_METRICS_CONFIG` pointing to `http://172.20.0.1:8086` (the kind docker bridge gateway), but the logs show it trying `localhost:8086`. The InfluxDB container (`solana-monitoring-influxdb-1`) is running on the host, but the validator can't reach it. **Impact:** No metrics collection. Cannot use Grafana dashboards to diagnose performance issues or track sync progress over time. ## BUG-8: sysctl values not visible inside kind container **Observed:** 2026-03-03 ``` ERROR solana_core::system_monitor_service Failed to query value for net.core.rmem_max: no such sysctl WARN solana_core::system_monitor_service net.core.rmem_max: recommended=134217728, current=-1 too small ``` The host has correct sysctl values (`net.core.rmem_max = 134217728`), but `/proc/sys/net/core/` does not exist inside the kind node container. The validator reads `-1` and reports the buffer as too small. The network buffers themselves may still be effective (they're set on the host network namespace which the pod shares via `hostNetwork: true`), but this is unverified. If the buffers are not effective, it could limit shred ingestion throughput and contribute to slow repair. **Fix options:** - Set sysctls on the kind node container at creation time (`kind` supports `kubeadmConfigPatches` and sysctl configuration) - Verify empirically whether the host sysctls apply to hostNetwork pods by checking actual socket buffer sizes from inside the pod ## Validator replay stall (under investigation) **Observed:** 2026-03-03 The validator root has been stuck at slot 403,892,310 for 55+ minutes. The gap to the cluster tip is ~120,000 slots and growing. **Observed symptoms:** - Zero `Frozen` banks in log history — replay stage is not processing slots - All incoming slots show `bank_status: Unprocessed` - Repair only requests tip slots and two specific old slots (403,892,310, 403,909,228) — not the ~120k slot gap - Repair peer count is 3-12 per cycle (vs 1,000+ gossip peers) - Startup logs have rotated out (BUG-6), so initialization context is lost **Unknown:** - What snapshot the validator loaded at boot - Whether replay ever started or was blocked from the beginning - Whether the sysctl issue (BUG-8) is limiting repair throughput - Whether the missing metrics (BUG-7) would show what's happening internally