-
Notifications
You must be signed in to change notification settings - Fork 0
Description
Context
Setting up ShellForge as a governed Goose swarm on the jared box (WSL2, no local GPU), using a RunPod RTX 4090 for Ollama inference.
What we found during setup
-
shellforge setupdetects Linux + no GPU → skips Ollama + Goose entirely. It should offer remote Ollama configuration (setOLLAMA_HOSTto a remote endpoint) instead of just saying "use CLI drivers". The headless server use case with remote GPU is a valid deployment. -
shellforge setupStep 5 skips Goose onisServer=true— but Goose works fine headlessly with a remote Ollama. TheisServercheck gates out the primary local agent driver. -
shellforge run goosecallsgoose run --no-session -t <prompt>— need to verify this still works with Goose v1.28.0 (the-tflag for text prompt). -
agents.yamlmax_parallel auto-detect — on a CPU-only box with remote inference, RAM-based auto-detection is wrong. Should detect remote Ollama and set concurrency based on GPU VRAM instead. -
Version drift — installed binary is v0.6.1 but
main.gohardcodesvar version = "0.4.8". Release build must be overriding, but the source is stale.
Proposed fixes
-
shellforge setup: add "Remote Ollama" option when no local GPU — prompt forOLLAMA_HOST -
shellforge setup: offer Goose install on headless servers (it works headlessly) -
shellforge serve: support remote Ollama — skip localmustOllama()health check ifOLLAMA_HOSTis non-localhost -
agents.yaml: addinference: remoteoption that disables RAM-based concurrency - Test
goose run --no-session -twith v1.28.0 — flag may have changed - Sync
versionvar in main.go with release tags
Environment
- Box: jared (WSL2 Ubuntu, 8-core EPYC, 30GB RAM, no GPU)
- ShellForge: v0.6.1 (brew tap)
- Goose: v1.28.0 (Block AI agent,
~/.local/bin/goose) - Ollama: v0.18.3 CLI (server to run on RunPod RTX 4090)
- Target model: qwen3:8b