Ollama on Mac mini M4 (24GB Unified Memory)
Mac mini M4 with 24GB can run strong local workflows, but its memory behavior differs from discrete GPUs. Model weights, KV cache, and macOS all pull from one unified pool.
The result: context sizing is even more important, and background app usage can change model stability.
This is why two seemingly identical setups can feel different: one machine is running a clean native Ollama session, while the other is sharing memory with browsers, design apps, and container overhead.