Ollama v4 Brings Distributed Inference to Everyone#
The Ollama project has released v4.0, a landmark update that introduces native distributed inference support. Users can now split large language model workloads across multiple machines on the same network, making it feasible to run 70B+ parameter models on clusters of consumer hardware.
