Smart inference management across all your compute
GridLLM is an open-source distributed AI inference platform that turns any computer into a smart inference network. Connect Ollama instances across laptops, servers, and cloud resources for automatic load balancing and scaling.