Ollama Cloud nabs Google Gemma 4 support
Ollama's hosted inference platform now supports Google's Gemma 4 31B model, enabling developers to run flagship-tier AI through a simple CLI command. The integration eliminates local hardware constraints while preserving the familiar Ollama workflow and API compatibility.
Ollama Cloud is effectively becoming the "iCloud for LLMs," seamlessly abstracting away the compute barrier for massive open models. Gemma 4 31B ranks #3 globally on the Arena AI leaderboard, offering frontier-level performance in a dense architecture. The hybrid ":cloud" suffix allows developers to toggle between local and hosted execution without changing a single line of code. A 256K token context window and native multimodality make it a potent engine for complex agentic workflows. The shift to an Apache 2.0 license for Gemma 4 models drastically lowers the legal barrier for enterprise adoption. Subscription-based GPU access provides a cost-effective alternative to high-end local hardware upgrades for developers.
DISCOVERED
5d ago
2026-04-06
PUBLISHED
5d ago
2026-04-06
RELEVANCE
AUTHOR
DIY Smart Code