IBM Releases as Open Source Its ‘Granite 4.0 Nano’ AI Models that Can Run on the Browser

IBL News | New York

IBM last week released, under the Apache 2.0 license, four new Granite 4.0 Nano models, designed to be highly accessible and well-suited for developers building applications on consumer hardware, without relying on cloud computing.

With these models, IBM is entering a crowded and rapidly evolving market of small language models (SLMs), competing with offerings like Qwen3, Google’s Gemma, LiquidAI’s LFM2, and Mistral’s dense models in the sub-2B parameter space.

With this release, IBM is positioning Granite as a platform for building the next generation of lightweight, trustworthy AI systems.

The 350M variants can run comfortably on a modern laptop CPU with 8–16GB of RAM, while the 1.5B models typically require a GPU with at least 6–8GB of VRAM for smooth performance.

This is a fraction of the size of their server-bound counterparts from companies like OpenAI, Anthropic, and Google.

This Granite 4.0 Nano family includes four open-source models now available on Hugging Face:

  • Granite-4.0-H-1B (~1.5B parameters) – Hybrid-SSM architecture
  • Granite-4.0-H-350M (~350M parameters) – Hybrid-SSM architecture
  • Granite-4.0-1B – Transformer-based variant, parameter count closer to 2B
  • Granite-4.0-350M – Transformer-based variant

Overall, the Granite-4.0-1B achieved a leading average benchmark score of 68.3% across general knowledge, math, code, and safety domains.

For developers and researchers seeking performance without overhead, the Nano release means they don’t need 70 billion parameters to build something powerful.