IBL News | New York
IBM last week released, under the Apache 2.0 license, four new Granite 4.0 Nano models, designed to be highly accessible and well-suited for developers building applications on consumer hardware, without relying on cloud computing.
With these models, IBM is entering a crowded and rapidly evolving market of small language models (SLMs), competing with offerings like Qwen3, Google’s Gemma, LiquidAI’s LFM2, and Mistral’s dense models in the sub-2B parameter space.
With this release, IBM is positioning Granite as a platform for building the next generation of lightweight, trustworthy AI systems.
The 350M variants can run comfortably on a modern laptop CPU with 8–16GB of RAM, while the 1.5B models typically require a GPU with at least 6–8GB of VRAM for smooth performance.
This is a fraction of the size of their server-bound counterparts from companies like OpenAI, Anthropic, and Google.
This Granite 4.0 Nano family includes four open-source models now available on Hugging Face:
- Granite-4.0-H-1B (~1.5B parameters) – Hybrid-SSM architecture
- Granite-4.0-H-350M (~350M parameters) – Hybrid-SSM architecture
- Granite-4.0-1B – Transformer-based variant, parameter count closer to 2B
- Granite-4.0-350M – Transformer-based variant
Overall, the Granite-4.0-1B achieved a leading average benchmark score of 68.3% across general knowledge, math, code, and safety domains.
For developers and researchers seeking performance without overhead, the Nano release means they don’t need 70 billion parameters to build something powerful.

En Español












![OpenAI Released Apps that Work Inside ChatGPT and an SDK [Video]](https://iblnews.org/wp-content/uploads/2025/10/openaieventday-218x150.jpg)









