Key Takeaway
IBM has released the Granite 4.1 model family, its broadest Granite release to date, covering language, vision, speech, embedding and Guardian models for enterprise AI workflows. Within that release, the Granite Speech 4.1 2B models target multilingual ASR, speech translation and faster transcription under the Apache 2.0 license, while Granite 4.1 language models focus on instruction following, tool calling, predictable latency and lower operating costs.
IBM Launches Granite 4.1 – Key Points
The Story
IBM has released Granite 4.1, a broad collection of open enterprise AI models spanning small language models, speech, vision, embeddings and Guardian safety models. The release includes Granite Speech 4.1 2B, Granite Speech 4.1 2B-NAR and Granite Speech 4.1 2B-Plus, with different tradeoffs between transcription accuracy, throughput, latency and transcription richness. It also includes dense Granite 4.1 language models in 3B, 8B and 30B sizes, including an 8B instruct model that IBM says can match or outperform the earlier Granite 4.0 32B Mixture-of-Experts model. All Granite 4.1 models are released under the Apache 2.0 license and can be tried on watsonx, Hugging Face and other platforms.
The Facts
IBM calls Granite 4.1 its most expansive model release to date.
The collection covers language, vision, speech, embedding and Guardian models designed for enterprise-grade AI systems.
The release includes new Granite Speech 4.1 models.
Granite Speech 4.1 2B, Granite Speech 4.1 2B-NAR and Granite Speech 4.1 2B-Plus target speech recognition use cases with different tradeoffs between speed, latency and transcription detail.
Granite Speech 4.1 2B supports ASR and speech translation.
The standard model supports automatic speech recognition across English, French, German, Spanish, Portuguese and Japanese, and includes bidirectional automatic speech translation.
The NAR model is designed for higher-throughput ASR.
Granite Speech 4.1 2B-NAR focuses on automatic speech recognition, trades away some richer features and generates entire sequences at once rather than one token at a time.
IBM reports strong benchmark results for the standard speech model.
Granite Speech 4.1 2B achieves a 5.33% word-error rate, placing it among the top models on the OpenASR Leaderboard.
IBM links the NAR approach to better GPU utilization.
IBM Research says Granite Speech 4.1 2B-NAR’s structure results in considerably better GPU utilization and much higher throughput.
The Granite 4.1 language models come in 3B, 8B and 30B sizes.
They are dense, decoder-only models offered in base and instruct versions, with IBM reporting significant gains over similarly sized Granite 4.0 models.
IBM says Granite 4.1 8B can match or beat an older larger model.
IBM reports that Granite 4.1 8B instruct consistently matches or outperforms the Granite 4.0 32B Mixture-of-Experts model while using a simpler architecture that is more flexible for fine-tuning.
IBM released FP8 quantized versions.
These versions reduce disk footprint and GPU memory use by about 50%, with quantization applied to the weights and activations of linear operators inside the transformer blocks.
The language models were trained on about 15 trillion tokens.
IBM says training used a five-phase pre-training strategy, moving from broad pre-training toward higher-quality domain-specific data, with the final phase extending context length to 512K tokens.
Granite 4.1 targets non-reasoning enterprise workloads.
IBM positions the language models for instruction following, tool calling, stable token usage and predictable latency, rather than relying on long chains of thought.
Granite 4.1 also expands enterprise document, retrieval and guardrail tools.
Granite Vision 4.1 targets table, chart and key-value pair extraction, Granite Embedding Multilingual R2 scales retrieval support to more than 200 languages, and Granite Guardian 4.1 replaces Granite Guardian 3.3 8B with more risk definitions for evaluating model inputs and outputs.
Benchmarks / Evidence Check
IBM Research reports that Granite Speech 4.1 2B achieves a 5.33% word-error rate and places among the top models on the OpenASR Leaderboard.
For the broader Granite 4.1 family, IBM says the new language models improve over similarly sized Granite 4.0 models, and that Granite 4.1 8B instruct consistently matches or outperforms Granite 4.0 32B Mixture-of-Experts. IBM also says the models perform competitively with recent Gemma and Qwen dense decoder-only models on instruction following and tool calling when thinking is disabled.
Use Cases
Granite Speech 4.1 2B is the better fit for teams that need multilingual transcription, Japanese support and speech translation.
Granite Speech 4.1 2B-NAR is aimed at latency-sensitive ASR deployments where throughput matters more than richer transcription features.
Granite Speech 4.1 2B-Plus targets use cases where applications need speaker identity and precise timing, such as meeting transcription, call analysis or timestamped media workflows.
Granite 4.1 8B is relevant for users and developers who want an open instruct model that can be tested locally for instruction-following, tool-use and coding-style workflows.
Granite Vision 4.1 is built for enterprise document understanding, including tables, charts and key-value pair extraction from documents such as invoices.
Granite Guardian 4.1 is designed to monitor model inputs and outputs for safety, quality and correctness in AI pipelines.
Risks / Limitations
Granite Speech 4.1 2B-NAR has a narrower feature set than the standard speech model. It trades some transcription richness for substantially higher throughput.
IBM positions Granite 4.1 as a modular enterprise model family rather than a single all-purpose frontier model. Its value depends on matching the right model to the right workload.
For enterprise procurement, Mitch Ashley of The Futurum Group argues that open-weight providers also need operational wrappers such as ISO certification, indemnity and deployment predictability, not just model weights.
Numbers that Matter
- 2B parameters: Approximate size of the main Granite Speech 4.1 models.
- 5.33% WER: IBM-reported word-error rate for Granite Speech 4.1 2B.
- 3B, 8B and 30B: Granite 4.1 language model sizes.
- 15 trillion tokens: Approximate training scale for the Granite 4.1 language models.
- 512K tokens: Maximum context length IBM says the language models can reach.
- ~50%: Reported disk footprint and GPU memory reduction for FP8 quantized versions.
- 200+ languages: Retrieval support for Granite Embedding Multilingual R2.
- 97M parameters: Size of the smaller Granite Embedding Multilingual R2 model.
Why This Matters
IBM is positioning Granite 4.1 as a practical enterprise AI toolkit rather than a single model release. The speech models address transcription and translation, the language models target instruction following, tool calling and cost-efficient deployment, the vision models handle business documents, the embedding models support multilingual retrieval and Guardian adds moderation and risk signals directly into AI workflows.
This article was drafted with the assistance of generative AI. All facts and details were reviewed and confirmed by an editor prior to publication.
IBM 2026 X-Force Threat Index: North America hit 29% of cases in 2025; app-exploit attacks rose 44%, vulnerabilities drove 40% of incidents, extortion groups hit 109.
IBM and Anthropic form a strategic partnership to integrate Claude AI models into its enterprise software suite, starting with an IDE tested by 6,000 staff and reporting 45% productivity gains.
Explore the ‘Meta and IBM AI Alliance’ – a bold move to redefine AI development and open-source dynamics, aiming for a more inclusive, transparent future in AI technology.
Read a comprehensive monthly roundup of the latest AI news!






