ARMONK, N.Y., Feb. 26, 2025 /PRNewswire/ — IBM (NYSE: IBM) today debuted the next generation of its Granite large language model (LLM) family, Granite 3.2, in a continued effort to deliver small, efficient, practical enterprise AI for real-world impact.
All Granite 3.2 models are available under the permissive Apache 2.0 license on Hugging Face. Select models are available today on IBM watsonx.ai, Ollama, Replicate, and LM Studio, and expected soon in RHEL AI 1.5 – bringing advanced capabilities to businesses and the open-source community. Highlights include:
IBM’s strategy to deliver smaller, specialized AI models for enterprises continues to demonstrate efficacy in testing, with the Granite 3.1 8B model recently yielding high marks on accuracy in the Salesforce LLM Benchmark for CRM.
The Granite model family is supported by a robust ecosystem of partners, including leading software companies embedding the LLMs into their technologies.
“At CrushBank, we’ve seen first-hand how IBM’s open, efficient AI models deliver real value for enterprise AI – offering the right balance of performance, cost-effectiveness, and scalability,” said David Tan, CTO, CrushBank. “Granite 3.2 takes it further with new reasoning capabilities, and we’re excited to explore them in building new agentic solutions.”
Granite 3.2 is an important step in the evolution of IBM’s portfolio and strategy to deliver small, practical AI for enterprises. While chain of thought approaches for reasoning are powerful, they require substantial compute power that is not necessary for every task. That is why IBM has introduced the ability to turn chain of thought on or off programmatically. For simpler tasks, the model can operate without reasoning to reduce unnecessary compute overhead. Additionally, other reasoning techniques like inference scaling have shown that the Granite 3.2 8B model can match or exceed the performance of much larger models on standard math reasoning benchmarks. Evolving methods like inference scaling remains a key area of focus for IBM’s research teams.4
Alongside Granite 3.2 instruct, vision, and guardrail models, IBM is releasing the next generation of its TinyTimeMixers (TTM) models (sub 10M parameters), with capabilities for longer-term forecasting up to two years into the future. These make for powerful tools in long-term trend analysis, including finance and economics trends, supply chain demand forecasting and seasonal inventory planning in retail.
“The next era of AI is about efficiency, integration, and real-world impact – where enterprises can achieve powerful outcomes without excessive spend on compute,” said Sriram Raghavan, VP, IBM AI Research. “IBM’s latest Granite developments focus on open solutions demonstrate another step forward in making AI more accessible, cost-effective, and valuable for modern enterprises.”
To learn more about Granite 3.2, read this technical article.
New collaboration will establish a quantum-control–enabled center at the IQMP to accelerate and scale fault-tolerant…
With a seed round led by 10D Ventures, SENAI emerges from stealth to help government…
Shared industrial AI architecture combines Virtual Twins and AI infrastructure deployable at scale. Science-validated world…
Microchip’s MCPF1525 power module with PMBus™ delivers 25A DC-DC power, stackable up to 200A The…
New AI-powered platform – the first with full ERP integration – includes an AI agent…
The study validated the CORA system’s ability to rule out coronary artery disease with a…