ARMONK, N.Y., Feb. 26, 2025 /PRNewswire/ — IBM (NYSE: IBM) today debuted the next generation of its Granite large language model (LLM) family, Granite 3.2, in a continued effort to deliver small, efficient, practical enterprise AI for real-world impact.
All Granite 3.2 models are available under the permissive Apache 2.0 license on Hugging Face. Select models are available today on IBM watsonx.ai, Ollama, Replicate, and LM Studio, and expected soon in RHEL AI 1.5 – bringing advanced capabilities to businesses and the open-source community. Highlights include:
IBM’s strategy to deliver smaller, specialized AI models for enterprises continues to demonstrate efficacy in testing, with the Granite 3.1 8B model recently yielding high marks on accuracy in the Salesforce LLM Benchmark for CRM.
The Granite model family is supported by a robust ecosystem of partners, including leading software companies embedding the LLMs into their technologies.
“At CrushBank, we’ve seen first-hand how IBM’s open, efficient AI models deliver real value for enterprise AI – offering the right balance of performance, cost-effectiveness, and scalability,” said David Tan, CTO, CrushBank. “Granite 3.2 takes it further with new reasoning capabilities, and we’re excited to explore them in building new agentic solutions.”
Granite 3.2 is an important step in the evolution of IBM’s portfolio and strategy to deliver small, practical AI for enterprises. While chain of thought approaches for reasoning are powerful, they require substantial compute power that is not necessary for every task. That is why IBM has introduced the ability to turn chain of thought on or off programmatically. For simpler tasks, the model can operate without reasoning to reduce unnecessary compute overhead. Additionally, other reasoning techniques like inference scaling have shown that the Granite 3.2 8B model can match or exceed the performance of much larger models on standard math reasoning benchmarks. Evolving methods like inference scaling remains a key area of focus for IBM’s research teams.4
Alongside Granite 3.2 instruct, vision, and guardrail models, IBM is releasing the next generation of its TinyTimeMixers (TTM) models (sub 10M parameters), with capabilities for longer-term forecasting up to two years into the future. These make for powerful tools in long-term trend analysis, including finance and economics trends, supply chain demand forecasting and seasonal inventory planning in retail.
“The next era of AI is about efficiency, integration, and real-world impact – where enterprises can achieve powerful outcomes without excessive spend on compute,” said Sriram Raghavan, VP, IBM AI Research. “IBM’s latest Granite developments focus on open solutions demonstrate another step forward in making AI more accessible, cost-effective, and valuable for modern enterprises.”
To learn more about Granite 3.2, read this technical article.
The collaboration between Tower Semiconductor and Scintil Photonics offers a clear view into the future…
How virtual twins, accelerated computing and physics-based models are turning AI from smart software into…
For decades, cellular communication has relied on one almost unquestioned assumption: the base station is…
IBM (NYSE: IBM) today announced new results that its quantum computer can simulate real magnetic materials…
Arrow Electronics has announced the launch of a new evaluation platform, jointly designed in collaboration…
Siemens has unveiled Drivetrain Analyzer Onsite, a new on-premises analytics solution for industrial drive systems,…