Granite 4.1 LLMs: How They’re Built

Granite 4.1 LLMs: How They’re Built

Hugging Face Blog 14 min read

About this article

A Blog post by IBM Granite on Hugging Face

Back to Articles Granite 4.1 LLMs: How They’re Built Enterprise Article Published April 29, 2026 Upvote 7 +1 Yousaf Shah yousafshah Follow ibm-granite An in-depth technical walkthrough of data engineering, pre-training, supervised fine-tuning, and reinforcement learning behind the Granite 4.1 LLMs. Authors: Granite Team, IBM TL;DR — Granite 4.1 is a family of dense, decoder‑only LLMs (3B, 8B, and 30B) trained on ~15T tokens using a multi‑stage pre‑training pipeline, including long‑context extension of up to 512K tokens. The models are further refined with supervised fine‑tuning on ~4.1M high‑quality curated samples and reinforcement learning via on‑policy GRPO with DAPO loss (Yu et al., 2025). Notably, the 8B instruct model matches or surpasses the previous Granite 4.0‑H‑Small (32B‑A9B MoE) despite using a simpler dense architecture with fewer parameters. All Granite 4.1 models are released under the Apache 2.0 license. Links: Granite 4.1 HF Collection GitHub Repository Granite Docs Overview Building high‑quality small language models goes beyond simply scaling compute—it requires rigorous data curation throughout training. For Granite 4.1, we prioritized data quality over quantity, progressively refining the data mixture across five pre‑training stages. We further curated supervised fine‑tuning data using an LLM‑as‑Judge framework and applied a multi‑stage reinforcement learning pipeline to systematically strengthen performance in math, coding, instruction following, and ...

Originally published on April 29, 2026. Curated by AI News.

Related Articles

Llms

How strongly do you believe LLM judges on the for the ML papers?? [D]

I'm curious about your thoughts on these, as far as I've seen most of the comments are nitpicking about "missing ablations" while some co...

Reddit - Machine Learning · 1 min ·
Llms

Google just dropped "Deep Research Max" — We are officially entering the era of Autonomous Agents. RIP to the "Junior Analyst"?

The shift from "Chatbot" to "Agent" just hit warp speed. Google’s release of Deep Research Max isn't just another incremental update; it’...

Reddit - Artificial Intelligence · 1 min ·
Llms

Built a prompt injection proxy that beats OpenAI Moderation and LlamaGuard — try it in 30 seconds without leaving this

Built Arc Gate — sits in front of any OpenAI-compatible endpoint and blocks prompt injection before it reaches your model. Just change yo...

Reddit - Artificial Intelligence · 1 min ·
More Gemini features are coming to Google TV | TechCrunch
Llms

More Gemini features are coming to Google TV | TechCrunch

Google TV just got more Gemini features, including the ability to transform photos and videos with tools Nano Banana and Veo.

TechCrunch - AI · 4 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime