What's New in Mellea 0.4.0 + Granite Libraries Release

What's New in Mellea 0.4.0 + Granite Libraries Release

Hugging Face Blog 3 min read

About this article

A Blog post by IBM Granite on Hugging Face

Back to Articles What's New in Mellea 0.4.0 + Granite Libraries Release Enterprise Article Published March 20, 2026 Upvote 3 Abraham Daniels abedaniels Follow ibm-granite Overview We have released Mellea 0.4.0 alongside three Granite Libraries: granitelib-rag-r1.0,granitelib-core-r1.0,granitelib-guardian-r1.0. Together, these releases make it easier to build structured, verifiable, and safety-aware AI workflows on top of IBM Granite models. Mellea is an open-source Python library for writing generative programs -- replacing probabilistic prompt behavior with structured, maintainable AI workflows. Unlike general-purpose orchestration frameworks, Mellea is designed to make LLM-based programs maintainable and predictable through constrained decoding, structured repair loops, and composable pipelines (New to Mellea? Start with our introductory blog and meet the team) Mellea 0.4.0 Mellea 0.4.0 is the latest release of an open-source research project initiated and developed by IBM Research. Building on 0.3.0 foundational libraries and workflow primitives, 0.4.0 expands the library's integration surface and introduces new architectural patterns for structuring generative workflows. What’s included: Native integration with the Granite Libraries, offering a standardized API that relies on constrained decoding to guarantee schema correctness. Instruct-validate-repair pattern via rejection sampling strategies Observability hooks for event-driven callbacks to monitor and track workflo...

Originally published on March 20, 2026. Curated by AI News.

Related Articles

[2603.25112] Do LLMs Know What They Know? Measuring Metacognitive Efficiency with Signal Detection Theory
Llms

[2603.25112] Do LLMs Know What They Know? Measuring Metacognitive Efficiency with Signal Detection Theory

Abstract page for arXiv paper 2603.25112: Do LLMs Know What They Know? Measuring Metacognitive Efficiency with Signal Detection Theory

arXiv - AI · 4 min ·
[2603.24772] Evaluating Fine-Tuned LLM Model For Medical Transcription With Small Low-Resource Languages Validated Dataset
Llms

[2603.24772] Evaluating Fine-Tuned LLM Model For Medical Transcription With Small Low-Resource Languages Validated Dataset

Abstract page for arXiv paper 2603.24772: Evaluating Fine-Tuned LLM Model For Medical Transcription With Small Low-Resource Languages Val...

arXiv - Machine Learning · 4 min ·
[2603.25325] How Pruning Reshapes Features: Sparse Autoencoder Analysis of Weight-Pruned Language Models
Llms

[2603.25325] How Pruning Reshapes Features: Sparse Autoencoder Analysis of Weight-Pruned Language Models

Abstract page for arXiv paper 2603.25325: How Pruning Reshapes Features: Sparse Autoencoder Analysis of Weight-Pruned Language Models

arXiv - AI · 4 min ·
Llms

[D] Why evaluating only final outputs is misleading for local LLM agents

Been running local agents with Ollama + LangChain lately and noticed something kind of uncomfortable — you can get a completely correct f...

Reddit - Machine Learning · 1 min ·
More in Open Source Ai: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime