[2602.17826] Ontology-Guided Neuro-Symbolic Inference: Grounding Language Models with Mathematical Domain Knowledge

[2602.17826] Ontology-Guided Neuro-Symbolic Inference: Grounding Language Models with Mathematical Domain Knowledge

arXiv - Machine Learning 3 min read Article

Summary

This article explores the integration of formal domain ontologies into language models to enhance their reliability in mathematical reasoning, addressing limitations like hallucination and brittleness.

Why It Matters

As language models become increasingly used in critical fields, ensuring their reliability is paramount. This research highlights the potential of neuro-symbolic approaches to improve model performance through structured knowledge, particularly in high-stakes environments where accuracy is crucial.

Key Takeaways

  • Formal domain ontologies can enhance language model reliability.
  • Neuro-symbolic pipelines can improve performance in specialized fields.
  • Retrieval quality is critical; irrelevant context can degrade model output.
  • The study uses mathematics as a proof of concept for the proposed method.
  • Challenges remain in balancing context relevance and model performance.

Computer Science > Artificial Intelligence arXiv:2602.17826 (cs) [Submitted on 19 Feb 2026] Title:Ontology-Guided Neuro-Symbolic Inference: Grounding Language Models with Mathematical Domain Knowledge Authors:Marcelo Labre View a PDF of the paper titled Ontology-Guided Neuro-Symbolic Inference: Grounding Language Models with Mathematical Domain Knowledge, by Marcelo Labre View PDF HTML (experimental) Abstract:Language models exhibit fundamental limitations -- hallucination, brittleness, and lack of formal grounding -- that are particularly problematic in high-stakes specialist fields requiring verifiable reasoning. I investigate whether formal domain ontologies can enhance language model reliability through retrieval-augmented generation. Using mathematics as proof of concept, I implement a neuro-symbolic pipeline leveraging the OpenMath ontology with hybrid retrieval and cross-encoder reranking to inject relevant definitions into model prompts. Evaluation on the MATH benchmark with three open-source models reveals that ontology-guided context improves performance when retrieval quality is high, but irrelevant context actively degrades it -- highlighting both the promise and challenges of neuro-symbolic approaches. Comments: Subjects: Artificial Intelligence (cs.AI); Machine Learning (cs.LG); Symbolic Computation (cs.SC) Cite as: arXiv:2602.17826 [cs.AI]   (or arXiv:2602.17826v1 [cs.AI] for this version)   https://doi.org/10.48550/arXiv.2602.17826 Focus to learn more arXiv...

Related Articles

Llms

What if Claude purposefully made its own code leakable so that it would get leaked

What if Claude leaked itself by socially and architecturally engineering itself to be leaked by a dumb human submitted by /u/smurfcsgoawp...

Reddit - Artificial Intelligence · 1 min ·
Llms

Observer-Embedded Reality

Observer-Embedded Reality Consciousness, Complexity, Meaning, and the Limits of Human Knowledge A Conceptual Philosophy-of-Science Paper ...

Reddit - Artificial Intelligence · 1 min ·
Llms

I think we’re about to have a new kind of “SEO”… and nobody is talking about it.

More people are asking ChatGPT things like: “what’s the best CRM?” “is this tool worth it?” “alternatives to X” And they just… trust the ...

Reddit - Artificial Intelligence · 1 min ·
Llms

Why would Claude give me the same response over and over and give others different replies?

I asked Claude to "generate me a random word" so I could do some word play. Then I asked it again in a new prompt window on desktop after...

Reddit - Artificial Intelligence · 1 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime