[2602.18497] PIPE-RDF: An LLM-Assisted Pipeline for Enterprise RDF Benchmarking

[2602.18497] PIPE-RDF: An LLM-Assisted Pipeline for Enterprise RDF Benchmarking

arXiv - AI 3 min read Article

Summary

PIPE-RDF presents a novel pipeline for generating schema-specific NL-SPARQL benchmarks, enhancing RDF knowledge graph querying for enterprises by ensuring high validity and operational readiness.

Why It Matters

As enterprises increasingly rely on RDF knowledge graphs for data management, the PIPE-RDF pipeline addresses the gap in existing benchmarks that do not cater to proprietary schemas. This innovation supports better data querying and operational efficiency, crucial for businesses leveraging AI and data science.

Key Takeaways

  • PIPE-RDF generates tailored NL-SPARQL benchmarks for enterprises.
  • The pipeline achieves high validity rates, ensuring reliable data querying.
  • Structured artifacts and operational metrics are provided for practical deployment.

Computer Science > Databases arXiv:2602.18497 (cs) [Submitted on 15 Feb 2026] Title:PIPE-RDF: An LLM-Assisted Pipeline for Enterprise RDF Benchmarking Authors:Suraj Ranganath View a PDF of the paper titled PIPE-RDF: An LLM-Assisted Pipeline for Enterprise RDF Benchmarking, by Suraj Ranganath View PDF HTML (experimental) Abstract:Enterprises rely on RDF knowledge graphs and SPARQL to expose operational data through natural language interfaces, yet public KGQA benchmarks do not reflect proprietary schemas, prefixes, or query distributions. We present PIPE-RDF, a three-phase pipeline that constructs schema-specific NL-SPARQL benchmarks using reverse querying, category-balanced template generation, retrieval-augmented prompting, deduplication, and execution-based validation with repair. We instantiate PIPE-RDF on a fixed-schema company-location slice (5,000 companies) derived from public RDF data and generate a balanced benchmark of 450 question-SPARQL pairs across nine categories. The pipeline achieves 100% parse and execution validity after repair, with pre-repair validity rates of 96.5%-100% across phases. We report entity diversity metrics, template coverage analysis, and cost breakdowns to support deployment planning. We release structured artifacts (CSV/JSONL, logs, figures) and operational metrics to support model evaluation and system planning in real-world settings. Code is available at this https URL. Comments: Subjects: Databases (cs.DB); Artificial Intelligence (cs...

Related Articles

Llms

[P] Remote sensing foundation models made easy to use.

This project enables the idea of tasking remote sensing models to acquire embeddings like we task satellites to acquire data! https://git...

Reddit - Machine Learning · 1 min ·
Llms

I stopped using Claude like a chatbot — 7 prompt shifts that reclaimed 10 hours of my week

submitted by /u/ThereWas [link] [comments]

Reddit - Artificial Intelligence · 1 min ·
Llms

What features do you actually want in an AI chatbot that nobody has built yet?

Hey everyone 👋 I'm building a new AI chat app and before I build anything I want to hear from real users first. Current AI tools like Cha...

Reddit - Artificial Intelligence · 1 min ·
Llms

So, what exactly is going on with the Claude usage limits?

I'm extremely new to AI and am building a local agent for fun. I purchased a Claude Pro account because it helped me a lot in the past wh...

Reddit - Artificial Intelligence · 1 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime