[2602.15158] da Costa and Tarski meet Goguen and Carnap: a novel approach for ontological heterogeneity based on consequence systems

[2602.15158] da Costa and Tarski meet Goguen and Carnap: a novel approach for ontological heterogeneity based on consequence systems

arXiv - AI 3 min read Article

Summary

This paper introduces a novel approach to ontological heterogeneity, integrating concepts from Carnapian-Goguenism and consequence systems to enhance the understanding of ontology in artificial intelligence.

Why It Matters

The research addresses a critical gap in applied ontology by proposing a framework that reconciles different ontological perspectives. This is significant for AI development, as it can improve interoperability and knowledge representation across diverse systems, fostering advancements in AI applications.

Key Takeaways

  • Introduces da Costian-Tarskianism, a framework for ontological heterogeneity.
  • Utilizes consequence systems to relate different ontologies through morphisms.
  • Defines extended consequence systems and development graphs for enhanced ontology management.
  • Suggests future research directions in applied ontology.
  • Highlights the implications for AI interoperability and knowledge representation.

Computer Science > Artificial Intelligence arXiv:2602.15158 (cs) [Submitted on 16 Feb 2026] Title:da Costa and Tarski meet Goguen and Carnap: a novel approach for ontological heterogeneity based on consequence systems Authors:Gabriel Rocha View a PDF of the paper titled da Costa and Tarski meet Goguen and Carnap: a novel approach for ontological heterogeneity based on consequence systems, by Gabriel Rocha View PDF HTML (experimental) Abstract:This paper presents a novel approach for ontological heterogeneity that draws heavily from Carnapian-Goguenism, as presented by Kutz, Mossakowski and Lücke (2010). The approach is provisionally designated da Costian-Tarskianism, named after da Costa's Principle of Tolerance in Mathematics and after Alfred Tarski's work on the concept of a consequence operator. The approach is based on the machinery of consequence systems, as developed by Carnielli et al. (2008) and Citkin and Muravitsky (2022), and it introduces the idea of an extended consequence system, which is a consequence system extended with ontological axioms. The paper also defines the concept of an extended development graph, which is a graph structure that allows ontologies to be related via morphisms of extended consequence systems, and additionally via other operations such as fibring and splitting. Finally, we discuss the implications of this approach for the field of applied ontology and suggest directions for future research. Comments: Subjects: Artificial Intelligence...

Related Articles

[2511.06448] When AI Agents Collude Online: Financial Fraud Risks by Collaborative LLM Agents on Social Platforms
Llms

[2511.06448] When AI Agents Collude Online: Financial Fraud Risks by Collaborative LLM Agents on Social Platforms

Abstract page for arXiv paper 2511.06448: When AI Agents Collude Online: Financial Fraud Risks by Collaborative LLM Agents on Social Plat...

arXiv - AI · 4 min ·
[2510.20728] Co-Designing Quantum Codes with Transversal Diagonal Gates via Multi-Agent Systems
Ai Agents

[2510.20728] Co-Designing Quantum Codes with Transversal Diagonal Gates via Multi-Agent Systems

Abstract page for arXiv paper 2510.20728: Co-Designing Quantum Codes with Transversal Diagonal Gates via Multi-Agent Systems

arXiv - AI · 4 min ·
[2510.06800] FURINA: A Fully Customizable Role-Playing Benchmark via Scalable Multi-Agent Collaboration Pipeline
Llms

[2510.06800] FURINA: A Fully Customizable Role-Playing Benchmark via Scalable Multi-Agent Collaboration Pipeline

Abstract page for arXiv paper 2510.06800: FURINA: A Fully Customizable Role-Playing Benchmark via Scalable Multi-Agent Collaboration Pipe...

arXiv - AI · 4 min ·
[2509.12626] DoubleAgents: Human-Agent Alignment in a Socially Embedded Workflow
Ai Safety

[2509.12626] DoubleAgents: Human-Agent Alignment in a Socially Embedded Workflow

Abstract page for arXiv paper 2509.12626: DoubleAgents: Human-Agent Alignment in a Socially Embedded Workflow

arXiv - AI · 4 min ·
More in Ai Agents: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime