Large Language Model for OWL Proofs
This paper evaluates Large Language Models (LLMs) on their ability to construct and explain proofs using OWL (Web Ontology Language) ontologies, finding that while some models perform strongly, they struggle significantly with conclusions requiring complex derivation patterns, noisy input data, and incomplete premises. The study reveals that logical complexity, rather than the input format (formal logic vs. natural language), is the primary factor limiting LLM performance in these tasks.