Challenge Description
Despite the development of several ontology reasoning
optimizations, the traditional methods either do not scale well or only cover a
subset of OWL 2 language constructs. As an alternative, neuro-symbolic
approaches are gaining significant attention. However, the existing methods can
not deal with very expressive ontology languages. Other than that, some
SPARQL query engines also support reasoning, but their performance also is
still limited. To find and improve these performance bottlenecks of the reasoners,
we ideally need several real-world ontologies that span the broad spectrum in
terms of their size and expressivity. However, that is often not the case. One of
the potential reasons for the ontology developers to not build ontologies that vary
in terms of size and expressivity is the performance bottleneck of the reasoners.
SemREC aims to deal with this chicken and egg problem.
The second edition of this challenge includes the following tasks-
- Task-1 - Ontologies. Submit a real-world ontology that is a challenge in terms of the reasoning time or memory consumed during reasoning. We expect a detailed description of the ontology along with the analysis of the reasoning performance, the workarounds if any, that were used to make the ontology less challenging (for example, dropping of a few axioms, redesigning the ontology, etc.), and the (potential) applications in which the ontology could be used. We will be evaluating the submitted ontologies based on the time consumed for a reasoning task, such as classification, and the memory consumed during reasoning.
-
Task-2 - Systems
- Ontology/RDFS Reasoners. Submit an ontology/RDFS reasoner
that uses neural-symbolic techniques for reasoning and
optimization. In terms of technique used, the submissions could fall
under any of the below (or related) categories.
- Using learning-based techniques for performance optimization of traditional reasoning algorithms [6].
- Inductive reasoning techniques based on a subsymbolic representation of entities and relations learned through maximization of an objective function over valid triples [4, 5].
- Techniques that can learn the deductive reasoning aspect using the ontology axioms [1, 2, 3].
- Neural Multi-hop reasoners to deal with reasoning where multi-hop inference is required [7, 8].
- SPARQL query engines that support entailment regimes such as RDF, RDFS, or OWL 2. We expect a detailed description of the system, including an evaluation of the system on the provided datasets.
- Ontology/RDFS Reasoners. Submit an ontology/RDFS reasoner
that uses neural-symbolic techniques for reasoning and
optimization. In terms of technique used, the submissions could fall
under any of the below (or related) categories.
This challenge will be collocated with the 21st International Semantic Web Conference.
We have a discussion group for the challenge where we share the latest news with the participants and discuss issues related to the evaluation rounds.
Challenge Winners
- The CaLiGraph Ontology as a Challenge for OWL Reasoners
- Query Answering and Scaling Extensions of Konclude
- EmELvar : A Neuro-Symbolic Reasoner for the EL++ Description Logics
Recorded Presentations
- Task 1
- Task 2
- Task 3
Evaluations
See the details of evaluations here.Schedule
See full ISWC'21 program hereSemREC Presentations:
October 26, SEMANTIC WEB CHALLENGE 2 (EDT (US): 04:30-05:30. CEST (EU): 10:30-11:30. CST (China): 16:30-17:30):
- Challenge overview - 8 min. live.
- A Reasoner-Challenging Ontology from the Microelectronics Domain (presented by Frank Wawrzik). 10 min + 2 min live Q&A
- Reasoning Challenges on Gene Variants Data (presented by Asha Subramanian).10 min + 2 min live Q&A
- CaLiGraph Ontology (presented by Nicolas Heist). 10 min + 2 min live Q&A
- DACOC3 (presented by Johannes Frey). 10 min recorded + 2 min live Q&A
- Wrap-up - 2 min live.
- Task-1.
We expect a detailed description of the ontology along with the analysis of the reasoning performance, the workarounds, if any, that were used to make the ontology less challenging (for example, dropping of a few axioms, redesigning the ontology, etc.), and the (potential) applications in which the ontology could be used.
To generate the leaderboard, the submitted ontologies will be run on some of the traditional description logic reasoners such as Konclude, ELK, and Openllet . We will use reasoning (classification) time and the memory consumed as the primary metrics. We will use a timeout value of 6 hours, and the limit on the memory will be 96 GB. - Task-2.
- To generate the leaderboard, the neural-symbolic reasoners will be provided with training, validation, and test datasets. We provide an eval.py file for the subsumption task. This is provided only to give an idea of the kind of submission we expect from the participants. Participants are requested to make the changes mentioned in the file to evaluate it on their embeddings for the supported reasoning task (e.g., class subsumption, class membership, etc). We would require the class embeddings of your model along with a readme on the changes made on the evaluation file and how to use it. We plan to evaluate the submitted systems on a Linux-based GPU server. The participants will include all the results in the submitted papers and provide their trained models/embeddings. We will further be evaluating the provided models on another small test dataset. The evaluation metrics will be reasoning time, memory consumed, precision and recall.
- To generate the leaderboard, the SPARQL query engines will be evaluated for the scalability aspect in terms of load time, query response time, and memory consumed. We will provide the datasets, and the participants will discuss their evaluations in the submitted paper. For fair evaluation, we will be re-evaluating the submitted systems on our hardware.
Dataset Details
The submitted reasoning systems can vary in terms of their support for different OWL 2 profiles, subsets of Description Logics, reasoning tasks (such as classification, realization, or consistency checking for traditional reasoners, entailment, class membership, class subsumption, or axiom completion for neuro-symbolic reasoners). Ideally, the datasets must cover all these cases. However, due to the uncertainty with regard to the type of system submissions-- given the challenge is being organized for the first time, we are providing the OWL 2 Profile specific datasets only. If there are systems that partially support some profile then the participants can evaluate on part of the provided datasets.
We will use the following datasets for evaluating the reasoners submitted to Tasks 2 and 3.
Submission Details
The submissions can be either in the form of short papers of length 5 pages or long papers of length 10-12 pages. All the submissions must be in English and follow the 1-column CEUR-ART style (overleaf template). The proceedings will be published as a volume of CEUR-WS. Submissions should be made in the form of a pdf document on EasyChair.