LLMs4OL 2025: Large Language Models for Ontology Learning
The 2nd LLMs4OL Challenge @ ISWC 2025
ISWC 2025, Nara, Japan | 2-6 November
ISWC 2025, Nara, Japan | 2-6 November
In the LLMs4OL challenge, we have defined four main tasks. OL tasks revolve around ontology primitives:
1. Lexical entries L
2. Conceptual types T
3. A hierarchical taxonomy H_T
4. Non-taxonomic relations R within a heterarchy H_R
5. Axioms A for constraints and rules.
Key OL activities include corpus preparation, terminology extraction, term typing, taxonomy construction, relationship extraction, and axiom discovery. Together, these six tasks constitute the LLMs4OL task framework (see the following figure), aligning with the previously outlined LLMs4OL conceptual model.
Assuming the corpus preparation step is done by reusing ontologies publicly released in the community, for the second iteration of the LLMs4OL Challenge@ISWC-2025, we introduce four main tasks.
Task A - Text2Onto: Extract ontological terminologies and types from a raw text.
Task B - Term Typing: Discover the generalized type for a lexical term.
Task C - Taxonomy Discovery: Discover the taxonomic hierarchy between type pairs.
Task D - Non-Taxonomic Relation Extraction: Identify non-taxonomic, semantic relations between types.
The 1st LLMs4OL challenge marked a notable direction toward employing LLMs in OL, demonstrating their potential in more automated knowledge acquisition. The challenge was structured into two evaluation phases: a few-shot phase, where models were trained on a subset of ontologies before being tested on related but unseen data, and a zero-shot phase, which assessed generalizability by introducing entirely new ontologies. Participants tackled three tasks: Term Typing (assigning types to terms), Taxonomy Discovery (identifying hierarchical relations), and Non-Taxonomic Relation Extraction (detecting semantic relationships beyond hierarchy). The challenge attracted eight teams of participants, each contributing novel systems for OL within the Semantic Web community’s broader goal of making the web more intelligent and user-friendly. Participants explored various strategies, including prompt engineering, fine-tuning, and hybrid models that combined LLMs with rule-based or retrieval-augmented approaches.
Overall, we observed that LLMs excel in Term Typing and Taxonomy Discovery, where their ability to infer hierarchical relationships and generalize across diverse knowledge domains proved beneficial. However, hybrid approaches consistently outperformed pure LLM methods in simpler tasks, indicating that incorporating external knowledge or structured reasoning enhances performance. Dataset diversity significantly impacted results, underscoring the need for well-curated benchmarks that ensure robustness and fairness across different ontologies and domains.
While different prompting and fine-tuning strategies led to varied success rates, Non-Taxonomic Relation Extraction remained a key challenge. LLMs struggled with capturing complex domain-specific relationships, particularly those requiring deeper semantic understanding beyond surface-level lexical cues. This suggests that current LLM architectures, while powerful in generalization, may still require additional structured learning techniques, specialized fine-tuning, or external knowledge retrieval to fully grasp relational semantics.
The challenge results highlight both the strengths and limitations of LLMs in OL. While they provide a promising avenue for reducing manual efforts in knowledge engineering, future work should focus on improving reasoning capabilities, developing domain-adaptive fine-tuning methods, and integrating LLMs with symbolic reasoning frameworks to enhance precision in complex OL tasks.
For further details, refer to the "The 1st LLMs4OL Challenge @ ISWC 2024 Proceeding".
Participation in the LLMs4OL Challenge is flexible, allowing participants to choose one or more tasks among Task A, B, C, and D. Additionally, each task includes multiple ontologies, and participants can select one, multiple, or all available ontologies for their experiments. Participants are encouraged to implement LLM-based solutions with no restrictions on the choice of LLMs or prompting methods. However, preference is given to open-source LLMs to promote transparency and reproducibility.
The ontologies provided for each task are publicly available sources, and participants are strictly required not to recreate new datasets from these ontologies. The challenge organizers reserve a portion of each ontology for testing, and training on re-created datasets would lead to unfair evaluations, as it would no longer constitute a valid OL problem. Participants may introduce additional contextual information from the World Wide Web, but this must be explicitly documented in the system submission, either in the accompanying publication or the README file, as it will be considered part of the methodology.
The challenge consists of two main evaluation phases:
The seen-eval evaluation ontologies are as follows (with their respective domain).
Ontology for Biomedical Investigations (OBI) – Medicine – The OBI is a comprehensive, community-driven ontology that provides a structured framework for representing all aspects of biomedical and clinical investigations. It facilitates consistent annotation and integration of experimental data across diverse biomedical disciplines.
Material Ontology (MatOnto) – Material Science and Engineering – The MatOnto is a domain-specific ontology designed to represent knowledge about materials, their properties, structures, and processing methods, primarily for use in materials science and engineering applications.
Semantic Web for Earth and Environment Technology Ontology (SWEET) – Environment – The SWEET is an investigation in improving the discovery and use of Earth science data, through software understanding of the semantics of web resources. SWEET is a collection of ontologies conceptualizing a knowledge space for Earth system science and includes both orthogonal concepts (space, time, Earth realms, physical quantities, etc.) and integrative science knowledge concepts (phenomena, events, etc.).
Human Disease Ontology (DOID) – Medicine – The Disease Ontology has been developed as a standardized ontology for human disease with the purpose of providing the biomedical community with consistent, reusable, and sustainable descriptions of human disease terms, phenotype characteristics, and related medical vocabulary disease concepts.
Schema.org Ontology (SchemaOrg) – General Knowledge - Schema.org is a collaborative, community activity with a mission to create, maintain, and promote schemas for structured data on the Internet, on web pages, in email messages, and beyond.
PROcess Chemistry Ontology (PROCO) – Chemistry – PROCO is a formal ontology that aims to standardly represent entities and relations among entities in the domain of process chemistry.
Food Ontology (FoodON) – Agricultural – FoodOn, the food ontology, contains vocabulary for naming food materials and their anatomical and taxonomic origins, from raw harvested food to processed food products, for humans and domesticated animals. It provides a neutral and ontology-driven standard for government agencies, industry, nonprofits, and consumers to name and reference food products and their components throughout the food supply chain.
Plant Ontology (PO) – Agricultural – The Plant Ontology (PO) is a structured vocabulary and database resource that links plant anatomy, morphology, and growth and development to plant genomics data.
Gene Ontology (GO) – Biology and Life Sciences –The Gene Ontology (GO) provides structured controlled vocabularies for the annotation of gene products with respect to their molecular function, cellular component, and biological role.
Babaei Giglou, H., D’Souza, J., & Auer, S. (2024). LLMs4OL 2024 Overview: The 1st Large Language Models for Ontology Learning Challenge. Open Conference Proceedings, 4, 3–16. https://doi.org/10.52825/ocp.v4i.2473
Babaei Giglou, H., D’Souza, J., Auer, S. (2023). LLMs4OL: Large Language Models for Ontology Learning. In: Payne, T.R., et al. The Semantic Web – ISWC 2023. ISWC 2023. Lecture Notes in Computer Science, vol 14265. Springer, Cham. https://doi.org/10.1007/978-3-031-47240-4_22.
Babaei Giglou, H, LLMs4OL: Large Language Models for Ontology Learning, (2023). https://github.com/HamedBabaei/LLMs4OL.
The 1st LLMs4OL Challenge @ ISWC 2024 Proceeding. https://www.tib-op.org/ojs/index.php/ocp/issue/view/169.