News
Introduction
LM-KBC Challenge @ ISWC 2024
Task Description
Pretrained language models (LMs) like ChatGPT have advanced a range of semantic tasks and have also shown promise for knowledge extraction from the models itself. Although several works have explored this ability in a setting called probing or prompting, the viability of knowledge base construction from LMs remains underexplored. In the 3rd edition of this challenge, we invite participants to build actual disambiguated knowledge bases from LMs, for given subjects and relations. In crucial difference to existing probing benchmarks like LAMA (Petroni et al., 2019), we make no simplifying assumptions on relation cardinalities, i.e., a subject-entity can stand in relation with zero, one, or many object-entities. Furthermore, submissions need to go beyond just ranking predicted surface strings and materialize disambiguated entities in the output, which will be evaluated using established KB metrics of precision and recall.
s
)
and
relation
(r
), the task is to predict all the correct object-entities
({o1
, o2
, ...,
ok
})
using LM probing.
Special Features
This year, we impose a 10B parameter limit for participing systems, that ensures that no team can simply outperform others by monetary investment.
We will look at a smaller set of 5 relations (last year: 20), with very distinctive features:
- countryLandBordersCountry: Null values possible (e.g., Iceland)
- personHasCityOfDeath: Null values possible
- seriesHasNumberOfEpisodes: Object is numeric
- awardWonBy: Many objects per subject (e.g., 224 Physics Nobel prize winners)
- companyTradesAtStockExchange: Null values possible
Final result
Final Ranking
The following table lists the final ranking of participating systems. Systems without a accompanying description were not ranked, since we could not verify whether they relied on LLM knowledge. All paper submissions will receive reviews shortly.
Paper Submission # | User on CodaLab | Results in paper | Results on leaderboard | Rank | Authors |
---|---|---|---|---|---|
3 | davidebara | 0.91-0.94 | 0.9224 | 1 | Davide Mario Ricardo Bara |
4 | marcelomachado | 0.9083 | 0.9083 | 2 | Marcelo de Oliveira Costa Machado, João Marcello Bessa Rodrigues, Guilherme Lima, Viviane Torres da Silva |
6 | Thin | 0.698 | 0.6977 | 3 | Thin Prabhong, Natthawut Kertkeidkachorn, Areerat Trongratsameethong |
5 | NadeenFathallah | 0.653 | 0.6529 | 4 | Arunav Das, Nadeen Fathallah, Nicole Obretincheva |
2 | hannaabiakl-dsti | 0.6872 | 0.6872 | - | Hanna Abi Akl |
- | Borista | - | 0.9131 | - | - |
- | Rajaa | - | 0.5662 | - | - |
- | aunsiels | - | 0.5076 | - | - |
Calls
Call for Participants
Important Dates
Activity | Dates |
---|---|
Dataset (train and dev) release | 30 March 2024 |
Release of test dataset | |
Submission of test output and systems | |
Submission of system description | 2 August 2024 |
Winner announcement | 16 August 2024 |
Presentations@ISWC (hybrid) | November 11 or 12, 2024 |
Submission Details
Participants are required to submit:
- A system implementing the LM probing approach, uploaded to a public GitHub repo
- The output for the test dataset subject entites, in the same GitHub repo
- A system description in PDF format (5-12 pages, CEUR workshop style), mentioning the GitHub repo, submitted to OpenReview.
Optionally, system outputs can be submitted on Codalab:
Organization
Challenge Organizers
Jan-Christoph Kalo
VU Amsterdam
Tuan-Phong Nguyen
MPI for Informatics
Simon Razniewski
Bosch Center for AI
Bohui Zhang
King's College London
Contact
For general questions or discussion please use the Google Group.
Past Editions
Our challenge has been running since 2022. For more information on past editions, please visit the corresponding websites:
- 2nd Edition: LM-KBC 2023
- 1st Edition: LM-KBC 2022