Expanding Knowledge Graphs Through Text: Leveraging Large Language Models for Inductive Link Prediction
dc.contributor.author | Hamann, Felix | |
dc.contributor.author | Falk, Maurice | |
dc.contributor.author | Walker, Lukas | |
dc.contributor.editor | Klein, Maike | |
dc.contributor.editor | Krupka, Daniel | |
dc.contributor.editor | Winter, Cornelia | |
dc.contributor.editor | Gergeleit, Martin | |
dc.contributor.editor | Martin, Ludger | |
dc.date.accessioned | 2024-10-21T18:24:13Z | |
dc.date.available | 2024-10-21T18:24:13Z | |
dc.date.issued | 2024 | |
dc.description.abstract | Knowledge graphs (KG) play a crucial role for knowledge modeling in various domains such as web search, medical applications, or technical support, yet they are often incomplete. To mitigate this problem, knowledge graph completion (KGC) may be used to infer missing links of the graph. Taking it a step further, in an automated knowledge acquisition process, links for entirely new, unseen entities may be incorporated. This process is known as inductive link prediction (I-LP). Optionally, text as an external source of information is leveraged to infer the correct linkage of such entities. Depending on the context, this text either provides a comprehensive singular description of the entity or includes numerous incidental references to it. This paper presents a study that explores the application of LLAMA3 as a representative of the current generation of large language models (LLM) to I-LP. Through experimentation on popular benchmark datasets such as Wikidata5m, FB15k-237, WN18-RR, and IRT2, we evaluate the performance of LLMs for inserting new facts into a knowledge base, given textual references to the target object. These benchmarks, by design, exhibit significant variations in the quality of the associated text, as well as in the number of entities and links included. This paper explores several prompt formulations and studies whether pre-emptive retrieval of text helps. For automated link prediction, we implement the full cycle of prompt generation, answer processing, entity candidate lookup, and finally link prediction. Our results show that LLM-based inductive link prediction is outperformed by previously suggested models which fine-tune task-specific LM encoders. | en |
dc.identifier.doi | 10.18420/inf2024_123 | |
dc.identifier.isbn | 978-3-88579-746-3 | |
dc.identifier.pissn | 1617-5468 | |
dc.identifier.uri | https://dl.gi.de/handle/20.500.12116/45096 | |
dc.language.iso | en | |
dc.publisher | Gesellschaft für Informatik e.V. | |
dc.relation.ispartof | INFORMATIK 2024 | |
dc.relation.ispartofseries | Lecture Notes in Informatics (LNI) - Proceedings, Volume P-352 | |
dc.subject | Inductive Link Prediction | |
dc.subject | Knowledge Graph Completion | |
dc.subject | Large Language Models | |
dc.subject | Prompting | |
dc.title | Expanding Knowledge Graphs Through Text: Leveraging Large Language Models for Inductive Link Prediction | en |
dc.type | Text/Conference Paper | |
gi.citation.endPage | 1417 | |
gi.citation.publisherPlace | Bonn | |
gi.citation.startPage | 1407 | |
gi.conference.date | 24.-26. September 2024 | |
gi.conference.location | Wiesbaden | |
gi.conference.sessiontitle | AI@WORK |
Dateien
Originalbündel
1 - 1 von 1
Lade...
- Name:
- Hamann_et_al_Expanding_Knowledge_Graphs.pdf
- Größe:
- 518.09 KB
- Format:
- Adobe Portable Document Format