Please use this identifier to cite or link to this item: http://hdl.handle.net/20.500.12188/27403
Title: Enhancing Knowledge Graph Construction Using Large Language Models
Authors: Trajanoska, Milena
Stojanov, Riste 
Trajanov, Dimitar 
Keywords: ChatGPT, REBEL, LLMs, Relation-extraction, NLP, Sustainability
Issue Date: Jul-2023
Publisher: Ss Cyril and Methodius University in Skopje, Faculty of Computer Science and Engineering, Republic of North Macedonia
Series/Report no.: CIIT 2023 papers;28;
Conference: 20th International Conference on Informatics and Information Technologies - CIIT 2023
Abstract: The growing trend of Large Language Models (LLM) development has attracted significant attention, with mod els for various applications emerging consistently. However, the combined application of Large Language Models with semantic technologies for reasoning and inference is still a challenging task. This paper analyzes how the current advances in foundational LLM, like ChatGPT, can be compared with the specialized pretrained models, like REBEL, for joint entity and relation extraction. To evaluate this approach, we conducted several experiments using sustainability-related text as our use case. We created pipelines for the automatic creation of Knowledge Graphs from raw texts, and our findings indicate that using advanced LLM models can improve the accuracy of the process of creating these graphs from unstructured text. Furthermore, we explored the potential of automatic ontology creation using foundation LLM models, which resulted in even more relevant and accurate knowledge graphs.
URI: http://hdl.handle.net/20.500.12188/27403
Appears in Collections:Faculty of Computer Science and Engineering: Conference papers

Files in This Item:
File Description SizeFormat 
CIIT2023_paper_28.pdf9.19 MBAdobe PDFView/Open
Show full item record

Page view(s)

29
checked on Apr 26, 2024

Download(s)

23
checked on Apr 26, 2024

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.