Ве молиме користете го овој идентификатор да го цитирате или поврзете овој запис: http://hdl.handle.net/20.500.12188/33585
DC FieldValueLanguage
dc.contributor.authorFilipovska, Elenaen_US
dc.contributor.authorMladenovska, Anaen_US
dc.contributor.authorBajrami, Merxhanen_US
dc.contributor.authorDobreva, Jovanaen_US
dc.contributor.authorHillman, Velislavaen_US
dc.contributor.authorLameski, Petreen_US
dc.contributor.authorZdravevski, Eftimen_US
dc.date.accessioned2025-05-21T07:28:22Z-
dc.date.available2025-05-21T07:28:22Z-
dc.date.issued2024-09-08-
dc.identifier.urihttp://hdl.handle.net/20.500.12188/33585-
dc.description.abstractThe rapid growth of document volumes and com plexity in various domains necessitates advanced automated methods to enhance the efficiency and accuracy of information extraction and analysis. This paper aims to evaluate the efficiency and repeatability of OpenAI’s APIs and other Large Language Models (LLMs) in automating question-answering tasks across multiple documents, specifically focusing on analyzing Data Pri vacy Policy (DPP) documents of selected EdTech providers. We test how well these models perform on large-scale text processing tasks using the OpenAI’s LLM models (GPT 3.5 Turbo, GPT 4, GPT 4o) and APIs in several frameworks: direct API calls (i.e., one-shot learning), LangChain, and Retrieval Augmented Generation (RAG) systems. We also evaluate a local deployment of quantized versions (with FAISS) of LLM models (Llama-2- 13B-chat-GPTQ). Through systematic evaluation against pre defined use cases and a range of metrics, including response format, execution time, and cost, our study aims to provide insights into the optimal practices for document analysis. Our findings demonstrate that using OpenAI’s LLMs via API calls is a workable workaround for accelerating document analysis when using a local GPU-powered infrastructure is not a viable solution, particularly for long texts. On the other hand, the local deployment is quite valuable for maintaining the data within the private infrastructure. Our findings show that the quantized models retain substantial relevance even with fewer parameters than ChatGPT and do not impose processing restrictions on the number of tokens. This study offers insights on maximizing the use of LLMs for better efficiency and data governance in addition to confirming their usefulness in improving document analysis procedures.en_US
dc.publisherIEEEen_US
dc.subjectOpenAI, LangChain, RAG, GPT, QA, LLM, Llama, Large Language Models, Multi-document, one-shot learn ing, few-shot learning Q&Aen_US
dc.titleBenchmarking OpenAI’s APIs and other Large Language Models for Repeatable and Efficient Question Answering Across Multiple Documentsen_US
dc.typeProceedingsen_US
dc.relation.conference2024 19th Conference on Computer Science and Intelligence Systems (FedCSIS)en_US
item.fulltextWith Fulltext-
item.grantfulltextopen-
crisitem.author.deptFaculty of Computer Science and Engineering-
crisitem.author.deptFaculty of Computer Science and Engineering-
Appears in Collections:Faculty of Computer Science and Engineering: Conference papers
Files in This Item:
File SizeFormat 
3979.pdf288.48 kBAdobe PDFView/Open
Прикажи едноставен запис

Google ScholarTM

Проверете


Записите во DSpace се заштитени со авторски права, со сите права задржани, освен ако не е поинаку наведено.