Please use this identifier to cite or link to this item: http://hdl.handle.net/20.500.12188/25396
Title: Survey of nlp in pharmacology: Methodology, tasks, resources, knowledge, and tools
Authors: Trajanov, Dimitar 
Trajkovski, Vangel
Dimitrieva, Makedonka
Dobreva, Jovana
Jovanovik, Milos
Klemen, Matej
Žagar, Aleš
Robnik-Šikonja, Marko
Issue Date: 22-Aug-2022
Journal: arXiv preprint arXiv:2208.10228
Abstract: Natural language processing (NLP) is an area of artificial intelligence that applies information technologies to process the human language, understand it to a certain degree, and use it in various applications. This area has rapidly developed in the last few years and now employs modern variants of deep neural networks to extract relevant patterns from large text corpora. The main objective of this work is to survey the recent use of NLP in the field of pharmacology. As our work shows, NLP is a highly relevant information extraction and processing approach for pharmacology. It has been used extensively, from intelligent searches through thousands of medical documents to finding traces of adversarial drug interactions in social media. We split our coverage into five categories to survey modern NLP methodology, commonly addressed tasks, relevant textual data, knowledge bases, and useful programming libraries. We split each of the five categories into appropriate subcategories, describe their main properties and ideas, and summarize them in a tabular form. The resulting survey presents a comprehensive overview of the area, useful to practitioners and interested observers.
URI: http://hdl.handle.net/20.500.12188/25396
Appears in Collections:Faculty of Computer Science and Engineering: Journal Articles

Files in This Item:
File Description SizeFormat 
2208.10228.pdf4.95 MBAdobe PDFView/Open
Show full item record

Page view(s)

49
checked on Apr 28, 2024

Download(s)

11
checked on Apr 28, 2024

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.