Please use this identifier to cite or link to this item: http://hdl.handle.net/20.500.12188/21631
Title: Average Vibrational Potentials of Oscillators in Condensed-matter Environments using Hadoop
Authors: Koteska, Bojana 
Pejov, LJupcho 
Mishev, Anastas 
Keywords: Hadoop
Issue Date: 2014
Publisher: Faculty of Computer Science and Engineering, Skopje, North Macedonia
Source: Bojana Koteska, Ljupco Pejov, and Anastas Mishev. “Average Vibrational Potentials of Oscillators in Condensed-matter Environments using Hadoop”. In: Proceedings of the 11th Conference for Informatics and Information Technology. Bitola, Macedonia: Faculty of Computer Science and Engineering, Skopje, Macedonia, 2014, pp. 311–314. ISBN: 978-608-4699-04-0.
Conference: 11th International Conference on Informatics and Information Technologies CIIT 2014
Abstract: In physical sciences, when condensed matter systems (e.g. solids or liquids) are modeled with an explicit inclusion of dynamical effects, often the following computational problem arises. A given property of an embedded atomic/molecular system within condensed phase should be computed either at different possible structural arrangements and further average over configurations, or alternatively, it is possible to generate an averaged configuration of the dynamical surrounding that the system experiences and further compute the property of interest at that configuration. The problem of solving the average vibrational potentials of large number of oscillators in various condensed-matter environments (sampled from a statistical physics simulation) can be placed in the category of problems with large data sets. In this paper, a distributed and parallel processing of the large data sets needed for the generation of the averaged vibrational potential is efficiently performed by using the MapReduce programming model and Hadoop software library. Some of the reasons for choosing the Hadoop software library are: It is able to work on data pieces in parallel; The computing solutions enabled by Hadoop are scalable and flexible; The distributed file system enables rapid data transfer among nodes; Hadoop is fault-tolerant which means that if a node fails the job is redirected to another node. The main goal of this paper is to perform an efficient processing of the large data sets used in the scientific applications.
URI: http://hdl.handle.net/20.500.12188/21631
ISBN: 978-608-4699-04-0
Appears in Collections:Faculty of Computer Science and Engineering: Conference papers

Files in This Item:
File Description SizeFormat 
CIIT2014PREPRESS.pdf260.02 kBAdobe PDFView/Open
Show full item record

Page view(s)

46
checked on Apr 26, 2024

Download(s)

8
checked on Apr 26, 2024

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.