Please use this identifier to cite or link to this item: http://hdl.handle.net/20.500.12188/33586
DC FieldValueLanguage
dc.contributor.authorDimitrovski, Ivicaen_US
dc.contributor.authorSpasev, Vlatkoen_US
dc.contributor.authorKitanovski, Ivanen_US
dc.date.accessioned2025-05-21T07:33:48Z-
dc.date.available2025-05-21T07:33:48Z-
dc.date.issued2024-10-01-
dc.identifier.urihttp://hdl.handle.net/20.500.12188/33586-
dc.description.abstractAccurate semantic segmentation of remote sensing imagery is critical for various Earth observation applications, such as land cover mapping, urban planning, and environmental monitoring. However, individual data sources often present limitations for this task. Very High Resolution (VHR) aerial imagery provides rich spatial details but cannot capture temporal information about land cover changes. Conversely, Satellite Image Time Series (SITS) capture temporal dynamics, such as seasonal variations in vegetation, but with limited spatial resolution, making it difficult to distinguish fine-scale objects. This paper proposes a late fusion deep learning model (LF-DLM) for semantic segmentation that leverages the complementary strengths of both VHR aerial imagery and SITS. The proposed model consists of two independent deep learning branches. One branch integrates detailed textures from aerial imagery captured by UNetFormer with a Multi-Axis Vision Transformer (MaxViT) backbone. The other branch captures complex spatio-temporal dynamics from the Sentinel-2 satellite image time series using a U-Net with Temporal Attention Encoder (U-TAE). This approach leads to state-of-the-art results on the FLAIR dataset, a large-scale benchmark for land cover segmentation using multi-source optical imagery. The findings highlight the importance of multi-modality fusion in improving the accuracy and robustness of semantic segmentation in remote sensing applications.en_US
dc.relation.ispartofarXiv preprint arXiv:2410.00469en_US
dc.subjectEarth observation, semantic segmentation, remote sensing, multi-modality fusion, deep learningen_US
dc.titleDeep Multimodal Fusion for Semantic Segmentation of Remote Sensing Earth Observation Dataen_US
dc.typePreprinten_US
item.grantfulltextopen-
item.fulltextWith Fulltext-
crisitem.author.deptFaculty of Computer Science and Engineering-
crisitem.author.deptFaculty of Computer Science and Engineering-
Appears in Collections:Faculty of Computer Science and Engineering: Journal Articles
Files in This Item:
File SizeFormat 
2410.00469v1.pdf2.13 MBAdobe PDFView/Open
Show simple item record

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.