DSpace logo

Please use this identifier to cite or link to this item: http://dspace.bits-pilani.ac.in:8080/jspui/handle/123456789/16146
Full metadata record
DC FieldValueLanguage
dc.contributor.authorGoyal, Poonam-
dc.contributor.authorGoyal, Navneet-
dc.date.accessioned2024-10-21T06:10:26Z-
dc.date.available2024-10-21T06:10:26Z-
dc.date.issued2024-
dc.identifier.urihttps://ojs.aaai.org/index.php/AAAI/article/view/28686-
dc.identifier.urihttp://dspace.bits-pilani.ac.in:8080/jspui/handle/123456789/16146-
dc.description.abstractSatellite data bolstered by their increasing accessibility is leading to many endeavors of automated monitoring of the earth's surface for various applications. Such applications demand high spatial resolution images at a temporal resolution of a few days which entails the challenge of processing a huge volume of image time series data. To overcome this computing bottleneck, we present PatchNet, a bespoke adaptation of beam search and attention mechanism. PatchNet is an automated patch selection neural network that requires only a partial spatial traversal of an image time series and yet achieves impressive results. Satellite systems face a trade-off between spatial and temporal resolutions due to budget/technical constraints e.g., Landsat-8/9 or Sentinel-2 have high spatial resolution whereas, MODIS has high temporal resolution. To deal with the limitation of coarse temporal resolution, we propose FuSITSNet, a twofold feature-based generic fusion model with multimodal learning in a contrastive setting. It produces a learned representation after fusion of two satellite image time series leveraging finer spatial resolution of Landsat and finer temporal resolution of MODIS. The patch alignment module of FuSITSNet aligns the PatchNet processed patches of Landsat-8 with the corresponding MODIS regions to incorporate its finer resolution temporal features. The untraversed patches are handled by the cross-modality attention which highlights additional hot spot features from the two modalities. We conduct extensive experiments on more than 2000 counties of US for crop yield, snow cover, and solar energy prediction and show that even one-fourth spatial processing of image time series produces state-of-the-art results. FuSITSNet outperforms the predictions of single modality and data obtained using existing generative fusion models and allows for monitoring of dynamic phenomena using freely accessible images, thereby unlocking new opportunities.en_US
dc.language.isoenen_US
dc.publisherAssociation for the Advancement of Artificial Intelligenceen_US
dc.subjectComputer Scienceen_US
dc.subjectDMKM: Mining of Spatialen_US
dc.subjectTemporal or Spatio-Temporal Dataen_US
dc.subjectDeep Learning Algorithmsen_US
dc.subjectML: Multimodal Learningen_US
dc.subjectRepresentation Learningen_US
dc.titleEfficient Representation Learning of Satellite Image Time Series and Their Fusion for Spatiotemporal Applicationsen_US
dc.typeArticleen_US
Appears in Collections:Department of Computer Science and Information Systems

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.