Please use this identifier to cite or link to this item:
http://dspace.bits-pilani.ac.in:8080/jspui/handle/123456789/16146
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Goyal, Poonam | - |
dc.contributor.author | Goyal, Navneet | - |
dc.date.accessioned | 2024-10-21T06:10:26Z | - |
dc.date.available | 2024-10-21T06:10:26Z | - |
dc.date.issued | 2024 | - |
dc.identifier.uri | https://ojs.aaai.org/index.php/AAAI/article/view/28686 | - |
dc.identifier.uri | http://dspace.bits-pilani.ac.in:8080/jspui/handle/123456789/16146 | - |
dc.description.abstract | Satellite data bolstered by their increasing accessibility is leading to many endeavors of automated monitoring of the earth's surface for various applications. Such applications demand high spatial resolution images at a temporal resolution of a few days which entails the challenge of processing a huge volume of image time series data. To overcome this computing bottleneck, we present PatchNet, a bespoke adaptation of beam search and attention mechanism. PatchNet is an automated patch selection neural network that requires only a partial spatial traversal of an image time series and yet achieves impressive results. Satellite systems face a trade-off between spatial and temporal resolutions due to budget/technical constraints e.g., Landsat-8/9 or Sentinel-2 have high spatial resolution whereas, MODIS has high temporal resolution. To deal with the limitation of coarse temporal resolution, we propose FuSITSNet, a twofold feature-based generic fusion model with multimodal learning in a contrastive setting. It produces a learned representation after fusion of two satellite image time series leveraging finer spatial resolution of Landsat and finer temporal resolution of MODIS. The patch alignment module of FuSITSNet aligns the PatchNet processed patches of Landsat-8 with the corresponding MODIS regions to incorporate its finer resolution temporal features. The untraversed patches are handled by the cross-modality attention which highlights additional hot spot features from the two modalities. We conduct extensive experiments on more than 2000 counties of US for crop yield, snow cover, and solar energy prediction and show that even one-fourth spatial processing of image time series produces state-of-the-art results. FuSITSNet outperforms the predictions of single modality and data obtained using existing generative fusion models and allows for monitoring of dynamic phenomena using freely accessible images, thereby unlocking new opportunities. | en_US |
dc.language.iso | en | en_US |
dc.publisher | Association for the Advancement of Artificial Intelligence | en_US |
dc.subject | Computer Science | en_US |
dc.subject | DMKM: Mining of Spatial | en_US |
dc.subject | Temporal or Spatio-Temporal Data | en_US |
dc.subject | Deep Learning Algorithms | en_US |
dc.subject | ML: Multimodal Learning | en_US |
dc.subject | Representation Learning | en_US |
dc.title | Efficient Representation Learning of Satellite Image Time Series and Their Fusion for Spatiotemporal Applications | en_US |
dc.type | Article | en_US |
Appears in Collections: | Department of Computer Science and Information Systems |
Files in This Item:
There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.