Rahhal, Mohamad M. Al and Bazi, Yakoub and Abdullah, Taghreed and Mekhalfi, Mohamed L. and Zuair, Mansour (2020) Deep Unsupervised Embedding for Remote Sensing Image Retrieval Using Textual Cues. Applied Sciences, 10 (24). ISSN 2076-3417
Text
applsci-10-08931-v2.pdf - Published Version Restricted to Repository staff only Download (5MB) | Request a copy |
Abstract
Compared to image-image retrieval, text-image retrieval has been less investigated in the remote sensing community, possibly because of the complexity of appropriately tying textual data to respective visual representations. Moreover, a single image may be described via multiple sentences according to the perception of the human labeler and the structure/body of the language they use, which magnifies the complexity even further. In this paper, we propose an unsupervised method for text-image retrieval in remote sensing imagery. In the method, image representation is obtained via visual Big Transfer (BiT) Models, while textual descriptions are encoded via a bidirectional Long Short-Term Memory (Bi-LSTM) network. The training of the proposed retrieval architecture is optimized using an unsupervised embedding loss, which aims to make the features of an image closest to its corresponding textual description and different from other image features and vise-versa. To demonstrate the performance of the proposed architecture, experiments are performed on two datasets, obtaining plausible text/image retrieval outcomes.
Item Type: | Article |
---|---|
Uncontrolled Keywords: | remote sensing; big transfer (BiT); text-to-image retrieval; bidirectional long short-term memory network (B-LSTM); unsupervised embedding |
Subjects: | D Physical Science > Computer Science |
Divisions: | Department of > Computer Science |
Depositing User: | Mr Umendra uom |
Date Deposited: | 03 Apr 2021 05:24 |
Last Modified: | 22 Jun 2022 09:11 |
URI: | http://eprints.uni-mysore.ac.in/id/eprint/15460 |
Actions (login required)
View Item |