Deep semantic understanding of high resolution remote sensing image
- Publication Type:
- Conference Proceeding
- Citation:
- IEEE CITS 2016 - 2016 International Conference on Computer, Information and Telecommunication Systems, 2016
- Issue Date:
- 2016-08-16
Closed Access
Filename | Description | Size | |||
---|---|---|---|---|---|
07546397.pdf | Published version | 4.26 MB |
Copyright Clearance Process
- Recently Added
- In Progress
- Closed Access
This item is closed access and not available.
© 2016 IEEE. With the rapid development of remote sensing technology, huge quantities of high resolution remote sensing images are available now. Understanding these images in semantic level is of great significance. Hence, a deep multimodal neural network model for semantic understanding of the high resolution remote sensing images is proposed in this paper, which uses both visual and textual information of the high resolution remote sensing images to generate natural sentences describing the given images. In the proposed model, the convolution neural network is utilized to extract the image feature, which is then combined with the text descriptions of the images by RNN or LSTMs. And in the experiments, two new remote sensing image-captions datasets are built at first. Then different kinds of CNNs with RNN or LSTMs are combined to find which is the best combination for caption generation. The experiments results prove that the proposed method achieves good performances in semantic understanding of high resolution remote sensing images.
Please use this identifier to cite or link to this item: