Abstract
Urban land use is key to rational urban planning and management. Traditional land use classification methods rely heavily on domain experts, which is both expensive and inefficient. In this paper, deep neural network-based approaches are presented to label urban land use at pixel level using high-resolution aerial images and ground-level street view images. We use a deep neural network to extract semantic features from sparsely distributed street view images and interpolate them in the spatial domain to match the spatial resolution of the aerial images, which are then fused together through a deep neural network for classifying land use categories. Our methods are tested on a large publicly available aerial and street view images dataset of New York City, and the results show that using aerial images alone can achieve relatively high classification accuracy, the ground-level street view images contain useful information for urban land use classification, and fusing street image features with aerial images can improve classification accuracy. Moreover, we present experimental studies to show that street view images add more values when the resolutions of the aerial images are lower, and we also present case studies to illustrate how street view images provide useful auxiliary information to aerial images to boost performances.
Original language | English |
---|---|
Article number | 1553 |
Journal | Remote Sensing |
Volume | 10 |
Issue number | 10 |
DOIs | |
Publication status | Published - 1 Oct 2018 |
Keywords
- Aerial images
- Convolutional neural network (CNN)
- Data fusion
- Deep learning
- Land use classification
- Semantic segmentation
- Street view images
ASJC Scopus subject areas
- General Earth and Planetary Sciences