Quantum Dilated Convolutional Neural Network: History
Please note this is an old version of this entry, which may differ significantly from the current revision.

Road network extraction is a significant challenge in remote sensing (RS). Automated techniques for interpreting RS imagery offer a cost-effective solution for obtaining road network data quickly, surpassing traditional visual interpretation methods.

  • road extraction
  • remote sensing
  • convolutional neural networks

1. Introduction

Remote sensing images (RSI) find diverse applications in urban planning, building footprints extraction, and disaster management. Among the crucial aspects of urban areas is the structure of road, which plays a vital role in urban planning, automated navigation, transportation systems and unmanned vehicles [1]. Researchers in the field of RSI processing have a keen interest in extracting road networks, and high-resolution RS data is a valuable resource for real-time road network updates [2]. Thus, presenting a novel road-structure extraction approach from these images aids geospatial information systems (GIS) and intelligent transportation systems (ITS). However, several challenges complicate the extraction of roads from high-resolution RSI [3]. For example, extracting additional features from high-resolution images, such as tree shadows, vehicles on the road, and buildings alongside the road, presents difficulties [4]. Road networks exhibit intricate designs in RSI, with road segments appearing uneven. Accurate road structure extraction from aerial imagery is widely acknowledged as challenging due to the diverse road-type shadows and occlusion resulting from the proximity of trees and buildings. [5]. Previous studies have identified the five key factors for road extraction from aerial images as geometrical factors, including road curvature and length to breadth; radiometric factors [6]; road surface homogeneity and consistent gray color contrast; topological factors, as roads form interconnected networks without abrupt endings for topological reasons; and functional factors, such as the connecting of various regions within a city, including residential and commercial areas [7][8][9][10]. These factors collectively contribute to the road’s overall characteristics, but lighting conditions and obstructions can alter their appearance, adding to the complexities of road extraction [7][8][9][10]. Researchers have turned to artificial intelligence (AI) techniques, utilizing the important usefulness of deep convolutional neural networks (DCNNs) in diverse computer vision (CV) domains, to tackle the extraction of road networks from high-resolution RSI. Convolutional neural networks (CNNs) were first introduced by Yann Le Cun et al. in 1989 as a robust deep learning technique [11]. CNNs have demonstrated exceptional proficiency in the automated extraction of features from various types of data, thus proving their efficacy in computer vision tasks [12][13][14][15]. Simultaneously, progress has been made in quantum technologies.
The discipline of quantum machine learning is rapidly growing and has demonstrated its ability to enhance classical machine learning methods [16][17][18][19][20][21][22][23][24][25], including support vector machines, clustering, and principal component analysis. Quantum convolutional neural networks (QCNNs) are a notable field of study, representing a subset of variational quantum algorithms. QCNNs integrate quantum convolutional layers that employ parameterized quantum circuits to approximate intricate kernel functions within a high-dimensional Hilbert space. Liu et al. (2019) pioneered the development of the first QCNN model for image identification, drawing inspiration from regular CNNs [26]. This groundbreaking work has since sparked further investigation and research in the field, as evidenced by following publications [27][28][29][30][31][32], motivating the application of QCNN with improvement in its basic architecture for road extraction from HRSI.
Significant advancements have been made in extracting high-level features and improving the performance of numerous computer vision tasks, such as object detection, classification, and semantic segmentation [33][34]. These approaches demonstrate superior results compared to traditional methods, particularly when addressing the challenges posed by obstacles and shadow occlusion, geometrical factors, road curvature, length to breadth ratio, and radiometric factors [6] in road extraction from high-resolution imagery.

2. Quantum Dilated Convolutional Neural Network

Shao et al. [35] presented a novel road extraction network that incorporates an attention mechanism, aiming to address the task of automating the extraction of road networks from large volumes of remote sensing imagery (RSI). Their approach builds upon the U-Net architecture, which leverages spatial and spectral information and incorporates spatial and channel attention mechanisms. In addition, the researchers incorporated a residual dilated convolution module into their approach to capture road network data at various scales. They also integrated residual, densely connected blocks to effectively improve feature reuse and information flow. In a separate study [36], the researcher employed RADANet, an abbreviation for road-augmented deformable attention network, in order to effectively capture extensive interdependencies among particular road pixels. This was motivated by prior knowledge of road morphologies and advancements in deformable convolutions.
Li et al. [37] introduced a cascaded attention-enhanced framework designed to extract roadways with finer boundaries from remote sensing imagery (RSI). The proposed architecture integrates many levels of channel attention to enhance the fusion of multiscale features. Additionally, it incorporates a spatial attention residual block to effectively capture long-distance interactions within the multiscale characteristics. In addition, a lightweight encoder–decoder network is used in order to enhance the accuracy of road boundary extraction. Yan et al. [38] proposed an innovative approach to road surface extraction, incorporating a graph neural network (GNN) that operates on a pre-existing road graph composed of road centerlines. The suggested method exploits the GNN approach for vertex adjustment and employs CNN-based feature extraction to define road surface extraction as a two-sided width inference problem of the road graph. Rajamani et al. [39] aimed to develop an automated road recognition system and a building footprint extraction system using CNN from hyperspectral images. They employed polygon segmentation to detect and extract spectral features from hyperspectral data. CNN with different kernels was used to classify the retrieved spectral features into two categories: building footprints and road detection. The authors introduced a novel deep neural network approach, referred to as dual-decoder-U-Net (DDU-Net), in their study [40]. The authors incorporated global average pooling and cascading dilated convolutions to distill multiscale features. Additionally, a dilated convolution attention module (DCAM) was introduced between the encoder and decoder to expand the receptive field. The authors of reference [41] have proposed a novel road extraction network named DA-RoadNet, which integrates the ability to incorporate semantic reasoning. The primary architecture of DA-RoadNet consists of a shallow network that connects the encoder to the decoder. This network incorporates densely connected blocks in order to address the issue of road infrastructure data loss resulting from several down-sampling procedures. Hou et al. [42] proposed a route extraction approach for RSI using a complementary U-Net (C-UNet) with four modules. They introduced an MD-UNet (multi-scale dense dilated convolutional U-Net) to identify complementary road regions in the removed masks, after the standard U-Net was employed for rough road data extraction from RSI and generated the initial segmentation result.
The practical execution of many quantum circuits still poses challenges. QCNNs face computational difficulties due to the need to execute additional circuits for quantum operations and gradient calculations [28][29]. The utilization of quantum filters that possess trainable characteristics further exacerbates this concern. Unlike classical CNNs, QCNNs often lack vectorization capabilities on the majority of quantum devices, hence impeding their scalability [43][44].
To reduce the runtime complexity of QCNN, two main approaches are prominent. Firstly, dimension reduction techniques such as principal component analysis (PCA) and autoencoding can reduce the required qubits, but they may constrain the model’s expressiveness [45][46]. Secondly, the efficient conversion of classical data into quantum states is pursued through encoding methods. Amplitude encoding conserves qubits but relies on complex quantum circuits [47][48]. Conversely, angle encoding and its variants maintain consistent circuit depth but may be less efficient for high-dimensional data [32][49][50]. A hybrid encoding approach strikes a balance between qubit usage and circuit depth [46], while threshold-based encoding simplifies quantum convolution but may have limitations on real quantum devices [28].
Considering the various challenges that have been thoroughly examined and the subsequent advancements made, this study presents an unconventional quantum-classical architecture called the quantum dilated convolutional neural network (QDCNN) for road extraction with the Archimedes tuning process (ATP) from high-resolution remote sensing images. Initially, the proposed methodology benefited from previous architectures [26][28], and for the dilated convolutional neural network, it uses the architecture described in [51] and introduces a new strategy to decrease the computing expenses of QCNN in the use of a quanvolutional layer [28], drawing inspiration from dilated convolution techniques in deep learning. The utilization of dilated convolution, which was initially devised for discrete wavelet transformations [52], has become increasingly prominent in various fields, such as semantic segmentation [21][53][54][55][56][57], object localization [58], sound classification [59] and time-series forecasting [60][61]. The utilization of dilated convolution in QDCNNs effectively increases the filter context, resulting in improved computing efficiency without any additional parameters or complexity.

This entry is adapted from the peer-reviewed paper 10.3390/s23218783

References

  1. Xu, H.; He, H.; Zhang, Y.; Ma, L.; Li, J. A comparative study of loss functions for road segmentation in remotely sensed road datasets. Int. J. Appl. Earth Obs. Geoinf. 2023, 116, 103159.
  2. Chen, W.; Zhou, G.; Liu, Z.; Li, X.; Zheng, X.; Wang, L. NIGAN: A framework for mountain road extraction integrating remote sensing road-scene neighborhood probability enhancements and improved conditional generative adversarial network. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5626115.
  3. Chen, Z.; Wang, C.; Li, J.; Fan, W.; Du, J.; Zhong, B. Adaboost-like End-to-End multiple lightweight U-nets for road extraction from optical remote sensing images. Int. J. Appl. Earth Obs. Geoinf. 2021, 100, 102341.
  4. Behera, T.K.; Bakshi, S.; Sa, P.K.; Nappi, M.; Castiglione, A.; Vijayakumar, P.; Gupta, B.B. The NITRDrone dataset to address the challenges for road extraction from aerial images. J. Signal Process. Syst. 2023, 95, 197–209.
  5. Sultonov, F.; Park, J.H.; Yun, S.; Lim, D.W.; Kang, J.M. Mixer U-Net: An improved automatic road extraction from UAV imagery. Appl. Sci. 2022, 12, 1953.
  6. Bayramoğlu, Z.; Melis, U.Z.A.R. Performance analysis of rule-based classification and deep learning method for automatic road extraction. Int. J. Eng. Geosci. 2023, 8, 83–97.
  7. Li, J.; Meng, Y.; Dorjee, D.; Wei, X.; Zhang, Z.; Zhang, W. Automatic road extraction from remote sensing imagery using ensemble learning and postprocessing. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 10535–10547.
  8. Khan, M.J.; Singh, P.P. Advanced Road Extraction using CNN-based U-Net Model and Satellite Imagery. E-Prime-Adv. Electr. Eng. Electron. Energy 2023, 5, 100244.
  9. Chen, Z.; Deng, L.; Luo, Y.; Li, D.; Junior, J.M.; Gonçalves, W.N.; Nurunnabi, A.A.M.; Li, J.; Wang, C.; Li, D. Road extraction in remote sensing data: A survey. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102833.
  10. Khan, M.J.; Singh, P.P. Road Extraction from Remotely Sensed Data: A Review. AIJR Proc. 2021, 106–111.
  11. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324.
  12. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105.
  13. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.155.
  14. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9.
  15. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778.
  16. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587.
  17. Girshick, R. Fast R-CNN 2015. In Proceedings of the 2015 IEEE International Conference on Computer Vision, Boston, MA, USA, 7–12 June 2015; pp. 1440–1448.
  18. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788.
  19. He, K.; Gkioxari, G.; Doll, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969.
  20. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the 2015 International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241.
  21. Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the 2018 European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818.
  22. Sarker, I.H. Machine learning: Algorithms, real-world applications and research directions. SN Comput. Sci. 2021, 2, 160.
  23. Kerenidis, L.; Landman, J.; Luongo, A.; Prakash, A. q-means: A quantum algorithm for unsupervised machine learning. arXiv 2018, arXiv:1812.03584.
  24. Otterbach, J.; Manenti, R.; Alidoust, N.; Bestwick, A.; Block, M.; Caldwell, S.; Didier, N.; Fried, E.S.; Hong, S.; Karalekas, P. Unsupervised machine learning on a hybrid quantum computer. arXiv 2017, arXiv:1712.05771.
  25. Lloyd, S.; Mohseni, M.; Rebentrost, P. Quantum principal component analysis. Nat. Phys. 2013, 10, 7.
  26. Liu, J.; Lim, K.H.; Wood, K.L.; Huang, W.; Guo, C.; Huang, H.-L. Hybrid quantum-classical convolutional neural networks. arXiv 2019, arXiv:1911.02998.
  27. Cong, I.; Choi, S.; Lukin, M.D. Quantum convolutional neural networks. Nat. Phys. 2019, 15, 1273–1278.
  28. Henderson, M.; Shakya, S.; Pradhan, S.; Cook, T. Quanvolutional neural networks: Powering image recognition with quantum circuits. Quantum Mach. Intell. 2020, 2, 2.
  29. Oh, S.; Choi, J.; Kim, J. A tutorial on quantum convolutional neural networks (qcnn). In Proceedings of the 2020 International Conference on Information and Communication Technology Convergence (ICTC), Jeju Island, Republic of Korea, 19–21 October 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 236–239.
  30. Chen, S.Y.C.; Wei, T.C.; Zhang, C.; Yu, H.; Yoo, S. Quantum convolutional neural networks for high energy physics data analysis. arXiv 2020, arXiv:2012.12177.
  31. Houssein, E.H.; Abohashima, Z.; Elhoseny, M.; Mohamed, W.M. Hybrid quantum convolutional neural networks model for COVID-19 prediction using chest x-ray images. arXiv 2021, arXiv:2102.06535.
  32. Alam, M.; Kundu, S.; Topaloglu, R.O.; Ghosh, S. Iccad special session paper: Quantum-classical hybrid machine learning for image classification. arXiv 2021, arXiv:2109.02862.
  33. Yang, M.; Yuan, Y.; Liu, G. SDUNet: Road extraction via spatial enhanced and densely connected UNet. Pattern Recognit. 2022, 126, 108549.
  34. Li, Y.; Xiang, L.; Zhang, C.; Jiao, F.; Wu, C. A Guided Deep Learning Approach for Joint Road Extraction and Intersection Detection from RS Images and Taxi Trajectories. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 8008–8018.
  35. Shao, S.; Xiao, L.; Lin, L.; Ren, C.; Tian, J. Road Extraction Convolutional Neural Network with Embedded Attention Mechanism for Remote Sensing Imagery. Remote Sens. 2022, 14, 2061.
  36. Dai, L.; Zhang, G.; Zhang, R. RADANet: Road augmented deformable attention network for road extraction from complex high-resolution remote-sensing images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5602213.
  37. Li, S.; Liao, C.; Ding, Y.; Hu, H.; Jia, Y.; Chen, M.; Xu, B.; Ge, X.; Liu, T.; Wu, D. Cascaded residual attention enhanced road extraction from remote sensing images. ISPRS Int. J. Geo-Inf. 2022, 11, 9.
  38. Yan, J.; Ji, S.; Wei, Y. A combination of convolutional and graph neural networks for regularized road surface extraction. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13.
  39. Rajamani, T.; Sevugan, P.; Ragupathi, S. Automatic building footprint extraction and road detection from hyper-spectral imagery. J. Electron. Imaging 2023, 32, 011005.
  40. Wang, Y.; Peng, Y.; Li, W.; Alexandropoulos, G.C.; Yu, J.; Ge, D.; Xiang, W. DDU-Net: Dual-decoder-U-Net for road extraction using high-resolution remote sensing images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4409113.
  41. Wan, J.; Xie, Z.; Xu, Y.; Chen, S.; Qiu, Q. DA-RoadNet: A dual-attention network for road extraction from high resolution satellite imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 6302–6315.
  42. Hou, Y.; Liu, Z.; Zhang, T.; Li, Y. C-UNet: Complement UNet for remote sensing road extraction. Sensors 2021, 21, 2153.
  43. Linnainmaa, S. Taylor expansion of the accumulated rounding error. BIT Numer. Math. 1976, 16, 146–160.
  44. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning Representations by Back-propagating Errors. Nature 1986, 323, 533–536.
  45. Pramanik, S.; Chandra, M.G.; Sridhar, C.V.; Kulkarni, A.; Sahoo, P.; Vishwa, C.D.; Sharma, H.; Navelkar, V.; Poojary, S.; Shah, P.; et al. A quantum-classical hybrid method for image classification and segmentation. arXiv 2021, arXiv:2109.14431.
  46. Hur, T.; Kim, L.; Park, D.K. Quantum convolutional neural network for classical data classification. arXiv 2021, arXiv:2108.00661.
  47. Schuld, M.; Killoran, N. Quantum machine learning in feature hilbert spaces. Phys. Rev. Lett. 2019, 122, 040504.
  48. Mattern, D.; Martyniuk, D.; Willems, H.; Bergmann, F.; Paschke, A. Variational quanvolutional neural networks with enhanced image encoding. arXiv 2021, arXiv:2106.07327.
  49. Schuld, M.; Petruccione, F. Supervised Learning with Quantum Computers; Springer: Berlin/Heidelberg, Germany, 2018; Volume 17.
  50. LaRose, R.; Coyle, B. Robust data encodings for quantum classifiers. Phys. Rev. A 2020, 102, 032420.
  51. Wang, Y.; Kuang, N.; Zheng, J.; Xie, P.; Wang, M.; Zhao, C. Dilated Convolutional Network for Road Extraction in Remote Sensing Images. In Advances in Brain In-spired Cognitive Systems, Proceedings of the 10th International Conference, BICS 2019, Guangzhou, China, 13–14 July 2019; Springer International Publishing: Berlin/Heidelberg, Germany, 2019; Proceedings 10; pp. 263–272.
  52. Holschneider, M.; Kronland-Martinet, R.; Morlet, J.; Tchamitchian, P. A real-time algorithm for signal analysis with the help of the wavelet transform. In Wavelets, Time-Frequency Methods and Phase Space; Springer: Berlin/Heidelberg, Germany, 1989; Volume 1, pp. 286–297.
  53. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. arXiv 2016, arXiv:1511.07122.
  54. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.P.; Yuille, A.L. Semantic image segmentation with deep convolutional nets and fully connected crfs. arXiv 2015, arXiv:1412.7062.
  55. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848.
  56. Chen, L.-C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587.
  57. Hamaguchi, R.; Fujita, A.; Nemoto, K.; Imaizumi, T.; Hikosaka, S. Effective use of dilated convolutions for segmenting small object instances in remote sensing imagery. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1442–1450.
  58. Kudo, Y.; Aoki, Y. Dilated convolutions for image classification and object localization. In Proceedings of the 2017 Fifteenth IAPR International Conference on Machine Vision Applications (MVA), Nagoya, Japan, 8–12 May 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 452–455.
  59. Chen, Y.; Guo, Q.; Liang, X.; Wang, J.; Qian, Y. Environmental sound classification with dilated convolutions. Appl. Acoust. 2019, 148, 123–132.
  60. Borovykh, A.; Bohte, S.; Oosterlee, C.W. Conditional time series forecasting with convolutional neural networks. arXiv 2017, arXiv:1703.04691.
  61. Chen, Y.; Kang, Y.; Chen, Y.; Wang, Z. Probabilistic forecasting with temporal convolutional neural network. Neurocomputing 2020, 399, 491–501.
More
This entry is offline, you can click here to edit this entry!
Video Production Service