Next Article in Journal
Attributes’ Importance for Zero-Shot Pose-Classification Based on Wearable Sensors
Previous Article in Journal
A Hydrogen Gas Sensor Based on TiO2 Nanoparticles on Alumina Substrate
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Deep Learning to Identify Utility Poles with Crossarms and Estimate Their Locations from Google Street View Images

1
Department of Geography, University of Connecticut, Storrs, CT 06269, USA
2
Center for Environmental Science and Engineering, University of Connecticut, Storrs, CT 06269, USA
3
Connecticut State Data Center, University of Connecticut, Storrs, CT 06269, USA
4
Department of Natural Resources and the Environment, University of Connecticut, Storrs, CT 06269, USA
5
Eversource Energy Center, University of Connecticut, Storrs, CT 06269, USA
6
MIT Senseable City Lab, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
*
Author to whom correspondence should be addressed.
Submission received: 13 June 2018 / Revised: 21 July 2018 / Accepted: 27 July 2018 / Published: 1 August 2018
(This article belongs to the Section Remote Sensors)

Abstract

:
Traditional methods of detecting and mapping utility poles are inefficient and costly because of the demand for visual interpretation with quality data sources or intense field inspection. The advent of deep learning for object detection provides an opportunity for detecting utility poles from side-view optical images. In this study, we proposed using a deep learning-based method for automatically mapping roadside utility poles with crossarms (UPCs) from Google Street View (GSV) images. The method combines the state-of-the-art DL object detection algorithm (i.e., the RetinaNet object detection algorithm) and a modified brute-force-based line-of-bearing (LOB, a LOB stands for the ray towards the location of the target [UPC at here] from the original location of the sensor [GSV mobile platform]) measurement method to estimate the locations of detected roadside UPCs from GSV. Experimental results indicate that: (1) both the average precision (AP) and the overall accuracy (OA) are around 0.78 when the intersection-over-union (IoU) threshold is greater than 0.3, based on the testing of 500 GSV images with a total number of 937 objects; and (2) around 2.6%, 47%, and 79% of estimated locations of utility poles are within 1 m, 5 m, and 10 m buffer zones, respectively, around the referenced locations of utility poles. In general, this study indicates that even in a complex background, most utility poles can be detected with the use of DL, and the LOB measurement method can estimate the locations of most UPCs.

1. Introduction

Maintaining the electric grid is a challenging task and accurate maps of utility infrastructure are important for planning and operations, managing risk, and rapidly assessing damages after a storm [1]. However, the lack of exact locations of electric facilities is not uncommon [2]. For example, after hurricane Maria struck Puerto Rico in September of 2017, the lack of accurate maps for buildings, bridges, and electric facilities was considered as a main factor slowing recovery efforts [3]. Mapping utility poles is labor- and time-intense because the process is usually conducted using human interpretation of high spatial-resolution aerial imagery, ground-based field surveys, or unmanned aerial vehicles (UAVs)/helicopters [2,4]. The high degree of labor requirement makes mapping utility poles over large areas a daunting task.
Remote sensing (RS) provides promising solutions for automated detection and mapping of electric facilities. Utility mapping has been explored using optical sensors, on both satellite and aerial platforms [5,6,7,8,9,10,11,12]; synthetic aperture radars (SAR) [13,14], and light detection and ranging (LiDAR) [15,16,17,18,19]. Cetin and Bikdash [4] mapped utility poles using shadow information derived from aerial images and Sun et al. [20] mapped power poles using stereo images. Wang et al. [19] developed a semi-automated method to classify power lines from LiDAR data in urban areas with both precision and recall up to 98%. However, due to the small size of utility distribution poles, the spatial resolution of most satellite platforms is not sufficient for reliable detection [4,21]. The ultra-high spatial resolutions provided by UAV platforms have made them an emerging tool for surveying electric utilities [22,23,24].
Aerial optical imagery can reliably detect utility poles when the spatial resolution is about 30 cm or better. However, complex backgrounds, and varying sunlight intensities and other factors can prevent utility pole detection [21]. In addition, tree cover in forested areas, can prevent detection of utility poles from aerial images [25]. Mobile mapping systems (MPS) can obtain a better view of utility poles that are obscured by tree cover in aerial images [21]. Studies conducted by Cabo et al. [26] and Lehtomäki et al. [27] reported that most vertical pole-like objects are accurately detected from side views acquired from vehicle-based LiDAR. Cheng et al. [28] developed a voxel-based hierarchical method to extract power lines from vehicle borne LiDAR data in urban area and reported that the correctness and completeness of the exacted power line points were 99% and 93%. Guan et al. [29] mapped power lines and towers using a step-wise method (including off-road point extraction, power-transmission line extraction, and power-transmission line fitting) from vehicle borne LiDAR data with average completeness, correctness, and quality of 92%, 99%, and 91%. The disadvantages of MPS include their high costs of data collection and the massive sizes of their point cloud, which can be challenging to process efficiently [30].
Google Street View (GSV) is an open image collection of panoramic views with estimated accurate geolocation information using GPS, wheel encoder and inertial navigation sensor along streets acquired on mobile platforms (including car, trekker, tricycle, walking, and boat, etc.) [31], which have been increasingly used to enrich geographic information, such as urban greenery [32,33], land use classification [34,35], and shade provision of trees [36]. Some studies have been carried out using side-view imagery for electric utilities detection and survey. For example, Cheng and Song [37] applied a graph cut segmentation method with a set of prior rules to improve recognition of utility poles. Murthy et al. [38] detected utility poles using a template design from video surveillance on car. Barranco-Gutiérrez et al. [39] presented a method to detect utility poles from complex environment based on color, shape and photometric stereovision using dual cameras. Song and Li [40] developed a sequential local-to-global algorithm to detect power lines from optical images and tested on 160 pictures taken from the ground with 91.95% and 91.33% true positive rates for detecting straight lines and curved lines respectively. However, in general, it is difficult to distinguish utility poles from all vertical pole-like objects along roads even by human interpretation because of their small cross-sectional area [41]. In Sharma et al. [30], a five-stage detection algorithm (including segmentation, block-oriented quadrilateral extraction, quadrilateral shape determination, orientation-based spatial clustering of near-trapeziums, and context-based detection) was developed to detect utility poles in pure side-view images; 70% of poles from 212 frames ground truth were detected. However, the previously used methods are complicated because they involved utilizing a variety of models and algorithms, such as feature segmentation, extractions, filters, detection, and template match, among others.
Deep learning (DL) has shown its powerful ability in computer vision, natural language processing, and many other fields [42,43,44,45,46,47]. However, there are very few published studies on using DL to map or inspect power line components. In Nordeng et al. [48], a faster regional convolutional neural network (R-CNN) was used to detect dead-end body component from high voltage power lines with both detection accuracies and precisions up to 97%. Recently, Nguyen et al. [2] conducted a comprehensive review of automatic power line inspection from the prospective of computer vision and the limitations of current vision-based inspection systems, and they suggested that the DL vision-based unmanned aerial vehicle inspection might be a promising new solution. The objective of this study is to use DL on GSV imagery to map utility poles. We focused on the detection of utility poles with crossarms (UPCs) along roadsides. We used a DL-based object detection algorithm (the RetinaNet object detector [49]) to detect UPCs from GSV images and estimated the locations of detected utility poles using a modified brute-force-based line-of-bearing (LOB) measurement method. Therefore, the primary objective of this study is to explore the use of DL in conjunction with GSV for mapping roadside UPCs.

2. Data

2.1. Study Area and GIS Data

The town of Mansfield (CT, USA) was selected as the study area (Figure 1). Mansfield is a town in Tolland County, located in the Eastern Connecticut, with an area of 118.3 square kilometres based on the 2017 U.S. Census Bureau Topologically Integrated Geographic Encoding and Referencing products. In 2010, Mansfield’s population was 26,543 according to the 2010 U.S. decennial census.
The town boundary and road GIS datasets in vector file format were downloaded from the United State Census Bureau (https://www.census.gov/cgi-bin/geo/shapefiles/index.php). Roads within Mansfield town were extracted and then projected into the NAD83/UTM zone 18N (EPSG: 26918). The extracted road GIS dataset was further pre-processed in order to eliminate duplicated road segments through the following two steps: The first step was to merge all adjacent/overlapped road segments using the “dissolve” tool in ArcGIS software. In the second step, the “Multipart To Singlepart” tool in ArcGIS software was applied to dissolve road segments to generate viewpoints along roads for downloading corresponding GSV via Google API (see Section 2.2 Google Street View imagery). In this study, only major roads in the town were chosen (total length: 91.6 km). A total of 9290 viewpoints were created along the selected roads, at an interval of 10 m, using a geospatial tool called “Create Points on Lines” developed by Ian Broad (http://ianbroad.com/arcgis-toolbox-create-points-polylines-arcpy/). It is worth noting that there are differences between the generated viewpoints and the actual GSV viewpoints in terms of latitude and longitude values because the GSV vehicle does not follow the exactly same routes as the roads in the GIS dataset. A pre-processing solution was used to compromise this mismatching issue (see Section 2.2 Google Street View imagery). Reference UPCs (i.e., the ground truth data) for the whole town were manually mapped using 7.5 cm aerial imagery and checked using GSV. These validation data included 1039 poles, which were located within a 20 m buffer zone around selected roads.

2.2. Google Street View Imagery

GSV images can detect the orientation of detected UPCs in 360° GSV, similar as the bearing-only sensor, which is commonly used to measure the direction toward features from a robot. Static street view images were downloaded through the GSV image application programming interface (API) by providing uniform resource locators (URL) that embed the appropriate parameter information [50]. The GSV API automatically snaps the requested coordinates to the nearest available GSV viewpoint [50]. In order to acquire accurate latitude and longitude values for each viewpoint of GSV, the “streetview” Python package (https://github.com/robolyst/streetview) was used to convert the requested coordinates into the nearest available GSV coordinates via a list of panorama IDs (i.e., unique identification for each GSV panorama view with acquired date [year, month], latitude, and longitude). The most recent panorama ID was then used as location parameter input. Other required parameters for the URL include size (output size of GSV image), heading (cardinal direction in the range of 360°), fov (horizontal field-of-view angle), pitch (the up or down angle), and API key [50]. Four GSV images were obtained for each view point with a fov of 90° and headings of 0°, 90°, 180°, and 270°, respectively (Figure 2). We developed a Python script to automatically create the URLs and download the 37,160 GSV images needed to cover the study area. The GSV images were acquired between 2011 and 2016 (Figure 3); this range is due to the varying frequencies with which Google updates street view imagery. We cropped the downloaded images to eliminate the Google logos.

2.3. Annotation Data

A large amount of ground-truth data is essential for the deep supervised learning algorithms to be effective [51]. We created 3500 ground-truth data points by manually labelling UPCs in GSV images that were acquired by taking screen captures in Google Maps (Figure 4). In order to enhance the transferability of the method, each training/validation/test image was taken at a random utility pole location in the states of Connecticut, Massachusetts, Maine, New Hampshire, New York, and Texas. The “LabelImg” software (https://github.com/tzutalin/labelImg) was used to annotate the ground-truth images and produce outputs in the format of XML files. This format was used in the Pattern Analysis, Statistical Modelling and Computational Learning project (i.e., PASCAL VOC [http://host.robots.ox.ac.uk/pascal/VOC/]) and the ImageNet database (a large image database for visual object recognition research [http://www.image-net.org]). In general, a full dataset for DL includes a training dataset to “teach” the DL algorithm, a validation dataset for minimizing overfitting of the training data, and a test dataset for assessing the performance. Overall, 2500 annotated GSV images were used as training data, 500 images were used as validation data, and the remaining 500 images were used for accuracy assessment.

3. Methodology

3.1. General Procedure

Our proposed DL-based automatic mapping method for UPCs from GSV included three main steps (see Figure 5): The first step is to detect UPCs in the GSV images using a trained DL network. The second step is to calculate the azimuth from each viewpoint to the detected UPCs based on the known azimuth angles of the GSV images, relative to their view point locations, and the horizontal positions of the target in the images (Figure 5(2)) using the mean value of two X values of the bounding box. For example, suppose a detected UPC has a bounding box that is centered on column 358 in a GSV image that is centered at 0° azimuth relative to the image viewpoint. Each GSV image contains 640 columns and spans a 90° horizontal field-of-view; thus, each pixel spans 0.14°. The center of the UPC is 38 pixels to the right of the image center (at column 320) and so has an azimuth of 5.3° relative to the image viewpoint. The final step is to estimate the target locations based on the azimuths calculated from the second step (Figure 5(3)).

3.2. Deep Learning Algorithm

In this study, the RetinaNet object detector, as described in Focal Loss for Dense Object Detection [49], was used to detect utility poles because of its excellent performance from the prospective of accuracy and computational efficiency [49]. Compared to state-of-the-art two-stage detectors, such as faster R-CNN [52] and Mask R-CNN [53], the RetinaNet object detector is a recently proposed one-stage detector with simpler structure and faster speed, and it can achieve a better accuracy than two-stage detectors. The RetinaNet object detector is based on the one-stage RetinaNet network architecture, which was built on top of the feedforward residual learning network (ResNet) architecture with a Feature Pyramid Network (FPN) [45,49,54]. In terms of structure, a RetinaNet object detector consists of five main components: (1) ResNet as feedforward architecture; (2) FPN as backbone for producing convolutional feature pyramid; (3) region proposal networks (RPN) for generating proposal; (4) fully convolutional network (FCN) as classification subnetwork; and (5) FCN as box regression subnetwork. It is worth noting that the ResNet used in the RetinaNet network was pre-trained on ImageNet (i.e., the transfer learning). For this study, we detected UPCs in GSV images using a Keras (a high-level neural networks API) implementation of a RetinaNet object detector developed by the development team of keras-retinanet on Github (https://github.com/fizyr/keras-retinanet).

3.3. Utility Poles Position Inference

Some efforts have been spent in localizing street-level objects using multiple street-view pictures, such as manhole covers, traffic signs. For example, Timofte and Gool [55] developed pipeline to detect and localize manhole covers by using 4 pairs of downward looking stereo cameras and conducting two-steps localization (rough detection from single-view picture and accurate three dimensional (3D) localization from multi-view pictures). Soheilian et al. [56] detected and reconstructed 3D traffic signs and achieved average position accuracy of 3.5 cm by intersecting the corresponding rays from multi-view images and clustering traffic signs candidates. Hebbalaguppe et al. [57] proposed an automated updating system for telecom inventory using object detection and triangulation-based method (stereo-vision distance estimation with the SIFT feature matching algorithm) with GSV. Very recently, it was noticed that Krylov et al. [58] applied a CNN-based sematic segmentation model with a proposed geotagging method to estimate geographic objects’ locations from GSV with two classes—traffic lights and telegraph poles. In this study, we used the RetinaNet object detector and a modified brute-force-based line-of-bearing (LOB) measurement method to localize the position of UPCs from multiple-view GSV images. However, the specific methods we used are different from those used in Krylov et al. [58].

3.3.1. LOB Measurement

The outputs of UPCs detection in GSV images using DL are bounding boxes of detected UPCs, which result from implementation of odometry from monocular vision of GSV images as shown in Figure 6. Therefore, estimating locations of UPCs in pure GSV images is a multiple-source localization problem from passive angle measurements, which has been widely investigated [59,60]. The LOB-based approach is one of three main multiple-source localization approaches [61]. An LOB measurement was applied to estimate the location of a target (i.e., UPC) because detected UPCs are not signal sources such as propagating signal sources whose signal strength can be measured (Figure 6). An LOB measurement does not require as many strict conditions as other methods (e.g., synchronization and power transit) do.
In LOB localization, azimuths from viewpoints of multiple images to a given UPC allow the UPC location to be triangulated (see Figure 6). Ideally, intersection of multiple LOBs is the exact location of the target because the LOBs pass through the target (see Figure 6). However, numerous ghost nodes (i.e., false targets) occur when the LOB measurement is used in a dense emitter environment, as shown in our study for estimating locations of UPCs in GSV images [62] (see Figure 7). As a result, a modified brute-force-based three-station cross location algorithm was utilized to minimize the ghost node problem of multiple-source localization using LOB measurement (Figure 4: source localization from viewpoints A, B, and C), based on two assumptions that targets and sensors are on the xy plane, and all LOB measurements have the same precision [63]. More specifically, the LOB measurement method uses the following steps (Figure 7): (1) for a given viewpoint, find the closest neighboring viewpoints; we tested the performance of the algorithm using 2 to 8 of the closest neighboring viewpoints (i.e., the corresponding number of views is 3 to 9); (2) measure the angles between each pair of LOBs from all viewpoints [64]; (3) check if there are positive associations among LOBs (set as 50 m length) from current viewpoint and its nearest viewpoints [63]; (4) repeat the process from step (1) to step (3) for every intersection point. To be more specific, three positive detections from any three views within an angle threshold (β) produce a positive association among LOBs [63]. Therefore, theoretically, given that detection rates are constant, the number of estimated UPCs increases as the number of views increases based on the probability of combination. For example, assuming the number of all possibilities of UPC estimation is t ( t N ) and the detection rate is constant; then the probability of positive association with 7 views (i.e., C ( 7 ,   3 ) t ) is greater than the probability of positive association with 4 views (i.e., C ( 4 ,   3 ) t ).
In this study, nearest viewpoints were selected in order to conduct cross validation. A list of the closest neighboring viewpoints (2, 3, 4, 5, 6, 7, and 8 viewpoints; that is, 3, 4, 5, 6, 7, 8, and 9 views after including the view from the current viewpoint) and angle thresholds (1°, 2°, and 3°) were used for testing to decide if there is a positive association and which threshold performs the best. No more than 9 views were selected for testing mainly because of the length of LOB and the interval of GSV acquisition (10 m). The extreme scenario of 9 views is that 8 viewpoints are on a line and located on one side of the current viewpoint. 80 m is almost the maximum distance requirement for the intersection of two 50-m LOBs. A list of distances (3 m, 4 m, and 5 m) were applied to eliminate ghost nodes that are located too close to the center line of a road. These thresholds allow more relaxation of position inference rule in the LOB measurement. In the process of UPCs detection in GSV, almost parallel rays from camera to long-distance objects can intersect many other rays. Therefore, any object with a width less than 30 pixels was excluded in order to reduce the computation workload of eliminating ghost nodes because the used position inference method is an iterative method.

3.3.2. Multiple LOB Intersection Points Aggregation

LOB measurements with a modified brute-force-based three-station cross location algorithm produce multiple LOB intersection points as potential candidates for each utility pole. In order to estimate the most likely location of a utility pole, a geospatial aggregation algorithm with an aggregation distance of 10 m was used to estimate the centroid of clusters of LOB intersection points (see Figure 8). The geospatial aggregation algorithm consists of three main steps: (1) calculate Euclidean distance matrix of all LOB intersection points; (2) cluster LOB intersection points based on the Euclidean distances between LOB intersection points; and (3) calculate the centroid of each cluster of intersection points (Figure 8).

4. Experiments and Results

4.1. Experiments

For testing the performance of using deep learning to estimate the locations of UPCs in GSV images, we conducted experiments on a customized server, which is equipped with an Intel i5 CPU, 16 GB RAM, a GeForce GTX 970 graphic card, and a GeForce GTX 1080ti graphic card. To obtain an optimal parameter setting, we trained and validated the RetinaNet object detector with 50-layer, 101-layer, and 152-layer ResNets (denoted as RetinaNet-50, RetinaNet-101, and RetinaNet-152, respectively) using the same labeled samples. For each of the three trained detectors (or detector-training choices), a step of 2500, a batch size of 1, and an epoch number of 200 were selected for the learning process. In the training step of DL, the step size was decided based on the sizes of batch and training dataset. During training, random horizontal flips augmentation along the X direction (with a chance of 0.5) was used to introduce variation in the training data. In order to avoid overfitting, a validation dataset was used to evaluate the accuracy of the RetinaNet object detector with three training choices at the end of each epoch. The strategies for reducing overfitting were suspended during the validation process. Figure 9 shows that the average precision (AP) reached the peak value around the 25th epoch and the model tended to be convergent. Thus, the RetinaNet object detectors were trained for 25 epochs before applying to utility pole detection from GSV images for our study area (see Figure 9).

4.2. Results and Discussions

4.2.1. Accuracy Assessment of Object Detection

Figure 10 shows the accuracy of the RetinaNet object detector for detecting UPCs in GSV with 500 annotated GSV scenes including 937 reference objects. The count of false negative detection is zero in the confusion matrixes because non-utility pole objects were excluded from the training dataset. Overall, the RetinaNet-101 has the best overall accuracy (OA) among the three trained detectors. OAs of the RetinaNet-101 are around 78%, 72%, and 50% when the intersection-over-union (IoU) thresholds are greater than 0.3, 0.4, and 0.5, respectively. Additionally, compared with the other two trained detectors (RetinaNet-50 and RetinaNet-152), the RetinaNet-101 also has the highest recall and precision (Figure 10). The precisions of RetinaNet-101 are 0.95, 0.91, and 0.73 and its recalls are 0.81, 0.77, and 0.62, respectively, when IoUs thresholds are greater than 0.3, 0.4, and 0.5. It is worth mentioning that the OAs corresponding to different IoUs are presented because high IoUs may cause underestimation of the performance of the detection for UPCs in GSV in certain scenarios (Figure 11). For example, Figure 11 shows that the detector RetinaNet-101 was able to detect the target but still had an IoU less than the three selected IoU thresholds, which means the detected target was not considered as a positive detection. Based on the accuracy comparison, the RetinaNet-101 after training 25 epochs was finally chosen for detecting UPCs.

4.2.2. Accuracy Assessment of Location Estimation

Table 1 shows the accuracy values of UPC location estimation, measured as the percentage of the number of estimated locations of UPCs located within the buffer zones of reference utility poles. To evaluate the impacts of the number of views, the threshold of angle, and the threshold of distance to the center of a selected road, we considered seven views (i.e., 3, 4, 5, 6, 7, 8, 9), three thresholds of angle (i.e., 1°, 2°, and 3°), and three thresholds of distance to the center of the selected road (i.e., 3 m, 4 m, and 5 m). For the method we tested, around half of the estimated UPC locations were within the 6 m buffer zone of their reference locations, and up to 79% of the estimated locations were within the 10 m buffer zone of reference locations. However, around 12% of the estimated UPC locations were within the 2 m buffer zone of reference locations which suggests that the LOB approach does not provide consistently accurate UPC locations. In terms of the threshold of angle in the modified brute-force-based three-station cross-location algorithm, Table 1 shows that using more views and larger angle thresholds resulted in more estimated UPCs, which is attributed to the increase of relaxation of the modified brute-force-based three-station cross location algorithm. In the meantime, more estimated UPCs could also result in lower accuracy in UPC estimation because relaxation allows more ghost nodes to be estimated UPCs (see Table 1). From Table 1, one can see that the average percentage of the number of estimated locations of UPCs being within all buffer zones of reference utility poles for the results of 8 views is the highest (47.80%) compared to the results of other numbers of views. In contrast, using larger thresholds of distance to center of selected road resulted in less estimated UPCs. In general, the accuracy of the location estimation for UPCs is reasonable and the estimated data are valuable because the optical GSV imagery was the only data source used to conduct the localization.
Figure 12 shows a visual comparison of the distributions of estimated locations of UPCs with different numbers of views (i.e., 3, 4, 5, 6, 7, 8 and 9 views), an angle threshold of 2°, and a distance threshold of 4 m. Overall, the distributions of estimated UPC locations (Figure 12b–h) are almost the same as the distributions of the reference UPCs except for some missing UPCs (see Figure 13b–e). In particular, Figure 13b–e show that most estimated locations of UPCs are geographically close to the reference locations of UPCs. Our method failed to estimate UPC locations when a given UPC was not detected in at least three GSV images out of a certain number of views—three is the minimum number of images required to triangulate a position and eliminate ghost nodes (see Figure 7). This explains why the number of estimated UPCs increases as the number of views increases (see Table 1 and Figure 12b–h). Location mismatches that occurred (see Figure 13b–e) may have been caused by GSV image distortion, terrain relief, the position accuracy of GSV, UPC lean, or by limitations in the method we used. For example, ground locations of UPCs are different from the orthographic projected locations estimated from GSV images (see Figure 14) due to leaning UPCs. Our proposed method has great potential for the areas where GSV imagery is available and where a UPC distribution map with a ≤10 m accuracy is acceptable.

4.2.3. Parameter Sensitivity Analysis of Location Estimation

To quantify the degree to which the location inference method affects the location estimation, we conduct Sobol’s sensitivity analysis with a sampling size of 500 on the three main parameters for estimating locations of UPCs (i.e., number of views, angle threshold, and buffer distance). Parameter sensitivity for the location estimation method (as a function) is analyzed through conducting experiments using a variety of combinations of parameters (Table 1). During the sampling process in Sobol’s sensitivity analysis, we simplified the testing method (i.e., location inference method) by grouping results into categories because of the time intensive execution of the testing method with hundreds of sampling scenarios. Two sets of sensitivity analysis were conducted from the aspect of parameters versus the number of estimated UPCs (Table 2) and parameters versus the percentage of the number of estimated locations of UPCs being within a 5 m buffer zone of reference UPCs (Table 3).
Table 2 exhibits that, after measuring sensitivity across the three parameters, the number of views contributes the most to the number of estimated UPCs, followed by angle threshold. More UPCs can be estimated as the number of views increases (Table 1). The threshold of distance shows very limited impact on the number of estimated UPCs. For example, in the case of the number of views being 8, the corresponding average numbers of estimated UPCs for angle thresholds 1°, 2°, and 3° are 636, 773, and 842, respectively; in contrast, the corresponding average numbers of estimated UPCs for distance thresholds 3 m, 4 m, and 5 m are 763, 763, and 726, respectively. However, Sobol’s sensitivity analysis on the percentage of the number of estimated UPC locations being within a 5 m buffer zone of reference UPCs shows that the effect of distance threshold is much stronger, increasing from 0.0157 to 0.1826 in terms of total order sensitivity (Table 3). This indicates that increasing the threshold of distance can reduce the error of location estimation by eliminating ghost nodes generated from almost parallel LOBs. Increasing the other two parameters has a similar effect because increasing the number of views and threshold of angle allows more LOB intersections to suffice conditions of being candidates.

4.2.4. Limitations and Future Studies

This study is mainly an exploratory study, with only one method being used for each task of detection and location inference. Even though this study presents a great potential of using DL to map UPCs from GSV images, there are some limitations worth a mention. First, GSV is increasing the coverage and accelerating the update frequency, but it is still challenging to map up-to-date geographic information from GSV images. For example, the UPC locations may have changed since the time that the available GSV images were acquired, especially if the images are more than a few years old. Second, a large amount of training dataset is needed to achieve an acceptable accuracy with the Retina-101. Therefore, a comprehensive study about the minimum requirement of the training dataset for using DL to map UPCs or other geographic objects is needed. Third, we only validated the DL model on detecting utility poles with crossarms; however, a large percentage of utility poles do not have crossarms or other distinctive features. Fourth, accuracy of the LOB method is reduced when UPCs have a significant lean mainly because of terrain and optical distortion of GSV without compensation; better positions could have been obtained by detecting the base of the UPCs, but this would likely have resulted in false positives due to trees (Figure 14). Fifth, this method considers UPCs that are geographically very close to each other as one UPC because of the fundamental flaw of LOB measurement.

5. Conclusions

Mapping UPCs is a labor-intensive and time-consuming task. This study suggested a DL-based automatic mapping method to map UPCs from GSV, which uses a DL algorithm (the RetinaNet object detector) and the LOB measurement to estimate the locations of UPCs in GSV automatically. A case study was conducted to demonstrate the performance of the suggested method. The results show that (1) with properly-set parameters, the RetinaNet object detector is able to detect most UPCs from GSV (e.g., an OA of 0.78 can be achieved by RetinaNet-101 when IoU threshold is greater than 0.3); and (2) by combining LOB measurement and multiple points aggregation, the UPC position inference method can estimate the positions of UPCs at a reasonable accuracy (i.e., ≤10 m). In general, the suggested integrative method shows to be promising in our case study. With the wide availability of GSV images, the method might be a valuable way for automatically mapping UPCs and could be useful for mapping other geographic objects located along roads.

Author Contributions

W.Z. designed and conceived the study, and collected the data and performed the experiments. C.W. aided in designing and conceiving the study and revising the manuscript. W.L. and C.Z. contributed to the manuscript writing. X.L. and J.P. aided in designing the experiments and revising the manuscript.

Funding

This work was supported in part by the U.S. National Science Foundation [grant No. 1414108] and conducted during 2017 to 2018. Collection of the ground-truth utility pole locations was funded by Eversource Energy Connecticut.

Acknowledgments

The authors gratefully acknowledge the data and support of Eversource Energy Connecticut and the Eversource Energy Center at the University of Connecticut. The authors would like to thank Krista Rogers for her helpful reviews of this manuscript and thank Shahearn Philemon for visually interpreting reference utility poles from high-resolution aerial images. The authors also would like to thank the development team of keras-retinanet on Github, especially Hans Gaiser and Maarten de Vries.

Conflicts of Interest

The authors declare no conflict of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, and in the decision to publish the results.

References

  1. Nagura, S.; Masumoto, T.; Endo, K.; Wakasa, F.; Watanabe, S.; Ikeda, K. Development of mapping system for distribution facility management. In Electricity Distribution, Proceedings of 10th International Conference on Electricity Distribution, CIRED 1989, Brighton, UK, 8–12 May 1989; IET: London, UK, 1989. [Google Scholar]
  2. Nguyen, V.N.; Jenssen, R.; Roverso, D. Automatic autonomous vision-based power line inspection: A review of current status and the potential role of deep learning. Int. J. Electr. Power Energy Syst. 2018, 99, 107–120. [Google Scholar] [CrossRef]
  3. CITYLAB. Available online: https://www.citylab.com/environment/2017/10/how-open-source-mapping-helps-hurricane-recovery/542565/ (accessed on 1 February 2018).
  4. Cetin, B.; Bikdash, M.; McInerney, M. Automated electric utility pole detection from aerial images. In Proceedings of the IEEE Southeastcon 2009, Atlanta, GA, USA, 5–8 March 2009. [Google Scholar] [CrossRef]
  5. Bernstein, R.; Di Gesù, V.A. combined analysis to extract objects in remote sensing images. Pattern Recognit. Lett. 1999, 20, 1407–1414. [Google Scholar] [CrossRef]
  6. Golightly, I.; Jones, D. Corner detection and matching for visual tracking during power line inspection. Image Vis. Comput. 2003, 21, 827–840. [Google Scholar] [CrossRef]
  7. Jones, D.I.; Whitworth, C.C.; Earp, G.K.; Duller, A.W.G. A laboratory test-bed for an automated power line inspection system. Control Eng. Pract. 2005, 13, 835–851. [Google Scholar] [CrossRef]
  8. Khawaja, A.H.; Huang, Q.; Khan, Z.H. Monitoring of Overhead Transmission Lines: A Review from the Perspective of Contactless Technologies. Sens. Imaging 2017, 18, 24. [Google Scholar] [CrossRef]
  9. Li, W.H.; Tajbakhsh, A.; Rathbone, C.; Vashishtha, Y. Image processing to automate condition assessment of overhead line components. In Proceedings of the 2010 1st International Conference on Applied Robotics for the Power Industry, Montreal, QC, Canada, 5–7 October 2010. [Google Scholar] [CrossRef]
  10. Tong, W.G.; Li, B.S.; Yuan, J.S.; Zhao, S.T. Transmission line extraction and recognition from natural complex background. In Proceedings of the 2009 International Conference on Machine Learning and Cybernetics, Baoding, China, 12–15 July 2009. [Google Scholar] [CrossRef]
  11. Whitworth, C.C.; Duller, A.W.G.; Jones, D.I.; Earp, G.K. Aerial video inspection of overhead power lines. Power Eng. J. 2001, 15, 25–32. [Google Scholar] [CrossRef]
  12. Yan, G.; Li, C.; Zhou, G.; Zhang, W.; Li, X. Automatic extraction of power lines from aerial images. IEEE Geosci. Remote Sens. Lett. 2007, 4, 387–391. [Google Scholar] [CrossRef]
  13. Sarabandi, K.; Pierce, L.; Oh, Y.; Ulaby, F.T. Power lines: Radar measurements and detection algorithm for polarimetric SAR images. IEEE Trans. Aerosp. Electron. Syst. 1994, 30, 632–643. [Google Scholar] [CrossRef]
  14. Sarabandi, K.; Park, M. Extraction of power line maps from millimeter-wave polarimetric SAR images. IEEE Trans. Antennas Propag. 2000, 48, 1802–1809. [Google Scholar] [CrossRef] [Green Version]
  15. Jwa, Y.; Sohn, G. A piecewise catenary curve model growing for 3D power line reconstruction. Photogramm. Eng. Remote Sens. 2012, 78, 1227–1240. [Google Scholar] [CrossRef]
  16. Kim, E.; Medioni, G. Urban scene understanding from aerial and ground LIDAR data. Mach. Vis. Appl. 2011, 22, 691–703. [Google Scholar] [CrossRef]
  17. Kim, H.B.; Sohn, G. Point-based classification of power line corridor scene using random forests. Photogramm. Eng. Remote Sens. 2013, 79, 821–833. [Google Scholar] [CrossRef]
  18. McLaughlin, R.A. Extracting transmission lines from airborne LIDAR data. IEEE Geosci. Remote Sens. Lett. 2006, 3, 222–226. [Google Scholar] [CrossRef]
  19. Wang, Y.; Chen, Q.; Liu, L.; Zheng, D.; Li, C.; Li, K. Supervised Classification of Power Lines from Airborne LiDAR Data in Urban Areas. Remote Sens. 2017, 9, 771. [Google Scholar] [CrossRef]
  20. Sun, C.; Jones, R.; Talbot, H.; Wu, X.; Cheong, K.; Beare, R.; Buckley, M.; Berman, M. Measuring the distance of vegetation from powerlines using stereo vision. ISPRS J. Photogramm. Remote Sens. 2006, 60, 269–283. [Google Scholar] [CrossRef]
  21. Matikainen, L.; Lehtomäki, M.; Ahokas, E.; Hyyppä, J.; Karjalainen, M.; Jaakkola, A.; Kukko, A.; Heinonen, T. Remote sensing methods for power line corridor surveys. ISPRS J. Photogramm. Remote Sens. 2016, 119, 10–31. [Google Scholar] [CrossRef]
  22. Moore, A.J.; Schubert, M.; Rymer, N. Autonomous Inspection of Electrical Transmission Structures with Airborne UV Sensors-NASA Report on Dominion Virginia Power Flights of November 2016. NASA/TM-2017-219611, L-20808, NF1676L-26882, Nasa Technical Reports Server; Document ID: 20170004692; 1 May 2017. Available online: https://ntrs.nasa.gov/archive/nasa/casi.ntrs.nasa.gov/20170004692.pdf (accessed on 12 February 2017).
  23. Oh, J.; Lee, C. 3D power line extraction from multiple aerial images. Sensors 2017, 17, 2244. [Google Scholar] [CrossRef] [PubMed]
  24. Zhang, Y.; Yuan, X.; Li, W.; Chen, S. Automatic Power Line Inspection Using UAV Images. Remote Sens. 2017, 9, 824. [Google Scholar] [CrossRef]
  25. Zhu, L.; Hyyppä, J. Fully-automated power line extraction from airborne laser scanning point clouds in forest areas. Remote Sens. 2014, 6, 11267–11282. [Google Scholar] [CrossRef]
  26. Cabo, C.; Ordoñez, C.; García-Cortés, S.; Martínez, J. An algorithm for automatic detection of pole-like street furniture objects from Mobile Laser Scanner point clouds. ISPRS J. Photogramm. Remote Sens. 2014, 87, 47–56. [Google Scholar] [CrossRef]
  27. Lehtomäki, M.; Jaakkola, A.; Hyyppä, J.; Kukko, A.; Kaartinen, H. Detection of vertical pole-like objects in a road environment using vehicle-based laser scanning data. Remote Sens. 2010, 2, 641–664. [Google Scholar] [CrossRef]
  28. Cheng, L.; Tong, L.; Wang, Y.; Li, M. Extraction of urban power lines from vehicle-borne LiDAR data. Remote Sens. 2014, 6, 3302–3320. [Google Scholar] [CrossRef]
  29. Guan, H.; Yu, Y.; Li, J.; Ji, Z.; Zhang, Q. Extraction of power-transmission lines from vehicle-borne lidar data. Int. J. Remote Sens. 2016, 37, 229–247. [Google Scholar] [CrossRef]
  30. Sharma, H.; Adithya, V.; Dutta, T.; Balamuralidhar, P. Image Analysis-Based Automatic Utility Pole Detection for Remote Surveillance. In Proceedings of the 2015 International Conference on Digital Image Computing: Techniques and Applications (DICTA), Adelaide, Australia, 23–25 November 2015. [Google Scholar] [CrossRef]
  31. Anguelov, D.; Dulong, C.; Filip, D.; Frueh, C.; Lafon, S.; Lyon, R.; Ogale, A.; Vincent, L.; Weaver, J. Google street view: Capturing the world at street level. Computer 2010, 43, 32–38. [Google Scholar] [CrossRef]
  32. Li, X.; Zhang, C.; Li, W.; Ricard, R.; Meng, Q.; Zhang, W. Assessing street-level urban greenery using Google Street View and a modified green view index. Urban For. Urban Green. 2015, 14, 675–685. [Google Scholar] [CrossRef]
  33. Li, X.; Zhang, C.; Li, W.; Kuzovkina, Y.A.; Weiner, D. Who lives in greener neighborhoods? The distribution of street greenery and its association with residents’ socioeconomic conditions in Hartford, Connecticut, USA. Urban For. Urban Green. 2015, 14, 751–759. [Google Scholar] [CrossRef]
  34. Li, X.; Zhang, C.; Li, W. Building block level urban land-use information retrieval based on Google Street View images. GIsci. Remote Sens. 2017, 54, 819–835. [Google Scholar] [CrossRef]
  35. Zhang, W.; Li, W.; Zhang, C.; Hanink, D.M.; Li, X.; Wang, W. Parcel-based urban land use classification in megacity using airborne LiDAR, high resolution orthoimagery, and Google Street View. Comput. Environ. Urban Syst. 2017, 64, 215–228. [Google Scholar] [CrossRef]
  36. Li, X.; Ratti, C.; Seiferling, I. Quantifying the shade provision of street trees in urban landscape: A case study in Boston, USA, using Google Street View. Landsc. Urban Plan. 2018, 169, 81–91. [Google Scholar] [CrossRef]
  37. Cheng, W.; Song, Z. Power pole detection based on graph cut. In Proceedings of the 2008 Congress on Image and Signal Processing, Sanya, China, 27–30 May 2008. [Google Scholar] [CrossRef]
  38. Murthy, V.S.; Gupta, S.; Mohanta, D.K. Digital image processing approach using combined wavelet hidden Markov model for well-being analysis of insulators. IET Image Process. 2011, 5, 171–183. [Google Scholar] [CrossRef]
  39. Barranco-Gutiérrez, A.I.; Martínez-Díaz, S.; Gómez-Torres, J.L. An Approach for Utility Pole Recognition in Real Conditions. In Proceedings of the Image and Video Technology—PSIVT 2013 Workshops, Guanajuato, Mexico, 28–29 October 2013; Huang, F., Sugimoto, A., Eds.; Springer: Berlin/Heidelberg, Germany, 2013. [Google Scholar] [CrossRef]
  40. Song, B.; Li, X. Power line detection from optical images. Neurocomputing 2014, 129, 350–361. [Google Scholar] [CrossRef]
  41. Mills, S.J.; Castro, M.P.G.; Li, Z.; Cai, J.; Hayward, R.; Mejias, L.; Walker, R.A. Evaluation of aerial remote sensing techniques for vegetation management in power-line corridors. IEEE Trans. Geosci. Remote Sens. 2010, 48, 3379–3390. [Google Scholar] [CrossRef] [Green Version]
  42. Buduma, N.; Locascio, N. Fundamentals of Deep Learning: Designing Next-Generation Machine Intelligence Algorithms, 1st ed.; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2017; ISBN 1491925612. [Google Scholar]
  43. Géron, A. Hands-On Machine Learning with Scikit-Learn and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems, 1st ed.; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2017; ISBN 1491962291. [Google Scholar]
  44. Goodfellow, I.; Bengio, Y.; Courville, A.; Bengio, Y. Deep Learning; MIT Press: Cambridge, MA, USA, 2016; ISBN 0262035618. [Google Scholar]
  45. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar] [CrossRef]
  46. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  47. Patterson, J.; Gibson, A. Deep Learning: A Practitioner’s Approach, 1st ed.; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2017; ISBN 1491914254. [Google Scholar]
  48. Nordeng, I.E.; Hasan, A.; Olsen, D.; Neubert, J. DEBC Detection with Deep Learning. In Proceedings of the 20th Scandinavian Conference on Image Analysis, Tromsø, Norway, 12–14 June 2017; Puneet, S., Filippo Maria, B., Eds.; Springer: Cham, Switzerland, 2017. [Google Scholar] [CrossRef]
  49. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. arXiv, 2017; arXiv:1708.02002v2. [Google Scholar]
  50. Google. Available online: https://developers.google.com/maps/documentation/streetview/intro (accessed on 15 February 2018).
  51. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Advances in Neural Information Processing Systems 28; Cortes, C., Lawrence, N.D., Lee, D.D., Sugiyama, M., Garnett, R., Eds.; Neural Information Processing Systems Foundation, Inc.: Montréal, QC, Canada, 2015. [Google Scholar]
  53. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar] [CrossRef]
  54. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar] [CrossRef]
  55. Timofte, R.; Van Gool, L. Multi-view manhole detection, recognition, and 3D localisation. In Proceedings of the 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), Barcelona, Spain, 6–13 November 2011. [Google Scholar] [CrossRef]
  56. Soheilian, B.; Paparoditis, N.; Vallet, B. Detection and 3D reconstruction of traffic signs from multiple view color images. ISPRS J. Photogramm. Remote Sens. 2013, 77, 1–20. [Google Scholar] [CrossRef]
  57. Hebbalaguppe, R.; Garg, G.; Hassan, E.; Ghosh, H.; Verma, A. Telecom Inventory management via object recognition and localisation on Google Street View Images. In Proceedings of the 2017 IEEE Winter Conference on Applications of Computer Vision (WACV) (2017), Santa Rosa, CA, USA, 24–31 March 2017. [Google Scholar] [CrossRef]
  58. Krylov, V.A.; Kenny, E.; Dahyot, R. Automatic Discovery and Geotagging of Objects from Street View Imagery. Remote Sens. 2018, 10, 661. [Google Scholar] [CrossRef]
  59. Gavish, M.; Weiss, A.J. Performance analysis of bearing-only target location algorithms. IEEE Trans. Aerosp. Electron. Syst. 1992, 28, 817–828. [Google Scholar] [CrossRef]
  60. Zhang, H.; Jing, Z.; Hu, S. Localization of Multiple Emitters Based on the Sequential PHD Filter. Signal Process. 2010, 90, 34–43. [Google Scholar] [CrossRef]
  61. Reed, J.D.; da Silva, C.R.; Buehrer, R.M. Multiple-source localization using line-of-bearing measurements: Approaches to the data association problem. In Proceedings of the MILCOM 2008—2008 IEEE Military Communications Conference, San Diego, CA, USA, 16–19 November 2008. [Google Scholar] [CrossRef]
  62. Grabbe, M.T.; Hamschin, B.M.; Douglas, A.P. A measurement correlation algorithm for line-of-bearing geo-location. In Proceedings of the 2013 IEEE Aerospace Conference, Big Sky, MT, USA, 2–9 March 2013. [Google Scholar] [CrossRef]
  63. Tan, K.; Chen, H.; Cai, X. Research into the algorithm of false points elimination in three-station cross location. Shipboard Electron. Countermeas. 2009, 32, 79–81. [Google Scholar] [CrossRef]
  64. Reed, J. Approaches to Multiple-Source Localization and Signal Classification. Ph.D. Thesis, Virginia Polytechnic Institute and State University, Blacksburg, VA, USA, 5 May 2009. [Google Scholar]
Figure 1. (a) Location of the study area in the Northeastern United States; (b) location of the study area in the state of Connecticut.
Figure 1. (a) Location of the study area in the Northeastern United States; (b) location of the study area in the state of Connecticut.
Sensors 18 02484 g001
Figure 2. Two sets of GSV images, four per set, were obtained for a view point panorama ID (09RrbsPUqQaXgJjg25ey6w) and a viewpoint point panorama ID (BnGQHO076vHGvgRz37XVCQ) and an example of URL with the parameters highlighted in red color.
Figure 2. Two sets of GSV images, four per set, were obtained for a view point panorama ID (09RrbsPUqQaXgJjg25ey6w) and a viewpoint point panorama ID (BnGQHO076vHGvgRz37XVCQ) and an example of URL with the parameters highlighted in red color.
Sensors 18 02484 g002
Figure 3. Distribution of GSV images acquisition dates, between 2011 and 2016.
Figure 3. Distribution of GSV images acquisition dates, between 2011 and 2016.
Sensors 18 02484 g003
Figure 4. Bounding boxes (red boxes) of labelled utility poles in four GSV images (ad).
Figure 4. Bounding boxes (red boxes) of labelled utility poles in four GSV images (ad).
Sensors 18 02484 g004
Figure 5. The workflow of mapping utility poles from GSV with deep learning.
Figure 5. The workflow of mapping utility poles from GSV with deep learning.
Sensors 18 02484 g005
Figure 6. An example of using the bearing measurement to estimate a target location from a sensor from three locations.
Figure 6. An example of using the bearing measurement to estimate a target location from a sensor from three locations.
Sensors 18 02484 g006
Figure 7. An example of using the brute-force-based three-station cross location algorithm to eliminate ghost nodes from four views with an angle threshold of 5°.
Figure 7. An example of using the brute-force-based three-station cross location algorithm to eliminate ghost nodes from four views with an angle threshold of 5°.
Sensors 18 02484 g007
Figure 8. An example of aggregating multiple LOB intersection points.
Figure 8. An example of aggregating multiple LOB intersection points.
Sensors 18 02484 g008
Figure 9. Validating the RetinaNet object detector learning with 50-layer, 101-layer, and 152-layer ResNets, respectively, for optimizing the RetinaNet object detector.
Figure 9. Validating the RetinaNet object detector learning with 50-layer, 101-layer, and 152-layer ResNets, respectively, for optimizing the RetinaNet object detector.
Sensors 18 02484 g009
Figure 10. Confusion matrices for the RetinaNet object detector learning with 50-layer, 101-layer, and 152-layer ResNets after training 25 epochs on IoU with values greater than 0.3, 0.4, and 0.5 for choosing optimal parameters. (a) RetinaNet-50 with IoU > 0.3; (b) RetinaNet-50 with IoU > 0.4; (c) RetinaNet-50 with IoU > 0.5; (d) RetinaNet-101 with IoU > 0.3; (e) RetinaNet-101 with IoU > 0.4; (f) RetinaNet-101 with IoU > 0.5; (g) RetinaNet-152 with IoU > 0.3; (h) RetinaNet-152 with IoU > 0.4; (i) RetinaNet-152 with IoU > 0.5.
Figure 10. Confusion matrices for the RetinaNet object detector learning with 50-layer, 101-layer, and 152-layer ResNets after training 25 epochs on IoU with values greater than 0.3, 0.4, and 0.5 for choosing optimal parameters. (a) RetinaNet-50 with IoU > 0.3; (b) RetinaNet-50 with IoU > 0.4; (c) RetinaNet-50 with IoU > 0.5; (d) RetinaNet-101 with IoU > 0.3; (e) RetinaNet-101 with IoU > 0.4; (f) RetinaNet-101 with IoU > 0.5; (g) RetinaNet-152 with IoU > 0.3; (h) RetinaNet-152 with IoU > 0.4; (i) RetinaNet-152 with IoU > 0.5.
Sensors 18 02484 g010
Figure 11. Examples of accuracy assessment of utility poles detection with underestimated performance at different measured IoUs. (ad) are four different examples of true positive detections with different IoU values.
Figure 11. Examples of accuracy assessment of utility poles detection with underestimated performance at different measured IoUs. (ad) are four different examples of true positive detections with different IoU values.
Sensors 18 02484 g011
Figure 12. The distributions of reference utility poles with crossarms (UPCs) and estimated UPCs using different numbers of views, an angle threshold of 2°, and a distance threshold of 4 m in the study area. (a) Reference utility poles; (b) estimated UPCs using 3 views; (c) estimated UPCs using 4 views; (d) estimated UPCs using 5 views; (e) estimated UPCs using 6 views; (f) estimated UPCs using 7 views; (g) estimated UPCs using 8 views; (h) estimated UPCs using 9 views.
Figure 12. The distributions of reference utility poles with crossarms (UPCs) and estimated UPCs using different numbers of views, an angle threshold of 2°, and a distance threshold of 4 m in the study area. (a) Reference utility poles; (b) estimated UPCs using 3 views; (c) estimated UPCs using 4 views; (d) estimated UPCs using 5 views; (e) estimated UPCs using 6 views; (f) estimated UPCs using 7 views; (g) estimated UPCs using 8 views; (h) estimated UPCs using 9 views.
Sensors 18 02484 g012
Figure 13. Enlarged examples of distribution patterns of reference and estimated utility poles with crossarms in the study area. (a) the locations of enlarged examples; (b) enlarged area A; (c) enlarged area B; (d) enlarged area C; (e) enlarged area D.
Figure 13. Enlarged examples of distribution patterns of reference and estimated utility poles with crossarms in the study area. (a) the locations of enlarged examples; (b) enlarged area A; (c) enlarged area B; (d) enlarged area C; (e) enlarged area D.
Sensors 18 02484 g013
Figure 14. Examples of errors in estimating locations of utility poles from GSV, caused by the difference between ground locations (black solid lines) and their corresponding orthographic projected location (green dash lines). (ad) are four different examples of leaning UPCs with different leaning angles.
Figure 14. Examples of errors in estimating locations of utility poles from GSV, caused by the difference between ground locations (black solid lines) and their corresponding orthographic projected location (green dash lines). (ad) are four different examples of leaning UPCs with different leaning angles.
Sensors 18 02484 g014
Table 1. Accuracy assessment of location estimation of utility poles with crossarms (UPCs) based on 1039 reference UPCs.
Table 1. Accuracy assessment of location estimation of utility poles with crossarms (UPCs) based on 1039 reference UPCs.
Number of ViewsThreshold of Angle (°)Threshold of Distance to Center of Selected Road (m)Percentage of the Number of Estimated Locations of UPCs Being within a Certain Buffer Zone of Reference Utility Poles (%)Number of Estimated UPCs
<1 m<2 m<3 m<4 m<5 m<6 m<7 m<8 m<9 m<10 m
3131.758.0422.3835.6646.8554.962.5968.1874.8380.07286
141.838.4223.4436.6347.6255.6864.4770.776.9282.42273
151.927.6923.0837.3149.2357.6967.3172.6979.2385260
231.718.0519.5130.9844.8853.4158.7864.6372.9377.8410
241.758.2719.831.0845.3653.8860.1566.9274.9479.95399
251.858.7120.3232.7247.7656.7363.5969.6677.3182.32379
332.378.8420.4731.943.3250.8657.5464.0171.1275.86464
342.688.9520.3631.7743.1851.2358.1765.5572.9377.18447
352.569.32032.7944.8853.0260.2366.9874.6579.53430
41327.5619.5630.8940.6751.335863.1171.1176450
141.877.2620.3730.9142.6253.8660.8967.4574.7179.16427
152.027.8320.9632.3244.4455.8163.3870.4577.7883.33396
231.317.8617.8429.1339.4448.1254.9962.5267.7672.83611
241.27.7218.0129.541.6851.2958.3265.8771.776.33583
251.098.5618.5830.7843.5353.3760.4768.1274.3279.05549
331.357.7716.892938.4246.7955.3162.7868.7674.14669
341.398.1918.0830.4540.9649.1557.565.8471.5676.82647
351.479.6119.0630.9442.5150.9859.4568.474.4378.99614
5132.210.822.3435.3545.651.8357.5164.2971.9877.47546
142.461122.3536.3648.4855.361.9368.9475.1979.17528
152.6311.322.6337.3749.756.9763.8471.9277.7883.23495
232.311119.6230.342.1448.7757.5864.3671.2875.47693
242.5310.319.232.4445.8353.2761.6168.373.6677.98672
252.5210.219.6932.647.7255.2863.1570.0877.1781.57635
332.3710.419.0529.1739.5547.8356.3763.3469.5174.24761
342.8410.419.0830.7242.6351.4260.3566.9872.2676.73739
353.0111.120.2332.2843.953.5262.4170.0176.3380.63697
6132.512.223.2136.0646.4152.9260.2767.7873.1278.46599
142.8712.223.9937.6748.1454.5663.1870.7874.3278.55592
152.512.925.0438.6449.9157.0765.4773.77882.47559
232.4310.821.6233.9244.7352.359.8665.1471.7678.24740
242.4610.321.6134.7547.7456.564.7170.8674.6979.07731
252.4711.223.1136.6350.4459.8867.373.478.0582.27688
332.229.7520.4932.4741.2350.4957.963.2170.6275.43810
342.6310.121.1333.8844.555.3862.8868.6373.2576.63800
352.7610.822.9734.7846.1957.0964.4470.4776.1279.4762
7132.712.424.0138.3149.9256.663.2868.6874.7279.81629
142.9112.425.3641.0352.6758.1665.5972.5477.7182.23619
152.7413.227.0543.1555.6560.9669.0175.3480.6584.93584
232.29.9521.7134.2444.4452.0759.5665.3770.876.36774
242.619.5223.2136.6447.7256.1964.1570.0173.1477.84767
252.3710.623.2938.3551.0559.1467.7873.0877.8282.01717
332.7510.222.0433.5343.9552.2259.2866.3572.175.57835
343.1510.223.1235.2346.2555.3362.4768.7772.8876.15826
353.3111.523.9236.0148.656.8765.1470.9975.9579.64786
8133.0812.225.0840.9252.9258.7764.7770.4670.4676.77650
143.8712.926.6343.0355.4260.8466.4173.6878.1782.51646
154.0813.728.2245.0258.463.6270.4777.1682.5486.79613
233.4411.323.6636.0147.254.3361.5865.5270.7475.7786
243.6811.325.5137.445057.6164.4769.1672.2176.78788
253.4912.126.7139.1952.6259.4667.6572.8977.1880.54745
333.0511.223.1234.6245.4252.9360.0966.271.2475.7852
343.5111.224.0935.7947.656.0261.9969.3672.476.02855
353.7813.325.6138.6649.8857.4465.7371.7176.179.63820
9132.6711.724.6741.4652.1558.9965.5371.9276.3782.91673
142.851226.3942.7354.7261.6266.8773.0177.2182.01667
153.1413.828.5745.8458.8764.5270.876.7781.1685.71637
233.1811.322.0336.7247.3754.5961.5765.9770.2675.64817
243.0412.223.4537.348.9757.2362.3367.871.0875.7823
253.4712.525.0640.3651.859.1366.0771.4775.4578.92778
332.9511.422.0536.3647.1654.7760.9165.870.1175.34880
342.8312.223.4236.8849.2156.5662.4467.9971.1575.23884
353.412.924.7439.9851.1158.1564.9571.0475.3878.55853
Table 2. Sobol’s sensitivity analysis on the number of estimated UPCs based on tested results
Table 2. Sobol’s sensitivity analysis on the number of estimated UPCs based on tested results
Sobol (N = 500)
ParameterS1S1_confSTST_conf
Number of views0.67860.09790.68510.0730
Threshold of angle0.29210.0600.29620.0299
Threshold of distance0.01500.01600.01570.0021
S1: First order sensitivity; ST: Total order sensitivity; conf: corresponding confidence intervals with a confidence level of 95%.
Table 3. Sobol’s sensitivity analysis on the percentage of the number of estimated locations of UPCs being within a 5 m buffer zone of reference UPCs based on tested results.
Table 3. Sobol’s sensitivity analysis on the percentage of the number of estimated locations of UPCs being within a 5 m buffer zone of reference UPCs based on tested results.
Sobol (N = 500)
ParameterS1S1_confSTST_conf
Number of views0.50840.083340.60660.0779
Threshold of angle0.21850.056450.28410.03889
Threshold of distance0.153670.05000.18260.0274
S1: First order sensitivity; ST: Total order sensitivity; conf: corresponding confidence intervals with a confidence level of 95%.

Share and Cite

MDPI and ACS Style

Zhang, W.; Witharana, C.; Li, W.; Zhang, C.; Li, X.; Parent, J. Using Deep Learning to Identify Utility Poles with Crossarms and Estimate Their Locations from Google Street View Images. Sensors 2018, 18, 2484. https://0-doi-org.brum.beds.ac.uk/10.3390/s18082484

AMA Style

Zhang W, Witharana C, Li W, Zhang C, Li X, Parent J. Using Deep Learning to Identify Utility Poles with Crossarms and Estimate Their Locations from Google Street View Images. Sensors. 2018; 18(8):2484. https://0-doi-org.brum.beds.ac.uk/10.3390/s18082484

Chicago/Turabian Style

Zhang, Weixing, Chandi Witharana, Weidong Li, Chuanrong Zhang, Xiaojiang Li, and Jason Parent. 2018. "Using Deep Learning to Identify Utility Poles with Crossarms and Estimate Their Locations from Google Street View Images" Sensors 18, no. 8: 2484. https://0-doi-org.brum.beds.ac.uk/10.3390/s18082484

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop