3D target localization based on multi–unmanned aerial vehicle cooperation

Ground target three-dimensional positions measured from optical remote-sensing images taken by an unmanned aerial vehicle play an important role in related military and civil applications. The weakness of this system lies in its localization accuracy being unstable and its efficiency being low when using a single unmanned aerial vehicle. In this paper, a novel multi–unmanned aerial vehicle cooperative target localization measurement method is proposed to overcome these issues. In the target localization measurement stage, three or more unmanned aerial vehicles simultaneously observe the same ground target and acquire multiple remote-sensing images. According to the principle of perspective projection, the target point, its image point, and the camera’s optic center are collinear, and nonlinear observation equations are established. These equations are then converted to linear equations using a Taylor expansion. Robust weighted least-squares estimation is used to solve the equations with the objective function of minimizing the weighted square sum of re-projection errors from target points to multiple pairs of images, which can make the best use of the effective information and avoid interference from the observation data. An automatic calculation strategy using a weight matrix is designed, and the weight matrix and target-position coordinate value are updated in each iteration until the iteration stopping condition is satisfied. Compared with the stereo-image-pair cross-target localization method, the multi–unmanned aerial vehicle cooperative target localization method can use more observation information, which results in higher rendezvous accuracy and improved performance. Finally, the effectiveness and robustness of this method is verified by numerical simulation and flight testing. The results show that the proposed method can effectively improve the precision of the target’s localization and demonstrates great potential for providing more accurate target localization in engineering applications.


Introduction
In recent years, unmanned aerial vehicles (UAVs) have been successfully applied in various fields. Their value as surveillance platforms has been proven repeatedly in both military and civilian domains. [1][2][3][4][5] As an essential function of UAVs, target localization aims to solve the three-dimensional (3D) coordinate of the point on the ground by combining the data obtained by the measurement equipment carried on the UAV with the UAV's flight data. [6][7][8][9][10] Vision sensors have been widely used in UAV target localization due to their passive and non-cooperative characteristics. Vision sensors do not emit any energy to measure the target and do not need to communicate with the targets. [11][12][13] However, the monocular vision sensor imaging projects 3D objects onto the twodimensional (2D) image plane, and therefore it can only provide the bearing information of the target and not the range information. 14 In order to solve this problem, many schemes have been proposed. In their study, 15 Hosseinpoor et al. used a UAV with real-time kinematic (RTK) global positioning system (GPS) for estimation and localization, and processed the localization results through extended Kalman filtering. Meanwhile, the scale-invariant feature transform (SIFT) was used to extract the feature points of the same ground target in different frames in Han and de Souza's study. 16 The relative height of the target and the UAV was calculated through 3D reconstruction.
Quintero et al. 17 in their work studied the optimal routing of two camera-equipped fixed-wing aircraft cruising at fixed altitude and cooperatively tracking a single ground target. In Sharma et al.'s study, 18 the extended Kalman filter was used to improve the accuracy of target localization. However, this method needs to assume that the altitude of the ground is zero, and it only works on the flat ground. In Barber et al.'s work, 19 a vision-based target localization system for micro fixed-wing aircraft was introduced. In the flight test, the UAV located the stationary target while flying around the target.
In the above studies, one single UAV was used to achieve the target localization. However, a single UAV cannot be equipped with many sensors for target localization since it suffers drawbacks such as the payload weight limit. In recent years, with the development of control, communication, electronics, and other technologies, the cooperative work of multiple UAVs has been realized. [20][21][22][23][24][25] For this reason, a multi-UAV cooperative target localization scheme is proposed in this paper. The main contributions of this paper are summarized as follows: 1. We propose a novel multi-UAV cooperative target localization framework. The optoelectronic platform is allocated to each UAV, and each camera observes the same target point simultaneously and acquires multiple remote sensing images. 2. We establish a mathematical model of the multi-UAV cooperative target localization system, which is a set of linear equations derived using the constraints of three-point collinearity. 3. We formulate the objective in terms of minimizing the weighted square sum of re-projection errors from target points to multiple pairs of images and design an automatic calculation strategy for the weight matrix.
Furthermore, the effectiveness and robustness of our method is verified by numerical simulation and flight testing. The results show that our method can effectively improve the precision and enhance the antiinterference of the target localization system.

System composition and coordinate definition
The overall structure of the system In this study, multiple UAVs locate the same ground target cooperatively. Each UAV is equipped with a GPS, inertial measurement unit (IMU), and optoelectronic platform, as shown in Figure 1. The optoelectronic platform includes a camera and a camera pan-and-tilt system placed on the plane in a pod structure. During target localization, the video and telemetry information of each UAV are transmitted through a data link to the ground station. The manipulator can control the camera pan-and-tilt system to search for targets using a joystick, and when targets of interest appear on the screen, the optoelectronic platform locks onto the image and simultaneously measures the azimuth and elevation angles of the camera. Meanwhile, the UAV position and attitude are measured by GPS/IMU devices. The measured data of each UAV are unified in the reference coordinate system, and the target coordinates are solved by a multi-UAV cooperative target localization algorithm.

Coordinate system
Seven coordinate systems are used in this study. They are defined as follows: [26][27][28] 1. Geodetic coordinate system: based on international terrestrial reference system WGS-84. Geodetic coordinate system describes spatial position by latitude, longitude, and geodetic elevation (l, f, h). 2. Terrestrial rectangular coordinate system G(O g -X g Y g Z g ): the origin is located in the center of the reference ellipsoid. The axis Z g points to the north pole of the reference ellipsoid, and the axis X g is directed to the intersection point of the Greenwich meridian plane and equator. The axis Y g is perpendicular to the X g and Z g axes, following the right-hand rule. 3. UAV geographic coordinate system N(O n -X n Y n Z n ): the origin is the position (l, f, h) of a UAV, the X n points to the east, and the Y n points to the north. The axis Z n is perpendicular to the Y n and X n axes, following the right-hand rule.

UAV body coordinate system
the coordinate origin is at the center of mass of the UAV, X b points to the direction right above the UAV, Y b points to the UAV nose, and Z b , X b , and Y b form a right-handed coordinate system. 5. Camera coordinate system C(O c -X c Y c Z c ): the origin is the optical center of the camera, the X axis and Y axis are parallel to the X and Y axes of the image, and the Z axis is the optical axis of the camera, which is perpendicular to the image plane. 6. Image coordinate system: the origin is the intersection point of the camera's optical axis and the image plane, and the image coordinate system is a 2D rectangular coordinate system. The actual physical scale (mm, um, etc.) is the unit. 7. Pixel coordinate system: the origin is the upper left corner of the image, the X p and Y p represent the number of rows and columns of the pixel in the digital image. Figure 2 shows the relationship between camera coordinate system, image coordinate system, and pixel coordinate system.

Key technology of multi-UAV target localization
Principle of stereo-image-pair cross-target localization According to the perspective projection imaging principle, the three points of target point A, the corresponding image point a, and the camera's optic center S are collinear when the camera observes the ground target, as shown in Figure 3. Therefore, the following equation can be formulated according to the aforementioned geometric relation 29,30 x =À f where (x, y) is the image coordinate of the image point a, (X A , Y A , Z A ) is the coordinate of the target points in the reference coordinate system, (X S , Y S , Z S ) is the coordinate of the camera's optic center in the reference coordinate system, and f is the focal length of the camera. The rotation matrix of the coordinate-system conversion from the reference coordinate system to the camera coordinate system can be represented as R, and the translation matrix is T The terrestrial rectangular coordinate system is used as the reference coordinate system in this paper, and the conversion process from the reference coordinate system to the camera coordinate system is shown in Figure 4.
The coordinates of the target in the camera coordinate system can be calculated as T are the coordinates of the target in the camera coordinate system and terrestrial rectangular coordinate system, respectively; T N G (l, u, h) represents the origin of the UAV geographic coordinate system in the terrestrial rectangular coordinate system, and R N G (l, u), , and R C B (a, b) are the rotation matrices for coordinate conversion.
In the target localization process, the UAV position (l, u, h) and attitude (f, g, u) can be provided by GPS and IMU, respectively, and the azimuth and elevation  angle (a, b) can be obtained by the optoelectronic platform; thus, R and T can be obtained by equation (3).
We let two UAVs observe the same ground target point A at S 1 and S 2 , and obtain image points a 1 and a 2 , respectively, as shown in Figure 5. In the absence of error conditions, the rays S 1 a 1 and S 2 a 2 intersect at ground target point A. By combining equations (1) and (3), and taking the target point coordinate (X A , Y A , Z A ) as the unknown quantity, four equations can be formulated, and therefore (X A , Y A , Z A ) can be solved.

Multi-UAV cooperative target localization based on robust weighted least-squares estimation
The above target localization method is very sensitive to all kinds of noise. This is because for aerial surveys the object distance of the imaging system is much greater than the focal length. In this paper, based on the principle of stereo-image-pair cross-target localization, the same target point is measured multiple times by multiple UAVs, and the optimization algorithm is used to improve the precision and robustness of the target localization method.
Multiple UAVs acquire n (n . 2) images for the same target point from different angles. From equation (1) and the transformation between the camera coordinate system and reference coordinate system given by equation (3), and with the target coordinates (X A , Y A , Z A ) unknown, the following equation can be derived where (4) is a nonlinear equation that is transformed into a linear equation using a Taylor expansion. The first-order Taylor expansion of equation (4) at an initial value S 0 is Defining the vectors U = Z À H(X 0 ) and V = X À X 0 , equation (5) can be rewritten as the following linear system The objective function minimizes the sum of squares of errors According to least-squares estimation, the vector V can be estimated as follows During the process of UAV target localization, the attitude and height of the UAV and the azimuth and elevation angle of the camera are all different at each measurement point. In this case, even if the same type of camera is used, due to the camera's different external parameters, the localization accuracy of each  measurement point is different as is the contribution to the error. Therefore, the weight matrix P is introduced, and the weights for the X and Y axes of the same image are set to the same value P = The objective function to be solved can be expressed as According to weighted least-squares estimation, the vector can be estimated as follows Then, the vectors U = Z À H(X 0 ) and V = X À X 0 , and the estimated value of the variable X, can be obtained aŝ The initial value X 0 can be obtained based on the stereo-image-pair cross-target localization method.
When there is an increasing number of images, a robust estimation method is often introduced to improve target localization accuracy. In this paper, an iteration method is used to assign weights for images. The main steps are as follows 1. Set an initial value of the weight matrix The first estimate of target point coordinates can be derived from equation (12).

Set
Equation (10) can then be written as Calculate the standard deviation as follows 3. The weight factor v i is defined as where u i = Dd i = s , and k 0 and k 1 are empirical constants. In our experiment, k 0 = 0:5 and k 1 = 2:0.
4. Let p i = p i v i replace p i in equation (12), and recalculate the target point coordinates. 5. Repeat steps (2)-(4) until the difference between the two estimated results is less than the threshold value.

Experimental results and analysis
It is assumed that the truth values of the target point coordinates are X A = ½ l A = 39:5691518 u A = 105:3640668 h A = 100m T , and three groups of images with known interior and exterior elements (see Tables 1-3) are given, where f is the focal length of the camera; X s = l s u s h s ½ T are the coordinates of the UAV position; f, g, u are the attitude angles of the UAV; and a, b are the azimuth angle and the elevation angle of the camera, respectively.

Verification of algorithm correctness
To verify the correctness of the algorithm, equation (1) can be used to calculate the pixel coordinates x i of the target point in different images. Gauss errors of 2 m and 0.1°are added to the UAV position and attitude angle, respectively. Then, the target localization algorithm presented in this paper is used to calculate the new coordinates X i of the ground point A in three groups of images, and the results are finally compared with the truth values. In this example, all images are set with the same precision, which means the weight of all of them is 1. We use the Monte-Carlo method to conduct 1000 target localization experiments for these three groups and statistically analyze the localization errors. The results show that the target localization errors (root mean squared error, RMSE) of the three groups are 3.81, 7.74, and 4.95 m. This test verifies the correctness of the algorithm.

Verification of algorithm robustness
Taking the first group of data in Table 1 as an example to verify the robustness of this algorithm, 1 pixel (4 mm in size) of Gaussian noise is added to the coordinates of each image point, and a 2-mm gross error is added to the horizontal coordinate of the image point of image 1. Then, the method proposed in this paper is used to calculate the multi-UAV cooperative target localization. To display the precision of the localization intuitively, Dd is used to represent the displacement deviation of the target localization result from the truth value. Each image is assigned a weight of 1 in the initial calculation. Table 4 shows the weight of each calculation and the estimated result for the ground point.
As shown in Table 4, after several iterations the weight of the image containing the gross error decreases to 0 gradually, thus nullifying its effects during target localization. This method can avoid gross errors by redetermining the weight matrix during the target  localization process, thus demonstrating good robustness. On the contrary, the stereo-image-pair crosstarget localization method has poor robustness since the method does not make use of the redundant observation information and is greatly affected by the observation precision of single images. Fifteen (C 2 6 ) localization results were generated by the stereo-image-pair cross-target localization method and are shown in Table 5. Owing to the presence of a gross error, at least five estimated values among the stereo-image-pair cross-target localization results deviated largely from the truth value of the ground point.
It can be seen from Table 4 that after iteration, weight 1 is equal to 0, and weights 2-6 are equal to 1, which means that the final target localization value is determined by images 2-4. Compared with the results in Table 5, this localization accuracy is significantly better than that of stereo-image-pair cross-target localization method, which shows that the random error can be effectively suppressed. To sum up, the iterative weighting is to eliminate gross errors, and multi-UAV cooperative observation is to suppress random errors. Both of them improve the robustness of the localization algorithm.

Comparison of target localization precision of two algorithms
Since the target localization method proposed in this paper can effectively utilize the information of multiple images, it is not sensitive to measurement errors. To verify this feature, the proposed method was compared with the stereo-image-pair cross-target localization method. A curve of its localization precision changing with errors (0-100 pixels, tested at an interval of 1 pixel) of image point coordinates for the first group of images is plotted in Figure 6(a). Similarly, curves of the target localization precision changing with errors of the image point coordinate are plotted in Figure 6(b) and (c) for the other two groups of images (the group of images at large heights (errors of 0-10 pixels, tested at an interval of 0.1 pixel) and the group of images at mixed heights (errors of 0-20 pixels, tested at an interval of 0.2 pixel)). As shown in Figure 6, the localization precision of both the stereo-image-pair cross-target localization method and the method proposed in this paper decreases approximately linearly with increasing error of the image point coordinates. Meanwhile, for  each group of images at the same error level, a significantly greater localization precision is obtained when the proposed method is used. Upon adoption of the proposed target localization method, for the images at small heights, a localization precision of approximately 9 cm can be achieved when the image point coordinates are at the 10-pixel error level and a localization precision of 80 cm can still be achieved when the error is 100 pixels. For the images at large heights, a localization precision of approximately 1 m can be achieved when the image point coordinates are at the 10-pixel error level; meanwhile, for the images at mixed heights, a localization precision of approximately 80 cm can be achieved when the image point coordinates are at the 10-pixel error level. Similar to the case in which errors are added to the image point coordinates, target localization precision also shows a linearly decreasing trend with the errors in the exterior line element (UAV position) and angle elements (UAV attitude angle and camera angles), which will not be discussed in more detail here. As shown in Figure 6, it is concluded that at the same error level the target localization precision of the proposed method is much greater than that of the stereo-image-pair cross-target localization method.

Error analysis of proposed target localization algorithm
To analyze how the algorithm is affected by errors, the following simulation test was designed: 1. Truth values of ground point coordinates: 108 ground points with known coordinates uniformly distributed within a certain area were selected, as shown in Figure 7  As shown in Figure 8, for the images at small heights, after the errors in image point coordinates, exterior line elements, and angle elements were added individually, the components on the X, Y, and Z axes exhibit a similar deviation curve amplitude. However, the 10-m error of the line element led to the greatest deviation of the localization results from the truth values. As shown in Figure 9, for the images at large heights with an error being added to exterior line elements individually, the components on the X and Y axes exhibit a similar deviation amplitude and are much more stable than the component on the Z axis   and slightly more affected by the error. The 10-m error of the line element led to the greatest deviation of the localization result from the truth value. As shown in Figure 10, compared with the images at large heights, the component on the Z axis in the images at mixed heights is much less sensitive to the error in the exterior line element. Figures 8-10 show that the addition of the same level of noise into images at different heights will cause differences in the deviation degree of the localization results. Among the deviations, images at small heights have the smallest localization deviations, followed by the images at mixed heights, and then by the localization results of the images at large heights. When multi-UAV cooperative target localization is used at large heights, the component on the Z axis is more sensitive to errors, and thus the deviation of localization results is worse. To address this problem, information of the images at small heights can be used to effectively reduce the sensitivity of the component on the Z axis to errors, thus achieving greater localization precision.

Flight test
In this section, the proposed approach is validated using flight test data. In order to verify the feasibility of the algorithm, the flight test is performed in an outside field environment. One known target point is arranged on the ground, and six UAVs are launched to observe the target point simultaneously. Two of the medium UAVs are equipped with an optoelectronic platform (resolution ratio 192031080, a pixel size of 4.2 mm), while the other four small UAVs have vertical downlooking cameras (resolution ratio 12803720, a pixel size of 5.5 mm). Each UAV is equipped with a highprecision RTK GPS module, and its positioning error is less than 10 cm (circular error probable, CEP). The UAVs are shown in Figure 11. A ground receiving station receives the measurement data and video sequences from the six UAVs.
Four indices are used to evaluate the target localization accuracy: sB represents the longitude error, sL represents the latitude error, sH represents the altitude error, and ssum represents the error radius.
The flight test results show that the proposed target localization algorithm can achieve high precision. In the actual environment, the localization accuracy of low-flying UAVs is significantly higher than that of high-flying UAV, which is consistent with the simulation results. With the increase in UAV height, the localization result becomes very sensitive to all kinds of noise. This is because for aviation remote sensing systems the object distance is much greater than the focal length, and hence a slight error, such as from the GPS, IMU, or camera, will significantly enlarge the localization error.

Conclusion
To address the limitations of the existing UAV target localization method, a novel multi-UAV cooperative target localization method is proposed in this paper. A mathematical model for multi-UAV cooperative target localization based on collinear equations is established. In order to make the best use of the effective information and avoid interference from the observation data, a robust weighted least-squares estimation is used to solve the model. Finally, the effectiveness and robustness of this method is verified by numerical simulation and flight testing. The results show that the proposed method can effectively improve the precision of target localization, and it demonstrates great potential for providing more accurate target localization in engineering applications.

Declaration of conflicting interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.