Binocular stereovision omnidirectional motion handling robot

Binocular stereovision has become one of the development trends of machine vision and has been widely used in robot recognition and positioning. However, the current research on omnidirectional motion handling robots at home and abroad is too limited, and many problems cannot be solved well, such as single operating systems, complex algorithms, and low recognition rates. To make a high-efficiency handling robot with high recognition rate, this article studies the problem of robot image feature extraction and matching and proposes an improved speeded up robust features (SURF) algorithm that combines the advantages of both SURF and Binary Robust Independent Elementary Features. The algorithm greatly simplifies the complexity of the algorithm. Experiments show that the improved algorithm greatly improves the speed of matching and ensures the real-time and robustness of the algorithm. In this article, the problem of positioning the target workpiece of the robot is studied. The three-dimensional (3-D) reconstruction of the target workpiece position is performed to obtain the 3-D coordinates of the target workpiece position, thereby completing the positioning work. This article designs a software framework for real-time 3-D object reconstruction. A Bayesian-based matching algorithm combined with Delaunay triangulation is used to obtain the relationship between supported and nonsupported points, and 3-D reconstruction of target objects from sparse to dense matches is achieved.


Introduction
In the process of human cognition of the world, we perceive external information mainly through observation. Eighty pecent of the information is obtained through vision. With the rapid development of industrial technology, traditional industrial production methods can no longer meet the requirements of the times. It has great application value in industrial production. Machine vision works by simulating human vision functions. 1 Industrial charge-coupled device cameras are used to capture images of target scenes or objects. Different image processing methods are used in the corresponding optical imaging system to obtain the various results we want.
Many research teams at home and abroad have conducted in-depth research on design methods and control algorithms of handling robots. Clavien et al. 2 introduced the design of an independent contact region-based motion controller for AZIMUT-3, using new parameters of the motion state space and using the time scaling principle originally developed for the manipulator trajectory, designed the motion controller. Yorozu and Takahashi 3 propose a real-time omnidirectional automatic obstacle avoidance method based on sensors. This method is based on the synchronous control of the translation of the movable gap and the coverage area and the effective rotational movement, thereby achieving an autonomous omnidirectional mobile robot. Chang et al. 4 proposed a model and output feedback tracking control method for omnidirectional rehabilitation training walker. The results prove that choosing the proper design parameters will cause the root mean square of the tracking error to be arbitrarily small. Wang et al. 5 investigated a new omnidirectional laser speckle rheology (LSR) catheter for mapping the mechanical properties of luminal organs without the need for rotational motion. Evaluations performed by the authors in test phantoms with circumferentially varying mechanical properties demonstrate the ability of LSR catheters to perform accurate mechanical assessment of lumen organs. Hoang and Jo 6 proposed a method based on an omnidirectional camera and a laser rangefinder to estimate the pseudo three-dimensional (3-D) motion of a vehicle using a geometrically constrained minimum parameter set. Yang et al. 7 proposed a new type of omnidirectional visual inspection system for detecting morphological defects. Experimental results show that the system can perform high-quality measurement and reconstruction on the 3-D space of the pipeline and has achieved good detection results in terms of speed and accuracy. Zhang et al. 8 propose a space-saving steering system for four-wheel steering and four-wheel drive electric vehicles with higher maneuverability and flexibility. In this article, we use model-based methods to optimize task-specific readability models to generate clear robot behavior. 9 Busch et al. 10 developed and controlled a new low-cost omnidirectional interactive mobile robot platform with force feedback function that can assist and guide the patient's hand during the treatment process. The literature 11,12 studied the problem of stably controlling motion and tracking the trajectory of a mobile robot with three rollers and solved the stability and tracking of various programmed motions and trajectories of the robot. Andreyev and Peregudova 13 proposed a possible method to build a robot capable of participating in competitions, which can be used as a reference for current and new teams. the literature 14,15 model, analyze, and construct soft robots with locomotive capabilities controlled by body deformation. Finally, a prototype of the soft robot is established, and experiments and comparisons are performed to prove the validity of the theory.
Many research teams at home and abroad have conducted in-depth research on the theory of binocular stereovision and its application in different fields. the literature 16,17 proposed a forward design optimization method of binocular stereovision system based on highdimensional function modeling and optimization, and the feasibility of the method was verified by experiments. The literature 18,19 propose a self-calibrating single-lens 3-D video extensometer for noncontact, nondestructive, and high-precision strain measurement to prove the effectiveness of the proposed self-calibrating single-lens 3-D video extensometer. Wang et al. 20 designed calibration targets suitable for these two cameras. It can be seen from the "binocular" stereovision system reconstruction experiment results that the calibration data provided by this method can meet the system requirements. Similarity measurement is an essential step; however, there is always a trade-off between accuracy and computational efficiency. 21 The literature 22,23 used a high-speed binocular stereovision system to establish a combustion field diagnostic method and demonstrated the feasibility of the method. Yue-fei et al. 24 proposed a method based on stereovision, which can accurately measure the twist of helicopter rotor blades and simplify the operation process. In Xiong et al., 25 the test results show that the reconstruction by binary speckle has better performance. Tang et al. 26 achieved a selective baseline for stereovision by selecting two images from a onedimensional light field image to view a close-range object using a single master lens. Koyama et al. 27 discusse the occlusion problem and provide a solution for selfocclusion and occlusion between objects. The results show that the algorithm has superior performance and can recognize 80% or less occluded objects. The literature 28,29 proposed a binocular image stereo-matching algorithm for dynamic image sequences. The results show that the proposed algorithm can effectively suppress the noise generated by stereo matching of dynamic binocular vision images. Hui et al. 30 proposed a pedestrian detection algorithm based on binocular stereovision. Experimental results show that this method can effectively use binocular visual information to avoid monocular visual false alarms, improve detection accuracy, and increase detection speed. Jin et al. 31 established a kinematic model for a system with three segments connected by nonstandard connection types and their trajectories. The comparison of simulation results shows that the improved A* algorithm optimizes the path quality better than A* algorithm in complex environments. Xu and Liu 32 analyzed the availability of all visual cues with the help of semantic information and proposed a new visual odometer method. Experimental results confirm that the new method provides effective improvements in accuracy and robustness in complex dynamic scenes.
Although the application of robots is becoming more and more widespread, they are used in various fields of industrial production. 33 However, the working environment of most industrial robots is determined in advance, taught through a preprogrammed program, and then run according to a fixed trajectory. In general, the pose of the target workpiece is uncertain, such as occlusion and random placement. In addition, the actual pose and ideal pose of the target workpiece usually have certain errors. The existence of these errors may, as a result, the robot cannot work as expected, causing the robot to make a mistake. In general, the target workpiece conveying system in the omnidirectional motion conveying system does not have a fixed beat. There may be multiple target workpieces and the order is also changed. In this way, the fixed teaching industrial robot cannot meet the production requirements. Therefore, the traditional robot preteaching mode has limited the flexibility and flexibility of omnidirectional motion handling to a large extent, and binocular stereovision technology can just solve this problem.
Based on binocular stereovision, this article uses improved image feature extraction and matching algorithm of speeded up robust features (SURF) algorithm and combines point-based 3-D reconstruction method to achieve 3-D reconstruction of target workpiece position, obtain 3-D coordinates of target workpiece position, and complete the robot. This article studies the problem of robot image feature extraction and matching and proposes an improved SURF algorithm that combines the advantages of both SURF and Binary Robust Independent Elementary Features (BRIEF). Experiments show that the improved algorithm greatly improves the matching speed, ensures the real-time performance of the algorithm, and retains the robustness of the traditional SURF algorithm. This article studies the positioning of the target workpiece of the robot and combines the internal and external parameters obtained from the camera's stereo calibration and the 3-D reconstruction method based on points to 3-D reconstruct the target workpiece position to obtain the 3-D coordinates of the target workpiece position to complete the positioning work. This article designs a software framework for realtime 3-D object reconstruction. A Bayesian-based matching algorithm combined with Delaunay triangulation is used to obtain the relationship between supported and nonsupported points, and 3-D reconstruction of target objects from sparse to dense matches is achieved.

Robot feature extraction model based on improved SURF algorithm
In this article, the improved SURF algorithm uses the binary descriptor BRIEF operator at the feature description stage. The core theory of the operator is to randomly select a number of points around the feature points to make up the image block and then binarize its gray value. Combine the obtained result into a binary string and use it as the feature descriptor of the feature point. Any bit of the BRIEF descriptor is obtained by randomly selecting two binary points for comparison. Define a binary comparison criterion t of an image neighborhood p, as shown in the following formula, whose size is S Â S tðp; x; yÞ ¼ 1; pðxÞ < pðyÞ Among them, x and y are any two pixels in the image neighborhood P, and p(x) and p(y) are the gray values at x and y, respectively. There are many different methods for selecting nd (x i , y i ) pixel pairs from S Â S-sized image blocks. Calonder experiments use different sampling methods to study them and test the following sampling methods Final experiments proved that the second method yielded the best results, so the BRIEF descriptor used the second method. The descriptors generated by the above method have no direction, so to obtain the feature points with rotation invariance, this article uses the gray centroid method to obtain the feature point main direction q for the feature points extracted by the traditional SURF algorithm. For any feature point p, the moment of its neighborhood pixels can be expressed by the following formula x p y q Iðx; yÞ ð 2Þ Omnidirectional azimuth reconstruction algorithm based on target positioning

Gaussian mixture model
Generally, different sampling parts in the image correspond to different Gaussian models. The single Gaussian distribution density function is shown in the following formula Among them, u is the center point of the pixel, and S describes the covariance matrix. The method of establishing Gaussian mixture model (GMM) is as follows: Human experience gives the initial weight of each GMM, as shown in w 0 Then use the K-means algorithm to find the centers of modified GMM density functions, as shown in 2. Iteratively approximate the final parameters by using expectation maximization algorithm. 3. GMM is obtained from M single Gaussian functions by weighting and averaging. The result is shown in

GrabCut algorithm description
The GrabCut algorithm uses two incomplete labeling and iterative graph cutting strategies to reduce the amount of subsequent user interaction operations. For image Þof each pixel is obtained. This method uses a GMM to characterize the data, and the segmentation problem is transformed into a ¼ arg min a Eða; k; q; yÞ. Gibbs energy functions are shown by H Eða; k; q; yÞ ¼ U ða; k; q; yÞ þ V ða; yÞ and U ða; k; q; yÞ ¼ P n Dða; k; q; yÞ.
The data items are defined as shown in the following formula D a n ; k n ; q; y n ð Þ¼À logp y n ja n ; k n ; q; y n ð Þ À logp a n ; k n ð Þ where pða; kÞ is the weight of each Gaussian model. The main implementation process of the GrabCut algorithm is as follows: (1) First determine a rectangular frame and then use the Trimap algorithm to mark the entire image. The pixels outside the rectangular frame constitute the background area T B of the image and remember the unknown area T v ¼ T B of the image and the foreground area of the image is T F ¼ ; (2) the opacity of the image is then set according to related theories, where the opacity a p ¼ 0 of the pixels in the T B area is recorded and the opacity a p ¼ 1 of the pixels in the T U area; (3) T B the pixels in the region are used to establish the initial background GMM color data model, and the pixels in the T U region are used to establish the initial foreground GMM color data model; (4) the GMM labels of all pixels in the studied image are updated in real time; (5) the real-time update the GMM parameter q; (6) use the max-flow/mincut algorithm to process the Gibbs energy function to minimize its value; and (7) the algorithm continues to execute until the iteration abort condition is reached, otherwise it returns (4).

Stereo-matching algorithm for omnidirectional motion based on Bayes theory
By scaling the local surface shape index map of the 3-D target, the multiscale characteristics of the target surface are simulated, and feature points are defined at each scale, so that stable features can be extracted regardless of the target's scale change. The Gaussian function can produce smooth characteristics that meet the special criteria of scale-space requirements. As the Gaussian parameter s decreases, the details of the data features appear in turn.
For any image f (x, y, z), the result after Gaussian smoothing can be expressed as Lðx; y; sÞ ¼ Gðx; y; sÞ Ã f ðx; yÞ; Gðx; y; sÞ ¼ 1 Since the information in the Gaussian scale space is redundant, a Gaussian scale-space consisting of a set of Gaussian smooth images is first established and then a point-to-point subtraction operation is performed on adjacent scales to generate a Gaussian differential (difference of Gaussian) scale space, that is Dðx; y; z; sÞ ¼ Lðx; y; ksÞÀ Lðx; y; sÞ To get key descriptor L li ; R ri ð Þ, you need to meet Among them, D LR1 represents the point R i closest to L i in the figure, D LR2 represents the point R i closest to L i in the figure, and t represents the threshold. When the above formula is satisfied, this point is considered to be the best matching point. Under this condition, mismatch will still occur. Using random sample consensus (RANSAC) operator to remove mismatch, we can get accurate support point matching pair Let the image pair be [imagel, imager], the set of matched support points is S ¼ s 1 ; s 2 ; . . .
where p d n jS; O 0 n À Á represents the prior distribution function and p O r n jO l n ; d n À Á represents the image likelihood function. Interpolate each triangle in the Delaunay triangle mesh obtained using the support points to get the plane where i represents the pixel value corresponding to the plane. The Bayesian theory is used to obtain the posterior probability distribution of Markov random fields by using the disparity laws of different regions where D is the parallax distribution of the image, I is the stereo image pair, and E is the global energy function.

Data sources
The binocular camera used in this article is a small-scale industrial digital camera with Giga Giga network; the two cameras are fixed and installed in parallel on the axis. The camera is connected to the computer through two Giga network cables, and the corresponding software development kit (SDK) is installed to complete the camera setting, so that the images collected by the camera can be transferred into the computer with high fidelity. In the actual experiment, the installation of binocular camera cannot be completely parallel, so it is necessary to calibrate it and adjust the camera to achieve the optimal layout. To ensure the accuracy of the experiment, the chessboard calibration board of visco is selected in this experiment. The corresponding model is cp-400-o-30, and the calibration board is 12 Â 9 square array. The size of each square is 30 mm Â 30 mm, and the error accuracy is within +0.2 mm. In this experiment, the binocular camera collects 17 images of calibration plate with rotation and offset and saves the corresponding calibration images. The corner points of each pair of calibration plate images collected by the left and right cameras are extracted. The corner points are mainly the inner corner points of the calibration plate, namely the most central 6 Â 9 line corner points. See Table 1 for the specific parameters of 17 images.

Selection of positioning system for handling robot
(1) Robot motion system control After obtaining the calibration relationship of the system camera and successfully obtaining the 3-D coordinates of the target workpiece position, this article uses a binocular stereovision-based visual servo system control method to guide the robot to grab the target workpiece. The basic principle of binocular stereovision-based visual servoing system is to estimate the target pose in 3-D Cartesian space by extracting the characteristics of the target image, combining the known target geometric model and the internal and external parameters of the camera. Based on the pose deviation, a visual controller is designed to obtain the desired movement speed of the end of the manipulator in Cartesian space. The Jacobian matrix of the robot is used to calculate the movement speed of the joint space. The robot controls the movement of the robot.
(2) Error analysis and compensation system In the process of motion, industrial robots generate inertial forces of movement, which cause the rods of the robot to deform elastically, causing vibration, and affecting the positioning accuracy of the robot to a certain extent. This article analyzes the elastic vibration error, adopts the binocular stereovision-based elastic vibration error compensation strategy, and uses the elastic vibration error model of the handling robot to solve the deformation of the flexible rods in each position. The least square method is used to correct the trajectory, the correction result is as the compensation amount of the ideal joint, and the influence of the flexible deformation of the rod on the positioning accuracy of the industrial robot is further reduced.

Data sources and equipment
The omnidirectional motion handling robot positioning system based on binocular stereovision studied in this article uses a six-axis industrial robotic arm based on YCD-I six-degree-of-freedom mechanical servo system. Through corresponding theoretical research, the robot's positioning of the workpiece is finally realized. The hardware system includes a six-axis robot, HNY-CV-002 parallel binocular camera, light source, computer, and workpiece.
Binocular stereovision is the simultaneous acquisition of images of the same target by two cameras, which have strict requirements on the calculation speed and transmission speed of image data. This article uses the HNY-CV-002 parallel binocular camera (see Table 2). It uses a universal serial bus 3.0 communication interface with a frame rate of 30 frames/s and a transmission speed greater than 100 megabytes/s. The main parameters of HNY-CV-002 camera are presented in Table 2.

Experimental environment and software implementation process
This article uses Visual Studio 2010 software platform to realize the vision positioning technology of natural stacked workpieces and combines Halcon 12 and Open CV 3.1.0 machine vision algorithm library to write and verify the algorithm of the design system of this article. At the same time, the entire software system is integrated into the software operation interface using the Qt graphical user interface (GUI), which improves the degree of system visualization. Second, the camera driver and SDK development were redeveloped with Halcon software. The algorithm implementation and system GUI design were in Cþþ and Qt languages, respectively. The entire system is divided into a template part and a recognition part. The template part mainly uses the point cloud of a single workpiece as a template, and the recognition part is the recognition of the target workpiece in the point cloud image to be measured. The interface is divided into three operation modules: camera calibration module, 3-D reconstruction module, and target recognition and positioning module. The camera calibration module uses a binocular camera to collect the image of the calibration plate, extracts the corners of the central part of the checkerboard calibration plate, and outputs the calibration result. The 3-D reconstruction part reconstructs the speckle map of the collected workpiece and successively outputs supporting point cloud, dense point cloud, and point cloud package map. Target recognition and positioning according to the input point cloud template of a single workpiece, the point cloud image to be measured is identified and located, and the recognition result and pose information under coarse and fine extraction are output. 34 The software implementation process of this system is shown in Figure 1.

Results and discussions
Analysis of robot feature extraction results

(1) Contrast test of scale þ rotation change
From the experimental results in Figure 2, it can be seen that, in terms of matching accuracy, the matching accuracy of the two algorithms is not much different, and both remain around 90%; in terms of the time consumed for feature description, the improved binary descriptor used in SURF algorithm in this article compared to traditional SURF, the operating speed is increased by about 1.62 times.
(2) Contrast test of image blur In the image blur change comparison experiment, Figure 3(a) is the matching image between the original image and the 3 Â 3 Gaussian template after the blur change and experimental result data. Figure 3(b) is the original image and the 5 Â 5 Gaussian template matching map between blurred images and experimental result data. From the above experimental results, it can be known that under fuzzy changes, both algorithms can maintain a high matching accuracy, both of which are above 90%. In particular, the matching accuracy of the improved SURF algorithm in this article is above 95%, which is higher than the traditional SURF algorithm about 3.6 percentage points. In terms of algorithm operation speed, compared with the traditional SURF algorithm, the operation speed of the improved SURF algorithm in this article is increased by about 1.73 times.    reduced, from 179 to 134 and from 96 to 71, which reduces the matching by about 25%. Yes, but the matching accuracy of the two algorithms is relatively stable, both of which are above 90%. In addition, compared with the traditional SURF algorithm, the matching accuracy of the improved SURF algorithm is improved by about 2.51 percentage points. In terms of algorithm running speed, compared with the traditional SURF algorithm, the improved SURF algorithm in this article runs 1.86 times faster.
In this paper, the improved SURF algorithm is described by replacing the feature point description method in the traditional SURF algorithm with the binary descriptor BRIEF operator. The improved SURF algorithm still uses the traditional SURF algorithm for feature point detection at the feature point detection stage, so that the obtained feature points have scale-invariant features; and the gray point centroid method is used to calculate the main direction of the feature points to make the detected features points have directionality, which makes the algorithm rotation invariant. In the feature point description stage, the binary descriptor BRIEF operator is used for feature description. This description method has fast calculation speed and reduces the algorithm's operation time. Finally, through a large number of experiments, it is proved that compared with the traditional SURF algorithm, in the application scenario where image changes exist, the improved SURF algorithm not only improves the realtime performance of the algorithm but also maintains the same performance as the traditional SURF. The SURF algorithm can meet the fast and accurate matching of the target workpiece image by the omnidirectional motion handling robot.

Analysis of robot target positioning and reconstruction results
According to the binocular stereo camera calibration and 3-D reconstruction theory of target workpiece feature points described above, this section designs a 3-D reconstruction error analysis experiment to judge and analyze the binocular stereo camera calibration accuracy and 3-D reconstruction of target workpiece feature points. The binocular stereovision system studied in this article uses the HNY-CV-002 parallel binocular camera with a baseline distance of 55 mm. The camera's left and right camera coordinate system only has translation in the x-axis direction, and the internal parameters of the left and right cameras basically the same. The experiments performed 3-D reconstruction on the corner points of the extracted feature map. The accuracy of the 3-D reconstruction method was verified by calculating the distance between two adjacent corner points in the reconstructed feature map. The length of each square in the actual black and white grid is 32 mm. Calculate the distance between adjacent corner points in the black and white grids in the 3-D reconstruction result and then perform the same with the side length of the grid in the actual black and white grid comparative analyses, the experimental results are presented in Table 3. The error curve is shown in Figure 5.
From the results in Table 3 and Figure 5, it can be known that the average error of the reconstruction results obtained by the 3-D reconstruction method used in this article is 0.276 mm when the actual side length of the grid is 32 mm. It can be inferred that the 3-D reconstruction method has high accuracy and can meet the requirements of the flexible loading and unloading robot for grasping the target workpiece. By analyzing the errors, we can know that the reasons for the above errors are as follows: (1) The inevitable errors generated during the stereo calibration of the binocular camera, such as the errors in the extraction of the corner points of the black and white grid plates during calibration. (2) During the process of collecting images through the camera, due to the influence of environmental factors, such as light and noise interference. (3) The camera installation problem is impossible for the left and right cameras to be absolutely parallel.
This article mainly studies the positioning and reconstruction of the target workpiece position. First, the stereo-matching technology of binocular stereovision is introduced. Second, the segmentation method is used to eliminate the influence of background factors to obtain the  foreground target. Then, the foreground target in the left and right images workpiece is stereo matching. This article uses the improved SURF feature stereo-matching method to obtain the stereo-matching result. Then, by using the point-based 3-D reconstruction method, the target workpiece position is 3-D reconstructed to obtain the 3-D coordinates of the target workpiece position. Finally, the accuracy of the 3-D reconstruction method is verified by 3-D reconstruction error experiments.

Analysis of robot stereo-matching results
In the experiment, two pairs of binocular cameras and a digital speckle projection were used. The generated digital speckle was irradiated to the surface of the object by the projector, and the left and right camera pictures were collected by the camera. In this experiment, a gypsum portrait is used as the target object, because there is more complex texture information in the gypsum portrait and using this as a target workpiece can better reflect the superiority of the algorithm. First, after the binocular camera is calibrated, the target object is placed at a designated position on the work platform, and a speckle projector projects a digital speckle image onto it. After the images are collected by the binocular camera, the support points and parallax are calculated using SURF feature detection and epipolar constraint criteria. Then, the disparity value is used as the prior probability value of the Bayes dense match. Due to the uniqueness of Delaunay triangles, combined with the dense matching algorithm of Bayes theory, disparity values of nonsupporting points are obtained. Finally, the rotation matrix and translation matrix obtained by camera calibration are used to obtain the spatial 3-D coordinates of the unsupported points to obtain the dense point cloud image. The reconstructed point cloud image is stitched to obtain a relatively complete object 3-D point cloud model.
In Figure 6 it can be seen that the point cloud information of the support point reconstruction using the SURF and epipolar constraint rules can clearly describe the contour information of the target object, and the reconstructed edge information is relatively smooth. The point cloud information obtained by the Bayes matching algorithm clearly expresses the details and texture features of the target object, has higher accuracy, and reduces the mismatch rate, and its accuracy is 0.1 mm.
By comparing the SURF and digital speckle reconstruction method (referred to as method 2 for short) and the reconstruction method described in this article (referred to as method 1), the comparison is made from two aspects of running time and logarithm of matching points. It can be seen from the figure that although method 1 is 0.04 s slower than method 2, the number of matching point pairs obtained by method 1 reconstruction method is 16 times that of method 2, and the obtained 3-D point cloud density is greater, which is more beneficial to the object.
In this article, binocular stereovision reconstruction methods are compared and different binocular stereomatching methods are introduced. Aiming at the problem that most binocular stereo-matching algorithms cannot meet the requirements of reconstruction accuracy and real-time performance, a SURF feature detection algorithm is used to extract feature points from speckle images, and the Bayes stereo-matching algorithm and Delaunay triangle method are used to obtain the denseness of the target object. The disparity information is combined with camera calibration parameters to obtain a dense point cloud image. Experimental verification shows that the algorithm can obtain high-density point cloud information with a slight loss of speed and fully characterize the texture and contour information of the target object.

Conclusions
Based on binocular stereovision, this article uses improved image feature extraction and matching algorithm of SURF algorithm and combines point-based 3-D reconstruction method to achieve 3-D reconstruction of target workpiece position, obtain 3-D coordinates of target workpiece position, and complete the robot.
This article studies the problem of robot image feature extraction and matching and proposes an improved SURF algorithm that combines the advantages of both SURF and BRIEF. The feature description phase of the algorithm uses binary descriptor BRIEF operator for feature description, which greatly simplifies the complexity of the algorithm. At the same time, Hamming distance is used to match feature points. Ratio detection method and RANSAC algorithm are used to eliminate false matches. It is shown that the improved algorithm greatly improves the matching speed, ensures the real-time performance of the algorithm, and retains the robustness of the traditional SURF algorithm.
In this article, the problem of positioning the target workpiece of the robot is studied. The stereo coordinate matching method based on the improved SURF algorithm is used to obtain the image coordinates of the matching point pair. Then, the image coordinates of the target workpiece position marker points are calculated based on the average method. Finally, the camera stereo is combined. The internal and external parameters obtained by calibration and point-based 3-D reconstruction can reconstruct the position of the target workpiece and obtain the 3-D coordinates of the position of the target workpiece, so as to complete the positioning work.
This paper also uses a combination of binocular camera and laser speckle to design a software framework for real-time 3-D object reconstruction. A Bayesian-based matching algorithm combined with Delaunay triangulation is used to obtain the relationship between supported and unsupported points. Under the premise of less loss of reconstruction efficiency, the accuracy of point cloud reconstruction is improved.

Declaration of conflicting interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.

Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This work was supported by the Key Research and Development Program of Shaanxi (Program No.2019GY-123).