Open Access Paper
28 December 2022 Research on pose and position estimation based on mask RCNN edge extraction for AR auxiliary assembly system
Teng Zhu, Xiaoyin Guo, Hansheng Yan, Yongbin Chen, Bin Shu
Author Affiliations +
Proceedings Volume 12506, Third International Conference on Computer Science and Communication Technology (ICCSCT 2022); 125066F (2022) https://doi.org/10.1117/12.2661820
Event: International Conference on Computer Science and Communication Technology (ICCSCT 2022), 2022, Beijing, China
Abstract
To solve the registration problem of complex texture objects in monocular images in an AR auxiliary assembly system, a 3D object registration method based on Mask RCNN edge extraction is proposed in this paper. This method mainly adopts the edge contour feature of long plate workpiece and enhances the robustness of the algorithm by constructing the matching relationship between 3D model and 2D feature. The experimental result shows that compared with the contour template matching of the Holcon, the Mask CNN-based pose estimation method proposed in this paper can effectively improve the effectiveness of the auxiliary system.

1.

INTRODUCTION

Augmented reality (AR) is a technology that uses computers and related display devices to superimpose virtual information on users’ vision to enhance real scene information1. Instead of text information superposition, the main point of those current AR researches is how to “place” a 3D virtual model in the real world accurately and stably2. It is necessary to calculate the pose and position of the target in the virtual scene so that the 2D projection can be displayed without deviation, Then the 3D virtual object can be seamlessly added to the real environment3.

Currently, target pose detection is one of the research hotspots in the field of computer vision4,5. The methods are mainly divided into two categories: target reconstruction and matching based on binocular vision combined with depth information6, feature matching based on monocular cameras and image edges7, and feature points or texture8. Among them, binocular vision methods can effectively recover depth information, but they require a large amount of computation and are difficult to meet the needs of real-time tracking9. However, the real-time and economic requirements of the actual production environment make the monocular vision become a more concerning research direction.

Aiming at the problem that the classical experimental algorithms do not match the application requirements, a novel 3D position and pose monitoring method was proposed in this paper based on Mask Region Convolutional Neural Network (Mask RCNN) and line feature. Experiment results show that the algorithm can overcome the problems of unstable illumination, easily occluded natural characteristics and metal surface materials in the actual production environment.

2.

PRODUCTION ENVIRONMENT AND PRINCIPLE OF MONOCULAR POSE TRACKING

The research object of this paper is an AR auxiliary system in the process of antenna board wiring of the large base station. Its main structure includes AR glasses, graphics processing terminal, scene surveillance camera, workbench and target workpiece. For cost control and practical work needs, no manual marking is allowed in monitoring scenes such as antenna board and workbench. At the same time, a common optical camera is used for scene monitoring, as shown in Figure 1a.

Figure 1.

Research environment and theoretical basis.

00217_PSISDG12506_125066F_page_2_1.jpg

In this AR auxiliary system, the overhead camera is the main information acquisition device. Its images were used to position and pose transformation information of the target antenna panel and control the virtual model to complete the corresponding movement in the AR scene. Although monocular vision can hardly obtain depth information, it has the combined advantages of fast computation, low cost and moderate accuracy. Considering the long and narrow shape of the target plate, the actual 6-D solution can be reduced to the constrained 4-D case, i.e., both the depth displacement and the long-edge turning angle can be considered within a very small variation.

3.

LINE EDGE EXTRACTION BASED ON MASK RCNN

The basic structure of Mask RCNN and fast RCNN adopts the same two-state steps. First, the method finds the Region Proposal Network (RPN), then classifies and locates each ROI found by the RPN, and finally finds the binary mask, which is different from other network frameworks that find the mask first. The process of Mask RCNN for the AR auxiliary system is shown in Figure 2.

Figure 2.

Frame of the Mask RCNN instance split.

00217_PSISDG12506_125066F_page_2_2.jpg

Compared with fast RCNN, Mask RCNN adds a mask prediction branch10, improves ROI pooling and proposes ROI alignment, which can realize one-to-one correspondence between output and input pixels. The main process of the training sector is as follows: first, an image was input to the model, which will have 2000 regions to be detected and proposed on. Then the features of every region are extracted one by one through a convolution neural network. Finally, these extracted features are classified by a support vector machine (SVM) to obtain the category of the object, and the size of the target bounding box is adjusted through bounding box regression.

The surface parts and wiring of the antenna board will change with the progress of the process. Therefore, this paper mainly uses the edge line of the antenna board as the matching feature, and the edge segment extraction method is completed by using the excellent target recognition and segmentation ability of the Mask RCNN neural network11. The process of edge extraction includes training set selection, feature extraction, SVM classification, parameter training, neural network segmentation and edge fitting. The intermediate results are shown in Figure 3.

Figure 3.

Mask-RCNN line edge extraction.

00217_PSISDG12506_125066F_page_3_1.jpg

After the Line edges were found. Rotation variation of the target sheet along the Z axis in the world coordinate system θ. It can be calculated by two long edges that are basically parallel in the target contour line extracted from the template image and the key frame image. The calculation formula is:

00217_PSISDG12506_125066F_page_3_2.jpg

where k1 and k2 respectively refer to the average slope of the two parallel long edges of the template graph and the key frame, The direction of θ is judged by the size of k1 and k2. According to the right-hand rule, if k2 > k1, it is clockwise, otherwise, it is counterclockwise. The flip angle ϒ of the target sheet along the X axis in the world coordinate system can be obtained from the curve fitting the included angle of the contour edge line and the turning angle φ, which can be obtained by the following formulas.

00217_PSISDG12506_125066F_page_3_4.jpg

where v1 and v2 is the direction vector of line edge k1 and k2. And the flip angle ϒ And Z axis displacement z is:

00217_PSISDG12506_125066F_page_3_5.jpg

4.

EXPERIMENT AND DISCUSS

The robustness of the ANN-based algorithm in the actual production environment was tested in the experimental part. And the fitting accuracy and solution speed were compared with the local template matching method of Holcon 18.11. In the intelligent production case studied in this paper, 7 groups of sampling data between [-15°, 15°] are selected Homogeneous as polynomial fitting reference. The test part of the pose solution method mainly tests the robustness under three interference factors: occlusion, illumination and overturning. The results are shown in Figure 4.

Figure 4.

Shape matching results in a variety of complex working environments.

00217_PSISDG12506_125066F_page_4_1.jpg

As can be seen from Figure 4, the Mask CNN-based pose estimation algorithm adopted in this paper has good adaptability to different illumination and occlusion interference. At the same time, to verify the accuracy of this algorithm in the pose tracking process, the fitting results are compared with the real values according to seven groups of sampled data between [-15°, 15°]. The comparison results show that the error in the Y direction is not more than 5 pixels, the error in the X direction is not more than 2 pixels, and the fitting accuracy rotation of the angle φ can reach 1°, the turning angle can reach 2°, and the error with the real value is less than that of the local contour template matching method of Holcon. For the speed of the pose solution, the average time of the Mask CNN-based method is 643ms while the Holcon matching is 864ms, which improves the efficiency by more than 25%.

Finally, Figure 5 shows the augmented reality fusion scenario from the perspective of the AR cabling auxiliary system. It can be seen from the figure that since the target plate is a large target, the visual impact of the error of the polynomial fitting algorithm on the human eye is difficult to detect, which does not affect the operator’s wiring operation.

Figure 5.

View scene in AR glass.

00217_PSISDG12506_125066F_page_4_2.jpg

5.

CONCLUSION

The edge line feature extracted by Mask RCNN was used to solve the three-dimensional pose of the target workpiece, combined with the analysis of the pose change in the actual production environment, improving the pose solution speed by reducing the dimension and reducing the search space, completes the fast and robust extraction of the edge of the antenna board by using the anti-interference of Mask RCNN neural network. The experimental results show that the accuracy of the pose solution method based on Mask RCNN can reach 2-5 pixels, and the rotation angle fitting accuracy is better than 1°, both of which exceed the local template matching algorithm of the software Holcon. And the key frame relocation efficiency is improved by more than 25%.

ACKNOWLEDGEMENTS

The basic and applied basic research project of Guangzhou basic research plan, “Research on three-dimensional tracking and registration technology of multi person collaborative virtual simulation experiment platform” (20210201298), the special project for high-level talents of Guangdong industry and Trade Vocational and technical College “Research on the key technology of rapid generation method of VR scene and model” (2021-gc-06), and special project of National Virtual Simulation Training Base “Research on Key technologies of Remote Sensing virtual simulation training system”, the special project of Guangdong science and technology innovation strategy “Research on VR/AR key technologies in online commodity display in the post epidemic era” (pdjh2021a0741).

REFERENCES

[1] 

Chen, Y. C., He, H. W., Chen, H. E., et al., “Improving registration of augmented reality by incorporating DCNNS into Visual SLAM,” International Journal of Pattern Recognition & Artificial Intelligence, 32 (12), 1855022 (2018). https://doi.org/10.1142/S0218001418550224 Google Scholar

[2] 

Wu, Y. M., He, H. W., Sun, J., et al., “Vision-based hand tracking and gesture recognition for augmented assembly system,” Key Engineering Materials, 392 (13), 1030 –1036 (2009). Google Scholar

[3] 

David, P., DeMenthon, D., Duraiswami, R., et al., “Simultaneous pose and correspondence determination using line features,” in IEEE Computer Society Conference on Computer Vision & Pattern Recognition, 1211499 (2003). Google Scholar

[4] 

Chandra, A., Manojkumar, P. C., Tharun, M. B., et al., “Object detection, 3-D coordinate extraction and pose estimation for an autonomous medical robot,” in 2021 5th Inter. Conf. on Computer, Communication and Signal Processing (ICCCSP), 9465501 (2021). Google Scholar

[5] 

Liu, D., Arai, S., Xu, Y., et al., “6D Pose estimation of occlusion-free objects for robotic bin-picking using PPF-MEAM With 2D images (Occlusion-Free PPF-MEAM) IEEE Access,” 9 50857 –50871 (2021). Google Scholar

[6] 

Zhu, J., Pan, Z., Chao, S., et al., “Handling occlusions in video-based augmented reality using depth information,” Computer Animation & Virtual Worlds, 21 (5), 509 –521 (2021). https://doi.org/10.1002/cav.v21:5 Google Scholar

[7] 

Zhu, M., Derpanis, K. G., Yang, Y., et al., “Single image 3D object detection and pose estimation for grasping,” in IEEE Inter. Conf. on Robotics & Automation, 3936 –3943 (2014). Google Scholar

[8] 

Shu, C., Liang, L., Liang, W., et al., “3D pose tracking with multi-template warping and SIFT correspondences,” IEEE Transactions on Circuits & Systems for Video Technology, 26 (99), 1 –10 (2015). Google Scholar

[9] 

Mkhoyan, T., Visser, C. and Breuker, R. D., “Parallel real-time tracking and 3D reconstruction with TBB for intelligent control and smart sensing framework,” AIAA Scitech 2020 Forum, 2020 –2252 (2020). Google Scholar

[10] 

Girshick, R., “Fast R-CNN,” in 2015 IEEE Inter. Conf. on Computer Vision (ICCV), 1440 –1448 (2015). Google Scholar

[11] 

He, K., Gkioxari, G., Dollar, P., et al., “Mask R-CNN,” in 2017 IEEE Inter. Conf. on Computer Vision (ICCV), (2017). https://doi.org/10.1109/ICCV.2017.322 Google Scholar
© (2022) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Teng Zhu, Xiaoyin Guo, Hansheng Yan, Yongbin Chen, and Bin Shu "Research on pose and position estimation based on mask RCNN edge extraction for AR auxiliary assembly system", Proc. SPIE 12506, Third International Conference on Computer Science and Communication Technology (ICCSCT 2022), 125066F (28 December 2022); https://doi.org/10.1117/12.2661820
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Autoregressive models

Augmented reality

Feature extraction

Detection and tracking algorithms

3D modeling

Virtual reality

Cameras

Back to Top