Open Access Paper
2 February 2023 Assembly training system on HoloLens using embedded algorithm
Yujin Qin, Shuxia Wang, Qiang Zhang, Yao Cheng, Jiaxu Huang, Weiping He
Author Affiliations +
Proceedings Volume 12462, Third International Symposium on Computer Engineering and Intelligent Communications (ISCEIC 2022); 124620H (2023) https://doi.org/10.1117/12.2660940
Event: International Symposium on Computer Engineering and Intelligent Communications (ISCEIC 2022), 2022, Xi'an, China
Abstract
In this article, we demonstrate an implementation on Microsoft HoloLens, deep learning supported in the context of object detection. The main aim of the training system is to create the more accurate object detection model for augmented reality using deep learning models for image recognition directly on the HoloLens 2. In terms of the object detection approach, a deep learning model called YOLOv5 has been used for the implementation of this system. This article uses the Windows ML API to implement machine learning in augmented reality applications. A simple and easy method of drawing lines between specified 2D coordinates on a canvas is proposed. The module division and development steps of the development of augmented reality training system are given. Our system provides the annotation of augmented object detected and its bounding box via HoloLens. It allows to detect the new object in a few milliseconds. Preliminary results show a great rate of object detection and reasonable detection time.

1.

INTRODUCTION

In the era of intelligent manufacturing, the digital, The high-performance assembly of major and complex products such as aero-engines has attracted extensive research by domestic and foreign scholars. Due to the large number of parts and the complicated information of paper process documents, the stationary rocket engine is mainly assembled manually at present. In order to ensure the quality and efficiency of engine assembly, assembly training and cognitive operation of new employees are very important. However, the cognitive training of aero-engine assembly mainly adopts the traditional mode of “ experienced people teach newbies ”, which requires large labor costs and learning cycles.

With the updating of computer software and hardware, mixed reality technology has begun to be widely used in education, industry, medical and other industries. Mixed reality technology which can deeply integrate information of different dimensions breaks through traditional space limitations and interaction methods. Mixed reality includes virtual reality and augmented reality. The simulation training system developed by virtual reality technology [1] has the characteristics, such as immersive learning features, high security and so on. However, most of the existing virtual training systems can only achieve independent interactive learning, which cannot meet the needs of operators for scene perception. Augmented reality solves the above problems by ingeniously integrating virtual objects with real scenes. In addition, object detection has good results in object recognition of assembly parts and perception of assembly environment. The integrated training system which uses technologies related to augmented reality and object detection develops a new way of human-computer interaction。It can provide theoretical and technical support for problems existing in the learning training and visual guidance.

This article designs and develops a training system for solid rocket motor mandrel based on augmented reality and deep learning. The system solves the problem of long training time and low efficiency in on-site assembly and training of complex products such as aero-engines. Moreover, we achieve a high degree of fusion between machine learning models and augmented reality devices, which significantly improves field perception in assembly cognitive training.

2.

RELATED WORKS

At present, the target detection based on deep learning has surpassed the traditional target detection algorithm, and has greatly improved the accuracy and recognition efficiency. Deep learning algorithms for object detection include CNN, RCNN, fast-RCNN, and YOLO, etc. The second generation YOLO2 of the YOLO algorithm in 2017: Better, Faster, Stronger [2], which realizes the recognition of various types of objects. The original neural network structure is improved by replacing the fully connected layer with the convolutional layer. In 2018, Redmon proposed YOLOv3 [3], which uses the FPN structure for multi-scale feature extraction. In order to improve the detection effect of mAP and objects, the algorithm replaces Softmax with Logistic to form a deeper network level and multi-scale detection. In 2021, Li Yangfan et al. [4] proposed an improved YOLOv4 spatial infrared weak target detection method, with an average recognition accuracy rate of over 93.25% and a detection speed of 38.99ms/frame. In 2021, Song Xin of Dalian University of Technology [5] used the X-YOLOv3 network structure to identify parts of gears and bearings in the workshop in a simulated workshop. The results show that this method can reduce the work pressure of operators.

AR represents an innovative tool that can ensure the assembly efficient and correct transfer of knowledge. HoloLens device is recognized as the best solution for augmented reality applications, and image recognition and processing has become an inevitable trend in the augmented reality environment. In 2016, Zhang Lijuan [6] of Huazhong University of Science and Technology combined mobile phones and computers, and introduced augmented reality technology into the equipment maintenance system of the Expo Center. This method greatly facilitates the management of the equipment by the staff. In 2017, Zhang Shupeng[7] applied AR virtual interaction technology to the work of knowledge popularization, which effectively improved the learning efficiency of users. In 2018, Feng Danlin[8] of the Civil Aviation University of China used HoloLens to develop a mixed reality interactive demonstration system for civil aviation engines, which can intuitively understand the structure and principles of the engine. However, this interactive system is only a virtual engine model and cannot be used in combination with real engines and virtual engines. In 2019, Derakhshani et al. [9] proposed a technique of adding local activation, which effectively improves the MAP of the YOLO detector without affecting the computational speed. In 2020, Zhou Xiang of Fuzhou University and others [10] proposed an augmented reality system based on the image recognition BRISK-SURF algorithm. The system has high recognition accuracy for transformations such as illumination and occlusion, which further improves the computing speed of the system. In 2021, Zeng Xiao of Sichuan University [11] proposed a target detection system based on TCP/IP remote communication. The system breaks through the traditional interaction method, but there are problems of high information delay and data loss.

In order to further reduce the information delay in the interaction process and improve the on-site perception in the assembly process, this paper uses the YOLO algorithm to develop a solid rocket motor mandrel on-site training system. The system implements the independent target detection function of HoloLens2 through deep learning algorithm processing on the Microsoft HoloLens device. The training system can update the enhanced information corresponding to the assembly parts according to the detection results, so that the user can interact more intuitively.

3.

SYSTEM DESIGN

This section will describe the main working process of the system, the overall scheme of the solid rocket motor mandrel field training system is shown in the figure1. In terms of hardware, this article uses the processor of Microsoft HoloLens2 (Qualcomm Snapdragon 850) to perform the calculation of the machine model; Meanwhile, the HoloLens2 camera acts as an input module for object detection in the real world. The camera mounted on the front of the device allows the app to intuitively detect what the user is seeing. Developers can obtain device metadata by accessing and manipulating the camera. In the training system, we need access to the HoloLens research mode streams, as well as OpenCV to enable real-time processing of sensor frames streamed from the Microsoft HoloLens device in a C# Unity environment using IL2CPP Windows Runtime Support.

Figure 1.

Software and hardware system diagram.

00017_PSISDG12462_124620H_page_2_1.jpg

The training system construction process mainly includes two stages: machine learning model pre-training and system detection and recognition.

In terms of machine learning model pre-training, we use the automatic acquisition program we created to collect and frame the on-site assembly parts firstly. Secondly, performing Mosaic data augmentation on raw images and creating a dataset. This operation can enrich the background of the detected object and improve the detection effect of small target parts. Finally, we use the YOLOv5 network to perform transfer learning and training on the features of the assembled parts.

The operation process of the training system is shown in Figure 2. Firstly, capturing the HoloLens2 media frame and using the algorithm to obtain the latest video frame. Then we pass the video frame into the main program and wait for the asynchronous command. Secondly, the main program loads the pre-trained machine learning model and label files. The latest video frame is evaluated as the network input, and the network prediction results are output. Finally, according to the prediction results, the UI interface guidance information of the HoloLens holographic image is updated.

Figure 2.

System Structure Diagram.

00017_PSISDG12462_124620H_page_3_1.jpg

4.

SYSTEM COMPONENTS

In order to facilitate the subsequent upgrade of the training system, according to the idea of modular design, the on-site training system of the solid rocket motor core mold in this paper can be divided into four parts: 1: Model training based on YOLOv5n. 2: Integrating machine learning models into UWP applications. 3: Capture of HoloLens media frames. 4: the display and update of GUI interface.

4.1

Model training based on YOLOv5n

The data set is made by using the pictures collected on site, and the model is trained on the parts data set. In order to run better on HoloLens devices, this article uses the pre-trained weights of the lightweight model YOLOv5n, with a size of 3.9MB. Secondary training of YoloV5N-CBAM model is carried out on solid rocket motor parts dataset. The relevant parameters in model training are set as follows:

Table 1.

Relevant parameters of model secondary training.

ItemParameterItemParameter
Img_size640*640epoch200
Batch_size4momentum0.937
Class121r00.01
Warmuo_epochs3.0Warmup_bias_1r0.1

In the table, the ing_size represents the input image size; epoch represents the number of model iterations; batch_size represents the number of images for each training of the model. Momentum means bias correction for the optimizer, class refers to the category of engine parts, 1r0 represents the initial learning rate, warmup_epochs represents the warmup training period, and warmup_bias_lr represents the learning rate for the warmup training period. In the process of model training, the performance of the computer graphics card is directly related to the number of images trained in each iteration and the training time.

Figure 3.

Loss function change diagram.

00017_PSISDG12462_124620H_page_4_1.jpg

It can be seen from the figure that the frame regression loss value in the training set has the fastest decline rate in the warm-up learning cycle. As the number of iterations increases, the decreasing speed of the frame regression loss value gradually slows down until it becomes stable. The initial value of the target detection loss value is very small, and it slowly decreases with the increase of epoch until it falls below 0.01 and then slowly converges. The classification loss value only needs 20 epoch to start to converge. In the validation set, the three types of loss values decreased rapidly and gradually became stable in the first 20 epochs. The model training ends after 200 epochs, and we get 3.7MB of weight data for both.

4.2

Integrating machine learning models into UWP applications

This module converts a pre-trained machine learning model to an Open Neural Network Exchange (ONNX) model via the tf2onnx tool and adds it for integration into the application. Finally, an executable application is generated and deployed to the HoloLens device, as shown in Figure 4.

Figure 4.

Integrated development process.

00017_PSISDG12462_124620H_page_4_2.jpg

Windows ML is a high-performance and reliable API, which is used for hardware-accelerated ML inference on Windows devices. Therefore, this article uses the Windows Machine Learning API to deploy TensorFlow models in Windows applications. The process includes: (1) We can access the ONNX model through the LoadModel method and store it in memory; (2)Convert the incoming original image to VideoFrame format; (3)We need create a session for binding the model, and bind the model to the specified device, so that the model can be executed on the specific device of the computer.A machine learning model has input and output features that are used to pass information into and out of the model. (4) Evaluate the model and communicate its predictions.

4.3

Capturing of HoloLens media frames

This module uses MediaFrameReader in conjunction with MediaCapture to obtain media frames from one of the available sources, including color, depth, infrared cameras, etc. This function is designed to provide raw input to programs that process media frames in real time.

Use the MediaCapture class to capture photos and videos, and use MediaFrameReader to process media frames. Since the captured image needs to be used immediately in the app, the image should be captured into a SoftwareBitmap. Finally, creating a method keeps getting the latest video frame from the media frame reader.

4.4

The display and update of GUI interface

After obtaining the prediction results of the deep learning model for the assembly parts, this article uses HoloLens to interactively display the prediction results. The display of the prediction results of the parts can further improve the operation effect of the training system. In order to achieve this purpose, it is first necessary to use HoloLens to determine the boundary area of the recognized object. Subsequently, data is indexed in the unique assembly parts material library so that identification information can be updated in real time.

Since there is no appropriate method in Unity to draw a rectangle (or bounding box) on a canvas, this section extends the Texture2D methods in Unity. We use the Unity Vector2.Lerp method, which is used to linearly interpolate between two vectors by a specified increment. Therefore, this method achieve the aim of drawing lines between specified 2D coordinates on a canvas and construct a bounding box, complete with label and confidence. Among them, the incremental frac of linear interpolation can be expressed as:

00017_PSISDG12462_124620H_page_5_1.jpg

During the generation of the bounding box, HoloLens utilizes the embedded processor to index a library of materials specific to the assembly components. The material library includes attribute information for assembly components, assembly guide information, and associated parts. After obtaining the specific object information, the label and UI information is updated, thereby improving the interaction efficiency during the training process and expanding the dimension of the information.

5.

EXPERIMENTS AND RESULTS

5.1

Recall and Precision

In order to evaluate the recognition effect and operating efficiency of the training system, this article evaluates the recognition effect of assembly parts through the recall rate and precision rate of the training model. The recall rate and precision rate can be expressed as:

00017_PSISDG12462_124620H_page_5_2.jpg
00017_PSISDG12462_124620H_page_5_3.jpg

First, sort the n test results in descending order of confidence. Second, TP means that the correct part recognition frame is predicted. Among them, the FN array represents the positive samples that are predicted to be negative, and the FP array represents the negative samples which are predicted to be positive.

Since there are many parts involved in the training system in this paper, the AP value of each part is calculated by Formula (4) to judge the recognition efficiency of each part. The AP value of a certain category can be expressed as:

00017_PSISDG12462_124620H_page_6_1.jpg

The map which is mean of AP for all categories can be expressed as:

00017_PSISDG12462_124620H_page_6_2.jpg

Figure 5.

Change trend graph of mAP_0.5 and mAP_0.5:0.95 during training.

00017_PSISDG12462_124620H_page_6_3.jpg

The training results in this article are shown in Figure5. As the number of iterations increases, the values of mAP_0.5 and mAP_0.5:0.95 increase rapidly in the first 20 iteration step cycles, and eventually they tend to be stable. As can be seen, the maximum value of mAP_0.5 is finally stable at 0.98. At the same time, the maximum value of mAP_0.5:0.95 is stable at around 0.82. The results of each category of Recall and Precision obtained after training are shown in Table 2.

Table 2.

Recall and precision for each object.

ClassificationRecallPrecision
Exhaust78.2%96.8%
Mandrel I71.9%93.5%
Mandrel II62.5%95.6%
Upperpanel61.4%96.8%
Bolt75.9%98.2%
Shell86.3%97.3%
Headring69.4%92.1%
Downpanel85.1%93.1%
Frock76.8%98.6%

Further, the detection time of the system directly reflects the sensitivity and overall performance of the training system. This article deploys and runs the system on the HoloLens2 device. To use the Unity timer function to systematically time the inference process of the experimental network model. Under the current enhanced equipment conditions, the final detection time of the integrated system is shown in Table 3.

Table 3.

Detection time for each object.

ClassificationDetection time (/s)
Exhaust0.362
Mandrel I0.387
Mandrel II0.420
Upperpanel0.298
Bolt0.421
Shell0.364
Headring0.358
Downpanel0.402
Frock0.319

In order to test the robustness of the system, this paper conducts ensemble experiments on 1000 kinds of machine learning models. The experimental results s how that the average detection time is about 0.475s. This time indicates that the method performs well for a certain number of categories.

6.

CONCLUSION AND FUTURE WORK

In this paper, we built an augmented reality system that supports object detection and visual human-computer interaction by combining the latest augmented reality device HoloLens2 and the advanced target detection algorithm YOLOv5. This method embeds the machine learning model into the HoloLens device through the Windows ML API. The experimental results show that the method proposed in this paper has high recall rate and high precision rate on target detection, and the accuracy rate on the test set and the average accuracy rate of all categories have better performance. At the same time, using the HoloLens2 augmented reality helmet gets rid of the traditional human-computer interaction methods and provides theoretical and technical support for visual display. However, this paper lacks control experiments in the design of experiments, and the changes to the algorithm itself are small. The experiments will continue to be improved in future work.

REFERENCES

[1] 

Jiang, L., “Development of Virtual Reality Training System,” Virtual, Online, Association for Computing Machinery, Singapore (2021). Google Scholar

[2] 

Redmon, Joseph, and Ali Farhadi, “YOLO9000: better, faster, stronger,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 7263 –7271 (2017). Google Scholar

[3] 

Redmon, Joseph, and Ali Farhadi, “Yolov3: An incremental improvement,” arXiv preprint arXiv:1804.02767, (2018). Google Scholar

[4] 

Yangfan Liu, Lihua Cao, Ning Li,Yunfeng Zhang, “Space infrared weak target detection based on YOLOv4 [J],” LCD and Display, 36 (04), 615 –623 (2021). Google Scholar

[5] 

Xin Song, Research on Parts Recognition in Simulated Workshop Based on Convolutional Neural Network [D], (2021). Google Scholar

[6] 

Lijuan Chen, Hanbin Luo, Hongyan Xin, “Application of Augmented Reality Technology in Equipment Maintenance System of Wuhan International Expo Center [J],” Construction Technology, 45 (06), 37 –40 (2016). Google Scholar

[7] 

Shupeng Zhang, Wenjunn Hou, Ximeng Wang, “Research on the design of popular science knowledge learning method based on augmented reality virtual-real interaction [J],” Packaging Engineering, 38 (20), 48 –55 (2017). Google Scholar

[8] 

Danlin Feng, “Development of mixed reality interactive demonstration system for civil aviation engine based on HoloLens [J],” Digital Technology & Application, 36 (11), 152 –154+156 (2018). Google Scholar

[9] 

Derakhshani, Mohammad Mahdi, Saeed Masoudnia, Amir Hossein Shaker, Omid Mersa, Mohammad Amin Sadeghi, Mohammad Rastegari, and Babak N.Araabi, “Assisted Ex-citation of Activations: A Learning Technique to Improve Object Detectors,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 9201 –9210 (2019). Google Scholar

[10] 

Xiang Zhou, Lilin Tang, Ding Lin, “Natural feature virtual-real registration method based on binary robust invariant scale keypoint-accelerated robust feature[J],” Journal of Computer Applications, 40 (05), 1403 –1408 (2020). Google Scholar

[11] 

Xiao Zeng, “Research on target detection technology based on HoloLens2 [J],” Computer software and computer applications, (14), 92 –95 (2021). Google Scholar
© (2023) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Yujin Qin, Shuxia Wang, Qiang Zhang, Yao Cheng, Jiaxu Huang, and Weiping He "Assembly training system on HoloLens using embedded algorithm", Proc. SPIE 12462, Third International Symposium on Computer Engineering and Intelligent Communications (ISCEIC 2022), 124620H (2 February 2023); https://doi.org/10.1117/12.2660940
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Augmented reality

Machine learning

Data modeling

Instrument modeling

Target detection

Back to Top