In optoelectronic systems, infrared target tracking is a critical function. Due to occlusion causing template drift in infrared target tracking, correlation filtering algorithms have poor performance for infrared targets. Although target trackers based on Siamese convolutional neural networks exhibit excellent tracking performance, their complex architecture and high computational complexity hinder their real-time application on embedded chips. Furthermore, the output of anchor-based trackers is often unstable, which can be detrimental to the closed-loop control of optoelectronic devices. Therefore, this paper proposes a tracker based on the Siamese network for infrared small target tracking and presents an innovative lightweight approach to enhance real-time performance while minimizing accuracy loss. With regards to different backbone network structures, the computational complexity required for embedded computing was initially analyzed. Subsequently, artificial intelligence interpretability methods were employed to assess the performance of distinct networks, select the optimal backbone network, and ultimately striking a balance between accuracy and speed. Ultimately, real-time operation is achieved on embedded devices.
Tracking specific objects in images or videos is one of the most attractive problems in visual tasks. It is widely employed in security monitoring, automatic driving, military operations and other scenes. Recently, object tracker based on convolution neural network, especially Siamese network, obtains high accuracy and has been deeply studied. However, in practical application scenarios of visual tracking, when meets clutter background or the object is occluded, the accuracy of the tracking task will drop rapidly, and the tracker loses the target in extreme cases. It is particularly necessary to quickly and accurately relocate the target. Therefore, an anti-interference tracker based on Siamese convolution neural network is developed. Benefiting from the adaptive tracking confidence parameter, once the tracking effect of the tracker has dropped significantly during the tracking process, the location of the object will be corrected immediately. Experimental results show that the proposed method has the ability to relocate and track the target after occlusion or loss effectively.
As an important application of computer vision, visual tracking has being a fundamental topic. Compared with visible image, infrared image has the characteristic of low resolution, blurred contour and single color feature. Thus, it is still a challenge for infrared object tracking. Further, it is difficult to balance the real-time performance and accuracy. This paper proposed a tracker based on the SiamRPN tracker with a deeper and lightweight MobileNet V2 structure as the backbone network. During network training, the weights are updated by the scheme of the model-independent metalearning method. The computed model only passes a few gradient descents on the first frame can obtains the most suitable for the current frame. In the end, the tracker is tested on various datasets. Experimental results show that the tracker can achieve a balance between tracking accuracy and inference speed, which is crucial for deployment on mobile devices.
Visual object tracking is one of the most attractive issue in computer vision. Recently, deep neural network has been widely developed in object tracking and showing great accuracy. In general, the accuracy of tracking task decreases dramatically when the background becomes complex or occluded. Thus, a robust tracking method based on convolutional neural network and anti-occlusion mechanic is presented. Benefit from the adaptive tracking confidence parameter T, the tracking effect is evaluated during tracking. Once the target is occluded, the location of the target object is corrected immediately. Experimental results demonstrate that the proposed framework achieves state-of-the-art performance.
Dynamic range compression and contrast enhancement are the key steps of infrared imaging. Reasonable dynamic range compression should not destroy the gray distribution relationship between adjacent pixels. Most of the existing dynamic range compression algorithms do not take maintaining the gray distribution relationship between adjacent pixels as the basic principle of algorithm design. After dynamic range compression, the gray distribution of adjacent pixels can not be consistent with that before compression, which may lead to gradient reversal, edge halo, and some algorithms have the problem that the whole image is smooth, but the details are lost seriously. An infrared image dynamic range compression algorithm with the characteristics of neighborhood gra y distribution preservation is proposed based on the principle of keeping the gray distribution of neighboring pixels. The algorithm is based on the commonly used segmented linear transformation algorithm. In order to minimize the loss of image details in dynamic range compression, local factors are introduced into the global transformation to reduce the loss of overall image details. The specific method is to add the description operator of gray distribution of adjacent pixels in the calculation of transformation parameters. The algorithm effectively improves the image details, and can obtain good display effect for the original infrared image with high dynamic range. The experimental results show that the algorithm is better than the segmented linear transformation algorithm in displaying the original infrared image with high dynamic range.
Visual object tracking is one of the most attractive issue in computer vision. Recently, deep neural network has been widely developed in object tracking and showing great accuracy. In general, the accuracy of tracking task decreases dramatically when the background becomes complex or occluded. Here, we propose an end-to-end lightweight Siamese convolution neural network to achieve fast and robust target tracking especially for infrared target. The network structure replaces the hand-crafted features by the multi-layers deep convolution features of the target, so that higher precision can be achieved. Specifically, object location is updated in every frame by refreshing a response-map. However, the success rate of tracking task decreases dramatically when the background becomes complex or occluded. Consequently, a simple and robust anti-occlusion tracking method is presented. The tracking accuracy is evaluated during tracking process by computing the tracking confidence parameters. The parameters are composed of two parts: target confusion degree which indicates the degree of background interference and target occlusion degree which indicates the degree of target occlusion. Once the target is occluded, the location of the target object is corrected immediately. Experimental results demonstrate that the proposed framework achieves state-of-the-art performance on the popular OTB50 and OTB100 benchmarks.
A panoramic surveillance system is designed to achieve continuous monitoring of the surrounding environment. The image acquisition module of the system is composed of five fixed-focal-length cameras and one variable-focal-length camera, which realizes 360 degree environmental surveillance. An adaptive threshold is used to dynamically update the background template in order to better accommodate various weather changes. Further, a pixel-level video moving target detection algorithm is applied to effectively detect whether an intruding target exists and determine the direction of the target. It shows the advantages of less computation and preferable detection accuracy. Once an intrusive target is found, the deep convolution neural network SSD is employed to recognize the specific target quickly. As common sense, visual object tracking is one of the most attractive issue in computer vision. Recently, deep neural network has been widely developed in object tracking and shown great achievement. Here, we propose an end-to-end lightweight siamese convolution neural network to achieve fast and robust target tracking. The experiment result shows panoramic surveillance system can effectively and robustly perform security tasks such as panoramic imaging, target recognition and fast target tracking. At the same time, the deep convolution neural network can recognize and track the target accurately and quickly, which meets the real-time and accuracy requirements of practical task.
Prevailing object detection algorithms such as RCNN, YOLO and SSD usually are not suitable for high definition surveillance systems because of the fixed size of network input and masses of object candidate regions in the select search process. This paper proposes a fast moving object detection and recognition method for video surveillance system, which applies background extraction and frame difference to fulfill select search process, followed by a pretrained CNN model inference to complete object recognition. Proposed method was proved to be fast and effective in our experimental results which is more suitable for moving object detection for video surveillance system, compared to current other object detection algorithms.
A panoramic monitoring system is designed to achieve continuous monitoring of the surrounding environment. The image acquisition module of the system is composed of five fixed-focal-length cameras and a variable-focal-length camera, which realizes 360 degree environmental monitoring. Usually, the background of continuous photography changes due to fluctuations of ambient light, humidity and wind. Therefore, a dynamic adaptive threshold is used to dynamically update the background template in order to better accommodate various weather changes. Further, a motion-aware algorithm based on background updates is applied to effectively detect whether an intruding target exists and determine the direction of the target. Once an intrusive target is found, the deep convolution neural network Yolo is employed to recognize the target quickly. It shows the advantages of less computation and preferable detection accuracy. In addition, according to the preset warning level, when the intrusion target needs to be alarmed, the target orientation is transmitted to the platform through the central control processing unit, so that the variable-focal-length camera can take real-time snapshots. we propose an end-to-end lightweight siamese convolution neural network to achieve fast and robust target tracking. The network structure replaces the hand-crafted features by the multi-layers deep convolution features of the target, so that higher precision can be achieved. The experiment result shows panoramic surveillance system can effectively and robustly perform security tasks such as panoramic imaging, target recognition and fast target tracking.
Visual object tracking is one of the most attractive issue in computer vision. Recently, deep neural network has been widely developed in object tracking and showing great accuracy. In general, the accuracy of tracking task decreases dramatically when the background becomes complex or occluded. Thus, a robust tracking method based on convolutional neural network and anti-occlusion mechanic is presented. Benefit from the adaptive tracking confidence parameter T, the tracking effect is evaluated during tracking. Once the target is occluded, the location of the target object is corrected immediately. Experimental results demonstrate that the proposed framework achieves state-of-the-art performance on the popular OTB50 and OTB100 benchmarks.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.