1
|
Liu F, Gao C, Chen F, Meng D, Zuo W, Gao X. Infrared Small and Dim Target Detection With Transformer Under Complex Backgrounds. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2023; 32:5921-5932. [PMID: 37883292 DOI: 10.1109/tip.2023.3326396] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/28/2023]
Abstract
The infrared small and dim (S&D) target detection is one of the key techniques in the infrared search and tracking system. Since the local regions similar to infrared S&D targets spread over the whole background, exploring the correlation amongst image features in large-range dependencies to mine the difference between the target and background is crucial for robust detection. However, existing deep learning-based methods are limited by the locality of convolutional neural networks, which impairs the ability to capture large-range dependencies. Additionally, the S&D appearance of the infrared target makes the detection model highly possible to miss detection. To this end, we propose a robust and general infrared S&D target detection method with the transformer. We adopt the self-attention mechanism of the transformer to learn the correlation of image features in a larger range. Moreover, we design a feature enhancement module to learn discriminative features of S&D targets to avoid miss-detections. After that, to avoid the loss of the target information, we adopt a decoder with the U-Net-like skip connection operation to contain more information of S&D targets. Finally, we get the detection result by a segmentation head. Extensive experiments on two public datasets show the obvious superiority of the proposed method over state-of-the-art methods, and the proposed method has a stronger generalization ability and better noise tolerance.
Collapse
|
2
|
Zhang N, Liu W, Xia X. Video Global Motion Compensation Based on Affine Inverse Transform Model. SENSORS (BASEL, SWITZERLAND) 2023; 23:7750. [PMID: 37765806 PMCID: PMC10534421 DOI: 10.3390/s23187750] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/21/2023] [Revised: 09/02/2023] [Accepted: 09/05/2023] [Indexed: 09/29/2023]
Abstract
Global motion greatly increases the number of false alarms for object detection in video sequences against dynamic backgrounds. Therefore, before detecting the target in the dynamic background, it is necessary to estimate and compensate the global motion to eliminate the influence of the global motion. In this paper, we use the SURF (speeded up robust features) algorithm combined with the MSAC (M-Estimate Sample Consensus) algorithm to process the video. The global motion of a video sequence is estimated according to the feature point matching pairs of adjacent frames of the video sequence and the global motion parameters of the video sequence under the dynamic background. On this basis, we propose an inverse transformation model of affine transformation, which acts on each adjacent frame of the video sequence in turn. The model compensates the global motion, and outputs a video sequence after global motion compensation from a specific view for object detection. Experimental results show that the algorithm proposed in this paper can accurately perform motion compensation on video sequences containing complex global motion, and the compensated video sequences achieve higher peak signal-to-noise ratio and better visual effects.
Collapse
Affiliation(s)
- Nan Zhang
- School of Electrical and Control Engineering, Shaanxi University of Science and Technology, Xi’an 710021, China;
| | - Weifeng Liu
- School of Electrical and Control Engineering, Shaanxi University of Science and Technology, Xi’an 710021, China;
| | - Xingyu Xia
- School of Automation, Hangzhou Dianzi University, Hangzhou 310018, China;
| |
Collapse
|
3
|
Melville-Smith A, Finn A, Uzair M, Brinkworth RSA. Exploration of motion inhibition for the suppression of false positives in biologically inspired small target detection algorithms from a moving platform. BIOLOGICAL CYBERNETICS 2022; 116:661-685. [PMID: 36305942 PMCID: PMC9691501 DOI: 10.1007/s00422-022-00950-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/20/2021] [Accepted: 10/14/2022] [Indexed: 06/16/2023]
Abstract
Detecting small moving targets against a cluttered background in visual data is a challenging task. The main problems include spatio-temporal target contrast enhancement, background suppression and accurate target segmentation. When targets are at great distances from a non-stationary camera, the difficulty of these challenges increases. In such cases the moving camera can introduce large spatial changes between frames which may cause issues in temporal algorithms; furthermore targets can approach a single pixel, thereby affecting spatial methods. Previous literature has shown that biologically inspired methods, based on the vision systems of insects, are robust to such conditions. It has also been shown that the use of divisive optic-flow inhibition with these methods enhances the detectability of small targets. However, the location within the visual pathway the inhibition should be applied was ambiguous. In this paper, we investigated the tunings of some of the optic-flow filters and use of a nonlinear transform on the optic-flow signal to modify motion responses for the purpose of suppressing false positives and enhancing small target detection. Additionally, we looked at multiple locations within the biologically inspired vision (BIV) algorithm where inhibition could further enhance detection performance, and look at driving the nonlinear transform with a global motion estimate. To get a better understanding of how the BIV algorithm performs, we compared to other state-of-the-art target detection algorithms, and look at how their performance can be enhanced with the optic-flow inhibition. Our explicit use of the nonlinear inhibition allows for the incorporation of a wider dynamic range of inhibiting signals, along with spatio-temporal filter refinement, which further increases target-background discrimination in the presence of camera motion. Extensive experiments shows that our proposed approach achieves an improvement of 25% over linearly conditioned inhibition schemes and 2.33 times the detection performance of the BIV model without inhibition. Moreover, our approach achieves between 10 and 104 times better detection performance compared to any conventional state-of-the-art moving object detection algorithm applied to the same, highly cluttered and moving scenes. Applying the nonlinear inhibition to other algorithms showed that their performance can be increased by up to 22 times. These findings show that the application of optic-flow- based signal suppression should be applied to enhance target detection from moving platforms. Furthermore, they indicate where best to look for evidence of such signals within the insect brain.
Collapse
Affiliation(s)
- Aaron Melville-Smith
- Defense and Systems Institute, UniSA STEM, University of South Australia, Adelaide, SA 5095 Australia
| | - Anthony Finn
- Defense and Systems Institute, UniSA STEM, University of South Australia, Adelaide, SA 5095 Australia
| | - Muhammad Uzair
- Defense and Systems Institute, UniSA STEM, University of South Australia, Adelaide, SA 5095 Australia
| | | |
Collapse
|
4
|
Abstract
In industrial production, accidents caused by the unsafe behavior of operators often bring serious economic losses. Therefore, how to use artificial intelligence technology to monitor the unsafe behavior of operators in a production area in real time has become a research topic of great concern. Based on the YOLOv5 framework, this paper proposes an improved YOLO network to detect unsafe behaviors such as not wearing safety helmets and smoking in industrial places. First, the proposed network uses a novel adaptive self-attention embedding (ASAE) model to improve the backbone network and reduce the loss of context information in the high-level feature map by reducing the number of feature channels. Second, a new weighted feature pyramid network (WFPN) module is used to replace the original enhanced feature-extraction network PANet to alleviate the loss of feature information caused by too many network layers. Finally, the experimental results on the self-constructed behavior dataset show that the proposed framework has higher detection accuracy than traditional methods. The average detection accuracy of smoking increased by 3.3%, and the average detection accuracy of not wearing a helmet increased by 3.1%.
Collapse
|
5
|
Domingos LCF, Santos PE, Skelton PSM, Brinkworth RSA, Sammut K. A Survey of Underwater Acoustic Data Classification Methods Using Deep Learning for Shoreline Surveillance. SENSORS 2022; 22:s22062181. [PMID: 35336352 PMCID: PMC8954367 DOI: 10.3390/s22062181] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Subscribe] [Scholar Register] [Received: 12/22/2021] [Revised: 02/06/2022] [Accepted: 02/09/2022] [Indexed: 02/04/2023]
Abstract
This paper presents a comprehensive overview of current deep-learning methods for automatic object classification of underwater sonar data for shoreline surveillance, concentrating mostly on the classification of vessels from passive sonar data and the identification of objects of interest from active sonar (such as minelike objects, human figures or debris of wrecked ships). Not only is the contribution of this work to provide a systematic description of the state of the art of this field, but also to identify five main ingredients in its current development: the application of deep-learning methods using convolutional layers alone; deep-learning methods that apply biologically inspired feature-extraction filters as a preprocessing step; classification of data from frequency and time–frequency analysis; methods using machine learning to extract features from original signals; and transfer learning methods. This paper also describes some of the most important datasets cited in the literature and discusses data-augmentation techniques. The latter are used for coping with the scarcity of annotated sonar datasets from real maritime missions.
Collapse
Affiliation(s)
- Lucas C. F. Domingos
- Department of Electrical and Electronics Engineering, Centro Universitário FEI, Sao Bernardo do Campo 09850-901, SP, Brazil;
- Department of Computer Vision, Instituto de Pesquisas Eldorado, Campinas 13083-898, SP, Brazil
- Correspondence:
| | - Paulo E. Santos
- Department of Electrical and Electronics Engineering, Centro Universitário FEI, Sao Bernardo do Campo 09850-901, SP, Brazil;
- Centre for Defence Engineering Research and Training, College of Science and Engineering, Flinders University, Tonsley, SA 5042, Australia; (P.S.M.S.); (R.S.A.B.); (K.S.)
| | - Phillip S. M. Skelton
- Centre for Defence Engineering Research and Training, College of Science and Engineering, Flinders University, Tonsley, SA 5042, Australia; (P.S.M.S.); (R.S.A.B.); (K.S.)
| | - Russell S. A. Brinkworth
- Centre for Defence Engineering Research and Training, College of Science and Engineering, Flinders University, Tonsley, SA 5042, Australia; (P.S.M.S.); (R.S.A.B.); (K.S.)
| | - Karl Sammut
- Centre for Defence Engineering Research and Training, College of Science and Engineering, Flinders University, Tonsley, SA 5042, Australia; (P.S.M.S.); (R.S.A.B.); (K.S.)
| |
Collapse
|
6
|
Fang J, Finn A, Wyber R, Brinkworth RSA. Acoustic detection of unmanned aerial vehicles using biologically inspired vision processing. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2022; 151:968. [PMID: 35232118 DOI: 10.1121/10.0009350] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/20/2021] [Accepted: 01/04/2022] [Indexed: 06/14/2023]
Abstract
Robust detection of acoustically quiet, slow-moving, small unmanned aerial vehicles is challenging. A biologically inspired vision approach applied to the acoustic detection of unmanned aerial vehicles is proposed and demonstrated. The early vision system of insects significantly enhances signal-to-noise ratios in complex, cluttered, and low-light (noisy) scenes. Traditional time-frequency analysis allows acoustic signals to be visualized as images using spectrograms and correlograms. The signals of interest in these representations of acoustic signals, such as linearly related harmonics or broadband correlation peaks, essentially offer equivalence to meaningful image patterns immersed in noise. By applying a model of the photoreceptor stage of the hoverfly vision system, it is shown that the acoustic patterns can be enhanced and noise greatly suppressed. Compared with traditional narrowband and broadband techniques, the bio-inspired processing can extend the maximum detectable distance of the small and medium-sized unmanned aerial vehicles by between 30% and 50%, while simultaneously increasing the accuracy of flight parameter and trajectory estimations.
Collapse
Affiliation(s)
- Jian Fang
- Science, Technology, Engineering and Mathematics, University of South Australia, Adelaide, South Australia 5095, Australia
| | - Anthony Finn
- Science, Technology, Engineering and Mathematics, University of South Australia, Adelaide, South Australia 5095, Australia
| | - Ron Wyber
- Midspar Systems, Farrer Place, Oyster Bay, New South Wales 2225, Australia
| | - Russell S A Brinkworth
- College of Science and Engineering, Flinders University, Clovelly Park, South Australia 5042, Australia
| |
Collapse
|
7
|
James JV, Cazzolato BS, Grainger S, Wiederman SD. Nonlinear, neuronal adaptation in insect vision models improves target discrimination within repetitively moving backgrounds. BIOINSPIRATION & BIOMIMETICS 2021; 16:066015. [PMID: 34555824 DOI: 10.1088/1748-3190/ac2988] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/23/2021] [Accepted: 09/23/2021] [Indexed: 06/13/2023]
Abstract
Neurons which respond selectively to small moving targets, even against a cluttered background, have been identified in several insect species. To investigate what underlies these robust and highly selective responses, researchers have probed the neuronal circuitry in target-detecting, visual pathways. Observations in flies reveal nonlinear adaptation over time, composed of a fast onset and gradual decay. This adaptive processing is seen in both of the independent, parallel pathways encoding either luminance increments (ON channel) or decrements (OFF channel). The functional significance of this adaptive phenomenon has not been determined from physiological studies, though the asymmetrical time course suggests a role in suppressing responses to repetitive stimuli. We tested this possibility by comparing an implementation of fast adaptation against alternatives, using a model of insect 'elementary small target motion detectors'. We conducted target-detecting simulations on various natural backgrounds, that were shifted via several movement profiles (and target velocities). Using performance metrics, we confirmed that the fast adaptation observed in neuronal systems enhances target detection against a repetitively moving background. Such background movement would be encountered via natural ego-motion as the insect travels through the world. These findings show that this form of nonlinear, fast-adaptation (suitably implementable via cellular biophysics) plays a role analogous to background subtraction techniques in conventional computer vision.
Collapse
Affiliation(s)
- John V James
- School of Mechanical Engineering, University of Adelaide, Adelaide SA, Australia
- Adelaide Medical School, University of Adelaide, Adelaide SA, Australia
| | - Benjamin S Cazzolato
- School of Mechanical Engineering, University of Adelaide, Adelaide SA, Australia
| | - Steven Grainger
- School of Mechanical Engineering, University of Adelaide, Adelaide SA, Australia
| | | |
Collapse
|