1
|
Xie Z, Qiu R, Wang S, Tan X, Xie Y, Ma L. PIG: Prompt Images Guidance for Night-Time Scene Parsing. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2024; 33:3921-3934. [PMID: 38913509 DOI: 10.1109/tip.2024.3415963] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/26/2024]
Abstract
Night-time scene parsing aims to extract pixel-level semantic information in night images, aiding downstream tasks in understanding scene object distribution. Due to limited labeled night image datasets, unsupervised domain adaptation (UDA) has become the predominant method for studying night scenes. UDA typically relies on paired day-night image pairs to guide adaptation, but this approach hampers dataset construction and restricts generalization across night scenes in different datasets. Moreover, UDA, focusing on network architecture and training strategies, faces difficulties in handling classes with few domain similarities. In this paper, we leverage Prompt Images Guidance (PIG) to enhance UDA with supplementary night knowledge. We propose a Night-Focused Network (NFNet) to learn night-specific features from both target domain images and prompt images. To generate high-quality pseudo-labels, we propose Pseudo-label Fusion via Domain Similarity Guidance (FDSG). Classes with fewer domain similarities are predicted by NFNet, which excels in parsing night features, while classes with more domain similarities are predicted by UDA, which has rich labeled semantics. Additionally, we propose two data augmentation strategies: the Prompt Mixture Strategy (PMS) and the Alternate Mask Strategy (AMS), aimed at mitigating the overfitting of the NFNet to a few prompt images. We conduct extensive experiments on four night-time datasets: NightCity, NightCity+, Dark Zurich, and ACDC. The results indicate that utilizing PIG can enhance the parsing accuracy of UDA. The code is available at https://github.com/qiurui4shu/PIG.
Collapse
|
2
|
Li H, Liu C, Yang Y. LayerNet: A One-Step Layered Network for Semantic Segmentation at Night. IEEE COMPUTER GRAPHICS AND APPLICATIONS 2023; 43:9-21. [PMID: 37028057 DOI: 10.1109/mcg.2023.3253167] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
We have collected a novel, nighttime scene dataset, called Rebecca, including 600 real images captured at night with pixel-level semantic annotations, which is currently scarce and can be invoked as a new benchmark. In addition, we proposed a one-step layered network, named LayerNet, to combine local features rich in appearance information in the shallow layer, global features abundant in semantic information in the deep layer, and middle-level features in between by explicitly modeling multistage features of objects in the nighttime. In addition, a multihead decoder and a well-designed hierarchical module are utilized to extract and fuse features of different depths. Numerous experiments show that our dataset can significantly improve the segmentation ability of the existing models for nighttime images. Meanwhile, our LayerNet achieves the state-of-the-art accuracy on Rebecca (65.3% mIOU). The dataset is available at https://github.com/Lihao482/REebecca.
Collapse
|
3
|
Xie Z, Wang S, Xu K, Zhang Z, Tan X, Xie Y, Ma L. Boosting Night-Time Scene Parsing With Learnable Frequency. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2023; 32:2386-2398. [PMID: 37071518 DOI: 10.1109/tip.2023.3267044] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Night-Time Scene Parsing (NTSP) is essential to many vision applications, especially for autonomous driving. Most of the existing methods are proposed for day-time scene parsing. They rely on modeling pixel intensity-based spatial contextual cues under even illumination. Hence, these methods do not perform well in night-time scenes as such spatial contextual cues are buried in the over-/under-exposed regions in night-time scenes. In this paper, we first conduct an image frequency-based statistical experiment to interpret the day-time and night-time scene discrepancies. We find that image frequency distributions differ significantly between day-time and night-time scenes, and understanding such frequency distributions is critical to NTSP problem. Based on this, we propose to exploit the image frequency distributions for night-time scene parsing. First, we propose a Learnable Frequency Encoder (LFE) to model the relationship between different frequency coefficients to measure all frequency components dynamically. Second, we propose a Spatial Frequency Fusion module (SFF) that fuses both spatial and frequency information to guide the extraction of spatial context features. Extensive experiments show that our method performs favorably against the state-of-the-art methods on the NightCity, NightCity+ and BDD100K-night datasets. In addition, we demonstrate that our method can be applied to existing day-time scene parsing methods and boost their performance on night-time scenes. The code is available at https://github.com/wangsen99/FDLNet.
Collapse
|
4
|
Mei H, Yang X, Yu L, Zhang Q, Wei X, Lau RWH. Large-Field Contextual Feature Learning for Glass Detection. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2023; 45:3329-3346. [PMID: 35984803 DOI: 10.1109/tpami.2022.3181973] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Glass is very common in our daily life. Existing computer vision systems neglect it and thus may have severe consequences, e.g., a robot may crash into a glass wall. However, sensing the presence of glass is not straightforward. The key challenge is that arbitrary objects/scenes can appear behind the glass. In this paper, we propose an important problem of detecting glass surfaces from a single RGB image. To address this problem, we construct the first large-scale glass detection dataset (GDD) and propose a novel glass detection network, called GDNet-B, which explores abundant contextual cues in a large field-of-view via a novel large-field contextual feature integration (LCFI) module and integrates both high-level and low-level boundary features with a boundary feature enhancement (BFE) module. Extensive experiments demonstrate that our GDNet-B achieves satisfying glass detection results on the images within and beyond the GDD testing set. We further validate the effectiveness and generalization capability of our proposed GDNet-B by applying it to other vision tasks, including mirror segmentation and salient object detection. Finally, we show the potential applications of glass detection and discuss possible future research directions.
Collapse
|
5
|
Zhang S, Ren W, Tan X, Wang ZJ, Liu Y, Zhang J, Zhang X, Cao X. Semantic-Aware Dehazing Network With Adaptive Feature Fusion. IEEE TRANSACTIONS ON CYBERNETICS 2023; 53:454-467. [PMID: 34797770 DOI: 10.1109/tcyb.2021.3124231] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Despite that convolutional neural networks (CNNs) have shown high-quality reconstruction for single image dehazing, recovering natural and realistic dehazed results remains a challenging problem due to semantic confusion in the hazy scene. In this article, we show that it is possible to recover textures faithfully by incorporating semantic prior into dehazing network since objects in haze-free images tend to show certain shapes, textures, and colors. We propose a semantic-aware dehazing network (SDNet) in which the semantic prior is taken as a color constraint for dehazing, benefiting the acquisition of a reasonable scene configuration. In addition, we design a densely connected block to capture global and local information for dehazing and semantic prior estimation. To eliminate the unnatural appearance of some objects, we propose to fuse the features from shallow and deep layers adaptively. Experimental results demonstrate that our proposed model performs favorably against the state-of-the-art single image dehazing approaches.
Collapse
|
6
|
Peng Y, Xu Q, Lin S, Wang X, Xiang G, Huang S, Zhang H, Fan C. The Application of Electroencephalogram in Driving Safety: Current Status and Future Prospects. Front Psychol 2022; 13:919695. [PMID: 35936295 PMCID: PMC9354986 DOI: 10.3389/fpsyg.2022.919695] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/13/2022] [Accepted: 06/13/2022] [Indexed: 12/02/2022] Open
Abstract
The driver is one of the most important factors in the safety of the transportation system. The driver's perceptual characteristics are closely related to driving behavior, while electroencephalogram (EEG) as the gold standard for evaluating human perception is non-deceptive. It is essential to study driving characteristics by analyzing the driver's brain activity pattern, effectively acquiring driver perceptual characteristics, creating a direct connection between the driver's brain and external devices, and realizing information interchange. This paper first introduces the theories related to EEG, then reviews the applications of EEG in scenarios such as fatigue driving, distracted driving, and emotional driving. The limitations of existing research have been identified and the prospect of EEG application in future brain-computer interface automotive assisted driving systems have been proposed. This review provides guidance for researchers to use EEG to improve driving safety. It also offers valuable suggestions for future research.
Collapse
Affiliation(s)
- Yong Peng
- Key Laboratory of Traffic Safety on Track of Ministry of Education, School of Traffic & Transportation Engineering, Central South University, Changsha, China
| | - Qian Xu
- Key Laboratory of Traffic Safety on Track of Ministry of Education, School of Traffic & Transportation Engineering, Central South University, Changsha, China
| | - Shuxiang Lin
- Key Laboratory of Traffic Safety on Track of Ministry of Education, School of Traffic & Transportation Engineering, Central South University, Changsha, China
| | - Xinghua Wang
- Key Laboratory of Traffic Safety on Track of Ministry of Education, School of Traffic & Transportation Engineering, Central South University, Changsha, China
| | - Guoliang Xiang
- Key Laboratory of Traffic Safety on Track of Ministry of Education, School of Traffic & Transportation Engineering, Central South University, Changsha, China
| | - Shufang Huang
- School of Business and Trade, Hunan Industry Polytechnic, Changsha, China
| | - Honghao Zhang
- School of Mechanical Engineering, Shandong University, Jinan, China
| | - Chaojie Fan
- Key Laboratory of Traffic Safety on Track of Ministry of Education, School of Traffic & Transportation Engineering, Central South University, Changsha, China
| |
Collapse
|
7
|
Types, Risk Factors, Consequences, and Detection Methods of Train Driver Fatigue and Distraction. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:8328077. [PMID: 35371223 PMCID: PMC8970922 DOI: 10.1155/2022/8328077] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/10/2022] [Revised: 02/13/2022] [Accepted: 02/19/2022] [Indexed: 11/17/2022]
Abstract
Train drivers' inattention, including fatigue and distraction, impairs their ability to drive and is the major risk factor for human-caused train accidents. Many experts have undertaken numerous studies on train driver exhaustion and distraction, but a systematic study is still missing. Through a systematic review, this work aims to outline the types, risk factors, consequences, and detection methods of train driver fatigue and distraction. The effects of central nervous fatigue and cognitive distraction in train drivers during driving are caused by rest and sleep schedules, workload, automation levels, and mobile phones. Furthermore, train drivers' fatigue and distraction can cause loss of concentration and slow reaction, resulting in dangerous driving behaviour such as speeding and SPAD. Researchers have combined subjective reporting, physiological parameters, and physical factors to construct detection algorithms with good results to detect train driver fatigue and distraction. This review offers recommendations for researchers looking into train driver fatigue and distraction. And it can also make valuable recommendations for future studies about railway traffic safety.
Collapse
|
8
|
|
9
|
Wang Q, Tan X, Ma L, Liu C. Dual Windows Are Significant: Learning from Mediastinal Window and Focusing on Lung Window. ARTIF INTELL 2022. [DOI: 10.1007/978-3-031-20497-5_16] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/23/2022]
|