1
|
Abstract
Wind turbines are causing unprecedented numbers of bat fatalities. Many fatalities involve tree-roosting bats, but reasons for this higher susceptibility remain unknown. To better understand behaviors associated with risk, we monitored bats at three experimentally manipulated wind turbines in Indiana, United States, from July 29 to October 1, 2012, using thermal cameras and other methods. We observed bats on 993 occasions and saw many behaviors, including close approaches, flight loops and dives, hovering, and chases. Most bats altered course toward turbines during observation. Based on these new observations, we tested the hypotheses that wind speed and blade rotation speed influenced the way that bats interacted with turbines. We found that bats were detected more frequently at lower wind speeds and typically approached turbines on the leeward (downwind) side. The proportion of leeward approaches increased with wind speed when blades were prevented from turning, yet decreased when blades could turn. Bats were observed more frequently at turbines on moonlit nights. Taken together, these observations suggest that bats may orient toward turbines by sensing air currents and using vision, and that air turbulence caused by fast-moving blades creates conditions that are less attractive to bats passing in close proximity. Tree bats may respond to streams of air flowing downwind from trees at night while searching for roosts, conspecifics, and nocturnal insect prey that could accumulate in such flows. Fatalities of tree bats at turbines may be the consequence of behaviors that evolved to provide selective advantages when elicited by tall trees, but are now maladaptive when elicited by wind turbines.
Collapse
|
Research Support, U.S. Gov't, Non-P.H.S. |
11 |
121 |
2
|
Jain AK, Ross A. Bridging the gap: from biometrics to forensics. Philos Trans R Soc Lond B Biol Sci 2016; 370:rstb.2014.0254. [PMID: 26101280 DOI: 10.1098/rstb.2014.0254] [Citation(s) in RCA: 44] [Impact Index Per Article: 4.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022] Open
Abstract
Biometric recognition, or simply biometrics, refers to automated recognition of individuals based on their behavioural and biological characteristics. The success of fingerprints in forensic science and law enforcement applications, coupled with growing concerns related to border control, financial fraud and cyber security, has generated a huge interest in using fingerprints, as well as other biological traits, for automated person recognition. It is, therefore, not surprising to see biometrics permeating various segments of our society. Applications include smartphone security, mobile payment, border crossing, national civil registry and access to restricted facilities. Despite these successful deployments in various fields, there are several existing challenges and new opportunities for person recognition using biometrics. In particular, when biometric data is acquired in an unconstrained environment or if the subject is uncooperative, the quality of the ensuing biometric data may not be amenable for automated person recognition. This is particularly true in crime-scene investigations, where the biological evidence gleaned from a scene may be of poor quality. In this article, we first discuss how biometrics evolved from forensic science and how its focus is shifting back to its origin in order to address some challenging problems. Next, we enumerate the similarities and differences between biometrics and forensics. We then present some applications where the principles of biometrics are being successfully leveraged into forensics in order to solve critical problems in the law enforcement domain. Finally, we discuss new collaborative opportunities for researchers in biometrics and forensics, in order to address hitherto unsolved problems that can benefit society at large.
Collapse
|
Journal Article |
9 |
44 |
3
|
Lo SW, Wu JH, Lin FP, Hsu CH. Cyber surveillance for flood disasters. SENSORS (BASEL, SWITZERLAND) 2015; 15:2369-87. [PMID: 25621609 PMCID: PMC4367310 DOI: 10.3390/s150202369] [Citation(s) in RCA: 37] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/17/2014] [Accepted: 01/12/2015] [Indexed: 11/16/2022]
Abstract
Regional heavy rainfall is usually caused by the influence of extreme weather conditions. Instant heavy rainfall often results in the flooding of rivers and the neighboring low-lying areas, which is responsible for a large number of casualties and considerable property loss. The existing precipitation forecast systems mostly focus on the analysis and forecast of large-scale areas but do not provide precise instant automatic monitoring and alert feedback for individual river areas and sections. Therefore, in this paper, we propose an easy method to automatically monitor the flood object of a specific area, based on the currently widely used remote cyber surveillance systems and image processing methods, in order to obtain instant flooding and waterlogging event feedback. The intrusion detection mode of these surveillance systems is used in this study, wherein a flood is considered a possible invasion object. Through the detection and verification of flood objects, automatic flood risk-level monitoring of specific individual river segments, as well as the automatic urban inundation detection, has become possible. The proposed method can better meet the practical needs of disaster prevention than the method of large-area forecasting. It also has several other advantages, such as flexibility in location selection, no requirement of a standard water-level ruler, and a relatively large field of view, when compared with the traditional water-level measurements using video screens. The results can offer prompt reference for appropriate disaster warning actions in small areas, making them more accurate and effective.
Collapse
|
research-article |
10 |
37 |
4
|
Shorfuzzaman M, Hossain MS, Alhamid MF. Towards the sustainable development of smart cities through mass video surveillance: A response to the COVID-19 pandemic. SUSTAINABLE CITIES AND SOCIETY 2021; 64:102582. [PMID: 33178557 PMCID: PMC7644199 DOI: 10.1016/j.scs.2020.102582] [Citation(s) in RCA: 29] [Impact Index Per Article: 7.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/10/2020] [Revised: 10/13/2020] [Accepted: 10/27/2020] [Indexed: 05/22/2023]
Abstract
Sustainable smart city initiatives around the world have recently had great impact on the lives of citizens and brought significant changes to society. More precisely, data-driven smart applications that efficiently manage sparse resources are offering a futuristic vision of smart, efficient, and secure city operations. However, the ongoing COVID-19 pandemic has revealed the limitations of existing smart city deployment; hence; the development of systems and architectures capable of providing fast and effective mechanisms to limit further spread of the virus has become paramount. An active surveillance system capable of monitoring and enforcing social distancing between people can effectively slow the spread of this deadly virus. In this paper, we propose a data-driven deep learning-based framework for the sustainable development of a smart city, offering a timely response to combat the COVID-19 pandemic through mass video surveillance. To implementing social distancing monitoring, we used three deep learning-based real-time object detection models for the detection of people in videos captured with a monocular camera. We validated the performance of our system using a real-world video surveillance dataset for effective deployment.
Collapse
|
research-article |
4 |
29 |
5
|
Millan-Garcia L, Sanchez-Perez G, Nakano M, Toscano-Medina K, Perez-Meana H, Rojas-Cardenas L. An early fire detection algorithm using IP cameras. SENSORS 2012; 12:5670-86. [PMID: 22778607 PMCID: PMC3386706 DOI: 10.3390/s120505670] [Citation(s) in RCA: 28] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/29/2012] [Revised: 04/19/2012] [Accepted: 04/28/2012] [Indexed: 11/22/2022]
Abstract
The presence of smoke is the first symptom of fire; therefore to achieve early fire detection, accurate and quick estimation of the presence of smoke is very important. In this paper we propose an algorithm to detect the presence of smoke using video sequences captured by Internet Protocol (IP) cameras, in which important features of smoke, such as color, motion and growth properties are employed. For an efficient smoke detection in the IP camera platform, a detection algorithm must operate directly in the Discrete Cosine Transform (DCT) domain to reduce computational cost, avoiding a complete decoding process required for algorithms that operate in spatial domain. In the proposed algorithm the DCT Inter-transformation technique is used to increase the detection accuracy without inverse DCT operation. In the proposed scheme, firstly the candidate smoke regions are estimated using motion and color smoke properties; next using morphological operations the noise is reduced. Finally the growth properties of the candidate smoke regions are furthermore analyzed through time using the connected component labeling technique. Evaluation results show that a feasible smoke detection method with false negative and false positive error rates approximately equal to 4% and 2%, respectively, is obtained.
Collapse
|
|
13 |
28 |
6
|
He B, Yu S. Parallax-Robust Surveillance Video Stitching. SENSORS 2015; 16:s16010007. [PMID: 26712756 PMCID: PMC4732040 DOI: 10.3390/s16010007] [Citation(s) in RCA: 25] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/07/2015] [Revised: 11/30/2015] [Accepted: 12/17/2015] [Indexed: 11/16/2022]
Abstract
This paper presents a parallax-robust video stitching technique for timely synchronized surveillance video. An efficient two-stage video stitching procedure is proposed in this paper to build wide Field-of-View (FOV) videos for surveillance applications. In the stitching model calculation stage, we develop a layered warping algorithm to align the background scenes, which is location-dependent and turned out to be more robust to parallax than the traditional global projective warping methods. On the selective seam updating stage, we propose a change-detection based optimal seam selection approach to avert ghosting and artifacts caused by moving foregrounds. Experimental results demonstrate that our procedure can efficiently stitch multi-view videos into a wide FOV video output without ghosting and noticeable seams.
Collapse
|
|
10 |
25 |
7
|
FPGA implementation for real-time background subtraction based on Horprasert model. SENSORS 2012; 12:585-611. [PMID: 22368487 PMCID: PMC3279231 DOI: 10.3390/s120100585] [Citation(s) in RCA: 22] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/10/2011] [Revised: 12/28/2011] [Accepted: 01/03/2012] [Indexed: 11/17/2022]
Abstract
Background subtraction is considered the first processing stage in video surveillance systems, and consists of determining objects in movement in a scene captured by a static camera. It is an intensive task with a high computational cost. This work proposes an embedded novel architecture on FPGA which is able to extract the background on resource-limited environments and offers low degradation (produced because of the hardware-friendly model modification). In addition, the original model is extended in order to detect shadows and improve the quality of the segmentation of the moving objects. We have analyzed the resource consumption and performance in Spartan3 Xilinx FPGAs and compared to others works available on the literature, showing that the current architecture is a good trade-off in terms of accuracy, performance and resources utilization. With less than a 65% of the resources utilization of a XC3SD3400 Spartan-3A low-cost family FPGA, the system achieves a frequency of 66.5 MHz reaching 32.8 fps with resolution 1,024 × 1,024 pixels, and an estimated power consumption of 5.76 W.
Collapse
|
Research Support, Non-U.S. Gov't |
13 |
22 |
8
|
Zhou X, Liu X, Jiang A, Yan B, Yang C. Improving Video Segmentation by Fusing Depth Cues and the Visual Background Extractor (ViBe) Algorithm. SENSORS 2017; 17:s17051177. [PMID: 28531134 PMCID: PMC5470922 DOI: 10.3390/s17051177] [Citation(s) in RCA: 18] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/19/2017] [Revised: 05/12/2017] [Accepted: 05/18/2017] [Indexed: 11/16/2022]
Abstract
Depth-sensing technology has led to broad applications of inexpensive depth cameras that can capture human motion and scenes in three-dimensional space. Background subtraction algorithms can be improved by fusing color and depth cues, thereby allowing many issues encountered in classical color segmentation to be solved. In this paper, we propose a new fusion method that combines depth and color information for foreground segmentation based on an advanced color-based algorithm. First, a background model and a depth model are developed. Then, based on these models, we propose a new updating strategy that can eliminate ghosting and black shadows almost completely. Extensive experiments have been performed to compare the proposed algorithm with other, conventional RGB-D (Red-Green-Blue and Depth) algorithms. The experimental results suggest that our method extracts foregrounds with higher effectiveness and efficiency.
Collapse
|
|
8 |
18 |
9
|
Suriani NS, Hussain A, Zulkifley MA. Sudden event recognition: a survey. SENSORS 2013; 13:9966-98. [PMID: 23921828 PMCID: PMC3812589 DOI: 10.3390/s130809966] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/30/2013] [Revised: 07/26/2013] [Accepted: 07/26/2013] [Indexed: 11/16/2022]
Abstract
Event recognition is one of the most active research areas in video surveillance fields. Advancement in event recognition systems mainly aims to provide convenience, safety and an efficient lifestyle for humanity. A precise, accurate and robust approach is necessary to enable event recognition systems to respond to sudden changes in various uncontrolled environments, such as the case of an emergency, physical threat and a fire or bomb alert. The performance of sudden event recognition systems depends heavily on the accuracy of low level processing, like detection, recognition, tracking and machine learning algorithms. This survey aims to detect and characterize a sudden event, which is a subset of an abnormal event in several video surveillance applications. This paper discusses the following in detail: (1) the importance of a sudden event over a general anomalous event; (2) frameworks used in sudden event recognition; (3) the requirements and comparative studies of a sudden event recognition system and (4) various decision-making approaches for sudden event recognition. The advantages and drawbacks of using 3D images from multiple cameras for real-time application are also discussed. The paper concludes with suggestions for future research directions in sudden event recognition.
Collapse
|
Review |
12 |
16 |
10
|
Kalbo N, Mirsky Y, Shabtai A, Elovici Y. The Security of IP-Based Video Surveillance Systems. SENSORS 2020; 20:s20174806. [PMID: 32858840 PMCID: PMC7506579 DOI: 10.3390/s20174806] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/18/2020] [Revised: 08/13/2020] [Accepted: 08/20/2020] [Indexed: 11/16/2022]
Abstract
Over the last decade, video surveillance systems have become a part of the Internet of Things (IoT). These IP-based surveillance systems now protect industrial facilities, railways, gas stations, and even one's own home. Unfortunately, like other IoT systems, there are inherent security risks which can lead to significant violations of a user's privacy. In this review, we explore the attack surface of modern surveillance systems and enumerate the various ways they can be compromised with real examples. We also identify the threat agents, their attack goals, attack vectors, and the resulting consequences of successful attacks. Finally, we present current countermeasures and best practices and discuss the threat horizon. The purpose of this review is to provide researchers and engineers with a better understanding of a modern surveillance systems' security, to harden existing systems and develop improved security solutions.
Collapse
|
Review |
5 |
14 |
11
|
Nagothu D, Chen Y, Blasch E, Aved A, Zhu S. Detecting Malicious False Frame Injection Attacks on Surveillance Systems at the Edge Using Electrical Network Frequency Signals. SENSORS 2019; 19:s19112424. [PMID: 31141880 PMCID: PMC6603661 DOI: 10.3390/s19112424] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/29/2019] [Revised: 05/17/2019] [Accepted: 05/24/2019] [Indexed: 11/25/2022]
Abstract
Over the past few years, the importance of video surveillance in securing national critical infrastructure has significantly increased, with applications including the detection of failures and anomalies. Accompanied by the proliferation of video is the increasing number of attacks against surveillance systems. Among the attacks, False Frame Injection (FFI) attacks that replay video frames from a previous recording to mask the live feed has the highest impact. While many attempts have been made to detect FFI frames using features from the video feeds, video analysis is computationally too intensive to be deployed on-site for real-time false frame detection. In this paper, we investigated the feasibility of FFI attacks on compromised surveillance systems at the edge and propose an effective technique to detect the injected false video and audio frames by monitoring the surveillance feed using the embedded Electrical Network Frequency (ENF) signals. An ENF operates at a nominal frequency of 60 Hz/50 Hz based on its geographical location and maintains a stable value across the entire power grid interconnection with minor fluctuations. For surveillance system video/audio recordings connected to the power grid, the ENF signals are embedded. The time-varying nature of the ENF component was used as a forensic application for authenticating the surveillance feed. The paper highlights the ENF signal collection from a power grid creating a reference database and ENF extraction from the recordings using conventional short-time Fourier Transform and spectrum detection for robust ENF signal analysis in the presence of noise and interference caused in different harmonics. The experimental results demonstrated the effectiveness of ENF signal detection and/or abnormalities for FFI attacks.
Collapse
|
|
6 |
11 |
12
|
Li J, Jia X, Shao C. Predicting Driver Behavior during the Yellow Interval Using Video Surveillance. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2016; 13:ijerph13121213. [PMID: 27929447 PMCID: PMC5201354 DOI: 10.3390/ijerph13121213] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/31/2016] [Revised: 11/21/2016] [Accepted: 11/29/2016] [Indexed: 11/16/2022]
Abstract
At a signalized intersection, drivers must make a stop/go decision at the onset of the yellow signal. Incorrect decisions would lead to red light running (RLR) violations or crashes. This study aims to predict drivers' stop/go decisions and RLR violations during yellow intervals. Traffic data such as vehicle approaching speed, acceleration, distance to the intersection, and occurrence of RLR violations are gathered by a Vehicle Data Collection System (VDCS). An enhanced Gaussian Mixture Model (GMM) is used to extract moving vehicles from target lanes, and the Kalman Filter (KF) algorithm is utilized to acquire vehicle trajectories. The data collected from the VDCS are further analyzed by a sequential logit model, and the relationship between drivers' stop/go decisions and RLR violations is identified. The results indicate that the distance of vehicles to the stop line at the onset of the yellow signal is an important predictor for both drivers' stop/go decisions and RLR violations. In addition, vehicle approaching speed is a contributing factor for stop/go decisions. Furthermore, the accelerations of vehicles after the onset of the yellow signal are positively related to RLR violations. The findings of this study can be used to predict the probability of drivers' RLR violations and improve traffic safety at signalized intersections.
Collapse
|
|
9 |
10 |
13
|
Lee GB, Lee MJ, Lee WK, Park JH, Kim TH. Shadow Detection Based on Regions of Light Sources for Object Extraction in Nighttime Video. SENSORS 2017; 17:s17030659. [PMID: 28327515 PMCID: PMC5375945 DOI: 10.3390/s17030659] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/10/2017] [Revised: 03/06/2017] [Accepted: 03/18/2017] [Indexed: 11/28/2022]
Abstract
Intelligent video surveillance systems detect pre-configured surveillance events through background modeling, foreground and object extraction, object tracking, and event detection. Shadow regions inside video frames sometimes appear as foreground objects, interfere with ensuing processes, and finally degrade the event detection performance of the systems. Conventional studies have mostly used intensity, color, texture, and geometric information to perform shadow detection in daytime video, but these methods lack the capability of removing shadows in nighttime video. In this paper, a novel shadow detection algorithm for nighttime video is proposed; this algorithm partitions each foreground object based on the object’s vertical histogram and screens out shadow objects by validating their orientations heading toward regions of light sources. From the experimental results, it can be seen that the proposed algorithm shows more than 93.8% shadow removal and 89.9% object extraction rates for nighttime video sequences, and the algorithm outperforms conventional shadow removal algorithms designed for daytime videos.
Collapse
|
|
8 |
7 |
14
|
Fall Detection System-Based Posture-Recognition for Indoor Environments. J Imaging 2021; 7:jimaging7030042. [PMID: 34460698 PMCID: PMC8321307 DOI: 10.3390/jimaging7030042] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/16/2020] [Revised: 02/11/2021] [Accepted: 02/18/2021] [Indexed: 02/08/2023] Open
Abstract
The majority of the senior population lives alone at home. Falls can cause serious injuries, such as fractures or head injuries. These injuries can be an obstacle for a person to move around and normally practice his daily activities. Some of these injuries can lead to a risk of death if not handled urgently. In this paper, we propose a fall detection system for elderly people based on their postures. The postures are recognized from the human silhouette which is an advantage to preserve the privacy of the elderly. The effectiveness of our approach is demonstrated on two well-known datasets for human posture classification and three public datasets for fall detection, using a Support-Vector Machine (SVM) classifier. The experimental results show that our method can not only achieves a high fall detection rate but also a low false detection.
Collapse
|
Journal Article |
4 |
6 |
15
|
Robust feedback zoom tracking for digital video surveillance. SENSORS 2012; 12:8073-99. [PMID: 22969388 PMCID: PMC3436017 DOI: 10.3390/s120608073] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/09/2012] [Revised: 05/31/2012] [Accepted: 06/01/2012] [Indexed: 12/30/2022]
Abstract
Zoom tracking is an important function in video surveillance, particularly in traffic management and security monitoring. It involves keeping an object of interest in focus during the zoom operation. Zoom tracking is typically achieved by moving the zoom and focus motors in lenses following the so-called “trace curve”, which shows the in-focus motor positions versus the zoom motor positions for a specific object distance. The main task of a zoom tracking approach is to accurately estimate the trace curve for the specified object. Because a proportional integral derivative (PID) controller has historically been considered to be the best controller in the absence of knowledge of the underlying process and its high-quality performance in motor control, in this paper, we propose a novel feedback zoom tracking (FZT) approach based on the geometric trace curve estimation and PID feedback controller. The performance of this approach is compared with existing zoom tracking methods in digital video surveillance. The real-time implementation results obtained on an actual digital video platform indicate that the developed FZT approach not only solves the traditional one-to-many mapping problem without pre-training but also improves the robustness for tracking moving or switching objects which is the key challenge in video surveillance.
Collapse
|
|
13 |
5 |
16
|
Uddin MK, Bhuiyan A, Bappee FK, Islam MM, Hasan M. Person Re-Identification with RGB-D and RGB-IR Sensors: A Comprehensive Survey. SENSORS (BASEL, SWITZERLAND) 2023; 23:s23031504. [PMID: 36772548 PMCID: PMC9919319 DOI: 10.3390/s23031504] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/01/2022] [Revised: 01/14/2023] [Accepted: 01/21/2023] [Indexed: 05/27/2023]
Abstract
Learning about appearance embedding is of great importance for a variety of different computer-vision applications, which has prompted a surge in person re-identification (Re-ID) papers. The aim of these papers has been to identify an individual over a set of non-overlapping cameras. Despite recent advances in RGB-RGB Re-ID approaches with deep-learning architectures, the approach fails to consistently work well when there are low resolutions in dark conditions. The introduction of different sensors (i.e., RGB-D and infrared (IR)) enables the capture of appearances even in dark conditions. Recently, a lot of research has been dedicated to addressing the issue of finding appearance embedding in dark conditions using different advanced camera sensors. In this paper, we give a comprehensive overview of existing Re-ID approaches that utilize the additional information from different sensor-based methods to address the constraints faced by RGB camera-based person Re-ID systems. Although there are a number of survey papers that consider either the RGB-RGB or Visible-IR scenarios, there are none that consider both RGB-D and RGB-IR. In this paper, we present a detailed taxonomy of the existing approaches along with the existing RGB-D and RGB-IR person Re-ID datasets. Then, we summarize the performance of state-of-the-art methods on several representative RGB-D and RGB-IR datasets. Finally, future directions and current issues are considered for improving the different sensor-based person Re-ID systems.
Collapse
|
Review |
2 |
5 |
17
|
Ingle PY, Kim YG. Real-Time Abnormal Object Detection for Video Surveillance in Smart Cities. SENSORS 2022; 22:s22103862. [PMID: 35632270 PMCID: PMC9143895 DOI: 10.3390/s22103862] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/26/2022] [Revised: 05/17/2022] [Accepted: 05/17/2022] [Indexed: 01/26/2023]
Abstract
With the adaptation of video surveillance in many areas for object detection, monitoring abnormal behavior in several cameras requires constant human tracking for a single camera operative, which is a tedious task. In multiview cameras, accurately detecting different types of guns and knives and classifying them from other video surveillance objects in real-time scenarios is difficult. Most detecting cameras are resource-constrained devices with limited computational capacities. To mitigate this problem, we proposed a resource-constrained lightweight subclass detection method based on a convolutional neural network to classify, locate, and detect different types of guns and knives effectively and efficiently in a real-time environment. In this paper, the detection classifier is a multiclass subclass detection convolutional neural network used to classify object frames into different sub-classes such as abnormal and normal. The achieved mean average precision by the best state-of-the-art framework to detect either a handgun or a knife is 84.21% or 90.20% on a single camera view. After extensive experiments, the best precision obtained by the proposed method for detecting different types of guns and knives was 97.50% on the ImageNet dataset and IMFDB, 90.50% on the open-image dataset, 93% on the Olmos dataset, and 90.7% precision on the multiview cameras. This resource-constrained device has shown a satisfactory result, with a precision score of 85.5% for detection in a multiview camera.
Collapse
|
|
3 |
5 |
18
|
Wilkowski A, Stefańczyk M, Kasprzak W. Training Data Extraction and Object Detection in Surveillance Scenario. SENSORS 2020; 20:s20092689. [PMID: 32397277 PMCID: PMC7249100 DOI: 10.3390/s20092689] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/31/2020] [Revised: 05/02/2020] [Accepted: 05/05/2020] [Indexed: 11/16/2022]
Abstract
Police and various security services use video analysis for securing public space, mass events, and when investigating criminal activity. Due to a huge amount of data supplied to surveillance systems, some automatic data processing is a necessity. In one typical scenario, an operator marks an object in an image frame and searches for all occurrences of the object in other frames or even image sequences. This problem is hard in general. Algorithms supporting this scenario must reconcile several seemingly contradicting factors: training and detection speed, detection reliability, and learning from small data sets. In the system proposed here, we use a two-stage detector. The first region proposal stage is based on a Cascade Classifier while the second classification stage is based either on a Support Vector Machines (SVMs) or Convolutional Neural Networks (CNNs). The proposed configuration ensures both speed and detection reliability. In addition to this, an object tracking and background-foreground separation algorithm is used, supported by the GrabCut algorithm and a sample synthesis procedure, in order to collect rich training data for the detector. Experiments show that the system is effective, useful, and applicable to practical surveillance tasks.
Collapse
|
Journal Article |
5 |
5 |
19
|
Gualotuña T, Macías E, Suárez Á, C ERF, Rivadeneira A. Low Cost Efficient Deliverying Video Surveillance Service to Moving Guard for Smart Home. SENSORS 2018; 18:s18030745. [PMID: 29494551 PMCID: PMC5877210 DOI: 10.3390/s18030745] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/13/2018] [Revised: 02/13/2018] [Accepted: 02/27/2018] [Indexed: 11/30/2022]
Abstract
Low-cost video surveillance systems are attractive for Smart Home applications (especially in emerging economies). Those systems use the flexibility of the Internet of Things to operate the video camera only when an intrusion is detected. We are the only ones that focus on the design of protocols based on intelligent agents to communicate the video of an intrusion in real time to the guards by wireless or mobile networks. The goal is to communicate, in real time, the video to the guards who can be moving towards the smart home. However, this communication suffers from sporadic disruptions that difficults the control and drastically reduces user satisfaction and operativity of the system. In a novel way, we have designed a generic software architecture based on design patterns that can be adapted to any hardware in a simple way. The implanted hardware is of very low economic cost; the software frameworks are free. In the experimental tests we have shown that it is possible to communicate to the moving guard, intrusion notifications (by e-mail and by instant messaging), and the first video frames in less than 20 s. In addition, we automatically recovered the frames of video lost in the disruptions in a transparent way to the user, we supported vertical handover processes and we could save energy of the smartphone's battery. However, the most important thing was that the high satisfaction of the people who have used the system.
Collapse
|
|
7 |
5 |
20
|
Goldenberg SZ, Cryan PM, Gorresen PM, Fingersh LJ. Behavioral patterns of bats at a wind turbine confirm seasonality of fatality risk. Ecol Evol 2021; 11:4843-4853. [PMID: 33976852 PMCID: PMC8093663 DOI: 10.1002/ece3.7388] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2020] [Revised: 02/03/2021] [Accepted: 02/05/2021] [Indexed: 01/18/2023] Open
Abstract
Bat fatalities at wind energy facilities in North America are predominantly comprised of migratory, tree-dependent species, but it is unclear why these bats are at higher risk. Factors influencing bat susceptibility to wind turbines might be revealed by temporal patterns in their behaviors around these dynamic landscape structures. In northern temperate zones, fatalities occur mostly from July through October, but whether this reflects seasonally variable behaviors, passage of migrants, or some combination of factors remains unknown. In this study, we examined video imagery spanning one year in the state of Colorado in the United States, to characterize patterns of seasonal and nightly variability in bat behavior at a wind turbine. We detected bats on 177 of 306 nights representing approximately 3,800 hr of video and > 2,000 discrete bat events. We observed bats approaching the turbine throughout the night across all months during which bats were observed. Two distinct seasonal peaks of bat activity occurred in July and September, representing 30% and 42% increases in discrete bat events from the preceding months June and August, respectively. Bats exhibited behaviors around the turbine that increased in both diversity and duration in July and September. The peaks in bat events were reflected in chasing and turbine approach behaviors. Many of the bat events involved multiple approaches to the turbine, including when bats were displaced through the air by moving blades. The seasonal and nightly patterns we observed were consistent with the possibility that wind turbines invoke investigative behaviors in bats in late summer and autumn coincident with migration and that bats may return and fly close to wind turbines even after experiencing potentially disruptive stimuli like moving blades. Our results point to the need for a deeper understanding of the seasonality, drivers, and characteristics of bat movement across spatial scales.
Collapse
|
research-article |
4 |
5 |
21
|
Normalized Metadata Generation for Human Retrieval Using Multiple Video Surveillance Cameras. SENSORS 2016; 16:s16070963. [PMID: 27347961 PMCID: PMC4969833 DOI: 10.3390/s16070963] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/20/2016] [Revised: 06/16/2016] [Accepted: 06/21/2016] [Indexed: 11/16/2022]
Abstract
Since it is impossible for surveillance personnel to keep monitoring videos from a multiple camera-based surveillance system, an efficient technique is needed to help recognize important situations by retrieving the metadata of an object-of-interest. In a multiple camera-based surveillance system, an object detected in a camera has a different shape in another camera, which is a critical issue of wide-range, real-time surveillance systems. In order to address the problem, this paper presents an object retrieval method by extracting the normalized metadata of an object-of-interest from multiple, heterogeneous cameras. The proposed metadata generation algorithm consists of three steps: (i) generation of a three-dimensional (3D) human model; (ii) human object-based automatic scene calibration; and (iii) metadata generation. More specifically, an appropriately-generated 3D human model provides the foot-to-head direction information that is used as the input of the automatic calibration of each camera. The normalized object information is used to retrieve an object-of-interest in a wide-range, multiple-camera surveillance system in the form of metadata. Experimental results show that the 3D human model matches the ground truth, and automatic calibration-based normalization of metadata enables a successful retrieval and tracking of a human object in the multiple-camera video surveillance system.
Collapse
|
Journal Article |
9 |
4 |
22
|
Mohamed NA, Zulkifley MA, Ibrahim AA, Aouache M. Optimal Training Configurations of a CNN-LSTM-Based Tracker for a Fall Frame Detection System. SENSORS (BASEL, SWITZERLAND) 2021; 21:6485. [PMID: 34640803 PMCID: PMC8512416 DOI: 10.3390/s21196485] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/10/2021] [Revised: 09/17/2021] [Accepted: 09/23/2021] [Indexed: 12/27/2022]
Abstract
In recent years, there has been an immense amount of research into fall event detection. Generally, a fall event is defined as a situation in which a person unintentionally drops down onto a lower surface. It is crucial to detect the occurrence of fall events as early as possible so that any severe fall consequences can be minimized. Nonetheless, a fall event is a sporadic incidence that occurs seldomly that is falsely detected due to a wide range of fall conditions and situations. Therefore, an automated fall frame detection system, which is referred to as the SmartConvFall is proposed to detect the exact fall frame in a video sequence. It is crucial to know the exact fall frame as it dictates the response time of the system to administer an early treatment to reduce the fall's negative consequences and related injuries. Henceforth, searching for the optimal training configurations is imperative to ensure the main goal of the SmartConvFall is achieved. The proposed SmartConvFall consists of two parts, which are object tracking and instantaneous fall frame detection modules that rely on deep learning representations. The first stage will track the object of interest using a fully convolutional neural network (CNN) tracker. Various training configurations such as optimizer, learning rate, mini-batch size, number of training samples, and region of interest are individually evaluated to determine the best configuration to produce the best tracker model. Meanwhile, the second module goal is to determine the exact instantaneous fall frame by modeling the continuous object trajectories using the Long Short-Term Memory (LSTM) network. Similarly, the LSTM model will undergo various training configurations that cover different types of features selection and the number of stacked layers. The exact instantaneous fall frame is determined using an assumption that a large movement difference with respect to the ground level along the vertical axis can be observed if a fall incident happened. The proposed SmartConvFall is a novel technique as most of the existing methods still relying on detection rather than the tracking module. The SmartConvFall outperforms the state-of-the-art trackers, namely TCNN and MDNET-N trackers, with the highest expected average overlap, robustness, and reliability metrics of 0.1619, 0.6323, and 0.7958, respectively. The SmartConvFall also managed to produce the lowest number of tracking failures with only 43 occasions. Moreover, a three-stack LSTM delivers the lowest mean error with approximately one second delay time in locating the exact instantaneous fall frame. Therefore, the proposed SmartConvFall has demonstrated its potential and suitability to be implemented for a real-time application that could help to avoid any crucial fall consequences such as death and internal bleeding if the early treatment can be administered.
Collapse
|
research-article |
4 |
3 |
23
|
Li X, Onie S, Liang M, Larsen M, Sowmya A. Towards Building a Visual Behaviour Analysis Pipeline for Suicide Detection and Prevention. SENSORS (BASEL, SWITZERLAND) 2022; 22:4488. [PMID: 35746270 PMCID: PMC9228922 DOI: 10.3390/s22124488] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/15/2022] [Revised: 06/07/2022] [Accepted: 06/08/2022] [Indexed: 06/15/2023]
Abstract
Understanding human behaviours through video analysis has seen significant research progress in recent years with the advancement of deep learning. This topic is of great importance to the next generation of intelligent visual surveillance systems which are capable of real-time detection and analysis of human behaviours. One important application is to automatically monitor and detect individuals who are in crisis at suicide hotspots to facilitate early intervention and prevention. However, there is still a significant gap between research in human action recognition and visual video processing in general, and their application to monitor hotspots for suicide prevention. While complex backgrounds, non-rigid movements of pedestrians and limitations of surveillance cameras and multi-task requirements for a surveillance system all pose challenges to the development of such systems, a further challenge is the detection of crisis behaviours before a suicide attempt is made, and there is a paucity of datasets in this area due to privacy and confidentiality issues. Most relevant research only applies to detecting suicides such as hangings or jumps from bridges, providing no potential for early prevention. In this research, these problems are addressed by proposing a new modular design for an intelligent visual processing pipeline that is capable of pedestrian detection, tracking, pose estimation and recognition of both normal actions and high risk behavioural cues that are important indicators of a suicide attempt. Specifically, based on the key finding that human body gestures can be used for the detection of social signals that potentially precede a suicide attempt, a new 2D skeleton-based action recognition algorithm is proposed. By using a two-branch network that takes advantage of three types of skeleton-based features extracted from a sequence of frames and a stacked LSTM structure, the model predicts the action label at each time step. It achieved good performance on both the public dataset JHMDB and a smaller private CCTV footage collection on action recognition. Moreover, a logical layer, which uses knowledge from a human coding study to recognise pre-suicide behaviour indicators, has been built on top of the action recognition module to compensate for the small dataset size. It enables complex behaviour patterns to be recognised even from smaller datasets. The whole pipeline has been tested in a real-world application of suicide prevention using simulated footage from a surveillance system installed at a suicide hotspot, and preliminary results confirm its effectiveness at capturing crisis behaviour indicators for early detection and prevention of suicide.
Collapse
|
research-article |
3 |
3 |
24
|
Mazzola G, Lo Presti L, Ardizzone E, La Cascia M. A Dataset of Annotated Omnidirectional Videos for Distancing Applications. J Imaging 2021; 7:jimaging7080158. [PMID: 34460794 PMCID: PMC8404929 DOI: 10.3390/jimaging7080158] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/05/2021] [Revised: 08/05/2021] [Accepted: 08/17/2021] [Indexed: 11/16/2022] Open
Abstract
Omnidirectional (or 360°) cameras are acquisition devices that, in the next few years, could have a big impact on video surveillance applications, research, and industry, as they can record a spherical view of a whole environment from every perspective. This paper presents two new contributions to the research community: the CVIP360 dataset, an annotated dataset of 360° videos for distancing applications, and a new method to estimate the distances of objects in a scene from a single 360° image. The CVIP360 dataset includes 16 videos acquired outdoors and indoors, annotated by adding information about the pedestrians in the scene (bounding boxes) and the distances to the camera of some points in the 3D world by using markers at fixed and known intervals. The proposed distance estimation algorithm is based on geometry facts regarding the acquisition process of the omnidirectional device, and is uncalibrated in practice: the only required parameter is the camera height. The proposed algorithm was tested on the CVIP360 dataset, and empirical results demonstrate that the estimation error is negligible for distancing applications.
Collapse
|
|
4 |
3 |
25
|
Ciampi L, Foszner P, Messina N, Staniszewski M, Gennaro C, Falchi F, Serao G, Cogiel M, Golba D, Szczęsna A, Amato G. Bus Violence: An Open Benchmark for Video Violence Detection on Public Transport. SENSORS (BASEL, SWITZERLAND) 2022; 22:8345. [PMID: 36366043 PMCID: PMC9658862 DOI: 10.3390/s22218345] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/20/2022] [Revised: 10/25/2022] [Accepted: 10/28/2022] [Indexed: 06/16/2023]
Abstract
The automatic detection of violent actions in public places through video analysis is difficult because the employed Artificial Intelligence-based techniques often suffer from generalization problems. Indeed, these algorithms hinge on large quantities of annotated data and usually experience a drastic drop in performance when used in scenarios never seen during the supervised learning phase. In this paper, we introduce and publicly release the Bus Violence benchmark, the first large-scale collection of video clips for violence detection on public transport, where some actors simulated violent actions inside a moving bus in changing conditions, such as the background or light. Moreover, we conduct a performance analysis of several state-of-the-art video violence detectors pre-trained with general violence detection databases on this newly established use case. The achieved moderate performances reveal the difficulties in generalizing from these popular methods, indicating the need to have this new collection of labeled data, beneficial for specializing them in this new scenario.
Collapse
|
research-article |
3 |
2 |