1
|
Alsadik B, Spreeuwers L, Dadrass Javan F, Manterola N. Mathematical Camera Array Optimization for Face 3D Modeling Application. Sensors (Basel) 2023; 23:9776. [PMID: 38139622 PMCID: PMC10747194 DOI: 10.3390/s23249776] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/06/2023] [Revised: 11/20/2023] [Accepted: 12/07/2023] [Indexed: 12/24/2023]
Abstract
Camera network design is a challenging task for many applications in photogrammetry, biomedical engineering, robotics, and industrial metrology, among other fields. Many driving factors are found in the camera network design including the camera specifications, object of interest, and type of application. One of the interesting applications is 3D face modeling and recognition which involves recognizing an individual based on facial attributes derived from the constructed 3D model. Developers and researchers still face difficulty in reaching the required high level of accuracy and reliability needed for image-based 3D face models. This is caused among many factors by the hardware limitations and imperfection of the cameras and the lack of proficiency in designing the ideal camera-system configuration. Accordingly, for precise measurements, we still need engineering-based techniques to ascertain the specific level of deliverables quality. In this paper, an optimal geometric design methodology of the camera network is presented by investigating different multi-camera system configurations composed of four up to eight cameras. A mathematical nonlinear constrained optimization technique is applied to solve the problem and each camera system configuration is tested for a facial 3D model where a quality assessment is applied to conclude the best configuration. The optimal configuration is found to be a 7-camera array, comprising a pentagon shape enclosing two additional cameras, offering high accuracy. For those who prioritize point density, a 9-camera array with a pentagon and quadrilateral arrangement in the X-Z plane is a viable choice. However, a 5-camera array offers a balance between accuracy and the number of cameras.
Collapse
Affiliation(s)
- Bashar Alsadik
- Faculty of Geo-Information Science and Earth Observation (ITC), University of Twente, 7522 NB Enschede, The Netherlands;
| | - Luuk Spreeuwers
- Data Management and Biometrics (DMB), Faculty of Electrical Engineering, Mathematics and Computer Science (EEMCS), University of Twente, 7522 NB Enschede, The Netherlands; (L.S.); (N.M.)
| | - Farzaneh Dadrass Javan
- Faculty of Geo-Information Science and Earth Observation (ITC), University of Twente, 7522 NB Enschede, The Netherlands;
| | - Nahuel Manterola
- Data Management and Biometrics (DMB), Faculty of Electrical Engineering, Mathematics and Computer Science (EEMCS), University of Twente, 7522 NB Enschede, The Netherlands; (L.S.); (N.M.)
| |
Collapse
|
2
|
Di Gennaro TM, Waldmann J. Sensor Fusion with Asynchronous Decentralized Processing for 3D Target Tracking with a Wireless Camera Network. Sensors (Basel) 2023; 23:1194. [PMID: 36772236 PMCID: PMC9919314 DOI: 10.3390/s23031194] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/15/2022] [Revised: 01/13/2023] [Accepted: 01/16/2023] [Indexed: 06/18/2023]
Abstract
We present a method to acquire 3D position measurements for decentralized target tracking with an asynchronous camera network. Cameras with known poses have fields of view with overlapping projections on the ground and 3D volumes above a reference ground plane. The purpose is to track targets in 3D space without constraining motion to a reference ground plane. Cameras exchange line-of-sight vectors and respective time tags asynchronously. From stereoscopy, we obtain the fused 3D measurement at the local frame capture instant. We use local decentralized Kalman information filtering and particle filtering for target state estimation to test our approach with only local estimation. Monte Carlo simulation includes communication losses due to frame processing delays. We measure performance with the average root mean square error of 3D position estimates projected on the image planes of the cameras. We then compare only local estimation to exchanging additional asynchronous communications using the Batch Asynchronous Filter and the Sequential Asynchronous Particle Filter for further fusion of information pairs' estimates and fused 3D position measurements, respectively. Similar performance occurs in spite of the additional communication load relative to our local estimation approach, which exchanges just line-of-sight vectors.
Collapse
Affiliation(s)
- Thiago Marchi Di Gennaro
- Systems and Control Department, Weapons Systems Directorate, Brazilian Navy, Rio de Janeiro 20010-100, RJ, Brazil
| | - Jacques Waldmann
- Systems and Control Department, Eletronics Engineering Division, Instituto Tecnológico de Aeronáutica, São José dos Campos 12228-900, SP, Brazil
| |
Collapse
|
3
|
Ingle PY, Kim YG. Real-Time Abnormal Object Detection for Video Surveillance in Smart Cities. Sensors (Basel) 2022; 22:s22103862. [PMID: 35632270 PMCID: PMC9143895 DOI: 10.3390/s22103862] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/26/2022] [Revised: 05/17/2022] [Accepted: 05/17/2022] [Indexed: 01/26/2023]
Abstract
With the adaptation of video surveillance in many areas for object detection, monitoring abnormal behavior in several cameras requires constant human tracking for a single camera operative, which is a tedious task. In multiview cameras, accurately detecting different types of guns and knives and classifying them from other video surveillance objects in real-time scenarios is difficult. Most detecting cameras are resource-constrained devices with limited computational capacities. To mitigate this problem, we proposed a resource-constrained lightweight subclass detection method based on a convolutional neural network to classify, locate, and detect different types of guns and knives effectively and efficiently in a real-time environment. In this paper, the detection classifier is a multiclass subclass detection convolutional neural network used to classify object frames into different sub-classes such as abnormal and normal. The achieved mean average precision by the best state-of-the-art framework to detect either a handgun or a knife is 84.21% or 90.20% on a single camera view. After extensive experiments, the best precision obtained by the proposed method for detecting different types of guns and knives was 97.50% on the ImageNet dataset and IMFDB, 90.50% on the open-image dataset, 93% on the Olmos dataset, and 90.7% precision on the multiview cameras. This resource-constrained device has shown a satisfactory result, with a precision score of 85.5% for detection in a multiview camera.
Collapse
|
4
|
Kim H, Ishikawa M. Sub-Frame Evaluation of Frame Synchronization for Camera Network Using Linearly Oscillating Light Spot. Sensors (Basel) 2021; 21:6148. [PMID: 34577353 DOI: 10.3390/s21186148] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/28/2021] [Revised: 09/10/2021] [Accepted: 09/10/2021] [Indexed: 11/17/2022]
Abstract
Precisely evaluating the frame synchronization of the camera network is often required for accurate data fusion from multiple visual information. This paper presents a novel method to estimate the synchronization accuracy by using inherent visual information of linearly oscillating light spot captured in the camera images instead of using luminescence information or depending on external measurement instrument. The suggested method is compared to the conventional evaluation method to prove the feasibility. Our experiment result implies that the estimation accuracy of the frame synchronization can be achieved in sub-millisecond order.
Collapse
|
5
|
Mandischer N, Huhn T, Hüsing M, Corves B. Efficient and Consumer-Centered Item Detection and Classification with a Multi camera Network at High Ranges. Sensors (Basel) 2021; 21:s21144818. [PMID: 34300558 PMCID: PMC8309894 DOI: 10.3390/s21144818] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/25/2021] [Revised: 07/08/2021] [Accepted: 07/09/2021] [Indexed: 11/25/2022]
Abstract
In the EU project SHAREWORK, methods are developed that allow humans and robots to collaborate in an industrial environment. One of the major contributions is a framework for task planning coupled with automated item detection and localization. In this work, we present the methods used for detecting and classifying items on the shop floor. Important in the context of SHAREWORK is the user-friendliness of the methodology. Thus, we renounce heavy-learning-based methods in favor of unsupervised segmentation coupled with lenient machine learning methods for classification. Our algorithm is a combination of established methods adjusted for fast and reliable item detection at high ranges of up to eight meters. In this work, we present the full pipeline from calibration, over segmentation to item classification in the industrial context. The pipeline is validated on a shop floor of 40 sqm and with up to nine different items and assemblies, reaching a mean accuracy of 84% at 0.85 Hz.
Collapse
|
6
|
Truong AM, Philips W, Deligiannis N, Abrahamyan L, Guan J. Automatic Multi-Camera Extrinsic Parameter Calibration Based on Pedestrian Torsors. Sensors (Basel) 2019; 19:E4989. [PMID: 31731824 PMCID: PMC6891296 DOI: 10.3390/s19224989] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 09/25/2019] [Revised: 11/04/2019] [Accepted: 11/12/2019] [Indexed: 11/18/2022]
Abstract
Extrinsic camera calibration is essential for any computer vision task in a camera network. Typically, researchers place a calibration object in the scene to calibrate all the cameras in a camera network. However, when installing cameras in the field, this approach can be costly and impractical, especially when recalibration is needed. This paper proposes a novel, accurate and fully automatic extrinsic calibration framework for camera networks with partially overlapping views. The proposed method considers the pedestrians in the observed scene as the calibration objects and analyzes the pedestrian tracks to obtain extrinsic parameters. Compared to the state of the art, the new method is fully automatic and robust in various environments. Our method detect human poses in the camera images and then models walking persons as vertical sticks. We apply a brute-force method to determines the correspondence between persons in multiple camera images. This information along with 3D estimated locations of the top and the bottom of the pedestrians are then used to compute the extrinsic calibration matrices. We also propose a novel method to calibrate the camera network by only using the top and centerline of the person when the bottom of the person is not available in heavily occluded scenes. We verified the robustness of the method in different camera setups and for both single and multiple walking people. The results show that the triangulation error of a few centimeters can be obtained. Typically, it requires less than one minute of observing the walking people to reach this accuracy in controlled environments. It also just takes a few minutes to collect enough data for the calibration in uncontrolled environments. Our proposed method can perform well in various situations such as multi-person, occlusions, or even at real intersections on the street.
Collapse
Affiliation(s)
- Anh Minh Truong
- TELIN-IPI, Ghent University—imec, St-Pietersnieuwstraat 41, B-9000 Gent, Belgium;
| | - Wilfried Philips
- TELIN-IPI, Ghent University—imec, St-Pietersnieuwstraat 41, B-9000 Gent, Belgium;
| | - Nikos Deligiannis
- ETRO Department, Vrije Universiteit Brussel—imec, Pleinlaan 2, B-1050 Brussels, Belgium; (N.D.); (L.A.)
| | - Lusine Abrahamyan
- ETRO Department, Vrije Universiteit Brussel—imec, Pleinlaan 2, B-1050 Brussels, Belgium; (N.D.); (L.A.)
| | - Junzhi Guan
- CETC Key Laboratory of Aerospace Information Applications, Shijiazhuang 050000, China;
| |
Collapse
|
7
|
Guan J, Deboeverie F, Slembrouck M, Van Haerenborgh D, Van Cauwelaert D, Veelaert P, Philips W. Extrinsic Calibration of Camera Networks Based on Pedestrians. Sensors (Basel) 2016; 16:E654. [PMID: 27171080 DOI: 10.3390/s16050654] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/24/2016] [Revised: 04/25/2016] [Accepted: 05/04/2016] [Indexed: 11/17/2022]
Abstract
In this paper, we propose a novel extrinsic calibration method for camera networks by analyzing tracks of pedestrians. First of all, we extract the center lines of walking persons by detecting their heads and feet in the camera images. We propose an easy and accurate method to estimate the 3D positions of the head and feet w.r.t. a local camera coordinate system from these center lines. We also propose a RANSAC-based orthogonal Procrustes approach to compute relative extrinsic parameters connecting the coordinate systems of cameras in a pairwise fashion. Finally, we refine the extrinsic calibration matrices using a method that minimizes the reprojection error. While existing state-of-the-art calibration methods explore epipolar geometry and use image positions directly, the proposed method first computes 3D positions per camera and then fuses the data. This results in simpler computations and a more flexible and accurate calibration method. Another advantage of our method is that it can also handle the case of persons walking along straight lines, which cannot be handled by most of the existing state-of-the-art calibration methods since all head and feet positions are co-planar. This situation often happens in real life.
Collapse
|
8
|
Sanfeliu A, Andrade-Cetto J, Barbosa M, Bowden R, Capitán J, Corominas A, Gilbert A, Illingworth J, Merino L, Mirats JM, Moreno P, Ollero A, Sequeira J, Spaan MT. Decentralized sensor fusion for Ubiquitous Networking Robotics in Urban Areas. Sensors (Basel) 2010; 10:2274-314. [PMID: 22294927 PMCID: PMC3264480 DOI: 10.3390/s100302274] [Citation(s) in RCA: 31] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/12/2010] [Revised: 02/02/2010] [Accepted: 02/28/2010] [Indexed: 11/30/2022]
Abstract
In this article we explain the architecture for the environment and sensors that has been built for the European project URUS (Ubiquitous Networking Robotics in Urban Sites), a project whose objective is to develop an adaptable network robot architecture for cooperation between network robots and human beings and/or the environment in urban areas. The project goal is to deploy a team of robots in an urban area to give a set of services to a user community. This paper addresses the sensor architecture devised for URUS and the type of robots and sensors used, including environment sensors and sensors onboard the robots. Furthermore, we also explain how sensor fusion takes place to achieve urban outdoor execution of robotic services. Finally some results of the project related to the sensor network are highlighted.
Collapse
Affiliation(s)
- Alberto Sanfeliu
- Institut de Robòtica i Informàtica Industrial, CSIC-UPC, Barcelona, Spain; E-Mails: (J.A.-C.); (A.C.); (J.M.M.)
- Author to whom correspondence should be addressed; E-Mail: ; Tel.: +34-934015751; Fax: +34-934015750
| | - Juan Andrade-Cetto
- Institut de Robòtica i Informàtica Industrial, CSIC-UPC, Barcelona, Spain; E-Mails: (J.A.-C.); (A.C.); (J.M.M.)
| | - Marco Barbosa
- Instituto Superior Técnico & Institute for Systems and Robotics, Lisbon, Portugal; E-Mails: (M.B.); (P.M.); (J.S.); (M.T.J.S.)
| | - Richard Bowden
- Centre for Vision Speech and Signal Processing, University of Surrey, Guildford, UK; E-Mails: (R.B.); (A.G.); (J.I.)
| | - Jesús Capitán
- Robotics, Vision and Control Group, University of Seville, Seville, Spain; E-Mail:
| | - Andreu Corominas
- Institut de Robòtica i Informàtica Industrial, CSIC-UPC, Barcelona, Spain; E-Mails: (J.A.-C.); (A.C.); (J.M.M.)
| | - Andrew Gilbert
- Centre for Vision Speech and Signal Processing, University of Surrey, Guildford, UK; E-Mails: (R.B.); (A.G.); (J.I.)
| | - John Illingworth
- Centre for Vision Speech and Signal Processing, University of Surrey, Guildford, UK; E-Mails: (R.B.); (A.G.); (J.I.)
| | - Luis Merino
- Pablo de Olavide University, Seville, Spain; E-Mail:
| | - Josep M. Mirats
- Institut de Robòtica i Informàtica Industrial, CSIC-UPC, Barcelona, Spain; E-Mails: (J.A.-C.); (A.C.); (J.M.M.)
| | - Plínio Moreno
- Instituto Superior Técnico & Institute for Systems and Robotics, Lisbon, Portugal; E-Mails: (M.B.); (P.M.); (J.S.); (M.T.J.S.)
| | - Aníbal Ollero
- Robotics, Vision and Control Group, University of Seville, Seville, Spain; E-Mail:
- Center for Advanced Aerospace Technology, Seville, Spain
| | - João Sequeira
- Instituto Superior Técnico & Institute for Systems and Robotics, Lisbon, Portugal; E-Mails: (M.B.); (P.M.); (J.S.); (M.T.J.S.)
| | - Matthijs T.J. Spaan
- Instituto Superior Técnico & Institute for Systems and Robotics, Lisbon, Portugal; E-Mails: (M.B.); (P.M.); (J.S.); (M.T.J.S.)
| |
Collapse
|