1
|
Zhou P, Shi L, Liu X, Jin J, Zhang Y, Hou J. Light Field Depth Estimation via Stitched Epipolar Plane Images. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2024; 30:6866-6879. [PMID: 38113148 DOI: 10.1109/tvcg.2023.3344132] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/21/2023]
Abstract
Depth estimation is a fundamental problem in light field processing. Epipolar-plane image (EPI)-based methods often encounter challenges such as low accuracy in slope computation due to discretization errors and limited angular resolution. Besides, existing methods perform well in most regions but struggle to produce sharp edges in occluded regions and resolve ambiguities in texture-less regions. To address these issues, we propose the concept of stitched-EPI (SEPI) to enhance slope computation. SEPI achieves this by shifting and concatenating lines from different EPIs that correspond to the same 3D point. Moreover, we introduce the half-SEPI algorithm, which focuses exclusively on the non-occluded portion of lines to handle occlusion. Additionally, we present a depth propagation strategy aimed at improving depth estimation in texture-less regions. This strategy involves determining the depth of such regions by progressing from the edges towards the interior, prioritizing accurate regions over coarse regions. Through extensive experimental evaluations and ablation studies, we validate the effectiveness of our proposed method. The results demonstrate its superior ability to generate more accurate and robust depth maps across all regions compared to state-of-the-art methods.
Collapse
|
2
|
Fu J, Zhang Y, Li Y, Li J, Xiong Z. Fast 3D reconstruction via event-based structured light with spatio-temporal coding. OPTICS EXPRESS 2023; 31:44588-44602. [PMID: 38178526 DOI: 10.1364/oe.507688] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/11/2023] [Accepted: 11/26/2023] [Indexed: 01/06/2024]
Abstract
Event-based structured light (SL) systems leverage bio-inspired event cameras, which are renowned for their low latency and high dynamics, to drive progress in high-speed structured light systems. However, existing event-based structured light methods concentrate on the independent construction of either time-domain or space-domain features for stereo matching, ignoring the spatio-temporal consistency towards depth. In this work, we build an event-based SL system that consists of a laser point projector and an event camera, and we devise a spatial-temporal coding strategy that realizes depth encoding in dual domains through a single shot. To exploit the spatio-temporal synergy, we further present STEM, a novel Spatio-Temporal Enhanced Matching approach for 3D reconstruction. STEM is comprised of two parts, the spatio-temporal enhancing (STE) algorithm and the spatio-temporal matching (STM) algorithm. Specifically, STE integrates the dual-domain information to increase the saliency of the temporal coding, providing a more robust basis for matching. STM is a stereo matching algorithm explicitly tailored to the unique characteristics of event data modality, which computes the disparity via a meticulously designed hybrid cost function. Experimental results demonstrate the superior performance of our proposed method, achieving a reconstruction rate of 16 fps and a low root mean square error of 0.56 mm at a distance of 0.72 m.
Collapse
|
3
|
Wang J, Li L, Xu P. Visual Sensing and Depth Perception for Welding Robots and Their Industrial Applications. SENSORS (BASEL, SWITZERLAND) 2023; 23:9700. [PMID: 38139548 PMCID: PMC10747874 DOI: 10.3390/s23249700] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/14/2023] [Revised: 11/30/2023] [Accepted: 12/06/2023] [Indexed: 12/24/2023]
Abstract
With the rapid development of vision sensing, artificial intelligence, and robotics technology, one of the challenges we face is installing more advanced vision sensors on welding robots to achieve intelligent welding manufacturing and obtain high-quality welding components. Depth perception is one of the bottlenecks in the development of welding sensors. This review provides an assessment of active and passive sensing methods for depth perception and classifies and elaborates on the depth perception mechanisms based on monocular vision, binocular vision, and multi-view vision. It explores the principles and means of using deep learning for depth perception in robotic welding processes. Further, the application of welding robot visual perception in different industrial scenarios is summarized. Finally, the problems and countermeasures of welding robot visual perception technology are analyzed, and developments for the future are proposed. This review has analyzed a total of 2662 articles and cited 152 as references. The potential future research topics are suggested to include deep learning for object detection and recognition, transfer deep learning for welding robot adaptation, developing multi-modal sensor fusion, integrating models and hardware, and performing a comprehensive requirement analysis and system evaluation in collaboration with welding experts to design a multi-modal sensor fusion architecture.
Collapse
Affiliation(s)
- Ji Wang
- Department of Chemical and Materials Engineering, University of Alberta, Edmonton, AB T6G 1H9, Canada
- School of Materials Science and Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
| | - Leijun Li
- Department of Chemical and Materials Engineering, University of Alberta, Edmonton, AB T6G 1H9, Canada
| | - Peiquan Xu
- School of Materials Science and Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
| |
Collapse
|
4
|
Gao S, Li J, Wang P, Xu X. Learning to detect 3D planes from a single plenoptic image. APPLIED OPTICS 2023; 62:5050-5056. [PMID: 37707205 DOI: 10.1364/ao.491874] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/03/2023] [Accepted: 06/06/2023] [Indexed: 09/15/2023]
Abstract
In this paper, we propose a deep neural network to detect 3D planar surfaces from single plenoptic images captured by a Lytro Illum camera. Different from learning methods based on a single RGB image, we train a network to exploit the light distribution information both in spatial and angular dimensions from multi-sub-aperture images. The features from each sub-aperture image are extracted by using parameter-sharing convolutional layers and then fused to jointly infer the parameters of planes, depths, and segmentation masks. The experiments demonstrate that our approach outperforms the existing state-of-the-art methods with significant margins in the plane detection metrics.
Collapse
|
5
|
Shi L, Liu C, He D, Zhao X, Qiu J. Matching entropy based disparity estimation from light field data. OPTICS EXPRESS 2023; 31:6111-6131. [PMID: 36823876 DOI: 10.1364/oe.479741] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/02/2022] [Accepted: 01/17/2023] [Indexed: 06/18/2023]
Abstract
A major challenge for matching-based disparity estimation from light field data is to prevent mismatches in occlusion and smooth regions. An effective matching window satisfying three characteristics: texture richness, disparity consistency, and anti-occlusion should be able to prevent mismatches to some extent. According to these characteristics, we propose matching entropy in the spatial domain of the light field to measure the amount of correct information in a matching window, which provides the criterion for matching window selection. Based on matching entropy regularization, we establish an optimization model for disparity estimation with a matching cost fidelity term. To find the optimum, we propose a two-step adaptive matching algorithm. First, the region type is adaptively determined to identify occluding, occluded, smooth, and textured regions. Then, the matching entropy criterion is used to adaptively select the size and shape of matching windows, as well as the visible viewpoints. The two-step process can reduce mismatches and redundant calculations by selecting effective matching windows. The experimental results on synthetic and real data show that the proposed method can effectively improve the accuracy of disparity estimation in occlusion and smooth regions and has strong robustness for different noise levels. Therefore, high-precision disparity estimation from 4D light field data is achieved.
Collapse
|
6
|
Zhou P, Wang Y, Xu Y, Cai Z, Zuo C. Phase-unwrapping-free 3D reconstruction in structured light field system based on varied auxiliary point. OPTICS EXPRESS 2022; 30:29957-29968. [PMID: 36242108 DOI: 10.1364/oe.468049] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/16/2022] [Accepted: 07/20/2022] [Indexed: 06/16/2023]
Abstract
Three-dimensional (3D) reconstruction is an essential task in structured light field (SLF) related techniques and applications. This paper presents a new method to reconstruct a 3D object point by using many auxiliary points adjacent to it. The relationship between two points in a SLF system is derived. Different from conventional "direct" methods that reconstruct 3D coordinates of the object point by using phase, slope, disparity etc., the proposed method is an "indirect" method as the 3D coordinates of auxiliary points are not needed. Based on the auxiliary point theory, the wrapped phase obtained by 4-step phase-shifting method is sufficient for 3D reconstruction, without the need for phase unwrapping. To the best of our knowledge, this is the first strategy that combines the intrinsic characteristics of structured light and light field for phase-unwrapping-free 3D reconstruction. This paper also analyzes the constraints between system architecture parameters and phase rectification, phase to depth ratio, and presents a relatively simple criterion to guide the system design. Experimental results show that, with an appropriate system architecture, the proposed method can realize accurate, unambiguous, and reliable 3D reconstruction without phase unwrapping.
Collapse
|
7
|
Meng W, Quanyao H, Yongkai Y, Yang Y, Qijian T, Xiang P, Xiaoli L. Large DOF microscopic fringe projection profilometry with a coaxial light-field structure. OPTICS EXPRESS 2022; 30:8015-8026. [PMID: 35299552 DOI: 10.1364/oe.452361] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/05/2022] [Accepted: 02/15/2022] [Indexed: 06/14/2023]
Abstract
Fringe projection profilometry (FPP) has been widely researched for three-dimensional (3D) microscopic measurement during recent decades. Nevertheless, some disadvantages arising from the limited depth of field and occlusion still exist and need to be further addressed. In this paper, light field imaging is introduced for microscopic fringe projection profilometry (MFPP) to obtain a larger depth of field. Meanwhile, this system is built with a coaxial structure to reduce occlusion, where the principle of triangulation is no longer applicable. In this situation, the depth information is estimated based on the epipolar plane image (EPI) of light field. In order to make a quantitative measurement, a metric calibration method which establishes the mapping between the slope of the line feature in EPI and the depth information is proposed for this system. Finally, a group of experiments demonstrate that the proposed LF-MFPP system can work well for depth estimation with a large DOF and reduced occlusion.
Collapse
|
8
|
Xiang S, Liu L, Deng H, Wu J, Yang Y, Yu L. Fast depth estimation with cost minimization for structured light field. OPTICS EXPRESS 2021; 29:30077-30093. [PMID: 34614738 DOI: 10.1364/oe.434548] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/21/2021] [Accepted: 08/24/2021] [Indexed: 06/13/2023]
Abstract
Depth estimation is a fundamental task in light field (LF) related applications. However, conventional light field suffers from the lack of features, which introduces depth ambiguity and heavy computation load to depth estimation. In this paper, we introduce phase light field (PLF), which uses sinusoidal fringes as patterns and the latent phases as the codes. With PLF and the re-formatted phase-epipolar-plane-images (phase EPIs), a global cost minimization framework is proposed to estimate the depth. In general, EPI-based depth estimation tests a set of candidate lines to find the optimal one with most similar intensities, and the slope of the optimal line is converted to disparity and depth. Based on this principle, for phase-EPI, we propose a cost with weighted phase variance in the candidate line, and we prove that the cost is a convex function. After that, the beetle antennae search (BAS) optimization algorithm is utilized to find the optimal line and thus depth can be obtained. Finally, a bilateral filter is incorporated to further improve the depth quality. Simulation and real experimental results demonstrate that, the proposed method can produce accurate depth maps, especially at boundary regions. Moreover, the proposed method achieves an acceleration of about 5.9 times over the state-of-the-art refocus method with comparable depth quality, and thus can facilitate practical applications.
Collapse
|
9
|
Li H, Yu Y, Peng J, Wu Y, Zhang Y. Resolution Improvement of Light Field Imaging via a Nematic Liquid Crystal Microlens with Added Multi-Walled Carbon Nanotubes. SENSORS 2020; 20:s20195557. [PMID: 32998348 PMCID: PMC7582984 DOI: 10.3390/s20195557] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/16/2020] [Revised: 09/21/2020] [Accepted: 09/25/2020] [Indexed: 11/23/2022]
Abstract
A relatively simple method to improve the image resolution of light field based on a liquid crystal (LC) microlens doped with multi-walled carbon nanotubes (MWCNTs) was developed and evaluated. As the nanoparticles were doped in LC, its electro-optical features could enhance, leading to a short response time compared to the pure LC microlens. With the maximum use of the proposed LC microlens, a method combining aperiodicity extraction and weighted average algorithm was adopted to realize the high-resolution light field imaging. The aperiodicity extraction method was proposed, which could effectively improve resolution of view angle image. For synthesizing the full resolution image at 0 Vrms and the extracted view angle image of light field imaging at 2.0 Vrms, the final high-resolution light field imaging could be obtained in a short time by weighted average algorithm. In this way, the common problem of low resolution in light field imaging could be solved. This proposed method was in good agreement with our experimental results. And it was also in line with the development of the trend of the smart imaging sensor combining algorithm with hardware.
Collapse
Affiliation(s)
- Hui Li
- School of Computer Science and Engineering, Wuhan Institute of Technology, Wuhan 430205, China; (Y.Y.); (J.P.); (Y.Z.)
- Hubei Key Laboratory of Intelligent Robot, Wuhan 430205, China
- School of Chemistry and Chemical Engineering, Huazhong University of Science and Technology, Wuhan 430074, China
- Correspondence: (H.L.); (Y.W.)
| | - Yi Yu
- School of Computer Science and Engineering, Wuhan Institute of Technology, Wuhan 430205, China; (Y.Y.); (J.P.); (Y.Z.)
- Hubei Key Laboratory of Intelligent Robot, Wuhan 430205, China
| | - Jing Peng
- School of Computer Science and Engineering, Wuhan Institute of Technology, Wuhan 430205, China; (Y.Y.); (J.P.); (Y.Z.)
- Hubei Key Laboratory of Intelligent Robot, Wuhan 430205, China
| | - Yuntao Wu
- School of Computer Science and Engineering, Wuhan Institute of Technology, Wuhan 430205, China; (Y.Y.); (J.P.); (Y.Z.)
- Hubei Key Laboratory of Intelligent Robot, Wuhan 430205, China
- Correspondence: (H.L.); (Y.W.)
| | - Yanduo Zhang
- School of Computer Science and Engineering, Wuhan Institute of Technology, Wuhan 430205, China; (Y.Y.); (J.P.); (Y.Z.)
- Hubei Key Laboratory of Intelligent Robot, Wuhan 430205, China
| |
Collapse
|
10
|
Marrugo AG, Gao F, Zhang S. State-of-the-art active optical techniques for three-dimensional surface metrology: a review [Invited]. JOURNAL OF THE OPTICAL SOCIETY OF AMERICA. A, OPTICS, IMAGE SCIENCE, AND VISION 2020; 37:B60-B77. [PMID: 32902422 DOI: 10.1364/josaa.398644] [Citation(s) in RCA: 51] [Impact Index Per Article: 12.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/27/2020] [Accepted: 07/07/2020] [Indexed: 05/27/2023]
Abstract
This paper reviews recent developments of non-contact three-dimensional (3D) surface metrology using an active structured optical probe. We focus primarily on those active non-contact 3D surface measurement techniques that could be applicable to the manufacturing industry. We discuss principles of each technology, and its advantageous characteristics as well as limitations. Towards the end, we discuss our perspectives on the current technological challenges in designing and implementing these methods in practical applications.
Collapse
|
11
|
Cai Z, Liu X, Pedrini G, Osten W, Peng X. Light-field depth estimation considering plenoptic imaging distortion. OPTICS EXPRESS 2020; 28:4156-4168. [PMID: 32122073 DOI: 10.1364/oe.385285] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/09/2019] [Accepted: 01/23/2020] [Indexed: 06/10/2023]
Abstract
Light-field imaging can simultaneously record spatio-angular information of light rays to carry out depth estimation via depth cues which reflect a coupling of the angular information and the scene depth. However, the unavoidable imaging distortion in a light-field imaging system has a side effect on the spatio-angular coordinate computation, leading to incorrectly estimated depth maps. Based on the previously established unfocused plenoptic metric model, this paper reports a study on the effect of the plenoptic imaging distortion on the light-field depth estimation. A method of light-field depth estimation considering the plenoptic imaging distortion is proposed. Besides, the accuracy analysis of the light-field depth estimation was performed by using standard components. Experimental results demonstrate that efficiently compensating the plenoptic imaging distortion results in a six-fold improvement in measuring accuracy and more consistency across the measuring depth range. Consequently, the proposed method is proved to be suitable for light-field depth estimation and three-dimensional measurement with high quality, enabling unfocused plenoptic cameras to be metrological tools in the potential application scenarios such as industry, biomedicine, entertainment, and many others.
Collapse
|
12
|
Choi S, Min SW. Monocular depth estimation method using a focus tunable lens. APPLIED OPTICS 2019; 58:G52-G60. [PMID: 31873485 DOI: 10.1364/ao.58.000g52] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/17/2019] [Accepted: 08/23/2019] [Indexed: 06/10/2023]
Abstract
We propose a method for the passive-type estimation of depth using a focus tunable lens. The proposed method utilizes a lens group and a focus tunable lens and charge coupled device. The target object is imaged by the group of lenses and measured by changing the focal length of the focus tunable lens. The method aims to measure depth information in the image domain and not in the object domain. An autofocusing algorithm finds the best focus position of the target object through the focus value calculated by the Sobel operator. We believe that the proposed method is applicable to depth measurement systems because it offers a simple configuration without any active light source and can operate in real time. The experiment, performed for comparison with theoretical calculations, confirms the feasibility of the proposed method.
Collapse
|