1
|
Greene MR, Balas BJ, Lescroart MD, MacNeilage PR, Hart JA, Binaee K, Hausamann PA, Mezile R, Shankar B, Sinnott CB, Capurro K, Halow S, Howe H, Josyula M, Li A, Mieses A, Mohamed A, Nudnou I, Parkhill E, Riley P, Schmidt B, Shinkle MW, Si W, Szekely B, Torres JM, Weissmann E. The visual experience dataset: Over 200 recorded hours of integrated eye movement, odometry, and egocentric video. J Vis 2024; 24:6. [PMID: 39377740 PMCID: PMC11466363 DOI: 10.1167/jov.24.11.6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2024] [Accepted: 08/13/2024] [Indexed: 10/09/2024] Open
Abstract
We introduce the Visual Experience Dataset (VEDB), a compilation of more than 240 hours of egocentric video combined with gaze- and head-tracking data that offer an unprecedented view of the visual world as experienced by human observers. The dataset consists of 717 sessions, recorded by 56 observers ranging from 7 to 46 years of age. This article outlines the data collection, processing, and labeling protocols undertaken to ensure a representative sample and discusses the potential sources of error or bias within the dataset. The VEDB's potential applications are vast, including improving gaze-tracking methodologies, assessing spatiotemporal image statistics, and refining deep neural networks for scene and activity recognition. The VEDB is accessible through established open science platforms and is intended to be a living dataset with plans for expansion and community contributions. It is released with an emphasis on ethical considerations, such as participant privacy and the mitigation of potential biases. By providing a dataset grounded in real-world experiences and accompanied by extensive metadata and supporting code, the authors invite the research community to use and contribute to the VEDB, facilitating a richer understanding of visual perception and behavior in naturalistic settings.
Collapse
Affiliation(s)
- Michelle R Greene
- Barnard College, Columbia University, New York, NY, USA
- Bates College, Lewiston, ME, USA
| | | | | | | | | | - Kamran Binaee
- University of Nevada, Reno, NV, USA
- Magic Leap, Plantation, FL, USA
| | | | | | - Bharath Shankar
- University of Nevada, Reno, NV, USA
- Unmanned Ground Systems, Chelmsford, MA, USA
| | - Christian B Sinnott
- University of Nevada, Reno, NV, USA
- Smith-Kettlewell Eye Research Institute, San Francisco, CA, USA
| | | | | | | | | | - Annie Li
- Bates College, Lewiston, ME, USA
| | | | | | - Ilya Nudnou
- North Dakota State University, Fargo, ND, USA
| | | | | | | | | | | | | | | | | |
Collapse
|
2
|
Espitia-Mora LA, Vélez-Guerrero MA, Callejas-Cuervo M. Development of a Low-Cost Markerless Optical Motion Capture System for Gait Analysis and Anthropometric Parameter Quantification. SENSORS (BASEL, SWITZERLAND) 2024; 24:3371. [PMID: 38894161 PMCID: PMC11174744 DOI: 10.3390/s24113371] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/12/2024] [Revised: 05/15/2024] [Accepted: 05/20/2024] [Indexed: 06/21/2024]
Abstract
Technological advancements have expanded the range of methods for capturing human body motion, including solutions involving inertial sensors (IMUs) and optical alternatives. However, the rising complexity and costs associated with commercial solutions have prompted the exploration of more cost-effective alternatives. This paper presents a markerless optical motion capture system using a RealSense depth camera and intelligent computer vision algorithms. It facilitates precise posture assessment, the real-time calculation of joint angles, and acquisition of subject-specific anthropometric data for gait analysis. The proposed system stands out for its simplicity and affordability in comparison to complex commercial solutions. The gathered data are stored in comma-separated value (CSV) files, simplifying subsequent analysis and data mining. Preliminary tests, conducted in controlled laboratory environments and employing a commercial MEMS-IMU system as a reference, revealed a maximum relative error of 7.6% in anthropometric measurements, with a maximum absolute error of 4.67 cm at average height. Stride length measurements showed a maximum relative error of 11.2%. Static joint angle tests had a maximum average error of 10.2%, while dynamic joint angle tests showed a maximum average error of 9.06%. The proposed optical system offers sufficient accuracy for potential application in areas such as rehabilitation, sports analysis, and entertainment.
Collapse
Affiliation(s)
| | | | - Mauro Callejas-Cuervo
- Software Research Group, Universidad Pedagógica y Tecnológica de Colombia, Tunja 150002, Colombia; (L.A.E.-M.); (M.A.V.-G.)
| |
Collapse
|
3
|
DuTell V, Gibaldi A, Focarelli G, Olshausen BA, Banks MS. High-fidelity eye, head, body, and world tracking with a wearable device. Behav Res Methods 2024; 56:32-42. [PMID: 35879503 PMCID: PMC10794349 DOI: 10.3758/s13428-022-01888-3] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 05/09/2022] [Indexed: 12/13/2022]
Abstract
We describe the design and performance of a high-fidelity wearable head-, body-, and eye-tracking system that offers significant improvement over previous such devices. This device's sensors include a binocular eye tracker, an RGB-D scene camera, a high-frame-rate scene camera, and two visual odometry sensors, for a total of ten cameras, which we synchronize and record from with a data rate of over 700 MB/s. The sensors are operated by a mini-PC optimized for fast data collection, and powered by a small battery pack. The device records a subject's eye, head, and body positions, simultaneously with RGB and depth data from the subject's visual environment, measured with high spatial and temporal resolution. The headset weighs only 1.4 kg, and the backpack with batteries 3.9 kg. The device can be comfortably worn by the subject, allowing a high degree of mobility. Together, this system overcomes many limitations of previous such systems, allowing high-fidelity characterization of the dynamics of natural vision.
Collapse
Affiliation(s)
- Vasha DuTell
- Wertheim School of Optometry and Vision Science, UC Berkeley, Minor Hall, Berkeley, CA, USA.
- Redwood Center for Theoretical Neuroscience, UC Berkeley, Evans Hall, Berkeley, CA, USA.
| | - Agostino Gibaldi
- Wertheim School of Optometry and Vision Science, UC Berkeley, Minor Hall, Berkeley, CA, USA
| | - Giulia Focarelli
- Wertheim School of Optometry and Vision Science, UC Berkeley, Minor Hall, Berkeley, CA, USA
| | - Bruno A Olshausen
- Wertheim School of Optometry and Vision Science, UC Berkeley, Minor Hall, Berkeley, CA, USA
- Redwood Center for Theoretical Neuroscience, UC Berkeley, Evans Hall, Berkeley, CA, USA
| | - Martin S Banks
- Wertheim School of Optometry and Vision Science, UC Berkeley, Minor Hall, Berkeley, CA, USA
| |
Collapse
|
4
|
Park I, Cho S. Fusion localization for indoor airplane inspection using visual inertial odometry and ultrasonic RTLS. Sci Rep 2023; 13:18117. [PMID: 37872183 PMCID: PMC10593739 DOI: 10.1038/s41598-023-43425-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/20/2023] [Accepted: 09/23/2023] [Indexed: 10/25/2023] Open
Abstract
In this paper, the fusion localization system for the visual inertial odometry (VIO) and ultrasonic real-time localization system (RTLS) for indoor airplane inspection using drones is proposed. In a hangar environment, either trilateration-based RTLS or vision-based localization shows disadvantages and neither can be used alone. In this research, we design a configuration of VIO suitable for hangar environment and outlier filter on ultrasonic RTLS for non-line of sight situations, so that both can be fused using graph optimization. The proposed solution can provide more accurate localization than the visual odometry-only system as well as continue estimating positions in the absence of RTLS data. Localization and real-time performance of the proposed algorithm are evaluated through experimentation in a hangar and a flight test in an outdoor space.
Collapse
|
5
|
Sankepalle DM, Anthony B, Mallidi S. Visual inertial odometry enabled 3D ultrasound and photoacoustic imaging. BIOMEDICAL OPTICS EXPRESS 2023; 14:2756-2772. [PMID: 37342691 PMCID: PMC10278605 DOI: 10.1364/boe.489614] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/10/2023] [Revised: 05/04/2023] [Accepted: 05/04/2023] [Indexed: 06/23/2023]
Abstract
There is an increasing need for 3D ultrasound and photoacoustic (USPA) imaging technology for real-time monitoring of dynamic changes in vasculature or molecular markers in various malignancies. Current 3D USPA systems utilize expensive 3D transducer arrays, mechanical arms or limited-range linear stages to reconstruct the 3D volume of the object being imaged. In this study, we developed, characterized, and demonstrated an economical, portable, and clinically translatable handheld device for 3D USPA imaging. An off-the-shelf, low-cost visual odometry system (the Intel RealSense T265 camera equipped with simultaneous localization and mapping technology) to track free hand movements during imaging was attached to the USPA transducer. Specifically, we integrated the T265 camera into a commercially available USPA imaging probe to acquire 3D images and compared it to the reconstructed 3D volume acquired using a linear stage (ground truth). We were able to reliably detect 500 µm step sizes with 90.46% accuracy. Various users evaluated the potential of handheld scanning, and the volume calculated from the motion-compensated image was not significantly different from the ground truth. Overall, our results, for the first time, established the use of an off-the-shelf and low-cost visual odometry system for freehand 3D USPA imaging that can be seamlessly integrated into several photoacoustic imaging systems for various clinical applications.
Collapse
Affiliation(s)
| | - Brian Anthony
- Institute of Medical Engineering and Sciences, Department of Mechanical Engineering, Massachusetts Institute of Technology, Cambridge, MA, 02139, USA
| | - Srivalleesha Mallidi
- Department of Biomedical Engineering, Tufts University, Medford, MA, 02155, USA
- Wellman Center for Photomedicine, Harvard Medical School, Boston, MA, 02115, USA
| |
Collapse
|
6
|
Sinnott CB, Hausamann PA, MacNeilage PR. Natural statistics of human head orientation constrain models of vestibular processing. Sci Rep 2023; 13:5882. [PMID: 37041176 PMCID: PMC10090077 DOI: 10.1038/s41598-023-32794-z] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/24/2022] [Accepted: 04/02/2023] [Indexed: 04/13/2023] Open
Abstract
Head orientation relative to gravity determines how gravity-dependent environmental structure is sampled by the visual system, as well as how gravity itself is sampled by the vestibular system. Therefore, both visual and vestibular sensory processing should be shaped by the statistics of head orientation relative to gravity. Here we report the statistics of human head orientation during unconstrained natural activities in humans for the first time, and we explore implications for models of vestibular processing. We find that the distribution of head pitch is more variable than head roll and that the head pitch distribution is asymmetrical with an over-representation of downward head pitch, consistent with ground-looking behavior. We further suggest that pitch and roll distributions can be used as empirical priors in a Bayesian framework to explain previously measured biases in perception of both roll and pitch. Gravitational and inertial acceleration stimulate the otoliths in an equivalent manner, so we also analyze the dynamics of human head orientation to better understand how knowledge of these dynamics can constrain solutions to the problem of gravitoinertial ambiguity. Gravitational acceleration dominates at low frequencies and inertial acceleration dominates at higher frequencies. The change in relative power of gravitational and inertial components as a function of frequency places empirical constraints on dynamic models of vestibular processing, including both frequency segregation and probabilistic internal model accounts. We conclude with a discussion of methodological considerations and scientific and applied domains that will benefit from continued measurement and analysis of natural head movements moving forward.
Collapse
Affiliation(s)
| | - Peter A Hausamann
- Department of Electrical and Computer Engineering, Technical University of Munich, 80333, Munich, Germany
| | | |
Collapse
|
7
|
Sinnott C, Hausamann PA, MacNeilage PR. Natural statistics of human head orientation constrain models of vestibular processing. RESEARCH SQUARE 2023:rs.3.rs-2412413. [PMID: 36711500 PMCID: PMC9882651 DOI: 10.21203/rs.3.rs-2412413/v1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 01/12/2023]
Abstract
Head orientation relative to gravity determines how gravity-dependent environmental structure is sampled by the visual system, as well as how gravity itself is sampled by the vestibular system. Therefore, both visual and vestibular sensory processing should be shaped by the statistics of head orientation relative to gravity. Here we report the statistics of human head orientation during unconstrained natural activities in humans for the first time, and we explore implications for models of vestibular processing. We find that the distribution of head pitch is more variable than head roll and that the head pitch distribution is asymmetrical with an over-representation of downward head pitch, consistent with ground-looking behavior. We further show that pitch and roll distributions can be used as empirical priors in a Bayesian framework to explain previously measured biases in perception of both roll and pitch. We also analyze the dynamics of human head orientation to better understand how gravitational and inertial acceleration are processed by the vestibular system. Gravitational acceleration dominates at low frequencies and inertial acceleration dominates at higher frequencies. The change in relative power of gravitational and inertial components as a function of frequency places empirical constraints on dynamic models of vestibular processing. We conclude with a discussion of methodological considerations and scientific and applied domains that will benefit from continued measurement and analysis of natural head movements moving forward.
Collapse
Affiliation(s)
- Christian Sinnott
- University of Nevada, Department of Psychology, Reno, 89557, United States of America,
| | - Peter A. Hausamann
- Technical University of Munich, Department of Electrical and Computer Engineering, Munich, 80333, Germany
| | - Paul R. MacNeilage
- University of Nevada, Department of Psychology, Reno, 89557, United States of America
| |
Collapse
|
8
|
Schyga J, Hinckeldeyn J, Kreutzfeldt J. Meaningful Test and Evaluation of Indoor Localization Systems in Semi-Controlled Environments. SENSORS (BASEL, SWITZERLAND) 2022; 22:2797. [PMID: 35408410 PMCID: PMC9003439 DOI: 10.3390/s22072797] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 02/28/2022] [Revised: 03/30/2022] [Accepted: 03/31/2022] [Indexed: 06/14/2023]
Abstract
Despite their enormous potential, the use of indoor localization systems (ILS) remains seldom. One reason is the lack of market transparency and stakeholders' trust in the systems' performance as a consequence of insufficient use of test and evaluation (T&E) methodologies. The heterogeneous nature of ILS, their influences, and their applications pose various challenges for the design of a methodology that provides meaningful results. Methodologies for building-wide testing exist, but their use is mostly limited to associated indoor localization competitions. In this work, the T&E 4iLoc Framework is proposed-a methodology for T&E of indoor localization systems in semi-controlled environments based on a system-level and black-box approach. In contrast to building-wide testing, T&E in semi-controlled environments, such as test halls, is characterized by lower costs, higher reproducibility, and better comparability of the results. The limitation of low transferability to real-world applications is addressed by an application-driven design approach. The empirical validation of the T&E 4iLoc Framework, based on the examination of a contour-based light detection and ranging (LiDAR) ILS, an ultra wideband ILS, and a camera-based ILS for the application of automated guided vehicles in warehouse operation, demonstrates the benefits of T&E with the T&E 4iLoc Framework.
Collapse
|
9
|
Salchow-Hömmen C, Skrobot M, Jochner MCE, Schauer T, Kühn AA, Wenger N. Review-Emerging Portable Technologies for Gait Analysis in Neurological Disorders. Front Hum Neurosci 2022; 16:768575. [PMID: 35185496 PMCID: PMC8850274 DOI: 10.3389/fnhum.2022.768575] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/31/2021] [Accepted: 01/07/2022] [Indexed: 01/29/2023] Open
Abstract
The understanding of locomotion in neurological disorders requires technologies for quantitative gait analysis. Numerous modalities are available today to objectively capture spatiotemporal gait and postural control features. Nevertheless, many obstacles prevent the application of these technologies to their full potential in neurological research and especially clinical practice. These include the required expert knowledge, time for data collection, and missing standards for data analysis and reporting. Here, we provide a technological review of wearable and vision-based portable motion analysis tools that emerged in the last decade with recent applications in neurological disorders such as Parkinson's disease and Multiple Sclerosis. The goal is to enable the reader to understand the available technologies with their individual strengths and limitations in order to make an informed decision for own investigations and clinical applications. We foresee that ongoing developments toward user-friendly automated devices will allow for closed-loop applications, long-term monitoring, and telemedical consulting in real-life environments.
Collapse
Affiliation(s)
- Christina Salchow-Hömmen
- Department of Neurology With Experimental Neurology, Charité-Universitätsmedizin Berlin, Berlin, Germany
| | - Matej Skrobot
- Department of Neurology With Experimental Neurology, Charité-Universitätsmedizin Berlin, Berlin, Germany
| | - Magdalena C E Jochner
- Department of Neurology With Experimental Neurology, Charité-Universitätsmedizin Berlin, Berlin, Germany
| | - Thomas Schauer
- Control Systems Group, Technische Universität Berlin, Berlin, Germany
| | - Andrea A Kühn
- Department of Neurology With Experimental Neurology, Charité-Universitätsmedizin Berlin, Berlin, Germany
- Berlin School of Mind and Brain, Charité-Universitätsmedizin Berlin, Berlin, Germany
- NeuroCure Clinical Research Centre, Charité-Universitätsmedizin Berlin, Berlin, Germany
- German Center for Neurodegenerative Diseases, DZNE, Berlin, Germany
| | - Nikolaus Wenger
- Department of Neurology With Experimental Neurology, Charité-Universitätsmedizin Berlin, Berlin, Germany
| |
Collapse
|