1
|
Lesport Q, Joerger G, Kaminski HJ, Girma H, McNett S, Abu-Rub M, Garbey M. Eye Segmentation Method for Telehealth: Application to the Myasthenia Gravis Physical Examination. SENSORS (BASEL, SWITZERLAND) 2023; 23:7744. [PMID: 37765800 PMCID: PMC10536520 DOI: 10.3390/s23187744] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/19/2023] [Revised: 08/28/2023] [Accepted: 09/04/2023] [Indexed: 09/29/2023]
Abstract
Due to the precautions put in place during the COVID-19 pandemic, utilization of telemedicine has increased quickly for patient care and clinical trials. Unfortunately, teleconsultation is closer to a video conference than a medical consultation, with the current solutions setting the patient and doctor into an evaluation that relies entirely on a two-dimensional view of each other. We are developing a patented telehealth platform that assists with diagnostic testing of ocular manifestations of myasthenia gravis. We present a hybrid algorithm combining deep learning with computer vision to give quantitative metrics of ptosis and ocular muscle fatigue leading to eyelid droop and diplopia. The method works both on a fixed image and frame by frame of the video in real-time, allowing capture of dynamic muscular weakness during the examination. We then use signal processing and filtering to derive robust metrics of ptosis and l ocular misalignment. In our construction, we have prioritized the robustness of the method versus accuracy obtained in controlled conditions in order to provide a method that can operate in standard telehealth conditions. The approach is general and can be applied to many disorders of ocular motility and ptosis.
Collapse
Affiliation(s)
- Quentin Lesport
- Department of Surgery, School of Medicine and Health Sciences, George Washington University, Washington, DC 20037, USA;
| | | | - Henry J. Kaminski
- Department of Neurology & Rehabilitation Medicine, School of Medicine and Health Sciences, George Washington University, Washington, DC 20037, USA; (H.J.K.); (H.G.); (S.M.); (M.A.-R.)
| | - Helen Girma
- Department of Neurology & Rehabilitation Medicine, School of Medicine and Health Sciences, George Washington University, Washington, DC 20037, USA; (H.J.K.); (H.G.); (S.M.); (M.A.-R.)
| | - Sienna McNett
- Department of Neurology & Rehabilitation Medicine, School of Medicine and Health Sciences, George Washington University, Washington, DC 20037, USA; (H.J.K.); (H.G.); (S.M.); (M.A.-R.)
| | - Mohammad Abu-Rub
- Department of Neurology & Rehabilitation Medicine, School of Medicine and Health Sciences, George Washington University, Washington, DC 20037, USA; (H.J.K.); (H.G.); (S.M.); (M.A.-R.)
| | - Marc Garbey
- Department of Surgery, School of Medicine and Health Sciences, George Washington University, Washington, DC 20037, USA;
- Care Constitution Corp., Newark, DE 19702, USA;
- LaSIE, UMR CNRS 7356, Université de la Rochelle, 17000 La Rochelle, France
| |
Collapse
|
2
|
Wilson VAD, Bethell EJ, Nawroth C. The use of gaze to study cognition: limitations, solutions, and applications to animal welfare. Front Psychol 2023; 14:1147278. [PMID: 37205074 PMCID: PMC10185774 DOI: 10.3389/fpsyg.2023.1147278] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/18/2023] [Accepted: 04/17/2023] [Indexed: 05/21/2023] Open
Abstract
The study of gaze responses, typically using looking time paradigms, has become a popular approach to improving our understanding of cognitive processes in non-verbal individuals. Our interpretation of data derived from these paradigms, however, is constrained by how we conceptually and methodologically approach these problems. In this perspective paper, we outline the application of gaze studies in comparative cognitive and behavioral research and highlight current limitations in the interpretation of commonly used paradigms. Further, we propose potential solutions, including improvements to current experimental approaches, as well as broad-scale benefits of technology and collaboration. Finally, we outline the potential benefits of studying gaze responses from an animal welfare perspective. We advocate the implementation of these proposals across the field of animal behavior and cognition to aid experimental validity, and further advance our knowledge on a variety of cognitive processes and welfare outcomes.
Collapse
Affiliation(s)
- Vanessa A. D. Wilson
- Department of Comparative Cognition, Institute of Biology, University of Neuchâtel, Neuchâtel, Switzerland
- Department of Comparative Language Science, University of Zurich, Zurich, Switzerland
- Center for the Interdisciplinary Study of Language Evolution (ISLE), University of Zurich, Zurich, Switzerland
- *Correspondence: Vanessa A. D. Wilson,
| | - Emily J. Bethell
- Research Centre in Evolutionary Anthropology and Palaeoecology, Liverpool John Moores University, Liverpool, United Kingdom
| | - Christian Nawroth
- Institute of Behavioural Physiology, Research Institute for Farm Animal Biology (FBN), Dummerstorf, Germany
| |
Collapse
|
3
|
Gibertoni G, Borghi G, Rovati L. Vision-Based Eye Image Classification for Ophthalmic Measurement Systems. SENSORS (BASEL, SWITZERLAND) 2022; 23:386. [PMID: 36616983 PMCID: PMC9823474 DOI: 10.3390/s23010386] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/07/2022] [Revised: 12/20/2022] [Accepted: 12/23/2022] [Indexed: 06/17/2023]
Abstract
The accuracy and the overall performances of ophthalmic instrumentation, where specific analysis of eye images is involved, can be negatively influenced by invalid or incorrect frames acquired during everyday measurements of unaware or non-collaborative human patients and non-technical operators. Therefore, in this paper, we investigate and compare the adoption of several vision-based classification algorithms belonging to different fields, i.e., Machine Learning, Deep Learning, and Expert Systems, in order to improve the performance of an ophthalmic instrument designed for the Pupillary Light Reflex measurement. To test the implemented solutions, we collected and publicly released PopEYE as one of the first datasets consisting of 15 k eye images belonging to 22 different subjects acquired through the aforementioned specialized ophthalmic device. Finally, we discuss the experimental results in terms of classification accuracy of the eye status, as well as computational load analysis, since the proposed solution is designed to be implemented in embedded boards, which have limited hardware resources in computational power and memory size.
Collapse
Affiliation(s)
- Giovanni Gibertoni
- Department of Engineering “Enzo Ferrari”, University of Modena and Reggio Emilia, 41125 Modena, Italy
| | - Guido Borghi
- Department of Computer Science and Engineering, University of Bologna, 40126 Bologna, Italy
| | - Luigi Rovati
- Department of Engineering “Enzo Ferrari”, University of Modena and Reggio Emilia, 41125 Modena, Italy
| |
Collapse
|
4
|
Valenzuela W, Saavedra A, Zarkesh-Ha P, Figueroa M. Motion-Based Object Location on a Smart Image Sensor Using On-Pixel Memory. SENSORS (BASEL, SWITZERLAND) 2022; 22:s22176538. [PMID: 36080999 PMCID: PMC9460117 DOI: 10.3390/s22176538] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/19/2022] [Revised: 08/18/2022] [Accepted: 08/24/2022] [Indexed: 05/27/2023]
Abstract
Object location is a crucial computer vision method often used as a previous stage to object classification. Object-location algorithms require high computational and memory resources, which poses a difficult challenge for portable and low-power devices, even when the algorithm is implemented using dedicated digital hardware. Moving part of the computation to the imager may reduce the memory requirements of the digital post-processor and exploit the parallelism available in the algorithm. This paper presents the architecture of a Smart Imaging Sensor (SIS) that performs object location using pixel-level parallelism. The SIS is based on a custom smart pixel, capable of computing frame differences in the analog domain, and a digital coprocessor that performs morphological operations and connected components to determine the bounding boxes of the detected objects. The smart-pixel array implements on-pixel temporal difference computation using analog memories to detect motion between consecutive frames. Our SIS can operate in two modes: (1) as a conventional image sensor and (2) as a smart sensor which delivers a binary image that highlights the pixels in which movement is detected between consecutive frames and the object bounding boxes. In this paper, we present the design of the smart pixel and evaluate its performance using post-parasitic extraction on a 0.35 µm mixed-signal CMOS process. With a pixel-pitch of 32 µm × 32 µm, we achieved a fill factor of 28%. To evaluate the scalability of the design, we ported the layout to a 0.18 µm process, achieving a fill factor of 74%. On an array of 320×240 smart pixels, the circuit operates at a maximum frame rate of 3846 frames per second. The digital coprocessor was implemented and validated on a Xilinx Artix-7 XC7A35T field-programmable gate array that runs at 125 MHz, locates objects in a video frame in 0.614 µs, and has a power consumption of 58 mW.
Collapse
Affiliation(s)
- Wladimir Valenzuela
- Department of Electrical Engineering, Faculty of Engineering, Universidad de Concepción, Concepción 4070386, Chile
| | - Antonio Saavedra
- Embedded Systems Architecture Group, Institute of Computer Engineering and Microelectronics, Electrical Engineering and Computer Science Faculty, Technische Universität Berlin, 10623 Berlin, Germany
| | - Payman Zarkesh-Ha
- Department of Electrical and Computer Engineering (ECE), School of Engineering, University of New Mexico, Albuquerque, NM 87131-1070, USA
| | - Miguel Figueroa
- Department of Electrical Engineering, Faculty of Engineering, Universidad de Concepción, Concepción 4070386, Chile
| |
Collapse
|
5
|
Li T, An X, Di Y, He J, Liu S, Ming D. Segmentation Method of Cerebral Aneurysms Based on Entropy Selection Strategy. ENTROPY (BASEL, SWITZERLAND) 2022; 24:1062. [PMID: 36010726 PMCID: PMC9407399 DOI: 10.3390/e24081062] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/18/2022] [Revised: 07/21/2022] [Accepted: 07/28/2022] [Indexed: 11/16/2022]
Abstract
The segmentation of cerebral aneurysms is a challenging task because of their similar imaging features to blood vessels and the great imbalance between the foreground and background. However, the existing 2D segmentation methods do not make full use of 3D information and ignore the influence of global features. In this study, we propose an automatic solution for the segmentation of cerebral aneurysms. The proposed method relies on the 2D U-Net as the backbone and adds a Transformer block to capture remote information. Additionally, through the new entropy selection strategy, the network pays more attention to the indistinguishable blood vessels and aneurysms, so as to reduce the influence of class imbalance. In order to introduce global features, three continuous patches are taken as inputs, and a segmentation map corresponding to the central patch is generated. In the inference phase, using the proposed recombination strategy, the segmentation map was generated, and we verified the proposed method on the CADA dataset. We achieved a Dice coefficient (DSC) of 0.944, an IOU score of 0.941, recall of 0.946, an F2 score of 0.942, a mAP of 0.896 and a Hausdorff distance of 3.12 mm.
Collapse
Affiliation(s)
- Tingting Li
- Academy of Medical Engineering and Translational Medicine, Tianjin University, Tianjin 300110, China; (T.L.); (Y.D.); (J.H.); (S.L.); (D.M.)
| | - Xingwei An
- Academy of Medical Engineering and Translational Medicine, Tianjin University, Tianjin 300110, China; (T.L.); (Y.D.); (J.H.); (S.L.); (D.M.)
- Tianjin Center for Brain Science, Tianjin 300110, China
| | - Yang Di
- Academy of Medical Engineering and Translational Medicine, Tianjin University, Tianjin 300110, China; (T.L.); (Y.D.); (J.H.); (S.L.); (D.M.)
| | - Jiaqian He
- Academy of Medical Engineering and Translational Medicine, Tianjin University, Tianjin 300110, China; (T.L.); (Y.D.); (J.H.); (S.L.); (D.M.)
| | - Shuang Liu
- Academy of Medical Engineering and Translational Medicine, Tianjin University, Tianjin 300110, China; (T.L.); (Y.D.); (J.H.); (S.L.); (D.M.)
- Tianjin Center for Brain Science, Tianjin 300110, China
| | - Dong Ming
- Academy of Medical Engineering and Translational Medicine, Tianjin University, Tianjin 300110, China; (T.L.); (Y.D.); (J.H.); (S.L.); (D.M.)
- Tianjin Center for Brain Science, Tianjin 300110, China
- Department of Biomedical Engineering, School of Precision Instruments and Optoelectronics Engineering, Tianjin University, Tianjin 300110, China
| |
Collapse
|
6
|
Masoumian A, Rashwan HA, Cristiano J, Asif MS, Puig D. Monocular Depth Estimation Using Deep Learning: A Review. SENSORS (BASEL, SWITZERLAND) 2022; 22:5353. [PMID: 35891033 PMCID: PMC9325018 DOI: 10.3390/s22145353] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/10/2022] [Revised: 07/01/2022] [Accepted: 07/15/2022] [Indexed: 06/15/2023]
Abstract
In current decades, significant advancements in robotics engineering and autonomous vehicles have improved the requirement for precise depth measurements. Depth estimation (DE) is a traditional task in computer vision that can be appropriately predicted by applying numerous procedures. This task is vital in disparate applications such as augmented reality and target tracking. Conventional monocular DE (MDE) procedures are based on depth cues for depth prediction. Various deep learning techniques have demonstrated their potential applications in managing and supporting the traditional ill-posed problem. The principal purpose of this paper is to represent a state-of-the-art review of the current developments in MDE based on deep learning techniques. For this goal, this paper tries to highlight the critical points of the state-of-the-art works on MDE from disparate aspects. These aspects include input data shapes and training manners such as supervised, semi-supervised, and unsupervised learning approaches in combination with applying different datasets and evaluation indicators. At last, limitations regarding the accuracy of the DL-based MDE models, computational time requirements, real-time inference, transferability, input images shape and domain adaptation, and generalization are discussed to open new directions for future research.
Collapse
Affiliation(s)
- Armin Masoumian
- Department of Computer Engineering and Mathematics, University of Rovira i Virgili, 43007 Tarragona, Spain; (H.A.R.); (J.C.); (D.P.)
- Department of Electrical and Computer Engineering, University of California, Riverside, CA 92521, USA;
| | - Hatem A. Rashwan
- Department of Computer Engineering and Mathematics, University of Rovira i Virgili, 43007 Tarragona, Spain; (H.A.R.); (J.C.); (D.P.)
| | - Julián Cristiano
- Department of Computer Engineering and Mathematics, University of Rovira i Virgili, 43007 Tarragona, Spain; (H.A.R.); (J.C.); (D.P.)
| | - M. Salman Asif
- Department of Electrical and Computer Engineering, University of California, Riverside, CA 92521, USA;
| | - Domenec Puig
- Department of Computer Engineering and Mathematics, University of Rovira i Virgili, 43007 Tarragona, Spain; (H.A.R.); (J.C.); (D.P.)
| |
Collapse
|
7
|
Bethell EJ, Khan W, Hussain A. A deep transfer learning model for head pose estimation in rhesus macaques during cognitive tasks: towards a nonrestraint noninvasive 3Rs approach. Appl Anim Behav Sci 2022. [DOI: 10.1016/j.applanim.2022.105708] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
8
|
Stateful-Service-Based Pupil Recognition in Natural Light Environments. Healthcare (Basel) 2022; 10:healthcare10050789. [PMID: 35627927 PMCID: PMC9140742 DOI: 10.3390/healthcare10050789] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/04/2022] [Revised: 04/20/2022] [Accepted: 04/21/2022] [Indexed: 02/01/2023] Open
Abstract
Smartphones are currently extensively used worldwide, and advances in hardware quality have engendered improvements in smartphone image quality, which is occasionally comparable to the quality of medical imaging systems. This paper proposes two algorithms for pupil recognition: a stateful-service-based pupil recognition mechanism and color component low-pass filtering algorithm. The PRSSM algorithm can determine pupil diameters in images captured in indoor natural light environments, and the CCLPF algorithm can determine pupil diameters in those captured outdoors under sunlight. The PRSSM algorithm converts RGB colors into the hue saturation value color space and performs adaptive thresholding, morphological operations, and contour detection for effectively analyzing the diameter of the pupil. The CCLPF algorithm derives the average matrix for the red components of eye images captured in outdoor environments. It also performs low-pass filtering, morphological and contour detection operations, and rule-of-thumb correction. This algorithm can effectively analyze pupil diameter in outdoor natural light. Traditional ruler-based measurements of pupil diameter were used as the reference to verify the accuracy of the PRSSM and CCLPF algorithms and to compare their accuracy with that of the other algorithm. The errors in pupil diameter data were smaller for the PRSSM and CCLPF algorithms than for the other algorithm.
Collapse
|
9
|
Automatic Asbestos Control Using Deep Learning Based Computer Vision System. APPLIED SCIENCES-BASEL 2021. [DOI: 10.3390/app112210532] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
The paper discusses the results of the research and development of an innovative deep learning-based computer vision system for the fully automatic asbestos content (productivity) estimation in rock chunk (stone) veins in an open pit and within the time comparable with the work of specialists (about 10 min per one open pit processing place). The discussed system is based on the applying of instance and semantic segmentation of artificial neural networks. The Mask R-CNN-based network architecture is applied to the asbestos-containing rock chunks searching images of an open pit. The U-Net-based network architecture is applied to the segmentation of asbestos veins in the images of selected rock chunks. The designed system allows an automatic search and takes images of the asbestos rocks in an open pit in the near-infrared range (NIR) and processes the obtained images. The result of the system work is the average asbestos content (productivity) estimation for each controlled open pit. It is validated to estimate asbestos content as the graduated average ratio of the vein area value to the selected rock chunk area value, both determined by the trained neural network. For both neural network training tasks the training, validation, and test datasets are collected. The designed system demonstrates an error of about 0.4% under different weather conditions in an open pit when the asbestos content is about 1.5–4%. The obtained accuracy is sufficient to use the system as a geological service tool instead of currently applied visual-based estimations.
Collapse
|
10
|
Khan W, Alusi S, Tawfik H, Hussain A. The relationship between non-motor features and weight-loss in the premanifest stage of Huntington's disease. PLoS One 2021; 16:e0253817. [PMID: 34197537 PMCID: PMC8248657 DOI: 10.1371/journal.pone.0253817] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2021] [Accepted: 06/14/2021] [Indexed: 11/19/2022] Open
Abstract
Weight-loss is an integral part of Huntington's disease (HD) that can start before the onset of motor symptoms. Investigating the underlying pathological processes may help in the understanding of this devastating disease as well as contribute to its management. However, the complex behavior and associations of multiple biological factors is impractical to be interpreted by the conventional statistics or human experts. For the first time, we combine a clinical dataset, expert knowledge and machine intelligence to model the multi-dimensional associations between the potentially relevant factors and weight-loss activity in HD, specifically at the premanifest stage. The HD dataset is standardized and transformed into required knowledge base with the help of clinical HD experts, which is then processed by the class rule mining and self-organising maps to identify the significant associations. Statistical results and experts' report indicate a strong association between severe weight-loss in HD at the premanifest stage and measures of certain cognitive, psychiatric functional ability factors. These results suggest that the mechanism underlying weight-loss in HD is, at least partly related to dysfunction of certain areas of the brain, a finding that may have not been apparent otherwise. These associations will aid the understanding of the pathophysiology of the disease and its progression and may in turn help in HD treatment trials.
Collapse
Affiliation(s)
- Wasiq Khan
- School of Computer Science and Mathematics, Liverpool John Moores University, Liverpool, United Kingdom
| | - Sundus Alusi
- Department of Neurology, The Walton Centre NHS Foundation Trust, Liverpool, United Kingdom
| | - Hissam Tawfik
- School of Built Environment, Engineering and Computing, Leeds Beckett University, Leeds, United Kingdom
| | - Abir Hussain
- School of Computer Science and Mathematics, Liverpool John Moores University, Liverpool, United Kingdom
| |
Collapse
|
11
|
Abstract
Simultaneous localization and mapping (SLAM) systems have been generally limited to static environments. Moving objects considerably reduce the location accuracy of SLAM systems, rendering them unsuitable for several applications. Using a combined vision camera and inertial measurement unit (IMU) to separate moving and static objects in dynamic scenes, we improve the location accuracy and adaptability of SLAM systems in these scenes. We develop a moving object-matched feature points elimination algorithm that uses IMU data to eliminate matches on moving objects but retains them on stationary objects. Moreover, we develop a second algorithm to validate the IMU data to avoid erroneous data from influencing image feature points matching. We test the new algorithms with public datasets and in a real-world experiment. In terms of the root mean square error of the location absolute pose error, the proposed method exhibited higher positioning accuracy for the public datasets than the traditional algorithms. Compared with the closed-loop errors obtained by OKVIS-mono and VINS-mono, those obtained in the practical experiment were lower by 50.17% and 56.91%, respectively. Thus, the proposed method eliminates the matching points on moving objects effectively and achieves feature point matching results that are realistic.
Collapse
|
12
|
Schilling A, Maier A, Gerum R, Metzner C, Krauss P. Quantifying the separability of data classes in neural networks. Neural Netw 2021; 139:278-293. [PMID: 33862387 DOI: 10.1016/j.neunet.2021.03.035] [Citation(s) in RCA: 12] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/13/2020] [Revised: 03/23/2021] [Accepted: 03/24/2021] [Indexed: 11/18/2022]
Abstract
We introduce the Generalized Discrimination Value (GDV) that measures, in a non-invasive manner, how well different data classes separate in each given layer of an artificial neural network. It turns out that, at the end of the training period, the GDV in each given layer L attains a highly reproducible value, irrespective of the initialization of the network's connection weights. In the case of multi-layer perceptrons trained with error backpropagation, we find that classification of highly complex data sets requires a temporal reduction of class separability, marked by a characteristic 'energy barrier' in the initial part of the GDV(L) curve. Even more surprisingly, for a given data set, the GDV(L) is running through a fixed 'master curve', independently from the total number of network layers. Finally, due to its invariance with respect to dimensionality, the GDV may serve as a useful tool to compare the internal representational dynamics of artificial neural networks with different architectures for neural architecture search or network compression; or even with brain activity in order to decide between different candidate models of brain function.
Collapse
Affiliation(s)
- Achim Schilling
- Laboratory of Sensory and Cognitive Neuroscience, Aix-Marseille University, Marseille, France; Neuroscience Lab, University Hospital Erlangen, Germany; Cognitive Computational Neuroscience Group, University Erlangen-Nürnberg (FAU), Germany
| | - Andreas Maier
- Chair of Machine Intelligence, University Erlangen-Nürnberg (FAU), Germany
| | - Richard Gerum
- Department of Physics and Center for Vision Research, York University, Toronto, Ontario, Canada
| | - Claus Metzner
- Neuroscience Lab, University Hospital Erlangen, Germany; Chair of Biophysics, University Erlangen-Nürnberg (FAU), Germany
| | - Patrick Krauss
- Neuroscience Lab, University Hospital Erlangen, Germany; Cognitive Computational Neuroscience Group, University Erlangen-Nürnberg (FAU), Germany; Cognitive Neuroscience Center, University of Groningen, The Netherlands.
| |
Collapse
|
13
|
A Fast and Effective System for Analysis of Optokinetic Waveforms with a Low-Cost Eye Tracking Device. Healthcare (Basel) 2020; 9:healthcare9010010. [PMID: 33374811 PMCID: PMC7824545 DOI: 10.3390/healthcare9010010] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/02/2020] [Revised: 12/19/2020] [Accepted: 12/21/2020] [Indexed: 11/16/2022] Open
Abstract
Optokinetic nystagmus (OKN) is an involuntary eye movement induced by motion of a large proportion of the visual field. It consists of a "slow phase (SP)" with eye movements in the same direction as the movement of the pattern and a "fast phase (FP)" with saccadic eye movements in the opposite direction. Study of OKN can reveal valuable information in ophthalmology, neurology and psychology. However, the current commercially available high-resolution and research-grade eye tracker is usually expensive. Methods & Results: We developed a novel fast and effective system combined with a low-cost eye tracking device to accurately quantitatively measure OKN eye movement. Conclusions: The experimental results indicate that the proposed method achieves fast and promising results in comparisons with several traditional approaches.
Collapse
|
14
|
Experimental Verification of Objective Visual Fatigue Measurement Based on Accurate Pupil Detection of Infrared Eye Image and Multi-Feature Analysis. SENSORS 2020; 20:s20174814. [PMID: 32858920 PMCID: PMC7506756 DOI: 10.3390/s20174814] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/08/2020] [Revised: 08/21/2020] [Accepted: 08/24/2020] [Indexed: 12/15/2022]
Abstract
As the use of electronic displays increases rapidly, visual fatigue problems are also increasing. The subjective evaluation methods used for visual fatigue measurement have individual difference problems, while objective methods based on bio-signal measurement have problems regarding motion artifacts. Conventional eye image analysis-based visual fatigue measurement methods do not accurately characterize the complex changes in the appearance of the eye. To solve this problem, in this paper, an objective visual fatigue measurement method based on infrared eye image analysis is proposed. For accurate pupil detection, a convolutional neural network-based semantic segmentation method was used. Three features are calculated based on the pupil detection results: (1) pupil accommodation speed, (2) blink frequency, and (3) eye-closed duration. In order to verify the calculated features, differences in fatigue caused by changes in content color components such as gamma, color temperature, and brightness were compared with a reference video. The pupil detection accuracy was confirmed to be 96.63% based on the mean intersection over union. In addition, it was confirmed that all three features showed significant differences from the reference group; thus, it was verified that the proposed analysis method can be used for the objective measurement of visual fatigue.
Collapse
|