1
|
Zhu X, Peng X. Strategic assessment model of smart stadiums based on genetic algorithms and literature visualization analysis: A case study from Chengdu, China. Heliyon 2024; 10:e31759. [PMID: 38828338 PMCID: PMC11140808 DOI: 10.1016/j.heliyon.2024.e31759] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2023] [Revised: 05/17/2024] [Accepted: 05/21/2024] [Indexed: 06/05/2024] Open
Abstract
This paper leverages Citespace and VOSviewer software to perform a comprehensive bibliometric analysis on a corpus of 384 references related to smart sports venues, spanning from 1998 to 2022. The analysis encompasses various facets, including author network analysis, institutional network analysis, temporal mapping, keyword clustering, and co-citation network analysis. Moreover, this paper constructs a smart stadiums strategic assessment model (SSSAM) to compensate for confusion and aimlessness by genetic algorithms (GA). Our findings indicate an exponential growth in publications on smart sports venues year over year. Arizona State University emerges as the institution with the highest number of collaborative publications, Energy and Buildings becomes the publication with the most documents. While, Wang X stands out as the scholar with the most substantial contribution to the field. In scrutinizing the betweenness centrality indicators, a paradigm shift in research hotspots becomes evident-from intelligent software to the domains of the Internet of Things (IoT), intelligent services, and artificial intelligence (AI). The SSSAM model based on artificial neural networks (ANN) and GA algorithms also reached similar conclusions through a case study of the International University Sports Federation (FISU), building Information Modeling (BIM), cloud computing and artificial intelligence Internet of Things (AIoT) are expected to develop in the future. Three key themes developed over time. Finally, a comprehensive knowledge system with common references and future hot spots is proposed.
Collapse
Affiliation(s)
- Xi Zhu
- College of Physical Education, Southwest Jiaotong University, Chengdu, 611756, Sichuan, China
| | - Xiaobo Peng
- School of Physical Education, Chengdu Normal University, Chengdu, 611130, Sichuan, China
| |
Collapse
|
2
|
Kim JH, Hong H, Lee K, Jeong Y, Ryu H, Kim H, Jang SH, Park HK, Han JY, Park HJ, Bae H, Oh BM, Kim WS, Lee SY, Lee SU. AI in evaluating ambulation of stroke patients: severity classification with video and functional ambulation category scale. Top Stroke Rehabil 2024:1-9. [PMID: 38841903 DOI: 10.1080/10749357.2024.2359342] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2023] [Accepted: 05/18/2024] [Indexed: 06/07/2024]
Abstract
BACKGROUND The evaluation of gait function and severity classification of stroke patients are important to determine the rehabilitation goal and the level of exercise. Physicians often qualitatively evaluate patients' walking ability through visual gait analysis using naked eye, video images, or standardized assessment tools. Gait evaluation through observation relies on the doctor's empirical judgment, potentially introducing subjective opinions. Therefore, conducting research to establish a basis for more objective judgment is crucial. OBJECTIVE To verify a deep learning model that classifies gait image data of stroke patients according to Functional Ambulation Category (FAC) scale. METHODS Gait vision data from 203 stroke patients and 182 healthy individuals recruited from six medical institutions were collected to train a deep learning model for classifying gait severity in stroke patients. The recorded videos were processed using OpenPose. The dataset was randomly split into 80% for training and 20% for testing. RESULTS The deep learning model attained a training accuracy of 0.981 and test accuracy of 0.903. Area Under the Curve(AUC) values of 0.93, 0.95, and 0.96 for discriminating among the mild, moderate, and severe stroke groups, respectively. CONCLUSION This confirms the potential of utilizing human posture estimation based on vision data not only to develop gait parameter models but also to develop models to classify severity according to the FAC criteria used by physicians. To develop an AI-based severity classification model, a large amount and variety of data is necessary and data collected in non-standardized real environments, not in laboratories, can also be used meaningfully.
Collapse
Affiliation(s)
- Jeong-Hyun Kim
- Department of Rehabilitation Medicine, Seoul Metropolitan Government Boramae Medical Center, Seoul, South Korea
| | - Hyeon Hong
- Department of Rehabilitation Medicine, Seoul Metropolitan Government Boramae Medical Center, Seoul, South Korea
| | - Kyuwon Lee
- Department of Rehabilitation Medicine, Seoul Metropolitan Government Boramae Medical Center, Seoul, South Korea
| | - Yeji Jeong
- Department of Rehabilitation Medicine, Seoul Metropolitan Government Boramae Medical Center, Seoul, South Korea
| | - Hokyoung Ryu
- Department of Graduate School of Technology and Innovation Management, Hanyang University, Seoul, South Korea
| | - Hyundo Kim
- Department of Intelligence Computing, Hanyang University, Seoul, South Korea
| | - Seong-Ho Jang
- Department of Rehabilitation Medicine, Hanyang University, Guri Hospital, Gyeonggi-do, South Korea
| | - Hyeng-Kyu Park
- Department of Physical & Rehabilitation Medicine, Regional Cardiocerebrovascular Center, Center for Aging and Geriatrics, Chonnam National University Medical School & Hospital, Gwangju, South Korea
| | - Jae-Young Han
- Department of Physical & Rehabilitation Medicine, Regional Cardiocerebrovascular Center, Center for Aging and Geriatrics, Chonnam National University Medical School & Hospital, Gwangju, South Korea
| | - Hye Jung Park
- Department of Rehabilitation Medicine, Seoul St. Mary's Hospital, College of Medicine, The Catholic University of Korea, Seoul, South Korea
| | - Hasuk Bae
- Department of Rehabilitation Medicine, Ewha Woman's University, Seoul, South Korea
| | - Byung-Mo Oh
- Department of Rehabilitation, Seoul National University Hospital, Seoul, South Korea
| | - Won-Seok Kim
- Department of Rehabilitation Medicine, Seoul National University College of Medicine, Seoul, South Korea
| | - Sang Yoon Lee
- Department of Rehabilitation Medicine, Seoul National University College of Medicine, SMG-SNU Boramae Medical Center, Seoul, South Korea
| | - Shi-Uk Lee
- Department of Rehabilitation Medicine, Seoul Metropolitan Government Boramae Medical Center, Seoul, South Korea
- Department of Physical Medicine & Rehabilitation, College of Medicine, Seoul National University, Seoul, South Korea
| |
Collapse
|
3
|
Beltrán Beltrán R, Richter J, Köstermeyer G, Heinkel U. Climbing Technique Evaluation by Means of Skeleton Video Stream Analysis. SENSORS (BASEL, SWITZERLAND) 2023; 23:8216. [PMID: 37837046 PMCID: PMC10574944 DOI: 10.3390/s23198216] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/22/2023] [Revised: 09/21/2023] [Accepted: 09/26/2023] [Indexed: 10/15/2023]
Abstract
Due to the growing interest in climbing, increasing importance has been given to research in the field of non-invasive, camera-based motion analysis. While existing work uses invasive technologies such as wearables or modified walls and holds, or focuses on competitive sports, we for the first time present a system that uses video analysis to automatically recognize six movement errors that are typical for novices with limited climbing experience. Climbing a complete route consists of three repetitive climbing phases. Therefore, a characteristic joint arrangement may be detected as an error in a specific climbing phase, while this exact arrangement may not considered to be an error in another climbing phase. That is why we introduced a finite state machine to determine the current phase and to check for errors that commonly occur in the current phase. The transition between the phases depends on which joints are being used. To capture joint movements, we use a fourth-generation iPad Pro with LiDAR to record climbing sequences in which we convert the climber's 2-D skeleton provided by the Vision framework from Apple into 3-D joints using the LiDAR depth information. Thereupon, we introduced a method that derives whether a joint moves or not, determining the current phase. Finally, the 3-D joints are analyzed with respect to defined characteristic joint arrangements to identify possible motion errors. To present the feedback to the climber, we imitate a virtual mentor by realizing an application on the iPad that creates an analysis immediately after the climber has finished the route by pointing out the detected errors and by giving suggestions for improvement. Quantitative tests with three experienced climbers that were able to climb reference routes without any errors and intentionally with errors resulted in precision-recall curves evaluating the error detection performance. The results demonstrate that while the number of false positives is still in an acceptable range, the number of detected errors is sufficient to provide climbing novices with adequate suggestions for improvement. Moreover, our study reveals limitations that mainly originate from incorrect joint localizations caused by the LiDAR sensor range. With human pose estimation becoming increasingly reliable and with the advance of sensor capabilities, these limitations will have a decreasing impact on our system performance.
Collapse
Affiliation(s)
- Raul Beltrán Beltrán
- Professorship Circuit and System Design, Chemnitz University of Technology, Reichenhainer Straße 70, 09126 Chemnitz, Germany; (J.R.); (U.H.)
| | - Julia Richter
- Professorship Circuit and System Design, Chemnitz University of Technology, Reichenhainer Straße 70, 09126 Chemnitz, Germany; (J.R.); (U.H.)
| | - Guido Köstermeyer
- Department Sportwissenschaft und Sport, Friedrich-Alexander-Universität Erlangen-Nürnberg, Schlossplatz 4, 91054 Erlangen, Germany;
| | - Ulrich Heinkel
- Professorship Circuit and System Design, Chemnitz University of Technology, Reichenhainer Straße 70, 09126 Chemnitz, Germany; (J.R.); (U.H.)
| |
Collapse
|
4
|
Baumgartner T, Paassen B, Klatt S. Extracting spatial knowledge from track and field broadcasts for monocular 3D human pose estimation. Sci Rep 2023; 13:14031. [PMID: 37640789 PMCID: PMC10462612 DOI: 10.1038/s41598-023-41142-0] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2022] [Accepted: 08/22/2023] [Indexed: 08/31/2023] Open
Abstract
Collecting large datasets for investigations into human locomotion is an expensive and labor-intensive process. Methods for 3D human pose estimation in the wild are becoming increasingly accurate and could soon be sufficient to assist with the collection of datasets for analysis into running kinematics from TV broadcast data. In the domain of biomechanical research, small differences in 3D angles play an important role. More precisely, the error margins of the data collection process need to be smaller than the expected variation between athletes. In this work, we propose a method to infer the global geometry of track and field stadium recordings using lane demarcations. By projecting estimated 3D skeletons back into the image using this global geometry, we show that current state-of-the-art 3D human pose estimation methods are not (yet) accurate enough to be used in kinematics research.
Collapse
Affiliation(s)
- Tobias Baumgartner
- Institute for Exercise Training and Sport Informatics, German Sport University, Cologne, Germany.
| | - Benjamin Paassen
- German Research Center for Artificial Intelligence, Berlin, Germany
| | - Stefanie Klatt
- Institute for Exercise Training and Sport Informatics, German Sport University, Cologne, Germany
| |
Collapse
|
5
|
Kim S, Park S, Lee S, Seo SH, Kim HS, Cha Y, Kim JT, Kim JW, Ha YC, Yoo JI. Assessing physical abilities of sarcopenia patients using gait analysis and smart insole for development of digital biomarker. Sci Rep 2023; 13:10602. [PMID: 37391464 PMCID: PMC10313812 DOI: 10.1038/s41598-023-37794-7] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2023] [Accepted: 06/28/2023] [Indexed: 07/02/2023] Open
Abstract
The aim of this study is to compare variable importance across multiple measurement tools, and to use smart insole and artificial intelligence (AI) gait analysis to create variables that can evaluate the physical abilities of sarcopenia patients. By analyzing and comparing sarcopenia patients with non sarcopenia patients, this study aims to develop predictive and classification models for sarcopenia and discover digital biomarkers. The researchers used smart insole equipment to collect plantar pressure data from 83 patients, and a smart phone to collect video data for pose estimation. A Mann-Whitney U was conducted to compare the sarcopenia group of 23 patients and the control group of 60 patients. Smart insole and pose estimation were used to compare the physical abilities of sarcopenia patients with a control group. Analysis of joint point variables showed significant differences in 12 out of 15 variables, but not in knee mean, ankle range, and hip range. These findings suggest that digital biomarkers can be used to differentiate sarcopenia patients from the normal population with improved accuracy. This study compared musculoskeletal disorder patients to sarcopenia patients using smart insole and pose estimation. Multiple measurement methods are important for accurate sarcopenia diagnosis and digital technology has potential for improving diagnosis and treatment.
Collapse
Affiliation(s)
- Shinjune Kim
- Department of Biomedical Research Institute, Inha University Hospital, Incheon, Republic of Korea
| | - Seongjin Park
- Department of Biomedical Research Institute, Gyeongsang National University Hospital, Jinju, Republic of Korea
| | - Sangyeob Lee
- Department of Biomedical Research Institute, Gyeongsang National University Hospital, Jinju, Republic of Korea
| | - Sung Hyo Seo
- Department of Biomedical Research Institute, Gyeongsang National University Hospital, Jinju, Republic of Korea
| | - Hyeon Su Kim
- Department of Biomedical Research Institute, Inha University Hospital, Incheon, Republic of Korea
| | - Yonghan Cha
- Department of Orthopaedic Surgery, Daejeon Eulji Medical Center, Daejeon, Republic of Korea
| | - Jung-Taek Kim
- Department of Orthopedic Surgery, Ajou University School of Medicine, Suwon, Republic of Korea
| | - Jin-Woo Kim
- Department of Orthopaedic Surgery, Nowon Eulji Medical Center, Seoul, Republic of Korea
| | - Yong-Chan Ha
- Department of Orthopaedic Surgery, Bumin Medical Center, Seoul, Republic of Korea
| | - Jun-Il Yoo
- Department of Orthopedic Surgery, Inha University Hospital, 27, Inhang-ro, Jung-gu, Incheon, Republic of Korea.
| |
Collapse
|
6
|
Mundt M, Born Z, Goldacre M, Alderson J. Estimating Ground Reaction Forces from Two-Dimensional Pose Data: A Biomechanics-Based Comparison of AlphaPose, BlazePose, and OpenPose. SENSORS (BASEL, SWITZERLAND) 2022; 23:s23010078. [PMID: 36616676 PMCID: PMC9823796 DOI: 10.3390/s23010078] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/11/2022] [Revised: 12/12/2022] [Accepted: 12/16/2022] [Indexed: 05/14/2023]
Abstract
The adoption of computer vision pose estimation approaches, used to identify keypoint locations which are intended to reflect the necessary anatomical landmarks relied upon by biomechanists for musculoskeletal modelling, has gained increasing traction in recent years. This uptake has been further accelerated by keypoint use as inputs into machine learning models used to estimate biomechanical parameters such as ground reaction forces (GRFs) in the absence of instrumentation required for direct measurement. This study first aimed to investigate the keypoint detection rate of three open-source pose estimation models (AlphaPose, BlazePose, and OpenPose) across varying movements, camera views, and trial lengths. Second, this study aimed to assess the suitability and interchangeability of keypoints detected by each pose estimation model when used as inputs into machine learning models for the estimation of GRFs. The keypoint detection rate of BlazePose was distinctly lower than that of AlphaPose and OpenPose. All pose estimation models achieved a high keypoint detection rate at the centre of an image frame and a lower detection rate in the true sagittal plane camera field of view, compared with slightly anteriorly or posteriorly located quasi-sagittal plane camera views. The three-dimensional ground reaction force, instantaneous loading rate, and peak force for running could be estimated using the keypoints of all three pose estimation models. However, only AlphaPose and OpenPose keypoints could be used interchangeably with a machine learning model trained to estimate GRFs based on AlphaPose keypoints resulting in a high estimation accuracy when OpenPose keypoints were used as inputs and vice versa. The findings of this study highlight the need for further evaluation of computer vision-based pose estimation models for application in biomechanical human modelling, and the limitations of machine learning-based GRF estimation models that rely on 2D keypoints. This is of particular relevance given that machine learning models informing athlete monitoring guidelines are being developed for application related to athlete well-being.
Collapse
Affiliation(s)
- Marion Mundt
- UWA Minderoo Tech & Policy Lab, Law School, The University of Western Australia, Crawley, WA 6009, Australia
- Correspondence:
| | - Zachery Born
- UWA Minderoo Tech & Policy Lab, Law School, The University of Western Australia, Crawley, WA 6009, Australia
| | - Molly Goldacre
- UWA Minderoo Tech & Policy Lab, Law School, The University of Western Australia, Crawley, WA 6009, Australia
| | - Jacqueline Alderson
- UWA Minderoo Tech & Policy Lab, Law School, The University of Western Australia, Crawley, WA 6009, Australia
- Sports Performance Research Institute New Zealand (SPRINZ), Auckland University of Technology, Auckland 1010, New Zealand
| |
Collapse
|
7
|
Link J, Schwinn L, Pulsmeyer F, Kautz T, Eskofier BM. xLength: Predicting Expected Ski Jump Length Shortly after Take-Off Using Deep Learning. SENSORS (BASEL, SWITZERLAND) 2022; 22:8474. [PMID: 36366174 PMCID: PMC9657424 DOI: 10.3390/s22218474] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/23/2022] [Revised: 10/26/2022] [Accepted: 10/26/2022] [Indexed: 06/16/2023]
Abstract
With tracking systems becoming more widespread in sports research and regular training and competitions, more data are available for sports analytics and performance prediction. We analyzed 2523 ski jumps from 205 athletes on five venues. For every jump, the dataset includes the 3D trajectory, 3D velocity, skis' orientation, and metadata such as wind, starting gate, and ski jumping hill data. Using this dataset, we aimed to predict the expected jump length (xLength) inspired by the expected goals metric in soccer (xG). We evaluate the performance of a fully connected neural network, a convolutional neural network (CNN), a long short-term memory (LSTM), and a ResNet architecture to estimate the xLength. For the prediction of the jump length one second after take-off, we achieve a mean absolute error (MAE) of 5.3 m for the generalization to new athletes and an MAE of 5.9 m for the generalization to new ski jumping hills using ResNet architectures. Additionally, we investigated the influence of the input time after the take-off on the predictions' accuracy. As expected, the MAE becomes smaller with longer inputs. Due to the real-time transmission of the sensor's data, xLength can be updated during the flight phase and used in live TV broadcasting. xLength could also be used as an analysis tool for experts to quantify the quality of the take-off and flight phases.
Collapse
|
8
|
Nguyen HC, Nguyen TH, Scherer R, Le VH. Unified End-to-End YOLOv5-HR-TCM Framework for Automatic 2D/3D Human Pose Estimation for Real-Time Applications. SENSORS (BASEL, SWITZERLAND) 2022; 22:5419. [PMID: 35891099 PMCID: PMC9315644 DOI: 10.3390/s22145419] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/16/2022] [Revised: 07/06/2022] [Accepted: 07/16/2022] [Indexed: 06/15/2023]
Abstract
Three-dimensional human pose estimation is widely applied in sports, robotics, and healthcare. In the past five years, the number of CNN-based studies for 3D human pose estimation has been numerous and has yielded impressive results. However, studies often focus only on improving the accuracy of the estimation results. In this paper, we propose a fast, unified end-to-end model for estimating 3D human pose, called YOLOv5-HR-TCM (YOLOv5-HRet-Temporal Convolution Model). Our proposed model is based on the 2D to 3D lifting approach for 3D human pose estimation while taking care of each step in the estimation process, such as person detection, 2D human pose estimation, and 3D human pose estimation. The proposed model is a combination of best practices at each stage. Our proposed model is evaluated on the Human 3.6M dataset and compared with other methods at each step. The method achieves high accuracy, not sacrificing processing speed. The estimated time of the whole process is 3.146 FPS on a low-end computer. In particular, we propose a sports scoring application based on the deviation angle between the estimated 3D human posture and the standard (reference) origin. The average deviation angle evaluated on the Human 3.6M dataset (Protocol #1-Pro #1) is 8.2 degrees.
Collapse
Affiliation(s)
- Hung-Cuong Nguyen
- Faculty of Engineering Technology, Hung Vuong University, Viet Tri City 35100, Vietnam; (H.-C.N.); (T.-H.N.)
| | - Thi-Hao Nguyen
- Faculty of Engineering Technology, Hung Vuong University, Viet Tri City 35100, Vietnam; (H.-C.N.); (T.-H.N.)
| | - Rafal Scherer
- Department of Intelligent Computer Systems, Czestochowa University of Technology, 42-218 Czestochowa, Poland;
| | - Van-Hung Le
- Faculty of Basic science, Tan Trao University, Tuyen Quang City 22000, Vietnam
| |
Collapse
|
9
|
Analysis of Competition and Training Videos of Speed Climbing Athletes Using Feature and Human Body Keypoint Detection Algorithms. SENSORS 2022; 22:s22062251. [PMID: 35336423 PMCID: PMC8955718 DOI: 10.3390/s22062251] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/07/2022] [Revised: 03/01/2022] [Accepted: 03/03/2022] [Indexed: 11/16/2022]
Abstract
Compared to 25 years ago, the climbing sport itself has changed dramatically. From a rock climbing modification to a separation in three independent disciplines, the requirements to athletes and trainers increased rapidly. To ensure continuous improvement of the sport itself, the usage of measurement and sensor technology is unavoidable. Especially in the field of the discipline speed climbing, which will be performed as a single discipline at the Olympic Games 2024 in Paris, the current state of the art of movement analysis only consists of video analysis and the benefit of the experience of trainers. Therefore, this paper presents a novel method, which supports trainers and athletes and enables analysis of motion sequences and techniques. Prerecorded video footage is combined with existing feature and human body keypoint detection algorithms and standardized boundary conditions. Therefore, several image processing steps are necessary to convert the recorded movement of different speed climbing athletes to significant parameters for detailed analysis. By studying climbing trials of professional athletes and the used techniques in different sections of the speed climbing wall, the aim among others is to get comparable results and detect mistakes. As a conclusion, the presented method enables powerful analysis of speed climbing training and competition and serves with the aid of a user-friendly designed interface as a support for trainers and athletes for the evaluation of motion sequences.
Collapse
|
10
|
Li J, Wang Z, Qi B, Zhang J, Yang H. MEMe: A Mutually Enhanced Modeling Method for Efficient and Effective Human Pose Estimation. SENSORS (BASEL, SWITZERLAND) 2022; 22:632. [PMID: 35062592 PMCID: PMC8780536 DOI: 10.3390/s22020632] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/20/2021] [Revised: 01/06/2022] [Accepted: 01/11/2022] [Indexed: 12/10/2022]
Abstract
In this paper, a mutually enhanced modeling method (MEMe) is presented for human pose estimation, which focuses on enhancing lightweight model performance, but with low complexity. To obtain higher accuracy, a traditional model scale is largely expanded with heavy deployment difficulties. However, for a more lightweight model, there is a large performance gap compared to the former; thus, an urgent need for a way to fill it. Therefore, we propose a MEMe to reconstruct a lightweight baseline model, EffBase transferred intuitively from EfficientDet, into the efficient and effective pose (EEffPose) net, which contains three mutually enhanced modules: the Enhanced EffNet (EEffNet) backbone, the total fusion neck (TFNeck), and the final attention head (FAHead). Extensive experiments on COCO and MPII benchmarks show that our MEMe-based models reach state-of-the-art performances, with limited parameters. Specifically, in the same conditions, our EEffPose-P0 with 256 × 192 can use only 8.98 M parameters to achieve 75.4 AP on the COCO val set, which outperforms HRNet-W48, but with only 14% of its parameters.
Collapse
Affiliation(s)
- Jie Li
- Key Laboratory of Optical Engineering, Chinese Academy of Sciences, Chengdu 610209, China; (J.L.); (Z.W.); (J.Z.); (H.Y.)
- Institute of Optics and Electronics, Chinese Academy of Sciences, Chengdu 610209, China
- University of Chinese Academy of Sciences, Beijing 100039, China
| | - Zhixing Wang
- Key Laboratory of Optical Engineering, Chinese Academy of Sciences, Chengdu 610209, China; (J.L.); (Z.W.); (J.Z.); (H.Y.)
- Institute of Optics and Electronics, Chinese Academy of Sciences, Chengdu 610209, China
- University of Chinese Academy of Sciences, Beijing 100039, China
- School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu 610209, China
| | - Bo Qi
- Key Laboratory of Optical Engineering, Chinese Academy of Sciences, Chengdu 610209, China; (J.L.); (Z.W.); (J.Z.); (H.Y.)
- Institute of Optics and Electronics, Chinese Academy of Sciences, Chengdu 610209, China
- University of Chinese Academy of Sciences, Beijing 100039, China
| | - Jianlin Zhang
- Key Laboratory of Optical Engineering, Chinese Academy of Sciences, Chengdu 610209, China; (J.L.); (Z.W.); (J.Z.); (H.Y.)
- Institute of Optics and Electronics, Chinese Academy of Sciences, Chengdu 610209, China
- University of Chinese Academy of Sciences, Beijing 100039, China
| | - Hu Yang
- Key Laboratory of Optical Engineering, Chinese Academy of Sciences, Chengdu 610209, China; (J.L.); (Z.W.); (J.Z.); (H.Y.)
- Institute of Optics and Electronics, Chinese Academy of Sciences, Chengdu 610209, China
| |
Collapse
|
11
|
Single Camera-Based Remote Physical Therapy: Verification on a Large Video Dataset. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12020799] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
In recent years, several systems have been developed to capture human motion in real-time using common RGB cameras. This approach has great potential to become widespread among the general public as it allows the remote evaluation of exercise at no additional cost. The concept of using these systems in rehabilitation in the home environment has been discussed, but no work has addressed the practical problem of detecting basic body parts under different sensing conditions on a large scale. In this study, we evaluate the ability of the OpenPose pose estimation algorithm to perform keypoint detection of anatomical landmarks under different conditions. We infer the quality of detection based on the keypoint confidence values reported by the OpenPose. We used more than two thousand unique exercises for the evaluation. We focus on the influence of the camera view and the influence of the position of the trainees, which are essential in terms of the use for home exercise. Our results show that the position of the trainee has the greatest effect, in the following increasing order of suitability across all camera views: lying position, position on the knees, sitting position, and standing position. On the other hand, the effect of the camera view was only marginal, showing that the side view is having slightly worse results. The results might also indicate that the quality of detection of lower body joints is lower across all conditions than the quality of detection of upper body joints. In this practical overview, we present the possibilities and limitations of current camera-based systems in telerehabilitation.
Collapse
|