1
|
Chen X, Kamavuako EN. Vision-Based Methods for Food and Fluid Intake Monitoring: A Literature Review. SENSORS (BASEL, SWITZERLAND) 2023; 23:6137. [PMID: 37447988 PMCID: PMC10346353 DOI: 10.3390/s23136137] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/10/2023] [Revised: 05/28/2023] [Accepted: 06/28/2023] [Indexed: 07/15/2023]
Abstract
Food and fluid intake monitoring are essential for reducing the risk of dehydration, malnutrition, and obesity. The existing research has been preponderantly focused on dietary monitoring, while fluid intake monitoring, on the other hand, is often neglected. Food and fluid intake monitoring can be based on wearable sensors, environmental sensors, smart containers, and the collaborative use of multiple sensors. Vision-based intake monitoring methods have been widely exploited with the development of visual devices and computer vision algorithms. Vision-based methods provide non-intrusive solutions for monitoring. They have shown promising performance in food/beverage recognition and segmentation, human intake action detection and classification, and food volume/fluid amount estimation. However, occlusion, privacy, computational efficiency, and practicality pose significant challenges. This paper reviews the existing work (253 articles) on vision-based intake (food and fluid) monitoring methods to assess the size and scope of the available literature and identify the current challenges and research gaps. This paper uses tables and graphs to depict the patterns of device selection, viewing angle, tasks, algorithms, experimental settings, and performance of the existing monitoring systems.
Collapse
Affiliation(s)
- Xin Chen
- Department of Engineering, King’s College London, London WC2R 2LS, UK;
| | - Ernest N. Kamavuako
- Department of Engineering, King’s College London, London WC2R 2LS, UK;
- Faculté de Médecine, Université de Kindu, Site de Lwama II, Kindu, Maniema, Democratic Republic of the Congo
| |
Collapse
|
2
|
Elbassuoni S, Ghattas H, El Ati J, Zoughby Y, Semaan A, Akl C, Trabelsi T, Talhouk R, Ben Gharbia H, Shmayssani Z, Mourad A. Capturing children food exposure using wearable cameras and deep learning. PLOS DIGITAL HEALTH 2023; 2:e0000211. [PMID: 36972212 PMCID: PMC10042366 DOI: 10.1371/journal.pdig.0000211] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/21/2022] [Accepted: 02/09/2023] [Indexed: 03/29/2023]
Abstract
Children's dietary habits are influenced by complex factors within their home, school and neighborhood environments. Identifying such influencers and assessing their effects is traditionally based on self-reported data which can be prone to recall bias. We developed a culturally acceptable machine-learning-based data-collection system to objectively capture school-children's exposure to food (including food items, food advertisements, and food outlets) in two urban Arab centers: Greater Beirut, in Lebanon, and Greater Tunis, in Tunisia. Our machine-learning-based system consists of 1) a wearable camera that captures continuous footage of children's environment during a typical school day, 2) a machine learning model that automatically identifies images related to food from the collected data and discards any other footage, 3) a second machine learning model that classifies food-related images into images that contain actual food items, images that contain food advertisements, and images that contain food outlets, and 4) a third machine learning model that classifies images that contain food items into two classes, corresponding to whether the food items are being consumed by the child wearing the camera or whether they are consumed by others. This manuscript reports on a user-centered design study to assess the acceptability of using wearable cameras to capture food exposure among school children in Greater Beirut and Greater Tunis. We then describe how we trained our first machine learning model to detect food exposure images using data collected from the Web and utilizing the latest trends in deep learning for computer vision. Next, we describe how we trained our other machine learning models to classify food-related images into their respective categories using a combination of public data and data acquired via crowdsourcing. Finally, we describe how the different components of our system were packed together and deployed in a real-world case study and we report on its performance.
Collapse
Affiliation(s)
- Shady Elbassuoni
- Computer Science Department, American University of Beirut, Beirut, Lebanon
| | - Hala Ghattas
- Center for Research on Population and Health, American University of Beirut, Beirut, Lebanon
- Department of Health Promotion, Education, and Behavior, University of South Carolina, Columbia, South Carolina, United States of America
| | - Jalila El Ati
- SURVEN Research Laboratory, National Institute of Nutrition and Food Technology, Tunis, Tunisia
| | - Yorgo Zoughby
- Computer Science Department, American University of Beirut, Beirut, Lebanon
| | - Aline Semaan
- Department of Public Health, Institute of Tropical Medicine, Antwerp, Belgium
| | - Christelle Akl
- Center for Research on Population and Health, American University of Beirut, Beirut, Lebanon
| | - Tarek Trabelsi
- SURVEN Research Laboratory, National Institute of Nutrition and Food Technology, Tunis, Tunisia
| | - Reem Talhouk
- School of Design, Northumbria University, Newcastle upon Tyne, United Kingdom
| | - Houda Ben Gharbia
- SURVEN Research Laboratory, National Institute of Nutrition and Food Technology, Tunis, Tunisia
| | | | - Aya Mourad
- Computer Science Department, American University of Beirut, Beirut, Lebanon
| | | |
Collapse
|
3
|
Peruzzi G, Pozzebon A, Van Der Meer M. Fight Fire with Fire: Detecting Forest Fires with Embedded Machine Learning Models Dealing with Audio and Images on Low Power IoT Devices. SENSORS (BASEL, SWITZERLAND) 2023; 23:783. [PMID: 36679579 PMCID: PMC9863941 DOI: 10.3390/s23020783] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/07/2022] [Revised: 01/03/2023] [Accepted: 01/05/2023] [Indexed: 06/17/2023]
Abstract
Forest fires are the main cause of desertification, and they have a disastrous impact on agricultural and forest ecosystems. Modern fire detection and warning systems rely on several techniques: satellite monitoring, sensor networks, image processing, data fusion, etc. Recently, Artificial Intelligence (AI) algorithms have been applied to fire recognition systems, enhancing their efficiency and reliability. However, these devices usually need constant data transmission along with a proper amount of computing power, entailing high costs and energy consumption. This paper presents the prototype of a Video Surveillance Unit (VSU) for recognising and signalling the presence of forest fires by exploiting two embedded Machine Learning (ML) algorithms running on a low power device. The ML models take audio samples and images as their respective inputs, allowing for timely fire detection. The main result is that while the performances of the two models are comparable when they work independently, their joint usage according to the proposed methodology provides a higher accuracy, precision, recall and F1 score (96.15%, 92.30%, 100.00%, and 96.00%, respectively). Eventually, each event is remotely signalled by making use of the Long Range Wide Area Network (LoRaWAN) protocol to ensure that the personnel in charge are able to operate promptly.
Collapse
Affiliation(s)
- Giacomo Peruzzi
- Department of Information Engineering, University of Padova, 35131 Padova, Italy
| | - Alessandro Pozzebon
- Department of Information Engineering, University of Padova, 35131 Padova, Italy
| | - Mattia Van Der Meer
- Department of Information Engineering and Mathematics, University of Siena, 53100 Siena, Italy
| |
Collapse
|
4
|
Thomas DM, Kleinberg S, Brown AW, Crow M, Bastian ND, Reisweber N, Lasater R, Kendall T, Shafto P, Blaine R, Smith S, Ruiz D, Morrell C, Clark N. Machine learning modeling practices to support the principles of AI and ethics in nutrition research. Nutr Diabetes 2022; 12:48. [PMID: 36456550 PMCID: PMC9715415 DOI: 10.1038/s41387-022-00226-y] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/22/2022] [Revised: 10/28/2022] [Accepted: 11/15/2022] [Indexed: 12/03/2022] Open
Abstract
BACKGROUND Nutrition research is relying more on artificial intelligence and machine learning models to understand, diagnose, predict, and explain data. While artificial intelligence and machine learning models provide powerful modeling tools, failure to use careful and well-thought-out modeling processes can lead to misleading conclusions and concerns surrounding ethics and bias. METHODS Based on our experience as reviewers and journal editors in nutrition and obesity, we identified the most frequently omitted best practices from statistical modeling and how these same practices extend to machine learning models. We next addressed areas required for implementation of machine learning that are not included in commercial software packages. RESULTS Here, we provide a tutorial on best artificial intelligence and machine learning modeling practices that can reduce potential ethical problems with a checklist and guiding principles to aid nutrition researchers in developing, evaluating, and implementing artificial intelligence and machine learning models in nutrition research. CONCLUSION The quality of AI/ML modeling in nutrition research requires iterative and tailored processes to mitigate against potential ethical problems or to predict conclusions that are free of bias.
Collapse
Affiliation(s)
- Diana M. Thomas
- grid.419884.80000 0001 2287 2270Department of Mathematical Sciences, United States Military Academy, West Point, NY 10996 USA
| | - Samantha Kleinberg
- grid.217309.e0000 0001 2180 0654Department of Computer Science, Stevens Institute of Technology, Hoboken, NJ 07030 USA
| | - Andrew W. Brown
- grid.241054.60000 0004 4687 1637Department of Biostatistics, University of Arkansas for Medical Sciences, Little Rock, AR 72205 USA ,grid.488749.eArkansas Children’s Research Institute, Little Rock, AR 72202 USA
| | - Mason Crow
- grid.419884.80000 0001 2287 2270Department of Mathematical Sciences, United States Military Academy, West Point, NY 10996 USA
| | - Nathaniel D. Bastian
- grid.419884.80000 0001 2287 2270Army Cyber Institute, United States Military Academy, West Point, NY 10996 USA
| | - Nicholas Reisweber
- grid.419884.80000 0001 2287 2270Department of Mathematical Sciences, United States Military Academy, West Point, NY 10996 USA
| | - Robert Lasater
- grid.419884.80000 0001 2287 2270Department of Mathematical Sciences, United States Military Academy, West Point, NY 10996 USA
| | - Thomas Kendall
- grid.419884.80000 0001 2287 2270Department of Mathematical Sciences, United States Military Academy, West Point, NY 10996 USA
| | - Patrick Shafto
- grid.430387.b0000 0004 1936 8796Department of Mathematics and Computer Science, Rutgers University, Newark, NJ 07102 USA
| | - Raymond Blaine
- grid.419884.80000 0001 2287 2270Department of Electrical Engineering and Computer Science, United States Military Academy, West Point, NY 10996 USA
| | - Sarah Smith
- grid.419884.80000 0001 2287 2270Department of Electrical Engineering and Computer Science, United States Military Academy, West Point, NY 10996 USA
| | - Daniel Ruiz
- grid.419884.80000 0001 2287 2270Department of Electrical Engineering and Computer Science, United States Military Academy, West Point, NY 10996 USA
| | - Christopher Morrell
- grid.419884.80000 0001 2287 2270Department of Electrical Engineering and Computer Science, United States Military Academy, West Point, NY 10996 USA
| | - Nicholas Clark
- grid.419884.80000 0001 2287 2270Department of Mathematical Sciences, United States Military Academy, West Point, NY 10996 USA
| |
Collapse
|
5
|
Ghosh T, Sazonov E. A Comparative Study of Deep Learning Algorithms for Detecting Food Intake. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2022; 2022:2993-2996. [PMID: 36085821 DOI: 10.1109/embc48229.2022.9871278] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
The choice of appropriate machine learning algorithms is crucial for classification problems. This study compares the performance of state-of-the-art time-series deep learning algorithms for classifying food intake using sensor signals. The sensor signals were collected with the help of a wearable sensor system (the Automatic Ingestion Monitor v2, or AIM-2). AIM-2 used an optical and 3-axis accelerometer sensor to capture temporalis muscle activation. Raw signals from those sensors were used to train five classifiers (multilayer perceptron (MLP), time Convolutional Neural Network (time-CNN), Fully Convolutional Neural Network (FCN), Residual Neural Network (ResNet), and Inception network) to differentiate food intake (eating and drinking) from other activities. Data were collected from 17 pilot subjects over the course of 23 days in free-living conditions. A leave one subject out cross-validation scheme was used for training and testing. Time-CNN, FCN, ResNet, and Inception achieved average balanced classification accuracy of 88.84%, 90.18%, 93.47%, and 92.15%, respectively. The results indicate that ResNet outperforms other state-of-the-art deep learning algorithms for this specific problem.
Collapse
|
6
|
Ramesh A, Raju VB, Rao M, Sazonov E. Food Detection and Segmentation from Egocentric Camera Images. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2021; 2021:2736-2740. [PMID: 34891816 DOI: 10.1109/embc46164.2021.9630823] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Tracking an individual's food intake provides useful insight into their eating habits. Technological advancements in wearable sensors such as the automatic capture of food images from wearable cameras have made the tracking of food intake efficient and feasible. For accurate food intake monitoring, an automated food detection technique is needed to recognize foods from unstaged real-world images. This work presents a novel food detection and segmentation pipeline to detect the presence of food in images acquired from an egocentric wearable camera, and subsequently segment the food image. An ensemble of YOLOv5 detection networks is trained to detect and localize food items among other objects present in captured images. The model achieves an overall 80.6% mean average precision on four objects-Food, Beverage, Screen, and Person. Post object detection, the predicted food objects which are sufficiently sharp were considered for segmentation. The Normalized-Graph-Cut algorithm was used to segment the different parts of the food resulting in an average IoU of 82%.Clinical relevance- The automatic monitoring of food intake using wearable devices can play a pivotal role in the treatment and prevention of eating disorders, obesity, malnutrition and other related issues. It can aid in understanding the pattern of nutritional intake and make personalized adjustments to lead a healthy life.
Collapse
|
7
|
Chen G, Jia W, Zhao Y, Mao ZH, Lo B, Anderson AK, Frost G, Jobarteh ML, McCrory MA, Sazonov E, Steiner-Asiedu M, Ansong RS, Baranowski T, Burke L, Sun M. Food/Non-Food Classification of Real-Life Egocentric Images in Low- and Middle-Income Countries Based on Image Tagging Features. Front Artif Intell 2021; 4:644712. [PMID: 33870184 PMCID: PMC8047062 DOI: 10.3389/frai.2021.644712] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2020] [Accepted: 02/26/2021] [Indexed: 11/25/2022] Open
Abstract
Malnutrition, including both undernutrition and obesity, is a significant problem in low- and middle-income countries (LMICs). In order to study malnutrition and develop effective intervention strategies, it is crucial to evaluate nutritional status in LMICs at the individual, household, and community levels. In a multinational research project supported by the Bill & Melinda Gates Foundation, we have been using a wearable technology to conduct objective dietary assessment in sub-Saharan Africa. Our assessment includes multiple diet-related activities in urban and rural families, including food sources (e.g., shopping, harvesting, and gathering), preservation/storage, preparation, cooking, and consumption (e.g., portion size and nutrition analysis). Our wearable device ("eButton" worn on the chest) acquires real-life images automatically during wake hours at preset time intervals. The recorded images, in amounts of tens of thousands per day, are post-processed to obtain the information of interest. Although we expect future Artificial Intelligence (AI) technology to extract the information automatically, at present we utilize AI to separate the acquired images into two binary classes: images with (Class 1) and without (Class 0) edible items. As a result, researchers need only to study Class-1 images, reducing their workload significantly. In this paper, we present a composite machine learning method to perform this classification, meeting the specific challenges of high complexity and diversity in the real-world LMIC data. Our method consists of a deep neural network (DNN) and a shallow learning network (SLN) connected by a novel probabilistic network interface layer. After presenting the details of our method, an image dataset acquired from Ghana is utilized to train and evaluate the machine learning system. Our comparative experiment indicates that the new composite method performs better than the conventional deep learning method assessed by integrated measures of sensitivity, specificity, and burden index, as indicated by the Receiver Operating Characteristic (ROC) curve.
Collapse
Affiliation(s)
- Guangzong Chen
- Department of Electrical and Computer Engineering, University of Pittsburgh, PA, United States
| | - Wenyan Jia
- Department of Electrical and Computer Engineering, University of Pittsburgh, PA, United States
| | - Yifan Zhao
- Department of Electrical and Computer Engineering, University of Pittsburgh, PA, United States
| | - Zhi-Hong Mao
- Department of Electrical and Computer Engineering, University of Pittsburgh, PA, United States
| | - Benny Lo
- Hamlyn Centre, Imperial College London, London, United Kingdom
| | - Alex K. Anderson
- Department of Foods and Nutrition, University of Georgia, Athens, GA, United States
| | - Gary Frost
- Section for Nutrition Research, Department of Metabolism, Digestion and Reproduction, Imperial College London, London, United Kingdom
| | - Modou L. Jobarteh
- Section for Nutrition Research, Department of Metabolism, Digestion and Reproduction, Imperial College London, London, United Kingdom
| | - Megan A. McCrory
- Department of Health Sciences, Boston University, Boston, MA, United States
| | - Edward Sazonov
- Department of Electrical and Computer Engineering, University of Alabama, Tuscaloosa, AL, United States
| | | | - Richard S. Ansong
- Department of Nutrition and Food Science, University of Ghana, Legon-Accra, Ghana
| | - Thomas Baranowski
- USDA/ARS Children's Nutrition Research Center, Department of Pediatrics, Baylor College of Medicine, Houston, TX, United States
| | - Lora Burke
- School of Nursing, University of Pittsburgh, Pittsburgh, PA, United States
| | - Mingui Sun
- Department of Electrical and Computer Engineering, University of Pittsburgh, PA, United States
- Department of Neurosurgery, University of Pittsburgh, Pittsburgh, PA, United States
| |
Collapse
|