Volumetric Food Quantification Using Computer Vision on a Depth-Sensing Smartphone: Preclinical Study
BackgroundQuantification of dietary intake is key to the prevention and management of numerous metabolic disorders. Conventional approaches are challenging, laborious, and lack accuracy. The recent advent of depth-sensing smartphones in conjunction with computer vision could...
Main Authors: | , , , , , , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
JMIR Publications
2020-03-01
|
Series: | JMIR mHealth and uHealth |
Online Access: | http://mhealth.jmir.org/2020/3/e15294/ |
id |
doaj-0a887b9d85454e81a7c2130efcb3b497 |
---|---|
record_format |
Article |
spelling |
doaj-0a887b9d85454e81a7c2130efcb3b4972021-05-02T19:28:42ZengJMIR PublicationsJMIR mHealth and uHealth2291-52222020-03-0183e1529410.2196/15294Volumetric Food Quantification Using Computer Vision on a Depth-Sensing Smartphone: Preclinical StudyHerzig, DavidNakas, Christos TStalder, JanineKosinski, ChristopheLaesser, CélineDehais, JoachimJaeggi, RaphaelLeichtle, Alexander BenediktDahlweid, Fried-MichaelStettler, ChristophBally, Lia BackgroundQuantification of dietary intake is key to the prevention and management of numerous metabolic disorders. Conventional approaches are challenging, laborious, and lack accuracy. The recent advent of depth-sensing smartphones in conjunction with computer vision could facilitate reliable quantification of food intake. ObjectiveThe objective of this study was to evaluate the accuracy of a novel smartphone app combining depth-sensing hardware with computer vision to quantify meal macronutrient content using volumetry. MethodsThe app ran on a smartphone with a built-in depth sensor applying structured light (iPhone X). The app estimated weight, macronutrient (carbohydrate, protein, fat), and energy content of 48 randomly chosen meals (breakfasts, cooked meals, snacks) encompassing 128 food items. The reference weight was generated by weighing individual food items using a precision scale. The study endpoints were (1) error of estimated meal weight, (2) error of estimated meal macronutrient content and energy content, (3) segmentation performance, and (4) processing time. ResultsIn both absolute and relative terms, the mean (SD) absolute errors of the app’s estimates were 35.1 g (42.8 g; relative absolute error: 14.0% [12.2%]) for weight; 5.5 g (5.1 g; relative absolute error: 14.8% [10.9%]) for carbohydrate content; 1.3 g (1.7 g; relative absolute error: 12.3% [12.8%]) for fat content; 2.4 g (5.6 g; relative absolute error: 13.0% [13.8%]) for protein content; and 41.2 kcal (42.5 kcal; relative absolute error: 12.7% [10.8%]) for energy content. Although estimation accuracy was not affected by the viewing angle, the type of meal mattered, with slightly worse performance for cooked meals than for breakfasts and snacks. Segmentation adjustment was required for 7 of the 128 items. Mean (SD) processing time across all meals was 22.9 seconds (8.6 seconds). ConclusionsThis study evaluated the accuracy of a novel smartphone app with an integrated depth-sensing camera and found highly accurate volume estimation across a broad range of food items. In addition, the system demonstrated high segmentation performance and low processing time, highlighting its usability.http://mhealth.jmir.org/2020/3/e15294/ |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Herzig, David Nakas, Christos T Stalder, Janine Kosinski, Christophe Laesser, Céline Dehais, Joachim Jaeggi, Raphael Leichtle, Alexander Benedikt Dahlweid, Fried-Michael Stettler, Christoph Bally, Lia |
spellingShingle |
Herzig, David Nakas, Christos T Stalder, Janine Kosinski, Christophe Laesser, Céline Dehais, Joachim Jaeggi, Raphael Leichtle, Alexander Benedikt Dahlweid, Fried-Michael Stettler, Christoph Bally, Lia Volumetric Food Quantification Using Computer Vision on a Depth-Sensing Smartphone: Preclinical Study JMIR mHealth and uHealth |
author_facet |
Herzig, David Nakas, Christos T Stalder, Janine Kosinski, Christophe Laesser, Céline Dehais, Joachim Jaeggi, Raphael Leichtle, Alexander Benedikt Dahlweid, Fried-Michael Stettler, Christoph Bally, Lia |
author_sort |
Herzig, David |
title |
Volumetric Food Quantification Using Computer Vision on a Depth-Sensing Smartphone: Preclinical Study |
title_short |
Volumetric Food Quantification Using Computer Vision on a Depth-Sensing Smartphone: Preclinical Study |
title_full |
Volumetric Food Quantification Using Computer Vision on a Depth-Sensing Smartphone: Preclinical Study |
title_fullStr |
Volumetric Food Quantification Using Computer Vision on a Depth-Sensing Smartphone: Preclinical Study |
title_full_unstemmed |
Volumetric Food Quantification Using Computer Vision on a Depth-Sensing Smartphone: Preclinical Study |
title_sort |
volumetric food quantification using computer vision on a depth-sensing smartphone: preclinical study |
publisher |
JMIR Publications |
series |
JMIR mHealth and uHealth |
issn |
2291-5222 |
publishDate |
2020-03-01 |
description |
BackgroundQuantification of dietary intake is key to the prevention and management of numerous metabolic disorders. Conventional approaches are challenging, laborious, and lack accuracy. The recent advent of depth-sensing smartphones in conjunction with computer vision could facilitate reliable quantification of food intake.
ObjectiveThe objective of this study was to evaluate the accuracy of a novel smartphone app combining depth-sensing hardware with computer vision to quantify meal macronutrient content using volumetry.
MethodsThe app ran on a smartphone with a built-in depth sensor applying structured light (iPhone X). The app estimated weight, macronutrient (carbohydrate, protein, fat), and energy content of 48 randomly chosen meals (breakfasts, cooked meals, snacks) encompassing 128 food items. The reference weight was generated by weighing individual food items using a precision scale. The study endpoints were (1) error of estimated meal weight, (2) error of estimated meal macronutrient content and energy content, (3) segmentation performance, and (4) processing time.
ResultsIn both absolute and relative terms, the mean (SD) absolute errors of the app’s estimates were 35.1 g (42.8 g; relative absolute error: 14.0% [12.2%]) for weight; 5.5 g (5.1 g; relative absolute error: 14.8% [10.9%]) for carbohydrate content; 1.3 g (1.7 g; relative absolute error: 12.3% [12.8%]) for fat content; 2.4 g (5.6 g; relative absolute error: 13.0% [13.8%]) for protein content; and 41.2 kcal (42.5 kcal; relative absolute error: 12.7% [10.8%]) for energy content. Although estimation accuracy was not affected by the viewing angle, the type of meal mattered, with slightly worse performance for cooked meals than for breakfasts and snacks. Segmentation adjustment was required for 7 of the 128 items. Mean (SD) processing time across all meals was 22.9 seconds (8.6 seconds).
ConclusionsThis study evaluated the accuracy of a novel smartphone app with an integrated depth-sensing camera and found highly accurate volume estimation across a broad range of food items. In addition, the system demonstrated high segmentation performance and low processing time, highlighting its usability. |
url |
http://mhealth.jmir.org/2020/3/e15294/ |
work_keys_str_mv |
AT herzigdavid volumetricfoodquantificationusingcomputervisiononadepthsensingsmartphonepreclinicalstudy AT nakaschristost volumetricfoodquantificationusingcomputervisiononadepthsensingsmartphonepreclinicalstudy AT stalderjanine volumetricfoodquantificationusingcomputervisiononadepthsensingsmartphonepreclinicalstudy AT kosinskichristophe volumetricfoodquantificationusingcomputervisiononadepthsensingsmartphonepreclinicalstudy AT laesserceline volumetricfoodquantificationusingcomputervisiononadepthsensingsmartphonepreclinicalstudy AT dehaisjoachim volumetricfoodquantificationusingcomputervisiononadepthsensingsmartphonepreclinicalstudy AT jaeggiraphael volumetricfoodquantificationusingcomputervisiononadepthsensingsmartphonepreclinicalstudy AT leichtlealexanderbenedikt volumetricfoodquantificationusingcomputervisiononadepthsensingsmartphonepreclinicalstudy AT dahlweidfriedmichael volumetricfoodquantificationusingcomputervisiononadepthsensingsmartphonepreclinicalstudy AT stettlerchristoph volumetricfoodquantificationusingcomputervisiononadepthsensingsmartphonepreclinicalstudy AT ballylia volumetricfoodquantificationusingcomputervisiononadepthsensingsmartphonepreclinicalstudy |
_version_ |
1721488164819828736 |