TY - GEN
T1 - A comparison between audio and IMU data to detect chewing events based on an earable device
AU - Lotfi, Roya
AU - Tzanetakis, George
AU - Eskicioglu, Rasit
AU - Irani, Pourang
N1 - Publisher Copyright:
© 2020 Association for Computing Machinery.
PY - 2020/5/27
Y1 - 2020/5/27
N2 - The feasibility of collecting various data from built-in wearable sensors has enticed many researchers to use these devices for analyzing human activities and behaviors. In particular, audio, video, and motion data have been utilized for automatic dietary monitoring. In this paper, we investigate the feasibility of detecting chewing activities based on audio and inertial sensor data obtained from an ear-worn device, eSense. We process each sensor data separately and determine the accuracy of each sensing modality for chewing detection when using MFCC and Spectral Centroid as features and Logistic Regression, Decision Tree, and Random Forest as classifiers. We also measure the performance of chewing detection when fusing features extracted from both audio and inertial sensor data. We evaluate the chewing detection algorithm by running a pilot study inside a lab environment on a total of 5 participants. This consists of 130 minutes audio and inertial measurement unit (IMU) data. The results of this study indicate that an in-ear IMU with an accuracy of 95% outperforms audio data in detecting chewing and fusing both modalities improves the accuracy to 97%.
AB - The feasibility of collecting various data from built-in wearable sensors has enticed many researchers to use these devices for analyzing human activities and behaviors. In particular, audio, video, and motion data have been utilized for automatic dietary monitoring. In this paper, we investigate the feasibility of detecting chewing activities based on audio and inertial sensor data obtained from an ear-worn device, eSense. We process each sensor data separately and determine the accuracy of each sensing modality for chewing detection when using MFCC and Spectral Centroid as features and Logistic Regression, Decision Tree, and Random Forest as classifiers. We also measure the performance of chewing detection when fusing features extracted from both audio and inertial sensor data. We evaluate the chewing detection algorithm by running a pilot study inside a lab environment on a total of 5 participants. This consists of 130 minutes audio and inertial measurement unit (IMU) data. The results of this study indicate that an in-ear IMU with an accuracy of 95% outperforms audio data in detecting chewing and fusing both modalities improves the accuracy to 97%.
KW - Audio
KW - Chewing detection
KW - Earables
KW - IMU
KW - MFCC
KW - Machine learning pipeline
KW - Spectral centroid
UR - http://www.scopus.com/inward/record.url?scp=85123041634&partnerID=8YFLogxK
U2 - 10.1145/3396339.3396362
DO - 10.1145/3396339.3396362
M3 - Conference contribution
AN - SCOPUS:85123041634
T3 - ACM International Conference Proceeding Series
BT - Proceedings of the 11th Augmented Human International Conference, AH 2020
PB - Association for Computing Machinery
T2 - 11th Augmented Human International Conference, AH 2020
Y2 - 27 May 2020 through 29 May 2020
ER -