The importance of automated and objective monitoring of dietary behavior is becoming increasingly accepted. The advancements in sensor technology along with recent achievements in machine-learning--based signal-processing algorithms have enabled the development of dietary monitoring solutions that yield highly accurate results. A common bottleneck for developing and training machine learning algorithms is obtaining labeled data for training supervised algorithms, and in particular ground truth annotations. Manual ground truth annotation is laborious, cumbersome, can sometimes introduce errors, and is sometimes impossible in free-living data collection. As a result, there is a need to decrease the labeled data required for training. Additionally, unlabeled data, gathered in-the-wild from existing wearables (such as Bluetooth earbuds) can be used to train and fine-tune eating-detection models. In this work, we focus on training a feature extractor for audio signals captured by an in-ear microphone for the task of eating detection in a self-supervised way. We base our approach on the SimCLR method for image classification, proposed by Chen et al. from the domain of computer vision. Results are promising as our self-supervised method achieves similar results to supervised training alternatives, and its overall effectiveness is comparable to current state-of-the-art methods. Code is available at \url{https://github.com/mug-auth/ssl-chewing}.
翻译:对饮食行为进行自动化和客观监测的重要性日益得到人们的接受。传感器技术的进步以及基于机器学习的信号处理算法的最近成就使得能够开发产生高度准确结果的饮食监测解决方案。开发和培训机器学习算法的一个常见瓶颈是,为培训受监督的算法,特别是地面真相说明,正在获取标签数据。人工地面真相说明是困难的、繁琐的,有时在自由生活数据收集中可能引入错误,有时甚至不可能。因此,有必要减少培训所需的标签数据。此外,从现有磨损器(如蓝牙耳膜)中收集的无标签数据可以用来培训和微小的饮食监测方法。在这项工作中,我们侧重于培训用近耳麦克风捕捉到的语音信号的特征提取器,以便以自我监控的方式进行饮食检测。我们的方法基于Chen et al提出的SimCLR图像分类方法。此外,从计算机视野的域域域(如蓝牙耳耳耳耳等)收集的无标签数据,用于培训和微调饮食检测模型模式。结果有希望作为我们目前监督的替代方法。我们现有的自我监督方法。可比较性标准方法。结果是可用于监督式的软件。