Iteration of training and evaluating a machine learning model is an important process to improve its performance. However, while teachable interfaces enable blind users to train and test an object recognizer with photos taken in their distinctive environment, accessibility of training iteration and evaluation steps has received little attention. Iteration assumes visual inspection of the training photos, which is inaccessible for blind users. We explore this challenge through MyCam, a mobile app that incorporates automatically estimated descriptors for non-visual access to the photos in the users' training sets. We explore how blind participants (N=12) interact with MyCam and the descriptors through an evaluation study in their homes. We demonstrate that the real-time photo-level descriptors enabled blind users to reduce photos with cropped objects, and that participants could add more variations by iterating through and accessing the quality of their training sets. Also, Participants found the app simple to use indicating that they could effectively train it and that the descriptors were useful. However, subjective responses were not reflected in the performance of their models, partially due to little variation in training and cluttered backgrounds.
翻译:培训和评价一个机器学习模式的迭代是提高其性能的一个重要过程。然而,尽管可教学界面使盲人用户能够用在自己独特的环境中拍摄的照片来培训和测试一个对象识别器,但培训迭代和评估步骤的可及性却很少引起注意。迭代假设对培训照片进行视觉检查,而盲人用户无法接触这些照片。我们通过MyCam来探讨这一挑战。MyCam是一个手机应用程序,它包含非视觉访问用户培训成套照片的自动估计描述器。我们探索盲人参与者(N=12)如何通过在他们家中进行的评估研究与MyCam和描述器互动。我们证明,实时的光级描述器使盲人用户能够减少与作物对象的照片,而参与者可以通过循环和获取其培训设备的质量来增加更多的差异。此外,与会者发现使用这一应用程序简单,表明他们能够有效地培训这些照片,而且描述器有用。然而,主观反应并没有反映在他们模型的性能表现中,部分原因是由于培训和背景的模糊性差。