Resource-constrained IoT devices increasingly rely on deep learning models, however, these models experience significant accuracy drops due to domain shifts when encountering variations in lighting, weather, and seasonal conditions. While cloud-based retraining can address this issue, many IoT deployments operate with limited connectivity and energy constraints, making traditional fine-tuning approaches impractical. We explore this challenge through the lens of wildlife ecology, where camera traps must maintain accurate species classification across changing seasons, weather, and habitats without reliable connectivity. We introduce WildFit, an autonomous in-situ adaptation framework that leverages the key insight that background scenes change more frequently than the visual characteristics of monitored species. WildFit combines background-aware synthesis to generate training samples on-device with drift-aware fine-tuning that triggers model updates only when necessary to conserve resources. Our background-aware synthesis surpasses efficient baselines by 7.3% and diffusion models by 3.0% while being orders of magnitude faster, our drift-aware fine-tuning achieves Pareto optimality with 50% fewer updates and 1.5% higher accuracy, and the end-to-end system outperforms domain adaptation approaches by 20-35% while consuming only 11.2 Wh over 37 days-enabling battery-powered deployment.
翻译:资源受限的物联网设备日益依赖深度学习模型,然而,当遇到光照、天气和季节条件变化时,这些模型会因域偏移而出现显著的精度下降。虽然基于云端的重训练可以解决此问题,但许多物联网部署因连接受限和能量约束而无法采用传统的微调方法。我们通过野生动物生态学的视角探讨这一挑战:相机陷阱必须在缺乏可靠连接的情况下,在变化的季节、天气和栖息地环境中保持准确的物种分类。我们提出了WildFit,一种自主原位自适应框架,其核心洞见在于背景场景的变化频率高于被监测物种的视觉特征变化。WildFit结合了背景感知合成技术(在设备端生成训练样本)与漂移感知微调机制(仅在必要时触发模型更新以节约资源)。我们的背景感知合成方法在效率上超越基线方法7.3%,优于扩散模型3.0%,且速度提升数个数量级;漂移感知微调以减少50%的更新次数和提升1.5%的精度实现了帕累托最优;端到端系统在37天内仅消耗11.2 Wh电能(支持电池供电部署)的同时,性能超越领域自适应方法20-35%。