Mosquito-borne diseases (MBDs), such as dengue virus, chikungunya virus, and West Nile virus, cause over one million deaths globally every year. Because many such diseases are spread by the Aedes and Culex mosquitoes, tracking these larvae becomes critical in mitigating the spread of MBDs. Even as citizen science grows and obtains larger mosquito image datasets, the manual annotation of mosquito images becomes ever more time-consuming and inefficient. Previous research has used computer vision to identify mosquito species, and the Convolutional Neural Network (CNN) has become the de-facto for image classification. However, these models typically require substantial computational resources. This research introduces the application of the Vision Transformer (ViT) in a comparative study to improve image classification on Aedes and Culex larvae. Two ViT models, ViT-Base and CvT-13, and two CNN models, ResNet-18 and ConvNeXT, were trained on mosquito larvae image data and compared to determine the most effective model to distinguish mosquito larvae as Aedes or Culex. Testing revealed that ConvNeXT obtained the greatest values across all classification metrics, demonstrating its viability for mosquito larvae classification. Based on these results, future research includes creating a model specifically designed for mosquito larvae classification by combining elements of CNN and transformer architecture.
翻译:蚊子传染疾病(MBDs),如登革热病毒、Chikungunyya病毒和西尼罗病毒(West Nero病毒),每年在全球造成100多万人死亡。由于许多这类疾病是由Aedes和Cullike蚊子传播的,因此跟踪这些幼虫对于减缓MBDs的传播至关重要。即使公民科学增长并获得更多的蚊子成像数据集,对蚊子图像的人工注解也变得越来越耗时和低效。以前的研究利用计算机视野来识别蚊子物种,而革命性分类神经网络(CNN)已成为图像分类的脱法。然而,这些模型通常需要大量的计算资源。这一研究在一项比较研究中引入了视野变异变器(Vive Tranger)的应用,以改善Aedes和Culex larvae的图像分类。两个ViT模型,即Vit-Base和CvT-13,以及两个CNNCM模型,ResNet-18和CONNEXT,接受了关于蚊子变形图像数据的培训,并比较了用来区分蚊子变形变形变形模型的最有效模型的模式,具体用来区分蚊子变形为ADRevlevevevaveleva 的模型的模型,展示了所有AX的模型。测试了这些模型,展示了这些模型,展示了整个CON-Nex-Nex的模型。测试了这些模型。