We describe our two new datasets with images described by humans. Both the datasets were collected using Amazon Mechanical Turk, a crowdsourcing platform. The two datasets contain significantly more descriptions per image than other existing datasets. One is based on a popular image description dataset called the UIUC Pascal Sentence Dataset, whereas the other is based on the Abstract Scenes dataset con- taining images made from clipart objects. In this paper we describe our interfaces, analyze some properties of and show example descriptions from our two datasets.
翻译:我们用人类描述的图像描述我们的两个新的数据集。 这两套数据集都是用众包平台亚马逊机械土耳其( 众包平台) 收集的。 这两套数据集包含的每个图像描述比其他现有数据集要多得多。 其中一套基于一个流行图像描述数据集,称为UIUC Pascal句数据集,而另一套基于“ 抽象场景” 数据集,由剪辑对象制成的图像。 在本文中,我们描述我们的界面,分析我们两个数据集的一些属性,并展示其示例描述。