The architecture of deep convolutional networks (CNNs) has evolved for years, becoming more accurate and faster. However, it is still challenging to design reasonable network structures that aim at obtaining the best accuracy under a limited computational budget. In this paper, we propose a Tree block, named after its appearance, which extends the One-Shot Aggregation (OSA) module while being more lightweight and flexible. Specifically, the Tree block replaces each of the $3\times3$ Conv layers in OSA into a stack of shallow residual block (SRB) and $1\times1$ Conv layer. The $1\times1$ Conv layer is responsible for dimension increasing and the SRB is fed into the next step. By doing this, when aggregating the same number of subsequent feature maps, the Tree block has a deeper network structure while having less model complexity. In addition, residual connection and efficient channel attention(ECA) is added to the Tree block to further improve the performance of the network. Based on the Tree block, we build efficient backbone models calling TreeNets. TreeNet has a similar network architecture to ResNet, making it flexible to replace ResNet in various computer vision frameworks. We comprehensively evaluate TreeNet on common-used benchmarks, including ImageNet-1k for classification, MS COCO for object detection, and instance segmentation. Experimental results demonstrate that TreeNet is more efficient and performs favorably against the current state-of-the-art backbone methods.
翻译:深层革命网络(CNNs)的架构多年来不断演变,变得更加准确和更快。然而,设计合理的网络结构,目的是在有限的计算预算下获得最准确性,仍然具有挑战性。在本文件中,我们提议了一个以外观命名的树块,以扩展单片聚合模块,同时更轻和灵活。具体地说,树块将OSA的3美元3时3分3分的Conv层替换成一堆浅层(SRB)和1\times1美元Conv层。 1\times1美元Conv层是用来增加维度和将SRB纳入下一步的。我们这样做时,在汇总相同数量的随后的地貌地图时,树块的网络结构更深一些,同时不那么复杂。此外,树块连接和高效的频道注意力(ECA)被添加到树块块中,以进一步提高网络的性能。基于树块块,我们建立高效的主干模型,叫TreNet。TreNet有一个类似的网络结构结构结构结构结构与ResNet相似,包括全面检测系统网络结构结构结构部分,以灵活地取代了SRONet。