Weak supervision, e.g., in the form of partial labels or image tags, is currently attracting significant attention in CNN segmentation as it can mitigate the lack of full and laborious pixel/voxel annotations. Enforcing high-order (global) inequality constraints on the network output, for instance, on the size of the target region, can leverage unlabeled data, guiding training with domain-specific knowledge. Inequality constraints are very flexible because they do not assume exact prior knowledge. However,constrained Lagrangian dual optimization has been largely avoided in deep networks, mainly for computational tractability reasons.To the best of our knowledge, the method of Pathak et al. is the only prior work that addresses deep CNNs with linear constraints in weakly supervised segmentation. It uses the constraints to synthesize fully-labeled training masks (proposals)from weak labels, mimicking full supervision and facilitating dual optimization.We propose to introduce a differentiable term, which enforces inequality constraints directly in the loss function, avoiding expensive Lagrangian dual iterates and proposal generation. From constrained-optimization perspective, our simple approach is not optimal as there is no guarantee that the constraints are satisfied. However, surprisingly,it yields substantially better results than the proposal-based constrained CNNs, while reducing the computational demand for training.In the context of cardiac images, we reached a segmentation performance close to full supervision using a fraction (0.1%) of the full ground-truth labels and image-level tags.While our experiments focused on basic linear constraints such as the target-region size and image tags, our framework can be easily extended to other non-linear constraints.Therefore, it has the potential to close the gap between weakly and fully supervised learning in semantic image segmentation.
翻译:例如,以部分标签或图像标签的形式进行的微弱监督,目前吸引CNN截断部分的注意力,因为这样可以缓解缺乏完整和艰苦的像素/voxel说明的情况。在网络输出中执行高顺序(全球)不平等限制,例如目标区域的规模,可以利用未贴标签的数据,指导特定领域知识的培训。不平等限制非常灵活,因为它们不假定准确的先前知识。然而,在深层次网络中,不严的Lagrangian双重优化在很大程度上避免了,主要是为了计算可移动性的原因。在我们的知识中,Pathak et al. 的方法是以前解决深度CNN(全球)不平等的限制,在监管不力的截断部分中受到线限制。它利用限制来合成全标签化的培训面具(建议),充分监控和促进双重优化。我们提议引入一个不同的术语,在损失功能中直接实施不均匀的内存限制,避免成本昂贵的Lagrang 双轴和提议生成。从限制到不精确的图像,Pathak-road road road road road road road roduction roduction roduction roduder roduction roduder roduder roduder roduis is is lais lais laut laut laut laut lautes laut laut laut laut laut lautt lat lat lat lautt lauttal lautt lautt lauts lauts laut laut lax laut lax lax lax lax laut lax lauts laut laut lauts laut laut laut laut laut laut laut lauts lauts lauts lauts lauts 完全 laut laut lauts lauts laut 完全 laut laut laut 完全 lax laut laut laut lax lauts lax latal lax 完全 完全 完全 完全