Skip to yearly menu bar Skip to main content


Poster

ReMaX: Relaxing for Better Training on Efficient Panoptic Segmentation

Shuyang Sun · WEIJUN WANG · Andrew Howard · Qihang Yu · Philip Torr · Liang-Chieh Chen

Great Hall & Hall B1+B2 (level 1) #209
[ ]
Wed 13 Dec 3 p.m. PST — 5 p.m. PST

Abstract:

This paper presents a new mechanism to facilitate the training of mask transformers for efficient panoptic segmentation, democratizing its deployment. We observe that due to the high complexity in the training objective of panoptic segmentation, it will inevitably lead to much higher penalization on false positive. Such unbalanced loss makes the training process of the end-to-end mask-transformer based architectures difficult, especially for efficient models. In this paper, we present ReMaX that adds relaxation to mask predictions and class predictions during the training phase for panoptic segmentation. We demonstrate that via these simple relaxation techniques during training, our model can be consistently improved by a clear margin without any extra computational cost on inference. By combining our method with efficient backbones like MobileNetV3-Small, our method achieves new state-of-the-art results for efficient panoptic segmentation on COCO, ADE20K and Cityscapes. Code and pre-trained checkpoints will be available at https://github.com/google-research/deeplab2.

Chat is not available.