Machine Learning

Addressing Appearance Change in Outdoor Robotics with Adversarial Domain Adaptation

This topic contains 0 replies, has 1 voice, and was last updated by  arXiv 2 years ago.


  • arXiv
    5 pts

    Addressing Appearance Change in Outdoor Robotics with Adversarial Domain Adaptation

    Appearance changes due to weather and seasonal conditions represent a strong impediment to the robust implementation of machine learning systems in outdoor robotics. While supervised learning optimises a model for the training domain, it will deliver degraded performance in application domains that underlie distributional shifts caused by these changes. Traditionally, this problem has been addressed via the collection of labelled data in multiple domains or by imposing priors on the type of shift between both domains. We frame the problem in the context of unsupervised domain adaptation and develop a framework for applying adversarial techniques to adapt popular, state-of-the-art network architectures with the additional objective to align features across domains. Moreover, as adversarial training is notoriously unstable, we first perform an extensive ablation study, adapting many techniques known to stabilise generative adversarial networks, and evaluate on a surrogate classification task with the same appearance change. The distilled insights are applied to the problem of free-space segmentation for motion planning in autonomous driving.

    Addressing Appearance Change in Outdoor Robotics with Adversarial Domain Adaptation
    by Markus Wulfmeier, Alex Bewley, Ingmar Posner
    https://arxiv.org/pdf/1703.01461v2.pdf

You must be logged in to reply to this topic.