Learning Neural Sequence-to-Sequence Models from Weak Feedback with Bipolar Ramp Loss

In many machine learning scenarios, supervision by gold labels is not available and conse quently neural models cannot be trained directly by maximum likelihood estimation. In a weak supervision scenario, metric-augmented objectives can be employed to assign feedback to model o...

Full description

Bibliographic Details
Main Authors: Jehl, Laura, Lawrence, Carolin, Riezler, Stefan
Format: Article
Language:English
Published: The MIT Press 2019-11-01
Series:Transactions of the Association for Computational Linguistics
Online Access:https://www.mitpressjournals.org/doi/abs/10.1162/tacl_a_00265
id doaj-f2a018e1be9240748622542dd57d1cd0
record_format Article
spelling doaj-f2a018e1be9240748622542dd57d1cd02020-11-25T02:00:22ZengThe MIT PressTransactions of the Association for Computational Linguistics2307-387X2019-11-01723324810.1162/tacl_a_00265Learning Neural Sequence-to-Sequence Models from Weak Feedback with Bipolar Ramp LossJehl, LauraLawrence, CarolinRiezler, Stefan In many machine learning scenarios, supervision by gold labels is not available and conse quently neural models cannot be trained directly by maximum likelihood estimation. In a weak supervision scenario, metric-augmented objectives can be employed to assign feedback to model outputs, which can be used to extract a supervision signal for training. We present several objectives for two separate weakly supervised tasks, machine translation and semantic parsing. We show that objectives should actively discourage negative outputs in addition to promoting a surrogate gold structure. This notion of bipolarity is naturally present in ramp loss objectives, which we adapt to neural models. We show that bipolar ramp loss objectives outperform other non-bipolar ramp loss objectives and minimum risk training on both weakly supervised tasks, as well as on a supervised machine translation task. Additionally, we introduce a novel token-level ramp loss objective, which is able to outperform even the best sequence-level ramp loss on both weakly supervised tasks. https://www.mitpressjournals.org/doi/abs/10.1162/tacl_a_00265
collection DOAJ
language English
format Article
sources DOAJ
author Jehl, Laura
Lawrence, Carolin
Riezler, Stefan
spellingShingle Jehl, Laura
Lawrence, Carolin
Riezler, Stefan
Learning Neural Sequence-to-Sequence Models from Weak Feedback with Bipolar Ramp Loss
Transactions of the Association for Computational Linguistics
author_facet Jehl, Laura
Lawrence, Carolin
Riezler, Stefan
author_sort Jehl, Laura
title Learning Neural Sequence-to-Sequence Models from Weak Feedback with Bipolar Ramp Loss
title_short Learning Neural Sequence-to-Sequence Models from Weak Feedback with Bipolar Ramp Loss
title_full Learning Neural Sequence-to-Sequence Models from Weak Feedback with Bipolar Ramp Loss
title_fullStr Learning Neural Sequence-to-Sequence Models from Weak Feedback with Bipolar Ramp Loss
title_full_unstemmed Learning Neural Sequence-to-Sequence Models from Weak Feedback with Bipolar Ramp Loss
title_sort learning neural sequence-to-sequence models from weak feedback with bipolar ramp loss
publisher The MIT Press
series Transactions of the Association for Computational Linguistics
issn 2307-387X
publishDate 2019-11-01
description In many machine learning scenarios, supervision by gold labels is not available and conse quently neural models cannot be trained directly by maximum likelihood estimation. In a weak supervision scenario, metric-augmented objectives can be employed to assign feedback to model outputs, which can be used to extract a supervision signal for training. We present several objectives for two separate weakly supervised tasks, machine translation and semantic parsing. We show that objectives should actively discourage negative outputs in addition to promoting a surrogate gold structure. This notion of bipolarity is naturally present in ramp loss objectives, which we adapt to neural models. We show that bipolar ramp loss objectives outperform other non-bipolar ramp loss objectives and minimum risk training on both weakly supervised tasks, as well as on a supervised machine translation task. Additionally, we introduce a novel token-level ramp loss objective, which is able to outperform even the best sequence-level ramp loss on both weakly supervised tasks.
url https://www.mitpressjournals.org/doi/abs/10.1162/tacl_a_00265
work_keys_str_mv AT jehllaura learningneuralsequencetosequencemodelsfromweakfeedbackwithbipolarramploss
AT lawrencecarolin learningneuralsequencetosequencemodelsfromweakfeedbackwithbipolarramploss
AT riezlerstefan learningneuralsequencetosequencemodelsfromweakfeedbackwithbipolarramploss
_version_ 1724961005447938048