Overprotective Training Environments Fall Short at Testing Time: Let Models Contribute to Their Own Training

Fiche du document

Date

3 septembre 2021

Discipline
Périmètre
Langue
Identifiants
Collection

OpenEdition Books

Organisation

OpenEdition

Licences

https://www.openedition.org/12554 , info:eu-repo/semantics/openAccess



Sujets proches En

Skills training

Citer ce document

Alberto Testoni et al., « Overprotective Training Environments Fall Short at Testing Time: Let Models Contribute to Their Own Training », Accademia University Press, ID : 10.4000/books.aaccademia.8934


Métriques


Partage / Export

Résumé 0

Despite important progress, conversational systems often generate dialogues that sound unnatural to humans. We conjecture that the reason lies in their different training and testing conditions: agents are trained in a controlled “lab” setting but tested in the “wild”. During training, they learn to generate an utterance given the human dialogue history. On the other hand, during testing, they must interact with each other, and hence deal with noisy data. We propose to fill this gap by training the model with mixed batches containing both samples of human and machine-generated dialogues. We assess the validity of the proposed method on GuessWhat?!, a visual referential game.

document thumbnail

Par les mêmes auteurs

Sur les mêmes sujets

Sur les mêmes disciplines

Exporter en