Unsupervised and weakly supervised approaches for answer selection tasks with scarce annotations

Abstract: Addressing Answer Selection (AS) tasks with complex neural networks typically requires a large amount of annotated data to increase the accuracy of the models. In this work, we are interested in simple models that can potentially give good performance on datasets with no or few annotations. First, we propose new unsupervised baselines that leverage distributed word and sentence representations. Second, we compare the ability of our neural architectures to learn from few annotated examples in a weakly supervised scheme and we demonstrate how these methods can benefit from a pre-training on an external dataset. With an emphasis on results reproducibility, we show that our simple methods can reach or approach state-of-the-art performances on four common AS datasets.

Standort
Deutsche Nationalbibliothek Frankfurt am Main
Umfang
Online-Ressource
Sprache
Englisch

Erschienen in
Unsupervised and weakly supervised approaches for answer selection tasks with scarce annotations ; volume:9 ; number:1 ; year:2019 ; pages:136-144 ; extent:9
Open computer science ; 9, Heft 1 (2019), 136-144 (gesamt 9)

Urheber
Vallee, Emmanuel
Charlet, Delphine
Galassi, Francesca
Marzinotto, Gabriel
Clérot, Fabrice
Meyer, Frank

DOI
10.1515/comp-2019-0008
URN
urn:nbn:de:101:1-2410301502393.631274753975
Rechteinformation
Open Access; Der Zugriff auf das Objekt ist unbeschränkt möglich.
Letzte Aktualisierung
15.08.2025, 07:20 MESZ

Datenpartner

Dieses Objekt wird bereitgestellt von:
Deutsche Nationalbibliothek. Bei Fragen zum Objekt wenden Sie sich bitte an den Datenpartner.

Beteiligte

  • Vallee, Emmanuel
  • Charlet, Delphine
  • Galassi, Francesca
  • Marzinotto, Gabriel
  • Clérot, Fabrice
  • Meyer, Frank

Ähnliche Objekte (12)