Unifying Parsing and Tree-Structured Models for Generating Sentence Semantic Representations

Fiche du document

Date

10 juillet 2022

Discipline
Type de document
Périmètre
Langue
Identifiants
Relations

Ce document est lié à :
info:eu-repo/semantics/altIdentifier/doi/10.18653/v1/2022.naacl-srw.33

Collection

Archives ouvertes

Licence

info:eu-repo/semantics/OpenAccess


Sujets proches En

Pattern Model

Citer ce document

Antoine Simoulin et al., « Unifying Parsing and Tree-Structured Models for Generating Sentence Semantic Representations », HAL-SHS : linguistique, ID : 10.18653/v1/2022.naacl-srw.33


Métriques


Partage / Export

Résumé En

We introduce a novel tree-based model that learns its composition function together with its structure. The architecture produces sentence embeddings by composing words according to an induced syntactic tree. The parsing and the composition functions are explicitly connected and, therefore, learned jointly. As a result, the sentence embedding is computed according to an interpretable linguistic pattern and may be used on any downstream task. We evaluate our encoder on downstream tasks, and we observe that it outperforms tree-based models relying on external parsers. In some configurations, it is even competitive with Bert base model. Our model is capable of supporting multiple parser architectures. We exploit this property to conduct an ablation study by comparing different parser initializations. We explore to which extent the trees produced by our model compare with linguistic structures and how this initialization impacts downstream performances. We empirically observe that downstream supervision troubles producing stable parses and preserving linguistically relevant structures.

document thumbnail

Par les mêmes auteurs

Sur les mêmes sujets

Sur les mêmes disciplines

Exporter en