Téléchargement | - Voir la version finale : Syntax encoding with application in Authorship Attribution (PDF, 478 Kio)
|
---|
DOI | Trouver le DOI : https://doi.org/10.18653/v1/D18-1294 |
---|
Auteur | Rechercher : Zhang, Richong; Rechercher : Hu, Zhiyuan; Rechercher : Guo, Hongyu1; Rechercher : Mao, Yongyi |
---|
Affiliation | - Conseil national de recherches du Canada. Technologies numériques
|
---|
Format | Texte, Article |
---|
Conférence | Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Oct. 31 - Nov. 4, 2018, Brussels, Belgium |
---|
Résumé | We propose a novel strategy to encode the syntax parse tree of sentence into a learnable distributed representation. The proposed syntax encoding scheme is provably information-lossless. In specific, an embedding vector is constructed for each word in the sentence, encoding the path in the syntax tree corresponding to the word. The one-to-one correspondence between these “syntax-embedding” vectors and the words (hence their embedding vectors) in the sentence makes it easy to integrate such a representation with all word-level NLP models. We empirically show the benefits of the syntax embeddings on the Authorship Attribution domain, where our approach improves upon the prior art and achieves new performance records on five benchmarking data sets. |
---|
Date de publication | 2018-11-04 |
---|
Maison d’édition | Association for Computational Linguistics |
---|
Licence | |
---|
Dans | |
---|
Langue | anglais |
---|
Publications évaluées par des pairs | Oui |
---|
Exporter la notice | Exporter en format RIS |
---|
Signaler une correction | Signaler une correction (s'ouvre dans un nouvel onglet) |
---|
Identificateur de l’enregistrement | 8f70be2b-f362-4e2e-a318-a7ef5f93e3cd |
---|
Enregistrement créé | 2021-06-17 |
---|
Enregistrement modifié | 2021-06-18 |
---|