Abstract
Performance in language modelling has been significantly improved by training recurrent neural networks on large corpora. This progress has come at the cost of interpretability and an understanding of how these architectures function, making principled development of better language models more difficult. We look inside a state-of-the-art neural language model to analyse how this model represents high-level lexico-semantic information. In particular, we investigate how the model represents words by extracting activation patterns where they occur in the text, and compare these representations directly to human semantic knowledge.
Original language | English |
---|---|
Title of host publication | Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP |
Publisher | Association for Computational Linguistics |
Pages | 362-364 |
Number of pages | 3 |
ISBN (Electronic) | 9781948087711 |
DOIs | |
Publication status | Published - Nov 2018 |
Event | 1st Workshop on BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, co-located with the 2018 Conference on Empirical Methods in Natural Language Processing, EMNLP 2018 - Brussels, Belgium Duration: 01 Nov 2018 → … |
Publication series
Name | EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, |
---|---|
Publisher | Association for Computational Linguistics |
Conference
Conference | 1st Workshop on BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, co-located with the 2018 Conference on Empirical Methods in Natural Language Processing, EMNLP 2018 |
---|---|
Country/Territory | Belgium |
City | Brussels |
Period | 01/11/2018 → … |
Bibliographical note
Funding Information:This work was partly funded by a Microsoft Azure for Research Award.
Publisher Copyright:
© 2018 Association for Computational Linguistics
ASJC Scopus subject areas
- Computer Science Applications
- Computational Theory and Mathematics
- Information Systems
Fingerprint
Dive into the research topics of 'Representation of Word Meaning in the Intermediate Projection Layer of a Neural Language Model'. Together they form a unique fingerprint.Student theses
-
Interpretable semantic representations from neural language models and computer vision
Derby, S. (Author), Murphy, B. (Supervisor), Miller, P. (Supervisor) & Devereux, B. (Supervisor), Jul 2022Student thesis: Doctoral Thesis › Doctor of Philosophy
File