A Structured Self-attentive Sentence Embedding

Zhouhan Lin,Minwei Feng,C. D. Santos,Mo Yu,Bing Xiang,Bowen Zhou,Yoshua Bengio

Published 2017 in International Conference on Learning Representations

ABSTRACT

This paper proposes a new model for extracting an interpretable sentence embedding by introducing self-attention. Instead of using a vector, we use a 2-D matrix to represent the embedding, with each row of the matrix attending on a different part of the sentence. We also propose a self-attention mechanism and a special regularization term for the model. As a side effect, the embedding comes with an easy way of visualizing what specific parts of the sentence are encoded into the embedding. We evaluate our model on 3 different tasks: author profiling, sentiment classification, and textual entailment. Results show that our model yields a significant performance gain compared to other sentence embedding methods in all of the 3 tasks.

PUBLICATION RECORD

  • Publication year

    2017

  • Venue

    International Conference on Learning Representations

  • Publication date

    2017-03-09

  • Fields of study

    Computer Science

  • Identifiers
  • External record

    Open on Semantic Scholar

  • Source metadata

    Semantic Scholar

CITATION MAP

EXTRACTION MAP

CLAIMS

  • No claims are published for this paper.

CONCEPTS

  • No concepts are published for this paper.

REFERENCES

Showing 1-38 of 38 references · Page 1 of 1

CITED BY

Showing 1-100 of 2274 citing papers · Page 1 of 23