End-to-end dialogue generation has achieved promising results without using handcrafted features and attributes specific for each task and corpus. However, one of the fatal drawbacks in such approaches is that they are unable to generate informative utterances, so it limits their usage from some real-world conversational applications. This paper attempts at generating diverse and informative responses with a variational generation model, which contains a joint attention mechanism conditioning on the information from both dialogue contexts and extra knowledge.
Knowledge-Grounded Response Generation with Deep Attentional Latent-Variable Model
Hao-Tong Ye,K. Lo,Shang-Yu Su,Yun-Nung (Vivian) Chen
Published 2019 in Computer Speech and Language
ABSTRACT
PUBLICATION RECORD
- Publication year
2019
- Venue
Computer Speech and Language
- Publication date
2019-03-23
- Fields of study
Computer Science
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-35 of 35 references · Page 1 of 1
CITED BY
Showing 1-28 of 28 citing papers · Page 1 of 1