Skip to content

Latest commit

 

History

History
15 lines (14 loc) · 693 Bytes

dialogpt-transformers-for-dialogues.md

File metadata and controls

15 lines (14 loc) · 693 Bytes
title date categories tags
DialoGPT: Transformers for Dialogues
2021-03-16
buffer
deep-learning
dialogpt
dialogue
machine-learning
text
transformer

DialoGPT is “a tunable gigaword-scale neural network model for generation of conversational responses, trained on Reddit data”. It uses a Transformer based architecture for doing so, because of their great empirical success. Doing so, the creators have attempted to resolve challenges present with neural response generation – i.e. generating texts relevant to the prompt. These are related to the fact that conversations are informal, noisy, and contain abbreviations or errors.