|
Base model: [microsoft/DialoGPT-small](https://huggingface.co/microsoft/DialoGPT-small) |
|
|
|
Fine tuned for dialogue response generation on the [Persuasion For Good Dataset](https://gitlab.com/ucdavisnlp/persuasionforgood) (Wang et al., 2019) |
|
|
|
Three additional special tokens were added during the fine-tuning process: |
|
- <|pad|> padding token |
|
- <|user|> speaker control token to prompt user responses |
|
- <|system|> speaker control token to prompt system responses |
|
|
|
The following Dialogues were excluded: |
|
- Those with donation amounts outside of the task range of [$0, $2]. |
|
- Those where a donation of 0 was made at the end of the task but a non-zero amount was pledged in the dialogue. |
|
- Those with more than 800 words. |
|
|
|
Stats: |
|
- Training set: 519 dialogues |
|
- Validation set: 58 dialogues |
|
- ~20 utterances per dialogue |