# MoEL
**Repository Path**: PhantomVein/MoEL
## Basic Information
- **Project Name**: MoEL
- **Description**: No description available
- **Primary Language**: Python
- **License**: MIT
- **Default Branch**: master
- **Homepage**: None
- **GVP Project**: No
## Statistics
- **Stars**: 1
- **Forks**: 0
- **Created**: 2019-12-12
- **Last Updated**: 2021-03-06
## Categories & Tags
**Categories**: Uncategorized
**Tags**: None
## README
## MoEL: Mixture of Empathetic Listeners
[](https://opensource.org/licenses/MIT)
This is the PyTorch implementation of the paper:
**MoEL: Mixture of Empathetic Listeners**. [**Zhaojiang Lin**](https://zlinao.github.io/), Andrea Madotto, Jamin Shin, Peng Xu, Pascale Fung ***EMNLP 2019*** [[PDF]](https://arxiv.org/pdf/1908.07687.pdf)
This code has been written using PyTorch >= 0.4.1. If you use any source codes or datasets included in this toolkit in your work, please cite the following paper. The bibtex is listed below:
@article{lin2019moel,
title={MoEL: Mixture of Empathetic Listeners},
author={Lin, Zhaojiang and Madotto, Andrea and Shin, Jamin and Xu, Peng and Fung, Pascale},
journal={arXiv preprint arXiv:1908.07687},
year={2019}
}
## Abstract
Previous research on empathetic dialogue systems has mostly focused on generating responses given certain emotions. However, being empathetic not only requires the ability of generating emotional responses, but more importantly, requires the understanding of user emotions and replying appropriately. In this paper, we propose a novel end-to-end approach for modeling empathy in dialogue systems: Mixture of Empathetic Listeners (MoEL). Our model first captures the user emotions and outputs an emotion distribution. Based on this, MoEL will softly combine the output states of the appropriate Listener(s), which are each optimized to react to certain emotions, and generate an empathetic response. Human evaluations on empathetic-dialogues dataset confirm that MoEL outperforms multitask training baseline in terms of empathy, relevance, and fluency. Furthermore, the case study on generated responses of different Listeners shows high interpretability of our model.
## MoEL Architecture: