The project consists of building an intent classifier which purpose is to pedict the sequence of labels in a dialogue
Abstract :
The hype around OpenAI's ChatGPT has more than ever sparked interest in AI-based bots where labeling and classification of utterances are a centerpiece in order to improve user experience. Broadly, Dialogue Acts (DA) and Emotion/Sentiment (E/S) tasks are identified through sequence labeling systems that are trained in a supervised manner. In this work, we propose four encoder-decoder models to learn generic representations adapted to the spoken dialog, which we evaluate on six datasets of different sizes of the Sequence labellIng Evaluation benChmark fOr spoken laNguagE benchmark (SILICONE) benchmark. Designed models are represented with either a hierarchical encoder or non-hierarchical encoders both based on pre-trained transformers (BERT/XLNet). We notice the failure of the models to learn some datasets due to their inherent properties but in general, the BERT-GRU architecture is the best model regarding accuracy.
- Clone the repository
git clone https://github.com/konkinit/intent_classification.git
- Upgrade
pip
and install the dependencies
python -m pip install --upgrade pip
pip install -r requirements.txt
-
Run the script
./src/utils/get_datasets.py
until all the experiment datasets of SILICONE are downloaded -
Run the notebook
./notebooks/experimental_results.ipynb
We design 4 models based on the below encoder-decoder architecture where
The models we dsigned have been applied to some datasets of SILICONE to obtain the following results:
Architecture | ||||||
---|---|---|---|---|---|---|
BERT + MLP | 37.4 | 63.5 | 69.1 | 86.1 | 52.0 | 57.8 |
BERT + GRU | 44.0 | 81.9 | 69.3 | 86.7 | 60.5 | 70.3 |
XLNet + MLP | 39.1 | 61.7 | 69.3 | 85.7 | 52.3 | 53.7 |
XLNet + GRU | 58.7 | 78.3 | 69.3 | 85.3 | 51.2 | 63.9 |