MILU is a joint neural model that allows you to simultaneously predict multiple dialog act items (a dialog act item takes a form of domain-intent(slot, value). Since it is common that, in a multi-domain setting, an utterance has multiple dialog act items, MILU is likely to yield higher performance than conventional single-intent models.
If you want to perform end-to-end evaluation, you can include the trained model by adding the model path (serialization_dir/model.tar.gz) to your ConvLab spec file.
Data
We use the multiwoz data (data/multiwoz/[train|val|test].json.zip).
MILU on datasets in unified format
We support training MILU on datasets that are in our unified format.
For
non-categorical
dialogue acts whose values are in the utterances, we use
slot tagging
to extract the values.
For
categorical
and
binary
dialogue acts whose values may not be presented in the utterances, we treat them as
intents
of the utterances.
Takes MultiWOZ 2.1 (unified format) as an example,
$ python train.py unified_datasets/configs/multiwoz21_user_context3.jsonnet -s serialization_dir
$ python evaluate.py serialization_dir/model.tar.gz test --cuda-device {CUDA_DEVICE} --output_file output/multiwoz21_user/output.json
# to generate output/multiwoz21_user/predictions.json that merges test data and model predictions.
$ python unified_datasets/merge_predict_res.py -d multiwoz21 -s user -p output/multiwoz21_user/output.json
Note that the config file is different from the above. You should set:
"use_unified_datasets": true
in
dataset_reader
and
model
"dataset_name": "multiwoz21"
in
dataset_reader
"train_data_path": "train"
"validation_data_path": "validation"
"test_data_path": "test"
Predict
See
nlu.py
under
multiwoz
and
unified_datasets
directories.
Performance on unified format datasets
To illustrate that it is easy to use the model for any dataset that in our unified format, we report the performance on several datasets in our unified format. We follow
README.md
and config files in
unified_datasets/
to generate
predictions.json
, then evaluate it using
../evaluate_unified_datasets.py
. Note that we use almost the same hyper-parameters for different datasets, which may not be optimal.
MultiWOZ 2.1
Taskmaster-1
Taskmaster-2
Taskmaster-3
Model
Acc
F1
Acc
F1
Acc
F1
Acc
F1
MILU
72.9
85.2
72.9
49.2
79.1
68.7
85.4
80.3
MILU (context=3)
76.6
87.9
72.4
48.5
78.9
68.4
85.1
80.1
Acc: whether all dialogue acts of an utterance are correctly predicted
F1: F1 measure of the dialogue act predictions over the corpus.
References
@inproceedings{lee2019convlab,
title={ConvLab: Multi-Domain End-to-End Dialog System Platform},
author={Lee, Sungjin and Zhu, Qi and Takanobu, Ryuichi and Li, Xiang and Zhang, Yaoqin and Zhang, Zheng and Li, Jinchao and Peng, Baolin and Li, Xiujun and Huang, Minlie and Gao, Jianfeng},
booktitle={Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics},
year={2019}
}
Runs of ConvLab milu on huggingface.co
0
Total runs
0
24-hour runs
0
3-day runs
0
7-day runs
0
30-day runs
More Information About milu huggingface.co Model
milu huggingface.co
milu huggingface.co is an AI model on huggingface.co that provides milu's model effect (), which can be used instantly with this ConvLab milu model. huggingface.co supports a free trial of the milu model, and also provides paid use of the milu. Support call milu model through api, including Node.js, Python, http.
milu huggingface.co is an online trial and call api platform, which integrates milu's modeling effects, including api services, and provides a free online trial of milu, you can try milu online for free by clicking the link below.
milu is an open source model from GitHub that offers a free installation service, and any user can find milu on GitHub to install. At the same time, huggingface.co provides the effect of milu install, users can directly use milu installed effect in huggingface.co for debugging and trial. It also supports api for free installation.