Faggioli G., Ferrante M., Ferro N., Perego R., Tonellotto N.
Conversational agents Conversational evaluations Evaluation approach Evaluation framework Evaluation tool Multi-turn Offline Performance Rapid growth
The rapid growth in number and complexity of conversational agents has highlighted the need for suitable evaluation tools to describe their performance. Current offline conversational evaluation approaches rely on collections composed of multiturn conversations, each including a sequence of utterances. Such sequences represent a snapshot of reality: a single dialog between the user and a hypothetical system on a specific topic. We argue that this paradigm is not realistic enough: multiple users will ask diverse questions in variable order, even for a conversation on the same topic. In this work1 we propose a dependency-aware utterances sampling strategy to augment data available in conversational collections while maintaining temporal dependencies within conversations. Using the sampled conversations, we show that the current evaluation framework favours specific systems while penalizing others, leading to biased evaluation. We further show how to exploit dependency-aware utterances permutations in our current evaluation framework and increase the power of statistical evaluation tools such as ANOVA.
Source: SEBD 2022 - Italian Symposium on Advanced Database Systems, Pisa, Italy, 19-22/07/2022
@inproceedings{oai:it.cnr:prodotti:490718, title = {Improving conversational evaluation via a dependency-aware permutation strategy}, author = {Faggioli G. and Ferrante M. and Ferro N. and Perego R. and Tonellotto N.}, booktitle = {SEBD 2022 - Italian Symposium on Advanced Database Systems, Pisa, Italy, 19-22/07/2022}, year = {2022} }