1. Multi-IF: Benchmarking LLMs on Multi-Turn and Multilingual Instructions Following
- Author
-
He, Yun, Jin, Di, Wang, Chaoqi, Bi, Chloe, Mandyam, Karishma, Zhang, Hejia, Zhu, Chen, Li, Ning, Xu, Tengyu, Lv, Hongjiang, Bhosale, Shruti, Zhu, Chenguang, Sankararaman, Karthik Abinav, Helenowski, Eryk, Kambadur, Melanie, Tayade, Aditya, Ma, Hao, Fang, Han, and Wang, Sinong
- Subjects
Computer Science - Computation and Language - Abstract
Large Language Models (LLMs) have demonstrated impressive capabilities in various tasks, including instruction following, which is crucial for aligning model outputs with user expectations. However, evaluating LLMs' ability to follow instructions remains challenging due to the complexity and subjectivity of human language. Current benchmarks primarily focus on single-turn, monolingual instructions, which do not adequately reflect the complexities of real-world applications that require handling multi-turn and multilingual interactions. To address this gap, we introduce Multi-IF, a new benchmark designed to assess LLMs' proficiency in following multi-turn and multilingual instructions. Multi-IF, which utilizes a hybrid framework combining LLM and human annotators, expands upon the IFEval by incorporating multi-turn sequences and translating the English prompts into another 7 languages, resulting in a dataset of 4,501 multilingual conversations, where each has three turns. Our evaluation of 14 state-of-the-art LLMs on Multi-IF reveals that it presents a significantly more challenging task than existing benchmarks. All the models tested showed a higher rate of failure in executing instructions correctly with each additional turn. For example, o1-preview drops from 0.877 at the first turn to 0.707 at the third turn in terms of average accuracy over all languages. Moreover, languages with non-Latin scripts (Hindi, Russian, and Chinese) generally exhibit higher error rates, suggesting potential limitations in the models' multilingual capabilities. We release Multi-IF prompts and the evaluation code base to encourage further research in this critical area.
- Published
- 2024