Back to Search Start Over

LongSafetyBench: Long-Context LLMs Struggle with Safety Issues

Authors :
Huang, Mianqiu
Liu, Xiaoran
Zhou, Shaojun
Zhang, Mozhi
Tan, Chenkun
Wang, Pengyu
Guo, Qipeng
Xu, Zhe
Li, Linyang
Lei, Zhikai
Li, Linlin
Liu, Qun
Zhou, Yaqian
Qiu, Xipeng
Huang, Xuanjing
Publication Year :
2024

Abstract

With the development of large language models (LLMs), the sequence length of these models continues to increase, drawing significant attention to long-context language models. However, the evaluation of these models has been primarily limited to their capabilities, with a lack of research focusing on their safety. Existing work, such as ManyShotJailbreak, has to some extent demonstrated that long-context language models can exhibit safety concerns. However, the methods used are limited and lack comprehensiveness. In response, we introduce \textbf{LongSafetyBench}, the first benchmark designed to objectively and comprehensively evaluate the safety of long-context models. LongSafetyBench consists of 10 task categories, with an average length of 41,889 words. After testing eight long-context language models on LongSafetyBench, we found that existing models generally exhibit insufficient safety capabilities. The proportion of safe responses from most mainstream long-context LLMs is below 50\%. Moreover, models' safety performance in long-context scenarios does not always align with that in short-context scenarios. Further investigation revealed that long-context models tend to overlook harmful content within lengthy texts. We also proposed a simple yet effective solution, allowing open-source models to achieve performance comparable to that of top-tier closed-source models. We believe that LongSafetyBench can serve as a valuable benchmark for evaluating the safety capabilities of long-context language models. We hope that our work will encourage the broader community to pay attention to the safety of long-context models and contribute to the development of solutions to improve the safety of long-context LLMs.

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2411.06899
Document Type :
Working Paper