Back to Search Start Over

Learnable Linguistic Watermarks for Tracing Model Extraction Attacks on Large Language Models

Authors :
Bai, Minhao
Pang, Kaiyi
Huang, Yongfeng
Publication Year :
2024

Abstract

In the rapidly evolving domain of artificial intelligence, safeguarding the intellectual property of Large Language Models (LLMs) is increasingly crucial. Current watermarking techniques against model extraction attacks, which rely on signal insertion in model logits or post-processing of generated text, remain largely heuristic. We propose a novel method for embedding learnable linguistic watermarks in LLMs, aimed at tracing and preventing model extraction attacks. Our approach subtly modifies the LLM's output distribution by introducing controlled noise into token frequency distributions, embedding an statistically identifiable controllable watermark.We leverage statistical hypothesis testing and information theory, particularly focusing on Kullback-Leibler Divergence, to differentiate between original and modified distributions effectively. Our watermarking method strikes a delicate well balance between robustness and output quality, maintaining low false positive/negative rates and preserving the LLM's original performance.<br />Comment: not decided

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2405.01509
Document Type :
Working Paper