Back to Search Start Over

Rule Extrapolation in Language Models: A Study of Compositional Generalization on OOD Prompts

Authors :
Mészáros, Anna
Ujváry, Szilvia
Brendel, Wieland
Reizinger, Patrik
Huszár, Ferenc
Publication Year :
2024

Abstract

LLMs show remarkable emergent abilities, such as inferring concepts from presumably out-of-distribution prompts, known as in-context learning. Though this success is often attributed to the Transformer architecture, our systematic understanding is limited. In complex real-world data sets, even defining what is out-of-distribution is not obvious. To better understand the OOD behaviour of autoregressive LLMs, we focus on formal languages, which are defined by the intersection of rules. We define a new scenario of OOD compositional generalization, termed rule extrapolation. Rule extrapolation describes OOD scenarios, where the prompt violates at least one rule. We evaluate rule extrapolation in formal languages with varying complexity in linear and recurrent architectures, the Transformer, and state space models to understand the architectures' influence on rule extrapolation. We also lay the first stones of a normative theory of rule extrapolation, inspired by the Solomonoff prior in algorithmic information theory.

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2409.13728
Document Type :
Working Paper