Back to Search Start Over

ContraCLM: Contrastive Learning For Causal Language Model

Authors :
Jain, Nihal
Zhang, Dejiao
Ahmad, Wasi Uddin
Wang, Zijian
Nan, Feng
Li, Xiaopeng
Tan, Ming
Nallapati, Ramesh
Ray, Baishakhi
Bhatia, Parminder
Ma, Xiaofei
Xiang, Bing
Source :
ACL 2023
Publication Year :
2022

Abstract

Despite exciting progress in causal language models, the expressiveness of the representations is largely limited due to poor discrimination ability. To remedy this issue, we present ContraCLM, a novel contrastive learning framework at both token-level and sequence-level. We assess ContraCLM on a variety of downstream tasks. We show that ContraCLM enhances discrimination of the representations and bridges the gap with the encoder-only models, which makes causal language models better suited for tasks beyond language generation. Specifically, we attain $44\%$ relative improvement on the Semantic Textual Similarity tasks and $34\%$ on Code-to-Code Search tasks. Furthermore, by improving the expressiveness of the representations, ContraCLM also boosts the source code generation capability with $9\%$ relative improvement on execution accuracy on the HumanEval benchmark.<br />Comment: 10 pages

Details

Database :
arXiv
Journal :
ACL 2023
Publication Type :
Report
Accession number :
edsarx.2210.01185
Document Type :
Working Paper