DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models
View PDF
HTML (experimental)
Abstract:Despite their impressive capabilities, large language models (LLMs) are prone to hallucinations, i.e., generating content that deviates from facts seen during pretraining. We propose a simple decoding strategy for reducing hallucinations with pretrained LLMs that does not require conditioning on retrieved external knowledge nor additional fine-tuning. Our approach obtains the next-token distribution by contrasting the differences in logits obtained from proj...
Read more at arxiv.org