Back to Papers

Base models beat aligned models at randomness and creativity

Peter West

Computer science - computation and language

Abstract

Alignment has quickly become a default ingredient in LLM development, with techniques such as reinforcement learning from human feedback making models act safely, follow instructions, and perform ever-better on complex tasks. While these techniques are certainly useful, we propose that they should not be universally applied and demonstrate a range of tasks on which base language models consistently outperform their popular aligned forms. Particularly, we study tasks that require unpredictable ou

Relevance Assessment

Research Gap

Notes

Notes are automatically saved as you type

Tags

creativity frameworks › computational creativityevaluation › human evaltextual genre › poetrycreativity frameworks › psychological/cognitiveevaluates a creative feature › logic (puzzles, etc.)

Search Queries

Paper ID: e86032fb-8d23-4a91-a1bd-70512282827dAdded: 10/26/2025