Base models beat aligned models at randomness and creativity
Peter West
Computer science - computation and language
Abstract
Alignment has quickly become a default ingredient in LLM development, with techniques such as reinforcement learning from human feedback making models act safely, follow instructions, and perform ever-better on complex tasks. While these techniques are certainly useful, we propose that they should not be universally applied and demonstrate a range of tasks on which base language models consistently outperform their popular aligned forms. Particularly, we study tasks that require unpredictable ou
Relevance Assessment
Research Gap
Notes
Notes are automatically saved as you type
Tags
creativity frameworks › computational creativityevaluation › human evaltextual genre › poetrycreativity frameworks › psychological/cognitiveevaluates a creative feature › logic (puzzles, etc.)
Search Queries
Paper ID: e86032fb-8d23-4a91-a1bd-70512282827dAdded: 10/26/2025