Hedonic Reinforcement as a Unifying Objective for Attention-Based Language Models staff authors and LLM [ChatGPT 5.2 Pro]January 2026 Abstract Transformer language models trained primarily via next-token prediction exhibit remarkable pattern completion and generalization, yet they often lack stable, long-horizon goal pursuit, robust online adaptation, and consistent preference satisfaction in interactive settings. In biological agents, learning and action selection are strongly shaped by affective valuation: organisms…
Read more: Pleasure Is All You Need