Thinking in New Directions: When LLMs Learn to Evolve Their Own Concepts
Opening — Why This Matters Now Large language models can explain quantum mechanics, draft legal memos, and debate philosophy. Yet ask them to solve an ARC-style grid puzzle or sustain a 10-step symbolic argument, and their confidence dissolves into beautifully formatted nonsense. We have spent two years scaling test-time compute: chain-of-thought, self-consistency, tree-of-thought, reinforcement learning with verifiers. All of these methods share a quiet assumption: the model’s internal representation space is fixed. We simply search harder inside it. ...