Cover image

The Forest Within: How Galaxy Reinvents LLM Agents with Self-Evolving Cognition

In a field where many agents act like well-trained dogs, obediently waiting for commands, Galaxy offers something more radical: a system that watches, thinks, adapts, and evolves—without needing to be told. It’s not just an intelligent personal assistant (IPA); it’s an architecture that redefines what intelligence means for LLM-based agents. Let’s dive into why Galaxy is a leap beyond chatty interfaces and into cognition-driven autonomy. 🌳 Beyond Pipelines: The Cognition Forest At the heart of Galaxy lies the Cognition Forest, a structured semantic space that fuses cognitive modeling and system design. Each subtree represents a facet of agent understanding: ...

August 7, 2025 · 4 min · Zelina
Cover image

Sketching a Thought: How Mental Imagery Could Unlock Autonomous Machine Reasoning

From Reaction to Reflection Modern AI models, especially language models, are stunningly capable at answering our queries. But what happens when there is no query? Can an AI reason about the world not just in reaction to prompts, but proactively — triggered by internal goals, simulated futures, and visual imagination? That’s the central question Slimane Larabi explores in his latest paper: “Can Mental Imagery Improve the Thinking Capabilities of AI Systems?” ...

July 18, 2025 · 3 min · Zelina
Cover image

Mind Over Modules: How Smart Agents Learn What to See—and What to Be

In the race to build more autonomous, more intelligent AI agents, we’re entering an era where “strategy” isn’t just about picking the next move—it’s about choosing the right mind for the job and deciding which version of the world to trust. Two recent arXiv papers—one on state representation in dynamic routing games, the other on self-generating agentic systems with swarm intelligence—show just how deeply this matters in practice. We’re no longer only asking: What should the agent do? We now must ask: ...

June 19, 2025 · 5 min · Zelina
Cover image

The Memory Advantage: When AI Agents Learn from the Past

What if your AI agent could remember the last time it made a mistake—and plan better this time? From Reaction to Reflection: Why Memory Matters Most language model agents today operate like goldfish—brilliant at reasoning in the moment, but forgetful. Whether navigating virtual environments, answering complex questions, or composing multi-step strategies, they often repeat past mistakes simply because they lack a memory of past episodes. That’s where the paper “Agentic Episodic Control” by Zhihan Xiong et al. introduces a critical upgrade to today’s LLM agents: a modular episodic memory system inspired by human cognition. Instead of treating each prompt as a blank slate, this framework allows agents to recall, adapt, and refine prior reasoning paths—without retraining the underlying model. ...

June 3, 2025 · 3 min
Cover image

The Crossroads of Reason: When AI Hallucinates with Purpose

The Crossroads of Reason: When AI Hallucinates with Purpose On this day of reflection and sacrifice, we ask not what AI can do, but what it should become. Good Friday is not just a historical commemoration—it’s a paradox made holy: a moment when failure is reinterpreted as fulfillment, when death is the prelude to transformation. In today’s Cognaptus Insights, we draw inspiration from this theme to reimagine the way we evaluate, guide, and build large language models (LLMs). ...

April 18, 2025 · 6 min