Voaige is an AI research lab applying computational principles from cognitive and systems neuroscience to LLMs.
Most improvements to AI systems target either the agent layer or the model layer. Test Time Cognition operates between them, modulating inference without touching weights or prompt context.
No changes to model weights or agent configuration. Works across closed-source and open-source models alike.
Reinforcement learning is, at its core, memorized search. A way of compressing the results of exhaustive trial-and-error into policy parameters, so that at inference time, the right action can be retrieved without re-running the search. It works. But it has a fundamental ceiling: the quality of the compressed policy is bounded by the quality of the search that produced it.
Hard problems, genuine reasoning, novel planning, open-ended generalization, cannot be fully compressed this way. They require search at test time, not just at training time. This is why scaling training alone is not enough. But naively scaling test-time compute is not enough either. What is needed is carefully designed inference: systems that know when to search, where to search, and how to do so efficiently.
The brain solved this problem long ago. Biological cognition doesn't brute-force search over exponential spaces; it navigates them through hierarchical abstraction, selective attention, and early pruning, performing structured, resource-constrained search on 20 watts. What makes this possible is an underlying capacity for adaptation: assessing difficulty on the fly, allocating compute where uncertainty is high, scaling back where it is not.
At Voaige, we study these computational principles. Our goal is to understand how the brain conducts search, what algorithmic strategies it employs, and where the real gains come from, then implement principled approximations of those mechanisms using the representational machinery of large language models.
This is what we call Test Time Cognition: search-capable inference that is architecturally grounded in neuroscience, not just scaled compute.
Voaige is a drop-in OpenAI-compatible endpoint. No changes to your model, your agent, or your prompts. Instant accuracy gains, at lower cost per task.
Mini-SWE-agent · Terminal-Bench 2.0 · April 2026.
We introduce our first test-time cognition algorithm and evaluate it on agentic coding tasks across multiple model configurations and agent harnesses.
Read results → Research essayWe hold model weights and prompts fixed and explore everything else, discovering levers within LLMs that improve reasoning at test time.
Read essay → Research essayAI today encodes intelligence into weights and retrieves it at inference. Hard problems demand more: active reasoning that balances accuracy with efficiency.
Read essay →