
What is the Bicameral Model?
The Bicameral Model is a neural architecture that couples two parallel language models through a trainable neural interface operating directly on their intermediate hidden states, enabling real-time latent-channel coordination without generating text tokens between them.
Why It Matters
Traditional multi-agent systems communicate by passing text between models β a slow, token-expensive process that loses information in translation. The Bicameral Model replaces this with a continuous latent channel, proving that agents can coordinate at the representation level. In experiments, coupling two 0.5B parameter models with a calculator auxiliary raised arithmetic accuracy from 36% to 96%. On ZebraLogic, two 0.6B models achieved 1.7Γ the performance of an unaugmented baseline β without any additional training on either base model.
How It Works
A primary model drives the main reasoning task. An auxiliary model runs in lockstep, specializing in tool execution, constraint checking, or code computation. A small trainable neural interface bridges the two: it reads the intermediate hidden states of both models and injects corrective signals back into the primary model's representation stream.
The two base models are typically frozen β only the interface is trained. This gives three key properties:
- Parallel execution β both models process every token simultaneously, not sequentially
- Latent communication β coordination happens at the hidden-state level, bypassing the token bottleneck
- β each model optimizes for a different function without interfering with the other