The Three Brains Concept

A Framework for Self-Improving, Curiosity-Driven AI

A Low-Cost Approach to Autonomous Learning, Reasoning, and Real-World Interaction

Dayyan James

dayyan.james@gmail.com


Abstract

This paper presents a framework for designing a low-cost, self-improving AI model that develops natural curiosity, enhances its reasoning abilities, and interacts with the real world to refine its understanding. Unlike conventional AI systems, which passively process inputs based on predefined training sets, this model actively seeks knowledge, experiments with problem-solving, and autonomously improves itself over long durations. It employs a multi-agent cognitive structure, a hierarchical memory storage system, and quantum-inspired hybrid storage to achieve high efficiency while maintaining adaptability. The AI is initially trained through a combination of existing LLM models and live video and audio feeds captured from smart glasses, allowing it to learn from real-world human interactions. Over time, the AI transitions from passive learning to active engagement, refining its behaviour through an experimental feedback loop.


Inspiration

This framework has been inspired by numerous works in the field of AI and human cognition, with most of the concepts being drawn from human cognition. From Roger Wolcott Sperry’s Split Brain Experiments, Jeff Hawkins novel “A Thousand Brains”, Kevin Simler and Ronin Hanson’s “The Elephant in the Brain”, Dan Ariely’s “Predictably Irrational”, and the my own perception on how his ADHD brain functions.


 

Introduction

Problem Statement

Traditional AI models face several limitations, including, lack of true curiosity, struggles with self-improvement, limited error-driven learning and high computational costs. That is, they do not actively seek information beyond their training data, rely on periodic retraining instead of continuous learning, aren’t strong on efficiently identifying, correcting and learning from mistakes and are expensive to scale.

This paper proposes a framework for an AI model that develops natural curiosity about its environment and human interactions while learning and improving over time by refining its decision-making capabilities. It also aims to create an AI that collaborates with its human trainer benefiting both itself and the user all while operating on relatively low-cost hardware, making continuous learning feasible.

Key Innovations

AI System Architecture

Multi-Agent Cognitive Architecture

ComponentFunction
Cognitive BrainActs as the primary decision-making unit, driving curiosity and determining what to learn next.
Logical & Reasoning BrainVerifies information, corrects errors, and applies structured thinking to solve problems.
Creative & Experimental BrainGenerates alternative solutions, tests hypotheses, and refines knowledge through experimentation.
Hierarchical Memory StorageBalances immediate recall with long-term knowledge archiving.
Quantum-Inspired Hybrid StorageOptimizes data storage efficiency by encoding superposition-like information across two disks.

The AI functions like a self-improving organism, where:

  1. The Cognitive Brain asks questions and directs curiosity.
  2. The Logical Brain verifies facts and evaluates past mistakes.
  3. The Creative Brain explores alternative problem-solving approaches.
  4. Learnings are stored efficiently in hierarchical memory.

Training Strategy

Two-Stage Training Process

StageObjectiveAI Behaviour
Stage 1: Learning PhasePassive observationAI collects real-world experiences via smart glasses, processes interactions, and archives learnings. Relies on existing LLM models to speed up interpretation.
Stage 2: Interactive PhaseActive engagementAI begins live interactions, asking questions, providing insights, and refining its knowledge based on real-time feedback.

Live Human Interaction Data Collection Using Smart Glasses

Memory Optimization & Self-Improvement

Implementation Details

Hardware & Processing Setup

AI Processing Pipeline

  1. Data Collection: Smart glasses capture and transmit key moments.
  2. Real-Time Categorization: AI filters and prioritizes learnings.
  3. Memory Optimization: Important concepts stay in fast memory; detailed learnings are archived.
  4. Self-Improvement: AI continuously tests, refines, and updates its models.

AI Experimentation & Testing

Expected Outcomes & Future Applications

Implementation Strategy

Multi Brain Architecture

A. Multi-Agent AI System (MAS)

Using a multi-agent system, each brain is an independent AI model communicating through a message-passing system.

Logical Brain: A rule-based system or logic-driven model (e.g., Prolog, Bayesian networks for structured reasoning).

Creative Brain: A generative AI model that explores new ideas (diffusion models for image/video processing and reinforcement learning models trained for innovation).

Cognitive Brain: A controller AI that acts as a mediator, selecting which outputs to use.

These agents will communicate asynchronously, allowing for dynamic collaboration.

B. Communication Layer Between Brains

Using message queue system (e.g., RabbitMQ, Kafka) where:

  1. The Cognitive Brain submits a query to both the Logical Brain and Creative Brain.
  2. Each brain processes the query separately and submits its response.
  3. The Cognitive Brain chooses the best action from their outputs.
    1. Accepting one of the outcomes
    2. Passing outcomes between logical and creative brains for logical verification or further experimentation respectively.

Implementation Example:

C. Cognitive Brain: Decision-Making Framework

The Cognitive Brain uses:

Example Logic


D. Logical & Reasoning Brain: Fact-Checking & Structured Thinking

This brain ensures that the AI is grounded in reality.

Implementation Example


E. Creative & Experimental Brain: Out-of-the-Box Thinking

Implementation Example

Final Integration: How the Three Brains Work Together

  1. User Input → Cognitive Brain

    • Example: "What’s the best way to create an alcohol-free perfume with a unique scent profile?"
  2. Cognitive Brain sends the request to both:

    • Logical Brain → "Analyses industry standards, chemical stability, and safety considerations."
    • Creative Brain → "Proposes wild experimental scent combinations and new formulation ideas."
  3. Logical Brain outputs: "MCT oil is a stable base, but diffusion longevity is reduced. Consider using jojoba oil."

  4. Creative Brain outputs: "Try pairing ozone notes with blue cypress and white musk for an ethereal scent."

  5. Cognitive Brain: Evaluates both responses, assigns a weighted score, and decides to experiment with both approaches.

  6. Final Output: A hybrid recommendation that blends structured formulation with creative scent combinations.

Tools & Tech Stack for Building This System

AI Models & Frameworks

Communication & Processing

Decision Optimization

Quantum Inspired Storage

The goal is to allow AI to store, manipulate and reason with superposition-like states on classical hardware with the ultimate goal that the Logic and Reasoning brain can use it for storing fuzzy logic, Q-bit data for uncertainty models; the Creative brain can experiment with Q-bit states to determine how different configurations impact decision making and the Cognitive brain can read the results from both logical and creative brains (making the final decision in Binary from the Q-bit uncertainty). This may be achieved through a dual disk system controlled via a custom FPGA, with the disks simultaneously reading and writing to the same address, encoding:

  1. Classic Binary States (0 and 1)
    1. 1 is written as 1 on both disks
    2. 0 is written as 0 on both disks
  2. Quantum Superposition Equivalent (q-bit state)
    1. Q-bit state written as 0 on Disk A and 1 on Disk B

The system interprets a q-bit state though a third logic state (Q) where:

  1. Q = 0 is classical binary mode
  2. Q= 1 is superpostion-like mode

Read operations must check the Q-bit first to determine the correct interpretation. This is to avoid situation where there is an error or a redundancy check.

Hardware

FPGA
FPGA (Field-Programmable Gate Array) will act as an interface between the AI system and this Quantum inspired storage. The FPGA will handle simultaneous reads/writes to both disks, interpret Q states correctly and communicate with AI over PCIe (or ethernet for scalability later)

Possible hardware options are Xilinx Zynq UltraScale+, Intel Stratix 10, Lattice ECP5.

Disks

High-speed RAMDisks with custom RAM-based buffer storage for short-term dynamic Q-bit states

Integration Layer

PCIe (or ethernet for scalability) for AI to connect to the FPGA and a Neural Processing Unit to offload AI’s encoding and retrieval of Q-bit states.

Logic Design

  1. Encode data onto two disks
    1. 1 written as 1 on both disks
    2. 0 written as 0 on both disks
    3. Q written as 1 on Disk A and 0 on Disk B
  2. Reads data from both disks
    1. If both values match, return classical binary
    2. If mismatched interpret as Q-bit

Logic Diagram

 

              AI <-> FPGA <-> {Disk A, Disk B}

FPGA acts as a data coordinator where AI can query the superposition-like storage, and process the output using probabilistic algorithms.

Adaptive Hierarchical Memory Storage

Multi-Level Memory Architecture

Memory Type

Storage Medium

Data Type

Retrieval Speed

Retention Strategy

Working Memory (Fast, active)

RAM / High-Speed NVMe SSD

High-level concepts, current tasks

Instant

Frequently updated, short-term

Relevant Memory

High-Speed NVMe SSD

Details from commonly used recent concepts, recent completed learnings

Fast

Updated with new information, data indexed

Archived Memory

Vector Database (Pinecone, FAISS) object storage HDDs

Detailed learnings, older information

Slow on demand

Data indexed and compressed. Main concepts compressed, with details heavily compressed or combined with like data.

Relearning Mechanism

Self-learning module

Newly discovered methods

Adaptive

Choose between relearning vs retrieval.

 

Intelligent Memory Compression

Archived learnings are compressed with key metadata tags for faster retrieval, vector embedding indexing used to enable associative search for past experiences and incremental updates are made to the archive ensuring relevance.

Decision Model

Decision Factors

  1. Task Relevance: If memory closely matches current needs, retrieve detailed data. If marginally relevant, weigh the cost of retrieval vs. relearning.

    1. Retrieval vs. Relearning Cost: estimate computational expense of fetching archived data and time priority, if the relearning costs less or offers improvement learn anew.

    2. Knowledge Expiry: validate whether achieved knowledge is outdated and update relearning if so.

Example

Scenario

AI Memory Action

I need to combine ingredient A and B

Checks Active Memory for prior experience

Find a past record that says “specific formulation needed”

Decision: Retrieve archived data or relearn

Retrieval is faster

Loads exact temperature and pH conditions from archived memory

Relearning is faster

AI experiments and updates is knowledge