It learns by interacting with its environment and inventing explanatory theories.
In the example below, our prototype agent learns a toy robotics task from only 3 attempts.
No neural networks, and no pre-training.
Guesses theories about what's there and how it works.
Tests its theories by interacting with its environment.
Learns from its mistakes by revising its theories.
Based on principles from philosophy. No deep learning or statistics.
Sight, sound, etc.
One recording is the "training" episode; the other is the "testing" episode.
The agent's core principles were developed independent of the task, and are the foundation for solving more complex tasks.
The sensory input at t0 is the first "data" ever seen by the agent.
First, the agent looks for any new problems in its theories and attempts to reconcile them.
In the first timestep, the agent has no theories yet, so there are no problems to reconcile.
We choose the goal for now, and we can modify the goal at any time.
It chooses a strategy for reaching its goal. If it doesn't have any strategies, it needs to invent one.
This time, the agent's sensory input reveals a problem with its theories. It reconciles the problem by rejecting one or more theories.
This time, it can rule out more theories because it has more sensory experience.
t3, it reaches the goal for the first time.But it doesn't yet understand how it reached the goal...
It discards the problematic theories despite reaching the goal.
Let's skip ahead to when it reaches the goal for the third time, at t8.
t8, the agent reaches the goal for the third time...Importantly, it didn't experience any problems on the way to the goal.
It knows how to reach its goal without encountering any problems.
A problem can come from an error in its theories or from not-knowing how to reach its goal.
If the agent doesn't discover any new problems, then it has no "fuel" to continue learning.
When the agent's theories are good-enough to reliably reach the goal without encountering problems, then learning is complete.
The agent retains all of its knowledge from training.
It needs to conjecture theories about what's there rather than how it works.
The agent got lucky and guessed its situation on the first try.
This indicates a threshold level of understanding of this situation, simulation, and task.
Tasks that require less abstraction and reasoning, e.g. assembly, welding, factory tasks.
Tasks that require more abstraction and reasoning, e.g. language, math, decision-making, multi-step processes.
Tasks that require creating and testing new designs or new theories.
What if LLMs are a dead end? We need new ideas — we can't afford to go all-in on one direction.
If we could answer this question, we could "write down" the code for AGI — today.
It won't be the case that someone writes down the code for AGI before answering this question.
From what I've seen so far, I think a small, dedicated team could create AGI within 10 years — if they're searching in the right direction.
Reach out to Collin Kindrom for more information, or with any other questions or ideas.
collinkindrom@gmail.com