I have experience with world class performance for 1000x less and I’ll do the same with DroidX via the The Droid Flywheel Coldstart.
How? All that matters is the derivative rate of improvement.
The smallest possible system with the highest derivative rate of improvement (acceleration) will eventually dominate. It is not expensive to build these systems and they are hyper efficient to scale.
The way you maximize acceleration is by maximizing:
- rate of feedback
- signal of feedback
The higher both of these, the greater the rate of compounding improvement.
The Opportunity
Both of these dimensions are incredibly weak (relative to potential) in modern AI. Cycles are long and signal is fuzzy because the data is fuzzy. With language, how do you know what is correct? How does a model source its own learning? How does it know how to explore? See The Fallacies of AI .
How do we go from error in performance to improved performance as fast as possible? Modern system fail to even detect “error” or know how to improve. Loss calculations are not enough.
Ideally, this cycle is measured in seconds. In LLMs, you have a big long multi-month or longer cycles where the model grows stale. You have to crawl the internet and collect chat sessions to get data. You have to hire hundreds or thousands of people to manually give feedback to the system. You have to try and learn from a thumbs up on pages of text.
It is hard to even identify a real cycle because it’s hard to even define improvement in many systems. The best we can do is look at their “versions”. In the last half decade models like ChatGPT have had about 5 major evolutions.
You can look at charts like this analysis from Stanford and rough out an annual improvement on the order of 10% or so each year.
This sounds impressive but it’s actually quite low (relative to potential). We can increase the cycle time by 1000x — to less than half a day.
Sound impossible? How could all these smart people with billions not do this already?
It’s not a question of logistics, it’s a question of framing and legacy.
Our Approach
So, how do you maximize differential of improvement?
First, I’ll tell you how it won’t be done:
- hiring people to provide feedback to your models
- using the same systems everyone else is
- force feeding your models data that you sourced
- starting with ungrounded models like language
- spraying billions at the problem
How will it be done?
- give AI a body — reduce learning ambiguity dimensionality by using unambiguous environments like physics that can be self-tested against
- provide infinite high signal data — sim or the physical world offers infinite data on demand
- theorize — self-generate curriculums and test them synthetically, prioritize samples based on prior performance
- be open-loop — ask for help when stuck
- crowdsource tutors — use viral incentives
- experiment — test in the real world
Our approach is to do as Turing did when he invented the computer, he re-framed the problem he was working on so that it could be solvable. The Universal Learning definitions below can be represented mathematically and thus are solvable.
A Universal Learning Machine comes in a few flavors, this is the ladder:
- Bounded Universal Learning — If a machine can generate it’s own curriculum (a theory) and then test it (experiment) then it can learn recursively and learn everything coupled to the dimensions it understands.
- Weak Universal Learning — The ability to expand the dimensions of understanding by plugging in new sensors and new actuators that are given to it ad hoc. Think something like a universal transformer.
- Strong Universal Learning — Invent new sensors and actuators. At this point, the machine truly eclipses the capabilities of humanity.
We will build the smallest possible example of bounded universal learning system in the fastest and highest signal environment — the real world. Then expand up the ladder of the Universal Learning Machine to ever more abstract environments ending with Strong Universal Learning and Droid Capitalism.
Think of it like this, we will produce “baby” Universal Learning Machine and then grow it up while the rest of the world is trying to squeeze an adult through the birth canal.
Throw in a complex viral loop ala The Droid Flywheel Coldstart to drive federated tutoring (rather than in-house limited tutoring) and we have something truly special.
Closing
In a decade there will be tens of millions of droids operating under with an early form of Droid capitalism and Weak Universal Learning. In 50 years there will be at least 3-to-1 per human and arguable Strong Universal Learning.
While the rest of the world is caught up with the slow world of abstract thinking we will blitz krieg the fast world of physical evolution in silicon.
Welcome to the Droid revolution.