talked to several friends about o3 this week. Their summarized response is basically “holy crap is this actually happening?”
Yes, this is actually happening. The next few years are going to be insane. This is historic stuff, galactic even.
What’s ridiculous is that there’s no sophisticated discussion about what’s happening. AI labs can’t talk about it. The news barely touches it. The government doesn’t understand it.
The fact that a social media meme app newsfeed is how we discuss the future of humanity feels like some absurdist sitcom, but here we are.
Below is a bunch of my thoughts about what’s happening — my contribution to the X idea abyss.
Note, these thoughts are HALF-BAKED and FUN SPECULATION. I haven’t had enough time to think through / research all of them and I’ll be wrong about many. But I do hope these are interesting to some people out there who are trying to process what’s happening.
Enjoy.
A big question is how hard it will be to make research-caliber synthetic data. I’d guess not that hard. Phd-level math and researcher-level math look qualitatively different to us, but might look the same in kind to an AI, just requiring a couple more magnitudes of RL. I give mathematicians 700 days. (That sounds crazy, but o6 not beating mathematicians sounds equally crazy, so I’m more than 50/50 on this prediction, like all the other predictions in this post). That’s 700 days until humans are no longer the top dogs at math in the known universe.
Will AI automate all software engineers away soon? No. Software engineering is more than making PRs based on hyper clear prompts. Unlike mathematicians, software engineers constantly interface with the physical world, namely other humans. Engineers have to work with customers to understand their needs and with teammates to understand their needs. When engineers are designing an architecture or writing the code, they’re doing it with a ton of organizational context. o4 won’t be able to do that. But o4 will help the engineers who do have the context move 10x faster.
If software engineers are 10x faster then maybe we need fewer? Well, if you take a specific company then yes they might need fewer software engineers bc they can achieve the same output with a leaner team. However, the whole world’s need for software engineers might go up bc the world can def use 10x more quality software. So I think we’ll see a golden age of applications from leaner companies. Personalized microapps for every person and business.
We should take a step back and recognize that software turns itself on its head every generation. Software has and always will be about converting needs into pure logic. That conversion process has risen in the abstraction levels from binary to python. The difference now is that it’s rising to english.
Moving to english opens up coding to the non-technical. But the best builders will still always be the ones who can move up and down abstraction levels.
In short, bc software engineering is really all about understanding and fixing organization’s needs through code, the day software engineering is fully automated is the day all organizations are.
All the labs will quickly follow OpenAI with test-time compute models and some can make up for worse algorithms initially with more compute. They’ll play catch up just like they did with GPT-4. To make these models there’s a mix of common knowledge and each lab’s secret sauce. Unclear how much secret sauce OpenAI has with the o-class models, but their rate of improvement suggests it’s an algorithmic advance (which is easier to replicate) and not some unique mix of data (harder to replicate).
In the age of test-time compute, it’s not clear to me whether having more compute or better models is more important. On the one hand, you can make up for a worse model by throwing more test-time compute at it. On the other hand a slightly better model might save an exponential amount of compute.
It would be kindof funny if Xai catches up to OpenAI because they’re simply better at spinning up massive clusters.
Regardless, there’s not going to be a model moat that lasts longer than a year, bc labs swap researchers like baseball cards, and, perhaps more importantly, the researchers between labs party and sleep with each other. Plus I think researchers are too idealistic to not share information if things got out of hand.
Kindof crazy situation we have here. The AI race is like the nuclear race, but where the Americans and Soviets party together in Los Alamos on weekends and bait each other on twitter with “bet you’re not gonna have the biggest nuke in 2025 lols :)”
The AI race will continue to feel hippy and fun-loving until the government steps in and/or something really bad happens.
o-class models incentivize massive buildout bc they have clear gains with every order of magnitude more compute. Compute providers couldn’t have asked for a better scaling law. I’m guessing this law is what Sam saw when he wanted a multi-trillion dollar compute cluster.
This might not actually be great for Nvidia. o-class models make inference more important than training. I think super optimized inference chips are easier to build than training chips, so Nvidia doesn’t have as much of a moat there.
Very speculative: what if o-class models unlock the aggregated compute from the whole world to train the best models? Like how cool would it be if opensource beats closed souce bc we band together our macbook pros into an inference gigacluster.
It’ll start with anything theoretical in the name. Theoretical physics is up first. If math is actually solved (sounds ridiculous even writing this, but that doesn’t make it not likely), then theoretical physics can’t be that far behind. It too lives in the symbolic realm at which LLMs will be superhuman.
What happens when we have a million AI von neumann’s working day and night in the fields of Lousiana (Meta’s upcoming datacenter)? How quickly will they read every physics paper written by the thousands over the past century and immediately spit out more correct tokens?
Obviously this is the part of the story that is hard to predict. Theoretical physics, chemistry, biology — what if these are a joke to an LLM trained with RL? What reasonable argument at this point do we have that it won’t be? Yes we haven’t seen true innovation from these models yet, but they’ve been mostly at high school / college level and those age groups don’t invent new physics. We’re at phd-level now so we might start seeing some inventiveness.
Governments are not going to sit back and let the Earth be mined by automated robots run by a couple SF companies (regulation). And if the governments are too incompetent to stop them, then angry jobless people might resort to violence (terrorism). Unless people are so brain rotten from AI enhanced media that we can’t function as a society (societal collapse).
If war happens, I think it won’t be a bottleneck, rather an accelerant.
Things are gonna get serious. 2025 might be the last year where AI is this wild thing SF tech twitter memes about, before the normies in suits get involved, so let’s enjoy roon and sama while we can.
We have 5000 years of evidence of humans using the latest technology to kill each other. The post-WW2 peace is an anomaly that could fall apart the second the US missteps or when an adversary thinks a first-strike is necessary to stop the AI acceleration. When the weapons get more lethal, more autonomous, the stakes get higher.
The other big risk is AI causing societal chaos. AI generated media could cause mass confusion, mass hysteria, mass brain rot. An authoritarian country could win the AI race and use the new tech to deprive us all of freedom for 1000s of years.
Another risk is that the AI goes rogue. Meaning it causes something extinction level that we didn’t predict. Especially with RL being back in the game, AI is now discovering its own optimizations instead of trying to match human data (matching humans is safer). But so far the underlying brain of these models is still an LLM and LLMs have show to just understand people. Like if you include in the prompt “make sure not to do anything that could kill us”, burden is on you at this point to claim that it’s still likely to kill us. Of course I haven’t considered all the arguments here, but when I have nightmares about an AI dystopia, I see Chinese and Russian flags, not OpenAI’s logo.
The science-fiction world I’ve always wanted is coming. It’s coming a bit faster than expected — hence the fear — but of all the possible paths to get there, I’m not sure how much better the best path would be. This is a pretty great timeline.
Top of mind things that I hope are coming within a decade:
These don’t seem like science fiction anymore, they feel like nearby science reality.
The only way the future isn’t spectacular now is if we the people mess it up. Like our public opinion, our downstream policies, our societal stability, our international cooperation — these are the roadblocks that could prevent this spectacular future.
But our public opinion, our downstream policies, our societal stability, our international cooperation — this is completely uncertain. That means we collectively are the custodians of the future.
It falls upon each of us to help our world navigate these wild times ahead so that we get a great future and not a horrible one.
Perhaps the transition people need to make is from getting meaning through individual success to getting meaning through collective success. Many of our current jobs will be automated soon. We’ll have to adapt. If you derive meaning from a specific skill, yes that skill might no longer be necessary in 5 years and you’re out of luck. But if you can derive meaning from helping the world however you can, well that isn’t ever going away.
You also might need to accept an unstable life in an unstable world. It’s gonna get weird. You’re prob not gonna have two kids and a dog in the suburbs. You might have two cyborg kids and an AI dog on an interstellar ark.
We’re living on the eve of AGI, and on this Christmas eve I ask that you help make the AGI transition go well, so that I can say hi to you on Christmas eve 3024 AD, on a planet four light years away orbiting Altman Centauri.