‘We will learn quickly and course-correct’ — Sam Altman says this is OpenAI’s future, but it’s not the one it started with
Sam Altman posts a new ‘Our Principles’ document, and it’s a big shift
Sign up for breaking news, reviews, opinion, top tech deals, and more.
You are now subscribed
Your newsletter sign-up was successful
Join the club
Get full access to premium articles, exclusive features and a growing list of member rewards.
OpenAI has just published a new document called ‘Our Principles’ written by Sam Altman, and at first glance it reads like a simple corporate manifesto update. But the more I read it, the more I start to think that it isn't what it appears to be.
Something has definitely changed here, compared to OpenAI’s previous statements, so let’s unpack what it is. We can start by taking a look at what it is not saying, as well as what it is saying.
The race of AGI — is it still happening?
For a start, the post is very deliberately credited to Sam Altman, as if it’s something of a personal mission statement, which made me want to compare it to his previous blogs on AI.
Article continues belowWhat immediately struck me as curious about the new principles document, compared to his old blogs, is the lack of reference to Artificial General Intelligence (AGI). Achieving AGI for the benefit of humanity was, after all, the whole goal of creating OpenAI in the first place, but it is only mentioned in passing a few times in the new document. It seems to have been replaced by talking about broader AI deployment instead.
Eleven months ago, Altman was talking in much stronger terms about AGI in his personal blog: “We are past the event horizon;”, he wrote, “the takeoff has started. Humanity is close to building digital superintelligence, and at least so far it’s much less weird than it seems like it should be.”
Reading it back now it sounds like AGI was about to happen at any moment. Compared to the latest document, the language has softened significantly. If the takeoff had started, it seems the AGI rocket is still on the landing pad.
AI: The risks vs the rewards
While the broader benefits of AI are stressed — “a lot of the things we’ve only let ourselves dream about in sci-fi could become reality”, Altman says — there’s also the acknowledgement that these good outcomes aren’t guaranteed.
Sign up for breaking news, reviews, opinion, top tech deals, and more.
“Power in the future can either be held by a small handful of companies using and controlling superintelligence, or it can be held in a decentralized way by people”, Altman says. “We believe the latter is much better, and our goal is to put truly general AI in the hands of as many people as possible.”
Later the document raises more of the dangers of AI, particularly regarding pathogens:
"No AI lab can ensure a good future alone. For an obvious example, there may be extremely capable models that make it easier to create a new pathogen, and we need a society-wide approach to defend against this with pathogen-agnostic countermeasures.”
I was left wondering whether to be excited or scared for the future. And that feeling of contradiction started to increase the more I read...
Competitive spirit
While the document initially reads as collaborative, and there is a lot of talk about getting AI into the hands of everyone: “We want a future where everyone can have an excellent life”, Altman says, the question I have is how that happens.
Altman is talking about distribution, and in practice that usually means shipping faster, integrating it into more products and reaching more users — all things which define competition in tech. For that to happen in practice, it feels like OpenAI is going to have to become more competitive than it already is.
And when Altman says later, “We will learn quickly and course-correct”, he means shipping, learning, improving and repeating, which sits uneasily next to the new documents heavy safety framing.
Even when he says “We deserve an enormous amount of scrutiny” — it sounds humble on the surface, but strategically it’s perhaps more about OpenAI's willingness to justify decisions publicly. It’s the kind of language companies tend to use when they’re under increasing scrutiny.
In fact, the more I read the whole document, the more it starts to sound contradictory. Safety implies restraint and deliberation while scale implies speed and iteration.
After reading the document I’m still not entirely sure what OpenAI’s principles actually are. Principles should be clear and emphatic, while this document feels softer and more flexible, giving OpenAI more room for maneuver than it had before.
It’s hard not to think of the famous quote by Groucho Marx “Those are my principles, and if you don’t like them… well, I have others.”
Follow TechRadar on Google News and add us as a preferred source to get our expert news, reviews, and opinion in your feeds.

➡️ Read our full guide to the best business laptops
1. Best overall:
Dell 14 Premium
2. Best on a budget:
Acer Aspire 5
3. Best MacBook:
Apple MacBook Pro 14-inch (M4)

Graham is the Senior Editor for AI at TechRadar. With over 25 years of experience in both online and print journalism, Graham has worked for various market-leading tech brands including Computeractive, PC Pro, iMore, MacFormat, Mac|Life, Maximum PC, and more. He specializes in reporting on everything to do with AI and has appeared on BBC TV shows like BBC One Breakfast and on Radio 4 commenting on the latest trends in tech. Graham has an honors degree in Computer Science and spends his spare time podcasting and blogging.
You must confirm your public display name before commenting
Please logout and then login again, you will then be prompted to enter your display name.