Can you run OpenAI's new gpt-oss AI models on your laptop or phone? Here's what you'll need and how to do it
Powerful on-device AI can be yours – if your PC can handle it

As you may have seen, OpenAI has just released two new AI models – gpt‑oss‑20b and gpt‑oss-120b – which are the first open‑weight models from the firm since GPT‑2.
These two models – one is more compact, and the other much larger – are defined by the fact that you can run them locally. They'll work on your desktop PC or laptop – right on the device, with no need to go online or tap the power of the cloud, provided your hardware is powerful enough.
So, you can download either the 20b version – or, if your PC is a powerful machine, the 120b spin – and play around with it on your computer, check how it works (in text-to-text fashion) and how the model thinks (its whole process of reasoning is broken down into steps). And indeed, you can tweak and build on these open models, though safety guardrails and censorship measures will, of course, be in place.
But what kind of hardware do you need to run these AI models? In this article, I'm examining the PC spec requirements for both gpt‑oss‑20b – the more restrained model packing 21 billion parameters – and gpt‑oss-120b, which offers 117 billion parameters. The latter is designed for data center use, but it will run on a high-end PC, whereas gpt‑oss‑20b is the model designed specifically for consumer devices.
Indeed, when announcing these new AI models, Sam Altman referenced 20b working on not just run-of-the-mill laptops, but also smartphones – but suffice it to say, that's an ambitious claim, which I'll come back to later.
These models can be downloaded from Hugging Face (here's gpt‑oss‑20b and here’s gpt‑oss-120b) under the Apache 2.0 license, or for the merely curious, there's an online demo you can check out (no download necessary).
The smaller gpt-oss-20b model
Minimum RAM needed: 16GB
Sign up for breaking news, reviews, opinion, top tech deals, and more.
The official documentation from OpenAI simply lays out a requisite amount of RAM for these AI models, which in the case of this more compact gpt-oss-20b effort is 16GB.
This means you can run gpt-oss-20b on any laptop or PC that has 16GB of system memory (or 16GB of video RAM, or a combo of both). However, it's very much a case of the more, the merrier – or faster, rather. The model might chug along with that bare minimum of 16GB, and ideally, you'll want a bit more on tap.
As for CPUs, AMD recommends the use of a Ryzen AI 300 series CPU paired with 32GB of memory (and half of that, 16GB, set to Variable Graphics Memory). For the GPU, AMD recommends any RX 7000 or 9000 model that has 16GB of memory – but these aren't hard-and-fast requirements as such.
Really, the key factor is simply having enough memory – the mentioned 16GB allocation, and preferably having all of that on your GPU. This allows all the work to take place on the graphics card, without being slowed down by having to offload some of it to the PC's system memory. Although the so-called Mixture of Experts, or MoE, design OpenAI has used here helps to minimize any such performance drag, thankfully.
Anecdotally, to pick an example plucked from Reddit, gpt-oss-20b runs fine on a MacBook Pro M3 with 18GB.
The bigger gpt-oss-120b model
RAM needed: 80GB
It's the same overall deal with the beefier gpt-oss-120b model, except as you might guess, you need a lot more memory. Officially, this means 80GB, although remember that you don't have to have all of that RAM on your graphics card. That said, this large AI model is really designed for data center use on a GPU with 80GB of memory on board.
However, the RAM allocation can be split. So, you can run gpt-OSS-120b on a computer with 64GB of system memory and a 24GB graphics card (an Nvidia RTX 3090 Ti, for example, as per this Redditor), which makes a total of 88GB of RAM pooled.
AMD's recommendation in this case, CPU-wise, is for its top-of-the-range Ryzen AI Max+ 395 processor coupled with 128GB of system RAM (and 96GB of that allocated as Variable Graphics Memory).
In other words, you're looking at a seriously high-end workstation laptop or desktop (maybe with multiple GPUs) for gpt-oss-120b. However, you may be able to get away with a bit less than the stipulated 80GB of memory, going by some anecdotal reports - though I wouldn't bank on it by any means.
How to run these models on your PC
Assuming you meet the system requirements outlined above, you can run either of these new gpt-oss releases on Ollama, which is OpenAI's platform of choice for using these models.
Head here to grab OIlama for your PC (Windows, Mac, or Linux) - click the button to download the executable, and when it's finished downloading, double click the executable file to run it, and click Install.
Next, run the following two commands in Ollama to obtain and then run the model you want. In the example below, we're running gpt-oss-20b, but if you want the larger model, just replace 20b with 120b.
ollama pull gpt-oss:20b
ollama run gpt-oss:20b
If you prefer another option rather than Ollama, you could use LM Studio instead, using the following command. Again, you can switch 20b for 120b, or vice-versa, as appropriate:
lms get openai/gpt-oss-20b
Windows 11 (or 10) users can exercise the option of Windows AI Foundry (hat tip to The Verge).
In this case, you'll need to install Foundry Local - there's a caveat here, though, and it's that this is still in preview - check out this guide for the full instructions on what to do. Also, note that right now you'll need an Nvidia graphics card with 16GB of VRAM on-board (though other GPUs, like AMD Radeon models, will be supported eventually - remember, this is still a preview release).
Furthermore, macOS support is "coming soon," we're told.
What about smartphones?
As noted at the outset, while Sam Altman said that the smaller AI model runs on a phone, that statement is pushing it.
True enough, Qualcomm did issue a press release (as spotted by Android Authority) about gpt-oss-20b running on devices with a Snapdragon chip, but this is more about laptops – Copilot+ PCs that have Snapdragon X silicon – rather than smartphone CPUs.
Running gpt-oss-20b isn't a realistic proposition for today's phones, though it may be possible in a technical sense (assuming your phone has 16GB+ RAM). Even so, I doubt the results would be impressive.
However, we're not far away from getting these kinds of models running properly on mobiles, and this will surely be in the cards for the near-enough future.
You might also like
Darren is a freelancer writing news and features for TechRadar (and occasionally T3) across a broad range of computing topics including CPUs, GPUs, various other hardware, VPNs, antivirus and more. He has written about tech for the best part of three decades, and writes books in his spare time (his debut novel - 'I Know What You Did Last Supper' - was published by Hachette UK in 2013).
You must confirm your public display name before commenting
Please logout and then login again, you will then be prompted to enter your display name.