OpenAI Unveils Two New AI Models: gpt-oss-120b and gpt-oss-20b — and run locally on Snapdragon PCs and NVIDIA RTX GPUs

The OpenAI logo appears on the screen of a smartphone.
The two models announced today are the first open-weight models from OpenAI since GPT-2. (Image credit: Getty Images | NurPhoto)

OpenAI just released two generative AI models: gpt-oss-120b and gpt-oss-20b. While neither of them is the highly anticipated GPT-5, they mark a significant step forward for OpenAI.

Both gpt-oss-120b and gpt-oss-20b are open-weight models, meaning the parameters used to train the models are available to the public. OpenAI has not released an open-weight model since GPT-2 in 2019.

AI without the cloud

The new 2025 Zenbook A14 from ASUS featuring Qualcomm Snapdragon X processors and announced at CES 2025.

Many PCs with Snapdragon processors can run one of the latest OpenAI models locally. (Image credit: Future | Daniel Rubino)

I'm going to focus a bit more on gpt-oss-20b because it is capable of running on PCs powered by Snapdragon processors. That's a big deal for Snapdragon-powered computing, which is often limited by compatibility issues.

This time, it is Snapdragon-powered PCs that are supported on day one (though Qualcomm cannot claim exclusivity to supporting the model).

Qualcomm discussed the milestone:

"OpenAI has open-sourced its first reasoning model, gpt-oss-20b, a chain-of-thought reasoning model that runs directly on devices with flagship Snapdragon processors. OpenAI’s sophisticated models have been previously confined to the cloud. Today marks the first time the company is making its model available for on-device inference."

The gpt-oss-20b, which delivers similar results to o3-mini on certain benchmarks, can run on devices with only 16GB of RAM.

That means the model can run on many of the best Copilot+ PCs.

NVIDIA GeForce RTX 5090 Founders Edition installed in a PC with a 600W adapter

PCs powered by an RTX 5090 GPU can get performance of up to 256 tokens per second when running OpenAI's new models. (Image credit: Windows Central | Ben Wilson)

Today was a milestone for Qualcomm, but NVIDIA was not left out. Both gpt-oss-120b and gpt-oss-20b were trained on NVIDIA H100 GPUs. The GPU giant shared the news in a blog post.

The more powerful of the models required 2.1 million hours of training on those GPUs. The less powerful gpt-oss-20b only took around one-tenth of that time, as reported by Hawkdive.

NVIDIA GeForce RTX GPUs can run the gpt-oss-20b model locally. Specifically, the model is supported on systems with at least 16GB of VRAM.

You'll need an NVIDIA RTX PRO GPU if you want to run the more powerful gpt-oss-120b model.

NVIDIA is now a $4 trillion company thanks in large part to its dominance in AI. The company is so dominant in the AI space that President Trump gave up any plans to break it up to increase competition.

Sean Endicott
News Writer and apps editor

Sean Endicott is a tech journalist at Windows Central, specializing in Windows, Microsoft software, AI, and PCs. He's covered major launches, from Windows 10 and 11 to the rise of AI tools like ChatGPT. Sean's journey began with the Lumia 930, leading to strong ties with app developers. Outside writing, he coaches American football, utilizing Microsoft services to manage his team. He studied broadcast journalism at Nottingham Trent University and is active on X @SeanEndicott_ and Threads @sean_endicott_.

You must confirm your public display name before commenting

Please logout and then login again, you will then be prompted to enter your display name.