AInvest Newsletter
Daily stocks & crypto headlines, free to your inbox
OpenAI has released two open-source large language models that deliver performance comparable to its premium commercial offerings, enabling local execution on consumer hardware. The models, named gpt-oss-120b and gpt-oss-20b, are available under the Apache 2.0 license, allowing unrestricted use, modification, and commercialization by individuals and organizations, including potential competitors[1].
The 120-billion-parameter gpt-oss-120b operates on a single 80GB GPU, while the 20-billion-parameter gpt-oss-20b functions on devices with at least 16GB of VRAM. Both models support context lengths of up to 128,000 tokens, matching the capabilities of GPT-4o. Despite their large scale, the models activate only a subset of parameters per token—5.1 billion for the 120B model and 3.6 billion for the 20B model—via a mixture-of-experts architecture, enabling efficient deployment[1].
Performance benchmarks show that the gpt-oss-120b outperforms similarly sized open-source models in reasoning tasks and tool use. On Codeforces competition coding, it achieved an Elo rating of 2622 with tools and 2463 without, approaching the performance of OpenAI’s o3 model. The model also reached 96.6% accuracy on the AIME 2024 math competition, surpassing o4-mini’s 87.3%, and scored 57.6% on the HealthBench evaluation—higher than o3’s 50.1%. The smaller gpt-oss-20b also showed strong performance, with scores of 2516 Elo, 95.2% on AIME 2024, and 42.5% on HealthBench[1].
OpenAI trained the models using reinforcement learning and techniques from its o3 and other advanced systems. The company emphasized that both models support three levels of reasoning effort—low, medium, and high—allowing developers to balance performance and latency with simple adjustments in the system message. The models also feature unsupervised chain-of-thought reasoning, a design choice aimed at enabling ongoing monitoring of model behavior, deception, and misuse[1].
Safety was a central focus of the release. OpenAI filtered out harmful data related to chemical, biological, radiological, and nuclear threats during pre-training. The post-training phase incorporated deliberate alignment and instruction hierarchy to prevent unsafe responses and defend against prompt injections. Additionally, the models underwent adversarial fine-tuning and evaluation by three independent expert groups to test their potential for misuse. The results indicated that even after such fine-tuning, the models did not reach dangerous capability levels according to OpenAI’s Preparedness Framework[1].
The release of the models coincides with growing speculation about the imminent launch of GPT-5. OpenAI CEO Sam Altman hinted at significant updates in the coming days, stating, “We have a lot of new stuff for you over the next few days. Something big-but-small today. And then a big upgrade later this week.”[1].
OpenAI’s latest open-weight models mark a strategic move in the open-source AI space, following GPT-2 in 2019. The company aims to provide strong performance on consumer hardware while maintaining safety and control over harmful outputs, even after modifications by third parties.
Source:
[1] OpenAI Drops Two Open Source AI Models That Run Locally and Match Premium Offerings (https://decrypt.co/333617/openai-two-open-source-ai-models-run-locally-match-premium)

Quickly understand the history and background of various well-known coins

Dec.02 2025

Dec.02 2025

Dec.02 2025

Dec.02 2025

Dec.02 2025
Daily stocks & crypto headlines, free to your inbox
Comments
No comments yet