NVIDIA Open-Sources Open Code Reasoning Models (32B, 14B, 7B)

NVIDIA continues to push the boundaries of open AI development by open-sourcing its Open Code Reasoning (OCR) model suite — a trio of high-performance large language models purpose-built for code reasoning and problem-solving. The 32B, 14B, and 7B variants, all released under the Apache 2.0 license.
Benchmarked to Beat the Best
The Open Code Reasoning (OCR) models come with notable benchmark achievements, outperforming OpenAI’s o3-Mini and o1 (low) models on the LiveCodeBench benchmark. LiveCodeBench is a comprehensive evaluation suite for code reasoning tasks such as debugging, code generation, and logic completion in real-world developer environments. In direct comparison, NVIDIA’s 32B OCR model tops the leaderboard in reasoning capability for open models.
This leap in performance is attributed not only to model architecture, but to NVIDIA’s custom “OCR dataset” — a high-quality, code-centric training corpus designed to emphasize instruction-following, reasoning, and multi-step code problem solving. According to NVIDIA, this results in a 30% improvement in token efficiency, allowing the models to produce accurate code and logical outputs with fewer tokens.
A Model Lineup for Every Use Case
The Open Code Reasoning suite comes in three parameter scales:
- OpenCodeReasoning-Nemotron-32B
- OpenCodeReasoning-Nemotron-14B
- OpenCodeReasoning-Nemotron-7B
Each model balances scale with performance. The 32B variant delivers state-of-the-art results for high-performance inference and research; the 14B model provides strong reasoning capabilities with reduced compute requirements, and the 7B variant is ideal for resource-constrained environments while retaining competitive performance on benchmarks.
All models are trained using the Nemotron architecture, NVIDIA’s transformer-based backbone optimized for multilingual, multi-task learning. The model weights and configurations are available on Hugging Face:
Compatible with Open Inference Ecosystems
A key feature of these models is out-of-the-box compatibility with popular inference frameworks:
- llama.cpp for lightweight CPU/GPU inference
- vLLM for optimized GPU serving and speculative decoding
- Transformers by Hugging Face for training and evaluation pipelines
- TGI (Text Generation Inference) for scalable API deployment
This flexibility allows developers, researchers, and enterprises to plug these models into existing code AI infrastructure with minimal overhead.
A Step Forward for Open Code Intelligence
With this release, NVIDIA contributes significantly to the growing ecosystem of open code models. By targeting code reasoning — a domain historically dominated by proprietary models — and releasing under a fully open and permissive license, NVIDIA empowers the broader AI and developer community to build, fine-tune, and deploy advanced reasoning models in production.
The Open Code Reasoning suite adds to NVIDIA’s growing portfolio of open LLMs and strengthens its stance on accessible, transparent AI development. Whether you’re building developer copilots, automated code review agents, or code generation services, these models offer a high-performing, cost-effective, and community-friendly alternative to closed solutions.
Check out the 32B Model, 14B Model, 7B Model and 32B Instruction-Tuned Variant. Also, don’t forget to follow us on Twitter.
Here’s a brief overview of what we’re building at Marktechpost:

Sana Hassan, a consulting intern at Marktechpost and dual-degree student at IIT Madras, is passionate about applying technology and AI to address real-world challenges. With a keen interest in solving practical problems, he brings a fresh perspective to the intersection of AI and real-life solutions.