Model notes
OpenReasoning Nemotron 7B
Reasoning-tuned dense Nemotron checkpoint that tracks the familiar Qwen2.5 7B memory shape while targeting stronger math and code performance.
7.6B dense • 131,072 context • 4 KV heads
Architecture
Model spec
Architecture
Total params
Active params
Layers
Hidden size
Attention heads
KV heads
KV-bearing layers
Context length
Modality
License
Why it matters
Why memory behaves this way
Research highlight
OpenReasoning-Nemotron-7B is post-trained for deliberate reasoning but keeps the dense grouped-query Qwen2.5 backbone, so fit behavior remains straightforward.
Memory note
Resident weights set the floor, and the grouped KV layout keeps long-context cache growth moderate relative to older full-head dense models.
Checkpoints
Official profiles
Official BF16 checkpoint
BF16 checkpoint
NVIDIA publishes OpenReasoning-Nemotron-7B as a Hugging Face Transformers checkpoint derived from Qwen2.5-7B, so v1 models it with the same dense grouped-query memory geometry.
Sources