NVIDIA’s New Reasoning Models



AI Summary

Summary of Nvidia’s GTC 2025 Keynote by Jensen Huang

  • Event Overview: The NVIDIA GTC 2025 conference held in San Jose included a keynote by CEO Jensen Huang, focusing on data center advancements.

  • Key Themes:

    • Emphasis on the potential of reasoning models and increased token inference.
    • Alignment of announcements with the needs of investors rather than just developers.
  • Product Updates:

    • Presentation of the new Llama Neotron models based on the Llama 3 family:
      • Llama 3.3 Neotron Super 49b V1: A distilled version of the Llama 3.3 70b model.
      • Llama 3.1 Neotron Nano: An 8B version with enhanced reasoning capabilities.
  • Training Enhancements:

    • New approaches to post-training and reinforcement learning to enhance the reasoning capabilities of models.
    • Nvidia released a dataset with around 20 million samples for training reasoning models, mostly generated using the Deep Seek R1 model and others with permissive licenses.
  • Model Trials:

    • Users can trial the models via Nvidia’s API with options to turn reasoning on or off, affecting the length and detail of the output responses.
    • Observations were made on model performance, particularly noting that the 49b model was more effective than the 8B model in reasoning tasks.
  • Final Thoughts:

    • The dataset provided could assist in training personal reasoning models, offering significant value to developers.
    • Future comparisons with other models, like the qwQ32, are anticipated to assess performance viability.