Back to Models

nvidia/GR00T-N1.7-3B

nvidiageneral

Model Overview

Description:

NVIDIA Isaac GR00T N1.7 is an open foundation model for generalized humanoid robot reasoning and skills. This cross-embodiment model takes multimodal input, including language and images, to perform manipulation tasks in diverse environments. Developers and researchers can post-train GR00T N1.7 with real or synthetic data for their specific humanoid robot or task.

Isaac GR00T N1.7 is the medium-sized version of our model built using pre-trained vision and language encoders, and uses a flow matching action transformer to model a chunk of actions conditioned on vision, language and proprioception.

A detailed description of the Isaac GR00T N1.X architecture is provided in the GROOT N1 White Paper (https://arxiv.org/abs/2503.14734).

This model is ready for commercial/non-commercial use.

Model Developer: NVIDIA

Model Versions

The Isaac GR00T N1.7 model family includes the following 4 models:

GR00T N1.7 – SimplerEnv Bridge

Description
N1.7 post-trained model using the Bridge Dataset in SimplerEnv.

Post-Training Data
https://huggingface.co/datasets/IPEC-COMMUNITY/bridge_orig_lerobot

Dataset Summary
A LeRobot-format conversion of BridgeData V2, originally containing 60,096 trajectories of robot manipulation across 24 environments.

GR00T N1.7 – SimplerEnv Fractal

Description
N1.7 post-trained model using the Fractal Dataset in SimplerEnv.

Post-Training Data
https://huggingface.co/datasets/IPEC-COMMUNITY/bridge_orig_lerobot

Dataset Summary
A LeRobot-format conversion of BridgeData V2, originally containing 60,096 trajectories of robot manipulation across 24 environments.

GR00T N1.7 – Droid

Description
N1.7 post-trained model using the DROID Dataset.

Post-Training Data
https://droid-dataset.github.io/

Dataset Summary
A large-scale “in-the-wild” robot manipulation dataset with approximately 76,000 demonstration trajectories (~350 hours) of interaction data, collected across 564 distinct scenes in 52 buildings, covering 86 manipulation tasks from natural-language instructions.

GR00T N1.7 – LIBERO

Description
N1.7 post-trained model using the LIBERO Dataset.

Post-Training Data
https://github.com/Lifelong-Robot-Learning/LIBERO

Dataset Summary
A benchmark for lifelong robot learning, providing 130 language-conditioned manipulation tasks grouped into multiple task suites.
Includes human-teleoperated demonstrations designed to evaluate knowledge transfer and continual learning in robotic agents.

License

This model is released under the NVIDIA Open Model License Agreement.

Deployment Geography:

Global

Use Case:

Researchers, Academics, Open-Source Community: AI-driven robotics research and algorithm development. Developers: Integrate and customize AI for various robotic applications. Startups & Companies: Accelerate robotics development and reduce training costs.

Release Date:

Computational Load (Internal Only: For NVIDIA Models Only)

Cumulative Compute: Follow Instructions Estimated Energy and Emissions for Model Training: Follow Instructions Total kWh: 64 GB200 nodes * 4 gpus per node x 1200W x 0.001 x 0.8 x 120 hours * 1.4 = 41288 kWh Total Emission: 410.5 * 41288 * 0.000001 = 16.949 tCO2e

Model Architecture:

GR00T-N1.7 VLM backbone is now Cosmos-Reason2-2B

Network Architecture:

The schematic diagram is shown in the illustration above. Red, Green, Blue (RGB) camera frames are processed through a pre-trained vision transformer (SigLip2). Text is encoded by a pre-trained transformer (T5) Robot proprioception is encoded using a multi-layer perceptron (MLP) indexed by the embodiment ID. To handle variable-dimension proprioception, inputs are padded to a configurable max length before feeding into the MLP. Actions are encoded and velocity predictions decoded by an MLP, one per unique embodiment. The flow matching transformer is implemented as a diffusion transformer (DiT), in which the diffusion step conditioning is implemented using adaptive layernorm (AdaLN).

Model Architecture

Number of Model Parameters: 3,000,000,000

Input:

Input Type(s): -Vision: Image Frames -State: Robot Proprioception -Language Instruction: Text -Embodiment ID: Integer

Input Format: -Vision: Variable number of uint8 image frames, coming from robot cameras -State: Floating Point -Language Instruction: String -Embodiment ID: Integer indicating which of the training embodiments is observed

Input Parameters: -Vision: Two-Dimensional (2D) - Red, Green, Blue (RGB) -State: One-Dimensional (1D) - Floating number vector -Language Instruction: One-Dimensional (1D) - String -Embodiment ID: One-Dimensional (1D) - Integer

Output:

Output Type(s): Actions

Output Format Continuous-value vectors

Output Parameters: [Two-Dimensional (2D)]

Other Properties Related to Output: Continuous-value vectors correspond to different motor controls on a robot, which depends on Degrees of Freedom of the robot embodiment.

Our AI models are designed and/or optimized to run on NVIDIA GPU-accelerated systems. By leveraging NVIDIA’s hardware (e.g. GPU cores) and software frameworks (e.g., CUDA libraries), the model achieves faster training and inference times compared to CPU-only solutions.

Software Integration:

Runtime Engine(s): PyTorch

Supported Hardware Microarchitecture Compatibility: All of the below:

  • NVIDIA Ampere
  • NVIDIA Blackwell
  • NVIDIA Jetson
  • NVIDIA Hopper
  • NVIDIA Lovelace

[Preferred/Supported] Operating System(s):

  • Linux

The integration of foundation and fine-tuned models into AI systems requires additional testing using use-case-specific data to ensure safe and effective deployment. Following the V-model methodology, iterative testing and validation at both unit and system levels are essential to mitigate risks, meet technical and functional requirements, and ensure compliance with safety and ethical standards before deployment.

Model Version

GR00T N1.7 EA

Training and Evaluation Datasets:

The total size (in number of data points): 21.6 million
Total number of datasets: 13

Training Dataset:

GR00T Pretraining Data

Data Collection Method by dataset: Hybrid: Human, Robot, Simulated.

Labeling Method by dataset: Hybrid: Human, Automated.

Properties:

  • Cross-embodiment: Data collected on various robot embodiments
  • Sensor types: RGB camera, robot proprioception, robot actuator data

Evaluation:

We evaluate in both simulation and real robot benchmarks, as defined in the White Paper (https://arxiv.org/abs/2503.14734).

Data Collection Method by dataset: Hybrid: Human, Robot, Simulated.

Labeling Method by dataset: Hybrid: Human, Automated.

  • Sim evaluation benchmarks for upper body control
  • 9 DexMG Whitepaper tasks
  • 24 RoboCasa simulated mobile manipulator tasks
  • 24 Digital Cousin simulated GR-1 humanoid manipulation tasks
  • For sim, we automatically measure the success rate in each manipulation behavior.
  • For real robot
    • Grocery packing task
    • Novel objects (unseen from training data)
    • Industrial multi-robot coordination with handoffs
    • Evaluated by human observers in the lab

System Requirements and Performance

This section discusses the various configurations and inference runtimes for GR00T N1.7 tasks. We discuss both latency and speedup.

GR00T N1.7 Inference Timing (4 denoising steps, 1 camera):

DeviceModeData ProcessingBackboneAction HeadE2EFrequencyE2E Speedup
dGPU
H100 80GB HBM3PyTorch Eager6.2 ms31.3 ms48.2 ms85.8 ms11.7 Hz1.00x
torch.compile6.2 ms30.4 ms12.0 ms48.6 ms20.6 Hz1.77x
TensorRT (Full Pipeline)6.2 ms8.8 ms12.3 ms27.9 ms35.9 Hz3.08x
H20 96GB HBM3PyTorch Eager5.33 ms30.8 ms47.3 ms83.4 ms12.0 Hz1.00x
torch.compile5.33 ms31.1 ms13.3 ms49.7 ms20.1 Hz1.68x
TensorRT (Full Pipeline)5.33 ms14.2 ms14.5 ms34.0 ms29.4 Hz2.45x
RTX Pro 6000 BlackwellPyTorch Eager4.8 ms29.3 ms44.0 ms78.4 ms12.8 Hz1.00x
torch.compile4.8 ms29.4 ms16.5 ms50.7 ms19.7 Hz1.55x
TensorRT (Full Pipeline)4.8 ms9.9 ms13.2 ms27.9 ms35.9 Hz2.81x
RTX Pro 5000 72GBPyTorch Eager8.85 ms54.01 ms63.19 ms126.4 ms7.9 Hz1.00x
torch.compile8.85 ms55.74 ms20.38 ms84.9 ms11.8 Hz1.49x
TensorRT (Full Pipeline)8.85 ms14.37 ms17.33 ms40.5 ms24.7 Hz3.13x
L40PyTorch Eager6.6 ms42.8 ms78.9 ms128.3 ms7.8 Hz1.00x
torch.compile6.6 ms42.7 ms19.8 ms69.0 ms14.5 Hz1.86x
TensorRT (Full Pipeline)6.6 ms13.1 ms18.8 ms38.4 ms26.0 Hz3.34x
L20PyTorch Eager5.7 ms47.58 ms86.92 ms140.3 ms7.1 Hz1.00x
torch.compile5.7 ms47.2 ms20.18 ms73.1 ms13.7 Hz1.92x
TensorRT (Full Pipeline)5.7 ms17.27 ms19.79 ms42.8 ms23.3 Hz3.28x
Jetson / Spark
DGX SparkPyTorch Eager13.14 ms38.22 ms74.94 ms126.4 ms7.9 Hz1.00x
torch.compile13.14 ms39.23 ms56.49 ms108.8 ms9.2 Hz1.16x
TensorRT (Full Pipeline)13.14 ms33.43 ms52.37 ms98.6 ms10.1 Hz1.28x
AGX ThorPyTorch Eager8.21 ms55.26 ms81.65 ms144.9 ms6.9 Hz1.00x
torch.compile8.21 ms55.59 ms64.66 ms128.4 ms7.8 Hz1.13x
TensorRT (Full Pipeline)8.21 ms28.89 ms56.64 ms93.8 ms10.7 Hz1.54x
OrinPyTorch Eager9.45 ms127.6 ms205.39 ms342.8 ms2.9 Hz1.00x
torch.compile9.45 ms128.59 ms78.94 ms217.0 ms4.6 Hz1.58x
TensorRT (DiT-only)9.45 ms128.38 ms78.6 ms216.5 ms4.6 Hz1.58x

Note: Orin uses DiT-only TensorRT (--inference-mode tensorrt) because TRT 10.3 does not support the backbone engine. All other platforms use the full pipeline (--inference-mode trt_full_pipeline).

Inference:

Engine: PyTorch Test Hardware: A6000

Ethical Considerations:

NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse.

Users are responsible for model inputs and outputs. Users are responsible for ensuring safe integration of this model, including implementing guardrails as well as other safety mechanisms, prior to deployment.

For more detailed information on ethical considerations for this model, please see the Model Card++ Explainability, Bias, Safety & Security, and Privacy Subcards.

Please report model quality, risk, security vulnerabilities or NVIDIA AI Concerns here.

Visit Website

0 reviews

5
0
4
0
3
0
2
0
1
0
Likes23
Downloads
📝

No reviews yet

Be the first to review nvidia/GR00T-N1.7-3B!

Model Info

Providernvidia
Categorygeneral
Reviews0
Avg. Rating / 5.0

Community

Likes23
Downloads

Rating Guidelines

★★★★★Exceptional
★★★★Great
★★★Good
★★Fair
Poor