Skip to content

Conversation

@guyueh1
Copy link
Contributor

@guyueh1 guyueh1 commented Jan 13, 2026

What does this PR do ?

As title

Issues

List issues that this PR closes (syntax):

Usage

  • You can potentially add a usage example below
# Add a code snippet demonstrating how to use this

Before your PR is "Ready for review"

Pre checks:

  • Make sure you read and followed Contributor guidelines
  • Did you write any new necessary tests?
  • Did you run the unit tests and functional tests locally? Visit our Testing Guide for how to run tests
  • Did you add or update any necessary documentation? Visit our Document Development Guide for how to write, build and test the docs.

Additional Information

  • ...

Summary by CodeRabbit

  • Documentation
    • Added new performance benchmark sections for H100 BF16, H100 FP8, and GB200 BF16 configurations with system details and dataset references
    • Restructured benchmark tables for improved clarity with new Algorithm column
    • Updated benchmark metrics and performance data across all sections

✏️ Tip: You can customize this high-level summary in your review settings.

Signed-off-by: Guyue Huang <guyueh@nvidia.com>
@guyueh1 guyueh1 requested a review from a team as a code owner January 13, 2026 23:54
@github-actions github-actions bot added the documentation Improvements or additions to documentation label Jan 13, 2026
@guyueh1 guyueh1 requested a review from snowmanwwg January 13, 2026 23:55
@coderabbitai
Copy link
Contributor

coderabbitai bot commented Jan 13, 2026

📝 Walkthrough

Walkthrough

Documentation update to the performance summary benchmark file. Adds three new benchmark sections for H100 BF16/FP8 and GB200 BF16 configurations with restructured table format including an Algorithm column and updated metrics for training and generation parameters.

Changes

Cohort / File(s) Summary
Performance Summary Documentation
docs/about/performance-summary.md
Adds three new benchmark sections (H100 BF16, H100 FP8, GB200 BF16) with expanded table structure; replaces compact header with Algorithm-inclusive format; updates metric rows (T-Max Length, G-Average Seq len, #GPUs, G-GBS, T-GBS, Tokens/sec per GPU, Total Step time) across GRPO and DAPO algorithm entries.

Estimated code review effort

🎯 2 (Simple) | ⏱️ ~12 minutes

Possibly related PRs

Suggested reviewers

  • terrykong
🚥 Pre-merge checks | ✅ 4
✅ Passed checks (4 passed)
Check name Status Explanation
Description Check ✅ Passed Check skipped - CodeRabbit’s high-level summary is enabled.
Title check ✅ Passed The title 'docs: v0.5 performance results update' directly and clearly reflects the main change: updating documentation with v0.5 performance results as shown in the raw summary.
Docstring Coverage ✅ Passed No functions found in the changed files to evaluate docstring coverage. Skipping docstring coverage check.
Test Results For Major Changes ✅ Passed This pull request contains only minor changes—it updates performance benchmark documentation without introducing any code modifications, new features, breaking changes, or significant refactoring.

✏️ Tip: You can configure your own custom pre-merge checks in the settings.


Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share

Comment @coderabbitai help to get the list of available commands and usage tips.

Copy link
Contributor

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 1

🤖 Fix all issues with AI agents
In @docs/about/performance-summary.md:
- Line 46: The section header "## Nemo RL v0.4" is inconsistent with the PR
title referencing v0.5; update the header text in the markdown (replace "Nemo RL
v0.4" with "Nemo RL v0.5") so the documentation version matches the PR and other
references to the release.
📜 Review details

Configuration used: Path: .coderabbit.yaml

Review profile: CHILL

Plan: Pro

📥 Commits

Reviewing files that changed from the base of the PR and between e95efb9 and 6babe73.

📒 Files selected for processing (1)
  • docs/about/performance-summary.md
🧰 Additional context used
📓 Path-based instructions (2)
docs/**/*.md

📄 CodeRabbit inference engine (CODING_GUIDELINES.md)

Update docs/index.md when a new markdown doc is added under docs/**/*.md or a markdown file is renamed, ensuring the document appears in the most appropriate section

Files:

  • docs/about/performance-summary.md
!(**/tests/**|**/test_*.py|**/test_*.sh)

📄 CodeRabbit inference engine (CODING_GUIDELINES.md)

Add the NVIDIA copyright header to all Python files and shell scripts (excluding tests). The header should include the current year

Files:

  • docs/about/performance-summary.md
🧠 Learnings (1)
📚 Learning: 2025-11-24T17:24:47.707Z
Learnt from: CR
Repo: NVIDIA-NeMo/RL PR: 0
File: coderabbit-custom-pre-merge-checks-unique-id-file-non-traceable-F7F2B60C-1728-4C9A-8889-4F2235E186CA.txt:0-0
Timestamp: 2025-11-24T17:24:47.707Z
Learning: If a change could affect performance, the PR description should include before-and-after performance numbers, as well as the configuration and context in which they apply

Applied to files:

  • docs/about/performance-summary.md
⏰ Context from checks skipped due to timeout of 90000ms. You can increase the timeout in your CodeRabbit configuration to a maximum of 15 minutes (900000ms). (6)
  • GitHub Check: build-container / main
  • GitHub Check: Lint check
  • GitHub Check: build-container / main
  • GitHub Check: Lint check
  • GitHub Check: Post submodule check comment / Comment on PR
  • GitHub Check: Post automodel integration comment / Comment on PR
🔇 Additional comments (2)
docs/about/performance-summary.md (2)

48-78: LGTM! Well-structured benchmark sections.

The H100 BF16 and FP8 benchmark sections are well-organized with clear metadata and properly formatted tables. The addition of the "Algorithm" column effectively distinguishes between GRPO and DAPO results, and the dataset references on line 49 appropriately mention both algorithms used.


79-96: LGTM! GB200 benchmark section is well-structured.

The GB200 BF16 benchmark section follows the same clear structure as the H100 sections with appropriate system metadata and properly formatted tables. The results showcase performance on the new GB200-NVL72 system.


---

## Nemo RL v0.4
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🔴 Critical

Version mismatch: Section header says "v0.4" but PR title indicates "v0.5".

The section header states "Nemo RL v0.4" while the PR title is "v0.5 performance results update". Please verify and correct the version number to ensure documentation accuracy.

📝 Proposed fix if this should be v0.5
-## Nemo RL v0.4
+## Nemo RL v0.5
📝 Committable suggestion

‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.

Suggested change
## Nemo RL v0.4
## Nemo RL v0.5
🤖 Prompt for AI Agents
In @docs/about/performance-summary.md at line 46, The section header "## Nemo RL
v0.4" is inconsistent with the PR title referencing v0.5; update the header text
in the markdown (replace "Nemo RL v0.4" with "Nemo RL v0.5") so the
documentation version matches the PR and other references to the release.

@guyueh1
Copy link
Contributor Author

guyueh1 commented Jan 18, 2026

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

documentation Improvements or additions to documentation r0.5.0

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant