-
Notifications
You must be signed in to change notification settings - Fork 237
Ynankani/update windows benchmark md #762
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
… info Signed-off-by: unknown <ynankani@nvidia.com>
… info Signed-off-by: unknown <ynankani@nvidia.com>
📝 WalkthroughWalkthroughDocumentation expanded with three new metric subsections (MMLU Scores, Perplexity, KL-divergence) under Accuracy Comparison, including explanatory text, comparison tables, and evaluation guide links. Changes
Estimated code review effort🎯 1 (Trivial) | ⏱️ ~5 minutes 🚥 Pre-merge checks | ✅ 2 | ❌ 1❌ Failed checks (1 inconclusive)
✅ Passed checks (2 passed)
✏️ Tip: You can configure your own custom pre-merge checks in the settings. Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. Comment |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 0
🧹 Nitpick comments (1)
examples/windows/Benchmark.md (1)
58-58: Consider adding a note about different test configurations.The Perplexity and KL-divergence sections use RTX 5090 with v0.39.0, while earlier sections use RTX 4090 with v0.19.0. While the configurations are clearly stated in each section, consider adding a brief note explaining that different benchmarks were run at different times with different hardware/software versions to help readers understand why the configurations differ.
Also applies to: 78-78
📜 Review details
Configuration used: Path: .coderabbit.yaml
Review profile: CHILL
Plan: Pro
📒 Files selected for processing (1)
examples/windows/Benchmark.md
⏰ Context from checks skipped due to timeout of 90000ms. You can increase the timeout in your CodeRabbit configuration to a maximum of 15 minutes (900000ms). (4)
- GitHub Check: linux
- GitHub Check: wait-checks / wait
- GitHub Check: code-quality
- GitHub Check: build-docs
🔇 Additional comments (3)
examples/windows/Benchmark.md (3)
27-28: LGTM! Good organizational improvement.Adding the explicit subsection numbering (1.2.1) improves document structure and makes it consistent with the new sections that follow.
45-68: All verification checks pass: the external documentation links are accessible and the internal evaluation guide path exists atexamples/windows/accuracy_benchmark/perplexity_metrics/README.md.
70-92: All references verified—no issues found.The KL-divergence section is correct: the internal evaluation guide path (
./accuracy_benchmark/kl_divergence_metrics/README.md) exists, and both external documentation links are accessible.
Codecov Report✅ All modified and coverable lines are covered by tests. Additional details and impacted files@@ Coverage Diff @@
## main #762 +/- ##
=======================================
Coverage 74.62% 74.62%
=======================================
Files 192 192
Lines 18989 18989
=======================================
Hits 14171 14171
Misses 4818 4818 ☔ View full report in Codecov by Sentry. 🚀 New features to boost your workflow:
|
What does this PR do?
Type of change: ? documentation
Overview: Md update to add perplexity and kl divergence benchmark info.
Before your PR is "Ready for review"
Summary by CodeRabbit
✏️ Tip: You can customize this high-level summary in your review settings.