Conversation
Signed-off-by: Jingyu Xin <jingyux@nvidia.com>
|
Auto-sync is disabled for draft pull requests in this repository. Workflows must be run manually. Contributors can view more details about this message here. |
|
Important Review skippedAuto incremental reviews are disabled on this repository. Please check the settings in the CodeRabbit UI or the You can disable this status message by setting the
📝 WalkthroughWalkthroughThis PR adds conditional Transformer Engine (TE) support to the Megatron LoRA plugin by introducing TE-based LoRA wrapper classes and corresponding quantized variants with proper registry registration. Comprehensive tests validate LoRA and quantization behavior on TE-integrated Megatron models across multiple GPU workers. Changes
Estimated code review effort🎯 3 (Moderate) | ⏱️ ~25 minutes 🚥 Pre-merge checks | ✅ 2 | ❌ 1❌ Failed checks (1 warning)
✅ Passed checks (2 passed)
✏️ Tip: You can configure your own custom pre-merge checks in the settings. ✨ Finishing touches🧪 Generate unit tests (beta)
Important Action Needed: IP Allowlist UpdateIf your organization protects your Git platform with IP whitelisting, please add the new CodeRabbit IP address to your allowlist:
Reviews will stop working after February 8, 2026 if the new IP is not added to your allowlist. Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. Comment |
Codecov Report✅ All modified and coverable lines are covered by tests. Additional details and impacted files☔ View full report in Codecov by Sentry. 🚀 New features to boost your workflow:
|
Signed-off-by: Jingyu Xin <jingyux@nvidia.com>
Edwardf0t1
left a comment
There was a problem hiding this comment.
LGTM.
@jingyu-ml Could you add more context about why we need this change and what use cases are supported?
This is because not every model uses Megatron’s non-TE parallel linear layer. We need to make PEFT easier to use when a model relies on the TE parallel linear layer. |
Signed-off-by: Jingyu Xin <jingyux@nvidia.com>
What does this PR do?
Type of change: new feature
Overview:
Added optional Transformer Engine (TE) support in modelopt/torch/peft/lora/plugins/megatron.py by importing TEColumnParallelLinear and TERowParallelLinear behind a try/except guard (HAS_TE).
Testing
Before your PR is "Ready for review"
Additional Information
Summary by CodeRabbit
New Features
Tests