We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent b345941 commit 886f8cbCopy full SHA for 886f8cb
1 file changed
transformer_engine/pytorch/graph.py
@@ -507,11 +507,6 @@ def hook_fn(
507
else:
508
grad_inputs = None
509
del outputs, grad_inputs
510
- # The following code is added specifically for MCore's special requirements,
511
- # aimed at preventing warmup from altering the control flow.
512
- for module in func.modules():
513
- if hasattr(module, "is_first_microbatch"):
514
- module.is_first_microbatch = True
515
torch.cuda.synchronize()
516
517
# All captures here share a mempool. To avoid replays corrupting each other's memory,
0 commit comments