Skip to content

Commit

Permalink
Revert changes to nemotron recipe; will put those changes in performa…
Browse files Browse the repository at this point in the history
…nce scripts in a separate PR

Signed-off-by: Guyue Huang <[email protected]>
  • Loading branch information
guyueh1 committed Jan 24, 2025
1 parent 5bf3f74 commit 6a218f1
Show file tree
Hide file tree
Showing 2 changed files with 0 additions and 6 deletions.
2 changes: 0 additions & 2 deletions nemo/collections/llm/recipes/nemotron4_15b.py
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,6 @@
from nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger
from nemo.collections.llm.recipes.nemotron import nemotron_model, nemotron_trainer
from nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing
from nemo.collections.llm.recipes.tp_overlap_configs.userbuffers import userbuffers_bf16_b200_h6144_tp2_mbs1_seqlen4096
from nemo.lightning.pytorch.callbacks.megatron_comm_overlap import MegatronCommOverlapCallback
from nemo.utils.exp_manager import TimingCallback

Expand Down Expand Up @@ -203,7 +202,6 @@ def pretrain_performance_optimizations(recipe: run.Partial) -> run.Partial:
run.Config(
MegatronCommOverlapCallback,
tp_comm_overlap=True,
tp_comm_overlap_cfg=userbuffers_bf16_b200_h6144_tp2_mbs1_seqlen4096,
)
)
return recipe
Expand Down
4 changes: 0 additions & 4 deletions nemo/collections/llm/recipes/nemotron4_340b.py
Original file line number Diff line number Diff line change
Expand Up @@ -25,9 +25,6 @@
from nemo.collections.llm.recipes.log.default import default_log, default_resume, tensorboard_logger
from nemo.collections.llm.recipes.nemotron import nemotron_model, nemotron_trainer
from nemo.collections.llm.recipes.optim.adam import distributed_fused_adam_with_cosine_annealing
from nemo.collections.llm.recipes.tp_overlap_configs.userbuffers import (
userbuffers_bf16_b200_h18432_tp8_mbs1_seqlen4096,
)
from nemo.lightning.pytorch.callbacks.megatron_comm_overlap import MegatronCommOverlapCallback
from nemo.utils.exp_manager import TimingCallback

Expand Down Expand Up @@ -212,7 +209,6 @@ def pretrain_performance_optimizations(recipe: run.Partial) -> run.Partial:
run.Config(
MegatronCommOverlapCallback,
tp_comm_overlap=True,
tp_comm_overlap_cfg=userbuffers_bf16_b200_h18432_tp8_mbs1_seqlen4096,
defer_embedding_wgrad_compute=True,
wgrad_deferral_limit=22,
overlap_param_gather_with_optimizer_step=False, # Currently disabled due to an issue with checkpointing
Expand Down

0 comments on commit 6a218f1

Please sign in to comment.