Text Generation
Transformers
Safetensors
PyTorch
nvidia
nemotron-h
suhara commited on
Commit
bedad02
·
verified ·
1 Parent(s): 6d8fd0b

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -34,3 +34,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ nemo/weights/.metadata filter=lfs diff=lfs merge=lfs -text
38
+ nemo/weights/__0_0.distcp filter=lfs diff=lfs merge=lfs -text
39
+ nemo/weights/__0_1.distcp filter=lfs diff=lfs merge=lfs -text
nemo/context/io.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"root": {"type": "ref", "key": "trainer_context_1"}, "objects": {"tuple_1": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [["Index(index=0)", "tensor_model_parallel_size"], ["Index(index=1)", "pipeline_model_parallel_size"], ["Index(index=2)", "virtual_pipeline_model_parallel_size"], ["Index(index=3)", "sequence_parallel"], ["Index(index=4)", "context_parallel_size"], ["Index(index=5)", "expert_model_parallel_size"], ["Index(index=6)", "expert_tensor_parallel_size"], ["Index(index=7)", "moe_extended_tp"], ["Index(index=8)", "use_te_rng_tracker"], ["Index(index=9)", "pipeline_dtype"], ["Index(index=10)", "microbatch_group_size_per_vp_stage"], ["Index(index=11)", "num_layers_in_first_pipeline_stage"], ["Index(index=12)", "num_layers_in_last_pipeline_stage"], ["Index(index=13)", "account_for_embedding_in_pipeline_split"], ["Index(index=14)", "account_for_loss_in_pipeline_split"]], "metadata": null}, "dict_1": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}}, "dict_2": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}}, "buildable_traverser_metadata_1": {"type": {"type": "pyref", "module": "fiddle._src.config", "name": "BuildableTraverserMetadata"}, "items": [["Attr(name='fn_or_cls')", {"type": "pyref", "module": "nemo.collections.llm.gpt.model.ssm", "name": "NemotronHConfig47B"}], ["Attr(name='argument_names')", {"type": "ref", "key": "tuple_1"}], ["Attr(name='argument_tags')", {"type": "ref", "key": "dict_1"}], ["Attr(name='argument_history')", {"type": "ref", "key": "dict_2"}]], "metadata": {"type": "pyref", "module": "fiddle._src.config", "name": "BuildableTraverserMetadata"}}, "nemotron_hconfig47_b_1": {"type": {"type": "pyref", "module": "fiddle._src.config", "name": "Config"}, "items": [["Attr(name='tensor_model_parallel_size')", {"type": "leaf", "value": 1, "paths": ["<root>.model.config.tensor_model_parallel_size"]}], ["Attr(name='pipeline_model_parallel_size')", {"type": "leaf", "value": 1, "paths": ["<root>.model.config.pipeline_model_parallel_size"]}], ["Attr(name='virtual_pipeline_model_parallel_size')", {"type": "leaf", "value": null, "paths": ["<root>.model.config.virtual_pipeline_model_parallel_size"]}], ["Attr(name='sequence_parallel')", {"type": "leaf", "value": false, "paths": ["<root>.model.config.sequence_parallel"]}], ["Attr(name='context_parallel_size')", {"type": "leaf", "value": 1, "paths": ["<root>.model.config.context_parallel_size"]}], ["Attr(name='expert_model_parallel_size')", {"type": "leaf", "value": 1, "paths": ["<root>.model.config.expert_model_parallel_size"]}], ["Attr(name='expert_tensor_parallel_size')", {"type": "leaf", "value": null, "paths": ["<root>.model.config.expert_tensor_parallel_size"]}], ["Attr(name='moe_extended_tp')", {"type": "leaf", "value": false, "paths": ["<root>.model.config.moe_extended_tp"]}], ["Attr(name='use_te_rng_tracker')", {"type": "leaf", "value": false, "paths": ["<root>.model.config.use_te_rng_tracker"]}], ["Attr(name='pipeline_dtype')", {"type": "leaf", "value": null, "paths": ["<root>.model.config.pipeline_dtype"]}], ["Attr(name='microbatch_group_size_per_vp_stage')", {"type": "leaf", "value": 1, "paths": ["<root>.model.config.microbatch_group_size_per_vp_stage"]}], ["Attr(name='num_layers_in_first_pipeline_stage')", {"type": "leaf", "value": null, "paths": ["<root>.model.config.num_layers_in_first_pipeline_stage"]}], ["Attr(name='num_layers_in_last_pipeline_stage')", {"type": "leaf", "value": null, "paths": ["<root>.model.config.num_layers_in_last_pipeline_stage"]}], ["Attr(name='account_for_embedding_in_pipeline_split')", {"type": "leaf", "value": false, "paths": ["<root>.model.config.account_for_embedding_in_pipeline_split"]}], ["Attr(name='account_for_loss_in_pipeline_split')", {"type": "leaf", "value": false, "paths": ["<root>.model.config.account_for_loss_in_pipeline_split"]}]], "metadata": {"type": "ref", "key": "buildable_traverser_metadata_1"}, "paths": ["<root>.model.config"]}, "tuple_2": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [["Index(index=0)", "optimizer"], ["Index(index=1)", "lr"], ["Index(index=2)", "min_lr"], ["Index(index=3)", "decoupled_lr"], ["Index(index=4)", "decoupled_min_lr"], ["Index(index=5)", "weight_decay"], ["Index(index=6)", "fp16"], ["Index(index=7)", "bf16"], ["Index(index=8)", "params_dtype"], ["Index(index=9)", "use_precision_aware_optimizer"], ["Index(index=10)", "main_grads_dtype"], ["Index(index=11)", "main_params_dtype"], ["Index(index=12)", "exp_avg_dtype"], ["Index(index=13)", "exp_avg_sq_dtype"], ["Index(index=14)", "loss_scale"], ["Index(index=15)", "initial_loss_scale"], ["Index(index=16)", "min_loss_scale"], ["Index(index=17)", "loss_scale_window"], ["Index(index=18)", "hysteresis"], ["Index(index=19)", "adam_beta1"], ["Index(index=20)", "adam_beta2"], ["Index(index=21)", "adam_eps"], ["Index(index=22)", "sgd_momentum"], ["Index(index=23)", "use_distributed_optimizer"], ["Index(index=24)", "overlap_param_gather_with_optimizer_step"], ["Index(index=25)", "optimizer_cpu_offload"], ["Index(index=26)", "optimizer_offload_fraction"], ["Index(index=27)", "use_torch_optimizer_for_cpu_offload"], ["Index(index=28)", "overlap_cpu_optimizer_d2h_h2d"], ["Index(index=29)", "pin_cpu_grads"], ["Index(index=30)", "pin_cpu_params"], ["Index(index=31)", "clip_grad"], ["Index(index=32)", "log_num_zeros_in_grad"], ["Index(index=33)", "barrier_with_L1_time"], ["Index(index=34)", "timers"], ["Index(index=35)", "config_logger_dir"]], "metadata": null}, "dict_3": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}}, "dict_4": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}}, "buildable_traverser_metadata_2": {"type": {"type": "pyref", "module": "fiddle._src.config", "name": "BuildableTraverserMetadata"}, "items": [["Attr(name='fn_or_cls')", {"type": "pyref", "module": "megatron.core.optimizer.optimizer_config", "name": "OptimizerConfig"}], ["Attr(name='argument_names')", {"type": "ref", "key": "tuple_2"}], ["Attr(name='argument_tags')", {"type": "ref", "key": "dict_3"}], ["Attr(name='argument_history')", {"type": "ref", "key": "dict_4"}]], "metadata": {"type": "pyref", "module": "fiddle._src.config", "name": "BuildableTraverserMetadata"}}, "optimizer_config_1": {"type": {"type": "pyref", "module": "fiddle._src.config", "name": "Config"}, "items": [["Attr(name='optimizer')", {"type": "leaf", "value": "adam", "paths": ["<root>.model.optim.config.optimizer"]}], ["Attr(name='lr')", {"type": "leaf", "value": 0.0001, "paths": ["<root>.model.optim.config.lr"]}], ["Attr(name='min_lr')", {"type": "leaf", "value": null, "paths": ["<root>.model.optim.config.min_lr"]}], ["Attr(name='decoupled_lr')", {"type": "leaf", "value": null, "paths": ["<root>.model.optim.config.decoupled_lr"]}], ["Attr(name='decoupled_min_lr')", {"type": "leaf", "value": null, "paths": ["<root>.model.optim.config.decoupled_min_lr"]}], ["Attr(name='weight_decay')", {"type": "leaf", "value": 0.01, "paths": ["<root>.model.optim.config.weight_decay"]}], ["Attr(name='fp16')", {"type": "leaf", "value": false, "paths": ["<root>.model.optim.config.fp16"]}], ["Attr(name='bf16')", {"type": "leaf", "value": false, "paths": ["<root>.model.optim.config.bf16"]}], ["Attr(name='params_dtype')", {"type": "pyref", "module": "torch", "name": "float32", "paths": ["<root>.model.optim.config.params_dtype", "<root>.model.optim.config.main_grads_dtype", "<root>.model.optim.config.main_params_dtype", "<root>.model.optim.config.exp_avg_dtype", "<root>.model.optim.config.exp_avg_sq_dtype"]}], ["Attr(name='use_precision_aware_optimizer')", {"type": "leaf", "value": false, "paths": ["<root>.model.optim.config.use_precision_aware_optimizer"]}], ["Attr(name='main_grads_dtype')", {"type": "pyref", "module": "torch", "name": "float32", "paths": ["<root>.model.optim.config.params_dtype", "<root>.model.optim.config.main_grads_dtype", "<root>.model.optim.config.main_params_dtype", "<root>.model.optim.config.exp_avg_dtype", "<root>.model.optim.config.exp_avg_sq_dtype"]}], ["Attr(name='main_params_dtype')", {"type": "pyref", "module": "torch", "name": "float32", "paths": ["<root>.model.optim.config.params_dtype", "<root>.model.optim.config.main_grads_dtype", "<root>.model.optim.config.main_params_dtype", "<root>.model.optim.config.exp_avg_dtype", "<root>.model.optim.config.exp_avg_sq_dtype"]}], ["Attr(name='exp_avg_dtype')", {"type": "pyref", "module": "torch", "name": "float32", "paths": ["<root>.model.optim.config.params_dtype", "<root>.model.optim.config.main_grads_dtype", "<root>.model.optim.config.main_params_dtype", "<root>.model.optim.config.exp_avg_dtype", "<root>.model.optim.config.exp_avg_sq_dtype"]}], ["Attr(name='exp_avg_sq_dtype')", {"type": "pyref", "module": "torch", "name": "float32", "paths": ["<root>.model.optim.config.params_dtype", "<root>.model.optim.config.main_grads_dtype", "<root>.model.optim.config.main_params_dtype", "<root>.model.optim.config.exp_avg_dtype", "<root>.model.optim.config.exp_avg_sq_dtype"]}], ["Attr(name='loss_scale')", {"type": "leaf", "value": null, "paths": ["<root>.model.optim.config.loss_scale"]}], ["Attr(name='initial_loss_scale')", {"type": "leaf", "value": 4294967296, "paths": ["<root>.model.optim.config.initial_loss_scale"]}], ["Attr(name='min_loss_scale')", {"type": "leaf", "value": 1.0, "paths": ["<root>.model.optim.config.min_loss_scale"]}], ["Attr(name='loss_scale_window')", {"type": "leaf", "value": 1000, "paths": ["<root>.model.optim.config.loss_scale_window"]}], ["Attr(name='hysteresis')", {"type": "leaf", "value": 2, "paths": ["<root>.model.optim.config.hysteresis"]}], ["Attr(name='adam_beta1')", {"type": "leaf", "value": 0.9, "paths": ["<root>.model.optim.config.adam_beta1"]}], ["Attr(name='adam_beta2')", {"type": "leaf", "value": 0.999, "paths": ["<root>.model.optim.config.adam_beta2"]}], ["Attr(name='adam_eps')", {"type": "leaf", "value": 1e-08, "paths": ["<root>.model.optim.config.adam_eps"]}], ["Attr(name='sgd_momentum')", {"type": "leaf", "value": 0.9, "paths": ["<root>.model.optim.config.sgd_momentum"]}], ["Attr(name='use_distributed_optimizer')", {"type": "leaf", "value": true, "paths": ["<root>.model.optim.config.use_distributed_optimizer"]}], ["Attr(name='overlap_param_gather_with_optimizer_step')", {"type": "leaf", "value": false, "paths": ["<root>.model.optim.config.overlap_param_gather_with_optimizer_step"]}], ["Attr(name='optimizer_cpu_offload')", {"type": "leaf", "value": false, "paths": ["<root>.model.optim.config.optimizer_cpu_offload"]}], ["Attr(name='optimizer_offload_fraction')", {"type": "leaf", "value": 0.0, "paths": ["<root>.model.optim.config.optimizer_offload_fraction"]}], ["Attr(name='use_torch_optimizer_for_cpu_offload')", {"type": "leaf", "value": false, "paths": ["<root>.model.optim.config.use_torch_optimizer_for_cpu_offload"]}], ["Attr(name='overlap_cpu_optimizer_d2h_h2d')", {"type": "leaf", "value": false, "paths": ["<root>.model.optim.config.overlap_cpu_optimizer_d2h_h2d"]}], ["Attr(name='pin_cpu_grads')", {"type": "leaf", "value": true, "paths": ["<root>.model.optim.config.pin_cpu_grads"]}], ["Attr(name='pin_cpu_params')", {"type": "leaf", "value": true, "paths": ["<root>.model.optim.config.pin_cpu_params"]}], ["Attr(name='clip_grad')", {"type": "leaf", "value": 1.0, "paths": ["<root>.model.optim.config.clip_grad"]}], ["Attr(name='log_num_zeros_in_grad')", {"type": "leaf", "value": false, "paths": ["<root>.model.optim.config.log_num_zeros_in_grad"]}], ["Attr(name='barrier_with_L1_time')", {"type": "leaf", "value": false, "paths": ["<root>.model.optim.config.barrier_with_L1_time"]}], ["Attr(name='timers')", {"type": "leaf", "value": null, "paths": ["<root>.model.optim.config.timers"]}], ["Attr(name='config_logger_dir')", {"type": "leaf", "value": "", "paths": ["<root>.model.optim.config.config_logger_dir"]}]], "metadata": {"type": "ref", "key": "buildable_traverser_metadata_2"}, "paths": ["<root>.model.optim.config"]}, "tuple_3": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [["Index(index=0)", "config"]], "metadata": null}, "dict_5": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}}, "dict_6": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}}, "buildable_traverser_metadata_3": {"type": {"type": "pyref", "module": "fiddle._src.config", "name": "BuildableTraverserMetadata"}, "items": [["Attr(name='fn_or_cls')", {"type": "pyref", "module": "nemo.lightning.pytorch.optim.megatron", "name": "MegatronOptimizerModule"}], ["Attr(name='argument_names')", {"type": "ref", "key": "tuple_3"}], ["Attr(name='argument_tags')", {"type": "ref", "key": "dict_5"}], ["Attr(name='argument_history')", {"type": "ref", "key": "dict_6"}]], "metadata": {"type": "pyref", "module": "fiddle._src.config", "name": "BuildableTraverserMetadata"}}, "megatron_optimizer_module_1": {"type": {"type": "pyref", "module": "fiddle._src.config", "name": "Config"}, "items": [["Attr(name='config')", {"type": "ref", "key": "optimizer_config_1"}]], "metadata": {"type": "ref", "key": "buildable_traverser_metadata_3"}, "paths": ["<root>.model.optim"]}, "tuple_4": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [["Index(index=0)", "pretrained_model_name"], ["Index(index=1)", "trust_remote_code"]], "metadata": null}, "dict_7": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}}, "dict_8": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}}, "buildable_traverser_metadata_4": {"type": {"type": "pyref", "module": "fiddle._src.config", "name": "BuildableTraverserMetadata"}, "items": [["Attr(name='fn_or_cls')", {"type": "pyref", "module": "nemo.collections.common.tokenizers.huggingface.auto_tokenizer", "name": "AutoTokenizer"}], ["Attr(name='argument_names')", {"type": "ref", "key": "tuple_4"}], ["Attr(name='argument_tags')", {"type": "ref", "key": "dict_7"}], ["Attr(name='argument_history')", {"type": "ref", "key": "dict_8"}]], "metadata": {"type": "pyref", "module": "fiddle._src.config", "name": "BuildableTraverserMetadata"}}, "auto_tokenizer_1": {"type": {"type": "pyref", "module": "fiddle._src.config", "name": "Config"}, "items": [["Attr(name='pretrained_model_name')", {"type": "leaf", "value": "nemo_tokenizer", "paths": ["<root>.model.tokenizer.pretrained_model_name"]}], ["Attr(name='trust_remote_code')", {"type": "leaf", "value": true, "paths": ["<root>.model.tokenizer.trust_remote_code"]}]], "metadata": {"type": "ref", "key": "buildable_traverser_metadata_4"}, "paths": ["<root>.model.tokenizer"]}, "tuple_5": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [["Index(index=0)", "config"], ["Index(index=1)", "optim"], ["Index(index=2)", "tokenizer"]], "metadata": null}, "dict_9": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}}, "dict_10": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}}, "buildable_traverser_metadata_5": {"type": {"type": "pyref", "module": "fiddle._src.config", "name": "BuildableTraverserMetadata"}, "items": [["Attr(name='fn_or_cls')", {"type": "pyref", "module": "nemo.collections.llm.gpt.model.ssm", "name": "MambaModel"}], ["Attr(name='argument_names')", {"type": "ref", "key": "tuple_5"}], ["Attr(name='argument_tags')", {"type": "ref", "key": "dict_9"}], ["Attr(name='argument_history')", {"type": "ref", "key": "dict_10"}]], "metadata": {"type": "pyref", "module": "fiddle._src.config", "name": "BuildableTraverserMetadata"}}, "mamba_model_1": {"type": {"type": "pyref", "module": "fiddle._src.config", "name": "Config"}, "items": [["Attr(name='config')", {"type": "ref", "key": "nemotron_hconfig47_b_1"}], ["Attr(name='optim')", {"type": "ref", "key": "megatron_optimizer_module_1"}], ["Attr(name='tokenizer')", {"type": "ref", "key": "auto_tokenizer_1"}]], "metadata": {"type": "ref", "key": "buildable_traverser_metadata_5"}, "paths": ["<root>.model"]}, "tuple_6": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [["Index(index=0)", "always_save_context"]], "metadata": null}, "dict_11": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [["Key(key='always_save_context')", {"type": "leaf", "value": true, "paths": ["<root>.trainer.strategy.kwargs['always_save_context']"]}]], "metadata": {"type": "ref", "key": "tuple_6"}, "paths": ["<root>.trainer.strategy.kwargs"]}, "tuple_7": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [["Index(index=0)", "ckpt_save_optimizer"], ["Index(index=1)", "kwargs"]], "metadata": null}, "dict_12": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}}, "dict_13": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}}, "buildable_traverser_metadata_6": {"type": {"type": "pyref", "module": "fiddle._src.config", "name": "BuildableTraverserMetadata"}, "items": [["Attr(name='fn_or_cls')", {"type": "pyref", "module": "nemo.lightning.pytorch.strategies.megatron_strategy", "name": "MegatronStrategy"}], ["Attr(name='argument_names')", {"type": "ref", "key": "tuple_7"}], ["Attr(name='argument_tags')", {"type": "ref", "key": "dict_12"}], ["Attr(name='argument_history')", {"type": "ref", "key": "dict_13"}]], "metadata": {"type": "pyref", "module": "fiddle._src.config", "name": "BuildableTraverserMetadata"}}, "megatron_strategy_1": {"type": {"type": "pyref", "module": "fiddle._src.config", "name": "Config"}, "items": [["Attr(name='ckpt_save_optimizer')", {"type": "leaf", "value": false, "paths": ["<root>.trainer.strategy.ckpt_save_optimizer"]}], ["Attr(name='kwargs')", {"type": "ref", "key": "dict_11"}]], "metadata": {"type": "ref", "key": "buildable_traverser_metadata_6"}, "paths": ["<root>.trainer.strategy"]}, "tuple_8": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [["Index(index=0)", "accelerator"], ["Index(index=1)", "strategy"], ["Index(index=2)", "devices"]], "metadata": null}, "dict_14": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}}, "dict_15": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}}, "buildable_traverser_metadata_7": {"type": {"type": "pyref", "module": "fiddle._src.config", "name": "BuildableTraverserMetadata"}, "items": [["Attr(name='fn_or_cls')", {"type": "pyref", "module": "nemo.lightning.pytorch.trainer", "name": "Trainer"}], ["Attr(name='argument_names')", {"type": "ref", "key": "tuple_8"}], ["Attr(name='argument_tags')", {"type": "ref", "key": "dict_14"}], ["Attr(name='argument_history')", {"type": "ref", "key": "dict_15"}]], "metadata": {"type": "pyref", "module": "fiddle._src.config", "name": "BuildableTraverserMetadata"}}, "trainer_1": {"type": {"type": "pyref", "module": "fiddle._src.config", "name": "Config"}, "items": [["Attr(name='accelerator')", {"type": "leaf", "value": "cpu", "paths": ["<root>.trainer.accelerator"]}], ["Attr(name='strategy')", {"type": "ref", "key": "megatron_strategy_1"}], ["Attr(name='devices')", {"type": "leaf", "value": 1, "paths": ["<root>.trainer.devices"]}]], "metadata": {"type": "ref", "key": "buildable_traverser_metadata_7"}, "paths": ["<root>.trainer"]}, "dict_16": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}, "paths": ["<root>.extra"]}, "tuple_9": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [["Index(index=0)", "model"], ["Index(index=1)", "trainer"], ["Index(index=2)", "extra"]], "metadata": null}, "dict_17": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}}, "dict_18": {"type": {"type": "pyref", "module": "builtins", "name": "dict"}, "items": [], "metadata": {"type": {"type": "pyref", "module": "builtins", "name": "tuple"}, "items": [], "metadata": null}}, "buildable_traverser_metadata_8": {"type": {"type": "pyref", "module": "fiddle._src.config", "name": "BuildableTraverserMetadata"}, "items": [["Attr(name='fn_or_cls')", {"type": "pyref", "module": "nemo.lightning.io.pl", "name": "TrainerContext"}], ["Attr(name='argument_names')", {"type": "ref", "key": "tuple_9"}], ["Attr(name='argument_tags')", {"type": "ref", "key": "dict_17"}], ["Attr(name='argument_history')", {"type": "ref", "key": "dict_18"}]], "metadata": {"type": "pyref", "module": "fiddle._src.config", "name": "BuildableTraverserMetadata"}}, "trainer_context_1": {"type": {"type": "pyref", "module": "fiddle._src.config", "name": "Config"}, "items": [["Attr(name='model')", {"type": "ref", "key": "mamba_model_1"}], ["Attr(name='trainer')", {"type": "ref", "key": "trainer_1"}], ["Attr(name='extra')", {"type": "ref", "key": "dict_16"}]], "metadata": {"type": "ref", "key": "buildable_traverser_metadata_8"}, "paths": ["<root>"]}}, "refcounts": {"tuple_1": 1, "dict_1": 1, "dict_2": 1, "buildable_traverser_metadata_1": 1, "nemotron_hconfig47_b_1": 1, "tuple_2": 1, "dict_3": 1, "dict_4": 1, "buildable_traverser_metadata_2": 1, "optimizer_config_1": 1, "tuple_3": 1, "dict_5": 1, "dict_6": 1, "buildable_traverser_metadata_3": 1, "megatron_optimizer_module_1": 1, "tuple_4": 1, "dict_7": 1, "dict_8": 1, "buildable_traverser_metadata_4": 1, "auto_tokenizer_1": 1, "tuple_5": 1, "dict_9": 1, "dict_10": 1, "buildable_traverser_metadata_5": 1, "mamba_model_1": 1, "tuple_6": 1, "dict_11": 1, "tuple_7": 1, "dict_12": 1, "dict_13": 1, "buildable_traverser_metadata_6": 1, "megatron_strategy_1": 1, "tuple_8": 1, "dict_14": 1, "dict_15": 1, "buildable_traverser_metadata_7": 1, "trainer_1": 1, "dict_16": 1, "tuple_9": 1, "dict_17": 1, "dict_18": 1, "buildable_traverser_metadata_8": 1, "trainer_context_1": 1}, "version": "0.0.1"}
nemo/context/model.yaml ADDED
@@ -0,0 +1,277 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _target_: nemo.collections.llm.gpt.model.ssm.MambaModel
2
+ config:
3
+ _cpu_offloading_context: null
4
+ _target_: nemo.collections.llm.gpt.model.ssm.NemotronHConfig47B
5
+ account_for_embedding_in_pipeline_split: false
6
+ account_for_loss_in_pipeline_split: false
7
+ activation_func:
8
+ _call_: false
9
+ _target_: nemo.collections.llm.gpt.model.ssm.NemotronHConfigBase.<lambda>
10
+ activation_func_fp8_input_store: false
11
+ add_bias_linear: false
12
+ add_qkv_bias: false
13
+ apply_query_key_layer_scaling: false
14
+ apply_residual_connection_post_layernorm: false
15
+ apply_rope_fusion: true
16
+ async_tensor_model_parallel_allreduce: false
17
+ attention_backend:
18
+ _call_: true
19
+ _target_: megatron.core.transformer.enums.AttnBackend
20
+ attention_dropout: 0.0
21
+ attention_softmax_in_fp32: false
22
+ autocast_dtype: null
23
+ barrier_with_L1_time: true
24
+ batch_p2p_comm: true
25
+ batch_p2p_sync: true
26
+ bf16: true
27
+ bias_activation_fusion: false
28
+ bias_dropout_fusion: true
29
+ calculate_per_token_loss: false
30
+ clone_scatter_output_in_embedding: true
31
+ config_logger_dir: ''
32
+ context_parallel_size: 1
33
+ cp_comm_type: null
34
+ cpu_offloading: false
35
+ cpu_offloading_activations: true
36
+ cpu_offloading_num_layers: 0
37
+ cpu_offloading_weights: true
38
+ cross_entropy_fusion_impl: native
39
+ cross_entropy_loss_fusion: true
40
+ cuda_graph_retain_backward_graph: false
41
+ cuda_graph_scope: full
42
+ cuda_graph_use_single_mempool: false
43
+ cuda_graph_warmup_steps: 3
44
+ data_step_fn:
45
+ _call_: false
46
+ _target_: nemo.collections.llm.gpt.model.base.gpt_data_step
47
+ deallocate_pipeline_outputs: true
48
+ defer_embedding_wgrad_compute: false
49
+ deterministic_mode: false
50
+ disable_parameter_transpose_cache: false
51
+ distribute_saved_activations: null
52
+ enable_autocast: false
53
+ enable_cuda_graph: false
54
+ expert_model_parallel_size: 1
55
+ expert_tensor_parallel_size: null
56
+ external_cuda_graph: false
57
+ ffn_hidden_size: 30720
58
+ finalize_model_grads_func: null
59
+ first_last_layers_bf16: true
60
+ flash_decode: false
61
+ forward_step_fn:
62
+ _call_: false
63
+ _target_: nemo.collections.llm.gpt.model.ssm.ssm_forward_step
64
+ fp16: false
65
+ fp16_lm_cross_entropy: false
66
+ fp32_residual_connection: false
67
+ fp8: null
68
+ fp8_amax_compute_algo: most_recent
69
+ fp8_amax_history_len: 1
70
+ fp8_dot_product_attention: false
71
+ fp8_interval: 1
72
+ fp8_margin: 0
73
+ fp8_multi_head_attention: false
74
+ fp8_recipe: delayed
75
+ fp8_wgrad: true
76
+ gated_linear_unit: false
77
+ get_attention_mask_from_fusion: false
78
+ grad_scale_func: null
79
+ grad_sync_func: null
80
+ gradient_accumulation_fusion: false
81
+ hidden_dropout: 0.0
82
+ hidden_size: 8192
83
+ hierarchical_context_parallel_sizes: null
84
+ hybrid_attention_ratio: 0.0
85
+ hybrid_mlp_ratio: 0.0
86
+ hybrid_override_pattern: M-M-M-M-M-M-M-M-M*-M-M-M-M-M-M-M-M-M-M*-M-M-M-M-M*-M-M-M-M-M*-M-M-M-M-M-M-M---MM---M-M*-M-M-M-M-M-
87
+ inference_rng_tracker: false
88
+ init_method: null
89
+ init_method_std: 0.02
90
+ init_model_with_meta_device: false
91
+ is_hybrid_model: true
92
+ kv_channels: null
93
+ layernorm_epsilon: 1.0e-05
94
+ layernorm_zero_centered_gamma: false
95
+ make_vocab_size_divisible_by: 128
96
+ mamba_head_dim: 64
97
+ mamba_nheads: 256
98
+ mamba_num_groups: 8
99
+ mamba_state_dim: 256
100
+ mapping_type: nvidia-hybrid-nemotronh
101
+ masked_softmax_fusion: true
102
+ memory_efficient_layer_norm: false
103
+ microbatch_group_size_per_vp_stage: 1
104
+ moe_aux_loss_coeff: 0
105
+ moe_enable_deepep: false
106
+ moe_expert_capacity_factor: null
107
+ moe_extended_tp: false
108
+ moe_ffn_hidden_size: null
109
+ moe_grouped_gemm: false
110
+ moe_input_jitter_eps: null
111
+ moe_layer_freq: 1
112
+ moe_layer_recompute: false
113
+ moe_pad_expert_input_to_capacity: false
114
+ moe_per_layer_logging: false
115
+ moe_permute_fusion: false
116
+ moe_router_bias_update_rate: 0.001
117
+ moe_router_dtype: null
118
+ moe_router_enable_expert_bias: false
119
+ moe_router_group_topk: null
120
+ moe_router_load_balancing_type: aux_loss
121
+ moe_router_num_groups: null
122
+ moe_router_pre_softmax: false
123
+ moe_router_score_function: softmax
124
+ moe_router_topk: 2
125
+ moe_router_topk_limited_devices: null
126
+ moe_router_topk_scaling_factor: null
127
+ moe_shared_expert_intermediate_size: null
128
+ moe_shared_expert_overlap: false
129
+ moe_token_dispatcher_type: allgather
130
+ moe_token_drop_policy: probs
131
+ moe_token_dropping: false
132
+ moe_use_legacy_grouped_gemm: false
133
+ moe_z_loss_coeff: null
134
+ mtp_loss_scaling_factor: null
135
+ mtp_num_layers: null
136
+ multi_latent_attention: false
137
+ no_sync_func: null
138
+ normalization: RMSNorm
139
+ num_attention_heads: 64
140
+ num_layers: 98
141
+ num_layers_at_end_in_bf16: 1
142
+ num_layers_at_start_in_bf16: 1
143
+ num_layers_in_first_pipeline_stage: null
144
+ num_layers_in_last_pipeline_stage: null
145
+ num_microbatches_with_partial_activation_checkpoints: null
146
+ num_moe_experts: null
147
+ num_query_groups: 8
148
+ output_layer_init_method: null
149
+ overlap_p2p_comm: false
150
+ overlap_p2p_comm_warmup_flush: false
151
+ parallel_output: true
152
+ param_sync_func: null
153
+ params_dtype:
154
+ _call_: false
155
+ _target_: torch.bfloat16
156
+ perform_initialization: true
157
+ persist_layer_norm: true
158
+ pipeline_dtype: null
159
+ pipeline_model_parallel_comm_backend: null
160
+ pipeline_model_parallel_size: 1
161
+ pipeline_model_parallel_split_rank: null
162
+ position_embedding_type: none
163
+ post_process: true
164
+ pre_process: true
165
+ qk_layernorm: false
166
+ recompute_granularity: null
167
+ recompute_method: null
168
+ recompute_num_layers: null
169
+ rotary_base: 10000
170
+ rotary_interleaved: false
171
+ rotary_percent: 1.0
172
+ seq_len_interpolation_factor: null
173
+ seq_length: 8192
174
+ sequence_parallel: false
175
+ share_embeddings_and_output_weights: false
176
+ softmax_scale: null
177
+ tensor_model_parallel_size: 1
178
+ test_mode: false
179
+ timers: null
180
+ tokenizer_library: tiktoken
181
+ tokenizer_model_path: null
182
+ tokenizer_name: TiktokenTokenizer
183
+ tp_comm_atomic_ag: false
184
+ tp_comm_atomic_rs: false
185
+ tp_comm_bootstrap_backend: nccl
186
+ tp_comm_bulk_dgrad: true
187
+ tp_comm_bulk_wgrad: true
188
+ tp_comm_overlap: false
189
+ tp_comm_overlap_ag: true
190
+ tp_comm_overlap_disable_fc1: false
191
+ tp_comm_overlap_disable_qkv: false
192
+ tp_comm_overlap_rs: true
193
+ tp_comm_overlap_rs_dgrad: false
194
+ tp_comm_split_ag: true
195
+ tp_comm_split_rs: true
196
+ tp_only_amax_red: false
197
+ use_cpu_initialization: false
198
+ use_custom_fsdp: false
199
+ use_ring_exchange_p2p: false
200
+ use_te_rng_tracker: false
201
+ variable_seq_lengths: false
202
+ virtual_pipeline_model_parallel_size: null
203
+ vocab_file: null
204
+ vocab_size: 131072
205
+ wgrad_deferral_limit: 0
206
+ window_size: null
207
+ model_transform: null
208
+ optim:
209
+ _target_: nemo.lightning.pytorch.optim.megatron.MegatronOptimizerModule
210
+ config:
211
+ _target_: megatron.core.optimizer.optimizer_config.OptimizerConfig
212
+ adam_beta1: 0.9
213
+ adam_beta2: 0.999
214
+ adam_eps: 1.0e-08
215
+ barrier_with_L1_time: false
216
+ bf16: false
217
+ clip_grad: 1.0
218
+ config_logger_dir: ''
219
+ decoupled_lr: null
220
+ decoupled_min_lr: null
221
+ exp_avg_dtype:
222
+ _call_: false
223
+ _target_: torch.float32
224
+ exp_avg_sq_dtype:
225
+ _call_: false
226
+ _target_: torch.float32
227
+ fp16: false
228
+ hysteresis: 2
229
+ initial_loss_scale: 4294967296
230
+ log_num_zeros_in_grad: false
231
+ loss_scale: null
232
+ loss_scale_window: 1000
233
+ lr: 0.0001
234
+ main_grads_dtype:
235
+ _call_: false
236
+ _target_: torch.float32
237
+ main_params_dtype:
238
+ _call_: false
239
+ _target_: torch.float32
240
+ min_loss_scale: 1.0
241
+ min_lr: null
242
+ optimizer: adam
243
+ optimizer_cpu_offload: false
244
+ optimizer_offload_fraction: 0.0
245
+ overlap_cpu_optimizer_d2h_h2d: false
246
+ overlap_param_gather_with_optimizer_step: false
247
+ params_dtype:
248
+ _call_: false
249
+ _target_: torch.float32
250
+ pin_cpu_grads: true
251
+ pin_cpu_params: true
252
+ sgd_momentum: 0.9
253
+ timers: null
254
+ use_distributed_optimizer: true
255
+ use_precision_aware_optimizer: false
256
+ use_torch_optimizer_for_cpu_offload: false
257
+ weight_decay: 0.01
258
+ lr_mult: 1.0
259
+ lr_scheduler: null
260
+ no_weight_decay_cond: null
261
+ scale_lr_cond: null
262
+ tokenizer:
263
+ _target_: nemo.collections.common.tokenizers.huggingface.auto_tokenizer.AutoTokenizer
264
+ additional_special_tokens: []
265
+ bos_token: null
266
+ cls_token: null
267
+ eos_token: null
268
+ include_special_tokens: false
269
+ mask_token: null
270
+ merges_file: null
271
+ pad_token: null
272
+ pretrained_model_name: nemo_tokenizer
273
+ sep_token: null
274
+ trust_remote_code: true
275
+ unk_token: null
276
+ use_fast: false
277
+ vocab_file: null
nemo/context/nemo_tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
nemo/context/nemo_tokenizer/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3277c00fe5fb3963b3cb7c07b7f183722d2af4d775a4aea7cfb3684d7cccbc2f
3
+ size 17078330
nemo/context/nemo_tokenizer/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
nemo/weights/.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c351cb6da43f06c2cbea80f6b64247dacaee17e4a7b0baee8c664709307b0e7
3
+ size 425519
nemo/weights/__0_0.distcp ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87418abcf047f4dd5aa2924d8de88ccff8225e555249043a8916206ceebb17da
3
+ size 46792303844
nemo/weights/__0_1.distcp ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9270a49378b0de2284768e37136f913abffd74caa9cf0684b0455e22d89704f7
3
+ size 46792361482
nemo/weights/common.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf2d72b4f4fef17ea0a99868f378f27ac9989452f3ffdd0691821ae1a3b9f285
3
+ size 3099
nemo/weights/metadata.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"sharded_backend": "torch_dist", "sharded_backend_version": 1, "common_backend": "torch", "common_backend_version": 1}