Skip to content

Commit ccb580d

Browse files
committed
fixes
Signed-off-by: adil-a <adil.asif2000@hotmail.com>
1 parent 7419674 commit ccb580d

4 files changed

Lines changed: 7 additions & 6 deletions

File tree

examples/llm_finetune/nemotron/nemotron_nano_9b_squad.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -38,6 +38,7 @@ model:
3838
_target_: nemo_automodel.NeMoAutoModelForCausalLM.from_pretrained
3939
pretrained_model_name_or_path: nvidia/NVIDIA-Nemotron-Nano-9B-v2
4040
trust_remote_code: true
41+
force_hf: true
4142

4243
# torch.compile configuration
4344
compile:

examples/llm_finetune/nemotron/nemotron_nano_9b_squad_peft.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -38,6 +38,7 @@ model:
3838
_target_: nemo_automodel.NeMoAutoModelForCausalLM.from_pretrained
3939
pretrained_model_name_or_path: nvidia/NVIDIA-Nemotron-Nano-9B-v2
4040
trust_remote_code: true
41+
force_hf: true
4142

4243
# torch.compile configuration
4344
compile:

examples/llm_finetune/nemotron/nemotron_nano_v3_hellaswag.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14,9 +14,9 @@
1414

1515

1616
# To run this recipe, please use the following command. Replace --nproc-per-node=8 with the number of GPUs on your node:
17-
# torchrun --nproc-per-node=8 examples/llm_finetune/finetune.py --config examples/llm_finetune/nemotron/nemotron_nano_v3_squad.yaml
17+
# torchrun --nproc-per-node=8 examples/llm_finetune/finetune.py --config examples/llm_finetune/nemotron/nemotron_nano_v3_hellaswag.yaml
1818
# For multi-node, use the following:
19-
# torchrun --nproc-per-node=8 --nnodes=2 examples/llm_finetune/finetune.py --config examples/llm_finetune/nemotron/nemotron_nano_v3_squad.yaml
19+
# torchrun --nproc-per-node=8 --nnodes=2 examples/llm_finetune/finetune.py --config examples/llm_finetune/nemotron/nemotron_nano_v3_hellaswag.yaml
2020

2121

2222
step_scheduler:

examples/llm_finetune/nemotron/nemotron_nano_v3_hellaswag_peft.yaml

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -13,10 +13,8 @@
1313
# limitations under the License.
1414

1515

16-
# To run this recipe, please use the following command. Replace --nproc-per-node=8 with the number of GPUs on your node:
17-
# torchrun --nproc-per-node=8 examples/llm_finetune/finetune.py --config examples/llm_finetune/nemotron/nemotron_nano_v3_squad_peft.yaml
18-
# For multi-node, use the following:
19-
# torchrun --nproc-per-node=8 --nnodes=2 examples/llm_finetune/finetune.py --config examples/llm_finetune/nemotron/nemotron_nano_v3_squad_peft.yaml
16+
# To run this recipe, please use the following command:
17+
# python examples/llm_finetune/finetune.py --config examples/llm_finetune/nemotron/nemotron_nano_v3_hellaswag_peft.yaml
2018

2119

2220
step_scheduler:
@@ -39,6 +37,7 @@ model:
3937
_target_: nemo_automodel.NeMoAutoModelForCausalLM.from_pretrained
4038
pretrained_model_name_or_path: nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16
4139
trust_remote_code: true
40+
force_hf: true
4241

4342
# torch.compile configuration
4443
compile:

0 commit comments

Comments
 (0)