Skip to content

Commit 4a511a8

Browse files
authored
Merge pull request #49 from aws/revert-48-release-1.5.0
Revert "Release 1.5.0"
2 parents 25cf042 + 51b3533 commit 4a511a8

10 files changed

+0
-750
lines changed

README.md

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -116,10 +116,6 @@ Nova Pro | Model Distillation for Post-Training | - | - | 1 | ml
116116
| DeepSeek R1 Distill Qwen 2 | LoRA | 32b | 8192 | 2 | ml.p5.48xlarge | GPU H100 | [link](recipes_collection/recipes/fine-tuning/deepseek/hf_deepseek_r1_distilled_qwen_32b_seq8k_gpu_lora.yaml) | [link](launcher_scripts/deepseek/run_hf_deepseek_r1_qwen_32b_seq8k_gpu_lora.sh) |
117117
| DeepSeek R1 Distill Qwen 2 | SFT | 32b | 16384 | 6 | ml.p5.48xlarge | GPU H100 | [link](recipes_collection/recipes/fine-tuning/deepseek/hf_deepseek_r1_distilled_qwen_32b_seq16k_gpu_fine_tuning.yaml) | [link](launcher_scripts/deepseek/run_hf_deepseek_r1_qwen_32b_seq16k_gpu_fine_tuning.sh) |
118118
| DeepSeek R1 Distill Qwen 2 | LoRA | 32b | 16384 | 2 | ml.p5.48xlarge | GPU H100 | [link](recipes_collection/recipes/fine-tuning/deepseek/hf_deepseek_r1_distilled_qwen_32b_seq16k_gpu_lora.yaml) | [link](launcher_scripts/deepseek/run_hf_deepseek_r1_qwen_32b_seq16k_gpu_lora.sh) |
119-
| GPT OSS | LoRA | 20b | 8192 | 1 | ml.p5.48xlarge | GPU H100 | [link](recipes_collection/recipes/fine-tuning/gpt_oss/hf_gpt_oss_20b_seq8k_gpu_lora.yaml) | [link](launcher_scripts/gpt_oss/run_hf_gpt_oss_20b_seq8k_gpu_lora.sh) |
120-
| GPT OSS | LoRA | 20b | 4096 | 1 | ml.p5.48xlarge | GPU H100 | [link](recipes_collection/recipes/fine-tuning/gpt_oss/hf_gpt_oss_20b_seq4k_gpu_lora.yaml) | [link](launcher_scripts/gpt_oss/run_hf_gpt_oss_20b_seq4k_gpu_lora.sh) |
121-
| GPT OSS | LoRA | 120b | 8192 | 8 | ml.p5.48xlarge | GPU H100 | [link](recipes_collection/recipes/fine-tuning/gpt_oss/hf_gpt_oss_120b_seq8k_gpu_lora.yaml) | [link](launcher_scripts/gpt_oss/run_hf_gpt_oss_120b_seq8k_gpu_lora.sh) |
122-
| GPT OSS | LoRA | 120b | 4096 | 4 | ml.p5.48xlarge | GPU H100 | [link](recipes_collection/recipes/fine-tuning/gpt_oss/hf_gpt_oss_120b_seq4k_gpu_lora.yaml) | [link](launcher_scripts/gpt_oss/run_hf_gpt_oss_120b_seq4k_gpu_lora.sh) |
123119
| Llama 3.1 | QLoRA | 405b | 131072 | 2 | ml.p5.48xlarge | GPU H100 | [link](recipes_collection/recipes/fine-tuning/llama/hf_llama3_405b_seq128k_gpu_qlora.yaml) | [link](launcher_scripts/llama/run_hf_llama3_405b_seq128k_gpu_qlora.sh) |
124120
| Llama 3.1 | QLoRA | 405b | 32768 | 2 | ml.p5.48xlarge | GPU H100 | [link](recipes_collection/recipes/fine-tuning/llama/hf_llama3_405b_seq32k_gpu_qlora.yaml) | [link](launcher_scripts/llama/run_hf_llama3_405b_seq32k_gpu_qlora.sh) |
125121
| Llama 3.1 | LoRA | 405b | 16384 | 6 | ml.p5.48xlarge | GPU H100 | [link](recipes_collection/recipes/fine-tuning/llama/hf_llama3_405b_seq16k_gpu_lora.yaml) | [link](launcher_scripts/llama/run_hf_llama3_405b_seq16k_gpu_lora.sh) |

launcher/nemo/stages.py

Lines changed: 0 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -303,9 +303,6 @@ def _make_launch_docker_container_text(self):
303303
if OmegaConf.select(self.cfg, "recipes.model.model_type", default=None) == "llama_v4":
304304
transformers_upgrade_cmd = "pip install transformers==4.51.3"
305305
post_launch_commands.append(transformers_upgrade_cmd)
306-
if OmegaConf.select(self.cfg, "recipes.model.model_type", default=None) == "gpt_oss":
307-
transformers_upgrade_cmd = "pip install transformers==4.55.0"
308-
post_launch_commands.append(transformers_upgrade_cmd)
309306

310307
launch_docker_container_text.append(f' "{image}" sleep infinity')
311308
launch_docker_container_text.append("")
@@ -432,10 +429,6 @@ def _make_train_script_text(self, stage_cfg_path=None, port=41000) -> str:
432429
transformers_upgrade_cmd = "pip install transformers==4.51.3"
433430
script_text.append("")
434431
script_text.append(transformers_upgrade_cmd)
435-
if OmegaConf.select(self.cfg, "recipes.model.model_type", default=None) == "gpt_oss":
436-
transformers_upgrade_cmd = "pip install transformers==4.55.0"
437-
script_text.append("")
438-
script_text.append(transformers_upgrade_cmd)
439432

440433
script_text.append("")
441434
script_text.append(self._make_custom_call_string(stage_cfg_path))
@@ -775,9 +768,6 @@ def update_stage_specific_k8s_values(self, values_template):
775768
if OmegaConf.select(self.cfg, "recipes.model.model_type", default=None) == "llama_v4":
776769
transformers_upgrade_cmd = "pip install transformers==4.51.3"
777770
values_template.trainingConfig.pre_script.append(transformers_upgrade_cmd)
778-
if OmegaConf.select(self.cfg, "recipes.model.model_type", default=None) == "gpt_oss":
779-
transformers_upgrade_cmd = "pip install transformers==4.55.0"
780-
values_template.trainingConfig.pre_script.append(transformers_upgrade_cmd)
781771

782772
return values_template
783773

launcher_scripts/gpt_oss/run_hf_gpt_oss_120b_seq4k_gpu_lora.sh

Lines changed: 0 additions & 28 deletions
This file was deleted.

launcher_scripts/gpt_oss/run_hf_gpt_oss_120b_seq8k_gpu_lora.sh

Lines changed: 0 additions & 28 deletions
This file was deleted.

launcher_scripts/gpt_oss/run_hf_gpt_oss_20b_seq4k_gpu_lora.sh

Lines changed: 0 additions & 28 deletions
This file was deleted.

launcher_scripts/gpt_oss/run_hf_gpt_oss_20b_seq8k_gpu_lora.sh

Lines changed: 0 additions & 28 deletions
This file was deleted.

recipes_collection/recipes/fine-tuning/gpt_oss/hf_gpt_oss_120b_seq4k_gpu_lora.yaml

Lines changed: 0 additions & 156 deletions
This file was deleted.

0 commit comments

Comments
 (0)