From 2cf39279c30f8e504c4abb697a34f60e11414090 Mon Sep 17 00:00:00 2001 From: guozhihua Date: Thu, 25 Sep 2025 11:10:48 +0800 Subject: [PATCH] [pytorch][sh]change poc llama2_70b_128k sh to llama31_70b_128k --- .../pretrain_llama31_70b_128k_256die_A2_pack_ptd.sh} | 2 +- .../pretrain_llama31_70b_128k_256die_A3_pack_ptd.sh} | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) rename tests/poc/{llama2_70b/pretrain_llama2_70b_128k_256die_A2_pack_ptd.sh => llama31_70b/pretrain_llama31_70b_128k_256die_A2_pack_ptd.sh} (97%) rename tests/poc/{llama2_70b/pretrain_llama2_70b_128k_256die_A3_pack_ptd.sh => llama31_70b/pretrain_llama31_70b_128k_256die_A3_pack_ptd.sh} (97%) diff --git a/tests/poc/llama2_70b/pretrain_llama2_70b_128k_256die_A2_pack_ptd.sh b/tests/poc/llama31_70b/pretrain_llama31_70b_128k_256die_A2_pack_ptd.sh similarity index 97% rename from tests/poc/llama2_70b/pretrain_llama2_70b_128k_256die_A2_pack_ptd.sh rename to tests/poc/llama31_70b/pretrain_llama31_70b_128k_256die_A2_pack_ptd.sh index 5a61738c0..c2e2b43a4 100644 --- a/tests/poc/llama2_70b/pretrain_llama2_70b_128k_256die_A2_pack_ptd.sh +++ b/tests/poc/llama31_70b/pretrain_llama31_70b_128k_256die_A2_pack_ptd.sh @@ -124,4 +124,4 @@ torchrun $DISTRIBUTED_ARGS pretrain_gpt.py \ ${DATA_ARGS} \ ${CKPT_ARGS} \ ${OUTPUT_ARGS} \ - --distributed-backend nccl | tee logs/pretrain_llama2_70b_128k_mcore_256die_A2_pack.log + --distributed-backend nccl | tee logs/pretrain_llama31_70b_128k_mcore_256die_A2_pack.log diff --git a/tests/poc/llama2_70b/pretrain_llama2_70b_128k_256die_A3_pack_ptd.sh b/tests/poc/llama31_70b/pretrain_llama31_70b_128k_256die_A3_pack_ptd.sh similarity index 97% rename from tests/poc/llama2_70b/pretrain_llama2_70b_128k_256die_A3_pack_ptd.sh rename to tests/poc/llama31_70b/pretrain_llama31_70b_128k_256die_A3_pack_ptd.sh index 706feb3d8..6ba29a877 100644 --- a/tests/poc/llama2_70b/pretrain_llama2_70b_128k_256die_A3_pack_ptd.sh +++ b/tests/poc/llama31_70b/pretrain_llama31_70b_128k_256die_A3_pack_ptd.sh @@ -131,4 +131,4 @@ torchrun $DISTRIBUTED_ARGS pretrain_gpt.py \ ${DATA_ARGS} \ ${CKPT_ARGS} \ ${OUTPUT_ARGS} \ - --distributed-backend nccl | tee logs/pretrain_llama2_70b_128k_mcore_256die_A3_pack.log + --distributed-backend nccl | tee logs/pretrain_llama31_70b_128k_mcore_256die_A3_pack.log -- Gitee