From d4674c29454af8a239fc53f526fe32874ce06ddb Mon Sep 17 00:00:00 2001 From: "liukangling (A)" Date: Fri, 7 Jun 2024 19:32:02 +0800 Subject: [PATCH] [built-in][Pytorch][MiniGPT4/CogVLM] log relocate --- .../foundation/CogVLM/cogvlm_utils/finetune_cogvlm_base_224.sh | 2 +- PyTorch/built-in/foundation/MiniGPT-4/test/finetune_gpt_1p.sh | 2 +- PyTorch/built-in/foundation/MiniGPT-4/test/pretrain_gpt_4p.sh | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/PyTorch/built-in/foundation/CogVLM/cogvlm_utils/finetune_cogvlm_base_224.sh b/PyTorch/built-in/foundation/CogVLM/cogvlm_utils/finetune_cogvlm_base_224.sh index c9f825c48c..40e9e8a063 100644 --- a/PyTorch/built-in/foundation/CogVLM/cogvlm_utils/finetune_cogvlm_base_224.sh +++ b/PyTorch/built-in/foundation/CogVLM/cogvlm_utils/finetune_cogvlm_base_224.sh @@ -45,7 +45,7 @@ gpt_options=" \ --split 1. \ --deepspeed_config test_config_bf16.json \ --skip-init \ - --seed 1234 + --seed 1234 > train_cogvlm.log 2>&1 & " diff --git a/PyTorch/built-in/foundation/MiniGPT-4/test/finetune_gpt_1p.sh b/PyTorch/built-in/foundation/MiniGPT-4/test/finetune_gpt_1p.sh index 2e1ade8092..bbd7e04462 100755 --- a/PyTorch/built-in/foundation/MiniGPT-4/test/finetune_gpt_1p.sh +++ b/PyTorch/built-in/foundation/MiniGPT-4/test/finetune_gpt_1p.sh @@ -13,4 +13,4 @@ WORLD_SIZE=$(($GPUS_PER_NODE*$NNODES)) DISTRIBUTED_ARGS="--nproc_per_node $GPUS_PER_NODE --nnodes $NNODES --node_rank $NODE_RANK --master_addr $MASTER_ADDR --master_port $MASTER_PORT" OPTIONS="run.max_epoch=2 run.iters_per_epoch=240 run.batch_size_train=10 run.batch_size_eval=10 " -torchrun $DISTRIBUTED_ARGS train.py --cfg-path train_configs/minigpt4_stage2_finetune.yaml --options ${OPTIONS} +torchrun $DISTRIBUTED_ARGS train.py --cfg-path train_configs/minigpt4_stage2_finetune.yaml --options ${OPTIONS} > minigpt4_finetune.log 2>&1 & diff --git a/PyTorch/built-in/foundation/MiniGPT-4/test/pretrain_gpt_4p.sh b/PyTorch/built-in/foundation/MiniGPT-4/test/pretrain_gpt_4p.sh index f53589b2be..10976fdcd2 100755 --- a/PyTorch/built-in/foundation/MiniGPT-4/test/pretrain_gpt_4p.sh +++ b/PyTorch/built-in/foundation/MiniGPT-4/test/pretrain_gpt_4p.sh @@ -14,4 +14,4 @@ WORLD_SIZE=$(($GPUS_PER_NODE*$NNODES)) DISTRIBUTED_ARGS="--nproc_per_node $GPUS_PER_NODE --nnodes $NNODES --node_rank $NODE_RANK --master_addr $MASTER_ADDR --master_port $MASTER_PORT" OPTIONS="run.max_epoch=4 run.iters_per_epoch=8000 run.batch_size_train=40 run.batch_size_eval=40 " -torchrun $DISTRIBUTED_ARGS train.py --cfg-path train_configs/minigpt4_stage1_pretrain.yaml --options ${OPTIONS} +torchrun $DISTRIBUTED_ARGS train.py --cfg-path train_configs/minigpt4_stage1_pretrain.yaml --options ${OPTIONS} > minigpt4_pretrain.log 2>&1 & -- Gitee