-
Notifications
You must be signed in to change notification settings - Fork 3
Expand file tree
/
Copy pathtrain_Exp_3_rawcode.sh
More file actions
executable file
·40 lines (37 loc) · 1.52 KB
/
train_Exp_3_rawcode.sh
File metadata and controls
executable file
·40 lines (37 loc) · 1.52 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
export PYTHONPATH=/PATH_PLACEHOLDER/NIPS_Material/
export VLLM_ATTENTION_BACKEND=XFORMERS
export HYDRA_FULL_ERROR=1
export CUDA_LAUNCH_BLOCKING=1
export CUDA_VISIBLE_DEVICES=0,1
base_model="Qwen/Qwen2.5-1.5B-Instruct"
project_name="compiler_autotuning_qwen"
sft_output_dir="./model_save/pure_rawcode_Exp3/"
sft_experiment_name="pure-llvmcode-$(basename $base_model)"
python3 -m examples.data_preprocess.compiler_autotuning_pure_llvmcode \
--llvm_ir_dir=examples/data_preprocess/llvmir_datasets/ \
--data_file=examples/data_preprocess/Experiment_3.csv \
--local_dir=./dataset/exp3_rawcode/
torchrun --standalone --nnodes=1 --nproc_per_node=2 \
-m verl.trainer.fsdp_sft_trainer \
data.train_files=./dataset/exp3_rawcode/train.parquet \
data.val_files=./dataset/exp3_rawcode/validation.parquet \
data.train_batch_size=32 \
data.micro_batch_size_per_gpu=8 \
data.prompt_key=extra_info \
data.response_key=extra_info \
optim.lr=1e-6 \
+data.prompt_dict_keys=['question'] \
+data.response_dict_keys=['answer'] \
data.micro_batch_size=8 \
data.max_length=15012 \
model.partial_pretrain=$base_model \
+model.torch_dtype=bfloat16 \
+model.attn_implementation=flash_attention_2 \
trainer.default_local_dir=$sft_output_dir \
trainer.project_name=$project_name \
trainer.experiment_name=$sft_experiment_name \
"trainer.logger=[console,wandb]" \
trainer.default_hdfs_dir=null \
trainer.total_epochs=1 \
ulysses_sequence_parallel_size=2 \
use_remove_padding=true