Skip to content
Snippets Groups Projects
Select Git revision
  • 12a1da7a103f1ed238d8f53172b86814b9e91237
  • master default protected
  • emoUS
  • add_default_vectorizer_and_pretrained_loading
  • clean_code
  • readme
  • issue127
  • generalized_action_dicts
  • ppo_num_dialogues
  • crossowoz_ddpt
  • issue_114
  • robust_masking_feature
  • scgpt_exp
  • e2e-soloist
  • convlab_exp
  • change_system_act_in_env
  • pre-training
  • nlg-scgpt
  • remapping_actions
  • soloist
20 results

run_tm3_user.sh

Blame
  • user avatar
    zqwerty authored
    12a1da7a
    History
    Code owners
    Assign users and groups as approvers for specific file changes. Learn more.
    run_tm3_user.sh 2.51 KiB
    n_gpus=1
    task_name="nlu"
    dataset_name="tm3"
    speaker="user"
    context_window_size=0
    data_dir="data/${task_name}/${dataset_name}/${speaker}/context_${context_window_size}"
    output_dir="output/${task_name}/${dataset_name}/${speaker}/context_${context_window_size}"
    cache_dir="../cache"
    logging_dir="${output_dir}/runs"
    train_file="${data_dir}/train.json"
    validation_file="${data_dir}/validation.json"
    test_file="${data_dir}/test.json"
    metric_name_or_path="nlu_metric.py"
    metric_for_best_model="overall_f1"
    source_prefix="${data_dir}/source_prefix.txt"
    source_column="context"
    target_column="dialogue_acts_seq"
    model_name_or_path="t5-small"
    per_device_train_batch_size=128
    per_device_eval_batch_size=64
    gradient_accumulation_steps=2
    lr=1e-3
    num_train_epochs=10
    
    python ../create_data.py --tasks ${task_name} --datasets ${dataset_name} --speaker ${speaker} --context_window_size ${context_window_size}
    
    python -m torch.distributed.launch \
        --nproc_per_node ${n_gpus} ../run_seq2seq.py \
        --task_name ${task_name} \
        --train_file ${train_file} \
        --source_column ${source_column} \
        --target_column ${target_column} \
        --source_prefix ${source_prefix} \
        --model_name_or_path ${model_name_or_path} \
        --do_train \
        --save_strategy epoch \
        --prediction_loss_only \
        --cache_dir ${cache_dir} \
        --output_dir ${output_dir} \
        --logging_dir ${logging_dir} \
        --overwrite_output_dir \
        --preprocessing_num_workers 4 \
        --per_device_train_batch_size ${per_device_train_batch_size} \
        --per_device_eval_batch_size ${per_device_eval_batch_size} \
        --gradient_accumulation_steps ${gradient_accumulation_steps} \
        --learning_rate ${lr} \
        --num_train_epochs ${num_train_epochs} \
        --debug underflow_overflow \
        --adafactor \
        --gradient_checkpointing
    
    python -m torch.distributed.launch \
        --nproc_per_node ${n_gpus} ../run_seq2seq.py \
        --task_name ${task_name} \
        --test_file ${test_file} \
        --source_column ${source_column} \
        --target_column ${target_column} \
        --source_prefix ${source_prefix} \
        --model_name_or_path ${output_dir} \
        --do_predict \
        --predict_with_generate \
        --metric_name_or_path ${metric_name_or_path} \
        --cache_dir ${cache_dir} \
        --output_dir ${output_dir} \
        --logging_dir ${logging_dir} \
        --overwrite_output_dir \
        --preprocessing_num_workers 4 \
        --per_device_eval_batch_size ${per_device_eval_batch_size} \
    
    python merge_predict_res.py -d ${dataset_name} -s ${speaker} -c ${context_window_size} -p ${output_dir}/generated_predictions.json