# conv-s2s
CUDA_VISIBLE_DEVICES=0,1 python test.py --task translation_bpe data-bin/writingPrompts-prompt2story_bpe500_fix -a fconv_self_att_wp \
    --lr 0.25 --clip-norm 0.1 --max-tokens 4000 --lr-scheduler reduce_lr_on_plateau \
    --decoder-attention True --encoder-attention False \
    --criterion story_cross_entropy_without_event --weight-decay .0000001 --label-smoothing 0 \
    --source-lang prompt --target-lang story \
    --gated-attention True --self-attention True --project-input True \
    --pretrained False --tensorboard-logdir ../fire_data/running/conv-s2s-d0.3_bpe500 \
    --log-interval  1000 --log-format simple --distributed-world-size 2 --dropout 0.3 \
    --skip-invalid-size-inputs-valid-test \
    --save-dir ../fire_data/running/conv-s2s-d0.3_bpe500 \
    --log-file logging.test --valid-subset test --restore-file checkpoint_best.pt

# conv-s2s fusion
CUDA_VISIBLE_DEVICES=5 python test.py --task translation_bpe data-bin/writingPrompts-prompt2story_bpe500_fix -a fconv_self_att_wp_fusion \
    --lr 0.25 --clip-norm 0.1 --max-tokens 4000 --lr-scheduler reduce_lr_on_plateau \
    --decoder-attention True --encoder-attention False \
    --criterion story_cross_entropy_without_event --weight-decay .0000001 --label-smoothing 0 \
    --source-lang prompt --target-lang story \
    --gated-attention True --self-attention True --project-input True \
    --tensorboard-logdir ../fire_data/running/conv-s2s-d0.3-fusion_bpe500 \
    --log-interval  1000 --log-format simple --distributed-world-size 1 --dropout 0.3 \
    --skip-invalid-size-inputs-valid-test \
    --save-dir ../fire_data/running/conv-s2s-d0.3-fusion_bpe500 --pretrained True --pretrained-checkpoint ../fire_data/running/conv-s2s-d0.3_bpe500/checkpoint_best.pt \
    --log-file logging.test --valid-subset test --restore-file checkpoint_best.pt

## event to story model
CUDA_VISIBLE_DEVICES=2,3 python test.py --task hierstory_bpe data-bin/writingPrompts-event2story_bpe500 \
    --save-dir ../fire_data/running/bilstm_transformer-lm_bpe500_format_pre --arch bilstm_h_transformer_lm_gpt_format \
    --source-lang event --target-lang new  --max-tokens 4000 \
    --optimizer adam --adam-betas '(0.9, 0.98)'  --lr 0.0005  --lr-scheduler inverse_sqrt \
    --warmup-updates 4000 --warmup-init-lr '1e-07' --min-lr '1e-09' \
    --criterion story_cross_entropy --weight-decay .0001 --label-smoothing 0 \
    --tensorboard-logdir ../fire_data/running/bilstm_transformer-lm_bpe500_format_pre \
    --log-interval  1000 --log-format simple --distributed-world-size 2 --dropout 0.3 \
    --skip-invalid-size-inputs-valid-test \
    --pretrained  --pretrained-checkpoint data-bin/writingPrompts-event2story_bpe500/gpt2model.pytorch \
    --log-file logging.test --valid-subset test --restore-file checkpoint_best.pt
CUDA_VISIBLE_DEVICES=2,3 python test.py --task hierstory_bpe data-bin/writingPrompts-srl2story_bpe500 \
    --save-dir ../fire_data/running/bilstm_transformer-lm_bpe500_format_pre_srl --arch bilstm_h_transformer_lm_gpt_format \
    --source-lang event --target-lang new  --max-tokens 4000 \
    --optimizer adam --adam-betas '(0.9, 0.98)'  --lr 0.0005  --lr-scheduler inverse_sqrt \
    --warmup-updates 4000 --warmup-init-lr '1e-07' --min-lr '1e-09' \
    --criterion story_cross_entropy --weight-decay .0001 --label-smoothing 0 \
    --tensorboard-logdir ../fire_data/running/bilstm_transformer-lm_bpe500_format_pre_srl \
    --log-interval  1000 --log-format simple --distributed-world-size 2 --dropout 0.3 \
    --skip-invalid-size-inputs-valid-test \
    --pretrained  --pretrained-checkpoint data-bin/writingPrompts-srl2story_bpe500/gpt2model.pytorch \
    --log-file logging.test --valid-subset test --restore-file checkpoint_best.pt

# story language model
CUDA_VISIBLE_DEVICES=0,1 python test.py --task language_modeling_bpe data-bin/writingPrompts-event2story_bpe500 \
  --save-dir ../fire_data/running/transformer-lm_bpe500_warmlr_format2 --arch transformer_lm_gpt_bpe_format \
  --max-tokens 4000 \
  --optimizer adam --adam-betas '(0.9, 0.98)'  --lr 0.0005  --lr-scheduler inverse_sqrt \
  --warmup-updates 4000 --warmup-init-lr '1e-07' --min-lr '1e-09' \
  --criterion story_cross_entropy_without_event --weight-decay .0001 --label-smoothing 0 \
  --tensorboard-logdir ../fire_data/running/transformer-lm_bpe500_warmlr_format2  \
  --log-interval  1000 --log-format simple --distributed-world-size 2 --dropout 0.3 \
  --skip-invalid-size-inputs-valid-test --tokens-per-sample 1024 --sample-break-mode eos \
  --log-file logging.test --valid-subset test --restore-file checkpoint_best.pt
CUDA_VISIBLE_DEVICES=0,1 python test.py --task language_modeling_bpe data-bin/writingPrompts-event2story_bpe500 \
  --save-dir ../fire_data/running/transformer-lm_bpe500_warmlr_format_pre2 --arch transformer_lm_gpt_bpe_format \
  --max-tokens 4000 \
  --optimizer adam --adam-betas '(0.9, 0.98)'  --lr 0.0005  --lr-scheduler inverse_sqrt \
  --warmup-updates 4000 --warmup-init-lr '1e-07' --min-lr '1e-09' \
  --criterion story_cross_entropy_without_event --weight-decay .0001 --label-smoothing 0 \
  --tensorboard-logdir ../fire_data/running/transformer-lm_bpe500_warmlr_format_pre2  \
  --log-interval  1000 --log-format simple --distributed-world-size 2 --dropout 0.3 \
  --skip-invalid-size-inputs-valid-test --tokens-per-sample 1024 --sample-break-mode eos \
  --pretrained  --pretrained-checkpoint data-bin/writingPrompts-event2story_bpe500/gpt2model.pytorch \
  --log-file logging.test --valid-subset test --restore-file checkpoint_best.pt
CUDA_VISIBLE_DEVICES=0,1 python test.py --task language_modeling_bpe data-bin/writingPrompts-srl2story_bpe500 \
  --save-dir ../fire_data/running/transformer-lm_bpe500_warmlr_format_pre2_srl --arch transformer_lm_gpt_bpe_format \
  --max-tokens 4000 \
  --optimizer adam --adam-betas '(0.9, 0.98)'  --lr 0.0005  --lr-scheduler inverse_sqrt \
  --warmup-updates 4000 --warmup-init-lr '1e-07' --min-lr '1e-09' \
  --criterion story_cross_entropy_without_event --weight-decay .0001 --label-smoothing 0 \
  --tensorboard-logdir ../fire_data/running/transformer-lm_bpe500_warmlr_format_pre2_srl  \
  --log-interval  1000 --log-format simple --distributed-world-size 2 --dropout 0.3 \
  --skip-invalid-size-inputs-valid-test --tokens-per-sample 1024 --sample-break-mode eos \
  --pretrained  --pretrained-checkpoint data-bin/writingPrompts-srl2story_bpe500/gpt2model.pytorch \
  --log-file logging.test --valid-subset test --restore-file checkpoint_best.pt

CUDA_VISIBLE_DEVICES=0,1,2,3 python test.py --user-dir coherence_story --task language_modeling_bpe data-bin/writingPrompts-event2story_bpe500 \
  --save-dir ../fire_data/running/transformer-lm_bpe500_warmlr_format_pre2_memory_gpt --arch transformer_lm_bpe_memory_gpt_format \
  --max-tokens 4000 \
  --optimizer adam --adam-betas '(0.9, 0.98)'  --lr 0.0005  --lr-scheduler inverse_sqrt \
  --warmup-updates 4000 --warmup-init-lr '1e-07' --min-lr '1e-09' \
  --criterion story_cross_entropy_without_event --weight-decay .0001 --label-smoothing 0 \
  --tensorboard-logdir ../fire_data/running/transformer-lm_bpe500_warmlr_format_pre2_memory_gpt  \
  --log-interval  1000 --log-format simple --distributed-world-size 4 --dropout 0.3 \
  --skip-invalid-size-inputs-valid-test --tokens-per-sample 1024 --sample-break-mode eos \
  --pretrained  --pretrained-checkpoint data-bin/writingPrompts-event2story_bpe500/gpt2model.pytorch \
  --log-file logging.test --valid-subset test --restore-file checkpoint_best.pt

CUDA_VISIBLE_DEVICES=0,1,2,3 python test.py --user-dir coherence_story --task language_modeling_bpe data-bin/writingPrompts-event2story_bpe500 \
  --save-dir ../fire_data/running/transformer-lm_bpe500_warmlr_format_pre2_memory_gate_gpt --arch transformer_lm_bpe_memory_gate_gpt_format \
  --max-tokens 4000 \
  --optimizer adam --adam-betas '(0.9, 0.98)'  --lr 0.0005  --lr-scheduler inverse_sqrt \
  --warmup-updates 4000 --warmup-init-lr '1e-07' --min-lr '1e-09' \
  --criterion story_cross_entropy_without_event --weight-decay .0001 --label-smoothing 0 \
  --tensorboard-logdir ../fire_data/running/transformer-lm_bpe500_warmlr_format_pre2_memory_gate_gpt  \
  --log-interval  1000 --log-format simple --distributed-world-size 4 --dropout 0.3 \
  --skip-invalid-size-inputs-valid-test --tokens-per-sample 1024 --sample-break-mode eos \
  --pretrained  --pretrained-checkpoint data-bin/writingPrompts-event2story_bpe500/gpt2model.pytorch \
  --log-file logging.test --valid-subset test --restore-file checkpoint_best.pt

# gan
CUDA_VISIBLE_DEVICES=4,5,6,7 python test_gan.py --user-dir coherence_story --task hierstory_bpe_gan data-bin/writingPrompts-srl2story_bpe500_fix \
  --save-dir ../out/bilstm_translm_coherence_gan --arch fairseqgan_gpt_format \
  --source-lang event --target-lang new  --max-tokens 1024 \
  --optimizer adam --adam-betas '(0.9, 0.98)'  --lr 0.0005  --lr-scheduler inverse_sqrt \
  --warmup-updates 4000 --warmup-init-lr '1e-07' --min-lr '1e-09' \
  --criterion story_adv --weight-decay .0001 --label-smoothing 0 \
  --tensorboard-logdir ../out/bilstm_translm_coherence_gan \
  --log-interval 100 --save-interval-updates 1000 --log-format simple --distributed-world-size 4 --dropout 0.3 \
  --skip-invalid-size-inputs-valid-test \
  --pretrained  --pretrained-checkpoint ../out/bilstm_transformer-lm_bpe500_format_pre_srl_hier_pos_fix/checkpoint_best.pt \
  --pretrained-checkpoint-dis ../fire_data/running/transformer-lm_bpe500_warmlr_format_pre2/checkpoint_best.pt