export CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 model_name_or_path="meta-llama/Llama-2-7b-chat-hf" # model_name_or_path="Yukang/LongAlpaca-7B-16k" # model_name_or_path="lmsys/longchat-7b-v1.5-32k" # model_name_or_path="syzymon/long_llama_code_7b_instruct" # super_tokenizer_name_or_path="/home/baaiks/ninglu/code/PluginTransformer/data/outputs/90k_0104+8/super_tokenizer" super_tokenizer_name_or_path="/home/baaiks/ninglu/code/PluginTransformer/data/outputs/90k_0104+12/super_tokenizer" # super_tokenizer_name_or_path="/home/baaiks/ninglu/code/PluginTransformer/data/outputs/90k_0111+8/super_tokenizer" # output_dir="data/results/lm/test_1" # mkdir -p ${output_dir} # python -m main.eval_lm \ # --model_name_or_path ${model_name_or_path} \ # --super_tokenizer_name_or_path ${super_tokenizer_name_or_path} \ # --super_tokenizer_num_hidden_layers 8 \ # --device_map "auto" \ # --dataset_list "pg19" \ # --output_dir ${output_dir} \ # --max_length 8192 \ # --target_length 512 \ # --compression_enable true \ # --compression_ratio 16 \ # | tee "${output_dir}/eval_lm.log" max_length_lst="4096 8192 16384 32768" compression_ratio_lst="16" for max_length in ${max_length_lst}; do for compression_ratio in ${compression_ratio_lst}; do output_dir="data/results/lm/pg19/llama-12-${compression_ratio}-${max_length}" mkdir -p ${output_dir} python -m main.eval_lm \ --model_name_or_path ${model_name_or_path} \ --super_tokenizer_name_or_path ${super_tokenizer_name_or_path} \ --super_tokenizer_num_hidden_layers 12 \ --device_map "auto" \ --dataset_list "pg19" \ --output_dir ${output_dir} \ --max_length ${max_length} \ --target_length 512 \ --compression_enable true \ --compression_ratio ${compression_ratio} \ --down_scale_method "uniform" \ | tee "${output_dir}/eval_lm.log" done done