Add a script for evaluating all models & fix errors in some configuration files (#3101)

* Add a script for evaluating all models & fix errors in some configuration files

* modify log dir name
pull/3113/head
Tongkai 2024-03-05 11:14:24 +08:00 committed by GitHub
parent b250aec615
commit 8a53cdd6cc
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
4 changed files with 114 additions and 11 deletions

View File

@ -1,5 +1,5 @@
repos:
- repo: https://github.com/PaddlePaddle/mirrors-yapf.git
- repo: https://github.com/cuicheng01/mirrors-yapf.git
rev: 0d79c0c469bab64f7229c9aca2b1186ef47f0e37
hooks:
- id: yapf

View File

@ -1,5 +1,5 @@
===========================train_params===========================
model_name:ResNeXt101_32x4d
model_name:ResNeXt152_32x4d
python:python3.7
gpu_list:0|0,1
-o Global.device:gpu
@ -13,7 +13,7 @@ train_infer_img_dir:./dataset/ILSVRC2012/val
null:null
##
trainer:amp_train
amp_train:tools/train.py -c ppcls/configs/ImageNet/ResNeXt/ResNeXt101_32x4d.yaml -o Global.seed=1234 -o DataLoader.Train.sampler.shuffle=False -o DataLoader.Train.loader.num_workers=0 -o DataLoader.Train.loader.use_shared_memory=False -o AMP.scale_loss=128 -o AMP.use_dynamic_loss_scaling=True -o AMP.level=O2 -o Global.eval_during_train=False -o Global.save_interval=2
amp_train:tools/train.py -c ppcls/configs/ImageNet/ResNeXt/ResNeXt152_32x4d.yaml -o Global.seed=1234 -o DataLoader.Train.sampler.shuffle=False -o DataLoader.Train.loader.num_workers=0 -o DataLoader.Train.loader.use_shared_memory=False -o AMP.scale_loss=128 -o AMP.use_dynamic_loss_scaling=True -o AMP.level=O2 -o Global.eval_during_train=False -o Global.save_interval=2
pact_train:null
fpgm_train:null
distill_train:null
@ -21,19 +21,19 @@ null:null
null:null
##
===========================eval_params===========================
eval:tools/eval.py -c ppcls/configs/ImageNet/ResNeXt/ResNeXt101_32x4d.yaml
eval:tools/eval.py -c ppcls/configs/ImageNet/ResNeXt/ResNeXt152_32x4d.yaml
null:null
##
===========================infer_params==========================
-o Global.save_inference_dir:./inference
-o Global.pretrained_model:
norm_export:tools/export_model.py -c ppcls/configs/ImageNet/ResNeXt/ResNeXt101_32x4d.yaml
norm_export:tools/export_model.py -c ppcls/configs/ImageNet/ResNeXt/ResNeXt152_32x4d.yaml
quant_export:null
fpgm_export:null
distill_export:null
kl_quant:null
export2:null
pretrained_model_url:https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/ResNeXt101_32x4d_pretrained.pdparams
pretrained_model_url:https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/ResNeXt152_32x4d_pretrained.pdparams
infer_model:../inference/
infer_export:True
infer_quant:Fasle

View File

@ -1,5 +1,5 @@
===========================train_params===========================
model_name:ResNeXt101_32x4d
model_name:ResNeXt152_32x4d
python:python3.7
gpu_list:0|0,1
-o Global.device:gpu
@ -13,7 +13,7 @@ train_infer_img_dir:./dataset/ILSVRC2012/val
null:null
##
trainer:norm_train
norm_train:tools/train.py -c ppcls/configs/ImageNet/ResNeXt/ResNeXt101_32x4d.yaml -o Global.seed=1234 -o DataLoader.Train.sampler.shuffle=False -o DataLoader.Train.loader.num_workers=0 -o DataLoader.Train.loader.use_shared_memory=False -o Global.eval_during_train=False -o Global.save_interval=2
norm_train:tools/train.py -c ppcls/configs/ImageNet/ResNeXt/ResNeXt152_32x4d.yaml -o Global.seed=1234 -o DataLoader.Train.sampler.shuffle=False -o DataLoader.Train.loader.num_workers=0 -o DataLoader.Train.loader.use_shared_memory=False -o Global.eval_during_train=False -o Global.save_interval=2
pact_train:null
fpgm_train:null
distill_train:null
@ -21,19 +21,19 @@ null:null
null:null
##
===========================eval_params===========================
eval:tools/eval.py -c ppcls/configs/ImageNet/ResNeXt/ResNeXt101_32x4d.yaml
eval:tools/eval.py -c ppcls/configs/ImageNet/ResNeXt/ResNeXt152_32x4d.yaml
null:null
##
===========================infer_params==========================
-o Global.save_inference_dir:./inference
-o Global.pretrained_model:
norm_export:tools/export_model.py -c ppcls/configs/ImageNet/ResNeXt/ResNeXt101_32x4d.yaml
norm_export:tools/export_model.py -c ppcls/configs/ImageNet/ResNeXt/ResNeXt152_32x4d.yaml
quant_export:null
fpgm_export:null
distill_export:null
kl_quant:null
export2:null
pretrained_model_url:https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/ResNeXt101_32x4d_pretrained.pdparams
pretrained_model_url:https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/ResNeXt152_32x4d_pretrained.pdparams
infer_model:../inference/
infer_export:True
infer_quant:Fasle

View File

@ -0,0 +1,103 @@
#!/bin/bash
source ./test_tipc/common_func.sh
# function: Extract dataset name from eval.log
extract_dataset_name() {
local file_path="$1"
if [ ! -f "$file_path" ]; then
return 1
fi
local result=$(grep "class_id_map_file" "$file_path" | \
awk -F 'class_id_map_file' '{print $2}' | \
grep -o 'ppcls/utils/.*_label_list.txt' | \
sed 's/^ppcls\/utils\///;s/_label_list.txt$//')
if [ "$result" = "imagenet1k" ]; then
echo "ImageNet1K"
else
echo "$result"
fi
}
# function: Extract top1 accuracy from the last line of eval.log
extract_top1() {
local file_path="$1"
if [ ! -f "$file_path" ]; then
return 1
fi
local last_line=$(tail -n 1 "$file_path")
local top1_acc=$(echo "$last_line" | \
awk -F 'top1:' '{print $2}' | \
awk -F ',' '{print $1}' | \
sed 's/^[[:space:]]*//;s/[[:space:]]*$//')
echo "$top1_acc"
}
log_dir="./eval_logs"
md_path="./eval_list.md"
mkdir -p "$log_dir"
touch "$md_path"
if [ ! -s "$md_path" ]; then
echo "| Model | Dataset | Top1 Acc |" >> "$md_path"
echo "| :--: | :--: | :--: |" >> "$md_path"
fi
target_dir="./test_tipc/configs"
skip_num=0
eval_num=0
# 1. Traverse the config directory of each model
for target_dir_path in "$target_dir"/*; do
# 2. Get the path of train_infer_python.txt
for txt_path in "$target_dir_path"/*train_infer_python.txt; do
if [ -f "$txt_path" ]; then
# 3. Use python3.9
sed -i "s/python3.7/python3.9/g" "$txt_path"
sed -i "s/python3.10/python3.9/g" "$txt_path"
# 4. Parser params
dataline=$(cat $txt_path)
IFS=$'\n'
lines=(${dataline})
model_name=$(func_parser_value "${lines[1]}")
python=$(func_parser_value "${lines[2]}")
use_gpu_key=$(func_parser_key "${lines[4]}")
use_gpu_value=$(func_parser_value "${lines[4]}")
eval_py=$(func_parser_value "${lines[23]}")
eval_key=$(func_parser_key "${lines[24]}")
eval_value=$(func_parser_value "${lines[24]}")
set_eval_params=$(func_set_params "${eval_key}" "${eval_value}")
set_use_gpu=$(func_set_params "${use_gpu_key}" "${use_gpu_value}")
log_path="${log_dir}/${model_name}_eval.log"
# 5. Skip models that already listed in the md file or don't have eval.py
if grep -q " $model_name " "$md_path" || [ "${eval_py}" == "null" ]; then
((skip_num++))
continue
fi
# 6. Prepare data -> eval -> remove weight file
# With "-o Arch.pretrained=True", the weights will be downloaded
# to "/root/.paddleclas/weights" and loaded automatically.
prepare_cmd="bash test_tipc/prepare.sh "$txt_path" 'lite_train_lite_infer'"
eval_cmd="${python} ${eval_py} ${set_use_gpu} ${set_eval_params} -o Arch.pretrained=True > ${log_path}"
rm_weight_cmd="rm -rf /root/.paddleclas/weights/*"
eval ${prepare_cmd}
eval ${eval_cmd}
status_check $? "${eval_cmd}" "./eval_status.log" "${model_name}"
eval ${rm_weight_cmd}
# 7. Write evaluation results to md file
dataset_name=$(extract_dataset_name "${log_path}")
top1_acc=$(extract_top1 "${log_path}")
echo "| $model_name | $dataset_name | $top1_acc |" >> "$md_path"
((eval_num++))
sleep 5
fi
done
done
model_num=$(expr $skip_num + $eval_num)
echo "Done."
echo "Skip ${skip_num} models."
echo "Eval ${eval_num} models."
echo "${model_num} models in total."