pull/11509/head
co63oc 2024-01-02 11:02:13 +08:00 committed by GitHub
parent 49ef54ee3c
commit 3b6f117c44
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
3 changed files with 3 additions and 3 deletions

View File

@ -127,7 +127,7 @@ def main(config, device, logger, vdl_writer):
run_sensitive_analysis=True:
Automatically compute the sensitivities of convolutions in a model.
The sensitivity of a convolution is the losses of accuracy on test dataset in
differenct pruned ratios. The sensitivities can be used to get a group of best
different pruned ratios. The sensitivities can be used to get a group of best
ratios with some condition.
run_sensitive_analysis=False:

View File

@ -139,7 +139,7 @@ def main(config, device, logger, vdl_writer):
if not (os.path.exists(os.path.join(inference_model_dir, "inference.pdmodel")) and \
os.path.exists(os.path.join(inference_model_dir, "inference.pdiparams")) ):
raise ValueError(
"Please set inference model dir in Global.inference_model or Global.pretrained_model for post-quantazition"
"Please set inference model dir in Global.inference_model or Global.pretrained_model for post-quantization"
)
if is_layoutxlm_ser:

View File

@ -127,7 +127,7 @@ def check_device(use_gpu, use_xpu=False, use_npu=False, use_mlu=False):
try:
if use_gpu and use_xpu:
print("use_xpu and use_gpu can not both be ture.")
print("use_xpu and use_gpu can not both be true.")
if use_gpu and not paddle.is_compiled_with_cuda():
print(err.format("use_gpu", "cuda", "gpu", "use_gpu"))
sys.exit(1)