Fix (#11448)
parent
49ef54ee3c
commit
3b6f117c44
|
@ -127,7 +127,7 @@ def main(config, device, logger, vdl_writer):
|
|||
run_sensitive_analysis=True:
|
||||
Automatically compute the sensitivities of convolutions in a model.
|
||||
The sensitivity of a convolution is the losses of accuracy on test dataset in
|
||||
differenct pruned ratios. The sensitivities can be used to get a group of best
|
||||
different pruned ratios. The sensitivities can be used to get a group of best
|
||||
ratios with some condition.
|
||||
|
||||
run_sensitive_analysis=False:
|
||||
|
|
|
@ -139,7 +139,7 @@ def main(config, device, logger, vdl_writer):
|
|||
if not (os.path.exists(os.path.join(inference_model_dir, "inference.pdmodel")) and \
|
||||
os.path.exists(os.path.join(inference_model_dir, "inference.pdiparams")) ):
|
||||
raise ValueError(
|
||||
"Please set inference model dir in Global.inference_model or Global.pretrained_model for post-quantazition"
|
||||
"Please set inference model dir in Global.inference_model or Global.pretrained_model for post-quantization"
|
||||
)
|
||||
|
||||
if is_layoutxlm_ser:
|
||||
|
|
|
@ -127,7 +127,7 @@ def check_device(use_gpu, use_xpu=False, use_npu=False, use_mlu=False):
|
|||
|
||||
try:
|
||||
if use_gpu and use_xpu:
|
||||
print("use_xpu and use_gpu can not both be ture.")
|
||||
print("use_xpu and use_gpu can not both be true.")
|
||||
if use_gpu and not paddle.is_compiled_with_cuda():
|
||||
print(err.format("use_gpu", "cuda", "gpu", "use_gpu"))
|
||||
sys.exit(1)
|
||||
|
|
Loading…
Reference in New Issue