From 0c93277f481b1697346f5342f6a6a3229d3af9a7 Mon Sep 17 00:00:00 2001 From: sibo2rr <1415419833@qq.com> Date: Mon, 14 Feb 2022 20:31:27 +0800 Subject: [PATCH] fix: remove outdated files --- docs/en/advanced_tutorials/index.rst | 2 +- docs/en/algorithm_introduction/index.rst | 2 +- docs/en/data_preparation/index.rst | 2 +- docs/en/doc_en.rst | 1 - docs/en/extension/VisualDL_en.md | 44 ------- docs/en/extension/index.rst | 13 -- .../en/extension/multi_machine_training_en.md | 11 -- docs/en/extension/paddle_hub_en.md | 6 - .../extension/paddle_mobile_inference_en.md | 114 ------------------ docs/en/extension/paddle_quantization_en.md | 12 -- docs/en/extension/paddle_serving_en.md | 64 ---------- docs/en/extension/train_with_DALI_en.md | 62 ---------- docs/en/faq_series/index.rst | 2 +- docs/en/image_recognition_pipeline/index.rst | 2 +- docs/en/index.rst | 1 - docs/en/inference_deployment/index.rst | 2 +- docs/en/installation/index.rst | 2 +- docs/en/installation/install_paddle_en.md | 6 +- docs/en/introduction/index.rst | 2 +- docs/en/models/index.rst | 38 +++--- docs/en/models_training/index.rst | 2 +- docs/en/others/index.rst | 2 +- docs/en/others/paddle_mobile_inference_en.md | 2 +- docs/en/quick_start/index.rst | 2 +- 24 files changed, 34 insertions(+), 362 deletions(-) delete mode 100644 docs/en/extension/VisualDL_en.md delete mode 100644 docs/en/extension/index.rst delete mode 100644 docs/en/extension/multi_machine_training_en.md delete mode 100644 docs/en/extension/paddle_hub_en.md delete mode 100644 docs/en/extension/paddle_mobile_inference_en.md delete mode 100644 docs/en/extension/paddle_quantization_en.md delete mode 100644 docs/en/extension/paddle_serving_en.md delete mode 100644 docs/en/extension/train_with_DALI_en.md diff --git a/docs/en/advanced_tutorials/index.rst b/docs/en/advanced_tutorials/index.rst index b0121559b..e741733e7 100644 --- a/docs/en/advanced_tutorials/index.rst +++ b/docs/en/advanced_tutorials/index.rst @@ -2,7 +2,7 @@ advanced_tutorials ================================ .. toctree:: - :maxdepth: 1 + :maxdepth: 2 DataAugmentation_en.md distillation/index diff --git a/docs/en/algorithm_introduction/index.rst b/docs/en/algorithm_introduction/index.rst index 951108910..e0fad7973 100644 --- a/docs/en/algorithm_introduction/index.rst +++ b/docs/en/algorithm_introduction/index.rst @@ -2,7 +2,7 @@ algorithm_introduction ================================ .. toctree:: - :maxdepth: 1 + :maxdepth: 2 image_classification_en.md metric_learning_en.md diff --git a/docs/en/data_preparation/index.rst b/docs/en/data_preparation/index.rst index 589157255..e668126ec 100644 --- a/docs/en/data_preparation/index.rst +++ b/docs/en/data_preparation/index.rst @@ -2,7 +2,7 @@ data_preparation ================================ .. toctree:: - :maxdepth: 1 + :maxdepth: 2 recognition_dataset_en.md classification_dataset_en.md diff --git a/docs/en/doc_en.rst b/docs/en/doc_en.rst index ee21fcc7b..aa25829f5 100644 --- a/docs/en/doc_en.rst +++ b/docs/en/doc_en.rst @@ -15,7 +15,6 @@ Welcome to PaddleClas! algorithm_introduction/index advanced_tutorials/index others/index - extension/index faq_series/index diff --git a/docs/en/extension/VisualDL_en.md b/docs/en/extension/VisualDL_en.md deleted file mode 100644 index 403a74f49..000000000 --- a/docs/en/extension/VisualDL_en.md +++ /dev/null @@ -1,44 +0,0 @@ -# Use VisualDL to visualize the training - -## Preface -VisualDL, a visualization analysis tool of PaddlePaddle, provides a variety of charts to show the trends of parameters, and visualizes model structures, data samples, histograms of tensors, PR curves , ROC curves and high-dimensional data distributions. It enables users to understand the training process and the model structure more clearly and intuitively so as to optimize models efficiently. For more information, please refer to [VisualDL](https://github.com/PaddlePaddle/VisualDL/). - -## Use VisualDL in PaddleClas -Now PaddleClas support use VisualDL to visualize the changes of learning rate, loss, accuracy in training. - -### Set config and start training -You only need to set the field `Global.use_visualdl` to `True` in train config: - -```yaml -# config.yaml -Global: -... - use_visualdl: True -... -``` - -PaddleClas will save the VisualDL logs to subdirectory `vdl/` under the output directory specified by `Global.output_dir`. And then you just need to start training normally: - -```shell -python3 tools/train.py -c config.yaml -``` - -### Start VisualDL -After starting the training program, you can start the VisualDL service in a new terminal session: - -```shell - visualdl --logdir ./output/vdl/ -``` - -In the above command, `--logdir` specify the directory of the VisualDL logs produced in training. VisualDL will traverse and iterate to find the subdirectories of the specified directory to visualize all the experimental results. You can also use the following parameters to set the IP and port number of the VisualDL service: - -* `--host`:ip, default is 127.0.0.1 -* `--port`:port, default is 8040 - -More information about the command,please refer to [VisualDL](https://github.com/PaddlePaddle/VisualDL/blob/develop/README.md#2-launch-panel). - -Then you can enter the address `127.0.0.1:8840` and view the training process in the browser: - - -![](../../images/VisualDL/train_loss.png) - diff --git a/docs/en/extension/index.rst b/docs/en/extension/index.rst deleted file mode 100644 index 880a85c1b..000000000 --- a/docs/en/extension/index.rst +++ /dev/null @@ -1,13 +0,0 @@ -extension -================================ - -.. toctree:: - :maxdepth: 1 - - train_with_DALI_en.md - VisualDL_en.md - paddle_mobile_inference_en.md - paddle_serving_en.md - paddle_quantization_en.md - paddle_hub_en.md - multi_machine_training_en.md diff --git a/docs/en/extension/multi_machine_training_en.md b/docs/en/extension/multi_machine_training_en.md deleted file mode 100644 index d4fb99784..000000000 --- a/docs/en/extension/multi_machine_training_en.md +++ /dev/null @@ -1,11 +0,0 @@ -# Distributed Training - -Distributed deep neural networks training is highly efficient in PaddlePaddle. -And it is one of the PaddlePaddle's core advantage technologies. -On image classification tasks, distributed training can achieve almost linear acceleration ratio. -[Fleet](https://github.com/PaddlePaddle/Fleet) is High-Level API for distributed training in PaddlePaddle. -By using Fleet, a user can shift from local machine paddlepaddle code to distributed code easily. -In order to support both single-machine training and multi-machine training, -[PaddleClas](https://github.com/PaddlePaddle/PaddleClas) uses the Fleet API interface. -For more information about distributed training, -please refer to [Fleet API documentation](https://github.com/PaddlePaddle/Fleet/blob/develop/README.md). diff --git a/docs/en/extension/paddle_hub_en.md b/docs/en/extension/paddle_hub_en.md deleted file mode 100644 index d9d833a01..000000000 --- a/docs/en/extension/paddle_hub_en.md +++ /dev/null @@ -1,6 +0,0 @@ -# Paddle Hub - -[PaddleHub](https://github.com/PaddlePaddle/PaddleHub) is a pre-trained model application tool for PaddlePaddle. -Developers can conveniently use the high-quality pre-trained model combined with Fine-tune API to quickly complete the whole process from model migration to deployment. -All the pre-trained models of [PaddleClas](https://github.com/PaddlePaddle/PaddleClas) have been collected by PaddleHub. -For further details, please refer to [PaddleHub website](https://www.paddlepaddle.org.cn/hub). diff --git a/docs/en/extension/paddle_mobile_inference_en.md b/docs/en/extension/paddle_mobile_inference_en.md deleted file mode 100644 index 86c8e040d..000000000 --- a/docs/en/extension/paddle_mobile_inference_en.md +++ /dev/null @@ -1,114 +0,0 @@ -# Paddle-Lite - -## Introduction - -[Paddle-Lite](https://github.com/PaddlePaddle/Paddle-Lite) is a set of lightweight inference engine which is fully functional, easy to use and then performs well. Lightweighting is reflected in the use of fewer bits to represent the weight and activation of the neural network, which can greatly reduce the size of the model, solve the problem of limited storage space of the mobile device, and the inference speed is better than other frameworks on the whole. - -In [PaddleClas](https://github.com/PaddlePaddle/PaddleClas), we uses Paddle-Lite to [evaluate the performance on the mobile device](../models/Mobile_en.md), in this section we uses the `MobileNetV1` model trained on the `ImageNet1k` dataset as an example to introduce how to use `Paddle-Lite` to evaluate the model speed on the mobile terminal (evaluated on SD855) - -## Evaluation Steps - -### Export the Inference Model - -* First you should transform the saved model during training to the special model which can be used to inference, the special model can be exported by `tools/export_model.py`, the specific way of transform is as follows. - -```shell -python tools/export_model.py -m MobileNetV1 -p pretrained/MobileNetV1_pretrained/ -o inference/MobileNetV1 -``` - -Finally the `model` and `parmas` can be saved in `inference/MobileNetV1`. - - -### Download Benchmark Binary File - -* Use the adb (Android Debug Bridge) tool to connect the Android phone and the PC, then develop and debug. After installing adb and ensuring that the PC and the phone are successfully connected, use the following command to view the ARM version of the phone and select the pre-compiled library based on ARM version. - -```shell -adb shell getprop ro.product.cpu.abi -``` - -* Download Benchmark_bin File - -```shell -wget -c https://paddle-inference-dist.bj.bcebos.com/PaddleLite/benchmark_0/benchmark_bin_v8 -``` - -If the ARM version is v7, the v7 benchmark_bin file should be downloaded, the command is as follow. - -```shell -wget -c https://paddle-inference-dist.bj.bcebos.com/PaddleLite/benchmark_0/benchmark_bin_v7 -``` - -### Inference benchmark - -After the PC and mobile phone are successfully connected, use the following command to start the model evaluation. - -``` -sh deploy/lite/benchmark/benchmark.sh ./benchmark_bin_v8 ./inference result_armv8.txt true -``` - -Where `./benchmark_bin_v8` is the path of the benchmark binary file, `./inference` is the path of all the models that need to be evaluated, `result_armv8.txt` is the result file, and the final parameter `true` means that the model will be optimized before evaluation. Eventually, the evaluation result file of `result_armv8.txt` will be saved in the current folder. The specific performances are as follows. - -``` -PaddleLite Benchmark -Threads=1 Warmup=10 Repeats=30 -MobileNetV1 min = 30.89100 max = 30.73600 average = 30.79750 - -Threads=2 Warmup=10 Repeats=30 -MobileNetV1 min = 18.26600 max = 18.14000 average = 18.21637 - -Threads=4 Warmup=10 Repeats=30 -MobileNetV1 min = 10.03200 max = 9.94300 average = 9.97627 -``` - -Here is the model inference speed under different number of threads, the unit is FPS, taking model on one threads as an example, the average speed of MobileNetV1 on SD855 is `30.79750FPS`. - -### Model Optimization and Speed Evaluation - -* In II.III section, we mention that the model will be optimized before evaluation, here you can first optimize the model, and then directly load the optimized model for speed evaluation - -* Paddle-Lite -In Paddle-Lite, we provides multiple strategies to automatically optimize the original training model, which contain Quantify, Subgraph fusion, Hybrid scheduling, Kernel optimization and so on. In order to make the optimization more convenient and easy to use, we provide opt tools to automatically complete the optimization steps and output a lightweight, optimal and executable model in Paddle-Lite, which can be downloaded on [Paddle-Lite Model Optimization Page](https://paddle-lite.readthedocs.io/zh/latest/user_guides/model_optimize_tool.html). Here we take `MacOS` as our development environment, download[opt_mac](https://paddlelite-data.bj.bcebos.com/model_optimize_tool/opt_mac) model optimization tools and use the following commands to optimize the model. - - -```shell -model_file="../MobileNetV1/model" -param_file="../MobileNetV1/params" -opt_models_dir="./opt_models" -mkdir ${opt_models_dir} -./opt_mac --model_file=${model_file} \ - --param_file=${param_file} \ - --valid_targets=arm \ - --optimize_out_type=naive_buffer \ - --prefer_int8_kernel=false \ - --optimize_out=${opt_models_dir}/MobileNetV1 -``` - -Where the `model_file` and `param_file` are exported model file and the file address respectively, after transforming successfully, the `MobileNetV1.nb` will be saved in `opt_models` - - - -Use the benchmark_bin file to load the optimized model for evaluation. The commands are as follows. - -```shell -bash benchmark.sh ./benchmark_bin_v8 ./opt_models result_armv8.txt -``` - -Finally the result is saved in `result_armv8.txt` and shown as follow. - -``` -PaddleLite Benchmark -Threads=1 Warmup=10 Repeats=30 -MobileNetV1_lite min = 30.89500 max = 30.78500 average = 30.84173 - -Threads=2 Warmup=10 Repeats=30 -MobileNetV1_lite min = 18.25300 max = 18.11000 average = 18.18017 - -Threads=4 Warmup=10 Repeats=30 -MobileNetV1_lite min = 10.00600 max = 9.90000 average = 9.96177 -``` - - -Taking the model on one threads as an example, the average speed of MobileNetV1 on SD855 is `30.84173FPS`. - -More specific parameter explanation and Paddle-Lite usage can refer to [Paddle-Lite docs](https://paddle-lite.readthedocs.io/zh/latest/)。 diff --git a/docs/en/extension/paddle_quantization_en.md b/docs/en/extension/paddle_quantization_en.md deleted file mode 100644 index e84e3a820..000000000 --- a/docs/en/extension/paddle_quantization_en.md +++ /dev/null @@ -1,12 +0,0 @@ -# Model Quantifization - -Int8 quantization is one of the key features in [PaddleSlim](https://github.com/PaddlePaddle/PaddleSlim). -It supports two kinds of training aware, **Dynamic strategy** and **Static strategy**, -layer-wise and channel-wise quantization, -and using PaddleLite to deploy models generated by PaddleSlim. - -By using this toolkit, [PaddleClas](https://github.com/PaddlePaddle/PaddleClas) quantized the mobilenet_v3_large_x1_0 model whose accuracy is 78.9% after distilled. -After quantized, the prediction speed is accelerated from 19.308ms to 14.395ms on SD855. -The storage size is reduced from 21M to 10M. -The top1 recognition accuracy rate is 75.9%. -For specific training methods, please refer to [PaddleSlim quant aware](../../../deploy/slim/README_en.md)。 diff --git a/docs/en/extension/paddle_serving_en.md b/docs/en/extension/paddle_serving_en.md deleted file mode 100644 index 3ad259526..000000000 --- a/docs/en/extension/paddle_serving_en.md +++ /dev/null @@ -1,64 +0,0 @@ -# Model Service Deployment - -## Overview -[Paddle Serving](https://github.com/PaddlePaddle/Serving) aims to help deep-learning researchers to easily deploy online inference services, supporting one-click deployment of industry, high concurrency and efficient communication between client and server and supporting multiple programming languages to develop clients. - -Taking HTTP inference service deployment as an example to introduce how to use PaddleServing to deploy model services in PaddleClas. - -## Serving Install - -It is recommends to use docker to install and deploy the Serving environment in the Serving official website, first, you need to pull the docker environment and create Serving-based docker. - -```shell -nvidia-docker pull hub.baidubce.com/paddlepaddle/serving:0.2.0-gpu -nvidia-docker run -p 9292:9292 --name test -dit hub.baidubce.com/paddlepaddle/serving:0.2.0-gpu -nvidia-docker exec -it test bash -``` - -In docker, you need to install some packages about Serving - -```shell -pip install paddlepaddle-gpu -pip install paddle-serving-client -pip install paddle-serving-server-gpu -``` - -* If the installation speed is too slow, you can add `-i https://pypi.tuna.tsinghua.edu.cn/simple` following pip to speed up the process. - -* If you want to deploy CPU service, you can install the cpu version of Serving, the command is as follow. - -```shell -pip install paddle-serving-server -``` - -### Export Model - -Exporting the Serving model using `tools/export_serving_model.py`, taking ResNet50_vd as an example, the command is as follow. - -```shell -python tools/export_serving_model.py -m ResNet50_vd -p ./pretrained/ResNet50_vd_pretrained/ -o serving -``` - -finally, the client configures, model parameters and structure file will be saved in `ppcls_client_conf` and `ppcls_model`. - - -### Service Deployment and Request - -* Using the following commands to start the Serving. - -```shell -python tools/serving/image_service_gpu.py serving/ppcls_model workdir 9292 -``` - -`serving/ppcls_model` is the address of the Serving model just saved, `workdir` is the work directory, and `9292` is the port of the service. - - -* Using the following script to send an identification request to the Serving and return the result. - -``` -python tools/serving/image_http_client.py 9292 ./docs/images/logo.png -``` - -`9292` is the port for sending the request, which is consistent with the Serving starting port, and `./docs/images/logo.png` is the test image, the final top1 label and probability are returned. - -* For more Serving deployment, such RPC inference service, you can refer to the Serving official website: [https://github.com/PaddlePaddle/Serving/tree/develop/python/examples/imagenet](https://github.com/PaddlePaddle/Serving/tree/develop/python/examples/imagenet) diff --git a/docs/en/extension/train_with_DALI_en.md b/docs/en/extension/train_with_DALI_en.md deleted file mode 100644 index a67a76166..000000000 --- a/docs/en/extension/train_with_DALI_en.md +++ /dev/null @@ -1,62 +0,0 @@ -# Train with DALI - -## Preface -[The NVIDIA Data Loading Library](https://docs.nvidia.com/deeplearning/dali/user-guide/docs/index.html) (DALI) is a library for data loading and pre-processing to accelerate deep learning applications. It can build Dataloader of Paddle. - -Since the Deep learning relies on a large amount of data in the training stage, these data need to be loaded and preprocessed. These operations are usually executed on the CPU, which limits the further improvement of the training speed, especially when the batch_size is large, which become the bottleneck of speed. DALI can use GPU to accelerate these operations, thereby further improve the training speed. - -## Installing DALI -DALI only support Linux x64 and version of CUDA is 10.2 or later. - -* For CUDA 10: - - pip install --extra-index-url https://developer.download.nvidia.com/compute/redist nvidia-dali-cuda100 - -* For CUDA 11.0: - - pip install --extra-index-url https://developer.download.nvidia.com/compute/redist nvidia-dali-cuda110 - -For more information about installing DALI, please refer to [DALI](https://docs.nvidia.com/deeplearning/dali/user-guide/docs/installation.html). - -## Using DALI -Paddleclas supports training with DALI in static graph. Since DALI only supports GPU training, `CUDA_VISIBLE_DEVICES` needs to be set, and DALI needs to occupy GPU memory, so it needs to reserve GPU memory for Dali. To train with DALI, just set the fields in the training config `use_dali = True`, or start the training by the following command: - -```shell -# set the GPUs that can be seen -export CUDA_VISIBLE_DEVICES="0" - -# set the GPU memory used for neural network training, generally 0.8 or 0.7, and the remaining GPU memory is reserved for DALI -export FLAGS_fraction_of_gpu_memory_to_use=0.80 - -python tools/static/train.py -c configs/ResNet/ResNet50.yaml -o use_dali=True -``` - -And you can train with muti-GPUs: - -```shell -# set the GPUs that can be seen -export CUDA_VISIBLE_DEVICES="0,1,2,3,4,5,6,7" - -# set the GPU memory used for neural network training, generally 0.8 or 0.7, and the remaining GPU memory is reserved for DALI -export FLAGS_fraction_of_gpu_memory_to_use=0.80 - -python -m paddle.distributed.launch \ - --gpus="0,1,2,3,4,5,6,7" \ - tools/static/train.py \ - -c ./configs/ResNet/ResNet50.yaml \ - -o use_dali=True -``` - -## Train with FP16 - -On the basis of the above, using FP16 half-precision can further improve the training speed, you can refer to the following command. - -```shell -export CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 -export FLAGS_fraction_of_gpu_memory_to_use=0.8 - -python -m paddle.distributed.launch \ - --gpus="0,1,2,3,4,5,6,7" \ - tools/static/train.py \ - -c configs/ResNet/ResNet50_fp16.yaml -``` diff --git a/docs/en/faq_series/index.rst b/docs/en/faq_series/index.rst index 106fc0708..69a9f2d2a 100644 --- a/docs/en/faq_series/index.rst +++ b/docs/en/faq_series/index.rst @@ -2,7 +2,7 @@ faq_series ================================ .. toctree:: - :maxdepth: 1 + :maxdepth: 2 faq_2021_s2_en.md faq_2021_s1_en.md diff --git a/docs/en/image_recognition_pipeline/index.rst b/docs/en/image_recognition_pipeline/index.rst index 1e87e37e1..95aa67d4c 100644 --- a/docs/en/image_recognition_pipeline/index.rst +++ b/docs/en/image_recognition_pipeline/index.rst @@ -2,7 +2,7 @@ image_recognition_pipeline ================================ .. toctree:: - :maxdepth: 1 + :maxdepth: 2 mainbody_detection_en.md feature_extraction_en.md diff --git a/docs/en/index.rst b/docs/en/index.rst index 30cc73b43..0bb50d3dc 100644 --- a/docs/en/index.rst +++ b/docs/en/index.rst @@ -5,7 +5,6 @@ :maxdepth: 2 models_training/index - extension/index introduction/index image_recognition_pipeline/index others/index diff --git a/docs/en/inference_deployment/index.rst b/docs/en/inference_deployment/index.rst index 16f9d5557..1e3c344e4 100644 --- a/docs/en/inference_deployment/index.rst +++ b/docs/en/inference_deployment/index.rst @@ -2,7 +2,7 @@ inference_deployment ================================ .. toctree:: - :maxdepth: 1 + :maxdepth: 2 export_model_en.md python_deploy_en.md diff --git a/docs/en/installation/index.rst b/docs/en/installation/index.rst index 832675a6b..39d432ae7 100644 --- a/docs/en/installation/index.rst +++ b/docs/en/installation/index.rst @@ -2,7 +2,7 @@ installation ================================ .. toctree:: - :maxdepth: 1 + :maxdepth: 2 install_paddle_en.md install_paddleclas_en.md diff --git a/docs/en/installation/install_paddle_en.md b/docs/en/installation/install_paddle_en.md index 3922b7325..c282f3ed6 100644 --- a/docs/en/installation/install_paddle_en.md +++ b/docs/en/installation/install_paddle_en.md @@ -1,4 +1,4 @@ -# Installation PaddlePaddle +# Install PaddlePaddle --- @@ -9,7 +9,7 @@ - [3. Install PaddlePaddle using pip](#3) - [4. Verify installation](#4) -At present, **PaddleClas** requires **PaddlePaddle** version **>=2.0**. Docker is recomended to run Paddleclas, for more detailed information about docker and nvidia-docker, you can refer to the [tutorial](https://docs.docker.com/get-started/). If you do not want to use docker, you can skip section [2. (Recommended) Prepare a docker environment](#2), and go into section [3. Install PaddlePaddle using pip](#3). +At present, **PaddleClas** requires **PaddlePaddle** version `>=2.0`. Docker is recomended to run Paddleclas, for more detailed information about docker and nvidia-docker, you can refer to the [tutorial](https://docs.docker.com/get-started/). If you do not want to use docker, you can skip section [2. (Recommended) Prepare a docker environment](#2), and go into section [3. Install PaddlePaddle using pip](#3). @@ -96,5 +96,5 @@ python -c "import paddle; print(paddle.__version__)" Note: * Make sure the compiled source code is later than PaddlePaddle2.0. -* Indicate **WITH_DISTRIBUTE=ON** when compiling, Please refer to [Instruction](https://www.paddlepaddle.org.cn/documentation/docs/zh/develop/install/Tables.html#id3) for more details. +* Indicate `WITH_DISTRIBUTE=ON` when compiling, Please refer to [Instruction](https://www.paddlepaddle.org.cn/documentation/docs/zh/develop/install/Tables.html#id3) for more details. * When running in docker, in order to ensure that the container has enough shared memory for dataloader acceleration of Paddle, please set the parameter `--shm-size=8g` at creating a docker container, if conditions permit, you can set it to a larger value. diff --git a/docs/en/introduction/index.rst b/docs/en/introduction/index.rst index 7b8fdf358..e22a647e8 100644 --- a/docs/en/introduction/index.rst +++ b/docs/en/introduction/index.rst @@ -2,7 +2,7 @@ introduction ================================ .. toctree:: - :maxdepth: 1 + :maxdepth: 2 function_intro_en.md more_demo/index diff --git a/docs/en/models/index.rst b/docs/en/models/index.rst index abeee98c3..4642eb1de 100644 --- a/docs/en/models/index.rst +++ b/docs/en/models/index.rst @@ -2,29 +2,29 @@ models ================================ .. toctree:: - :maxdepth: 1 + :maxdepth: 2 - PP-LCNet_en.md - SEResNext_and_Res2Net_en.md - ReXNet_en.md - Others_en.md - Twins_en.md - Inception_en.md - HarDNet_en.md + DPN_DenseNet_en.md + models_intro_en.md + RepVGG_en.md EfficientNet_and_ResNeXt101_wsl_en.md + ViT_and_DeiT_en.md + SwinTransformer_en.md + Others_en.md + SEResNext_and_Res2Net_en.md ESNet_en.md HRNet_en.md - RepVGG_en.md - RedNet_en.md - Mobile_en.md - ResNeSt_RegNet_en.md - ResNet_and_vd_en.md - models_intro_en.md + ReXNet_en.md + Inception_en.md TNT_en.md - ViT_and_DeiT_en.md - LeViT_en.md + RedNet_en.md DLA_en.md - PVTV2_en.md - DPN_DenseNet_en.md + ResNeSt_RegNet_en.md + PP-LCNet_en.md + HarDNet_en.md + ResNet_and_vd_en.md + LeViT_en.md + Mobile_en.md MixNet_en.md - SwinTransformer_en.md + Twins_en.md + PVTV2_en.md diff --git a/docs/en/models_training/index.rst b/docs/en/models_training/index.rst index 687cb7393..1d27e65bd 100644 --- a/docs/en/models_training/index.rst +++ b/docs/en/models_training/index.rst @@ -2,7 +2,7 @@ models_training ================================ .. toctree:: - :maxdepth: 1 + :maxdepth: 2 config_description_en.md recognition_en.md diff --git a/docs/en/others/index.rst b/docs/en/others/index.rst index e1c98e30c..d106343c8 100644 --- a/docs/en/others/index.rst +++ b/docs/en/others/index.rst @@ -2,7 +2,7 @@ others ================================ .. toctree:: - :maxdepth: 1 + :maxdepth: 2 transfer_learning_en.md train_with_DALI_en.md diff --git a/docs/en/others/paddle_mobile_inference_en.md b/docs/en/others/paddle_mobile_inference_en.md index 153c96ffa..4cb731790 100644 --- a/docs/en/others/paddle_mobile_inference_en.md +++ b/docs/en/others/paddle_mobile_inference_en.md @@ -1,4 +1,4 @@ -# Paddle-Lite +# Benchmark on Mobile --- diff --git a/docs/en/quick_start/index.rst b/docs/en/quick_start/index.rst index 4380b8b1f..2edd78dde 100644 --- a/docs/en/quick_start/index.rst +++ b/docs/en/quick_start/index.rst @@ -2,7 +2,7 @@ quick_start ================================ .. toctree:: - :maxdepth: 1 + :maxdepth: 2 quick_start_classification_new_user_en.md quick_start_classification_professional_en.md