mmengine/examples/llama2
Mashiro f4f2555324
Llama2 example (#1264)
2023-07-24 10:20:21 +08:00
..
README.md Llama2 example (#1264) 2023-07-24 10:20:21 +08:00
fsdp_finetune.py Llama2 example (#1264) 2023-07-24 10:20:21 +08:00
generate.py Llama2 example (#1264) 2023-07-24 10:20:21 +08:00

README.md

Train Llama2 in MMEngine

Setup env

Note: This example requires PyTorch 2.0+ and MMEngine 0.8.0+.

  • Install MMEngine

    git clone https://github.com/open-mmlab/mmengine.git
    cd mmengine
    pip install -e . -v
    
  • Install third-party dependencies

    pip install -U transformers accelerate tokenizers
    

Prepare data

mkdir data
wget https://raw.githubusercontent.com/tatsu-lab/stanford_alpaca/main/alpaca_data.json -O data/alpaca_data.json

Prepare model

Download model weights from https://huggingface.co/meta-llama/Llama-2-7b-hf

Train

torchrun --nproc-per-node 8 examples/llama2/fsdp_finetune.py data/alpaca_data.json ${model_weights}

Inference

python examples/llama2/generate.py ${checkpoints}