training w/ Fastchat inference w/ vllm
Create a new python environment. We will install everything with cuda 11.8. Other cuda versions and library versions likely won't work for now.
conda create -n llm python=3.10 -y
conda deactivate && conda activate llm
bash setup.sh
Play with vllm_test.py
Modify data_module.py
to support the data format, and use finetune.sh
or finetune_lora.sh