training w/ Fastchat inference w/ vllm
Create a new python environment. We will install everything with cuda 11.8. Other cuda versions and library versions likely won't work for now.
conda create -n llm python=3.10 -y
conda deactivate && conda activate llm
bash setup.shPlay with vllm_test.py
Modify data_module.py to support the data format, and use finetune.sh or finetune_lora.sh