We provide the inference demo script for WizardCoder-Family.
- According to the instructions of Llama-X, install the environment.
- Install these packages:
pip install transformers==4.31.0
pip install vllm==0.1.4
- Enjoy your demo:
CUDA_VISIBLE_DEVICES=1,2,3,4 python wizardcoder_demo.py \
--base_model "WizardLM/WizardCoder-Python-34B-V1.0" \
--n_gpus 4
Note: This script supports WizardLM/WizardCoder-Python-34B/13B/7B-V1.0
. If you want to inference with WizardLM/WizardCoder-15B/3B/1B-V1.0
, please change the stop_tokens = ['</s>']
to stop_tokens = ['<|endoftext|>']
in the script.
We provide the inference demo script for WizardMath-Family.
- According to the instructions of Llama-X, install the environment.
- Install these packages:
pip install transformers==4.31.0
pip install vllm==0.1.4
- Enjoy your demo:
CUDA_VISIBLE_DEVICES=0 python wizardmath_demo.py \
--base_model "xxx/path/to/wizardmath_7b_model" \
--n_gpus 1
We provide the inference demo script for WizardLM-Family.
- According to the instructions of Llama-X, install the environment.
- Install these packages:
pip install transformers==4.31.0
pip install vllm==0.1.4
- Enjoy your demo:
CUDA_VISIBLE_DEVICES=0 python wizardLM_demo.py \
--base_model "xxx/path/to/wizardLM_7b_model" \
--n_gpus 1