LLaMA weights(llama-13b-hf) + Vicuna weights (vicuna-13b-delta-v1.1) = Vicuna-13B

How to

pip3 install fschat
pip3 install git+https://github.com/huggingface/transformers
sudo apt install git git-lfs
git clone https://huggingface.co/myaniu/Vicuna-13B
python3 -m fastchat.serve.cli --model-path /path/to/Vicuna-13B

Downloads last month
10
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.