2023-04-18 13:01:24 +00:00
## How to Prepare Vicuna Weight
Vicuna is an open-source LLAMA-based LLM that has a performance close to ChatGPT.
We currently use the v0 version of Vicuna-13B.
2023-04-20 19:03:34 +00:00
To prepare Vicuna’ s weight, first download Vicuna’ s **delta** weight from [https://huggingface.co/lmsys/vicuna-13b-delta-v0 ](https://huggingface.co/lmsys/vicuna-13b-delta-v0 ).
In case you have git-lfs installed (https://git-lfs.com), this can be done by
2023-04-18 13:01:24 +00:00
```
git lfs install
2023-04-20 19:03:34 +00:00
git clone https://huggingface.co/lmsys/vicuna-13b-delta-v0 # more powerful, need at least 24G gpu memory
# or
git clone https://huggingface.co/lmsys/vicuna-7b-delta-v0 # smaller, need 12G gpu memory
2023-04-18 13:01:24 +00:00
```
Note that this is not directly the working weight, but the difference between the working weight and the original weight of LLAMA-13B. (Due to LLAMA’ s rules, we cannot distribute the weight of LLAMA.)
2023-04-20 19:03:34 +00:00
Then, you need to obtain the original LLAMA-7B or LLAMA-13B weights in the HuggingFace format
either following the instruction provided by HuggingFace
[here ](https://huggingface.co/docs/transformers/main/model_doc/llama ) or from the Internet.
2023-04-18 13:01:24 +00:00
When these two weights are ready, we can use tools from Vicuna’ s team to create the real working weight.
First, Install their library that is compatible with v0 Vicuna by
```
2023-04-18 19:54:09 +00:00
pip install git+https://github.com/lm-sys/FastChat.git@v0.1.10
2023-04-18 13:01:24 +00:00
```
Then, run the following command to create the final working weight
```
2023-04-20 19:03:34 +00:00
python -m fastchat.model.apply_delta --base /path/to/llama-13bOR7b-hf/ --target /path/to/save/working/vicuna/weight/ --delta /path/to/vicuna-13bOR7b-delta-v0/
2023-04-18 13:01:24 +00:00
```
Now you are good to go!