1 Star 0 Fork 1

鲲鹏 / MiniGPT-4

forked from Mr.Huang / MiniGPT-4 
加入 Gitee
与超过 1200万 开发者一起发现、参与优秀开源项目,私有仓库也完全免费 :)
免费加入
克隆/下载
PrepareVicuna.md 1.57 KB
一键复制 编辑 原始数据 按行查看 历史
Deyao Zhu 提交于 2023-04-20 22:03 . add checkpoint for vicuna 7b

How to Prepare Vicuna Weight

Vicuna is an open-source LLAMA-based LLM that has a performance close to ChatGPT. We currently use the v0 version of Vicuna-13B.

To prepare Vicuna’s weight, first download Vicuna’s delta weight from https://huggingface.co/lmsys/vicuna-13b-delta-v0. In case you have git-lfs installed (https://git-lfs.com), this can be done by

git lfs install
git clone https://huggingface.co/lmsys/vicuna-13b-delta-v0  # more powerful, need at least 24G gpu memory
# or
git clone https://huggingface.co/lmsys/vicuna-7b-delta-v0  # smaller, need 12G gpu memory

Note that this is not directly the working weight, but the difference between the working weight and the original weight of LLAMA-13B. (Due to LLAMA’s rules, we cannot distribute the weight of LLAMA.)

Then, you need to obtain the original LLAMA-7B or LLAMA-13B weights in the HuggingFace format either following the instruction provided by HuggingFace here or from the Internet.

When these two weights are ready, we can use tools from Vicuna’s team to create the real working weight. First, Install their library that is compatible with v0 Vicuna by

pip install git+https://github.com/lm-sys/FastChat.git@v0.1.10

Then, run the following command to create the final working weight

python -m fastchat.model.apply_delta --base /path/to/llama-13bOR7b-hf/  --target /path/to/save/working/vicuna/weight/  --delta /path/to/vicuna-13bOR7b-delta-v0/

Now you are good to go!

1
https://gitee.com/garenx/MiniGPT-4.git
git@gitee.com:garenx/MiniGPT-4.git
garenx
MiniGPT-4
MiniGPT-4
main

搜索帮助

53164aa7 5694891 3bd8fe86 5694891