Update README.md
Browse files
README.md
CHANGED
@@ -13,7 +13,7 @@ license: apache-2.0
|
|
13 |
|
14 |
This is the merged weights of [bunny-phi-2-siglip-lora](https://huggingface.co/BAAI/bunny-phi-2-siglip-lora).
|
15 |
|
16 |
-
Bunny is a family of lightweight but powerful multimodal models. It offers multiple plug-and-play vision encoders, like EVA-CLIP, SigLIP and language backbones, including Phi-1.5, StableLM-2 and Phi-2. To compensate for the decrease in model size, we construct more informative training data by curated selection from a broader data source. Remarkably, our Bunny-v1.0-3B model built upon SigLIP and Phi-2 outperforms the state-of-the-art MLLMs, not only in comparison with models of similar size but also against larger MLLM frameworks (7B), and even achieves performance on par with 13B models.
|
17 |
|
18 |
The model is pretrained on LAION-2M and finetuned on Bunny-695K.
|
19 |
More details about this model can be found in [GitHub](https://github.com/BAAI-DCAI/Bunny).
|
@@ -22,7 +22,13 @@ More details about this model can be found in [GitHub](https://github.com/BAAI-D
|
|
22 |
|
23 |
# Quickstart
|
24 |
|
25 |
-
Here we show a code snippet to show you how to use the model with transformers
|
|
|
|
|
|
|
|
|
|
|
|
|
26 |
|
27 |
```python
|
28 |
import torch
|
@@ -69,12 +75,6 @@ output_ids = model.generate(
|
|
69 |
print(tokenizer.decode(output_ids[input_ids.shape[1]:], skip_special_tokens=True).strip())
|
70 |
```
|
71 |
|
72 |
-
Before running the snippet, you need to install the following dependencies:
|
73 |
-
|
74 |
-
```shell
|
75 |
-
pip install torch transformers accelerate pillow
|
76 |
-
```
|
77 |
-
|
78 |
# License
|
79 |
This project utilizes certain datasets and checkpoints that are subject to their respective original licenses. Users must comply with all terms and conditions of these original licenses.
|
80 |
The content of this project itself is licensed under the Apache license 2.0.
|
|
|
13 |
|
14 |
This is the merged weights of [bunny-phi-2-siglip-lora](https://huggingface.co/BAAI/bunny-phi-2-siglip-lora).
|
15 |
|
16 |
+
Bunny is a family of lightweight but powerful multimodal models. It offers multiple plug-and-play vision encoders, like EVA-CLIP, SigLIP and language backbones, including Phi-1.5, StableLM-2, Qwen1.5 and Phi-2. To compensate for the decrease in model size, we construct more informative training data by curated selection from a broader data source. Remarkably, our Bunny-v1.0-3B model built upon SigLIP and Phi-2 outperforms the state-of-the-art MLLMs, not only in comparison with models of similar size but also against larger MLLM frameworks (7B), and even achieves performance on par with 13B models.
|
17 |
|
18 |
The model is pretrained on LAION-2M and finetuned on Bunny-695K.
|
19 |
More details about this model can be found in [GitHub](https://github.com/BAAI-DCAI/Bunny).
|
|
|
22 |
|
23 |
# Quickstart
|
24 |
|
25 |
+
Here we show a code snippet to show you how to use the model with transformers.
|
26 |
+
|
27 |
+
Before running the snippet, you need to install the following dependencies:
|
28 |
+
|
29 |
+
```shell
|
30 |
+
pip install torch transformers accelerate pillow
|
31 |
+
```
|
32 |
|
33 |
```python
|
34 |
import torch
|
|
|
75 |
print(tokenizer.decode(output_ids[input_ids.shape[1]:], skip_special_tokens=True).strip())
|
76 |
```
|
77 |
|
|
|
|
|
|
|
|
|
|
|
|
|
78 |
# License
|
79 |
This project utilizes certain datasets and checkpoints that are subject to their respective original licenses. Users must comply with all terms and conditions of these original licenses.
|
80 |
The content of this project itself is licensed under the Apache license 2.0.
|