Update README.md
Browse files
README.md
CHANGED
@@ -2,6 +2,10 @@
|
|
2 |
license: apache-2.0
|
3 |
---
|
4 |
|
|
|
|
|
|
|
|
|
5 |
<div align="left" style="line-height: 1;">
|
6 |
<a href="https://bagel-ai.org/" target="_blank" style="margin: 2px;">
|
7 |
<img alt="Homepage" src="https://img.shields.io/badge/BAGEL-Homepage-a468fe?color=a468fe&logoColor=white" style="display: inline-block; vertical-align: middle;"/>
|
@@ -9,18 +13,19 @@ license: apache-2.0
|
|
9 |
<a href="https://github.com/ByteDance-Seed/BAGEL/blob/main/BAGEL-Technical-Report.pdf" target="_blank" style="margin: 2px;">
|
10 |
<img alt="Technical Report" src="https://img.shields.io/badge/(upcoming)-Technical%20Report-brightgreen?logo=arxiv&logoColor=white" style="display: inline-block; vertical-align: middle;"/>
|
11 |
</a>
|
12 |
-
|
13 |
<a href="https://github.com/bytedance-seed/BAGEL" target="_blank" style="margin: 2px;">
|
14 |
-
<img alt="Github" src="https://img.shields.io/badge/
|
15 |
</a>
|
16 |
</div>
|
17 |
|
18 |
-
# 🥯 BAGEL • Unified Model for Multimodal Understanding and Generation
|
19 |
> We present **BAGEL**, an open‑source multimodal foundation model with 7B active parameters (14B total) trained on large‑scale interleaved multimodal data. BAGEL outperforms the current top‑tier open‑source VLMs like Qwen2.5-VL and InternVL-2.5 on standard multimodal understanding leaderboards, and delivers text‑to‑image quality that is competitive with strong specialist generators such as SD3.
|
20 |
Moreover, BAGEL demonstrates superior qualitative results in classical image‑editing scenarios than the leading open-source models. More importantly, it extends to free-form visual manipulation, multiview synthesis, and world navigation, capabilities that constitute "world-modeling" tasks beyond the scope of previous image-editing models.
|
21 |
Below is a showcase of BAGEL's qualitative performance.
|
22 |
|
23 |
|
|
|
|
|
|
|
24 |
|
25 |
|
26 |
## 📊 Benchmarks
|
@@ -40,11 +45,10 @@ Below is a showcase of BAGEL's qualitative performance.
|
|
40 |
### 3. Image Editing
|
41 |
| Benchmark | Step1X-Edit | Gemini-2-exp. | **BAGEL** | **BAGEL + CoT** |
|
42 |
| ------------------------ | ----------: | ------------: | --------: | --------------: |
|
43 |
-
| **GEdit-Bench-EN** (↑) |
|
44 |
| **IntelligentBench** (↑) | 14.9 | 57.6 | 44.0 | **55.3** |
|
45 |
|
46 |
## License
|
47 |
-
|
48 |
BAGEL is licensed under the Apache 2.0 license. It is finetuned from [Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct), and uses the [FLUX.1-schnell VAE model](https://huggingface.co/black-forest-labs/FLUX.1-schnell) and the [siglip-so400m-14-980-flash-attn2-navit](https://huggingface.co/HuggingFaceM4/siglip-so400m-14-980-flash-attn2-navit) model, all under Apache 2.0.
|
49 |
|
50 |
## ✍️ Citation
|
|
|
2 |
license: apache-2.0
|
3 |
---
|
4 |
|
5 |
+
|
6 |
+
|
7 |
+
# 🥯 BAGEL • Unified Model for Multimodal Understanding and Generation
|
8 |
+
|
9 |
<div align="left" style="line-height: 1;">
|
10 |
<a href="https://bagel-ai.org/" target="_blank" style="margin: 2px;">
|
11 |
<img alt="Homepage" src="https://img.shields.io/badge/BAGEL-Homepage-a468fe?color=a468fe&logoColor=white" style="display: inline-block; vertical-align: middle;"/>
|
|
|
13 |
<a href="https://github.com/ByteDance-Seed/BAGEL/blob/main/BAGEL-Technical-Report.pdf" target="_blank" style="margin: 2px;">
|
14 |
<img alt="Technical Report" src="https://img.shields.io/badge/(upcoming)-Technical%20Report-brightgreen?logo=arxiv&logoColor=white" style="display: inline-block; vertical-align: middle;"/>
|
15 |
</a>
|
|
|
16 |
<a href="https://github.com/bytedance-seed/BAGEL" target="_blank" style="margin: 2px;">
|
17 |
+
<img alt="Github" src="https://img.shields.io/badge/GitGub-Repo-536af5?color=536af5&logo=github" style="display: inline-block; vertical-align: middle;"/>
|
18 |
</a>
|
19 |
</div>
|
20 |
|
|
|
21 |
> We present **BAGEL**, an open‑source multimodal foundation model with 7B active parameters (14B total) trained on large‑scale interleaved multimodal data. BAGEL outperforms the current top‑tier open‑source VLMs like Qwen2.5-VL and InternVL-2.5 on standard multimodal understanding leaderboards, and delivers text‑to‑image quality that is competitive with strong specialist generators such as SD3.
|
22 |
Moreover, BAGEL demonstrates superior qualitative results in classical image‑editing scenarios than the leading open-source models. More importantly, it extends to free-form visual manipulation, multiview synthesis, and world navigation, capabilities that constitute "world-modeling" tasks beyond the scope of previous image-editing models.
|
23 |
Below is a showcase of BAGEL's qualitative performance.
|
24 |
|
25 |
|
26 |
+
This repository hosts the model weights for **BAGEL**.
|
27 |
+
For installation, usage instructions, and further documentation, please visit our [GitHub repository](https://github.com/bytedance-seed/BAGEL).
|
28 |
+
|
29 |
|
30 |
|
31 |
## 📊 Benchmarks
|
|
|
45 |
### 3. Image Editing
|
46 |
| Benchmark | Step1X-Edit | Gemini-2-exp. | **BAGEL** | **BAGEL + CoT** |
|
47 |
| ------------------------ | ----------: | ------------: | --------: | --------------: |
|
48 |
+
| **GEdit-Bench-EN** (↑) | **6.70** | 6.32 | 6.52 | – |
|
49 |
| **IntelligentBench** (↑) | 14.9 | 57.6 | 44.0 | **55.3** |
|
50 |
|
51 |
## License
|
|
|
52 |
BAGEL is licensed under the Apache 2.0 license. It is finetuned from [Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct), and uses the [FLUX.1-schnell VAE model](https://huggingface.co/black-forest-labs/FLUX.1-schnell) and the [siglip-so400m-14-980-flash-attn2-navit](https://huggingface.co/HuggingFaceM4/siglip-so400m-14-980-flash-attn2-navit) model, all under Apache 2.0.
|
53 |
|
54 |
## ✍️ Citation
|