XCLiu commited on
Commit
9a2bf6f
1 Parent(s): ca941de

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +71 -3
README.md CHANGED
@@ -1,3 +1,71 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ license_name: deepseek
4
+ license_link: LICENSE
5
+ pipeline_tag: any-to-any
6
+ library_name: transformers
7
+ tags:
8
+ - muiltimodal
9
+ - text-to-image
10
+ - unified-model
11
+ ---
12
+
13
+
14
+ ## 1. Introduction
15
+
16
+ We present JanusFlow, a powerful framework that unifies image understanding and generation in a single model.
17
+ JanusFlow introduces a minimalist architecture that integrates autoregressive
18
+ language models with rectified flow, a state-of-the-art method in generative modeling. Our
19
+ key finding demonstrates that rectified flow can be straightforwardly trained within the large
20
+ language model framework, eliminating the need for complex architectural modifications.
21
+ To further improve the performance of our unified model, we adopt two key strategies: (i)
22
+ decoupling the understanding and generation encoders, and (ii) aligning their representations
23
+ during unified training. Extensive experiments show that JanusFlow achieves comparable or
24
+ superior performance to specialized models in their respective domains, while significantly
25
+ outperforming existing unified approaches across standard benchmarks. This work represents
26
+ a step toward more efficient and versatile vision-language models.
27
+
28
+ [JanusFlow: Harmonizing Autoregression and Rectified Flow for Unified Multimodal Understanding and Generation](TBD)
29
+
30
+ [**Github Repository**](https://github.com/deepseek-ai/Janus)
31
+
32
+ <div align="center">
33
+ <img alt="image" src="teaser.png" style="width:90%;">
34
+ </div>
35
+
36
+
37
+ ### 2. Model Summary
38
+
39
+ Janus is a unified understanding and generation MLLM, which decouples visual encoding for multimodal understanding and generation.
40
+ Janus is constructed based on the DeepSeek-LLM-1.3b-base which is trained on an approximate corpus of 500B text tokens.
41
+ For multimodal understanding, it uses the [SigLIP-L](https://huggingface.co/timm/ViT-L-16-SigLIP-384) as the vision encoder, which supports 384 x 384 image input. For image generation, Janus uses the tokenizer from [here](https://github.com/FoundationVision/LlamaGen) with a downsample rate of 16.
42
+
43
+ <div align="center">
44
+ <img alt="image" src="arch.jpg" style="width:90%;">
45
+ </div>
46
+
47
+ ## 3. Quick Start
48
+
49
+ Please refer to [**Github Repository**](https://github.com/deepseek-ai/Janus)
50
+
51
+
52
+ ## 4. License
53
+
54
+ This code repository is licensed under [the MIT License](https://github.com/deepseek-ai/DeepSeek-LLM/blob/HEAD/LICENSE-CODE). The use of Janus models is subject to [DeepSeek Model License](https://github.com/deepseek-ai/DeepSeek-LLM/blob/HEAD/LICENSE-MODEL).
55
+ ## 5. Citation
56
+
57
+ ```
58
+ @misc{wu2024janus,
59
+ title={Janus: Decoupling Visual Encoding for Unified Multimodal Understanding and Generation},
60
+ author={Chengyue Wu and Xiaokang Chen and Zhiyu Wu and Yiyang Ma and Xingchao Liu and Zizheng Pan and Wen Liu and Zhenda Xie and Xingkai Yu and Chong Ruan and Ping Luo},
61
+ year={2024},
62
+ eprint={2410.13848},
63
+ archivePrefix={arXiv},
64
+ primaryClass={cs.CV},
65
+ url={https://arxiv.org/abs/2410.13848},
66
+ }
67
+ ```
68
+
69
+ ## 6. Contact
70
+
71
+ If you have any questions, please raise an issue or contact us at [[email protected]](mailto:[email protected]).