Update README.md
Browse files
README.md
CHANGED
@@ -41,29 +41,6 @@ We investigate domain adaptation of MLLMs through post-training, focusing on dat
|
|
41 |
|
42 |
**Code**: [https://github.com/bigai-ai/QA-Synthesizer](https://github.com/bigai-ai/QA-Synthesizer)
|
43 |
|
44 |
-
## About
|
45 |
-
|
46 |
-
AdaMLLM is our latest effort to enhance task generalization of (M)LLMs by scaling synthetic supervised tasks based on unsupervised contexts.
|
47 |
-
|
48 |
-
<p align='left'>
|
49 |
-
<img src="https://cdn-uploads.huggingface.co/production/uploads/650801ced5578ef7e20b33d4/HUN3Cr66w_xpj5_c7QQaI.png" width="1000">
|
50 |
-
</p>
|
51 |
-
|
52 |
-
- [AdaptLLM](https://huggingface.co/papers/2309.09530)
|
53 |
-
We employ rule-based methods to extract tasks from domain-specific corpora, reformatting them into reading comprehension tasks for continued pre-training. Our 7B finance model outperforms domain-specific models of much larger scales, such as BloombergGPT-50B.
|
54 |
-
|
55 |
-
- [Instruction Pre-Training](https://huggingface.co/papers/2406.14491)
|
56 |
-
We develop a general-purpose instruction synthesizer which significantly increases task diversity for LM pre-training, outperforming vanilla pre-training in both general pre-training from scratch and domain-adaptive continual pre-training.
|
57 |
-
|
58 |
-
- AdaMLLM
|
59 |
-
We extend supervised task synthesis to multimodality, introducing a unified visual instruction synthesizer to extract instruction-response pairs from image-caption data. Our synthetic tasks outperform those generated by manual rules, GPT-4, and GPT-4V in improving domain-specific performance for MLLMs.
|
60 |
-
|
61 |
-
Looking ahead, we envision further broadening the scope of supervised task synthesis, efficiently enhancing the general capabilities of trained models.
|
62 |
-
|
63 |
-
## Contact
|
64 |
-
Daixuan Cheng: `[email protected]`
|
65 |
-
|
66 |
-
|
67 |
## Citation
|
68 |
If you find our work helpful, please cite us.
|
69 |
|
@@ -77,16 +54,6 @@ If you find our work helpful, please cite us.
|
|
77 |
}
|
78 |
```
|
79 |
|
80 |
-
[Instruction Pre-Training](https://huggingface.co/papers/2406.14491) (EMNLP 2024)
|
81 |
-
```bibtex
|
82 |
-
@article{instructPT,
|
83 |
-
title={Instruction Pre-Training: Language Models are Supervised Multitask Learners},
|
84 |
-
author={Cheng, Daixuan and Gu, Yuxian and Huang, Shaohan and Bi, Junyu and Huang, Minlie and Wei, Furu},
|
85 |
-
journal={arXiv preprint arXiv:2406.14491},
|
86 |
-
year={2024}
|
87 |
-
}
|
88 |
-
```
|
89 |
-
|
90 |
[Adapt LLM to Domains](https://huggingface.co/papers/2309.09530) (ICLR 2024)
|
91 |
```bibtex
|
92 |
@inproceedings{
|
|
|
41 |
|
42 |
**Code**: [https://github.com/bigai-ai/QA-Synthesizer](https://github.com/bigai-ai/QA-Synthesizer)
|
43 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
44 |
## Citation
|
45 |
If you find our work helpful, please cite us.
|
46 |
|
|
|
54 |
}
|
55 |
```
|
56 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
57 |
[Adapt LLM to Domains](https://huggingface.co/papers/2309.09530) (ICLR 2024)
|
58 |
```bibtex
|
59 |
@inproceedings{
|